-
Notifications
You must be signed in to change notification settings - Fork 14
Open
Labels
enhancementNew feature or requestNew feature or request
Description
Change DEFAULT_LLM_MAX_TOKENS from 128000 to 8192 for reasoning (JSON response is always under 500 tokens). Add a separate TRANSCRIPT_MAX_TOKENS=128000 for transcript analysis if needed later.
Files to modify:
- src/amplihack/fleet/_constants.py
- src/amplihack/fleet/_backends.py
Requirements:
- Change DEFAULT_LLM_MAX_TOKENS = 8192 in _constants.py
- Add TRANSCRIPT_MAX_TOKENS = 128000 in _constants.py
- Update _backends.py to use DEFAULT_LLM_MAX_TOKENS for reasoning calls
- Update _backends.py to use TRANSCRIPT_MAX_TOKENS for transcript analysis calls
- Add docstrings explaining the distinction
- All 918+ fleet tests must pass
- Commit changes on this branch
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
enhancementNew feature or requestNew feature or request