mirror of
https://github.com/khoaliber/khoj.git
synced 2026-03-04 13:20:17 +00:00
Reduce max reasoning tokens for gemini models
A high reasoning tokens does not seem to help for standard Khoj use cases. And hopefully reducing it may avoid repetition loops by model.
This commit is contained in:
@@ -46,7 +46,7 @@ gemini_clients: Dict[str, genai.Client] = {}
|
||||
# This avoids premature response termination.
|
||||
MAX_OUTPUT_TOKENS_FOR_REASONING_GEMINI = 20000
|
||||
MAX_OUTPUT_TOKENS_FOR_STANDARD_GEMINI = 8000
|
||||
MAX_REASONING_TOKENS_GEMINI = 10000
|
||||
MAX_REASONING_TOKENS_GEMINI = 512
|
||||
|
||||
SAFETY_SETTINGS = [
|
||||
gtypes.SafetySetting(
|
||||
|
||||
Reference in New Issue
Block a user