mirror of
https://github.com/khoaliber/khoj.git
synced 2026-03-07 21:29:13 +00:00
Reduce max reasoning tokens for gemini models
A high reasoning tokens does not seem to help for standard Khoj use cases. And hopefully reducing it may avoid repetition loops by model.
This commit is contained in:
@@ -46,7 +46,7 @@ gemini_clients: Dict[str, genai.Client] = {}
|
|||||||
# This avoids premature response termination.
|
# This avoids premature response termination.
|
||||||
MAX_OUTPUT_TOKENS_FOR_REASONING_GEMINI = 20000
|
MAX_OUTPUT_TOKENS_FOR_REASONING_GEMINI = 20000
|
||||||
MAX_OUTPUT_TOKENS_FOR_STANDARD_GEMINI = 8000
|
MAX_OUTPUT_TOKENS_FOR_STANDARD_GEMINI = 8000
|
||||||
MAX_REASONING_TOKENS_GEMINI = 10000
|
MAX_REASONING_TOKENS_GEMINI = 512
|
||||||
|
|
||||||
SAFETY_SETTINGS = [
|
SAFETY_SETTINGS = [
|
||||||
gtypes.SafetySetting(
|
gtypes.SafetySetting(
|
||||||
|
|||||||
Reference in New Issue
Block a user