From 386a17371d445d5f0d1398fad76495d8625a1719 Mon Sep 17 00:00:00 2001 From: Debanjum Date: Tue, 26 Aug 2025 17:14:44 -0700 Subject: [PATCH] Fix identifying deepseek r1 model to process its thinking tokens --- src/khoj/processor/conversation/openai/utils.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/khoj/processor/conversation/openai/utils.py b/src/khoj/processor/conversation/openai/utils.py index 8b00cda8..5f408844 100644 --- a/src/khoj/processor/conversation/openai/utils.py +++ b/src/khoj/processor/conversation/openai/utils.py @@ -325,7 +325,7 @@ async def chat_completion_with_backoff( elif ( model_name.startswith("deepseek-chat") or model_name.startswith("deepseek-reasoner") - or "deepseek-r1" in model_name + or "deepseek-r1" in model_name.lower() ): # Official Deepseek models and some inference APIs like vLLM return structured thinking output. # Others like DeepInfra return it in response stream.