From f76af869f119107b208c530acad0fced6c733641 Mon Sep 17 00:00:00 2001 From: Debanjum Singh Solanky Date: Fri, 28 Jul 2023 19:14:04 -0700 Subject: [PATCH] Do not log the gpt4all chat response stream in khoj backend Stream floods stdout and does not provide useful info to user --- src/khoj/processor/conversation/gpt4all/chat_model.py | 1 - 1 file changed, 1 deletion(-) diff --git a/src/khoj/processor/conversation/gpt4all/chat_model.py b/src/khoj/processor/conversation/gpt4all/chat_model.py index 3cd6c076..9089a2ec 100644 --- a/src/khoj/processor/conversation/gpt4all/chat_model.py +++ b/src/khoj/processor/conversation/gpt4all/chat_model.py @@ -163,6 +163,5 @@ def llm_thread(g, messages: List[ChatMessage], model: GPT4All): prompted_message = templated_system_message + chat_history + templated_user_message response_iterator = model.generate(prompted_message, streaming=True, max_tokens=2000) for response in response_iterator: - logger.info(response) g.send(response) g.close()