diff --git a/src/khoj/processor/conversation/gpt.py b/src/khoj/processor/conversation/gpt.py index e1b54303..2ee93053 100644 --- a/src/khoj/processor/conversation/gpt.py +++ b/src/khoj/processor/conversation/gpt.py @@ -175,6 +175,8 @@ def converse( conversation_log, model, ) + truncated_messages = "\n".join({f"{message.content[:40]}..." for message in messages}) + logger.debug(f"Conversation Context for GPT: {truncated_messages}") # Get Response from GPT return chat_completion_with_backoff( diff --git a/src/khoj/processor/conversation/utils.py b/src/khoj/processor/conversation/utils.py index 4c3e7619..99084bf0 100644 --- a/src/khoj/processor/conversation/utils.py +++ b/src/khoj/processor/conversation/utils.py @@ -48,7 +48,7 @@ class ThreadedGenerator: item = self.queue.get() if item is StopIteration: time_to_response = perf_counter() - self.start_time - logger.info(f"Time to stream full response: {time_to_response:.3f}") + logger.info(f"Chat streaming took: {time_to_response:.3f} seconds") if self.completion_func: # The completion func effective acts as a callback. # It adds the aggregated response to the conversation history. It's constructed in api.py.