From 0e63a9037750d16c0bb14bbab16bfadfc0aeeb27 Mon Sep 17 00:00:00 2001 From: Saba Date: Sun, 4 Jun 2023 20:25:37 -0700 Subject: [PATCH] Fix the mechanism to retrieve the message content --- src/khoj/processor/conversation/utils.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/src/khoj/processor/conversation/utils.py b/src/khoj/processor/conversation/utils.py index 2df1f24f..5bccdf79 100644 --- a/src/khoj/processor/conversation/utils.py +++ b/src/khoj/processor/conversation/utils.py @@ -98,10 +98,10 @@ def generate_chatml_messages_with_context( # Truncate oldest messages from conversation history until under max supported prompt size by model encoder = tiktoken.encoding_for_model(model_name) - tokens = sum([len(encoder.encode(content)) for message in messages for content in message.content]) + tokens = sum([len(encoder.encode(message.content)) for message in messages]) while tokens > max_prompt_size[model_name] and len(messages) > 1: messages.pop() - tokens = sum([len(encoder.encode(content)) for message in messages for content in message.content]) + tokens = sum([len(encoder.encode(message.content)) for message in messages]) # Truncate last message if still over max supported prompt size by model if tokens > max_prompt_size[model_name]: