From aebdd174ecf031a6b7ea7a689dd738ae59bb7d7f Mon Sep 17 00:00:00 2001 From: Debanjum Date: Sat, 25 Jan 2025 18:32:02 +0700 Subject: [PATCH 1/2] Make title of PeopleAlsoAsk section of online results optional --- src/khoj/database/models/__init__.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/khoj/database/models/__init__.py b/src/khoj/database/models/__init__.py index 7ece32f8..1a3697b5 100644 --- a/src/khoj/database/models/__init__.py +++ b/src/khoj/database/models/__init__.py @@ -60,7 +60,7 @@ class PeopleAlsoAsk(PydanticBaseModel): link: Optional[str] = None question: Optional[str] = None snippet: Optional[str] = None - title: str + title: Optional[str] = None class KnowledgeGraph(PydanticBaseModel): From af49375884fc8bd59765caa30a9b62f2595fca42 Mon Sep 17 00:00:00 2001 From: Debanjum Date: Sat, 25 Jan 2025 18:34:30 +0700 Subject: [PATCH 2/2] Only show confusing fallback tokenizer used logs in high verbosity mode --- src/khoj/processor/conversation/utils.py | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/src/khoj/processor/conversation/utils.py b/src/khoj/processor/conversation/utils.py index 75a76918..1ab06fae 100644 --- a/src/khoj/processor/conversation/utils.py +++ b/src/khoj/processor/conversation/utils.py @@ -534,9 +534,10 @@ def truncate_messages( encoder = download_model(model_name).tokenizer() except: encoder = tiktoken.encoding_for_model(default_tokenizer) - logger.debug( - f"Fallback to default chat model tokenizer: {default_tokenizer}.\nConfigure tokenizer for model: {model_name} in Khoj settings to improve context stuffing." - ) + if state.verbose > 2: + logger.debug( + f"Fallback to default chat model tokenizer: {default_tokenizer}.\nConfigure tokenizer for model: {model_name} in Khoj settings to improve context stuffing." + ) # Extract system message from messages system_message = None