From 7071d081e97984f0afbe59fbe662d7808db92d6b Mon Sep 17 00:00:00 2001 From: Debanjum Singh Solanky Date: Mon, 16 Jan 2023 22:55:36 -0300 Subject: [PATCH] Increase max_tokens returned by GPT summarizer. Remove default params --- src/processor/conversation/gpt.py | 10 +--------- 1 file changed, 1 insertion(+), 9 deletions(-) diff --git a/src/processor/conversation/gpt.py b/src/processor/conversation/gpt.py index c67739d1..8a813a1c 100644 --- a/src/processor/conversation/gpt.py +++ b/src/processor/conversation/gpt.py @@ -10,7 +10,7 @@ import openai from src.utils.constants import empty_escape_sequences -def summarize(text, summary_type, model, user_query=None, api_key=None, temperature=0.5, max_tokens=100): +def summarize(text, summary_type, model, user_query=None, api_key=None, temperature=0.5, max_tokens=200): """ Summarize user input using OpenAI's GPT """ @@ -39,9 +39,7 @@ Summarize the notes in second person perspective:''' model=model, temperature=temperature, max_tokens=max_tokens, - top_p=1, frequency_penalty=0.2, - presence_penalty=0, stop="\"\"\"") # Extract, Clean Message from GPT's Response @@ -88,9 +86,7 @@ A:{ "search-type": "notes" }''' model=model, temperature=temperature, max_tokens=max_tokens, - top_p=1, frequency_penalty=0.2, - presence_penalty=0, stop=["\n"]) # Extract, Clean Message from GPT's Response @@ -159,9 +155,7 @@ A: { "intent": {"type": "generate", "activity": "chat", "query": "Can you dance model=model, temperature=temperature, max_tokens=max_tokens, - top_p=1, frequency_penalty=0.2, - presence_penalty=0, stop=["\n"]) # Extract, Clean Message from GPT's Response @@ -193,8 +187,6 @@ AI: Hi, I am an AI conversational companion created by OpenAI. How can I help yo model=model, temperature=temperature, max_tokens=max_tokens, - top_p=1, - frequency_penalty=0, presence_penalty=0.6, stop=["\n", "Human:", "AI:"])