-
Notifications
You must be signed in to change notification settings - Fork 122
[Bug] Context window is too large #323
Comments
When you instantiate |
Hi @igiloh-pinecone I think I'm running everything on defaults: llm = OpenAILLM() # Defaults to GPT3.5
chat_engine = ChatEngine(context_engine=context_engine, llm=llm) I see GPT3.5-turbo now has a context window of 16k tokens, I think in the past that was 4k but I could be wrong. Maybe that's what causing the issue? I can enter the values for the generated tokens, context tokens and max prompt tokens but I think canopy would keep that managed on its own when using defaults? |
hey @igiloh-pinecone any update on this by any chance? I keep running into the issue from time to time, the default value for the ChatEngine is 4096 already, so I don't think passing the same value as the default will make much difference? |
Hey @igiloh-pinecone I'm having the same issue of exceeding max tokens: I searched through issue and came across this open one. I went through the source code and I believe the issue is with Within
I'd would really appreciate it if this issue is fixed soon! Right now, I'm manually truncating my chat history to prevent this from happening. Thanks! |
Is this a new bug?
Current Behavior
I provide a window of context to the canopy chat, but it appears that it doesn't reduce that window sufficiently.
I'm using
chat_engine.chat(messages=messages, stream=False, namespace="namespace")
, are we expected to truncate the messages ourselves to a certain degree?The output of the error
Expected Behavior
I expected the knowledge base or context builder to reduce the history of messages to an acceptable window.
Steps To Reproduce
Relevant log output
Environment
Additional Context
No response
The text was updated successfully, but these errors were encountered: