Hello, i have a problem, after a few messages with my chat i have an errot:
error_code=context_length_exceeded error_message=“This model’s maximum context length is 8192 tokens. However, your messages resulted in 9066 tokens. Please reduce the length of the messages.” error_param=messages error_type=invalid_request_error message=‘OpenAI API error received’ stream_error=False
my main Chain looks like this:
chain = RetrievalQAWithSourcesChain.from_chain_type(
llm=llm,
chain_type="stuff",
retriever=retriever_prawo,
reduce_k_below_max_tokens=True,
chain_type_kwargs={
"verbose": True,
"prompt": prompt,
"memory": ConversationBufferMemory(
memory_key='history',
input_key='question'),
}
but reduce_k_below_max_tokens=True is not helping, also I’ve tried to use later in the code chain.max_tokens_limit = 8000, bu this also not working. Im using ChainLit to make a chat.
How can I prevent my chat from getting this error?