I have searched for a long while and got 2 ways to deal with this situations.
-
Splitting the document into multiple chunks and embed in MapReduce or Redefine way (which could not stuff the whole document into a message content window size) Summarization | 🦜️🔗 Langchain
-
Using the tools such as ChatGPT Prompt Splitter to split your document into multiple parts and send them by multiple messages. follow the blog article. Ways to automate breaking a large piece of input into chunks that fit in the 4096 token constraint? - API - OpenAI Developer Forum
While I have no idea which is the better. according to my testing. I thought the Map-Reduce way would generate a better result. As I know the method 2 seems like the Refine way. Their workflows are similar