I am trying to figure out the best route to be able to load a long text document (think a 60 page lease or medical paper). Then i want to ask questions about the text. Is this fine tuning? Seems like fine tuning would only work if i had sample responses.
I have a pipeline written down that creates embeddings for subdocuments, uses semantic search to find the relevant subdocument, and then uses text-davinci-003 to rephrase the subdocument for a specific audience. Does that seem to be helpful for your use-case?
I have replied in a private chat message with details of a workflow you can use.
We have also resolved the breaking down of documents for fine tuning, hallucination issue, and providing accurate citations, and video/html in responses. It is a mix of embedding and fine-tuning.
We are not quite ready for a public announcement yet and are proving limited BETA access on a case-by-case basis (Depends on use case)
I will share more to the community in coming days.