this is my first post on this forum. I need a bit of advice on the approach of how to create accurate chatbot trained on our data. My current approach is:
export pages / files from our docs
store them as files in open ai and attach to a vector store
create assistant with a file_search function. Model 4.1 nano / mini
test, test, test
result: lot of halucinations e.g. I am asking what is our latest release and it’s features. Releases are stored as separate files / pages with detail information for each LTS release.
I am not an AI / ML expert so bit of guidance on how to approach this problem would be helpful. Thank you,
j.
I would recommend playing with temperature and top_p, specifically lowering those in small increments, to see if this improves your results.
Also, prompt engineering is an overlooked aspect of LLM integration and may actually be a major factor of these issues. OpenAI has a good resource on prompting 4.1: GPT-4.1 Prompting Guide
How is this any different than OpenAI’s vector store?
Personally for me having full control over the whole system and each of the elements in it allows to get way better precision and exact fit into my apps. More work in the beginning but far better results and less work in the long run, especially if planning to scale/reuse.