I have an agent app that helps users with questions about exercise routines, health, etc. I have limited the agent’s responses to only use answers from a specific list of sources that are published after the year 2020. I’ve followed the highly recommended paper “Principled Instructions Are All You Need for
Questioning LLaMA-1/2, GPT-3.5/4” which virtually eliminates hallucinations at the start of the chat. The problem I am having is that about 50% of the time, the model (I am using GPT-4 as it follows instructions better than GPT 4 Turbo, 4o, etc) will give some random year (ie, 2020) for publication when in fact the article was published prior to 2020 (ie, 2005).
Example of the agent’s system prompt:
### Task & Resources
- **Mandatory Sources:** Utilize only the latest peer-reviewed publications (2019-present) from....
and further down…
### Source Publication Guidelines
**Penalties:**
Immediate penalties apply for:
- Citing sources published before 2019.
- Fabrication of sources used for answers.
Anyone got any tips or guidance on how to get the model to not hallucinate the publication dates?