How you motivate LLM Agent to read all the data before answer and to work only on the database and answer only based on it?

maybe guided thinking might be a better term.

but the temperature in the ChatGPT configuration is too high for it to be super reliable.

I don’t think you can wrangle chatgpt to read through 1600 pages of anything. You’d need to figure out a way to get it to recall the n most likely candidates and filter. Multi-step approaches will unfortunately yield compounding errors with the way the system is set up at the moment.

You might want to check out this thread: Using gpt-4 API to Semantically Chunk Documents

they discuss how to improve recall, and jr.2509’s post (Using gpt-4 API to Semantically Chunk Documents - #145 by jr.2509) deals specifically with regulatory documents. (they use embeddings, not custom gpts)