maybe guided thinking might be a better term.
but the temperature in the ChatGPT configuration is too high for it to be super reliable.
I don’t think you can wrangle chatgpt to read through 1600 pages of anything. You’d need to figure out a way to get it to recall the n most likely candidates and filter. Multi-step approaches will unfortunately yield compounding errors with the way the system is set up at the moment.
You might want to check out this thread: Using gpt-4 API to Semantically Chunk Documents
they discuss how to improve recall, and jr.2509’s post (Using gpt-4 API to Semantically Chunk Documents - #145 by jr.2509) deals specifically with regulatory documents. (they use embeddings, not custom gpts)