Gpt-4 32k vs GPT-4 Turbo api + Legal advice on using

the max i was able to get an output was 1k tokens which is a bit jarring since if i want proper Q/A i would have to run each chunk individually that will increase the input token cost 3-4x , any way to force more detailed answers since in my local llm’s i could always get it to generate around max output capacity