Accuracy and token limitation

Heyy. Can some one help me with this. I wanna summarize and categorize the call transcriptions of more than 1hour calls via llm’s. I’ve tried several methods by sending it to gpt 3.5 and gpt 4 via chatcompletions. But the token limit for one api call is 16k for 3.5 turbo and 8k for 4. But for an 1hr convo the tokens is like 25 to 30 k. I wont be able to run it via openai apis (exclude GPT-4-32K). Is there any way to run it on foundation models with the ability of GPT-4’s accuracy ? I’m more concerned about the accuracy and quality of the classification and summarization and I need the out in json format as it has to be pasted to sheets automatically. Please suggest me some methods.