I consider myself quite novice, but I’ve been using the API for many months now and made some progress with text generation.
Context: I use open-ai’s text generation to write introductory paragraphs based off a fairly small body of text content. Imagine somebody gives me the basic content of a 500-word article that was missing an introduction, and I am using ChatGPT API to generate the introductory paragraph. I’m doing stuff like that.
I have found out through many iterations of use that I have to limit my question scope or desired shaping of the AI response when sending input. In other words, I cannot get my ideal response unless I break up the input/output messaging into smaller concerns.
Can I assume this is normal and the conclusion most AI API users arrive at?
As a result there is a lot of round tripping to the API going on. For instance, based on my desire for an introductory paragraph above I will:
- Build list of blacklisted words and create tokens array for use with logit_bias (used throughout messages)
- Set system tone and expectations
- Send 500-word article and Ask for an outline
- Ask for Introductory paragraph based on outline
- Ask for grammatical fixes
- Ask for SEO optimization
- Ask for tone fixes
The above order of operations is a skeletal representation. There’s a lot more going on with each ask of the API and some cleaning up on my part.
The results I have been getting have been getting significantly better as I have broken things up into multiple messages; each one attempting to solve a smaller part of my overall goal.
The only downside is the growing latency. This is only for my personal use, but still…if I add a few more input/output cycles I can see the wait time getting increasingly slower, exceeding 1+ minute or so.
Am I approaching this the right way?
As a beginner with some experience; now what?
What is the next logical step or direction to take my learning/project in?
Thank you kindly.