So if I’m understanding this correctly, you’re trying to get an LLM to produce a fixed set of values, of which either are a 1 or a 0 ?
So you want all its outputs to strictly be [1,1,1,0,0,0] , or [1,0,1,0,1] for example, right?
Hmm… Are you using LangChain for this by any chance?
This doesn’t look like a prompt possible in a single-shot style. In theory, it should be possible for GPT to identify something within a sentence or phrase with a binary yes/no answer (1,0), but in terms of handling all of that in one prompt, making those determinations, and placing the correct values in order in a finite array like that correctly each time is not going to work.
Also, what is “it” here? LLMs need an almost superfluous level of clarity to operate optimally under their constraints. From your prompt, it’s not clear what “it” is (although I also understand this may have been a prompt in a different language at first, which could also cause communication breakdowns).
On a high-level reconfiguration of this setup, you would probably have to coax it into producing strictly a yes/no single word answer for each subconcept iteratively (meaning, you will not be able to feed it 5 subconcepts at once), and then go in post-processing to translate yes → 1, and no → 0.
This looks like something LangChain would actually work well for now that I see your prompt. However, note that it’s still going to be relatively complex. You will not be able to get away with 0 programming for this, if you want that exact formatted array.
These models are built to generate natural language. They can analyze patterns well, especially in language, so the use case can still work in theory. However, it is not built to produce numerical outputs in that fashion like that. You can either choose to work with yes/no in place of 1/0 and change the prompting approach, or create a program/some scripting work that will turn the model’s outputs into the format you want, but it will still need to be prompted differently and likely with multiple shots if you want to maximize clarity and verifiability.
Also, even if it did produce that format correctly, do not assume its determination was entirely accurate. It would need to be cross-checked and verified by a human, which, depending on your corpus size, can get exponentially high, and turn into a lot of work. As an unpaid intern handling corpus labeling at my uni at the time, I get that labeling corpus data is a pain.
Those are the limitations and capabilities right now from my experience. Once you decide on which approach you’d prefer, we’d love to continue assisting you. Otherwise, you will not be able to generate both accurate results and consistent formats through single-shot prompting alone in the way you provided.