Hi Community, long time user and first time poster - please be kind
I’m attempting to write a prompt which will read multiple answers to a question from different sources then provide a definitive comprehensive answer. I’d like to use this approach so that I can provide GTP3 with human picked reliable answers allowing GTP3 to read knowledge which the model hasn’t been trained on then summarise and provide an answer.
Example of the type of question and knowledge I’m providing:
Question
What colour is the sky?
Source 1
The sky is blue.
Source 2
I see a blue sky during the day.
Source 3
We can see sky as blue because it has the smallest wavelength when scattering light.
Source 4
Depends, at night the sky is black.
Source 5
Usually blue, but at sunset the sky turns warm and glowing shades of gold, pink and orange.
Source 6
I don’t have any evidence but I’m sure I saw the sky turn green and yellow once.
Example of what I’m expecting GTP3 to return based on reading these 6 sources:
Question:
What colour is the sky?
Expected typical answer:
During the day the sky is blue due to being the smallest wavelength when scattering light. When the sun sets the sky turns gold, pink and orange, then at night the sky is black. Some sources believe the sky is green and yellow however this is unreliable.
So far I’ve used a TLDR prompt, the Q&A and 2nd grader templates - usually just getting back “The sky is blue” regardless of the engine used and any combination of temperature and penalties - no mention of other colour information mentioned in the statements. Is this coming from GTP3 training data or reading the sources?
I’ve also tried implementing the answer endpoint by uploading each answer then asking the question - however I’m mainly getting back verbatim statements when I was expecting some re-writing and summarisation to occur - tweaking the temperature starts getting too creative and other random non-related information starts creeping in which is usually completely wrong.
The main benefit of this summarise sources approach is that I can provide citations of where the information came from to combat misinformation, allowing greater trust with the reader. GTP3 is being used in this instance as a summarise system (“GTP3 - please answer this question by reading the following statements and ignore what you already know”), not relying on it’s own knowledge which might be out of date or incorrect.
Any suggestions or feedback on the correct prompt, approach or methodology would be appreciated!
Rob