Hello @dara,
First of all, thank you for your input! I’d like to hear your reasoning since this is exactly why I created this post. I see now that I should have gone into more detail regarding a few key points I made to address the counter-arguments you gave me. As a side-note, I don’t believe this post does any injustice to the OpenAI crew. This post is meant to be a learning experience built upon and discussed by whoever wishes to provide their insights and perspective regarding GPT-3 and what GPT-3 is capable of.
Clarification:
Since I didn’t communicate this as clearly as I would’ve liked, allow me to further explain why I used the training data I did in order to conclude why I believe GPT-3 is capable of understanding math to any degree. My analysis aimed to conclude that GPT-3 has the core function necessary to understand a math problem and solve it with correct understanding. I designed the practice problems to see if GPT-3 was able to learn one area of mathematics. Once the language model demonstrates that, then theoretically what’s to stop it from learning more mathematical concepts?
During my analysis, I did have limitations but I did my best to work with what I had. Please review the limitations I faced when performing my analysis, as this may help understand why the training data wasn’t the best, but still valuable I believe:
-
I saw that GPT-3 had a tough time understanding numbers, however just like natural language, there is a pattern in the way numbers are expressed (word, numerical, or any other form). I’m sure that with the right training data, GPT-3 can accurately and confidently quantify the numbers expressed to it. I did look into how I could teach GPT-3 to quantify numbers properly, however, I found that I didn’t have as good of an understanding of how GPT-3 best understands input given to it that I do now and found myself re-constructing my training data over and over. On top of that, I realized that I would need way more training data as well. I am sure there’s a way to tackle it, but I do have other projects I’d like to work on with GPT-3 and I’d like to explore other topics to be as productive as possible.
-
Going forward, let’s assume that we successfully trained GPT-3 on how to properly quantify any number given to it. We then need to see if GPT-3 can pick up and understand basic arithmetic and algebra. We do see that GPT-3 was able to learn how to solve the problem. However, since I didn’t train GPT-3 to quantify numbers confidently, as @boris pointed out, the language model did have a harder time solving the equations since it wasn’t confident with the numbers we were giving to it. The screenshots that provided the Completion with the full spectrum probability graph show that the least confident parts of the input data were the numbers. Now, I’m unsure how much of this was due to the language model learning a new change in the pattern with each sample problem, but knowing that GPT-3 does have trouble quantifying numbers, I’m sure that affected the confidence level.
My Main Point:
We see that when GPT-3 fails to understand how big numbers are supposed to be quantified but we still introduce a higher-level concept like arithmetic prematurely, we automatically inherit the problems that the lower-level concept had. Sure, you can feed it more training data to account for the loss, but I suspect it’d take more training data to accurately train the model.
Given well-defined training data, i.e. a series of math problems, GPT-3 can understand the structure of the problem such that it correctly identifies how every part of the problem will affect the final answer. With that, any similar problem presented to GPT-3 will be understood already, since the model has trained enough to know how every part interacts with each other to produce the correct answer!
By ensuring that lower-level concepts are properly understood by the model, we prevent further errors when training GPT-3 with higher-level concepts. Yes, I am aware that I could’ve trained the model better but that would’ve taken more time and resources to perfect the model and I was trying to show y’all with my data how I could still get the main point across. Since we were able to get the model to briefly demonstrate a higher-level concept such as basic algebra so as long as we didn’t use big numbers, it shows that if we had perfected the lower-level concept that we would’ve been able to see the model do basic algebra with much higher numbers more confidently.
And also I’m not saying that there’s only one way of training GPT-3, I’m sure there are several other ways that would allow for the model to learn a concept! The key point is that as long as the model in learning correctly, then GPT-3 should be able to continuously build upon previous concepts to learn more advanced topics. My examples were meant to conceptualize how GPT-3 could be taught in order to achieve an understanding of higher-level mathematics correctly.
Summary:
To sum up what I’m saying, here is the rundown of why I believe GPT-3 can be good at math:
-
If we look past the poor confidence when I didn’t teach GPT-3 to properly quantify numbers, we see that GPT-3 was pretty good with answering basic algebra questions with the numbers it could quantify. I’m sure had we successfully taught GPT-3 to quantify numbers first, it would be much easier for us to teach higher-level mathematical concepts.
-
With the right training data, I’m sure we can teach GPT-3 how to quantify numbers correctly so that we can allow it to better learn more advanced topics.
-
If it can learn one mathematical concept, why can’t it learn another one? Theoretically, we should be able to build upon multiple concepts given that we have trained them correctly.
Questions I Now Have:
-
What are some things that I did or didn’t do that may have hindered its learning process when answering the math questions I gave it? Please exclude the example of teaching GPT-3 to quantify numbers correctly.
-
Since I only fed GPT-3 the question “If x is equivalent to [first number], what is [second number] plus x?”, I found that giving GPT-3 the same question but worded differently slightly threw off GPT-3’s overall confidence when answering the question. When training GPT-3 to get good at a concept, would feeding training data with the same problems simply worded differently help the language model separate the concept from the structure of the question itself?
-
In addition to the projects I have planned, are there any topics that you’d like to see explored?
-
Are there any topics that you found hard or almost impossible to teach GPT-3 that you’d like to share?
Please feel free to share your thoughts and ideas, fellow community members!