This post is meant for discussion and sharing of ideas. I just felt compelled to pushback on the tidal wave of party poopers writing papers on LLMs suggesting the fun is over. 2024 is still the ground floor.
I am a trained scientist. My expertise is in research on humans not AI (although I’m now doing both). I just have to post here because I’m disappointed with the level of LLM criticisms being published.
As a researcher I understand the mindset of scientists publishing on LLMs and I find them annoying these days. If you are like me, and find these critiques annoying let me help you understand why. Forgive me for using clear words when really, to be accurate I shouldn’t sound so definitive I no longer do lab research. I conduct and apply research findings in the real world and have lost the interest in taking like I don’t know anything (that is how scientists prefer to speak).
Why are the AI researchers being so annoying:
(1) Scientists disprove the last thoughts. You only publish what adds to the knowledge and therefore it must prove something new / disprove prior research. Every new paper is basically saying ‘ha, ha, you were wrong.’ Annoying.
(2) Scientists must run controlled experiments to prove anything at all and thus the best research is artificially constrained and does not mimic reality. This means they find strong evidence that when a human has 10 arms they cannot play the piano and therefore more arms doesn’t help you play the piano. Never happens in real life buddy: annoying.
(3) Only scientists in the same subfield can properly read and understand scientific papers written by a single set of authors which leads to a lot of misunderstandings. Scientists write without proper explanation or interest in explaining themselves to anyone who doesn’t already understand them: annoying.
All this to say that I am getting a little annoyed at all these inflammatory journal titles about LLMs.
For example, highly constrained unrealistic lab experiments on LLM training using synthetic data seem to underperform, according to the scientists.
I call BS.
If the scientists can’t figure out how to generate useful synthetic data and LLMs that can work with that data these scientists shouldn’t be publishing on the impact of synthetic data.
I am going to sound definitive again: there is a way to generate and consume synthetic data in order to train superior LLMs. Of course there is. It’s obvious. LLMs work like playing ping pong. The LLM hits back what you serve to it.
Again, just because the method used in one study, or even most studies show this quality gap - Does - Not - Mean - there is no viable method to get from point A to point B.
Always keep in mind that the collective wisdom of science at any given moment is wrong, or, scientific research would cease. So as long as we need more scientific research for any topic, the implication is we have been wrong for a long time and are collectively working to fix that.