Can only imagine how stressful their production engineers have it at the moment. The totally unprecedented overnight success bolstered by a massive hype train must have multiplied the demands on their infrastructure by several orders of magnitude. Very uncomfortable time for the rank engineers and all the pressure being applied by some very senior people and the expectation of customers.
Definitely a good reminder that itâs not just the capability of the model but of the infrastructure around it.
I donât think they are that concerned. Their goal is AGI as theyâve stated several times. If their goal was competing with OAI, probably wouldnât have released OS
He missed two points:
Measuring progress to AGI by the sudden drops of breakthroughs (in math, comp sci first)
AIâs first expert task should be the AI programmer, use the tool to build the tool
Lol what? Open source isnât âright behindâ, itâs literally LEADING the charge in most cases. Maybe do your homework before making such claims?
Linux runs most of the internet, powers 97% of smartphones (Android is Linux-based), dominates supercomputers (like, literally 100% of top 500), runs your smart TVs, powers most IoT devices, and even runs the Mars helicopters! Git revolutionized code collaboration to the point where even Microsoft (yeah, the Windows guys) had to embrace it and bought GitHub. The most cutting-edge AI research is happening in open source right now with projects like Stable Diffusion, LLaMA, and Whisper leading innovation. Kubernetes? Open source. Docker? Open source. PostgreSQL? Open source. The whole web runs on open-source tech like Apache, Nginx, and Node.js.
The whole âright behindâ thing is laughably wrong when literally every piece of modern tech infrastructure is built on open source foundations. Even âclosedâ companies like Meta and Google rely heavily on open source and have to contribute back to stay competitive.
And âdespite the walls put upâ? Please. Open source BREAKS walls. Thatâs literally the point. While big tech is busy building their walled gardens, open source communities are out there sharing knowledge and pushing boundaries.
Context matters.
This topic is about language models and in this category it is perfectly reasonably to claim that the proprietary models are still in the lead.
Deepseek R1 is a super interesting release for the open source community but it has nothing to do with other types of software and AI models you mentioned.
Actually, you missed the context. I was responding to RonaldGRuckus claiming âOpen source has constantly shown that itâs right behind despite the walls put upâ which is a general statement about open source, not specifically about LLMs.
But since you brought up language models even there your claim about proprietary models leading is debatable. Have you seen the recent benchmarks? Deepseek R1 outperforms GPT-4o in several key areas including coding, math, and reasoning tasks.
The open source community has gone from âno viable modelsâ to âcompetitive with state-of-the-artâ in less than a year. Thatâs not âbeing behindâ, thatâs catching up and overtaking at an unprecedented pace.
And again to respond to @anon10827405 where he mention âbeing able to see reasoning tokens is hugeâ.
Tokens is basically a chains of random characters, not readable text. So first, what you see is generated text, not token
And Deepseek actually show the reasoning text on their own client for showcase. But using API, it literally takes one line of code to hide the reasoning which is already neatly packaged in [reasoning] tags. Thatâs the whole point of structured output.
You canât compare model if you donât use models with API. That it.
And donât forget that OpenAI have been existing for the last 10 years while Deepseek was founded 18 months ago.
Iâm not sure what youâre arguing here. Itâs quite common for companies to take an open source project (or many), repackage it/them and then try to profit off it by making sure what theyâve made canât be surpassed.
This is what I was referring to, bring on an OpenAI forum.
Iâm not arguing that open source is useless. It truly is fundamental to a lot of the world
Itâs a good thing we can map it out
Also, R1 is not better than SOTA OpenAI models. You really canât be serious by comparing R1 to gpt4o.
Itâs very strange, youâre making it seem like Iâm bashing open source when all Iâve said is that despite the barriers put up, like hiding the reasoning tokens, open source has still managed to keep up with it
CODEI/O: Condensing Reasoning Patterns via Code Input-Output Prediction:
DeepSeek AI Introduces CODEI/O: A Novel Approach that Transforms Code-based Reasoning Patterns into Natural Language Formats to Enhance LLMsâ Reasoning Capabilities