That’s what it was. Even the free, open source models are vastly superior to the best of the best from just a year ago.
People got into their heads that AI is shit when it was shit and decided at that moment that it was going to be stuck in that state forever. They forget that AI is just software and software usually gets better over time. Especially open source software which is what all the big AI vendors are building their tools on top of.
I tried one for the first time yesterday. It was mediocre at best. Certainly not production code. It would take just as much effort to refine it as it would to just write it in the first place.
If you read AI critics, you will see people presenting solid financial evidence of the failure of AI companies to do what they promised. Remember Sam Altman promised AGI in 2025? I certainly do, and now so do you.
Do you have any concrete evidence that this financial flop will turn around before it runs out of money?
Assume all the big AI firms die: Anthropic, OpenAI, Microsoft, Google, and Meta. Poof! They’re gone!
Here would be my reaction: “So anyway… have you tried GLM-7? It’s amazing! Also, there’s a new workflow in ComfyUI I’ve been using that works great to generate…”
Generative AI is here to stay. You don’t need a trillion dollars worth of data centers for progress to continue. That’s just billionaires living in an AGI fantasy land.
I just added up how much it would cost (in theory—assuming everything is in-stock and ready to ship) to build out a data center capable of training something like qwen3.5:122b from scratch in a few months: $66M. That’s how much it would cost for 128 Nvidia B200 nodes (they have 8 GPUs each), infiniband networking, all-flash storage (SSDs), and 20 racks (the hardware).
If OpenAI went bankrupt, that would result in a glut of such hardware which would flood the market, so the cost would probably drop by 40-60%.
Right now, hardware like that is all being bought up and monopolized by Big AI. This has resulted in prices going up for all these things. In a normal market, it would not cost this much! Furthermore, the reason why Big AI is spending sooooo much fucking money on data centers is because they’re imagining demand. It’s not for training. Not anymore. They’re assuming they’re going to reach AGI any day now and when they do, they’ll need all that hardware to be the world’s “virtual employee” provider.
BTW: Anthropic has a different problem than the others with AGI dreams… Claude (for coding) is in such high demand that their biggest cost is inference. They can’t build out hardware fast enough to meet the demand (inference, specifically). For every dollar they make, they’re spending a dollar to build out infrastructure. Presumably—some day—they’ll actually be able to meet demand with what they’ve got and on that day they’ll basically be printing money. Assuming they can outrun their debts, of course.
Same places as usual: Academia and open source foundations.
That’s where 99% of all advancements in AI come from. You don’t actually think Big AI is paying as many people to do computer science and mathematics research as all the universities in the world (with computer science programs)?
It’s the same shit as always: Big companies commercialize advancements and discoveries made by scientist and researchers from academia (mostly) and give almost nothing back.
Big AI has partnerships with tons of schools and if it weren’t for that, they wouldn’t be advancing the technology as fast as they are. In fact, the only reason why many of these discoveries are made public at all is because of the agreements with the schools that require the discoveries/papers be published (so their school, professors, researchers, and students can get credit).
Like I was saying before: You don’t need a trillion dollars in data centers to do this stuff. Almost all the GPUs and special chips being used (and preordered, sigh) by Big AI are being used to serve their customers (at great expense). Not for training.
Training used to be expensive but so many advancements have been made this is no longer the case. Instead, most of the resources being used in “AI data centers” (and research) is all about making inference more efficient. That’s the step that comes after you give an AI a prompt.
Training a super modern AI model can be done with a university’s data center or a few hundred thousand to a few million dollars of rented GPUs/compute. It doesn’t even take that long!
Generative AI improves at a ridiculously fast rate. In nearly all the ways you could think of: Training, inference (e.g. figuring out user intent), knowledge, understanding, and weirder, fluffier stuff like “creativity” (the benchmarks of which are dubious, BTW).
You seem to be unaware that it only takes about four NVIDIA HGX H100 nodes (32 GPUs) to train something like qwen3.5:122b. That model is about as good as ChatGPT was six months to a year ago (for the usual use cases). That would take a long ass time though (over a year) so you’d want probably 50-100 HGX H100s (or lots of the newer, cheaper ARM-based hardware devices).
The weights for qwen3.5:122b are open. That means that if you’ve got the hardware (loads of universities and non-profits have waaaay TF more than 4 HGX H100 nodes) you can continue modern AI development. Everything you need is right there on Huggingface! Deepseek’s stuff is also open I think but I forget. Aside: In my head, I hold the qwen models as “the gold standard” based on many articles I’ve read about them but AI moves so fast, there might be better stuff out on any given day! I haven’t read AI news in like a week so I could be all wrong and qwen3.5 is now sooo obsolete, hehe (that’s how it feels to follow AI news, anyway 🤣).
Even more interesting: qwen3.5:122b isn’t just an LLM. It does visual reasoning (e.g. give it a picture of a plant and ask it to identify it, count the number of screws in an image, estimate distances, etc) as well as the usual LLM stuff. You can read all about it here:
…and if you install ollama and spend $20 on ollama.com’s cloud service you can actually try it out without having to own enough GPUs to cover the 245+GB requirement. I highly recommend that service! You can try out all the latest & greatest models on your local PC (or phone!) for any purpose you want for a $20. Whenever a new model is out they usually have it up on their servers within a day or two and it’s fast, too.
FYI: I’ve used ollama cloud to evaluate models for coding (web dev with Python back end) and qwen3.5:122b is fantastic. It’s not as good as Claude Opus 4.6 but it’s close (and cheap) enough that you can just make up for the mistakes with extra instances that check the output with a critical eye (the latest trick in AI-based coding to get good output).
For reference, the University of Texas at Austin has data centers with 4,000 NVIDIA Blackwell (B200/GB200) GPUs, Harvard has 1,144 GPUs, and the University of Cambridge & Bristol (in the UK) has some monstrous mix of Intel and AMD GPUs. All three are perfectly capable of training new models from scratch or using continuing development on existing open-weight models like Deepseek and Qwen.
Generative AI isn’t going anywhere. Furthermore, advancements in that space happen so fast that it’s likely that in a few years we won’t need so many GPUs/VRAM to train models. Especially if ternary models (and similar, like Google’s TurboQuant tech) take off.
I know this is a long comment but I want to point something else out: If OpenAI and Anthropic go bust, that would flood the market with cheap GPUs. It would be a total price collapse and you can bet your ass that clever universities and service providers (like Amazon compute, but 3rd party) would snap those up and bring down prices across the board.
Whether AI can reliably detect issues and generate working code is a whole different thing from CEO’s delusions and hyperbole to game the market. Their financial success is also irrelevant, in fact it’s better if the sub/token model fails and we are left with locally ran models.
Traditional software was developed by humans as an artifact that, and to the degree that humans improved the software for some task, got better, but it was not guaranteed. Windows 11 is proof of that, and there are a laundry list of regressions and bugs introduced into software developed by humans. I acknowledge you say usually and especially for open source — I lukewarm agree with that statement but disagree that large LLMs or other generative models will follow this trend, and merely want to point out that software usually introduces bugs as it’s developed, which are hopefully fixed by people who can reason over the code.
Which brings us to AI models, and really they should just be called transformer models; they are statistical tensor product machines. They are not software in a traditional sense. They are trained to match their training input in a statistical sense. If the input data is corrupted, the model will actually get worse over time, not better. If the data is biased, it will get worse over time, not better. With the amount of slop generated on the web, it is extraordinarily hard to denoise and decide what’s good data and what’s bad data that shouldn’t be used for training. Which means the scaling we’ve seen with increased data will not necessarily hold. And there’s not a clear indication that scaling the model size, which is largely already impractical, is having some synergistic or emergent effect as hoped and hyped.
Also, we’re really not in the infancy of AI. Maybe the infancy of widespread hype for it, but the idea of using tensor products for statistical learning algorithms goes back at least as far as Smolensky, maybe before, and that was what, 1990?
We are in the infancy of I’d say quantum style compute, so we really don’t have much to draw on beyond theoretical models.
Generative LLM models have largely plateaued in my opinion.
We’re in the infancy of AI in the sense that widespread use, testing and properly-funded development of these technologies only began a few years ago when massively parallelized processing became affordable enough, even though the concepts are older. You could say we’re in the infancy of practical AI, not theoretical.
That’s what it was. Even the free, open source models are vastly superior to the best of the best from just a year ago.
People got into their heads that AI is shit when it was shit and decided at that moment that it was going to be stuck in that state forever. They forget that AI is just software and software usually gets better over time. Especially open source software which is what all the big AI vendors are building their tools on top of.
We’re still in the infancy of generative AI.
I tried one for the first time yesterday. It was mediocre at best. Certainly not production code. It would take just as much effort to refine it as it would to just write it in the first place.
If you read AI critics, you will see people presenting solid financial evidence of the failure of AI companies to do what they promised. Remember Sam Altman promised AGI in 2025? I certainly do, and now so do you.
Do you have any concrete evidence that this financial flop will turn around before it runs out of money?
Assume all the big AI firms die: Anthropic, OpenAI, Microsoft, Google, and Meta. Poof! They’re gone!
Here would be my reaction: “So anyway… have you tried GLM-7? It’s amazing! Also, there’s a new workflow in ComfyUI I’ve been using that works great to generate…”
Generative AI is here to stay. You don’t need a trillion dollars worth of data centers for progress to continue. That’s just billionaires living in an AGI fantasy land.
Bullshit
I just added up how much it would cost (in theory—assuming everything is in-stock and ready to ship) to build out a data center capable of training something like qwen3.5:122b from scratch in a few months: $66M. That’s how much it would cost for 128 Nvidia B200 nodes (they have 8 GPUs each), infiniband networking, all-flash storage (SSDs), and 20 racks (the hardware).
If OpenAI went bankrupt, that would result in a glut of such hardware which would flood the market, so the cost would probably drop by 40-60%.
Right now, hardware like that is all being bought up and monopolized by Big AI. This has resulted in prices going up for all these things. In a normal market, it would not cost this much! Furthermore, the reason why Big AI is spending sooooo much fucking money on data centers is because they’re imagining demand. It’s not for training. Not anymore. They’re assuming they’re going to reach AGI any day now and when they do, they’ll need all that hardware to be the world’s “virtual employee” provider.
BTW: Anthropic has a different problem than the others with AGI dreams… Claude (for coding) is in such high demand that their biggest cost is inference. They can’t build out hardware fast enough to meet the demand (inference, specifically). For every dollar they make, they’re spending a dollar to build out infrastructure. Presumably—some day—they’ll actually be able to meet demand with what they’ve got and on that day they’ll basically be printing money. Assuming they can outrun their debts, of course.
I’m sick and tired of AI fans making statements like
without evidence.
Citation needed.
Um… Where would it go? I’ve got about 30 models on my machine right now and I download new ones to try out all the time.
Are you suggesting that they’d all just magically disappear one day‽
Where do you think the “new ones” are coming from?
Same places as usual: Academia and open source foundations.
That’s where 99% of all advancements in AI come from. You don’t actually think Big AI is paying as many people to do computer science and mathematics research as all the universities in the world (with computer science programs)?
It’s the same shit as always: Big companies commercialize advancements and discoveries made by scientist and researchers from academia (mostly) and give almost nothing back.
Big AI has partnerships with tons of schools and if it weren’t for that, they wouldn’t be advancing the technology as fast as they are. In fact, the only reason why many of these discoveries are made public at all is because of the agreements with the schools that require the discoveries/papers be published (so their school, professors, researchers, and students can get credit).
Like I was saying before: You don’t need a trillion dollars in data centers to do this stuff. Almost all the GPUs and special chips being used (and preordered, sigh) by Big AI are being used to serve their customers (at great expense). Not for training.
Training used to be expensive but so many advancements have been made this is no longer the case. Instead, most of the resources being used in “AI data centers” (and research) is all about making inference more efficient. That’s the step that comes after you give an AI a prompt.
Training a super modern AI model can be done with a university’s data center or a few hundred thousand to a few million dollars of rented GPUs/compute. It doesn’t even take that long!
Generative AI improves at a ridiculously fast rate. In nearly all the ways you could think of: Training, inference (e.g. figuring out user intent), knowledge, understanding, and weirder, fluffier stuff like “creativity” (the benchmarks of which are dubious, BTW).
Before we spin into a tangent about theory and “what ifs” etc, care to link me to all these great models from academics and open-source institutions?
Because right now, the only companies I see making advancements in “AI” are burning through obscene amounts of cash, with no end in sight.
And there is no evidence the cost of inference is going down, and even Anthropic admits training will continue burning resources.
You seem to be unaware that it only takes about four NVIDIA HGX H100 nodes (32 GPUs) to train something like qwen3.5:122b. That model is about as good as ChatGPT was six months to a year ago (for the usual use cases). That would take a long ass time though (over a year) so you’d want probably 50-100 HGX H100s (or lots of the newer, cheaper ARM-based hardware devices).
The weights for qwen3.5:122b are open. That means that if you’ve got the hardware (loads of universities and non-profits have waaaay TF more than 4 HGX H100 nodes) you can continue modern AI development. Everything you need is right there on Huggingface! Deepseek’s stuff is also open I think but I forget. Aside: In my head, I hold the qwen models as “the gold standard” based on many articles I’ve read about them but AI moves so fast, there might be better stuff out on any given day! I haven’t read AI news in like a week so I could be all wrong and qwen3.5 is now sooo obsolete, hehe (that’s how it feels to follow AI news, anyway 🤣).
Even more interesting: qwen3.5:122b isn’t just an LLM. It does visual reasoning (e.g. give it a picture of a plant and ask it to identify it, count the number of screws in an image, estimate distances, etc) as well as the usual LLM stuff. You can read all about it here:
https://ollama.com/library/qwen3.5:122b
…and if you install
ollamaand spend $20 on ollama.com’s cloud service you can actually try it out without having to own enough GPUs to cover the 245+GB requirement. I highly recommend that service! You can try out all the latest & greatest models on your local PC (or phone!) for any purpose you want for a $20. Whenever a new model is out they usually have it up on their servers within a day or two and it’s fast, too.FYI: I’ve used ollama cloud to evaluate models for coding (web dev with Python back end) and qwen3.5:122b is fantastic. It’s not as good as Claude Opus 4.6 but it’s close (and cheap) enough that you can just make up for the mistakes with extra instances that check the output with a critical eye (the latest trick in AI-based coding to get good output).
For reference, the University of Texas at Austin has data centers with 4,000 NVIDIA Blackwell (B200/GB200) GPUs, Harvard has 1,144 GPUs, and the University of Cambridge & Bristol (in the UK) has some monstrous mix of Intel and AMD GPUs. All three are perfectly capable of training new models from scratch or using continuing development on existing open-weight models like Deepseek and Qwen.
Generative AI isn’t going anywhere. Furthermore, advancements in that space happen so fast that it’s likely that in a few years we won’t need so many GPUs/VRAM to train models. Especially if ternary models (and similar, like Google’s TurboQuant tech) take off.
I know this is a long comment but I want to point something else out: If OpenAI and Anthropic go bust, that would flood the market with cheap GPUs. It would be a total price collapse and you can bet your ass that clever universities and service providers (like Amazon compute, but 3rd party) would snap those up and bring down prices across the board.
@ikidd@lemmy.world @ ingeanus@ttrpg.network do you two have a source for these supposed great models?
https://mastodon.social/@nixCraft/111695037458159431
Oh wow, comparing a thing to a completely different thing without demonstrating the comparison is valid.
Exactly the non-evidence I expected.
Whether AI can reliably detect issues and generate working code is a whole different thing from CEO’s delusions and hyperbole to game the market. Their financial success is also irrelevant, in fact it’s better if the sub/token model fails and we are left with locally ran models.
Traditional software was developed by humans as an artifact that, and to the degree that humans improved the software for some task, got better, but it was not guaranteed. Windows 11 is proof of that, and there are a laundry list of regressions and bugs introduced into software developed by humans. I acknowledge you say usually and especially for open source — I lukewarm agree with that statement but disagree that large LLMs or other generative models will follow this trend, and merely want to point out that software usually introduces bugs as it’s developed, which are hopefully fixed by people who can reason over the code.
Which brings us to AI models, and really they should just be called transformer models; they are statistical tensor product machines. They are not software in a traditional sense. They are trained to match their training input in a statistical sense. If the input data is corrupted, the model will actually get worse over time, not better. If the data is biased, it will get worse over time, not better. With the amount of slop generated on the web, it is extraordinarily hard to denoise and decide what’s good data and what’s bad data that shouldn’t be used for training. Which means the scaling we’ve seen with increased data will not necessarily hold. And there’s not a clear indication that scaling the model size, which is largely already impractical, is having some synergistic or emergent effect as hoped and hyped.
Also, we’re really not in the infancy of AI. Maybe the infancy of widespread hype for it, but the idea of using tensor products for statistical learning algorithms goes back at least as far as Smolensky, maybe before, and that was what, 1990?
We are in the infancy of I’d say quantum style compute, so we really don’t have much to draw on beyond theoretical models.
Generative LLM models have largely plateaued in my opinion.
We’re in the infancy of AI in the sense that widespread use, testing and properly-funded development of these technologies only began a few years ago when massively parallelized processing became affordable enough, even though the concepts are older. You could say we’re in the infancy of practical AI, not theoretical.
They should all be destroyed