We have faster memory, it's just all used in data center cards you can't buy (and can't afford to buy).
AMD actually used HBM2 memory in their Radeon VII card back in 2019 (!!) for $700. It had 16 GB of HBM2 memory with 1 TB/s throughput.
The RTX 5080 in conversion l comparison also has 16 GB of VRAM, but was released in 2025 and has 960 GB/s throughput. The RTX 5090 does have an edge at 1.8 TB/s bandwidth and 32 GB of VRAM but it also costs several times more. Imagine if GPUs had gone down the path of the Radeon VII.
That being said, the data center cards from both are monstrous.
The Nvidia B200 has 180 GB of VRAM (2x90GB) offering 8.2 TB/s bandwidth (4.1 TB/s x2) released in 2024. It just costs as much as a car, but that doesn't matter, because afaik you can't even buy them individually. I think you need to buy a server system from Nvidia or Dell that will come with like 8 of these and cost you like $600k.
AMD has the Mi series. Eg AMD MI325x. 288 GB of VRAM doing 10 TB/s bandwidth and released in 2024. Same story as Nvidia: buy from an OEM that will sell you a full system with 8x of these (and if you do get your hands on one of these you need a special motherboard for them since they don't do PCIe). Supposedly a lot cheaper than Nvidia, but still probably $250k.
These are not even the latest and greatest for either company. The B300 and Mi355x are even better.
It's a shame about the socket for the Mi series GPUs (and the Nvidia ones too). The Mi200 and Mi250x would be pretty cool to get second-hand. They are 64 GB and 128GB VRAM GPUs, but since they use OAP socket you need the special motherboard to run them. They're from 2021, so in a few years time they will likely be replaced, but as a regular joe you likely can't use them.
The systems exist, you just can't have them, but you can rent them in the cloud at about $2-4 per hour per GPU.
I have doubts about this. Perhaps the closed models have, but I wouldn't be so sure for the open ones.
GLM 5, for example, is running 16-bit weights natively. This makes their 755B model 1.5TB in size. It also makes their 40B active parameters ~80GB each.
Compare this to Kimi K2.5. 1T model, but it's 4-bit weights (int4), which makes the model ~560 GB. Their 32B active parameters are ~16 GB.
Sure, GLM 5 is the stronger model, but is that price worth paying with 2-3x longer generation times? What about 2-3x more memory required?
I think this barrel's bottom really hasn't been scraped.
I disagree. You write when you have something to say. A service like Grammarly tries to help you convey what you want to say, but better. What you want to say is still up to you.
Words paint the picture, but the meaning of the picture is what matters.
Children and young students, certainly. Adult students: almost 100%. If writing is your job, then by definition, and your problem is more often finding something to say, not writing it.
You’re not counting all the office workers who have to write reports or emails, or all the scammers who write those websites to manipulate SEO or show you ads.
Everyone should think twice about putting their name on AI garbage, or garbage of any kind. But wishing doesn’t stop it from happening, especially when companies are explicitly selling you on doing just that. Remember the Apple Intelligence office ads?
Besides not getting consent in the case of HeLa, which part do you find problematic? Cancerous cell's ability to self-clone/grow is as much a feature as it is a bug in this particular use case.
I ask as someone who's has personally experienced loss of several loved ones from cancer (as most people my age probably have), but doesn't share your aversion to this particular use case (research.)
Yeah I do feel the OA is being overly flippant with their use of human cells here, likely for PR sake, which would be an ethical breach for me personally. Overall though, I find most research cases for human cell lines to be in line with my personal ethics. Neuron lines can certainly be used for good or ill, and this case leans towards the latter, although understanding the human brain may justify this line of work in the long term. If only we didn't live in a militaristic late stage capitalist society...
Didn't Anthropic's case already set the precedent that training itself is fine? It's not like copyrighted novels are a large portion of human-generated text data. It's just the stuff that's easier to get because it's preserved in bulk.
Video transcription has more or less been solved. Imagine how much data Google has in YouTube transcripts. And the longer these AI chat bots operate the more data they manage to collect for training as well (I think Google making it so you can easily upvote or downvote a response by the bot is a good idea).
And why would countries adopt this? So that other countries can use this cartel to push their own agenda? If anything it seems like it would be in every country's best interests to make sure such an organization doesn't exist.
In my opinion one of the reasons why European economies have been struggling for a long time is because energy has been much more expensive than elsewhere. Part of it is the excise tax on gasoline because it drives up the price of everything.
Even to this day EU countries where people earn less than a third of what Americans earn still pay more for gasoline.
AMD actually used HBM2 memory in their Radeon VII card back in 2019 (!!) for $700. It had 16 GB of HBM2 memory with 1 TB/s throughput.
The RTX 5080 in conversion l comparison also has 16 GB of VRAM, but was released in 2025 and has 960 GB/s throughput. The RTX 5090 does have an edge at 1.8 TB/s bandwidth and 32 GB of VRAM but it also costs several times more. Imagine if GPUs had gone down the path of the Radeon VII.
That being said, the data center cards from both are monstrous.
The Nvidia B200 has 180 GB of VRAM (2x90GB) offering 8.2 TB/s bandwidth (4.1 TB/s x2) released in 2024. It just costs as much as a car, but that doesn't matter, because afaik you can't even buy them individually. I think you need to buy a server system from Nvidia or Dell that will come with like 8 of these and cost you like $600k.
AMD has the Mi series. Eg AMD MI325x. 288 GB of VRAM doing 10 TB/s bandwidth and released in 2024. Same story as Nvidia: buy from an OEM that will sell you a full system with 8x of these (and if you do get your hands on one of these you need a special motherboard for them since they don't do PCIe). Supposedly a lot cheaper than Nvidia, but still probably $250k.
These are not even the latest and greatest for either company. The B300 and Mi355x are even better.
It's a shame about the socket for the Mi series GPUs (and the Nvidia ones too). The Mi200 and Mi250x would be pretty cool to get second-hand. They are 64 GB and 128GB VRAM GPUs, but since they use OAP socket you need the special motherboard to run them. They're from 2021, so in a few years time they will likely be replaced, but as a regular joe you likely can't use them.
The systems exist, you just can't have them, but you can rent them in the cloud at about $2-4 per hour per GPU.
reply