Awesome!
I wonder if things will organize around a “unofficial” modding API like Harmony for Rimworld, Forge for Minecraft, SMAPI for Stardew Valley, and so on? I guess it depends if some hero dev team does it and there’s enough “demand” to build a bunch of stuff on it. But a “final” patch (so future random patches don’t break the community API) and community enthusiasm from Larian are really good omens.
Skyrim and some other games stayed more fragmented, others like CP2077 just never hit critical mass I guess. And the format of the modded content just isn’t the same for mega RPGs like this.
Yeah you are correct, I was venting lol.
Another factor is that fab choice design decisions were made way before the GPUs launched, when everything you said (TSMC’s lead/reliability, in particular) rang more true. Maybe Samsung or Intel could offer steep discounts for the lower performance (hence Nvidia/AMD could translate that to bigger dies), but that’s quite a fantasy I’m sure…
It all just sucks now.
The chips going to datacenters could have been consumer stuff instead.
This is true, but again, they do use different processes. The B100 (and I think the 5090) is TSMC 4NP, while the other chips use a lesser process. Hopper (the H100) was TSMC 4N, Ada Lovelace (RTX 4000) was TSMC N4. The 3000 series/A100 was straight up split between Samsung and TSMC. The AMD 7000 was a mix of older N5/N6 due to the MCM design.
Local AI benefits from platforms with unified memory that can be expanded.
This is tricky because expandable memory is orthogonal to bandwidth and power efficiency. Framework (ostensibly) had to use soldered memory for their Strix Halo box because it’s literally the only way to make the traces good enough: SO-DIMMs are absolutely not fast enough, and even LPCAMM apparently isn’t there yet.
AMD’s Ryzen AI MAX 300 chip
Funny thing is the community is quite lukewarm to the AMD APUs due to poor software support. It works okay… if you’re a python dev that can spend hours screwing with rocm to get things fast :/ But it’s quite slow/underutilized if you just run popular frameworks like ollama or the old diffusion ones.
It’s the main reason why I believe Apple’s memory upgrades cost a ton so that it isn’t a viable option financially for local AI applications.
Nah, Apple’s been gouging memory way before AI was a thing. It’s their thing, and honestly it kinda backfired because it made them so unaffordable for AI.
Also, Apple’s stuff is actually… Not great for AI anyway. The M-chips have relatively poor software support (no pytorch, MLX is barebones, leaving you stranded with GGML mostly). They don’t have much compute compared to a GPU or even an AMD APU, the NPU part is useless. Unified memory doesn’t help at all, it’s just that their stuff happens to have a ton of memory hanging off the GPU, which is useful.
Unfortunately, no one is buying a 7900 XTX for AI, mostly not a 5090 either. The 5090 didn’t even work till recently and still doesn’t work with many projects, doubly so for the 7900 XTX.
The fab capacity thing is an issue, but not as much as you’d think since the process nodes are different.
Again, I am trying to emphasize, a lot of this is just Nvidia being greedy as shit. They are skimping on VRAM/busses and gouging gamers because they can.
The Nvidia GPUs in data centers are separate (and even on separate nodes than, with different memory chips than) gaming GPUs. The sole exception is the 4090/5090 which do see some use in data center forms, but at low volumes. And this problem is pretty much nonexistent for AMD.
…No, it’s just straight up price gouging and anti competitiveness. It’s just Nvidia being Nvidia, AMD being anticompetitive too (their CEOs are like cousins twice removed), and Intel unfortunately not getting traction, even though Battlemage is excellent.
For local AI, the only thing that gets sucked up are 3060s, 3090s, and for the rich/desperate, 4090s/5090s, with anything else being a waste of money with too little VRAM. And this is a pretty small niche.
Yeah, you and @[email protected] have a point.
I am vastly oversimplifying a lot, but… Perhaps mobile gaming, on aggregate, is too shitty for its own good? It really looks that way whenever I sample the popular ones.
live service games make up a significant amount of what the average consumer wants, and those customers largely play on PC for all sorts of reasons
You are leaving out the elephant in the room: smartphones.
So, so, so many people game on smartphones. It’s technically the majority of the “gaming” market, especially live service games. A large segment of the population doesn’t even use PCs and does the majority of their computer stuff on smartphones or tablets, and that fraction seems to be getting bigger. Point being the future of the Windows PC market is no guarantee.
Hear me out:
6 core CCD. Clocked real slow, but with 3D cache like the 5600x3d.
The slightly cut 32 CU GPU. Clocked real slow.
32GB of that LPDDR5X. 24GB, if the config is possible?
…OLED? I feel like there’s a much better selection of tablet screens to borrow now. If not, use whatever SKU the switch does.
I can dream, can’t I? But modern laptop GPUs/CPUs are absurdly efficient if you underclock them a little.
The problem is the way they’re pushing the tools as magic lamps, and shoving them down everyone’s throats.
AI is a really neat tool that got dragged into an incredibly toxic system that enshittified it. Not a useful tool to help development, no, skip straight to replacing employees even if it doesn’t freaking work.
Funny thing is SWTOR has some great art, heartfelt voice acting and quests, great soundtrack and such, but at the end of the day it’s buried in a grindy.
On the other hand, I tried Fallout 76 (after it was patched up) drunk with friends, and it was boring as heck. The quests were so dull, gameplay so arbitrarily janky and grindy. Drunk! With friends! Do you know how low a bar that is :/
Funny thing about AMD is the MI300X is supposedly not selling well, largely because they priced gouge everything as bad as Nvidia, even where they aren’t competitive. Other than the Framework desktop, they are desperate to stay as uncompetitive in the GPU space as they possibly can, and not because the hardware is bad.
Wasn’t the Intel B580 a good launch, though? It seems to have gotten rave reviews, and it’s in stock, yet has exited the hype cycle.
When do you think that stopped though?
There’s a lot of love for Skyrim, but I feel like there was already deterioration in the side quest writing, even strictly looking at Oblivion/FO3, not Morrowind.
As for BioWare, even ME3 was starting to show some cracks, even if you set the ending aside. And I loved Mass Effect to death. Heck, I’m even a bigger Andromeda fan than most.
…Point being I think we clung to BioWare/Bethesda a little too hard even when the signs of deoxygenation were there.
People understandably love to hate Oblivion and Fallout 3, but I feel the side quest writing had heart, like groups of devs got to go wild within their own little dungeons. Their exploitable mechanics were kinda endearing.
…And I didn’t get that from Starfield? I really tried to overlook the nostalgia factor, but all the writing felt… corporate. Gameplay, animation, Bethesda jank without any of the fun. I abandoned it early and tried to see what I was missing on YouTube, but still don’t “get” what people see in that game.
If you want a big walking sandbox in that vein, I feel like No Man’s Sky would scratch the itch far better, no?
Meanwhile, BG3 and KC2 completely floored me. So did Cyberpunk 2077, though I only experienced it patched up and modded. Heck, even ME Andromeda felt more compelling to me.
I find this driveby comment rather significant.
It means they are trying to conform to the developers’ strengths, desires, interests. They’re shaping huge business decisions around them. That’s just good for everyone, as opposed to devs inefficiently, dispassionately grinding away at something they don’t like.
That’s huge. I’d also posit “happy devs means happy business.” And Larian has repeatedly expressed similar things.