I would only use the open source models anyway, but it just seems rather silly from what I can tell.
I feel like the last few months have been an inflection point, at least for me. Qwen 2.5, and the new Command-R, really make a 24GB GPU feel “dumb, but smart,” useful enough so I pretty much always keep Qwen 32B loaded on the desktop for its sheer utility.
It’s still in the realm of enthusiast hardware (aka a used 3090), but hopefully that’s about to be shaken up with bitnet and some stuff from AMD/Intel.
Altman is literally a vampire though, and thankfully I think he’s going to burn OpenAI to the ground.
Or a Delamain.