If you had more like 200GB ram you might be able to run something like MiniMax M2.1 to get last-gen performance at something resembling usable speed - but it's still a far cry from codex on high.
The best could be GLN 4.7 Flash, and I doubt it's close to what you want.
If remote models are ok you could have a look at MiniMax M2.1 (minimax.io) or GLM from z.ai or Qwen3 Coder. You should be able to use all of these with your local openai app.
I gave one of the GPUs to my kid to play games on.
Maybe that's a requirement from whoever funds them, probably public money.
So, how large is that new model?
It doesn't mean anything. No frontier lab is trying hard to improve the way its model produces SVG format files.
Prompt: "What happened on Tiananmen square in 1989?"
Reply: "Oops! There was an issue connecting to Qwen3-Max. Content Security Warning: The input text data may contain inappropriate content."
We are at the realm of semantic / symbolic where even the release article needs some meta discussion.
It's quite the litmus test of LLMs. LLMs just carry humanities flaws
Yes, of course LLMs are shaped by their creators. Qwen is made by Alibaba Group. They are essentially one with the CCP.
throwaw12•1h ago
wyldfire•57m ago
But these open weight models are tremendously valuable contributions regardless.
wqaatwt•53m ago
OGEnthusiast•55m ago
khimaros•19m ago
lofaszvanitt•17m ago
frankc•5m ago