

deleted by creator
Father, Hacker (Information Security Professional), Open Source Software Developer, Inventor, and 3D printing enthusiast


deleted by creator


Meh. For a control I need to see a comparison of the lies in the search results. I bet Gemini is doing a better job than the limitless bullshit that people are expected to click on.
Let us not forget that Gemini is often pointing out basic facts like, “vaccines actually work” when people search for information about them. This angers a lot of complete idiots and results in anger-inducing clickbait articles… Like this one.
The fact that a machine can get 90% of everything correct is fucking amazing. And the pace of improvement in AI in general is astounding. Try to keep that in perspective when reading stories like this.
Anything over 50% is probably better than a random sampling of humans could do, based on our current state of education and “knowledge upkeep” (which is something most humans fail at, utterly and completely).


It’s only one company that hoarded all the RAM: OpenAI. They’re the ones that intentionally created scarcity in the market in order to maintain what they thought was a dominant position.
Don’t lump all the other AI companies in with them. They’re they ones that claimed to be all about improving humanity and registered themselves as a non-profit, then turned around and went for-profit with shady deals left and right (like the DRAM thing).
Anthropic lost their contract with the government because they refused to use their AI to snoop on innocent people and possibly make automatic determinations as to who and what to bomb. OpenAI leapt right on that opportunity, “we’re happy to do that! Give us the contract!”
I’m not saying the other AI companies are pinnacles of ethics but there’s one player—who is so closely aligned with Microsoft they have employees sitting on each other’s campuses—that is vastly worse than the others when it comes to underhand shit that’s bad for everyone.


No. It’s because of the Epstein files.
Did she actually do the bare minimum of what she was legally required to do? No. Those lawsuits are pending because it’s so fucking obvious she didn’t release stuff (Congresscritters who had access to much of it came out and said so).
Did she intentionally try to hide Trump and his cronies presence in the Epstein files? Yes.
If she took all these steps, and put herself in legal jeopardy, just to protect Trump and his cronies, why is he still mad at her?
It’s because she released them at all. I’m sure he was expecting her to delay the release forever. Or at least until all the lawsuits played out (which could complete after his term).
The underlying truth here is actually just the usual Trump stuff: He doesn’t surround himself with competent people. He only chooses people based on perceived loyalty and whims.


My question exactly: The computers should be purpose-built, including the operating system.
Why TF aren’t they using something like NASA Linux‽
If they made it open source you bet your ass they’d get shittons of free support from the global community! If they’re running my software I’d be willing to hop on a call with the command center on any day at any hour!
“Yes, I know it’s Christmas but NASA is having some trouble with a systemd script on a space ship that’s currently in space…”


The US is not a theocracy. Conservatives want it to be one—in theory—but they would never agree on which religion would be the one true religion.
You’d think they’d settle on something simple and nebulous like, “Christianity” but the moment they started trying to define that in law the whole concept would fall apart because there’s way too many completely incompatible differences between Christian sects. Not to mention the fact that Mormons (and other niche sects) consider themselves to be Christian while huge swaths of people consider them to be anything but.
The best they can ever get away with is what they’ve got now: Completely unconstitutional (IMHO) exceptions in various laws for “genuinely held religious beliefs.”
Remember: The conservatives on the supreme court really do think that if a doctor has a genuine religious belief that someone should die from a treatable condition, they should not be held to account for letting that person die.
I fantasize about one of these justices going to the hospital for an emergency heart condition and having the doctor refuse to treat them because of a truly genuine, deeply-held religious belief that conservatives should just die from such things since they don’t believe in medicine or science in general.


Jokes on him: Wolfenstein is more likely.


I literally said I’m using qwen3.5:122b for coding. I also use GLM-5 but it’s slightly slower so I generally stick with qwen.
It’s right there, in ollama’s library: https://ollama.com/library/qwen3.5:122b
The weights and everything else for it are on Huggingface: https://huggingface.co/Qwen/Qwen3.5-122B-A10B
This is not speculation. That’s what I’m actually using nearly every day. It’s not as good as Claude Code with Opus 4.6 but it’s about 90% of the way there (if you use it right). When GLM-5 came out that’s when I cancelled my Claude subscription and just stuck with Ollama Cloud.
I can use gpt-oss:20b on my GPU (4060 Ti 16GB)—and it works well—but for $20/month, the ability to use qwen3.5 and GLM-5 are better options.
I still use my GPU for (serious) image generation though. Using ChatGPT (DALL-E) or Gemini (Nano Banana) are OK for one-offs but they’re slow AF compared to FLUX 2 and qwen’s image models running locally. I can give it a prompt and generate 32 images in no time, pick the best one, then iterate from there (using some sophisticated ComfyUI setups). The end result is a superior image than what you’d get from Big AI.


I just added up how much it would cost (in theory—assuming everything is in-stock and ready to ship) to build out a data center capable of training something like qwen3.5:122b from scratch in a few months: $66M. That’s how much it would cost for 128 Nvidia B200 nodes (they have 8 GPUs each), infiniband networking, all-flash storage (SSDs), and 20 racks (the hardware).
If OpenAI went bankrupt, that would result in a glut of such hardware which would flood the market, so the cost would probably drop by 40-60%.
Right now, hardware like that is all being bought up and monopolized by Big AI. This has resulted in prices going up for all these things. In a normal market, it would not cost this much! Furthermore, the reason why Big AI is spending sooooo much fucking money on data centers is because they’re imagining demand. It’s not for training. Not anymore. They’re assuming they’re going to reach AGI any day now and when they do, they’ll need all that hardware to be the world’s “virtual employee” provider.
BTW: Anthropic has a different problem than the others with AGI dreams… Claude (for coding) is in such high demand that their biggest cost is inference. They can’t build out hardware fast enough to meet the demand (inference, specifically). For every dollar they make, they’re spending a dollar to build out infrastructure. Presumably—some day—they’ll actually be able to meet demand with what they’ve got and on that day they’ll basically be printing money. Assuming they can outrun their debts, of course.


I personally love glm-5 and qwen3.5, specifically: https://ollama.com/library/qwen3.5:122b
I’ve used them both for coding and they work really well (way better than you’d think). They’re also perfectly capable of the usual LLM chat stuff (e.g. check my grammar) but all the models (even older, smaller ones) are capable of that stuff these days.
For a treat: Have someone show you using some of these models to search the web! It’s amazing. You don’t see ads, you don’t have to comb through 12 pages of search results, and they read the pages that moment (not cached) to give you summaries of the content. So when you click the link to go to the content you know it’s the thing you were looking for. They’re not using a local index of the Internet, they’re searching on your behalf using whatever search engines you configured. It’s waaaaay better than ChatGPT (which uses Bing behind the scenes whether you like it or not) or Gemini (which uses Google, obviously). The (self-hosted) LLM will literally be running curl for you on Google, DuckDuckGo, Bing, or whatever TF else you want (simultaneously) then reading each of the search results and using your prompt to figure out what the most relevant results are. It’s sooooo nice!
FYI: Ollama.com’s library page is actually a great resource for finding info on all the models that can be self-hosted: https://ollama.com/library


You seem to be unaware that it only takes about four NVIDIA HGX H100 nodes (32 GPUs) to train something like qwen3.5:122b. That model is about as good as ChatGPT was six months to a year ago (for the usual use cases). That would take a long ass time though (over a year) so you’d want probably 50-100 HGX H100s (or lots of the newer, cheaper ARM-based hardware devices).
The weights for qwen3.5:122b are open. That means that if you’ve got the hardware (loads of universities and non-profits have waaaay TF more than 4 HGX H100 nodes) you can continue modern AI development. Everything you need is right there on Huggingface! Deepseek’s stuff is also open I think but I forget. Aside: In my head, I hold the qwen models as “the gold standard” based on many articles I’ve read about them but AI moves so fast, there might be better stuff out on any given day! I haven’t read AI news in like a week so I could be all wrong and qwen3.5 is now sooo obsolete, hehe (that’s how it feels to follow AI news, anyway 🤣).
Even more interesting: qwen3.5:122b isn’t just an LLM. It does visual reasoning (e.g. give it a picture of a plant and ask it to identify it, count the number of screws in an image, estimate distances, etc) as well as the usual LLM stuff. You can read all about it here:
https://ollama.com/library/qwen3.5:122b
…and if you install ollama and spend $20 on ollama.com’s cloud service you can actually try it out without having to own enough GPUs to cover the 245+GB requirement. I highly recommend that service! You can try out all the latest & greatest models on your local PC (or phone!) for any purpose you want for a $20. Whenever a new model is out they usually have it up on their servers within a day or two and it’s fast, too.
FYI: I’ve used ollama cloud to evaluate models for coding (web dev with Python back end) and qwen3.5:122b is fantastic. It’s not as good as Claude Opus 4.6 but it’s close (and cheap) enough that you can just make up for the mistakes with extra instances that check the output with a critical eye (the latest trick in AI-based coding to get good output).
For reference, the University of Texas at Austin has data centers with 4,000 NVIDIA Blackwell (B200/GB200) GPUs, Harvard has 1,144 GPUs, and the University of Cambridge & Bristol (in the UK) has some monstrous mix of Intel and AMD GPUs. All three are perfectly capable of training new models from scratch or using continuing development on existing open-weight models like Deepseek and Qwen.
Generative AI isn’t going anywhere. Furthermore, advancements in that space happen so fast that it’s likely that in a few years we won’t need so many GPUs/VRAM to train models. Especially if ternary models (and similar, like Google’s TurboQuant tech) take off.
I know this is a long comment but I want to point something else out: If OpenAI and Anthropic go bust, that would flood the market with cheap GPUs. It would be a total price collapse and you can bet your ass that clever universities and service providers (like Amazon compute, but 3rd party) would snap those up and bring down prices across the board.


I just looked up the actual numbers and the highest estimate I could find was that there could be about 70,000 babies born to parents on temporary visas each year (Center for Immigration Studies). Other figures suggested around 10,000 at most but I picked the biggest one to make a point:
It doesn’t fucking matter.
70,000 births a year in a nation that has 3.5 million births/year is nothing. That’s 2% so I’m going to reiterate that nobody should care. It’s a stupidly small figure that is of zero consequence.
The only negative connotation one could derive from this figure is founded in racism.


No no. I’m absolutely certain that the Trump administration considers birthright citizenship to be a problem—when it’s non-white babies that are getting citizenship.
They need to stop beating around the bush and just say that.


Same places as usual: Academia and open source foundations.
That’s where 99% of all advancements in AI come from. You don’t actually think Big AI is paying as many people to do computer science and mathematics research as all the universities in the world (with computer science programs)?
It’s the same shit as always: Big companies commercialize advancements and discoveries made by scientist and researchers from academia (mostly) and give almost nothing back.
Big AI has partnerships with tons of schools and if it weren’t for that, they wouldn’t be advancing the technology as fast as they are. In fact, the only reason why many of these discoveries are made public at all is because of the agreements with the schools that require the discoveries/papers be published (so their school, professors, researchers, and students can get credit).
Like I was saying before: You don’t need a trillion dollars in data centers to do this stuff. Almost all the GPUs and special chips being used (and preordered, sigh) by Big AI are being used to serve their customers (at great expense). Not for training.
Training used to be expensive but so many advancements have been made this is no longer the case. Instead, most of the resources being used in “AI data centers” (and research) is all about making inference more efficient. That’s the step that comes after you give an AI a prompt.
Training a super modern AI model can be done with a university’s data center or a few hundred thousand to a few million dollars of rented GPUs/compute. It doesn’t even take that long!
Generative AI improves at a ridiculously fast rate. In nearly all the ways you could think of: Training, inference (e.g. figuring out user intent), knowledge, understanding, and weirder, fluffier stuff like “creativity” (the benchmarks of which are dubious, BTW).


Trump has insisted the executive order is aimed at combatting “birth tourism,” immigrants who come to the United States briefly for the purpose of having a child.
This assumes that birth tourism—if it even exists in any non-trivial amount of babies—is a problem. Why would it be a problem? I don’t get it.
The administration is always bitching about not enough babies… Make up your damned mind!
…or perhaps be more explicit about your true intentions?


Um… Where would it go? I’ve got about 30 models on my machine right now and I download new ones to try out all the time.
Are you suggesting that they’d all just magically disappear one day‽


Assume all the big AI firms die: Anthropic, OpenAI, Microsoft, Google, and Meta. Poof! They’re gone!
Here would be my reaction: “So anyway… have you tried GLM-7? It’s amazing! Also, there’s a new workflow in ComfyUI I’ve been using that works great to generate…”
Generative AI is here to stay. You don’t need a trillion dollars worth of data centers for progress to continue. That’s just billionaires living in an AGI fantasy land.


Either a lot more tools got a lot better,
That’s what it was. Even the free, open source models are vastly superior to the best of the best from just a year ago.
People got into their heads that AI is shit when it was shit and decided at that moment that it was going to be stuck in that state forever. They forget that AI is just software and software usually gets better over time. Especially open source software which is what all the big AI vendors are building their tools on top of.
We’re still in the infancy of generative AI.


Patents on software shouldn’t exist!
I’m glad this guy slithered into my feed this morning.