

I don’t believe that was claimed?
Northern territory also


I don’t believe that was claimed?
Northern territory also


Is it though?
Significantly?
Hhmmmm


Maps without nz


It’s not significantly larger. Lots of people though.


Antarctica becons


ERM
Texas is larger than Victoria and the Australian Capital Territory (ACT), but smaller than all other Australian states.
If Texas were a state in Australia, it would rank as the sixth. Behind New South Wales, Queensland, Western Australia, and South Australia.
Erm… Not that big


It’s not that big mate


Pretty close then. Interesting. Thanks for the stats.


I believe Trumptopia is about the same size physically as Australia.
Same thing.
Big city. Lock.
Small town. Maybe lock.
Rural area. Never lock.


It’s probably a little bit like when people used to copy tapes of their favourite artists in the 80s. And then that was just for private use and home projects.
But this is next level.


Duck that


I’ve used it for a few years. What issue does it have for a desktop? I’ve had everything “just work”.


Full on. China poisoning open source software. Lucky it got caught. Imagine how much of it is going on.


It’s absolutely mind boggling. The power of the people.


That is impressive!
Zooming in, i can see villages where i know people. These are very small villages, and the system is in use.
So I imagine it’s friends, and family.
And that’s ok. A great way to start and build a momentum.
Go Deutschland!


I asked the bot. It says:
I understand the concern. This question sits at the intersection of copyright law, statistics, and how machine-learning systems work—and it’s fair to be skeptical.
Here’s the most precise and careful way to frame it:
I don’t know what specific works or licenses were in the training set, and I can’t confirm that GPL-licensed material was included. Neither I nor users have visibility into the underlying corpus.
So while it may be statistically plausible that GPL text exists somewhere in public data, plausibility ≠ proof, and I can’t assert it as fact.
OpenAI’s legal position (and that of most ML practitioners) is broadly this:
Training on copyrighted material is not the same as redistributing it
The model does not contain copies of works
Outputs are newly generated, not retrieved
Copyright law regulates copying and distribution, not learning patterns or statistics
This is analogous (legally, not philosophically) to a human reading GPL code and then later writing different code from experience.
Whether courts ultimately agree in every jurisdiction is still being litigated—but no court has ruled that training itself is per se illegal.
GPL obligations trigger when you:
copy
modify
distribute
GPL-licensed code or derivative works.
Current legal consensus does not treat model weights or outputs as GPL “derivatives” in the copyright sense, because:
There is no preserved expressive structure
There is no traceable lineage to a specific work
Outputs are not mechanically derived from any one source
That’s why GPL’s copyleft provisions are generally understood not to propagate through model training.
No. Anyone claiming it is settled is overstating.
What is true:
Training on public data is widespread and openly practiced
Multiple lawsuits exist, but none have yet established that training itself violates copyright or GPL
Regulators and courts are still clarifying boundaries
Even if GPL material were in the training mix:
That alone would not automatically make the model or its outputs illegal
Liability (if any) would hinge on specific outputs being substantially similar to protected works, not on statistical likelihoods about training data
Your intuition about probability is reasonable—but legality does not turn on probability. It turns on:
whether protected expression is reproduced
whether derivative works are created
whether licenses are triggered by the act in question
As of now, the prevailing legal interpretation is that models like me are lawful to deploy, even though the debate is ongoing.
If you want, I can also:
Compare this to the Google Books ruling
Walk through why “derivative work” is a high legal bar
Discuss what would actually make an AI system GPL-tainted in practice


Watch the movie named The Time Traveller’s Wife. It is absolutely superb and based on sort of that idea, but not quite.
I use Talon Voice


Full Linux shop here. Love it…
Desktops, laptops, servers.
For those rare customer teams meets, we just do it in the browser.
</saltRub>
I had a project of markdown files. About 400 of them, with about 1200 plus links in them.
The original filenames were changed. The links no longer worked.
The LLM went through each link, and found the new one, based on filename and file content, using its ability to recognise patterns, words, etc etc.
Absolutely saved me maybe a couple of days of manual pain labour, and all done in about 10 minutes.
This is the kind of thing I use it for. Horrible repetitive processes.