

If they do it’s not by the actual training of AI.
Basically a deer with a human face. Despite probably being some sort of magical nature spirit, his interests are primarily in technology and politics and science fiction.
Spent many years on Reddit before joining the Threadiverse as well.


If they do it’s not by the actual training of AI.


Simple.wikipedia isn’t a summary of regular Wikpedia, it’s a whole separate thing. It’s intended to convey the same data, just in a simpler way.


The problem being discussed here is not the availability of Wikipedia’s data. It’s about the ongoing maintenance and development of that data going forward, in the future. Having a static copy of Wikipedia gathering dust on various peoples’ hard drives isn’t going to help that.


Wikipedia’s traditional self-sustaining model works like this: Volunteers (editors) write and improve articles for free, motivated by idealism and the desire to share knowledge. This high-quality content attracts a massive number of readers from search engines and direct visits. Among those millions of readers, a small percentage are inspired to become new volunteers/editors, replenishing the workforce. This cycle is “virtuous” because each part fuels the next: Great content leads to more readers which leads to more editors which leads to even better content. AI tools (like ChatGPT, Google AI Overviews, Perplexity, etc.) disrupt this cycle by intercepting the user before they reach Wikipedia.
A week or two back there was a post on Reddit where someone was advertising a project they’d put up on GitHub, and when I went to look at it I didn’t find any documentation explaining how it actually worked - just how to install it and run it.
So I gave Gemini the URL of the repository and asked it to generate a “Deep Research” report on how it worked. Got a very extensive and detailed breakdown, including some positives and negatives that weren’t mentioned in the existing readme.
America isn’t the only source of history books.
How would that layer distinguish AI from non-AI?


Oh boy, I bet the comments on this one will be useful.


That 47%-for-Trump figure cannot be explained by gerrymandering, it’s a state-wide figure.
Frankly, I’m trying to make a point here too. I’m pointing out how even when Americans are supposedly being “nice” they’re often still grounding their niceness in unspoken assumptions of American exceptionalism. A bit of America going “we’ll just go join Canada and surely Canada will want to have us” has the same underlying feeling that obviously everyone would like to be American and join America if they could.


We don’t want it. Minnesota voted 47% for Trump, and the Democrats are a right-wing party too from Canadian perspective, a massive block of far-right voters like this joining Canada would be devastating.
Americans need to sort their own shit out, not come crawling to Canada to do it for them. We’ve got our own issues to deal with.


LLMs were trained on our social media feeds, after all…


Of course, my mistake. The American electorate is helpless before the power of their elected officials.


It all comes down to the electorate. America is getting the government it’s voting for.


Ah, low numbers of seeds. Must’ve just not wanted to wait.


Which, as I said, seems strange. Why don’t those businesses just download the torrents?


Seems strange. Anna’s Archive makes their collection available for bulk download as torrent files, they shouldn’t need to “cut a deal” for access to that. Just download the torrent and now you’ve got the whole collection available locally.


Only 12 percent reported both lower costs and higher revenue, while 56 percent saw neither benefit. Twenty-six percent saw reduced costs, but nearly as many experienced cost increases.
So 38% saw benefits from AI, whereas “nearly” 26% saw cost increases from it. One could just as easily write the headline “More companies experience increased benefits from AI than experience increased costs” based on this data but that headline wouldn’t get so many clicks.


Youtube isn’t the way you think it should be, though.
It’s weird how AI has turned so much of the internet from its generally anti-copyright stance. I’ve seen threads in piracy and datahoarding communities that were riddled with “won’t someone please think of the copyright!” Posts raging about how awful AI was.
I maintain the same view I always have. Copyright is indeed broken, because of how overly restrictive and expansive it has become. Most people long ago lost sight of what it’s actually for.