<?xml version="1.0" encoding="UTF-8"?><rss version="2.0" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>ai &amp;mdash; jolek78&#39;s blog</title>
    <link>https://jolek78.writeas.com/tag:ai</link>
    <description>thoughts from a friendly human being</description>
    <pubDate>Sun, 19 Apr 2026 22:30:25 +0000</pubDate>
    <image>
      <url>https://i.snap.as/DEj7yFm4.png</url>
      <title>ai &amp;mdash; jolek78&#39;s blog</title>
      <link>https://jolek78.writeas.com/tag:ai</link>
    </image>
    <item>
      <title>Anna&#39;s Archive: Robin Hood of knowledge or</title>
      <link>https://jolek78.writeas.com/annas-archive-robin-hood-of-knowledge-or?pk_campaign=rss-feed</link>
      <description>&lt;![CDATA[3:00 AM. Another one of those nights where my brain decided sleep was overrated. After my usual nocturnal walk through the streets of a remote Scottish town—where even a fox observed me with that &#34;humans are weird&#34; look—I sat back down at my server. Just a quick scan of my RSS feeds, I told myself, then I can start work. When...&#xA;&#xA;  We backed up Spotify (metadata and music files). It&#39;s distributed in bulk torrents (~300TB), grouped by popularity.&#xA;  This release includes the largest publicly available music metadata database with 256 million tracks and 186 million unique ISRCs.&#xA;  It&#39;s the world&#39;s first &#34;preservation archive&#34; for music which is fully open (meaning it can easily be mirrored by anyone with enough disk space), with 86 million music files, representing around 99.6% of listens.&#xA;&#xA;The news came from Anna&#39;s Archive—the world&#39;s largest pirate library—which had just scraped Spotify&#39;s entire catalog. Not just metadata, but also the audio files. 86 million tracks, 300 terabytes. I stopped to reread those numbers, then thought: holy shit, how big is this thing? &#xA;&#xA;!--more--&#xA;&#xA;And so, while the rest of the world slept, I started digging. This is one of those stories that needs to be told—a story weaving together hacker idealism, technology, billions of dollars in AI training data, and an ethical paradox few want to truly confront.&#xA;&#xA;When Z-Library fell&#xA;November 3, 2022. The FBI seized Z-Library&#39;s domains, one of the world&#39;s largest pirate libraries. Two alleged operators were arrested in Argentina. The community panicked—Z-Library served millions of students, researchers, and readers. And suddenly, everything vanished.&#xA;&#xA;But someone was prepared. A group called PiLiMi (Pirate Library Mirror) had created complete backups of all shadow libraries for years. LibGen, Z-Library, Sci-Hub. Everything. When Z-Library fell, these backups were ready. But there was a problem: petabytes of unusable data with no way to search them.&#xA;&#xA;Enter Anna Archivist—a pseudonym, probably a collective—who understood something fundamental: preserving data is useless if it&#39;s not accessible. Days after Z-Library&#39;s seizure, Anna&#39;s Archive was online with a meta-search engine aggregating all shadow library catalogs, making them searchable and—crucially—virtually impossible to censor.&#xA;&#xA;The numbers&#xA;December 2025:&#xA;&#xA;61.3 million books (PDF, EPUB, MOBI, DjVu)&#xA;95.5 million academic papers&#xA;256 million music tracks (Spotify metadata)&#xA;86 million audio files (~300TB)&#xA;Total: ~1.1 Petabyte in unified torrents&#xA;&#xA;To put this in perspective: the sum of all academic knowledge produced by humanity, plus a gigantic slice of world literary production, plus now music. All indexed, searchable, downloadable. Free. And virtually impossible to shut down.&#xA;&#xA;Why it can&#39;t be killed&#xA;Remember Napster? Centralized servers, one lawsuit, shut down in a day. BitTorrent learned from that—decentralized everything. But Anna&#39;s Archive goes further, combining layers of resilience that make it practically immortal:&#xA;&#xA;Distributed Frontend: Multiple domain mirrors (.li, .se, .org, .gs), Tor hidden service, Progressive Web App that works offline. Block one, others continue.&#xA;&#xA;Distributed Database: Elasticsearch + PostgreSQL + public API. Anyone can download the entire database and host their own instance. No central server to attack.&#xA;&#xA;Distributed Files: This is the genius part. Anna&#39;s Archive hosts almost nothing directly. Instead:&#xA;&#xA;IPFS (InterPlanetary File System): Files identified by cryptographic hash, served by volunteer nodes worldwide&#xA;BitTorrent: Classic torrents with multiple trackers, self-sustaining swarms&#xA;HTTP Gateways: For normal users who just want to click-and-download, links redirect to public IPFS gateways&#xA;&#xA;Result: user downloads via normal HTTP, but content comes from a decentralized network. Can&#39;t shut down IPFS. Can&#39;t stop BitTorrent. Can block gateways, but hundreds exist and anyone can create new ones.&#xA;&#xA;OpSec: Domains registered via privacy-focused Icelandic registrar, bulletproof hosting in non-cooperative jurisdictions, Bitcoin payments, PGP-encrypted communications, zero personal information.&#xA;&#xA;The only way to stop Anna&#39;s Archive would be to shut down the internet. Or convince every single seeder to stop. Good luck.&#xA;&#xA;81.7 terabytes free for meta&#xA;And here&#39;s where it gets disturbing.&#xA;&#xA;February 2025. Documents from Kadrey v. Meta are unsealed—a class action by authors against Meta for using their pirated books to train Llama AI models. Internal emails reveal a shocking timeline:&#xA;&#xA;October 2022 - Melanie Kambadur, Senior Research Manager:&#xA;&#xA;  I don&#39;t think we should use pirated material. I really need to draw a line there.&#xA;&#xA;Eleonora Presani, Meta employee:&#xA;&#xA;  Using pirated material should be beyond our ethical threshold. SciHub, ResearchGate, LibGen are basically like PirateBay... they&#39;re distributing content that is protected by copyright and they&#39;re infringing it.&#xA;&#xA;January 2023 - Meeting with Mark Zuckerberg present:&#xA;&#xA;  [Zuckerberg] wants to move this stuff forward, and we need to find a way to unblock all this.&#xA;&#xA;April 2023 - Nikolay Bashlykov, Meta engineer:&#xA;&#xA;  Using Meta IP addresses to load through torrents pirate content... torrenting from a corporate laptop doesn&#39;t feel right.&#xA;&#xA;2023-2024: The Operation&#xA;&#xA;Meta downloaded:&#xA;&#xA;81.7 TB via Anna&#39;s Archive torrents (35.7 TB from Z-Library alone)&#xA;80.6 TB from LibGen&#xA;Total: ~162 TB of pirated books&#xA;&#xA;Method: BitTorrent client on separate infrastructure, VPN to obscure origin, active seeding to other peers. Result: 197,000 copyrighted books integrated into Llama training data.&#xA;&#xA;June 2025: the ruling&#xA;Judge Vince Chhabria (Northern District California) applied the four-factor fair use test. The decision is legally fascinating and ethically disturbing.&#xA;&#xA;Factor 1 - Transformative Use: Meta wins decisively. The judge ruled AI training is &#34;spectacularly transformative&#34;—fundamentally different from human reading. The purpose isn&#39;t to express the content but to learn statistical relationships between words.&#xA;&#xA;Factor 2 - Nature of Work: Neutral. Creative fiction gets more copyright protection than factual works, but this didn&#39;t tip the scales either way.&#xA;&#xA;Factor 3 - Amount Used: Meta wins. Even though they used entire books, the judge found this necessary for training. You can&#39;t cherry-pick sentences and expect an AI to learn language patterns.&#xA;&#xA;Factor 4 - Market Effect: This is where the judge&#39;s discomfort shows through:&#xA;&#xA;  Generative AI has the potential to flood the market with endless amounts of images, songs, articles, books... So by training generative AI models with copyrighted works, companies are creating something that often will dramatically undermine the market for those works, and thus dramatically undermine the incentive for human beings to create things the old-fashioned way.&#xA;&#xA;He sees the problem clearly. AI trained on copyrighted works will compete with and potentially destroy the market for those very works. But the plaintiffs couldn&#39;t prove specific economic harm with hard data.&#xA;&#xA;The final ruling: &#34;Given the state of the record, the Court has no choice but to grant summary judgment.&#34; Meta wins on these specific facts. But the judge adds a critical caveat: &#34;In most cases, training LLMs on copyrighted works without permission is likely infringing and not fair use.&#34;&#xA;&#xA;Meta didn&#39;t win because what they did was legitimate. They won because the authors&#39; lawyers didn&#39;t build a strong enough evidentiary case. It&#39;s a technical legal victory that sidesteps the ethical question entirely.&#xA;&#xA;The precedent this sets is chilling: AI companies can pirate with relative impunity if they have good lawyers and plaintiffs can&#39;t prove specific damages.&#xA;&#xA;The math&#xA;Scenario A (legal):&#xA;&#xA;Meta negotiates licenses with publishers&#xA;Cost: $50-100 million (conservative estimate)&#xA;Authors receive royalties&#xA;&#xA;Scenario B (what they did):&#xA;&#xA;Download 81.7 TB for free&#xA;Legal defense: ~$5 million&#xA;Win in court&#xA;Authors receive: $0&#xA;&#xA;Meta&#39;s savings: $45-95 million&#xA;&#xA;And now every AI company knows: download from Anna&#39;s Archive, risk a lawsuit with weak evidence, save tens of millions.&#xA;&#xA;Anna&#39;s Archive also revealed they provide &#34;SFTP bulk access to approximately 30 companies&#34;—primarily Chinese LLM startups and data brokers—who contribute money or data. DeepSeek publicly admitted using Anna&#39;s Archive data for training. No consequences in Chinese jurisdiction.&#xA;&#xA;Aaron Swartz and the question that haunts this story&#xA;There&#39;s a ghost here. His name is Aaron Swartz, and his story illuminates everything wrong with how we treat information access.&#xA;&#xA;2011: Aaron, 24, brilliant programmer, Reddit co-founder, and information freedom activist, connected to MIT&#39;s network and downloaded 4.8 million academic papers from JSTOR. His intent was to make publicly-funded research freely available. He wasn&#39;t enriching himself. He was acting on principle.&#xA;&#xA;The response was swift and brutal. Federal prosecutors threw the book at him: 13 felony charges, maximum penalty of 50 years in prison and $1 million in fines. For downloading academic papers. The prosecution was led by U.S. Attorney Carmen Ortiz, who called it &#34;stealing is stealing, whether you use a computer command or a crowbar.&#34;&#xA;&#xA;The pressure was immense. Aaron faced financial ruin, decades in prison, complete destruction of his life. In January 2013, at age 26, he hanged himself. His family and partner blamed the aggressive prosecution. The internet mourned a brilliant mind and passionate advocate crushed by prosecutorial overreach.&#xA;&#xA;Now consider the parallel:&#xA;&#xA;Aaron Swartz: 4.8 million papers → federal persecution, suicide at 26&#xA;&#xA;Meta: 162 TB (~162 million papers) → wins in court, saves $95 million&#xA;&#xA;Aaron was an individual acting on idealistic principles about information freedom. Meta is a trillion-dollar corporation acting on profit motives. Aaron faced the full weight of federal prosecution. Meta faced a civil lawsuit they successfully defended with their massive legal team.&#xA;&#xA;The system punishes idealism and rewards profit. The disparity isn&#39;t just unjust—it reveals something fundamental about who gets to break rules and who doesn&#39;t.&#xA;&#xA;The paradox no one wants to see&#xA;Anna&#39;s Archive claims to fight publishing monopolies and inequality in access to knowledge. But the reality:&#xA;&#xA;Who benefits most?&#xA;&#xA;Meta: 81.7 TB free, $95M saved&#xA;~30 AI companies: privileged access&#xA;Corporations with $100M+ compute budgets&#xA;&#xA;Resources needed to benefit:&#xA;&#xA;Storage/Bandwidth: trivial for Meta ($1000s)&#xA;Computing for training: MASSIVE ($10-100M)&#xA;Legal defense: MASSIVE ($millions)&#xA;&#xA;Only big tech can afford this. The result:&#xA;&#xA;Data: socialized (Anna&#39;s Archive, shared risk)&#xA;Profits: privatized (proprietary LLMs, paid APIs)&#xA;Costs: externalized (authors not compensated)&#xA;&#xA;But what about students in the Global South?&#xA;&#xA;This is where the story gets complicated, because the benefits are real and they matter immensely.&#xA;&#xA;Consider a medical student in India. Her family earns about $400/month. A single medical textbook costs $300-500. She needs fifteen of them. The math is impossible. Her options: don&#39;t graduate, or Anna&#39;s Archive. She chose the latter and completed her degree. She&#39;s now a practicing physician.&#xA;&#xA;Or take a PhD researcher in South Africa studying climate change impacts. The critical papers for his dissertation are behind Elsevier&#39;s paywall at $35 each. He needs twenty papers minimum—$700 his university can&#39;t afford. Without Sci-Hub (accessible through Anna&#39;s Archive), his dissertation would have been impossible. He completed it, published findings that inform local climate policy.&#xA;&#xA;An art history teacher in Argentina wanted to enrich her curriculum with Renaissance art analysis. The books she needed weren&#39;t available in local libraries. Importing them? Prohibitive between shipping costs and customs. Anna&#39;s Archive gave her access to rare texts that transformed her teaching.&#xA;&#xA;The data backs this up: literature review times for researchers in developing countries reduced 60-80%. Citation patterns show researchers in Nigeria, Bangladesh, Ecuador now cite contemporary research at parity with Harvard and Oxford. Publications from developing countries have increased. Methodological quality has improved. International collaborations have expanded.&#xA;&#xA;This matters. This changes lives. This is not hypothetical.&#xA;&#xA;The problem is: both things are simultaneously true.&#xA;&#xA;Anna&#39;s Archive saves academic careers in the Global South&#xA;Anna&#39;s Archive allows Meta to save $95 million&#xA;&#xA;But Meta downloaded more data in one week than all Indian students download in a year. How do we square that?&#xA;&#xA;The broken system that created this monster&#xA;To understand why Anna&#39;s Archive exists and why it&#39;s grown so explosively, you need to understand how fundamentally broken academic publishing has become.&#xA;&#xA;Here&#39;s the perverse cycle:&#xA;&#xA;Researcher writes paper (unpaid)&#xA;Other researchers peer review it (unpaid)&#xA;Publisher publishes it&#xA;Researcher&#39;s own university must pay to read it&#xA;Publisher profits: Elsevier and Wiley report 35-40% profit margins&#xA;&#xA;Today, over 70% of academic papers sit behind paywalls. Access costs $35-50 per paper for individuals, or $10,000-100,000+ per year for institutional subscriptions. Universities in developing countries simply cannot afford these subscriptions. Neither can most universities in developed countries—Harvard famously called journal subscription costs &#34;fiscally unsustainable&#34; in 2012.&#xA;&#xA;The system extracts free labor from researchers, locks up publicly-funded research behind paywalls, charges exorbitant fees to access it, and funnels enormous profits to publishers who add relatively little value. Academic institutions create the knowledge, do the quality control, and then pay again to access their own work.&#xA;&#xA;Sci-Hub and Anna&#39;s Archive didn&#39;t emerge from nowhere. They&#39;re responses to a genuinely broken system. The question is whether they&#39;re the right response—and who ultimately benefits most from that response.&#xA;&#xA;The architecture determines the ethics&#xA;Anna&#39;s Archive can&#39;t discriminate because:&#xA;&#xA;Open source philosophy: everyone or no one&#xA;Technical impossibility: how do you block Meta but not students?&#xA;Legal strategy: claiming &#34;non-hosting&#34; makes usage control impossible&#xA;&#xA;IPFS and BitTorrent are magnificent tools for resisting censorship. But resistance to censorship also means resistance to ethical control. You can&#39;t have one without the other.&#xA;&#xA;The system is structurally designed to be unkillable. Which also means it&#39;s structurally designed to serve whoever has the resources to benefit most.&#xA;&#xA;Where does it end?&#xA;December 2025: Anna&#39;s Archive announced they&#39;d scraped Spotify. The same preservation narrative, the same pattern. 256 million tracks, 86 million audio files, 300TB available to anyone with the infrastructure to use it.&#xA;&#xA;&#34;This Spotify scrape is our humble attempt to start such a &#39;preservation archive&#39; for music,&#34; they wrote. The justification mirrors the books argument: Spotify loses licenses, music disappears; platform risk if Spotify fails; regional blocks prevent access; long tail poorly preserved.&#xA;&#xA;All true. But who downloads 300TB of music? Not the kid in Malawi who just wants to listen to his favorite artist. ByteDance, training the next AI music generator. Startups building Spotify competitors. The same companies with compute budgets in the tens of millions.&#xA;&#xA;Anna&#39;s Archive is pivoting from text to multimedia, and each escalation follows a predictable pattern:&#xA;&#xA;Books → Justified by paywalls and academic access&#xA;Papers → Justified by broken academic publishing&#xA;Music → Justified by platform risk and preservation&#xA;Video? → What&#39;s the justification for the next step?&#xA;&#xA;With each escalation:&#xA;&#xA;The value for big tech increases exponentially&#xA;The proportion of benefit for individual students decreases&#xA;Mass piracy becomes normalized as &#34;preservation&#34;&#xA;The ethical questions get harder to answer&#xA;&#xA;And the international precedent is already being set. Japan&#39;s AI Minister (January 2025) stated explicitly: &#34;AI companies in Japan can use whatever they want for AI training... whether it is content obtained from illegal sites or otherwise.&#34;&#xA;&#xA;The message from governments: pirate freely if it serves AI supremacy. We&#39;re in a race to the bottom where copyright becomes meaningless for AI training, and the companies with the most resources benefit most.&#xA;&#xA;Conclusions: I don&#39;t know which way to turn&#xA;I started from that sleepless night, 256 million songs in an RSS feed, and ended up here with more questions than answers.&#xA;&#xA;Anna&#39;s Archive is a technological marvel—IPFS, BitTorrent, distributed databases creating something genuinely uncensorable. It&#39;s also a lifeline for millions of students and researchers locked out of knowledge by an exploitative publishing system. And simultaneously, it&#39;s the largest intellectual property expropriation operation in history, saving corporations hundreds of millions while creators receive nothing.&#xA;&#xA;All of these things are true at once. This isn&#39;t a simple story with heroes and villains.&#xA;&#xA;The academic publishing system is genuinely broken. Researchers create knowledge for free, review it for free, then their institutions must pay exorbitant fees to access it while publishers extract 35-40% profit margins. This system deserves to be disrupted.&#xA;&#xA;But Anna&#39;s Archive isn&#39;t disrupting it equitably. The architecture that makes it uncensorable also makes it impossible to distinguish between a student in Lagos accessing a textbook and Meta downloading 162TB for AI training. You can&#39;t have selective resistance to censorship—it&#39;s all or nothing.&#xA;&#xA;Aaron Swartz died fighting for information freedom with idealistic principles. Meta achieves the same result with corporate profit motives and walks away victorious. The system rewards power and punishes principle.&#xA;&#xA;Can this be fixed? Copyright reform moves at the speed of politics—years, decades. Compulsory licensing for AI training? Just beginning to be discussed. Open access mandates? Facing massive publisher resistance. Meanwhile, Anna&#39;s Archive operates at the speed of software, and data flows freely to those with $100M compute clusters.&#xA;&#xA;The question isn&#39;t whether Anna&#39;s Archive will be stopped—it won&#39;t be, that&#39;s the point of the architecture. The question is what world we&#39;re building where the same technology that liberates a medical student in India also bankrolls Meta&#39;s AI ambitions, and we can&#39;t separate one from the other.&#xA;&#xA;I don&#39;t have answers. I have a functioning IPFS node, a Tor relay, and the uncomfortable knowledge that every byte I help distribute might be saving a researcher&#39;s career or training someone&#39;s proprietary AI model. Probably both.&#xA;&#xA;Free for everyone. The problem is that &#34;everyone&#34; has very different resources to benefit from that freedom.&#xA;&#xA;Now, if you&#39;ll excuse me, I&#39;m going to check how much bandwidth my nodes are using. And reflect on whether participation is complicity or resistance. Maybe it&#39;s both. Maybe that&#39;s the point.&#xA;&#xA;a href=&#34;https://remark.as/p/jolek78/annas-archive-robin-hood-of-knowledge-or&#34;Discuss.../a&#xA;&#xA;#AnnaArchive #AI #Copyright #AaronSwartz #Meta #AcademicPublishing #IPFS #InformationFreedom&#xA;&#xA;div class=&#34;center&#34;&#xD;&#xA;· 🦣 a href=&#34;https://fosstodon.org/@jolek78&#34;Mastodon/a · 📸 a href=&#34;https://pixelfed.social/jolek78&#34;Pixelfed/a ·  📬 a href=&#34;mailto:jolek78@jolek78.dev&#34;Email/a ·&#xD;&#xA;· ☕ a href=&#34;https://liberapay.com/jolek78&#34;Support this work on Liberapay/a&#xD;&#xA;/div]]&gt;</description>
      <content:encoded><![CDATA[<p>3:00 AM. Another one of those nights where my brain decided sleep was overrated. After my usual nocturnal walk through the streets of a remote Scottish town—where even a fox observed me with that “humans are weird” look—I sat back down at my server. Just a quick scan of my RSS feeds, I told myself, then I can start work. When...</p>

<blockquote><p>We backed up Spotify (metadata and music files). It&#39;s distributed in bulk torrents (~300TB), grouped by popularity.
This release includes the largest publicly available music metadata database with 256 million tracks and 186 million unique ISRCs.
It&#39;s the world&#39;s first “preservation archive” for music which is fully open (meaning it can easily be mirrored by anyone with enough disk space), with 86 million music files, representing around 99.6% of listens.</p></blockquote>

<p>The news came from <a href="https://annas-archive.li/">Anna&#39;s Archive</a>—the world&#39;s largest pirate library—which had just scraped Spotify&#39;s entire catalog. Not just metadata, but also the audio files. 86 million tracks, 300 terabytes. I stopped to reread those numbers, then thought: holy shit, how big is this thing?</p>



<p>And so, while the rest of the world slept, I started digging. This is one of those stories that needs to be told—a story weaving together hacker idealism, technology, billions of dollars in AI training data, and an ethical paradox few want to truly confront.</p>

<h3 id="when-z-library-fell" id="when-z-library-fell">When Z-Library fell</h3>

<p>November 3, 2022. The FBI seized Z-Library&#39;s domains, one of the world&#39;s largest pirate libraries. Two alleged operators were arrested in Argentina. The community panicked—Z-Library served millions of students, researchers, and readers. And suddenly, everything vanished.</p>

<p>But someone was prepared. A group called PiLiMi (Pirate Library Mirror) had created complete backups of all shadow libraries for years. LibGen, Z-Library, Sci-Hub. Everything. When Z-Library fell, these backups were ready. But there was a problem: petabytes of unusable data with no way to search them.</p>

<p>Enter Anna Archivist—a pseudonym, probably a collective—who understood something fundamental: preserving data is useless if it&#39;s not accessible. Days after Z-Library&#39;s seizure, Anna&#39;s Archive was online with a meta-search engine aggregating all shadow library catalogs, making them searchable and—crucially—virtually impossible to censor.</p>

<h3 id="the-numbers" id="the-numbers">The numbers</h3>

<p>December 2025:</p>
<ul><li>61.3 million books (PDF, EPUB, MOBI, DjVu)</li>
<li>95.5 million academic papers</li>
<li>256 million music tracks (Spotify metadata)</li>
<li>86 million audio files (~300TB)</li>
<li>Total: ~1.1 Petabyte in unified torrents</li></ul>

<p>To put this in perspective: the sum of all academic knowledge produced by humanity, plus a gigantic slice of world literary production, plus now music. All indexed, searchable, downloadable. Free. And virtually impossible to shut down.</p>

<h3 id="why-it-can-t-be-killed" id="why-it-can-t-be-killed">Why it can&#39;t be killed</h3>

<p>Remember Napster? Centralized servers, one lawsuit, shut down in a day. BitTorrent learned from that—decentralized everything. But Anna&#39;s Archive goes further, combining layers of resilience that make it practically immortal:</p>

<p><strong>Distributed Frontend:</strong> Multiple domain mirrors (.li, .se, .org, .gs), Tor hidden service, Progressive Web App that works offline. Block one, others continue.</p>

<p><strong>Distributed Database:</strong> Elasticsearch + PostgreSQL + public API. Anyone can download the entire database and host their own instance. No central server to attack.</p>

<p><strong>Distributed Files:</strong> This is the genius part. Anna&#39;s Archive hosts almost nothing directly. Instead:</p>
<ul><li>IPFS (InterPlanetary File System): Files identified by cryptographic hash, served by volunteer nodes worldwide</li>
<li>BitTorrent: Classic torrents with multiple trackers, self-sustaining swarms</li>
<li>HTTP Gateways: For normal users who just want to click-and-download, links redirect to public IPFS gateways</li></ul>

<p>Result: user downloads via normal HTTP, but content comes from a decentralized network. Can&#39;t shut down IPFS. Can&#39;t stop BitTorrent. Can block gateways, but hundreds exist and anyone can create new ones.</p>

<p><strong>OpSec:</strong> Domains registered via privacy-focused Icelandic registrar, bulletproof hosting in non-cooperative jurisdictions, Bitcoin payments, PGP-encrypted communications, zero personal information.</p>

<p>The only way to stop Anna&#39;s Archive would be to shut down the internet. Or convince every single seeder to stop. Good luck.</p>

<h3 id="81-7-terabytes-free-for-meta" id="81-7-terabytes-free-for-meta">81.7 terabytes free for meta</h3>

<p>And here&#39;s where it gets disturbing.</p>

<p>February 2025. Documents from <em>Kadrey v. Meta</em> are unsealed—a class action by authors against Meta for using their pirated books to train Llama AI models. Internal emails reveal a shocking timeline:</p>

<p><strong>October 2022</strong> – Melanie Kambadur, Senior Research Manager:</p>

<blockquote><p>I don&#39;t think we should use pirated material. I really need to draw a line there.</p></blockquote>

<p>Eleonora Presani, Meta employee:</p>

<blockquote><p>Using pirated material should be beyond our ethical threshold. SciHub, ResearchGate, LibGen are basically like PirateBay... they&#39;re distributing content that is protected by copyright and they&#39;re infringing it.</p></blockquote>

<p><strong>January 2023</strong> – Meeting with Mark Zuckerberg present:</p>

<blockquote><p>[Zuckerberg] wants to move this stuff forward, and we need to find a way to unblock all this.</p></blockquote>

<p><strong>April 2023</strong> – Nikolay Bashlykov, Meta engineer:</p>

<blockquote><p>Using Meta IP addresses to load through torrents pirate content... torrenting from a corporate laptop doesn&#39;t feel right.</p></blockquote>

<p><strong>2023-2024: The Operation</strong></p>

<p>Meta downloaded:</p>
<ul><li>81.7 TB via Anna&#39;s Archive torrents (35.7 TB from Z-Library alone)</li>
<li>80.6 TB from LibGen</li>
<li>Total: ~162 TB of pirated books</li></ul>

<p>Method: BitTorrent client on separate infrastructure, VPN to obscure origin, active seeding to other peers. Result: 197,000 copyrighted books integrated into Llama training data.</p>

<h3 id="june-2025-the-ruling" id="june-2025-the-ruling">June 2025: the ruling</h3>

<p>Judge Vince Chhabria (Northern District California) applied the four-factor fair use test. The decision is legally fascinating and ethically disturbing.</p>

<p><strong>Factor 1 – Transformative Use:</strong> Meta wins decisively. The judge ruled AI training is “spectacularly transformative”—fundamentally different from human reading. The purpose isn&#39;t to express the content but to learn statistical relationships between words.</p>

<p><strong>Factor 2 – Nature of Work:</strong> Neutral. Creative fiction gets more copyright protection than factual works, but this didn&#39;t tip the scales either way.</p>

<p><strong>Factor 3 – Amount Used:</strong> Meta wins. Even though they used entire books, the judge found this necessary for training. You can&#39;t cherry-pick sentences and expect an AI to learn language patterns.</p>

<p><strong>Factor 4 – Market Effect:</strong> This is where the judge&#39;s discomfort shows through:</p>

<blockquote><p>Generative AI has the potential to flood the market with endless amounts of images, songs, articles, books... So by training generative AI models with copyrighted works, companies are creating something that often will dramatically undermine the market for those works, and thus dramatically undermine the incentive for human beings to create things the old-fashioned way.</p></blockquote>

<p>He sees the problem clearly. AI trained on copyrighted works will compete with and potentially destroy the market for those very works. But the plaintiffs couldn&#39;t prove specific economic harm with hard data.</p>

<p>The final ruling: “Given the state of the record, the Court has no choice but to grant summary judgment.” Meta wins on these specific facts. But the judge adds a critical caveat: “In most cases, training LLMs on copyrighted works without permission is likely infringing and not fair use.”</p>

<p>Meta didn&#39;t win because what they did was legitimate. They won because the authors&#39; lawyers didn&#39;t build a strong enough evidentiary case. It&#39;s a technical legal victory that sidesteps the ethical question entirely.</p>

<p>The precedent this sets is chilling: AI companies can pirate with relative impunity if they have good lawyers and plaintiffs can&#39;t prove specific damages.</p>

<h3 id="the-math" id="the-math">The math</h3>

<p><strong>Scenario A (legal):</strong></p>
<ul><li>Meta negotiates licenses with publishers</li>
<li>Cost: $50-100 million (conservative estimate)</li>
<li>Authors receive royalties</li></ul>

<p><strong>Scenario B (what they did):</strong></p>
<ul><li>Download 81.7 TB for free</li>
<li>Legal defense: ~$5 million</li>
<li>Win in court</li>
<li>Authors receive: $0</li></ul>

<p><strong>Meta&#39;s savings: $45-95 million</strong></p>

<p>And now every AI company knows: download from Anna&#39;s Archive, risk a lawsuit with weak evidence, save tens of millions.</p>

<p>Anna&#39;s Archive also revealed they provide “SFTP bulk access to approximately 30 companies”—primarily Chinese LLM startups and data brokers—who contribute money or data. DeepSeek publicly admitted using Anna&#39;s Archive data for training. No consequences in Chinese jurisdiction.</p>

<h3 id="aaron-swartz-and-the-question-that-haunts-this-story" id="aaron-swartz-and-the-question-that-haunts-this-story">Aaron Swartz and the question that haunts this story</h3>

<p>There&#39;s a ghost here. His name is Aaron Swartz, and his story illuminates everything wrong with how we treat information access.</p>

<p>2011: Aaron, 24, brilliant programmer, Reddit co-founder, and information freedom activist, connected to MIT&#39;s network and downloaded 4.8 million academic papers from JSTOR. His intent was to make publicly-funded research freely available. He wasn&#39;t enriching himself. He was acting on principle.</p>

<p>The response was swift and brutal. Federal prosecutors threw the book at him: 13 felony charges, maximum penalty of 50 years in prison and $1 million in fines. For downloading academic papers. The prosecution was led by U.S. Attorney Carmen Ortiz, who called it “stealing is stealing, whether you use a computer command or a crowbar.”</p>

<p>The pressure was immense. Aaron faced financial ruin, decades in prison, complete destruction of his life. In January 2013, at age 26, he hanged himself. His family and partner blamed the aggressive prosecution. The internet mourned a brilliant mind and passionate advocate crushed by prosecutorial overreach.</p>

<p>Now consider the parallel:</p>

<p><strong>Aaron Swartz: 4.8 million papers → federal persecution, suicide at 26</strong></p>

<p><strong>Meta: 162 TB (~162 million papers) → wins in court, saves $95 million</strong></p>

<p>Aaron was an individual acting on idealistic principles about information freedom. Meta is a trillion-dollar corporation acting on profit motives. Aaron faced the full weight of federal prosecution. Meta faced a civil lawsuit they successfully defended with their massive legal team.</p>

<p>The system punishes idealism and rewards profit. The disparity isn&#39;t just unjust—it reveals something fundamental about who gets to break rules and who doesn&#39;t.</p>

<h3 id="the-paradox-no-one-wants-to-see" id="the-paradox-no-one-wants-to-see">The paradox no one wants to see</h3>

<p>Anna&#39;s Archive claims to fight publishing monopolies and inequality in access to knowledge. But the reality:</p>

<p><strong>Who benefits most?</strong></p>
<ul><li>Meta: 81.7 TB free, $95M saved</li>
<li>~30 AI companies: privileged access</li>
<li>Corporations with $100M+ compute budgets</li></ul>

<p><strong>Resources needed to benefit:</strong></p>
<ul><li>Storage/Bandwidth: trivial for Meta ($1000s)</li>
<li>Computing for training: MASSIVE ($10-100M)</li>
<li>Legal defense: MASSIVE ($millions)</li></ul>

<p>Only big tech can afford this. The result:</p>
<ul><li>Data: socialized (Anna&#39;s Archive, shared risk)</li>
<li>Profits: privatized (proprietary LLMs, paid APIs)</li>
<li>Costs: externalized (authors not compensated)</li></ul>

<p><strong>But what about students in the Global South?</strong></p>

<p>This is where the story gets complicated, because the benefits are real and they matter immensely.</p>

<p>Consider a medical student in India. Her family earns about $400/month. A single medical textbook costs $300-500. She needs fifteen of them. The math is impossible. Her options: don&#39;t graduate, or Anna&#39;s Archive. She chose the latter and completed her degree. She&#39;s now a practicing physician.</p>

<p>Or take a PhD researcher in South Africa studying climate change impacts. The critical papers for his dissertation are behind Elsevier&#39;s paywall at $35 each. He needs twenty papers minimum—$700 his university can&#39;t afford. Without Sci-Hub (accessible through Anna&#39;s Archive), his dissertation would have been impossible. He completed it, published findings that inform local climate policy.</p>

<p>An art history teacher in Argentina wanted to enrich her curriculum with Renaissance art analysis. The books she needed weren&#39;t available in local libraries. Importing them? Prohibitive between shipping costs and customs. Anna&#39;s Archive gave her access to rare texts that transformed her teaching.</p>

<p>The data backs this up: literature review times for researchers in developing countries reduced 60-80%. Citation patterns show researchers in Nigeria, Bangladesh, Ecuador now cite contemporary research at parity with Harvard and Oxford. Publications from developing countries have increased. Methodological quality has improved. International collaborations have expanded.</p>

<p>This matters. This changes lives. This is not hypothetical.</p>

<p>The problem is: <em>both things are simultaneously true.</em></p>
<ol><li>Anna&#39;s Archive saves academic careers in the Global South</li>
<li>Anna&#39;s Archive allows Meta to save $95 million</li></ol>

<p>But Meta downloaded more data in one week than all Indian students download in a year. How do we square that?</p>

<h3 id="the-broken-system-that-created-this-monster" id="the-broken-system-that-created-this-monster">The broken system that created this monster</h3>

<p>To understand why Anna&#39;s Archive exists and why it&#39;s grown so explosively, you need to understand how fundamentally broken academic publishing has become.</p>

<p>Here&#39;s the perverse cycle:</p>
<ol><li>Researcher writes paper (unpaid)</li>
<li>Other researchers peer review it (unpaid)</li>
<li>Publisher publishes it</li>
<li>Researcher&#39;s own university must pay to read it</li>
<li>Publisher profits: Elsevier and Wiley report 35-40% profit margins</li></ol>

<p>Today, over 70% of academic papers sit behind paywalls. Access costs $35-50 per paper for individuals, or $10,000-100,000+ per year for institutional subscriptions. Universities in developing countries simply cannot afford these subscriptions. Neither can most universities in developed countries—Harvard famously called journal subscription costs “fiscally unsustainable” in 2012.</p>

<p>The system extracts free labor from researchers, locks up publicly-funded research behind paywalls, charges exorbitant fees to access it, and funnels enormous profits to publishers who add relatively little value. Academic institutions create the knowledge, do the quality control, and then pay again to access their own work.</p>

<p>Sci-Hub and Anna&#39;s Archive didn&#39;t emerge from nowhere. They&#39;re responses to a genuinely broken system. The question is whether they&#39;re the right response—and who ultimately benefits most from that response.</p>

<h3 id="the-architecture-determines-the-ethics" id="the-architecture-determines-the-ethics">The architecture determines the ethics</h3>

<p>Anna&#39;s Archive can&#39;t discriminate because:</p>
<ol><li>Open source philosophy: everyone or no one</li>
<li>Technical impossibility: how do you block Meta but not students?</li>
<li>Legal strategy: claiming “non-hosting” makes usage control impossible</li></ol>

<p>IPFS and BitTorrent are magnificent tools for resisting censorship. But resistance to censorship also means resistance to ethical control. You can&#39;t have one without the other.</p>

<p>The system is structurally designed to be unkillable. Which also means it&#39;s structurally designed to serve whoever has the resources to benefit most.</p>

<h3 id="where-does-it-end" id="where-does-it-end">Where does it end?</h3>

<p>December 2025: Anna&#39;s Archive announced they&#39;d scraped Spotify. The same preservation narrative, the same pattern. 256 million tracks, 86 million audio files, 300TB available to anyone with the infrastructure to use it.</p>

<p>“This Spotify scrape is our humble attempt to start such a &#39;preservation archive&#39; for music,” they wrote. The justification mirrors the books argument: Spotify loses licenses, music disappears; platform risk if Spotify fails; regional blocks prevent access; long tail poorly preserved.</p>

<p>All true. But who downloads 300TB of music? Not the kid in Malawi who just wants to listen to his favorite artist. ByteDance, training the next AI music generator. Startups building Spotify competitors. The same companies with compute budgets in the tens of millions.</p>

<p>Anna&#39;s Archive is pivoting from text to multimedia, and each escalation follows a predictable pattern:</p>
<ul><li><strong>Books</strong> → Justified by paywalls and academic access</li>
<li><strong>Papers</strong> → Justified by broken academic publishing</li>
<li><strong>Music</strong> → Justified by platform risk and preservation</li>
<li><strong>Video?</strong> → What&#39;s the justification for the next step?</li></ul>

<p>With each escalation:</p>
<ul><li>The value for big tech increases exponentially</li>
<li>The proportion of benefit for individual students decreases</li>
<li>Mass piracy becomes normalized as “preservation”</li>
<li>The ethical questions get harder to answer</li></ul>

<p>And the international precedent is already being set. Japan&#39;s AI Minister (January 2025) stated explicitly: “AI companies in Japan can use whatever they want for AI training... whether it is content obtained from illegal sites or otherwise.”</p>

<p>The message from governments: pirate freely if it serves AI supremacy. We&#39;re in a race to the bottom where copyright becomes meaningless for AI training, and the companies with the most resources benefit most.</p>

<h3 id="conclusions-i-don-t-know-which-way-to-turn" id="conclusions-i-don-t-know-which-way-to-turn">Conclusions: I don&#39;t know which way to turn</h3>

<p>I started from that sleepless night, 256 million songs in an RSS feed, and ended up here with more questions than answers.</p>

<p>Anna&#39;s Archive is a technological marvel—IPFS, BitTorrent, distributed databases creating something genuinely uncensorable. It&#39;s also a lifeline for millions of students and researchers locked out of knowledge by an exploitative publishing system. And simultaneously, it&#39;s the largest intellectual property expropriation operation in history, saving corporations hundreds of millions while creators receive nothing.</p>

<p>All of these things are true at once. This isn&#39;t a simple story with heroes and villains.</p>

<p>The academic publishing system is genuinely broken. Researchers create knowledge for free, review it for free, then their institutions must pay exorbitant fees to access it while publishers extract 35-40% profit margins. This system deserves to be disrupted.</p>

<p>But Anna&#39;s Archive isn&#39;t disrupting it equitably. The architecture that makes it uncensorable also makes it impossible to distinguish between a student in Lagos accessing a textbook and Meta downloading 162TB for AI training. You can&#39;t have selective resistance to censorship—it&#39;s all or nothing.</p>

<p>Aaron Swartz died fighting for information freedom with idealistic principles. Meta achieves the same result with corporate profit motives and walks away victorious. The system rewards power and punishes principle.</p>

<p>Can this be fixed? Copyright reform moves at the speed of politics—years, decades. Compulsory licensing for AI training? Just beginning to be discussed. Open access mandates? Facing massive publisher resistance. Meanwhile, Anna&#39;s Archive operates at the speed of software, and data flows freely to those with $100M compute clusters.</p>

<p>The question isn&#39;t whether Anna&#39;s Archive will be stopped—it won&#39;t be, that&#39;s the point of the architecture. The question is what world we&#39;re building where the same technology that liberates a medical student in India also bankrolls Meta&#39;s AI ambitions, and we can&#39;t separate one from the other.</p>

<p>I don&#39;t have answers. I have a functioning IPFS node, a Tor relay, and the uncomfortable knowledge that every byte I help distribute might be saving a researcher&#39;s career or training someone&#39;s proprietary AI model. Probably both.</p>

<p>Free for everyone. The problem is that “everyone” has very different resources to benefit from that freedom.</p>

<p>Now, if you&#39;ll excuse me, I&#39;m going to check how much bandwidth my nodes are using. And reflect on whether participation is complicity or resistance. Maybe it&#39;s both. Maybe that&#39;s the point.</p>

<p><a href="https://remark.as/p/jolek78/annas-archive-robin-hood-of-knowledge-or">Discuss...</a></p>

<p><a href="https://jolek78.writeas.com/tag:AnnaArchive" class="hashtag"><span>#</span><span class="p-category">AnnaArchive</span></a> <a href="https://jolek78.writeas.com/tag:AI" class="hashtag"><span>#</span><span class="p-category">AI</span></a> <a href="https://jolek78.writeas.com/tag:Copyright" class="hashtag"><span>#</span><span class="p-category">Copyright</span></a> <a href="https://jolek78.writeas.com/tag:AaronSwartz" class="hashtag"><span>#</span><span class="p-category">AaronSwartz</span></a> <a href="https://jolek78.writeas.com/tag:Meta" class="hashtag"><span>#</span><span class="p-category">Meta</span></a> <a href="https://jolek78.writeas.com/tag:AcademicPublishing" class="hashtag"><span>#</span><span class="p-category">AcademicPublishing</span></a> <a href="https://jolek78.writeas.com/tag:IPFS" class="hashtag"><span>#</span><span class="p-category">IPFS</span></a> <a href="https://jolek78.writeas.com/tag:InformationFreedom" class="hashtag"><span>#</span><span class="p-category">InformationFreedom</span></a></p>

<div class="center">
· 🦣 <a href="https://fosstodon.org/@jolek78">Mastodon</a> · 📸 <a href="https://pixelfed.social/jolek78">Pixelfed</a> ·  📬 <a href="mailto:jolek78@jolek78.dev">Email</a> ·
· ☕ <a href="https://liberapay.com/jolek78">Support this work on Liberapay</a>
</div>
]]></content:encoded>
      <guid>https://jolek78.writeas.com/annas-archive-robin-hood-of-knowledge-or</guid>
      <pubDate>Mon, 29 Dec 2025 14:24:51 +0000</pubDate>
    </item>
    <item>
      <title>ChatGPT didn&#39;t invent anything.</title>
      <link>https://jolek78.writeas.com/chatgpt-didnt-invent-anything?pk_campaign=rss-feed</link>
      <description>&lt;![CDATA[When the world woke up astonished in November 2022 to this &#34;magical&#34; chatbot, few realized that this magic was the result of decades of research. The history of artificial intelligence begins in 1943, when Warren McCulloch and Walter Pitts proposed the first mathematical model of an artificial neuron. In 1956, at the Dartmouth Conference, John McCarthy coined the term &#34;Artificial Intelligence&#34; and the discipline was officially born.&#xA;&#xA;The &#39;60s and &#39;70s were characterized by excessive optimism: people thought strong AI was just around the corner. Two &#34;AI winters&#34; followed – periods when funding disappeared and research slowed – because promises weren&#39;t materializing. But some continued working in the shadows. Geoffrey Hinton, Yann LeCun, Yoshua Bengio – those we now call the &#34;godfathers of deep learning&#34; – continued their studies on neural networks when no one believed in them anymore.&#xA;&#xA;!--more--&#xA;&#xA;The real breakthrough came with three converging factors: computational power (GPUs), enormous amounts of data, and better algorithms. In 2012, AlexNet won the ImageNet Challenge by an overwhelming margin, demonstrating that deep learning really worked. From there, an unstoppable acceleration.&#xA;&#xA;Once upon a time in the Carboniferous...&#xA;Before ChatGPT exploded, my only knowledge of AI came from science fiction books. Philip K. Dick and his reflections on what it means to be human. Cyberpunk in general, with its technological dystopias. Gibson&#39;s Sprawl trilogy, where AIs live in cyberspace like digital deities. Those pages were my only window to a future that seemed incredibly distant.&#xA;&#xA;When I hosted the podcast Caccia al Fotone (a nice thing, but now belonging to the Carboniferous period...), I delved deeper into the subject. I read several papers published on arXiv and dedicated two episodes to AI development. In 2019, during the pandemic period, I devoured &#34;Artificial Intelligence: A Guide for Thinking Humans&#34; by Melanie Mitchell – a book that also helped me write a &#34;thing&#34; (those who know, know; those who don&#39;t, never mind...) on the evolution of computer systems and surveillance capitalism.&#xA;&#xA;I thought I had a clear picture. I thought I was prepared.&#xA;&#xA;Mea culpa&#xA;Then ChatGPT arrived.&#xA;&#xA;November 2022. First approach: total amazement. I couldn&#39;t believe my eyes. I kept asking questions, and despite all the initial hallucinations I encountered, I continued to have that &#34;wow effect&#34; typical of a child finding the most beautiful shell on the seashore (forgive me Newton for stealing that phrase, but it&#39;s always too beautiful).&#xA;&#xA;And here&#39;s my mea culpa: I set aside all my protective filters that I generally have regarding privacy, open source, control over my data. I let myself go for hours of conversations on the most diverse topics. Until one night – one of many sleepless nights – I found myself discussing with that LLM about depression, various mental disorders, and how one or more abuses can influence a person&#39;s life.&#xA;&#xA;When I realized what was happening, I stopped abruptly. I deleted the conversation, canceled my OpenAI subscription and didn&#39;t touch any LLM for more than a month. I was entrusting my most intimate thoughts to a proprietary system controlled by a corporation. I was betraying every principle I believed in.&#xA;&#xA;But I work in IT. This is a huge revolution. I couldn&#39;t afford to fall behind, nor could I simply reject it on principle. I had to find an alternative. I began to study seriously.&#xA;&#xA;Local, always local&#xA;I encountered the first models I could test locally. I discovered Hugging Face, and it was like finding an oasis in the desert. I began studying transformers, the datasets developed by the community. And I was astounded.&#xA;&#xA;Transformers are the architecture that revolutionized AI. Presented in the 2017 paper &#34;Attention Is All You Need&#34;, they replaced old recurrent neural networks (RNNs) with a more elegant and efficient mechanism: the attention mechanism.&#xA;&#xA;In simple words: instead of processing text word by word in sequence, a transformer looks at all words simultaneously and calculates which ones are most relevant to the context. When you read &#34;The bank of the river was green,&#34; the attention mechanism understands that &#34;bank&#34; refers to the river and not the financial institution, because it evaluates the weight of each word relative to the others.&#xA;&#xA;This architecture made models like BERT, GPT, and all modern LLMs possible. It&#39;s scalable, parallelizable, and extremely powerful.&#xA;&#xA;Hugging Face and the Open Source revolution&#xA;Hugging Face is much more than a platform: it has become the Library of Alexandria of the artificial intelligence era. Founded in 2016, it now hosts over 500,000 pre-trained models, 250,000 datasets, and thousands of demo applications.&#xA;&#xA;Their transformers library has democratized access to AI. With a few lines of Python you can download and use models that would cost millions of dollars to train from scratch. Hugging Face isn&#39;t the only platform doing this – there are also Ollama, LM Studio, GPT4All – but it&#39;s certainly the most extensive and collaborative.&#xA;&#xA;Here, praise must be given to the developers: this community of people scattered around the world is doing extraordinary work. They release open source models, share knowledge, meticulously document everything. They&#39;re building a real alternative to Big Tech&#39;s monopoly on AI.&#xA;&#xA;History repeating&#xA;Watching this explosion of open models, global collaboration, shared code, I had a powerful déjà-vu. This is incredibly similar to the open source revolution that happened 30 years ago.&#xA;&#xA;In the &#39;90s, Linux and the free software movement challenged Microsoft&#39;s dominance and proprietary systems. Many said it was impossible, that free software would never work. Today Linux powers 96% of the world&#39;s servers, all Android smartphones, and much of the Internet infrastructure.&#xA;&#xA;Now the same thing is happening with AI. Llama, Mistral, Falcon, Mixtral – &#34;open weight/open source&#34; models that compete with (and often surpass) their proprietary counterparts. History repeats itself, and this time I know which side to be on.&#xA;&#xA;Another server in my homeLab&#xA;I resumed studying Python, a study I had left on standby years ago. I began experimenting with training local LLM models. I added old scripts to provide my writing style (yes, it seems incredible but every coder has their own style, and it says a lot about their personality). I used Llama 3 to improve my Bash coding.&#xA;&#xA;And when I was ready, I decided to make an important purchase: I bought a small server – to add to my homelab: Proxmox, pfSense, Nextcloud, WireGuard etc... – that I would transform into an OpenWebUI system.&#xA;&#xA;OpenWebUI is a self-hosted web interface for local language models. Like ChatGPT, but running entirely on local hardware, without sending a single byte to someone else&#39;s servers.&#xA;&#xA;For the nerds reading: the simplest way to install is obviously through Docker. Here&#39;s a basic example:&#xA;&#xA;docker run -d -p 3000:8080 \&#xA;  -v open-webui:/app/backend/data \&#xA;  --name open-webui \&#xA;  --restart always \&#xA;  ghcr.io/open-webui/open-webui:main&#xA;&#xA;Once installed, just connect OpenWebUI to Ollama (the runtime for local models), download your preferred models, and you&#39;re operational.&#xA;&#xA;GPU usage is fundamental: a medium-sized LLM requires a lot of RAM and computing power. A dedicated GPU (like an NVIDIA GTX of various types) makes an enormous difference. For those using AMD, there&#39;s ROCm. With 16GB of RAM and an 8GB GPU, you can comfortably run 7B parameter models quantized to 4-bit.&#xA;&#xA;My favorite combo? AMD, Debian, Docker, OpenWebUI, Ollama and Mistral.&#xA;&#xA;A revolution. and a choice to make&#xA;We&#39;re facing a revolution that we cannot avoid. AI is here, it&#39;s powerful, and it&#39;s evolving rapidly. There are two roads ahead of us.&#xA;&#xA;The first: avoid it now, close our eyes, hope it passes or that someone else deals with it. And then, in twenty years, find ourselves chasing an evolved AI, probably impossible to understand, completely in the hands of those who controlled it from the beginning. This is the path of least resistance, but also of maximum risk. It means ceding control, understanding, and ultimately power to whoever gets there first.&#xA;&#xA;The second: study it, analyze it, use it and understand it today to be able to handle it better tomorrow. Actively participate in its evolution. Contribute to the open source community, ensure that this technology remains accessible, understandable, in the hands of many instead of a few. This path requires effort, time, sometimes admitting we were wrong (as I did). But it&#39;s the only path that leads to actual agency over our technological future.&#xA;&#xA;The choice seems obvious when stated this way, but it&#39;s not easy in practice. It requires overcoming fear, investing time, challenging our assumptions. It means getting our hands dirty with code, running models locally, understanding how these systems actually work instead of treating them as black boxes.&#xA;&#xA;I made my choice that night when I deleted my ChatGPT conversation history. I chose not to be a passive consumer of AI technology controlled by corporations. I chose to understand, to build, to contribute to the alternative that&#39;s being constructed by thousands of developers around the world.&#xA;&#xA;The technology is already here. The question is: will it be controlled by a few companies optimizing for profit and control, or will it be a tool accessible to everyone, understandable, modifiable, improvable by the community?&#xA;&#xA;As I&#39;ve learned on this journey, choosing to understand – even when it&#39;s difficult, even when it means admitting you were wrong – is always better than passively submitting.&#xA;&#xA;AI is not magic. It&#39;s mathematics, code, hardware, and above all: it&#39;s made by people. And if it&#39;s made by people, it can be understood, modified and shaped by people. For the better, not for the worse.&#xA;&#xA;The revolution is happening. The only question is: are you participating, or are you watching?&#xA;&#xA;#AI #OpenSource #LocalLLM #Privacy #ChatGPT #HuggingFace #Ollama #SelfHosted #MachineLearning #DigitalSovereignty&#xA;&#xA;a href=&#34;https://remark.as/p/jolek78/chatgpt-didnt-invent-anything&#34;Discuss.../a&#xA;&#xA;div class=&#34;center&#34;&#xD;&#xA;· 🦣 a href=&#34;https://fosstodon.org/@jolek78&#34;Mastodon/a · 📸 a href=&#34;https://pixelfed.social/jolek78&#34;Pixelfed/a ·  📬 a href=&#34;mailto:jolek78@jolek78.dev&#34;Email/a ·&#xD;&#xA;· ☕ a href=&#34;https://liberapay.com/jolek78&#34;Support this work on Liberapay/a&#xD;&#xA;/div]]&gt;</description>
      <content:encoded><![CDATA[<p>When the world woke up astonished in November 2022 to this “magical” chatbot, few realized that this magic was the result of decades of research. The history of artificial intelligence begins in 1943, when Warren McCulloch and Walter Pitts proposed the first mathematical model of an artificial neuron. In 1956, at the Dartmouth Conference, John McCarthy coined the term “Artificial Intelligence” and the discipline was officially born.</p>

<p>The &#39;60s and &#39;70s were characterized by excessive optimism: people thought strong AI was just around the corner. Two “AI winters” followed – periods when funding disappeared and research slowed – because promises weren&#39;t materializing. But some continued working in the shadows. Geoffrey Hinton, Yann LeCun, Yoshua Bengio – those we now call the “godfathers of deep learning” – continued their studies on neural networks when no one believed in them anymore.</p>



<p>The real breakthrough came with three converging factors: computational power (GPUs), enormous amounts of data, and better algorithms. In 2012, AlexNet won the ImageNet Challenge by an overwhelming margin, demonstrating that deep learning really worked. From there, an unstoppable acceleration.</p>

<h3 id="once-upon-a-time-in-the-carboniferous" id="once-upon-a-time-in-the-carboniferous">Once upon a time in the Carboniferous...</h3>

<p>Before ChatGPT exploded, my only knowledge of AI came from science fiction books. Philip K. Dick and his reflections on what it means to be human. Cyberpunk in general, with its technological dystopias. Gibson&#39;s Sprawl trilogy, where AIs live in cyberspace like digital deities. Those pages were my only window to a future that seemed incredibly distant.</p>

<p>When I hosted the podcast Caccia al Fotone (a nice thing, but now belonging to the Carboniferous period...), I delved deeper into the subject. I read several papers published on arXiv and dedicated two episodes to AI development. In 2019, during the pandemic period, I devoured “Artificial Intelligence: A Guide for Thinking Humans” by Melanie Mitchell – a book that also helped me write a “thing” (those who know, know; those who don&#39;t, never mind...) on the evolution of computer systems and surveillance capitalism.</p>

<p>I thought I had a clear picture. I thought I was prepared.</p>

<h3 id="mea-culpa" id="mea-culpa">Mea culpa</h3>

<p>Then ChatGPT arrived.</p>

<p>November 2022. First approach: total amazement. I couldn&#39;t believe my eyes. I kept asking questions, and despite all the initial hallucinations I encountered, I continued to have that “wow effect” typical of a child finding the most beautiful shell on the seashore (forgive me Newton for stealing that phrase, but it&#39;s always too beautiful).</p>

<p>And here&#39;s my mea culpa: I set aside all my protective filters that I generally have regarding privacy, open source, control over my data. I let myself go for hours of conversations on the most diverse topics. Until one night – one of many sleepless nights – I found myself discussing with that LLM about depression, various mental disorders, and how one or more abuses can influence a person&#39;s life.</p>

<p>When I realized what was happening, I stopped abruptly. I deleted the conversation, canceled my OpenAI subscription and didn&#39;t touch any LLM for more than a month. I was entrusting my most intimate thoughts to a proprietary system controlled by a corporation. I was betraying every principle I believed in.</p>

<p>But I work in IT. This is a huge revolution. I couldn&#39;t afford to fall behind, nor could I simply reject it on principle. I had to find an alternative. I began to study seriously.</p>

<h3 id="local-always-local" id="local-always-local">Local, always local</h3>

<p>I encountered the first models I could test locally. I discovered <a href="https://huggingface.co">Hugging Face</a>, and it was like finding an oasis in the desert. I began studying transformers, the datasets developed by the community. And I was astounded.</p>

<p><strong>Transformers</strong> are the architecture that revolutionized AI. Presented in the 2017 paper <a href="https://arxiv.org/abs/1706.03762">“Attention Is All You Need”</a>, they replaced old recurrent neural networks (RNNs) with a more elegant and efficient mechanism: the attention mechanism.</p>

<p>In simple words: instead of processing text word by word in sequence, a transformer looks at all words simultaneously and calculates which ones are most relevant to the context. When you read “The bank of the river was green,” the attention mechanism understands that “bank” refers to the river and not the financial institution, because it evaluates the weight of each word relative to the others.</p>

<p>This architecture made models like BERT, GPT, and all modern LLMs possible. It&#39;s scalable, parallelizable, and extremely powerful.</p>

<h3 id="hugging-face-and-the-open-source-revolution" id="hugging-face-and-the-open-source-revolution">Hugging Face and the Open Source revolution</h3>

<p><a href="https://huggingface.co">Hugging Face</a> is much more than a platform: it has become the Library of Alexandria of the artificial intelligence era. Founded in 2016, it now hosts over 500,000 pre-trained models, 250,000 datasets, and thousands of demo applications.</p>

<p>Their <a href="https://github.com/huggingface/transformers">transformers library</a> has democratized access to AI. With a few lines of Python you can download and use models that would cost millions of dollars to train from scratch. Hugging Face isn&#39;t the only platform doing this – there are also <a href="https://ollama.com">Ollama</a>, <a href="https://lmstudio.ai">LM Studio</a>, <a href="https://gpt4all.io">GPT4All</a> – but it&#39;s certainly the most extensive and collaborative.</p>

<p>Here, praise must be given to the developers: this community of people scattered around the world is doing extraordinary work. They release open source models, share knowledge, meticulously document everything. They&#39;re building a real alternative to Big Tech&#39;s monopoly on AI.</p>

<h3 id="history-repeating" id="history-repeating">History repeating</h3>

<p>Watching this explosion of open models, global collaboration, shared code, I had a powerful déjà-vu. This is incredibly similar to the open source revolution that happened 30 years ago.</p>

<p>In the &#39;90s, Linux and the free software movement challenged Microsoft&#39;s dominance and proprietary systems. Many said it was impossible, that free software would never work. Today Linux powers 96% of the world&#39;s servers, all Android smartphones, and much of the Internet infrastructure.</p>

<p>Now the same thing is happening with AI. Llama, Mistral, Falcon, Mixtral – “open weight/open source” models that compete with (and often surpass) their proprietary counterparts. History repeats itself, and this time I know which side to be on.</p>

<h3 id="another-server-in-my-homelab" id="another-server-in-my-homelab">Another server in my homeLab</h3>

<p>I resumed studying Python, a study I had left on standby years ago. I began experimenting with training local LLM models. I added old scripts to provide my writing style (yes, it seems incredible but every coder has their own style, and it says a lot about their personality). I used Llama 3 to improve my Bash coding.</p>

<p>And when I was ready, I decided to make an important purchase: I bought a small server – to add to my homelab: Proxmox, pfSense, Nextcloud, WireGuard etc... – that I would transform into an <a href="https://openwebui.com">OpenWebUI</a> system.</p>

<p>OpenWebUI is a self-hosted web interface for local language models. Like ChatGPT, but running entirely on local hardware, without sending a single byte to someone else&#39;s servers.</p>

<p>For the nerds reading: the simplest way to install is obviously through Docker. Here&#39;s a basic example:</p>

<pre><code>docker run -d -p 3000:8080 \
  -v open-webui:/app/backend/data \
  --name open-webui \
  --restart always \
  ghcr.io/open-webui/open-webui:main
</code></pre>

<p>Once installed, just connect OpenWebUI to <a href="https://ollama.com">Ollama</a> (the runtime for local models), download your preferred models, and you&#39;re operational.</p>

<p>GPU usage is fundamental: a medium-sized LLM requires a lot of RAM and computing power. A dedicated GPU (like an NVIDIA GTX of various types) makes an enormous difference. For those using AMD, there&#39;s ROCm. With 16GB of RAM and an 8GB GPU, you can comfortably run 7B parameter models quantized to 4-bit.</p>

<p>My favorite combo? AMD, Debian, Docker, OpenWebUI, Ollama and Mistral.</p>

<h3 id="a-revolution-and-a-choice-to-make" id="a-revolution-and-a-choice-to-make">A revolution. and a choice to make</h3>

<p>We&#39;re facing a revolution that we cannot avoid. AI is here, it&#39;s powerful, and it&#39;s evolving rapidly. There are two roads ahead of us.</p>

<p><strong>The first:</strong> avoid it now, close our eyes, hope it passes or that someone else deals with it. And then, in twenty years, find ourselves chasing an evolved AI, probably impossible to understand, completely in the hands of those who controlled it from the beginning. This is the path of least resistance, but also of maximum risk. It means ceding control, understanding, and ultimately power to whoever gets there first.</p>

<p><strong>The second:</strong> study it, analyze it, use it and understand it today to be able to handle it better tomorrow. Actively participate in its evolution. Contribute to the open source community, ensure that this technology remains accessible, understandable, in the hands of many instead of a few. This path requires effort, time, sometimes admitting we were wrong (as I did). But it&#39;s the only path that leads to actual agency over our technological future.</p>

<p>The choice seems obvious when stated this way, but it&#39;s not easy in practice. It requires overcoming fear, investing time, challenging our assumptions. It means getting our hands dirty with code, running models locally, understanding how these systems actually work instead of treating them as black boxes.</p>

<p>I made my choice that night when I deleted my ChatGPT conversation history. I chose not to be a passive consumer of AI technology controlled by corporations. I chose to understand, to build, to contribute to the alternative that&#39;s being constructed by thousands of developers around the world.</p>

<p>The technology is already here. The question is: will it be controlled by a few companies optimizing for profit and control, or will it be a tool accessible to everyone, understandable, modifiable, improvable by the community?</p>

<p>As I&#39;ve learned on this journey, choosing to understand – even when it&#39;s difficult, even when it means admitting you were wrong – is always better than passively submitting.</p>

<p>AI is not magic. It&#39;s mathematics, code, hardware, and above all: it&#39;s made by people. And if it&#39;s made by people, it can be understood, modified and shaped by people. For the better, not for the worse.</p>

<p>The revolution is happening. The only question is: are you participating, or are you watching?</p>

<p><a href="https://jolek78.writeas.com/tag:AI" class="hashtag"><span>#</span><span class="p-category">AI</span></a> <a href="https://jolek78.writeas.com/tag:OpenSource" class="hashtag"><span>#</span><span class="p-category">OpenSource</span></a> <a href="https://jolek78.writeas.com/tag:LocalLLM" class="hashtag"><span>#</span><span class="p-category">LocalLLM</span></a> <a href="https://jolek78.writeas.com/tag:Privacy" class="hashtag"><span>#</span><span class="p-category">Privacy</span></a> <a href="https://jolek78.writeas.com/tag:ChatGPT" class="hashtag"><span>#</span><span class="p-category">ChatGPT</span></a> <a href="https://jolek78.writeas.com/tag:HuggingFace" class="hashtag"><span>#</span><span class="p-category">HuggingFace</span></a> <a href="https://jolek78.writeas.com/tag:Ollama" class="hashtag"><span>#</span><span class="p-category">Ollama</span></a> <a href="https://jolek78.writeas.com/tag:SelfHosted" class="hashtag"><span>#</span><span class="p-category">SelfHosted</span></a> <a href="https://jolek78.writeas.com/tag:MachineLearning" class="hashtag"><span>#</span><span class="p-category">MachineLearning</span></a> <a href="https://jolek78.writeas.com/tag:DigitalSovereignty" class="hashtag"><span>#</span><span class="p-category">DigitalSovereignty</span></a></p>

<p><a href="https://remark.as/p/jolek78/chatgpt-didnt-invent-anything">Discuss...</a></p>

<div class="center">
· 🦣 <a href="https://fosstodon.org/@jolek78">Mastodon</a> · 📸 <a href="https://pixelfed.social/jolek78">Pixelfed</a> ·  📬 <a href="mailto:jolek78@jolek78.dev">Email</a> ·
· ☕ <a href="https://liberapay.com/jolek78">Support this work on Liberapay</a>
</div>
]]></content:encoded>
      <guid>https://jolek78.writeas.com/chatgpt-didnt-invent-anything</guid>
      <pubDate>Tue, 28 Oct 2025 12:56:35 +0000</pubDate>
    </item>
  </channel>
</rss>