<?xml version="1.0" encoding="UTF-8"?><rss version="2.0" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>opensource &amp;mdash; jolek78&#39;s blog</title>
    <link>https://jolek78.writeas.com/tag:opensource</link>
    <description>thoughts from a friendly human being</description>
    <pubDate>Sun, 19 Apr 2026 22:27:44 +0000</pubDate>
    <image>
      <url>https://i.snap.as/DEj7yFm4.png</url>
      <title>opensource &amp;mdash; jolek78&#39;s blog</title>
      <link>https://jolek78.writeas.com/tag:opensource</link>
    </image>
    <item>
      <title>Reflections on an (impossible) escape from capitalism</title>
      <link>https://jolek78.writeas.com/reflections-on-an-impossible-escape-from-capitalism?pk_campaign=rss-feed</link>
      <description>&lt;![CDATA[It was an ordinary Friday evening. The parcel had arrived with the courier that morning, but I only opened it after dinner, with that silent ceremony I perform every time new hardware shows up - as if opening a box too quickly were a form of disrespect toward the object. Inside was a HUNSN 4K. Small, almost ridiculously small. A mini PC in a form factor that fit in the palm of a hand. I put it on the table, looked at it. Looked at it again. And then an uncomfortable thought occurred to me. I had ordered it from a Chinese reseller, paid with a credit card, through a completely traceable payment infrastructure, from one of the most centralised and surveilled commercial ecosystems in existence. To build a homelab that would let me escape centralised and surveilled ecosystems.&#xA;&#xA;!--more--&#xA;&#xA;The funny thing - funny in the sense that it makes you laugh, but badly - is that I&#39;m not alone. Every day, somewhere in the world, someone orders a mini PC, a Raspberry Pi, a managed Mikrotik switch, with the stated goal of taking back control of their digital life. They order it on Alibaba, pay with PayPal, wait for the courier. And they see nothing strange in any of this, because the contradiction is so structural it has become invisible. This article is an attempt to make it visible again. Without easy solutions, because I don&#39;t have any. And when have I ever…&#xA;&#xA;The Promise of the Homelab&#xA;&#xA;When, in 2019, I started self-hosting pretty much everything - Nextcloud (always on a Raspberry Pi, first RPi3 then RPi4), Jellyfin, Navidrome, FreshRSS, and about twenty-five other services on Proxmox LXC, each with its own isolated Docker daemon - I did it with a precise motivation: I wanted to know where my data lived, who could read it, and have the ability to switch it off myself if I ever felt like it. Not when a company decides to shut down a service, not when someone else changes the licence terms. Me. This came after a long period of reflection on myself, the work I was doing and still do, and the technological society I live in. It is an ideological choice before it is a technical one. Technology as a tool for autonomy rather than control; infrastructure as something you own instead of something that owns you. I hope no one is alarmed if I say that some of these reflections began with reading Theodore Kaczynski&#39;s Manifesto, before eventually landing, of course, on more authoritative sources.&#xA;&#xA;Yes, I&#39;m mad, but not quite that mad…&#xA;&#xA;When you pay a subscription to a cloud service, the transaction does not end the moment you authorise the electronic payment. Shoshana Zuboff, in The Age of Surveillance Capitalism, calls this mechanism behavioral surplus: the behavioural data extracted beyond what is needed to provide the service, then resold as predictive raw material.&#xA;&#xA;  Under the regime of surveillance capitalism, however, the first text does not stand alone; it trails a shadow close behind. The first text, full of promise, actually functions as the supply operation for the second text: the shadow text. Everything that we contribute to the first text, no matter how trivial or fleeting, becomes a target for surplus extraction. That surplus fills the pages of the second text. This one is hidden from our view: &#34;read only&#34; for surveillance capitalists. In this text our experience is dragooned as raw material to be accumulated and analyzed as means to others&#39; market ends. The shadow text is a burgeoning accumulation of behavioral surplus and its analyses, and it says more about us than we can know about ourselves. Worse still, it becomes increasingly difficult, and perhaps impossible, to refrain from contributing to the shadow text. It automatically feeds on our experience as we engage in the normal and necessary routines of social participation.&#xA;&#xA;You are not the customer of the system - you are its product. Your habits, your schedules, your preferences, your hesitations before clicking on something: all of this is collected, modelled, sold. The transaction is not monthly: it is continuous, invisible, and never ends as long as you use the service. With hardware, in principle, the transaction is one-time: you buy, you pay, it ends, it is yours. The disk is in your room, not on a server subject to government requests, security breaches, or business decisions that are nothing to do with you but impact your access to those services. This distinction - between a tool you use and a system that uses you - is the real stake of the homelab. It is not about saving money, it is not about performance. It is about who controls what.&#xA;&#xA;The problem is that building this infrastructure requires hardware, time, knowledge, and resources. The hardware comes from somewhere; the time, the knowledge, and the energy resources come from a privilege not granted to everyone.&#xA;&#xA;The Market I Hadn&#39;t Seen&#xA;&#xA;Search for &#34;mini PC homelab&#34; on any marketplace. What you find is a productive ecosystem that has exploded over the past five years in a way I honestly did not expect.&#xA;&#xA;MINISFORUM, Beelink, Trigkey, Geekom, GMKtec. Zimaboard, with its single-board aesthetic designed explicitly for those who want home racks. Raspberry Pi and the galaxy of clones - Orange Pi, Rock Pi, Banana Pi. Managed Mikrotik switches at accessible prices. 1U rack cases to mount under the desk. M.2 NVMe SSDs with TBW figures calculated for small-server workloads. Silent power supplies designed to run 24/7. A market built from scratch, that exists precisely because there is a community of people who want to run servers at home. r/homelab and r/selfhosted on Reddit have approximately 2.8 and 1.7 million members respectively - numbers publicly verifiable, and growing. YouTube is full of dedicated channels. There is an entire attention economy built around &#34;escaping&#34; the attention economy.&#xA;&#xA;But it is worth asking: who built this market, and why. MINISFORUM and Beelink do not exist out of ideological sympathy for the homelab movement. They exist because they identified a profitable segment and served it with industrial precision. Kate Crawford, in Atlas of AI, documents how technology supply chains follow niche demand with the same efficiency with which they follow mass demand: factories in Guangdong optimise production lines not for a worldview, but for a margin. The fact that the resulting product also satisfies an ideological need is, from the manufacturer&#39;s point of view, irrelevant.&#xA;&#xA;  The Victorian environmental disaster at the dawn of the global information society shows how the relations between technology and its materials, environments, and labor practices are interwoven. Just as Victorians precipitated ecological disaster for their early cables, so do contemporary mining and global supply chains further imperil the delicate ecological balance of our era.&#xA;&#xA;The mechanism had been described with theoretical precision back in 1999 by Luc Boltanski and Ève Chiapello in The New Spirit of Capitalism. Their thesis: capitalism is never defeated by criticism - it is incorporated. When a critique becomes widespread enough, the system absorbs it and transforms it into a market segment. The artistic critique of the 1960s - autonomy, authenticity, rejection of standardisation - became the marketing of the creative economy. The critique of digital centralisation - sovereignty, privacy, control - has become an online catalogue to browse through.&#xA;&#xA;Resistance has become a market segment. Every time someone buys a HUNSN to stop paying subscriptions to services they don&#39;t control, a factory in Guangdong sells a HUNSN. Capitalism has not been defeated - it has shifted (at least for a small slice of the population: the nerds, the hackers) the extraction point from subscriptions to hardware.&#xA;&#xA;The Accumulation Syndrome&#xA;&#xA;But there is a further level - more ridiculous and more personal - that homelab communities never discuss openly, yet anyone who has a homelab recognises immediately. The Raspberry Pi 4 bought &#34;for a project.&#34; The old ThinkPad kept because &#34;you never know.&#34; The 4TB disk salvaged from a decommissioned NAS - and &#34;it might come in handy.&#34; The second-hand switch picked up on eBay for eighteen euros because it was cheap and might be useful. The cables, the cables, the cables.&#xA;&#xA;r/homelab has a term for this: just in case hardware. It is the hardware of the imaginary future, of projects that only exist in your head, of configurations that one day - one day - you will finally test. In the meantime it occupies a shelf, draws current in standby, and generates a diffuse sense of possibility that is indistinguishable from the most classic consumerism. The underlying psychological mechanism has a precise name: compensatory consumption - consumption as a response to a perceived loss of autonomy or control. You buy hardware because buying hardware gives you the feeling of recovering agency over something. The aesthetic is different from traditional consumerism - no luxury logos, no recognisable status symbols - but the mechanism is identical.&#xA;&#xA;That said, there is a partially honest answer to all of this: the second-hand and refurbished market. The ThinkPad X230 on eBay, the Dell R720 server decommissioned from a datacentre, the disk from someone who upgraded their NAS. My ZFS NAS, to give one example, is a recycled old tower with four 1TB disks in RAIDZ - hardware that would otherwise have ended up in landfill, with a life cycle extended by years, without generating new production demand. It is closer to the ethics of repair than to compulsive buying. But it has its own internal contradiction: it requires even more technical competence than buying new - knowing how to assess wear, diagnose an unknown component, manage ten-year-old drivers. The barrier to entry rises further. And the refurbished market is itself now an organised commercial sector, with its own margins, its own platforms, its own pricing logic. It is not a clean way out. It is a less dirty way out.&#xA;&#xA;And then there is the energy question, which is usually ignored in homelab discussions and is instead the most uncomfortable of all - uncomfortable enough to deserve a more in-depth treatment later on. For now, suffice it to say: every machine on your shelf that &#34;draws current in standby&#34; is a line item in the energy bill that the homelab movement rarely accounts for.&#xA;&#xA;Not for Everyone. And It Should Not Be This Way.&#xA;&#xA;There is a second level of the paradox that is even more uncomfortable than the first. Building a homelab costs money - relatively little, but it costs. It requires physical space. It requires a decent connection. And it requires time. A lot of time. Not installation time - that is measurable, finite. The learning time that precedes everything else. To reach the point where you can build a functional infrastructure with Proxmox, LXC containers, centralised authentication, reverse proxy, automated backups - you need to have already spent years understanding how Linux works, how to reason about networks and permissions, how to read a log. I started with a Red Hat in 1997, and it took me almost thirty years to get where I am. I should know this. Yet it always escapes me. And that time did not fall from the sky. It is time I was able to dedicate because I had a certain kind of job, a certain stability, a certain amount of mental energy left at the end of the day. It is middle-class-with-a-stable-position time, not the time of someone working three warehouse shifts a week. Passion is not enough.&#xA;&#xA;Johan Söderberg documents this in Hacking Capitalism: the FOSS movement was born as resistance to capitalism, but reproduces within itself hierarchies of skill and merit that make it structurally exclusive. Freedom is technically available to anyone, but effective access requires resources distributed in anything but a democratic manner. Söderberg goes further than simply observing the exclusivity: the voluntary open source work produces use value - functioning software, documentation, community support - that capital then extracts as exchange value without remunerating those who produced it. Red Hat builds a billion-dollar company on a kernel written largely by volunteers. It is not just that not everyone can get in: it is that those who get in often work for someone without knowing it. The homelab inherits this problem and amplifies it.&#xA;&#xA;  The narrative of orthodox historical materialism corresponds with some very popular ideas in the computer underground. It is widely held that the infinite reproducibility of information made possible by computers (forces of production) has rendered intellectual property (relations of production, superstructure) obsolete. The storyline of post-industrial ideology is endorsed but with a different ending. Rather than culminating in global markets, technocracy and liberalism, as Daniel Bell and the futurists would have it; hackers are looking forward to a digital gift economy and high-tech anarchism. In a second turn of events, hackers have jumped on the distorted remains of Marxism presented in information-age literature, and, while missing out on the vocabulary, ended up promoting an upgraded Karl Kautsky-version of historical materialism.&#xA;&#xA;This is not a quirk of the homelab movement: it is a recurring structure in every technological wave. Langdon Winner, in his influential essay Do Artifacts Have Politics?, argued that technological choices are never neutral - they incorporate power structures, distribute access in non-random ways. Amateur radio in the 1920s, the personal computer in the 1980s, the internet in the 1990s: every time the promise was democratising, every time the actual distribution followed the lines of pre-existing privilege. Not out of malice, but out of structure. The irony is this: those who would most need digital autonomy - those who cannot afford subscriptions, those who live under governments that surveil communications, those most exposed to data collection - are exactly those least likely to be able to build a homelab. Not for lack of interest or intelligence. For lack of time, money, and years of privileged exposure to technology.&#xA;&#xA;Homelab communities do not usually talk about this. They talk about which mini PC to buy, how to optimise energy consumption, which distro to use as a base. The conversation about structural exclusivity exists, but at the margins - in Jacobin, in Logic Magazine, in EFF activism - while the centre of the discourse remains impermeable. It is not that no one speaks about it: it is that the peripheries speak about it, and the peripheries do not set the agenda. This entire conversation takes place in a room to which not everyone has a ticket. And those inside do not seem to find that particularly problematic.&#xA;&#xA;A Technological Cosplay?&#xA;&#xA;So is the whole thing a con? Is the homelab just anti-capitalist cosplay while you continue to fund the same supply chains? In part, yes.&#xA;&#xA;The HUNSN 4K was designed in China, assembled in China, shipped by container on ships burning bunker fuel. Global maritime transport is responsible for approximately 2.5% of global CO₂ emissions - a share that the IMO (International Maritime Organization) has been trying to reduce for years with slow progress and targets continually postponed. Then: distributed through Alibaba, paid with a credit card. Every piece of technology hardware carries an extractive chain that begins in lithium mines in Bolivia and cobalt mines in the Democratic Republic of the Congo, passes through factories in Guangdong, and ends in electronic waste processing centres in Ghana. The hardware travels that supply chain exactly like any other consumer device. Furthermore, hardware has a lifecycle. In five years the HUNSN 4K will be too slow, or it will break, or something will come out with energy efficiency too much better to ignore. And I will buy again. The mini PC market for homelabs depends on the obsolescence of previous purchases - exactly like any other consumer market.&#xA;&#xA;The critique of capitalism, when it is widespread enough, is not suppressed - it is incorporated. The system absorbs the values of resistance and transforms them into a market segment. Autonomy becomes a selling point. Decentralisation becomes a brand. The rebel who wanted to exit the system finds himself funding a new vertical of the same system, convinced he is making an ethical choice.&#xA;&#xA;The Counter-Shot&#xA;&#xA;But there is a structural difference that would be dishonest to ignore.&#xA;&#xA;When you pay a subscription to a cloud service, the cost is not just the monthly fee. It is the continuous cession of data, behaviours, habits. It is the behavioral surplus Zuboff talks about: you are not using a service, you are being used as raw material to train models, build profiles, sell advertising. The transaction never ends, in ways you often cannot see and cannot escape from as long as you use the service.&#xA;&#xA;With hardware, the transaction ends. The data stays on a physical disk in your room, not on a server subject to government requests, breaches, or business decisions that have nothing to do with you but impact your life. The software running on it - Proxmox, Debian, Nextcloud, Jellyfin - is open source; you can modify it. If something changes in a way you cannot accept, you can leave. This resilience has real value - but it is worth noting that it is asymmetric resilience: it works for those who have the skills to exercise it. For those who do not, the theoretical portability of their data from Nextcloud to something else requires exactly the same skills we have already identified as the barrier to entry. The freedom to leave is real. Access to that freedom, much less so.&#xA;&#xA;And then there is the energy question, which I have deferred long enough. The major hyperscalers - AWS, Google, Azure - operate with a PUE (Power Usage Effectiveness) between 1.1 and 1.2. For every watt of useful computation they dissipate barely 0.1–0.2 watts in heat and infrastructure. They have enormous economies of scale, optimised industrial cooling, significant investments in renewable energy, and above all: their servers run at very high utilisation rates. Almost always busy.&#xA;&#xA;A home homelab works in a radically different way. The machine runs 24/7 even when it is doing nothing - and for most of the time it is doing nothing. Navidrome serving three requests a day, FreshRSS fetching every hour, an LDAP container sitting listening without receiving connections. You are paying the energy cost of the infrastructure regardless of usage. The implicit PUE of a homelab, calculated honestly on the ratio between total consumption and actual workload, is much worse than that of a datacentre. IEA data (Data Centres and Data Transmission Networks, updated annually) shows that large cloud providers progressively improve energy efficiency thanks to economies of scale that no individual homelab can replicate. The flip side is that the same growth in demand that makes economies of scale possible negates the efficiency gains: Amazon&#39;s absolute emissions increased between 2023 and 2024 despite improved PUE. Efficiency improves. Total consumption grows anyway. This is Jevons&#39; Paradox: energy efficiency, instead of reducing consumption, increases it, because it lowers the marginal cost of use and stimulates demand that grows faster than the efficiency gains.&#xA;&#xA;  Note: The comparison is not as linear as the numbers suggest. PUE measures the internal efficiency of a datacentre, not the energy cost of the network traffic that data generates every time it leaves it - traffic that a homelab eliminates almost completely for internal services. Nor does it measure proportion: AWS is efficient at delivering services to millions of users, but that scale says nothing about the real cost of storing fifty gigabytes of personal data on a server designed for loads a thousand times greater. A HUNSN N100 in idle consumes less than 8 watts. The honest energy comparison is not homelab vs hyperscaler in the abstract - it is homelab vs proportional share of hyperscaler for your specific workload, a calculation that nobody can make with publicly available data.&#xA;&#xA;This does not automatically mean that the cloud is the ethically correct choice - the problem does not reduce to PUE, and surveillance has costs that are not measured in kilowatts. It means that anyone with SolarPunk values who chooses the homelab must reckon with a real contradiction: the choice of sovereignty may be, watt for watt, energetically more costly than the system one wants to escape. I have no clean answer, but ignoring the question would be dishonest. Söderberg acknowledges that the FOSS movement has produced concrete and undeniable gains - they simply are not enough, on their own, to subvert the dynamics of informational capitalism.&#xA;&#xA;In short: this is not a critique of the homelab, but it is a critique of the homelab presented as a sufficient revolutionary act.&#xA;&#xA;What Happens at Eleven PM - and Beyond&#xA;&#xA;That night, with the HUNSN 4K on the table, I pressed on. I installed Proxmox. I configured the network. I started bringing up containers one by one. And at some point - three hours had passed, I had three terminals open and was debugging nslcd to centralise LDAP authentication across all the containers - I realised something: I was doing all of this simply because I enjoyed it. Not to resist something. Not to advance an ideological agenda. Because there was a problem to solve and solving it gave me satisfaction. Mihaly Csikszentmihalyi describes this state in Flow as total absorption in a task calibrated to one&#39;s own competencies: time expands, attention narrows, awareness of context vanishes. It is not motivation - it is something more immediate. Debugging an authentication problem at eleven at night on a system I could have chosen not to build is, neuropsychologically, indistinguishable from pleasure. Not the satisfaction of having finished: the process itself. Moreover, for an AuDHD person like me, going into hyperfocus allows you to lose your sense of time entirely, and to literally escape from a world you viscerally loathe.&#xA;&#xA;Ah - you had not figured that out yet?&#xA;&#xA;When I had finished and closed everything, the satisfaction was still there. Along with a mildly uncomfortable awareness: I could probably have used a hosted service, lived just as well, and not lost three hours of a weeknight. But in the meantime I had understood how PAM worked, I had read documentation I had never opened before, I had implemented it on my homelab, I had learned something I hadn&#39;t known I wanted to know.&#xA;&#xA;And here the circle closes in a somewhat unsettling way. Söderberg speaks of voluntary open source work as the production of pure use value - the intrinsic pleasure of doing, understanding, building something that works. But it is exactly this use value that capital then extracts as exchange value: the competence I accumulate debugging LDAP at eleven at night is the same competence I bring to work the next day, that I put into articles like this one, that I share in communities where others use it to build their own homelabs. Technical pleasure is not neutral. It has a production chain. Not always visible, but real.&#xA;&#xA;This is what the homelab is, at least for me: a way of learning that produces, as a side effect, an infrastructure I control. The ideology is there, but it comes second. First comes the pleasure of understanding how something works. Or rather: ideology and pleasure are interchangeable, and often run in parallel - but this does not resolve any of the contradictions I described above. It leaves them all standing, in fact makes them stranger. Am I resisting capitalism, or am I just cultivating an expensive hobby with a political aesthetic?&#xA;&#xA;The Hacker Ethic&#xA;&#xA;The word &#34;hacker&#34; has had bad press for decades. In 1990s news bulletins it was a synonym for a hooded cybercriminal; in the jargon of security companies it became a marketing term to prepend to anything. Neither has much to do with what the word historically means. Steven Levy, in Hackers: Heroes of the Computer Revolution, reconstructs the culture that formed around the MIT and Stanford labs in the 1960s: a community of programmers for whom code was an aesthetic object, access to information a moral principle, and technical competence the only legitimate hierarchy. The principles Levy identifies as the &#34;hacker ethic&#34; are precise: access to computers - and to anything that can teach you how the world works - should be unlimited and total. All information should be free. Decentralised systems are preferable to centralised ones. Hackers should be judged by what they produce, not by titles, age, race, or position. You can create art and beauty with a computer.&#xA;&#xA;It is not a political manifesto in the traditional sense. It is something more visceral - a disposition toward the world, a way of standing before a system you do not yet understand: the correct response is to take it apart, understand how it works, and put it back together better than before.&#xA;&#xA;Pekka Himanen, in The Hacker Ethic and the Spirit of the Information Age - with a preface by Linus Torvalds and an epilogue by Manuel Castells, which already says something about the project&#39;s ambition - performs a more explicit theoretical operation. He builds the hacker ethic in direct opposition to the Protestant work ethic described by Max Weber: where Weber saw work as duty, discipline as virtue, and leisure as absence of production, Himanen identifies in the hacker a figure who works out of passion, considers play an integral part of work, and rejects the sharp separation between productive time and free time. The hacker does not work for money - money is a side effect, when it comes. They work because the problem is interesting. Because the elegant solution has value in itself. Because understanding how something works is, in and of itself, sufficient.&#xA;&#xA;  Hacker activity is also joyful. It often has its roots in playful explorations. Torvalds has described, in messages on the Net, how Linux began to expand from small experiments with the computer he had just acquired. In the same messages, he has explained his motivation for developing Linux by simply stating that &#34;it was/is fun working on it.&#34; Tim Berners-Lee, the man behind the Web, also describes how this creation began with experiments in linking what he called &#34;play programs.&#34; Wozniak relates how many characteristics of the Apple computer &#34;came from a game, and the fun features that were built in were only to do one pet project, which was to program … [a game called] Breakout and show it off at the club.&#34;&#xA;&#xA;Recognise something? I do. Those three hours debugging nslcd at eleven at night were not work in the Weberian sense - nobody was paying me, nobody had asked me to do it, there was no corporate objective to reach. They were hacking in the precise sense that Levy and Himanen describe: exploration motivated by curiosity, with the infrastructure as an object of study as much as of utility. The homelab is, culturally, a direct expression of the hacker ethic. It is no coincidence that homelab communities and open source communities overlap almost perfectly, that they use the same language, the same platforms, the same values. But here, as elsewhere in this article, the story gets complicated.&#xA;&#xA;The hacker ethic promises a pure meritocracy: you are judged by what you can do, not by who you are. It is an attractive idea. It is also, in practice, a partial fiction. Technical meritocracy presupposes that everyone starts from the same point - that skills are accessible to anyone who really wants to acquire them, that the time to acquire them is distributed equally, that mentorship networks and learning resources are available regardless of context. The homelab as hacker practice inherits both things: the genuine nature of curiosity as a driver, and structural exclusivity as an undeclared side effect. The pleasure of taking a system apart to understand how it works is real and should not be devalued. But that pleasure is available, in practice, to those who already have the ticket.&#xA;&#xA;Conclusions&#xA;&#xA;The HUNSN 4K runs, alongside the other &#34;little electronic contraptions,&#34; on a rack next to my armchair - the one where, at the end of the day, I indulge my guilty pleasure of reading a book in the company of my cats. Proxmox, the Nextcloud server, the ZFS NAS, a small MINISFORUM box running Ollama with some local open-weight LLM models, a Raspberry Pi 5 running the Tor Relay, and a HUNSN RJ15 with pfSense controlling incoming and outgoing traffic. An infrastructure, in short, that allows me to have something resembling digital sovereignty within the limits of the possible. The contradictions I have described do not resolve. They are held together, with effort, as any intellectually complex position on a complex system must be held together.&#xA;&#xA;The first: the market that made the accessible homelab possible is the same market the homelab is supposed to emancipate us from. If this explosion of affordable, efficient mini PCs had not happened - if capitalism had not decided to build exactly what we wanted - how many of us would have taken the same path? How much of our &#34;ethical choice&#34; depends on the existence of products designed and sold precisely for us?&#xA;&#xA;The second: does incorporated resistance truly lose its force, or does it remain resistance even when someone profits from it? Boltanski and Chiapello describe the incorporation mechanism, but do not argue that critique loses all effectiveness in the process. Perhaps the homelab is simultaneously a product of the system and a real, if partial, form of withdrawal from it. The two things are not mutually exclusive.&#xA;&#xA;The third: if digital autonomy requires decades of accumulated skills, enough free time to use them, and enough money to buy the hardware, are we building a democratic alternative? Or are we building an exclusive club with a rebel aesthetic, reproducing the same hierarchies of privilege it claims to want to fight?&#xA;&#xA;The fourth: the energy question has no clean answer, and Jevons&#39; Paradox makes it even more uncomfortable - because it works in both directions. The cloud improves efficiency and increases total consumption. A homelab consumes proportionally more, but does not fuel the demand that drives that total consumption upwards. Are we building digital sovereignty, or are we simply choosing where to position ourselves within a contradiction that cannot be resolved at the individual level?&#xA;&#xA;I don&#39;t know. But at least I know where my data is.&#xA;&#xA;Fun Fact&#xA;&#xA;This article was written in Markdown using a Flatnotes instance running as a CT container on Proxmox, while listening to a symphonic metal playlist served by Navidrome - another CT container - pulling OGG files from a ZFS NAS over an NFS share. The cited books were in EPUB format on Calibre Web. In the background, Nextcloud on a Raspberry Pi 4 was syncing and backing up everything. Spelling mistakes were corrected by Qwen2.5, an LLM model served by Ollama on the MINISFORUM box, accessible locally via oterm and Open WebUI. And all of this, controlled from a laptop running Linux.&#xA;&#xA;Coincidences? I don&#39;t think so.&#xA;&#xA;a href=&#34;https://remark.as/p/jolek78/reflections-on-an-impossible-escape-from-capitalism&#34;Discuss.../a&#xA;&#xA;#Homelab #SelfHosted #SurveillanceCapitalism #Privacy #OpenSource #HackerEthic #SolarPunk #DigitalSovereignty #FOSS #Linux&#xA;&#xA;div class=&#34;center&#34;&#xD;&#xA;· 🦣 a href=&#34;https://fosstodon.org/@jolek78&#34;Mastodon/a · 📸 a href=&#34;https://pixelfed.social/jolek78&#34;Pixelfed/a ·  📬 a href=&#34;mailto:jolek78@jolek78.dev&#34;Email/a ·&#xD;&#xA;· ☕ a href=&#34;https://liberapay.com/jolek78&#34;Support this work on Liberapay/a&#xD;&#xA;/div]]&gt;</description>
      <content:encoded><![CDATA[<p>It was an ordinary Friday evening. The parcel had arrived with the courier that morning, but I only opened it after dinner, with that silent ceremony I perform every time new hardware shows up – as if opening a box too quickly were a form of disrespect toward the object. Inside was a HUNSN 4K. Small, almost ridiculously small. A mini PC in a form factor that fit in the palm of a hand. I put it on the table, looked at it. Looked at it again. And then an uncomfortable thought occurred to me. I had ordered it from a Chinese reseller, paid with a credit card, through a completely traceable payment infrastructure, from one of the most centralised and surveilled commercial ecosystems in existence. To build a homelab that would let me escape centralised and surveilled ecosystems.</p>



<p>The funny thing – funny in the sense that it makes you laugh, but badly – is that I&#39;m not alone. Every day, somewhere in the world, someone orders a mini PC, a Raspberry Pi, a managed Mikrotik switch, with the stated goal of taking back control of their digital life. They order it on Alibaba, pay with PayPal, wait for the courier. And they see nothing strange in any of this, because the contradiction is so structural it has become invisible. This article is an attempt to make it visible again. Without easy solutions, because I don&#39;t have any. And when have I ever…</p>

<h2 id="the-promise-of-the-homelab" id="the-promise-of-the-homelab">The Promise of the Homelab</h2>

<p>When, in 2019, I started self-hosting pretty much everything – Nextcloud (always on a Raspberry Pi, first RPi3 then RPi4), Jellyfin, Navidrome, FreshRSS, and about twenty-five other services on Proxmox LXC, each with its own isolated Docker daemon – I did it with a precise motivation: I wanted to know where my data lived, who could read it, and have the ability to switch it off myself if I ever felt like it. Not when a company decides to shut down a service, not when someone else changes the licence terms. Me. This came after a long period of reflection on myself, the work I was doing and still do, and the technological society I live in. It is an ideological choice before it is a technical one. Technology as a tool for autonomy rather than control; infrastructure as something you own instead of something that owns you. I hope no one is alarmed if I say that some of these reflections began with reading Theodore Kaczynski&#39;s Manifesto, before eventually landing, of course, on more authoritative sources.</p>

<p>Yes, I&#39;m mad, but not quite that mad…</p>

<p>When you pay a subscription to a cloud service, the transaction does not end the moment you authorise the electronic payment. Shoshana Zuboff, in <em>The Age of Surveillance Capitalism</em>, calls this mechanism <em>behavioral surplus</em>: the behavioural data extracted beyond what is needed to provide the service, then resold as predictive raw material.</p>

<blockquote><p>Under the regime of surveillance capitalism, however, the first text does not stand alone; it trails a shadow close behind. The first text, full of promise, actually functions as the supply operation for the second text: the shadow text. Everything that we contribute to the first text, no matter how trivial or fleeting, becomes a target for surplus extraction. That surplus fills the pages of the second text. This one is hidden from our view: “read only” for surveillance capitalists. In this text our experience is dragooned as raw material to be accumulated and analyzed as means to others&#39; market ends. The shadow text is a burgeoning accumulation of behavioral surplus and its analyses, and it says more about us than we can know about ourselves. Worse still, it becomes increasingly difficult, and perhaps impossible, to refrain from contributing to the shadow text. It automatically feeds on our experience as we engage in the normal and necessary routines of social participation.</p></blockquote>

<p>You are not the customer of the system – you are its product. Your habits, your schedules, your preferences, your hesitations before clicking on something: all of this is collected, modelled, sold. The transaction is not monthly: it is continuous, invisible, and never ends as long as you use the service. With hardware, in principle, the transaction is one-time: you buy, you pay, it ends, it is yours. The disk is in your room, not on a server subject to government requests, security breaches, or business decisions that are nothing to do with you but impact your access to those services. This distinction – between a tool you use and a system that uses you – is the real stake of the homelab. It is not about saving money, it is not about performance. It is about who controls what.</p>

<p>The problem is that building this infrastructure requires hardware, time, knowledge, and resources. The hardware comes from somewhere; the time, the knowledge, and the energy resources come from a privilege not granted to everyone.</p>

<h2 id="the-market-i-hadn-t-seen" id="the-market-i-hadn-t-seen">The Market I Hadn&#39;t Seen</h2>

<p>Search for “mini PC homelab” on any marketplace. What you find is a productive ecosystem that has exploded over the past five years in a way I honestly did not expect.</p>

<p>MINISFORUM, Beelink, Trigkey, Geekom, GMKtec. Zimaboard, with its single-board aesthetic designed explicitly for those who want home racks. Raspberry Pi and the galaxy of clones – Orange Pi, Rock Pi, Banana Pi. Managed Mikrotik switches at accessible prices. 1U rack cases to mount under the desk. M.2 NVMe SSDs with TBW figures calculated for small-server workloads. Silent power supplies designed to run 24/7. A market built from scratch, that exists precisely because there is a community of people who want to run servers at home. r/homelab and r/selfhosted on Reddit have approximately 2.8 and 1.7 million members respectively – numbers publicly verifiable, and growing. YouTube is full of dedicated channels. There is an entire attention economy built around “escaping” the attention economy.</p>

<p>But it is worth asking: who built this market, and why. MINISFORUM and Beelink do not exist out of ideological sympathy for the homelab movement. They exist because they identified a profitable segment and served it with industrial precision. Kate Crawford, in <em>Atlas of AI</em>, documents how technology supply chains follow niche demand with the same efficiency with which they follow mass demand: factories in Guangdong optimise production lines not for a worldview, but for a margin. The fact that the resulting product also satisfies an ideological need is, from the manufacturer&#39;s point of view, irrelevant.</p>

<blockquote><p>The Victorian environmental disaster at the dawn of the global information society shows how the relations between technology and its materials, environments, and labor practices are interwoven. Just as Victorians precipitated ecological disaster for their early cables, so do contemporary mining and global supply chains further imperil the delicate ecological balance of our era.</p></blockquote>

<p>The mechanism had been described with theoretical precision back in 1999 by Luc Boltanski and Ève Chiapello in <em>The New Spirit of Capitalism</em>. Their thesis: capitalism is never defeated by criticism – it is incorporated. When a critique becomes widespread enough, the system absorbs it and transforms it into a market segment. The artistic critique of the 1960s – autonomy, authenticity, rejection of standardisation – became the marketing of the creative economy. The critique of digital centralisation – sovereignty, privacy, control – has become an online catalogue to browse through.</p>

<p>Resistance has become a market segment. Every time someone buys a HUNSN to stop paying subscriptions to services they don&#39;t control, a factory in Guangdong sells a HUNSN. Capitalism has not been defeated – it has shifted (at least for a small slice of the population: the nerds, the hackers) the extraction point from subscriptions to hardware.</p>

<h2 id="the-accumulation-syndrome" id="the-accumulation-syndrome">The Accumulation Syndrome</h2>

<p>But there is a further level – more ridiculous and more personal – that homelab communities never discuss openly, yet anyone who has a homelab recognises immediately. The Raspberry Pi 4 bought “for a project.” The old ThinkPad kept because “you never know.” The 4TB disk salvaged from a decommissioned NAS – and “it might come in handy.” The second-hand switch picked up on eBay for eighteen euros because it was cheap and might be useful. The cables, the cables, the cables.</p>

<p>r/homelab has a term for this: <em>just in case hardware</em>. It is the hardware of the imaginary future, of projects that only exist in your head, of configurations that one day – one day – you will finally test. In the meantime it occupies a shelf, draws current in standby, and generates a diffuse sense of possibility that is indistinguishable from the most classic consumerism. The underlying psychological mechanism has a precise name: <em>compensatory consumption</em> – consumption as a response to a perceived loss of autonomy or control. You buy hardware because buying hardware gives you the feeling of recovering agency over something. The aesthetic is different from traditional consumerism – no luxury logos, no recognisable status symbols – but the mechanism is identical.</p>

<p>That said, there is a partially honest answer to all of this: the second-hand and refurbished market. The ThinkPad X230 on eBay, the Dell R720 server decommissioned from a datacentre, the disk from someone who upgraded their NAS. My ZFS NAS, to give one example, is a recycled old tower with four 1TB disks in RAIDZ – hardware that would otherwise have ended up in landfill, with a life cycle extended by years, without generating new production demand. It is closer to the ethics of repair than to compulsive buying. But it has its own internal contradiction: it requires even more technical competence than buying new – knowing how to assess wear, diagnose an unknown component, manage ten-year-old drivers. The barrier to entry rises further. And the refurbished market is itself now an organised commercial sector, with its own margins, its own platforms, its own pricing logic. It is not a clean way out. It is a less dirty way out.</p>

<p>And then there is the energy question, which is usually ignored in homelab discussions and is instead the most uncomfortable of all – uncomfortable enough to deserve a more in-depth treatment later on. For now, suffice it to say: every machine on your shelf that “draws current in standby” is a line item in the energy bill that the homelab movement rarely accounts for.</p>

<h2 id="not-for-everyone-and-it-should-not-be-this-way" id="not-for-everyone-and-it-should-not-be-this-way">Not for Everyone. And It Should Not Be This Way.</h2>

<p>There is a second level of the paradox that is even more uncomfortable than the first. Building a homelab costs money – relatively little, but it costs. It requires physical space. It requires a decent connection. And it requires time. A lot of time. Not installation time – that is measurable, finite. The learning time that precedes everything else. To reach the point where you can build a functional infrastructure with Proxmox, LXC containers, centralised authentication, reverse proxy, automated backups – you need to have already spent years understanding how Linux works, how to reason about networks and permissions, how to read a log. I started with a Red Hat in 1997, and it took me almost thirty years to get where I am. I should know this. Yet it always escapes me. And that time did not fall from the sky. It is time I was able to dedicate because I had a certain kind of job, a certain stability, a certain amount of mental energy left at the end of the day. It is middle-class-with-a-stable-position time, not the time of someone working three warehouse shifts a week. Passion is not enough.</p>

<p>Johan Söderberg documents this in <em>Hacking Capitalism</em>: the FOSS movement was born as resistance to capitalism, but reproduces within itself hierarchies of skill and merit that make it structurally exclusive. Freedom is technically available to anyone, but effective access requires resources distributed in anything but a democratic manner. Söderberg goes further than simply observing the exclusivity: the voluntary open source work produces use value – functioning software, documentation, community support – that capital then extracts as <em>exchange value</em> without remunerating those who produced it. Red Hat builds a billion-dollar company on a kernel written largely by volunteers. It is not just that not everyone can get in: it is that those who get in often work for someone without knowing it. The homelab inherits this problem and amplifies it.</p>

<blockquote><p>The narrative of orthodox historical materialism corresponds with some very popular ideas in the computer underground. It is widely held that the infinite reproducibility of information made possible by computers (forces of production) has rendered intellectual property (relations of production, superstructure) obsolete. The storyline of post-industrial ideology is endorsed but with a different ending. Rather than culminating in global markets, technocracy and liberalism, as Daniel Bell and the futurists would have it; hackers are looking forward to a digital gift economy and high-tech anarchism. In a second turn of events, hackers have jumped on the distorted remains of Marxism presented in information-age literature, and, while missing out on the vocabulary, ended up promoting an upgraded Karl Kautsky-version of historical materialism.</p></blockquote>

<p>This is not a quirk of the homelab movement: it is a recurring structure in every technological wave. Langdon Winner, in his influential essay <em>Do Artifacts Have Politics?</em>, argued that technological choices are never neutral – they incorporate power structures, distribute access in non-random ways. Amateur radio in the 1920s, the personal computer in the 1980s, the internet in the 1990s: every time the promise was democratising, every time the actual distribution followed the lines of pre-existing privilege. Not out of malice, but out of structure. The irony is this: those who would most need digital autonomy – those who cannot afford subscriptions, those who live under governments that surveil communications, those most exposed to data collection – are exactly those least likely to be able to build a homelab. Not for lack of interest or intelligence. For lack of time, money, and years of privileged exposure to technology.</p>

<p>Homelab communities do not usually talk about this. They talk about which mini PC to buy, how to optimise energy consumption, which distro to use as a base. The conversation about structural exclusivity exists, but at the margins – in Jacobin, in Logic Magazine, in EFF activism – while the centre of the discourse remains impermeable. It is not that no one speaks about it: it is that the peripheries speak about it, and the peripheries do not set the agenda. This entire conversation takes place in a room to which not everyone has a ticket. And those inside do not seem to find that particularly problematic.</p>

<h2 id="a-technological-cosplay" id="a-technological-cosplay">A Technological Cosplay?</h2>

<p>So is the whole thing a con? Is the homelab just anti-capitalist cosplay while you continue to fund the same supply chains? In part, yes.</p>

<p>The HUNSN 4K was designed in China, assembled in China, shipped by container on ships burning bunker fuel. Global maritime transport is responsible for approximately 2.5% of global CO₂ emissions – a share that the IMO (International Maritime Organization) has been trying to reduce for years with slow progress and targets continually postponed. Then: distributed through Alibaba, paid with a credit card. Every piece of technology hardware carries an extractive chain that begins in lithium mines in Bolivia and cobalt mines in the Democratic Republic of the Congo, passes through factories in Guangdong, and ends in electronic waste processing centres in Ghana. The hardware travels that supply chain exactly like any other consumer device. Furthermore, hardware has a lifecycle. In five years the HUNSN 4K will be too slow, or it will break, or something will come out with energy efficiency too much better to ignore. And I will buy again. The mini PC market for homelabs depends on the obsolescence of previous purchases – exactly like any other consumer market.</p>

<p>The critique of capitalism, when it is widespread enough, is not suppressed – it is incorporated. The system absorbs the values of resistance and transforms them into a market segment. Autonomy becomes a selling point. Decentralisation becomes a brand. The rebel who wanted to exit the system finds himself funding a new vertical of the same system, convinced he is making an ethical choice.</p>

<h2 id="the-counter-shot" id="the-counter-shot">The Counter-Shot</h2>

<p>But there is a structural difference that would be dishonest to ignore.</p>

<p>When you pay a subscription to a cloud service, the cost is not just the monthly fee. It is the continuous cession of data, behaviours, habits. It is the behavioral surplus Zuboff talks about: you are not using a service, you are being used as raw material to train models, build profiles, sell advertising. The transaction never ends, in ways you often cannot see and cannot escape from as long as you use the service.</p>

<p>With hardware, the transaction ends. The data stays on a physical disk in your room, not on a server subject to government requests, breaches, or business decisions that have nothing to do with you but impact your life. The software running on it – Proxmox, Debian, Nextcloud, Jellyfin – is open source; you can modify it. If something changes in a way you cannot accept, you can leave. This resilience has real value – but it is worth noting that it is asymmetric resilience: it works for those who have the skills to exercise it. For those who do not, the theoretical portability of their data from Nextcloud to something else requires exactly the same skills we have already identified as the barrier to entry. The freedom to leave is real. Access to that freedom, much less so.</p>

<p>And then there is the energy question, which I have deferred long enough. The major hyperscalers – AWS, Google, Azure – operate with a PUE (Power Usage Effectiveness) between 1.1 and 1.2. For every watt of useful computation they dissipate barely 0.1–0.2 watts in heat and infrastructure. They have enormous economies of scale, optimised industrial cooling, significant investments in renewable energy, and above all: their servers run at very high utilisation rates. Almost always busy.</p>

<p>A home homelab works in a radically different way. The machine runs 24/7 even when it is doing nothing – and for most of the time it is doing nothing. Navidrome serving three requests a day, FreshRSS fetching every hour, an LDAP container sitting listening without receiving connections. You are paying the energy cost of the infrastructure regardless of usage. The implicit PUE of a homelab, calculated honestly on the ratio between total consumption and actual workload, is much worse than that of a datacentre. IEA data (<em>Data Centres and Data Transmission Networks</em>, updated annually) shows that large cloud providers progressively improve energy efficiency thanks to economies of scale that no individual homelab can replicate. The flip side is that the same growth in demand that makes economies of scale possible negates the efficiency gains: Amazon&#39;s absolute emissions increased between 2023 and 2024 despite improved PUE. Efficiency improves. Total consumption grows anyway. This is Jevons&#39; Paradox: energy efficiency, instead of reducing consumption, increases it, because it lowers the marginal cost of use and stimulates demand that grows faster than the efficiency gains.</p>

<blockquote><p><em>Note: The comparison is not as linear as the numbers suggest. PUE measures the internal efficiency of a datacentre, not the energy cost of the network traffic that data generates every time it leaves it – traffic that a homelab eliminates almost completely for internal services. Nor does it measure proportion: AWS is efficient at delivering services to millions of users, but that scale says nothing about the real cost of storing fifty gigabytes of personal data on a server designed for loads a thousand times greater. A HUNSN N100 in idle consumes less than 8 watts. The honest energy comparison is not homelab vs hyperscaler in the abstract – it is homelab vs proportional share of hyperscaler for your specific workload, a calculation that nobody can make with publicly available data.</em></p></blockquote>

<p>This does not automatically mean that the cloud is the ethically correct choice – the problem does not reduce to PUE, and surveillance has costs that are not measured in kilowatts. It means that anyone with SolarPunk values who chooses the homelab must reckon with a real contradiction: the choice of sovereignty may be, watt for watt, energetically more costly than the system one wants to escape. I have no clean answer, but ignoring the question would be dishonest. Söderberg acknowledges that the FOSS movement has produced concrete and undeniable gains – they simply are not enough, on their own, to subvert the dynamics of informational capitalism.</p>

<p>In short: this is not a critique of the homelab, but it is a critique of the homelab presented as a sufficient revolutionary act.</p>

<h2 id="what-happens-at-eleven-pm-and-beyond" id="what-happens-at-eleven-pm-and-beyond">What Happens at Eleven PM – and Beyond</h2>

<p>That night, with the HUNSN 4K on the table, I pressed on. I installed Proxmox. I configured the network. I started bringing up containers one by one. And at some point – three hours had passed, I had three terminals open and was debugging nslcd to centralise LDAP authentication across all the containers – I realised something: I was doing all of this simply because I enjoyed it. Not to resist something. Not to advance an ideological agenda. Because there was a problem to solve and solving it gave me satisfaction. Mihaly Csikszentmihalyi describes this state in <em>Flow</em> as total absorption in a task calibrated to one&#39;s own competencies: time expands, attention narrows, awareness of context vanishes. It is not motivation – it is something more immediate. Debugging an authentication problem at eleven at night on a system I could have chosen not to build is, neuropsychologically, indistinguishable from pleasure. Not the satisfaction of having finished: the process itself. Moreover, for an AuDHD person like me, going into hyperfocus allows you to lose your sense of time entirely, and to literally escape from a world you viscerally loathe.</p>

<p>Ah – you had not figured that out yet?</p>

<p>When I had finished and closed everything, the satisfaction was still there. Along with a mildly uncomfortable awareness: I could probably have used a hosted service, lived just as well, and not lost three hours of a weeknight. But in the meantime I had understood how PAM worked, I had read documentation I had never opened before, I had implemented it on my homelab, I had learned something I hadn&#39;t known I wanted to know.</p>

<p>And here the circle closes in a somewhat unsettling way. Söderberg speaks of voluntary open source work as the production of pure use value – the intrinsic pleasure of doing, understanding, building something that works. But it is exactly this use value that capital then extracts as exchange value: the competence I accumulate debugging LDAP at eleven at night is the same competence I bring to work the next day, that I put into articles like this one, that I share in communities where others use it to build their own homelabs. Technical pleasure is not neutral. It has a production chain. Not always visible, but real.</p>

<p>This is what the homelab is, at least for me: a way of learning that produces, as a side effect, an infrastructure I control. The ideology is there, but it comes second. First comes the pleasure of understanding how something works. Or rather: ideology and pleasure are interchangeable, and often run in parallel – but this does not resolve any of the contradictions I described above. It leaves them all standing, in fact makes them stranger. Am I resisting capitalism, or am I just cultivating an expensive hobby with a political aesthetic?</p>

<h2 id="the-hacker-ethic" id="the-hacker-ethic">The Hacker Ethic</h2>

<p>The word “hacker” has had bad press for decades. In 1990s news bulletins it was a synonym for a hooded cybercriminal; in the jargon of security companies it became a marketing term to prepend to anything. Neither has much to do with what the word historically means. Steven Levy, in <em>Hackers: Heroes of the Computer Revolution</em>, reconstructs the culture that formed around the MIT and Stanford labs in the 1960s: a community of programmers for whom code was an aesthetic object, access to information a moral principle, and technical competence the only legitimate hierarchy. The principles Levy identifies as the “hacker ethic” are precise: access to computers – and to anything that can teach you how the world works – should be unlimited and total. All information should be free. Decentralised systems are preferable to centralised ones. Hackers should be judged by what they produce, not by titles, age, race, or position. You can create art and beauty with a computer.</p>

<p>It is not a political manifesto in the traditional sense. It is something more visceral – a disposition toward the world, a way of standing before a system you do not yet understand: the correct response is to take it apart, understand how it works, and put it back together better than before.</p>

<p>Pekka Himanen, in <em>The Hacker Ethic and the Spirit of the Information Age</em> – with a preface by Linus Torvalds and an epilogue by Manuel Castells, which already says something about the project&#39;s ambition – performs a more explicit theoretical operation. He builds the hacker ethic in direct opposition to the Protestant work ethic described by Max Weber: where Weber saw work as duty, discipline as virtue, and leisure as absence of production, Himanen identifies in the hacker a figure who works out of passion, considers play an integral part of work, and rejects the sharp separation between productive time and free time. The hacker does not work for money – money is a side effect, when it comes. They work because the problem is interesting. Because the elegant solution has value in itself. Because understanding how something works is, in and of itself, sufficient.</p>

<blockquote><p>Hacker activity is also joyful. It often has its roots in playful explorations. Torvalds has described, in messages on the Net, how Linux began to expand from small experiments with the computer he had just acquired. In the same messages, he has explained his motivation for developing Linux by simply stating that “it was/is fun working on it.” Tim Berners-Lee, the man behind the Web, also describes how this creation began with experiments in linking what he called “play programs.” Wozniak relates how many characteristics of the Apple computer “came from a game, and the fun features that were built in were only to do one pet project, which was to program … [a game called] Breakout and show it off at the club.”</p></blockquote>

<p>Recognise something? I do. Those three hours debugging nslcd at eleven at night were not work in the Weberian sense – nobody was paying me, nobody had asked me to do it, there was no corporate objective to reach. They were hacking in the precise sense that Levy and Himanen describe: exploration motivated by curiosity, with the infrastructure as an object of study as much as of utility. The homelab is, culturally, a direct expression of the hacker ethic. It is no coincidence that homelab communities and open source communities overlap almost perfectly, that they use the same language, the same platforms, the same values. But here, as elsewhere in this article, the story gets complicated.</p>

<p>The hacker ethic promises a pure meritocracy: you are judged by what you can do, not by who you are. It is an attractive idea. It is also, in practice, a partial fiction. Technical meritocracy presupposes that everyone starts from the same point – that skills are accessible to anyone who really wants to acquire them, that the time to acquire them is distributed equally, that mentorship networks and learning resources are available regardless of context. The homelab as hacker practice inherits both things: the genuine nature of curiosity as a driver, and structural exclusivity as an undeclared side effect. The pleasure of taking a system apart to understand how it works is real and should not be devalued. But that pleasure is available, in practice, to those who already have the ticket.</p>

<h2 id="conclusions" id="conclusions">Conclusions</h2>

<p>The HUNSN 4K runs, alongside the other “little electronic contraptions,” on a rack next to my armchair – the one where, at the end of the day, I indulge my guilty pleasure of reading a book in the company of my cats. Proxmox, the Nextcloud server, the ZFS NAS, a small MINISFORUM box running Ollama with some local open-weight LLM models, a Raspberry Pi 5 running the Tor Relay, and a HUNSN RJ15 with pfSense controlling incoming and outgoing traffic. An infrastructure, in short, that allows me to have something resembling digital sovereignty within the limits of the possible. The contradictions I have described do not resolve. They are held together, with effort, as any intellectually complex position on a complex system must be held together.</p>

<p>The first: the market that made the accessible homelab possible is the same market the homelab is supposed to emancipate us from. If this explosion of affordable, efficient mini PCs had not happened – if capitalism had not decided to build exactly what we wanted – how many of us would have taken the same path? How much of our “ethical choice” depends on the existence of products designed and sold precisely for us?</p>

<p>The second: does incorporated resistance truly lose its force, or does it remain resistance even when someone profits from it? Boltanski and Chiapello describe the incorporation mechanism, but do not argue that critique loses all effectiveness in the process. Perhaps the homelab is simultaneously a product of the system and a real, if partial, form of withdrawal from it. The two things are not mutually exclusive.</p>

<p>The third: if digital autonomy requires decades of accumulated skills, enough free time to use them, and enough money to buy the hardware, are we building a democratic alternative? Or are we building an exclusive club with a rebel aesthetic, reproducing the same hierarchies of privilege it claims to want to fight?</p>

<p>The fourth: the energy question has no clean answer, and Jevons&#39; Paradox makes it even more uncomfortable – because it works in both directions. The cloud improves efficiency and increases total consumption. A homelab consumes proportionally more, but does not fuel the demand that drives that total consumption upwards. Are we building digital sovereignty, or are we simply choosing where to position ourselves within a contradiction that cannot be resolved at the individual level?</p>

<p>I don&#39;t know. But at least I know where my data is.</p>

<h2 id="fun-fact" id="fun-fact">Fun Fact</h2>

<p>This article was written in Markdown using a Flatnotes instance running as a CT container on Proxmox, while listening to a symphonic metal playlist served by Navidrome – another CT container – pulling OGG files from a ZFS NAS over an NFS share. The cited books were in EPUB format on Calibre Web. In the background, Nextcloud on a Raspberry Pi 4 was syncing and backing up everything. Spelling mistakes were corrected by Qwen2.5, an LLM model served by Ollama on the MINISFORUM box, accessible locally via oterm and Open WebUI. And all of this, controlled from a laptop running Linux.</p>

<p>Coincidences? I don&#39;t think so.</p>

<p><a href="https://remark.as/p/jolek78/reflections-on-an-impossible-escape-from-capitalism">Discuss...</a></p>

<p><a href="https://jolek78.writeas.com/tag:Homelab" class="hashtag"><span>#</span><span class="p-category">Homelab</span></a> <a href="https://jolek78.writeas.com/tag:SelfHosted" class="hashtag"><span>#</span><span class="p-category">SelfHosted</span></a> <a href="https://jolek78.writeas.com/tag:SurveillanceCapitalism" class="hashtag"><span>#</span><span class="p-category">SurveillanceCapitalism</span></a> <a href="https://jolek78.writeas.com/tag:Privacy" class="hashtag"><span>#</span><span class="p-category">Privacy</span></a> <a href="https://jolek78.writeas.com/tag:OpenSource" class="hashtag"><span>#</span><span class="p-category">OpenSource</span></a> <a href="https://jolek78.writeas.com/tag:HackerEthic" class="hashtag"><span>#</span><span class="p-category">HackerEthic</span></a> <a href="https://jolek78.writeas.com/tag:SolarPunk" class="hashtag"><span>#</span><span class="p-category">SolarPunk</span></a> <a href="https://jolek78.writeas.com/tag:DigitalSovereignty" class="hashtag"><span>#</span><span class="p-category">DigitalSovereignty</span></a> <a href="https://jolek78.writeas.com/tag:FOSS" class="hashtag"><span>#</span><span class="p-category">FOSS</span></a> <a href="https://jolek78.writeas.com/tag:Linux" class="hashtag"><span>#</span><span class="p-category">Linux</span></a></p>

<div class="center">
· 🦣 <a href="https://fosstodon.org/@jolek78">Mastodon</a> · 📸 <a href="https://pixelfed.social/jolek78">Pixelfed</a> ·  📬 <a href="mailto:jolek78@jolek78.dev">Email</a> ·
· ☕ <a href="https://liberapay.com/jolek78">Support this work on Liberapay</a>
</div>
]]></content:encoded>
      <guid>https://jolek78.writeas.com/reflections-on-an-impossible-escape-from-capitalism</guid>
      <pubDate>Sun, 05 Apr 2026 15:46:47 +0000</pubDate>
    </item>
    <item>
      <title>Legacy systems: problem or resource?</title>
      <link>https://jolek78.writeas.com/legacy-systems-problem-or-resource?pk_campaign=rss-feed</link>
      <description>&lt;![CDATA[Tuesday morning, 9 AM. After a routine patching session, a long-standing ZFS storage system running Solaris 11 suddenly stops talking to its Windows 10 clients. The culprit is the usual, maddening SMB dialect dance: Windows pushes for SMB 3 on security grounds, while Solaris&#39;s native service struggles through the negotiation. Two days of banging my head against the wall - hard - and then the discovery: OpenCSW. A community that maintains updated packages for Solaris where the vendor long since threw in the towel. Updated libraries, sorted dependencies, problem solved. There are volunteers out there patching critical systems better than the official vendor ever did. Worth knowing.&#xA;&#xA;!--more--&#xA;&#xA;Same film, next scene.&#xA;&#xA;Friday afternoon - because critical migrations always happen out of hours. I&#39;m migrating a system from Red Hat 7 to Red Hat 9. Why? To support the new version of Charon-SSP, the Stromasys emulator that lets SPARC hardware run on x86. All of this to keep alive a virtual machine running Solaris 9, an operating system from 2002 that went end-of-life in 2014. It&#39;s a layered structure, each level propping up the one below. One of those classic houses of cards you can&#39;t quite understand how it stays balanced.&#xA;&#xA;Welcome to the world of legacy systems. A world where &#34;modernising&#34; often means finding increasingly creative ways to change nothing at all, and where communities and old-school sysadmins are the ones guarding infrastructure that corporations abandoned long ago. Try asking Oracle for Solaris support: they&#39;ll laugh in your face.&#xA;&#xA;The numbers&#xA;&#xA;In January 2025, the UK government published a report that should have rattled a few chairs at Westminster. Twenty-eight percent of central government IT systems are classified as legacy - up from 26% in 2023. Estimated productivity losses? Forty-five billion pounds. In 2024, the NHS recorded 123 critical IT system crashes. One hundred and twenty-three.&#xA;&#xA;But wait, because the numbers get even more interesting when you look at the banking sector. COBOL - a programming language dating back to 1959 - still processes 95% of global ATM transactions, 43% of the world&#39;s banking systems, and around 3 trillion dollars of commerce every day. Every day. It&#39;s estimated there are still 220 billion lines of COBOL code in production.&#xA;&#xA;And Windows XP? The one Microsoft stopped supporting in 2014? Today, 1-2% of internet-connected devices still run it. Sounds small until you realise we&#39;re talking about millions of machines. And not your grandad&#39;s PC: we&#39;re talking about MRI scanners in hospitals, industrial control systems, bank ATMs. Critical devices that can&#39;t be updated because the software controlling them only runs on XP, and re-certifying the entire system would cost more than building a new one.&#xA;&#xA;Remember WannaCry in 2017? The ransomware that paralysed 75,000 computers in 99 countries? The NHS was devastated. And do you know how many Windows XP machines the NHS had in 2019 - two years after the attack, five years after end-of-support? 2,300.&#xA;&#xA;At this point in the story one might say &#34;right, the problem is clear: legacy systems are dangerous and need replacing.&#34; And that would be the easy narrative - the one that consultants selling &#34;digital transformation&#34; love, and vendors wanting to sell licences love. What if I told you that a Solaris 11 system, properly isolated in a VLAN, is significantly more stable and secure than a shiny new Ubuntu 24.04 LTS?&#xA;&#xA;Reality, as always, is more complicated.&#xA;&#xA;Problems upon problems&#xA;&#xA;Here&#39;s the fundamental issue: we use the word &#34;legacy&#34; as if it meant one thing, when it actually covers at least three completely different situations.&#xA;&#xA;Type 1: Unavoidable legacy&#xA;Solaris 9 on SPARC hardware controlling industrial machinery. Windows XP on MRI scanners. Systems where hardware and software are inseparable, where an upgrade would require replacing equipment worth millions, where re-certification for medical or industrial use would take years and fortunes. These systems are legacy out of necessity, not negligence. There&#39;s no fault here. There&#39;s only the reality of a technological ecosystem where certain devices have 20-30 year lifespans and the software controlling them can&#39;t be changed without changing everything else.&#xA;&#xA;Type 2: Avoidable legacy&#xA;CentOS 7, for instance. End of support: 30 June 2024. Available alternatives: AlmaLinux, Rocky Linux, migration to RHEL. Cost of migration? Economically: it depends. In time, resources, learning: enormous. How many CentOS 7 systems are still in production today? Too many. Why? Because nobody wants to pay RHEL licences, because &#34;we&#39;ll do it next quarter,&#34; because &#34;there are other important things to deal with,&#34; because &#34;if it ain&#39;t broke, don&#39;t fix it.&#34; This is legacy by choice - or rather, by inertia. It&#39;s an organisational decision, not a technical one.&#xA;&#xA;Type 3: Non-legacy perceived as legacy&#xA;Take COBOL on modern IBM mainframes. Today&#39;s mainframes aren&#39;t the ones from the 1970s - they&#39;re immensely powerful machines, with dedicated processors, hardware security, 99.999% uptime. The COBOL running on them is the same as ever, but the underlying infrastructure is current. Is the code legacy, or the platform? And if the platform is modern, can we still call it legacy? The distinction is fundamental because it determines the strategy. A Type 1 system needs to be isolated and protected. A Type 2 system needs to be migrated. A Type 3 system needs to be left alone. Try explaining that to a CTO who just finished reading a Gartner report on &#34;legacy modernisation.&#34;&#xA;&#xA;From a thread on TheLayoff:&#xA;&#xA;  &#34;FWIW, there&#39;s a very good chance that your electronic footprint on any given day has passed through a piece of SPARC equipment running Solaris, and that will continue to happen for a good portion of your lifetime.&#34;&#xA;&#xA;Would you believe me if I told you I&#39;ve seen original BSD systems with eleven years of uptime?&#xA;&#xA;The real problem isn&#39;t the machines&#xA;&#xA;Here we get to the heart of the matter. And the answer will surprise you: the real problem with legacy systems isn&#39;t technological. It&#39;s human.&#xA;&#xA;Let&#39;s talk about the &#34;COBOL Cowboys&#34; - retired programmers called back on consulting contracts when something breaks. They&#39;re the last generation that knows how those systems actually work. When they leave, they take decades of undocumented knowledge with them. According to Deloitte, companies have seen a 23% decline in mainframe workforce over the last five years, with 63% of those positions left unfilled. It&#39;s not that there&#39;s no money to hire - it&#39;s that there&#39;s nobody to hire. Young developers don&#39;t want to learn COBOL. It&#39;s &#34;unsexy.&#34; It&#39;s &#34;archaic.&#34; It&#39;s &#34;boomer stuff.&#34;&#xA;&#xA;From ComputerWeekly:&#xA;&#xA;  &#34;The retirement of the generation of experts who possess in-depth knowledge of Cobol systems is leading to a severe knowledge shortage. They have knowledge not only of the Cobol programming language, but also of the specific systems they have worked on and built over the years&#34; - Tijs van der Storm, CWI/University of Groningen&#xA;&#xA;And so we find ourselves in a paradoxical situation: systems processing trillions of dollars a day, managed by people who might die of old age before anyone learns to replace them. Knowledge transfer never happened. Documentation - where it exists - is outdated, incomplete, written in a language nobody understands anymore. And every year that passes, the gap widens.&#xA;&#xA;This is the real legacy problem. Not the systems. The people.&#xA;&#xA;When modernisation fails (spoiler: often)&#xA;&#xA;There&#39;s a story that people in the UK know well, but that strangely never comes up when &#34;digital transformation&#34; is being discussed. It&#39;s called the National Programme for IT, or NPfIT.&#xA;&#xA;Launched in 2002, it was the largest public sector IT project in British history. The goal? Modernise the entire NHS IT infrastructure. Initial budget: 6 billion pounds. Planned completion: 2010.&#xA;&#xA;In 2011, after nine years of delays, exploding costs, vendors abandoning the project, and a system that simply didn&#39;t work, the UK government announced the dismantling of NPfIT. Final estimated cost: over 10 billion pounds. For a system that was never completed.&#xA;&#xA;What went wrong? Practically everything. Top-down decisions made by politicians who didn&#39;t understand technology. Rigid contracts with vendors who didn&#39;t understand the NHS. Resistance from medical staff who hadn&#39;t been consulted. Continuously shifting requirements. Impossible integrations with existing systems.&#xA;&#xA;From TechMonitor:&#xA;&#xA;  &#34;A lack of digital and procurement capability within government has led to wasted expenditure and lack of progress on major digital transformation programmes.&#34;&#xA;&#xA;The lesson? &#34;Modernising&#34; is not automatically better than &#34;maintaining.&#34; Sometimes, the legacy system that works is preferable to the modern system that never will. But this lesson, apparently, we haven&#39;t learned. Because the dominant narrative remains the same: legacy = bad, modern = good. And consultants keep selling the shiny new thing.&#xA;&#xA;Strategies that actually work&#xA;&#xA;TL;DR: There is no single solution. There&#39;s a matrix of options ranging from virtualisation to isolation, from refactoring to API wrapping. The choice depends on the type of legacy, the budget, and the acceptable level of risk.&#xA;&#xA;The Gartner 7Rs (yes, they have a name for everything):&#xA;&#xA;Retire - Switch it off. Only works if nobody&#39;s actually using it.&#xA;Retain - Keep it as is. Sometimes the best choice.&#xA;Relocate - Move it to new infrastructure without changes.&#xA;Rehost - &#34;Lift and shift&#34; to cloud. Changes the hardware, not the software.&#xA;Replatform - Minimal changes to run on a modern platform.&#xA;Refactor - Rewrite parts of the code while maintaining functionality.&#xA;Rearchitect - Completely redesign. The riskiest and most expensive.&#xA;&#xA;Virtualisation and emulation&#xA;For systems on proprietary architectures (SPARC, VAX, Alpha, PA-RISC), solutions like Stromasys Charon emulate the original hardware on x86-64 platforms. The operating system and software don&#39;t change - only the iron underneath does. For legacy x86 systems (Windows XP, Server 2003, old Linux), standard virtualisation (Proxmox, VMware, KVM) allows you to &#34;freeze&#34; the environment and keep it running indefinitely. I&#39;ve seen Proxmox setups running Windows 3.11. I&#39;m not joking.&#xA;&#xA;Network isolation&#xA;If a system can&#39;t be patched, it can at least be isolated. Dedicated VLANs, restrictive firewalls, air-gap where possible. It doesn&#39;t fix the problem, but it limits the impact in case of compromise.&#xA;&#xA;API wrapping&#xA;Put a modern REST layer in front of a legacy system. The mainframe keeps doing what it knows how to do; the outside world talks to the API. This is the strategy many banks use to expose COBOL functionality to mobile applications.&#xA;&#xA;The public sector: a special case&#xA;&#xA;Those who work in the public sector know that the dynamics differ from the private sector in ways that make the legacy problem even more complex.&#xA;&#xA;Multi-year budgets. You can&#39;t decide in January to modernise a system and have the money by March. Funding cycles are long, rigid, subject to political priorities that change with every election.&#xA;&#xA;Procurement. Buying software in the public sector is a bureaucratic nightmare. Tenders, compliance requirements, impact assessments, GDPR, accessibility. A purchase that takes a week in the private sector takes months here.&#xA;&#xA;Compliance. Systems handling health, education, or tax data are subject to stringent regulatory requirements. You can&#39;t simply &#34;migrate to the cloud&#34; - you have to demonstrate that the cloud complies with an endless list of standards.&#xA;&#xA;Service continuity (which in my view is the core problem). If a private company&#39;s system goes down for a day, they lose money. If a system managing national exams, or medical prescriptions, or pension payments goes down, the consequences fall on real people with no alternatives. The risk of downtime during a migration is often simply unacceptable.&#xA;&#xA;And then there&#39;s the political dimension. Every government wants to announce its own &#34;digital revolution.&#34; Nobody wants to inherit the previous government&#39;s problems. And so projects get started, abandoned, restarted, re-abandoned, in an endless cycle of waste.&#xA;&#xA;NPfIT wasn&#39;t an exception. It was the rule.&#xA;&#xA;The uncomfortable question&#xA;&#xA;At this point, the question nobody wants to ask is this: what if some legacy systems were simply… better? Not better in an absolute sense, but better for their specific purpose?&#xA;&#xA;Let me tell you something. I worked for years in environments dealing with large-scale Oracle infrastructure - the company that sells &#34;cloud transformation&#34; and &#34;modern infrastructure&#34; to half the world. And among other things, you know what got managed day to day? Old ZFS storage. Stuff that, on paper, should have been &#34;modernised&#34; years ago. Those machines had been running since before Docker existed, before Kubernetes, before &#34;cloud native&#34; became a term. And they worked. Quietly. Without drama. Nobody was in any hurry to replace them. Why would they be? In pursuit of what advantage, exactly?&#xA;&#xA;The COBOL processing bank transactions has been optimised for sixty years. Every bug has been found and fixed. Every edge case has been handled. Every possible scenario has been tested in production billions of times. It&#39;s code that has achieved a kind of perfection through Darwinian evolution. Rewriting it in Python would mean starting from scratch. New bugs. New untested scenarios. Years of instability before reaching the same level of reliability.&#xA;&#xA;And in the meantime? In the meantime, the legacy system keeps working. There&#39;s a reason banks aren&#39;t in a rush to abandon mainframes. It&#39;s not ignorance. It&#39;s not laziness. It&#39;s that they&#39;ve done the maths and understood that the risk of the new outweighs the cost of the old. And the old administrators have retired. But this is an uncomfortable truth. It doesn&#39;t sell well in PowerPoint presentations. It doesn&#39;t generate consulting contracts. It doesn&#39;t make tech headlines.&#xA;&#xA;And so we keep talking about &#34;modernisation&#34; as if it were automatically a good thing. As if &#34;new&#34; meant &#34;better.&#34; As if technology had a moral direction.&#xA;&#xA;So what?&#xA;&#xA;Legacy doesn&#39;t mean old - it means abandoned. The problem is never technical - it&#39;s always organisational. And &#34;modernising&#34; is not automatically better than &#34;maintaining.&#34;&#xA;&#xA;If there&#39;s one lesson, it&#39;s this: be suspicious of anyone with simple answers to complex problems.&#xA;&#xA;Every time I hear some manager say &#34;we need to automate everything with AI,&#34; I think about the software pachyderms holding up half of critical infrastructure. I think about the time it would take to train a model on COBOL written in 1987 with no documentation. I think about how long it would take to migrate a Java 1.7 system running on Solaris 9. I think about the hours spent reverse-engineering platforms still running Lotus Notes. I think about the costs. I think about the risks. And then I think that those same managers don&#39;t have the budget to hire juniors willing - and why should they be, when the IT world is moving in a completely different direction - to learn systems that have been decommissioned for at least thirty years. And I laugh. Bitterly, but I laugh. Then I take a few drops of CBD to calm myself down.&#xA;&#xA;Before talking about artificial intelligence - and those who know me know I&#39;m not against AI at all - perhaps we should make sure that human intelligence doesn&#39;t retire, taking years of undocumented knowledge with it. But that, evidently, is a less sexy priority to put on the slides.&#xA;&#xA;Sources and further reading&#xA;&#xA;UK government reports&#xA;– NAO: &#34;The sustainability of government IT&#34; (January 2025)&#xA;https://www.nao.org.uk/reports/local-government-financial-sustainability-2025/&#xA;– NHS Digital: Infrastructure assessment reports&#xA;https://www.bma.org.uk/advice-and-support/nhs-delivery-and-workforce/the-future/building-the-future-healthcare-infras&#xA;&#xA;COBOL and mainframes&#xA;– Reuters: &#34;Banks scramble to fix old systems&#34; (Commonwealth Bank Australia cost analysis)&#xA;https://www.reuters.com/article/technology/banks-scramble-to-fix-old-systems-as-it-cowboys-ride-into-sunset-idUSKBN17C0CN/&#xA;– IBM: &#34;COBOL Modernization&#34;&#xA;https://www.ibm.com/think/topics/cobol-modernization&#xA;&#xA;Legacy virtualisation&#xA;– Stromasys: &#34;What are legacy systems&#34;&#xA;https://www.stromasys.com/resources/what-are-legacy-systems-challenges-benefits/&#xA;– Proxmox Forums: discussions on legacy system virtualisation&#xA;https://forum.proxmox.com/tags/legacy/&#xA;&#xA;Sector analysis&#xA;– Gartner: 7Rs of Application Modernization&#xA;https://www.techtarget.com/searchCloudComputing/tip/Use-the-7-Rs-to-develop-an-app-modernization-strategy&#xA;– Deloitte: Mainframe workforce decline study&#xA;https://www.deloitte.com/us/en/insights/topics/technology-management/tech-trends/2023/future-mainframe-technology-latest-trends.html&#xA;– WSJ: How AI Can Rev Up Mainframe Modernization&#xA;https://deloitte.wsj.com/cio/how-ai-can-rev-up-mainframe-modernization-2e3c1c4a&#xA;&#xA;Case studies: failures&#xA;– Computer Weekly: &#34;What went wrong with the National Programme for IT&#34;&#xA;https://www.computerweekly.com/opinion/Six-reasons-why-the-NHS-National-Programme-for-IT-failed&#xA;– NAO: Post-implementation review NPfIT&#xA;https://www.nao.org.uk/reports/review-of-the-final-benefits-statement-for-programmes-previously-managed-under-the-national-programme-for-it-in-the-nhs/&#xA;&#xA;Security&#xA;– WannaCry incident reports&#xA;https://any.run/malware-trends/wannacry/&#xA;– NHS Windows XP audit findings (2019)&#xA;https://www.verdict.co.uk/windows-xp-nhs/&#xA;&#xA;a href=&#34;https://remark.as/p/jolek78/legacy-systems-problem-or-resource&#34;Discuss.../a&#xA;&#xA;#LegacySystems #Sysadmin #COBOL #Solaris #Linux #PublicSector #DigitalTransformation #Mainframe #OpenSource #Infrastructure&#xA;&#xA;div class=&#34;center&#34;&#xD;&#xA;· 🦣 a href=&#34;https://fosstodon.org/@jolek78&#34;Mastodon/a · 📸 a href=&#34;https://pixelfed.social/jolek78&#34;Pixelfed/a ·  📬 a href=&#34;mailto:jolek78@jolek78.dev&#34;Email/a ·&#xD;&#xA;· ☕ a href=&#34;https://liberapay.com/jolek78&#34;Support this work on Liberapay/a&#xD;&#xA;/div]]&gt;</description>
      <content:encoded><![CDATA[<p>Tuesday morning, 9 AM. After a routine patching session, a long-standing ZFS storage system running Solaris 11 suddenly stops talking to its Windows 10 clients. The culprit is the usual, maddening SMB dialect dance: Windows pushes for SMB 3 on security grounds, while Solaris&#39;s native service struggles through the negotiation. Two days of banging my head against the wall – hard – and then the discovery: OpenCSW. A community that maintains updated packages for Solaris where the vendor long since threw in the towel. Updated libraries, sorted dependencies, problem solved. There are volunteers out there patching critical systems better than the official vendor ever did. Worth knowing.</p>



<p>Same film, next scene.</p>

<p>Friday afternoon – because critical migrations always happen out of hours. I&#39;m migrating a system from Red Hat 7 to Red Hat 9. Why? To support the new version of Charon-SSP, the Stromasys emulator that lets SPARC hardware run on x86. All of this to keep alive a virtual machine running Solaris 9, an operating system from 2002 that went end-of-life in 2014. It&#39;s a layered structure, each level propping up the one below. One of those classic houses of cards you can&#39;t quite understand how it stays balanced.</p>

<p>Welcome to the world of legacy systems. A world where “modernising” often means finding increasingly creative ways to change nothing at all, and where communities and old-school sysadmins are the ones guarding infrastructure that corporations abandoned long ago. Try asking Oracle for Solaris support: they&#39;ll laugh in your face.</p>

<h2 id="the-numbers" id="the-numbers">The numbers</h2>

<p>In January 2025, the UK government published a report that should have rattled a few chairs at Westminster. Twenty-eight percent of central government IT systems are classified as legacy – up from 26% in 2023. Estimated productivity losses? Forty-five billion pounds. In 2024, the NHS recorded 123 critical IT system crashes. One hundred and twenty-three.</p>

<p>But wait, because the numbers get even more interesting when you look at the banking sector. COBOL – a programming language dating back to 1959 – still processes 95% of global ATM transactions, 43% of the world&#39;s banking systems, and around 3 trillion dollars of commerce every day. Every day. It&#39;s estimated there are still 220 billion lines of COBOL code in production.</p>

<p>And Windows XP? The one Microsoft stopped supporting in 2014? Today, 1-2% of internet-connected devices still run it. Sounds small until you realise we&#39;re talking about millions of machines. And not your grandad&#39;s PC: we&#39;re talking about MRI scanners in hospitals, industrial control systems, bank ATMs. Critical devices that can&#39;t be updated because the software controlling them only runs on XP, and re-certifying the entire system would cost more than building a new one.</p>

<p>Remember WannaCry in 2017? The ransomware that paralysed 75,000 computers in 99 countries? The NHS was devastated. And do you know how many Windows XP machines the NHS had in 2019 – two years after the attack, five years after end-of-support? 2,300.</p>

<p>At this point in the story one might say “right, the problem is clear: legacy systems are dangerous and need replacing.” And that would be the easy narrative – the one that consultants selling “digital transformation” love, and vendors wanting to sell licences love. What if I told you that a Solaris 11 system, properly isolated in a VLAN, is significantly more stable and secure than a shiny new Ubuntu 24.04 LTS?</p>

<p>Reality, as always, is more complicated.</p>

<h2 id="problems-upon-problems" id="problems-upon-problems">Problems upon problems</h2>

<p>Here&#39;s the fundamental issue: we use the word “legacy” as if it meant one thing, when it actually covers at least three completely different situations.</p>

<p><strong>Type 1: Unavoidable legacy</strong>
Solaris 9 on SPARC hardware controlling industrial machinery. Windows XP on MRI scanners. Systems where hardware and software are inseparable, where an upgrade would require replacing equipment worth millions, where re-certification for medical or industrial use would take years and fortunes. These systems are legacy out of necessity, not negligence. There&#39;s no fault here. There&#39;s only the reality of a technological ecosystem where certain devices have 20-30 year lifespans and the software controlling them can&#39;t be changed without changing everything else.</p>

<p><strong>Type 2: Avoidable legacy</strong>
CentOS 7, for instance. End of support: 30 June 2024. Available alternatives: AlmaLinux, Rocky Linux, migration to RHEL. Cost of migration? Economically: it depends. In time, resources, learning: enormous. How many CentOS 7 systems are still in production today? Too many. Why? Because nobody wants to pay RHEL licences, because “we&#39;ll do it next quarter,” because “there are other important things to deal with,” because “if it ain&#39;t broke, don&#39;t fix it.” This is legacy by choice – or rather, by inertia. It&#39;s an organisational decision, not a technical one.</p>

<p><strong>Type 3: Non-legacy perceived as legacy</strong>
Take COBOL on modern IBM mainframes. Today&#39;s mainframes aren&#39;t the ones from the 1970s – they&#39;re immensely powerful machines, with dedicated processors, hardware security, 99.999% uptime. The COBOL running on them is the same as ever, but the underlying infrastructure is current. Is the code legacy, or the platform? And if the platform is modern, can we still call it legacy? The distinction is fundamental because it determines the strategy. A Type 1 system needs to be isolated and protected. A Type 2 system needs to be migrated. A Type 3 system needs to be left alone. Try explaining that to a CTO who just finished reading a Gartner report on “legacy modernisation.”</p>

<p>From a thread on TheLayoff:</p>

<blockquote><p>“FWIW, there&#39;s a very good chance that your electronic footprint on any given day has passed through a piece of SPARC equipment running Solaris, and that will continue to happen for a good portion of your lifetime.”</p></blockquote>

<p>Would you believe me if I told you I&#39;ve seen original BSD systems with eleven years of uptime?</p>

<h2 id="the-real-problem-isn-t-the-machines" id="the-real-problem-isn-t-the-machines">The real problem isn&#39;t the machines</h2>

<p>Here we get to the heart of the matter. And the answer will surprise you: the real problem with legacy systems isn&#39;t technological. It&#39;s human.</p>

<p>Let&#39;s talk about the “COBOL Cowboys” – retired programmers called back on consulting contracts when something breaks. They&#39;re the last generation that knows how those systems actually work. When they leave, they take decades of undocumented knowledge with them. According to Deloitte, companies have seen a 23% decline in mainframe workforce over the last five years, with 63% of those positions left unfilled. It&#39;s not that there&#39;s no money to hire – it&#39;s that there&#39;s nobody to hire. Young developers don&#39;t want to learn COBOL. It&#39;s “unsexy.” It&#39;s “archaic.” It&#39;s “boomer stuff.”</p>

<p>From ComputerWeekly:</p>

<blockquote><p>“The retirement of the generation of experts who possess in-depth knowledge of Cobol systems is leading to a severe knowledge shortage. They have knowledge not only of the Cobol programming language, but also of the specific systems they have worked on and built over the years” – Tijs van der Storm, CWI/University of Groningen</p></blockquote>

<p>And so we find ourselves in a paradoxical situation: systems processing trillions of dollars a day, managed by people who might die of old age before anyone learns to replace them. Knowledge transfer never happened. Documentation – where it exists – is outdated, incomplete, written in a language nobody understands anymore. And every year that passes, the gap widens.</p>

<p>This is the real legacy problem. Not the systems. The people.</p>

<h2 id="when-modernisation-fails-spoiler-often" id="when-modernisation-fails-spoiler-often">When modernisation fails (spoiler: often)</h2>

<p>There&#39;s a story that people in the UK know well, but that strangely never comes up when “digital transformation” is being discussed. It&#39;s called the National Programme for IT, or NPfIT.</p>

<p>Launched in 2002, it was the largest public sector IT project in British history. The goal? Modernise the entire NHS IT infrastructure. Initial budget: 6 billion pounds. Planned completion: 2010.</p>

<p>In 2011, after nine years of delays, exploding costs, vendors abandoning the project, and a system that simply didn&#39;t work, the UK government announced the dismantling of NPfIT. Final estimated cost: over 10 billion pounds. For a system that was never completed.</p>

<p>What went wrong? Practically everything. Top-down decisions made by politicians who didn&#39;t understand technology. Rigid contracts with vendors who didn&#39;t understand the NHS. Resistance from medical staff who hadn&#39;t been consulted. Continuously shifting requirements. Impossible integrations with existing systems.</p>

<p>From TechMonitor:</p>

<blockquote><p>“A lack of digital and procurement capability within government has led to wasted expenditure and lack of progress on major digital transformation programmes.”</p></blockquote>

<p>The lesson? “Modernising” is not automatically better than “maintaining.” Sometimes, the legacy system that works is preferable to the modern system that never will. But this lesson, apparently, we haven&#39;t learned. Because the dominant narrative remains the same: legacy = bad, modern = good. And consultants keep selling the shiny new thing.</p>

<h2 id="strategies-that-actually-work" id="strategies-that-actually-work">Strategies that actually work</h2>

<p>TL;DR: There is no single solution. There&#39;s a matrix of options ranging from virtualisation to isolation, from refactoring to API wrapping. The choice depends on the type of legacy, the budget, and the acceptable level of risk.</p>

<p><strong>The Gartner 7Rs (yes, they have a name for everything):</strong></p>
<ol><li><strong>Retire</strong> – Switch it off. Only works if nobody&#39;s actually using it.</li>
<li><strong>Retain</strong> – Keep it as is. Sometimes the best choice.</li>
<li><strong>Relocate</strong> – Move it to new infrastructure without changes.</li>
<li><strong>Rehost</strong> – “Lift and shift” to cloud. Changes the hardware, not the software.</li>
<li><strong>Replatform</strong> – Minimal changes to run on a modern platform.</li>
<li><strong>Refactor</strong> – Rewrite parts of the code while maintaining functionality.</li>
<li><strong>Rearchitect</strong> – Completely redesign. The riskiest and most expensive.</li></ol>

<p><strong>Virtualisation and emulation</strong>
For systems on proprietary architectures (SPARC, VAX, Alpha, PA-RISC), solutions like Stromasys Charon emulate the original hardware on x86-64 platforms. The operating system and software don&#39;t change – only the iron underneath does. For legacy x86 systems (Windows XP, Server 2003, old Linux), standard virtualisation (Proxmox, VMware, KVM) allows you to “freeze” the environment and keep it running indefinitely. I&#39;ve seen Proxmox setups running Windows 3.11. I&#39;m not joking.</p>

<p><strong>Network isolation</strong>
If a system can&#39;t be patched, it can at least be isolated. Dedicated VLANs, restrictive firewalls, air-gap where possible. It doesn&#39;t fix the problem, but it limits the impact in case of compromise.</p>

<p><strong>API wrapping</strong>
Put a modern REST layer in front of a legacy system. The mainframe keeps doing what it knows how to do; the outside world talks to the API. This is the strategy many banks use to expose COBOL functionality to mobile applications.</p>

<h2 id="the-public-sector-a-special-case" id="the-public-sector-a-special-case">The public sector: a special case</h2>

<p>Those who work in the public sector know that the dynamics differ from the private sector in ways that make the legacy problem even more complex.</p>

<p><strong>Multi-year budgets.</strong> You can&#39;t decide in January to modernise a system and have the money by March. Funding cycles are long, rigid, subject to political priorities that change with every election.</p>

<p><strong>Procurement.</strong> Buying software in the public sector is a bureaucratic nightmare. Tenders, compliance requirements, impact assessments, GDPR, accessibility. A purchase that takes a week in the private sector takes months here.</p>

<p><strong>Compliance.</strong> Systems handling health, education, or tax data are subject to stringent regulatory requirements. You can&#39;t simply “migrate to the cloud” – you have to demonstrate that the cloud complies with an endless list of standards.</p>

<p><strong>Service continuity</strong> (which in my view is the core problem). If a private company&#39;s system goes down for a day, they lose money. If a system managing national exams, or medical prescriptions, or pension payments goes down, the consequences fall on real people with no alternatives. The risk of downtime during a migration is often simply unacceptable.</p>

<p>And then there&#39;s the political dimension. Every government wants to announce its own “digital revolution.” Nobody wants to inherit the previous government&#39;s problems. And so projects get started, abandoned, restarted, re-abandoned, in an endless cycle of waste.</p>

<p>NPfIT wasn&#39;t an exception. It was the rule.</p>

<h2 id="the-uncomfortable-question" id="the-uncomfortable-question">The uncomfortable question</h2>

<p>At this point, the question nobody wants to ask is this: what if some legacy systems were simply… better? Not better in an absolute sense, but better for their specific purpose?</p>

<p>Let me tell you something. I worked for years in environments dealing with large-scale Oracle infrastructure – the company that sells “cloud transformation” and “modern infrastructure” to half the world. And among other things, you know what got managed day to day? Old ZFS storage. Stuff that, on paper, should have been “modernised” years ago. Those machines had been running since before Docker existed, before Kubernetes, before “cloud native” became a term. And they worked. Quietly. Without drama. Nobody was in any hurry to replace them. Why would they be? In pursuit of what advantage, exactly?</p>

<p>The COBOL processing bank transactions has been optimised for sixty years. Every bug has been found and fixed. Every edge case has been handled. Every possible scenario has been tested in production billions of times. It&#39;s code that has achieved a kind of perfection through Darwinian evolution. Rewriting it in Python would mean starting from scratch. New bugs. New untested scenarios. Years of instability before reaching the same level of reliability.</p>

<p>And in the meantime? In the meantime, the legacy system keeps working. There&#39;s a reason banks aren&#39;t in a rush to abandon mainframes. It&#39;s not ignorance. It&#39;s not laziness. It&#39;s that they&#39;ve done the maths and understood that the risk of the new outweighs the cost of the old. And the old administrators have retired. But this is an uncomfortable truth. It doesn&#39;t sell well in PowerPoint presentations. It doesn&#39;t generate consulting contracts. It doesn&#39;t make tech headlines.</p>

<p>And so we keep talking about “modernisation” as if it were automatically a good thing. As if “new” meant “better.” As if technology had a moral direction.</p>

<h2 id="so-what" id="so-what">So what?</h2>

<p>Legacy doesn&#39;t mean old – it means abandoned. The problem is never technical – it&#39;s always organisational. And “modernising” is not automatically better than “maintaining.”</p>

<p>If there&#39;s one lesson, it&#39;s this: be suspicious of anyone with simple answers to complex problems.</p>

<p>Every time I hear some manager say “we need to automate everything with AI,” I think about the software pachyderms holding up half of critical infrastructure. I think about the time it would take to train a model on COBOL written in 1987 with no documentation. I think about how long it would take to migrate a Java 1.7 system running on Solaris 9. I think about the hours spent reverse-engineering platforms still running Lotus Notes. I think about the costs. I think about the risks. And then I think that those same managers don&#39;t have the budget to hire juniors willing – and why should they be, when the IT world is moving in a completely different direction – to learn systems that have been decommissioned for at least thirty years. And I laugh. Bitterly, but I laugh. Then I take a few drops of CBD to calm myself down.</p>

<p>Before talking about artificial intelligence – and those who know me know I&#39;m not against AI at all – perhaps we should make sure that human intelligence doesn&#39;t retire, taking years of undocumented knowledge with it. But that, evidently, is a less sexy priority to put on the slides.</p>

<h2 id="sources-and-further-reading" id="sources-and-further-reading">Sources and further reading</h2>

<p><strong>UK government reports</strong>
– NAO: “The sustainability of government IT” (January 2025)
<a href="https://www.nao.org.uk/reports/local-government-financial-sustainability-2025/">https://www.nao.org.uk/reports/local-government-financial-sustainability-2025/</a>
– NHS Digital: Infrastructure assessment reports
<a href="https://www.bma.org.uk/advice-and-support/nhs-delivery-and-workforce/the-future/building-the-future-healthcare-infras">https://www.bma.org.uk/advice-and-support/nhs-delivery-and-workforce/the-future/building-the-future-healthcare-infras</a></p>

<p><strong>COBOL and mainframes</strong>
– Reuters: “Banks scramble to fix old systems” (Commonwealth Bank Australia cost analysis)
<a href="https://www.reuters.com/article/technology/banks-scramble-to-fix-old-systems-as-it-cowboys-ride-into-sunset-idUSKBN17C0CN/">https://www.reuters.com/article/technology/banks-scramble-to-fix-old-systems-as-it-cowboys-ride-into-sunset-idUSKBN17C0CN/</a>
– IBM: “COBOL Modernization”
<a href="https://www.ibm.com/think/topics/cobol-modernization">https://www.ibm.com/think/topics/cobol-modernization</a></p>

<p><strong>Legacy virtualisation</strong>
– Stromasys: “What are legacy systems”
<a href="https://www.stromasys.com/resources/what-are-legacy-systems-challenges-benefits/">https://www.stromasys.com/resources/what-are-legacy-systems-challenges-benefits/</a>
– Proxmox Forums: discussions on legacy system virtualisation
<a href="https://forum.proxmox.com/tags/legacy/">https://forum.proxmox.com/tags/legacy/</a></p>

<p><strong>Sector analysis</strong>
– Gartner: 7Rs of Application Modernization
<a href="https://www.techtarget.com/searchCloudComputing/tip/Use-the-7-Rs-to-develop-an-app-modernization-strategy">https://www.techtarget.com/searchCloudComputing/tip/Use-the-7-Rs-to-develop-an-app-modernization-strategy</a>
– Deloitte: Mainframe workforce decline study
<a href="https://www.deloitte.com/us/en/insights/topics/technology-management/tech-trends/2023/future-mainframe-technology-latest-trends.html">https://www.deloitte.com/us/en/insights/topics/technology-management/tech-trends/2023/future-mainframe-technology-latest-trends.html</a>
– WSJ: How AI Can Rev Up Mainframe Modernization
<a href="https://deloitte.wsj.com/cio/how-ai-can-rev-up-mainframe-modernization-2e3c1c4a">https://deloitte.wsj.com/cio/how-ai-can-rev-up-mainframe-modernization-2e3c1c4a</a></p>

<p><strong>Case studies: failures</strong>
– Computer Weekly: “What went wrong with the National Programme for IT”
<a href="https://www.computerweekly.com/opinion/Six-reasons-why-the-NHS-National-Programme-for-IT-failed">https://www.computerweekly.com/opinion/Six-reasons-why-the-NHS-National-Programme-for-IT-failed</a>
– NAO: Post-implementation review NPfIT
<a href="https://www.nao.org.uk/reports/review-of-the-final-benefits-statement-for-programmes-previously-managed-under-the-national-programme-for-it-in-the-nhs/">https://www.nao.org.uk/reports/review-of-the-final-benefits-statement-for-programmes-previously-managed-under-the-national-programme-for-it-in-the-nhs/</a></p>

<p><strong>Security</strong>
– WannaCry incident reports
<a href="https://any.run/malware-trends/wannacry/">https://any.run/malware-trends/wannacry/</a>
– NHS Windows XP audit findings (2019)
<a href="https://www.verdict.co.uk/windows-xp-nhs/">https://www.verdict.co.uk/windows-xp-nhs/</a></p>

<p><a href="https://remark.as/p/jolek78/legacy-systems-problem-or-resource">Discuss...</a></p>

<p><a href="https://jolek78.writeas.com/tag:LegacySystems" class="hashtag"><span>#</span><span class="p-category">LegacySystems</span></a> <a href="https://jolek78.writeas.com/tag:Sysadmin" class="hashtag"><span>#</span><span class="p-category">Sysadmin</span></a> <a href="https://jolek78.writeas.com/tag:COBOL" class="hashtag"><span>#</span><span class="p-category">COBOL</span></a> <a href="https://jolek78.writeas.com/tag:Solaris" class="hashtag"><span>#</span><span class="p-category">Solaris</span></a> <a href="https://jolek78.writeas.com/tag:Linux" class="hashtag"><span>#</span><span class="p-category">Linux</span></a> <a href="https://jolek78.writeas.com/tag:PublicSector" class="hashtag"><span>#</span><span class="p-category">PublicSector</span></a> <a href="https://jolek78.writeas.com/tag:DigitalTransformation" class="hashtag"><span>#</span><span class="p-category">DigitalTransformation</span></a> <a href="https://jolek78.writeas.com/tag:Mainframe" class="hashtag"><span>#</span><span class="p-category">Mainframe</span></a> <a href="https://jolek78.writeas.com/tag:OpenSource" class="hashtag"><span>#</span><span class="p-category">OpenSource</span></a> <a href="https://jolek78.writeas.com/tag:Infrastructure" class="hashtag"><span>#</span><span class="p-category">Infrastructure</span></a></p>

<div class="center">
· 🦣 <a href="https://fosstodon.org/@jolek78">Mastodon</a> · 📸 <a href="https://pixelfed.social/jolek78">Pixelfed</a> ·  📬 <a href="mailto:jolek78@jolek78.dev">Email</a> ·
· ☕ <a href="https://liberapay.com/jolek78">Support this work on Liberapay</a>
</div>
]]></content:encoded>
      <guid>https://jolek78.writeas.com/legacy-systems-problem-or-resource</guid>
      <pubDate>Wed, 28 Jan 2026 16:30:00 +0000</pubDate>
    </item>
    <item>
      <title>Kiwix: Wikipedia in your pocket</title>
      <link>https://jolek78.writeas.com/kiwix-wikipedia-in-your-pocket?pk_campaign=rss-feed</link>
      <description>&lt;![CDATA[A hackmeeting, many years ago. A conference on various open-source projects. They were talking about Kiwix. The audience seemed interested, nodding, asking questions. I sat in the back of the room with a doubt that seemed legitimate but that I didn&#39;t dare express out loud: &#34;what&#39;s the point of offline Wikipedia?&#34; I mean: the internet is everywhere. If you need to look something up on Wikipedia, you open your browser, search, read. Done. Why would anyone download gigabytes of data to consult an encyclopedia offline? It seemed like a solution in search of a problem. Something for nerds nostalgic for CD-ROM encyclopedias.&#xA;&#xA;It took me years to understand how naive I&#39;d been.&#xA;&#xA;!--more--&#xA;&#xA;Years in which I continued to follow the project from afar. Years in which I read stories of deployments in Africa, Asia, prisons, refugee camps. Years in which I understood that the internet isn&#39;t everywhere, it&#39;s a privilege, not a given. And even where it exists, it&#39;s not necessarily accessible, affordable, or free from censorship.&#xA;&#xA;Years later, when I set up my Proxmox server, one of the first containers I decided to install was Kiwix. Not because I needed it—my connection works fine, thanks for asking—but because I wanted to be part of that project, so to speak. Because I had understood that Kiwix wasn&#39;t just software. It&#39;s a philosophy. It&#39;s practical proof that another web is possible: decentralized, offline, in users&#39; hands. &#xA;&#xA;Simply a matter of fundamental rights&#xA;There&#39;s a moment in 2004 when Emmanuel Engelhart—a French computer engineer working between Germany and Switzerland—becomes a Wikipedia editor and asks himself an apparently simple question: &#34;What about those without internet access?&#34; It wasn&#39;t a rhetorical question. At the time, as today, billions of people lived (and live) in areas where connectivity is a luxury, where broadband is science fiction, where even a single megabyte of data costs more than a meal.&#xA;&#xA;Engelhart&#39;s answer was radical: if people can&#39;t reach Wikipedia, then Wikipedia must reach people. Even without the internet.&#xA;&#xA;You know that thing about &#34;if the mountain won&#39;t come to Muhammad...&#34;? Exactly that.&#xA;&#xA;And so, in 2007, together with Renaud Gaudin—a Malian information management expert—Engelhart launched Kiwix. Open source software that allowed downloading the entire Wikipedia (and much more) to consult it completely offline.&#xA;&#xA;In a 2014 interview, Engelhart stated:&#xA;&#xA;  The contents of Wikipedia should be available for everyone! Even without Internet access. This is why I have launched the Kiwix project. Our users are all over the world: sailors on the oceans, poor students thirsty for knowledge, globetrotters almost living in planes, world&#39;s citizens suffering from censorship or free minded prisoners. For all these people, Kiwix provides a simple and practical solution to ponder about the world.&#xA;&#xA;And:&#xA;&#xA;  Water is a common good. You understand why you have to care about water. Wikipedia is the same; it&#39;s a common good. We have to care about Wikipedia.&#xA;&#xA;Digital Sovereignty&#xA;Why talk about Kiwix today? Because it&#39;s not just a technical solution to a connectivity problem. Kiwix represents something deeper: digital sovereignty in its purest form.&#xA;&#xA;While projects like Mastodon, Matrix, Lemmy, and Pixelfed create distributed networks—many nodes communicating with each other in federation—Kiwix goes beyond, or perhaps beneath, depending on your perspective. It&#39;s so radically independent that it doesn&#39;t even need a network. It&#39;s local. Completely. A single Kiwix installation is an autonomous island that communicates with nothing and no one.&#xA;&#xA;No federation, no peer-to-peer, no cloud.&#xA;&#xA;You have Wikipedia on your Raspberry Pi? It&#39;s yours—or rather, it&#39;s yours thanks to the contribution of all Wikipedians. It works without internet, without external dependencies. You can copy it to a USB stick and give it to someone else. You can take it to the middle of the ocean, the desert, Antarctica. You can share it on a local computer network. And it will work. Always. The data is on your hardware, under your physical control.&#xA;&#xA;The birth of the project&#xA;Kiwix&#39;s 2007 launch didn&#39;t happen with grand announcements or marketing campaigns. It was open source software, released under GPL license, developed by two enthusiasts. That&#39;s it.&#xA;&#xA;The technological heart of the project was (and is) the ZIM format—&#34;Zeno IMproved&#34;—an open source archive format optimized for wiki-style content. Highly compressed, easily indexable, designed to be searchable even without connection. All of Wikipedia&#39;s content is converted to static HTML, compressed into ZIM, and made available for download.&#xA;&#xA;To give you an idea of scale: the entire English Wikipedia—6.4 million articles, images included—takes up about 97 GB in ZIM format. Seems like a lot? The sum of all human knowledge now fits on a microSD card that costs 15 euros. On a 1TB portable hard drive you can put Wikipedia in ten different languages, the entire Project Gutenberg library, all TED talks, complete Stack Exchange, and you&#39;ll still have space left over.&#xA;&#xA;Between 2007 and 2011, the team also released three CD/DVD versions with article selections. Today they seem like archaeological artifacts, but at the time they were the solution for bringing Wikipedia to African schools where the internet simply didn&#39;t exist.&#xA;&#xA;The XULRunner problem and the rebirth&#xA;Like every serious open source project, Kiwix had its &#34;winter.&#34; Between 2014 and 2020, the software disappeared from many Linux distribution repositories. The reason? XULRunner, the Mozilla framework Kiwix was based on, was deprecated and removed from package databases.&#xA;&#xA;For six years, Kiwix was technically &#34;dead&#34; for many Linux users. But the community didn&#39;t give up. The team worked to completely rethink the software&#39;s architecture, rewrite it from scratch, and modernize it. When it reemerged in 2020, it was stronger than before: progressive WebApp, browser extensions, native mobile support, Raspberry Pi integration.&#xA;&#xA;It&#39;s the usual open source story: an obstacle that would seem fatal becomes an opportunity to improve and grow. How many proprietary companies would have simply shut down? But in open source, software doesn&#39;t die as long as the code is available and someone believes in it.&#xA;&#xA;Where Kiwix saves lives (not hyperbole)&#xA;Numbers are important, but it&#39;s the stories that make us truly understand a project&#39;s impact.&#xA;&#xA;Kenya: the Thika Alumni Trust&#xA;In 2015, seven friends who had studied together in the &#39;60s at a high school in Thika return for a visit. The principal asks for help: they need 50 computers to create a lab. The problem? The internet connection is 100 kbps. Useless.&#xA;&#xA;The solution was to create completely offline digital learning environments using Kiwix. Today, that project has transformed education in 61 schools throughout Kenya, reaching over 70,000 children. They&#39;ve installed 164 microservers running Kiwix—probably one of the largest networks in the world.&#xA;&#xA;The results? In primary schools where the Trust operates, national exam results improved from 8 to 12%. In special needs units, where absenteeism reached 50%, attendance now exceeds 90%.&#xA;&#xA;Mary Mungai, principal of a school with special needs units, says: &#34;All our children have benefited tremendously from the digital libraries. We have children who refused to attend classes but now do so faithfully, some who couldn&#39;t read or write but now do very well on computers.&#34;&#xA;&#xA;Ghana: the Kiwix4Schools Project&#xA;In 2019, four Ghanaian students from Ashesi University launched Kiwix4Schools with a simple goal: bring digital education to rural schools. They installed Kiwix on 15 Raspberry Pi devices, reaching 2,000 students in 15 schools.&#xA;&#xA;The impact was immediate. Teachers reported students staying after school to explore content. Children who had never touched a computer were navigating Wikipedia articles. Science class changed completely when students could look up experiments, see diagrams, understand concepts beyond what the single available textbook offered.&#xA;&#xA;India: Internet blackouts and censorship&#xA;In 2019-2020, the Indian government imposed internet blackouts in Kashmir—the longest in a democracy&#39;s history. For months, millions of people were cut off from the digital world. Hospitals, schools, businesses paralyzed.&#xA;&#xA;But those who had Kiwix continued accessing medical information, educational content, technical documentation. It wasn&#39;t a complete solution, but it was a lifeline. It demonstrated that offline access isn&#39;t just for poor countries—it&#39;s a resilience tool even in developed nations with unstable political situations.&#xA;&#xA;The ZIM format: open everything&#xA;The genius of Kiwix lies in the ZIM format. It&#39;s not just a compression format—it&#39;s an open standard specifically designed for offline content distribution. Any developer can create ZIM files, any software can read them. There&#39;s no vendor lock-in, no proprietary license.&#xA;&#xA;But ZIM isn&#39;t just for Wikipedia. Today ZIM archives exist for:&#xA;&#xA;Project Gutenberg (50,000+ public domain books)&#xA;Stack Exchange (all sites, all Q&amp;As)&#xA;TED Talks (thousands of videos with subtitles)&#xA;Khan Academy&#xA;Ubuntu documentation&#xA;Arch Wiki&#xA;WikiMed (medical encyclopedia, used by 100,000 doctors and students)&#xA;&#xA;The format is completely open, documented, and anyone can create ZIM archives of their content. It&#39;s the open source spirit in its purest form.&#xA;&#xA;Everything works&#xA;In 2018, Kiwix formalized collaboration with the Wikimedia Foundation, receiving $275,000 to improve offline access. In 2023, came a $250,000 grant from the Wikimedia Endowment.&#xA;&#xA;Stephane Coillet-Matillon, Kiwix CEO, in December 2018 declared:&#xA;&#xA;  Our hope is that one day everyone will have access to the internet, and eliminate the need for other offline methods of access to information. But we know that there are still serious gaps in internet access globally that require solutions today. Kiwix is a tool to start fixing things right now.&#xA;&#xA;Today, in 2025:&#xA;&#xA;Over 10 million users in more than 220 countries&#xA;More than 10,000 websites crawled regularly&#xA;Available on all platforms: Android, iOS, Windows, macOS, Linux&#xA;Browser extensions for Firefox, Chrome, Edge&#xA;Partnership with Orange Foundation to reach 500,000 children in West Africa&#xA;&#xA;You can explore the entire catalog at library.kiwix.org.&#xA;&#xA;The philosophy behind the code&#xA;Here we arrive at the heart of the matter. Why is Kiwix important? Not just because it works, not just because it&#39;s helped millions of people. But because it represents a way of thinking about technology.&#xA;&#xA;Kiwix is:&#xA;&#xA;Open Source: all code on GitHub, GPL license. Anyone can study it, modify it, improve it.&#xA;Completely local: doesn&#39;t depend on central servers, cloud, or connections. Each installation is autonomous.&#xA;Privacy-first: no tracking, no telemetry, no data sent to third parties. Impossible—it&#39;s offline.&#xA;Community-driven: developed by volunteers, funded by donations.&#xA;Accessible: designed to work even on old or limited hardware.&#xA;&#xA;It&#39;s the antithesis of the Big Tech model. There&#39;s no company controlling access, no centralized database of who reads what, no algorithms deciding which information to show you. It&#39;s technology as it should be: serving the user, before corporations transformed it into a machine for extracting data and selling advertising.&#xA;&#xA;A &#34;dangerous&#34; precedent&#xA;There&#39;s an interesting paradox. Kiwix exists because the internet isn&#39;t accessible to everyone. But its success demonstrates that maybe we don&#39;t even need it to be—at least not the way we conceive it now.&#xA;&#xA;Think about it: if I can have Wikipedia, Stack Exchange, Project Gutenberg, Khan Academy on a 128GB SD card, why should I depend on an always-on internet connection? If I can sync updates once a month when I pass by the library with WiFi, why should I pay 50 euros a month for a home connection?&#xA;&#xA;Kiwix demonstrates that the &#34;always connected, always online, always tracked&#34; model isn&#39;t the only possible one. That an alternative exists where knowledge is local, accessible, controllable. The monopoly isn&#39;t inevitable.&#xA;&#xA;And this, for Big Tech, is dangerous. Because if people realize they can access information without going through Google, without being tracked, without seeing ads... well, the entire business model collapses. It&#39;s also no secret that the entire streaming model—everything, no one excluded: Spotify, YouTube, Netflix, etc.—is ecologically unsustainable. Downloading once and playing a thousand times (locally) is less wasteful than downloading zero times and playing a thousand times (remotely). If it can be done for Wikipedia, TED Talks, and Project Gutenberg, it can be done for everything else.&#xA;&#xA;But the biggest challenge remains the same: making Kiwix known. Because the software exists, works, is free. But how many people know they can have Wikipedia in their pocket without the internet? How many African schools know they can have a complete digital library for the cost of a Raspberry Pi?&#xA;&#xA;Conclusions: what I learned&#xA;Innovation often doesn&#39;t come from Silicon Valley. It comes from a young French engineer working in Germany asking a simple question. It comes from developers scattered around the world contributing in their free time. It comes from the community, not corporations.&#xA;&#xA;Open source works. Kiwix is almost twenty years old, has overcome technical crises that would have killed a proprietary project, has continued to grow with ridiculous budgets. Why? Because the community believes in it. Because the code is open. Because the mission is clear.&#xA;&#xA;Technology is political. Deciding that knowledge must be accessible offline is a political choice. Deciding to use open source licenses is a political choice. Deciding not to track users is a political choice.&#xA;&#xA;Kiwix shows us an alternative. That we don&#39;t have to choose between functionality and ethics. That another web is possible.&#xA;&#xA;And now, if you&#39;ll excuse me, I&#39;m going to add a Python ZIM library to my Kiwix container, because I&#39;m studying it—or rather, &#34;I have to study it&#34;—for a bunch of small projects I have in mind. AI server included.&#xA;&#xA;#Kiwix #SmallWeb #DigitalSovereignty #OpenSource #Wikipedia #Offline #Privacy #Education #Africa&#xA;&#xA;a href=&#34;https://remark.as/p/jolek78/kiwix-wikipedia-in-your-pocket&#34;Discuss.../a&#xA;&#xA;div class=&#34;center&#34;&#xD;&#xA;· 🦣 a href=&#34;https://fosstodon.org/@jolek78&#34;Mastodon/a · 📸 a href=&#34;https://pixelfed.social/jolek78&#34;Pixelfed/a ·  📬 a href=&#34;mailto:jolek78@jolek78.dev&#34;Email/a ·&#xD;&#xA;· ☕ a href=&#34;https://liberapay.com/jolek78&#34;Support this work on Liberapay/a&#xD;&#xA;/div]]&gt;</description>
      <content:encoded><![CDATA[<p>A hackmeeting, many years ago. A conference on various open-source projects. They were talking about <a href="https://kiwix.org">Kiwix</a>. The audience seemed interested, nodding, asking questions. I sat in the back of the room with a doubt that seemed legitimate but that I didn&#39;t dare express out loud: “what&#39;s the point of offline Wikipedia?” I mean: the internet is everywhere. If you need to look something up on Wikipedia, you open your browser, search, read. Done. Why would anyone download gigabytes of data to consult an encyclopedia offline? It seemed like a solution in search of a problem. Something for nerds nostalgic for CD-ROM encyclopedias.</p>

<p>It took me years to understand how naive I&#39;d been.</p>



<p>Years in which I continued to follow the project from afar. Years in which I read stories of deployments in Africa, Asia, prisons, refugee camps. Years in which I understood that the internet isn&#39;t everywhere, it&#39;s a privilege, not a given. And even where it exists, it&#39;s not necessarily accessible, affordable, or free from censorship.</p>

<p>Years later, when I set up my Proxmox server, one of the first containers I decided to install was Kiwix. Not because I needed it—my connection works fine, thanks for asking—but because I wanted to be part of that project, so to speak. Because I had understood that Kiwix wasn&#39;t just software. It&#39;s a philosophy. It&#39;s practical proof that another web is possible: decentralized, offline, in users&#39; hands.</p>

<h3 id="simply-a-matter-of-fundamental-rights" id="simply-a-matter-of-fundamental-rights">Simply a matter of fundamental rights</h3>

<p>There&#39;s a moment in 2004 when Emmanuel Engelhart—a French computer engineer working between Germany and Switzerland—becomes a Wikipedia editor and asks himself an apparently simple question: “What about those without internet access?” It wasn&#39;t a rhetorical question. At the time, as today, billions of people lived (and live) in areas where connectivity is a luxury, where broadband is science fiction, where even a single megabyte of data costs more than a meal.</p>

<p>Engelhart&#39;s answer was radical: if people can&#39;t reach Wikipedia, then Wikipedia must reach people. Even without the internet.</p>

<p>You know that thing about “if the mountain won&#39;t come to Muhammad...”? Exactly that.</p>

<p>And so, in 2007, together with Renaud Gaudin—a Malian information management expert—Engelhart launched Kiwix. Open source software that allowed downloading the entire Wikipedia (and much more) to consult it completely offline.</p>

<p>In a <a href="https://diff.wikimedia.org/2014/09/12/emmanuel-engelhart-inventor-of-kiwix/">2014 interview</a>, Engelhart stated:</p>

<blockquote><p>The contents of Wikipedia should be available for everyone! Even without Internet access. This is why I have launched the Kiwix project. Our users are all over the world: sailors on the oceans, poor students thirsty for knowledge, globetrotters almost living in planes, world&#39;s citizens suffering from censorship or free minded prisoners. For all these people, Kiwix provides a simple and practical solution to ponder about the world.</p></blockquote>

<p>And:</p>

<blockquote><p>Water is a common good. You understand why you have to care about water. Wikipedia is the same; it&#39;s a common good. We have to care about Wikipedia.</p></blockquote>

<h3 id="digital-sovereignty" id="digital-sovereignty">Digital Sovereignty</h3>

<p>Why talk about Kiwix today? Because it&#39;s not just a technical solution to a connectivity problem. Kiwix represents something deeper: digital sovereignty in its purest form.</p>

<p>While projects like Mastodon, Matrix, Lemmy, and Pixelfed create distributed networks—many nodes communicating with each other in federation—Kiwix goes beyond, or perhaps beneath, depending on your perspective. It&#39;s so radically independent that it doesn&#39;t even need a network. It&#39;s local. Completely. A single Kiwix installation is an autonomous island that communicates with nothing and no one.</p>

<p>No federation, no peer-to-peer, no cloud.</p>

<p>You have Wikipedia on your Raspberry Pi? It&#39;s yours—or rather, it&#39;s yours <em>thanks to the contribution</em> of all Wikipedians. It works without internet, without external dependencies. You can copy it to a USB stick and give it to someone else. You can take it to the middle of the ocean, the desert, Antarctica. You can share it on a local computer network. And it will work. Always. The data is on your hardware, under your physical control.</p>

<h3 id="the-birth-of-the-project" id="the-birth-of-the-project">The birth of the project</h3>

<p>Kiwix&#39;s 2007 launch didn&#39;t happen with grand announcements or marketing campaigns. It was open source software, released under GPL license, developed by two enthusiasts. That&#39;s it.</p>

<p>The technological heart of the project was (and is) the ZIM format—”Zeno IMproved”—an open source archive format optimized for wiki-style content. Highly compressed, easily indexable, designed to be searchable even without connection. All of Wikipedia&#39;s content is converted to static HTML, compressed into ZIM, and made available for download.</p>

<p>To give you an idea of scale: the entire English Wikipedia—6.4 million articles, images included—takes up about 97 GB in ZIM format. Seems like a lot? The sum of all human knowledge now fits on a microSD card that costs 15 euros. On a 1TB portable hard drive you can put Wikipedia in ten different languages, the entire Project Gutenberg library, all TED talks, complete Stack Exchange, and you&#39;ll still have space left over.</p>

<p>Between 2007 and 2011, the team also released three CD/DVD versions with article selections. Today they seem like archaeological artifacts, but at the time they were the solution for bringing Wikipedia to African schools where the internet simply didn&#39;t exist.</p>

<h3 id="the-xulrunner-problem-and-the-rebirth" id="the-xulrunner-problem-and-the-rebirth">The XULRunner problem and the rebirth</h3>

<p>Like every serious open source project, Kiwix had its “winter.” Between 2014 and 2020, the software disappeared from many Linux distribution repositories. The reason? XULRunner, the Mozilla framework Kiwix was based on, was deprecated and removed from package databases.</p>

<p>For six years, Kiwix was technically “dead” for many Linux users. But the community didn&#39;t give up. The team worked to completely rethink the software&#39;s architecture, rewrite it from scratch, and modernize it. When it reemerged in 2020, it was stronger than before: progressive WebApp, browser extensions, native mobile support, Raspberry Pi integration.</p>

<p>It&#39;s the usual open source story: an obstacle that would seem fatal becomes an opportunity to improve and grow. How many proprietary companies would have simply shut down? But in open source, software doesn&#39;t die as long as the code is available and someone believes in it.</p>

<h3 id="where-kiwix-saves-lives-not-hyperbole" id="where-kiwix-saves-lives-not-hyperbole">Where Kiwix saves lives (not hyperbole)</h3>

<p>Numbers are important, but it&#39;s the stories that make us truly understand a project&#39;s impact.</p>

<h4 id="kenya-the-thika-alumni-trust" id="kenya-the-thika-alumni-trust">Kenya: the Thika Alumni Trust</h4>

<p>In 2015, seven friends who had studied together in the &#39;60s at a high school in Thika return for a visit. The principal asks for help: they need 50 computers to create a lab. The problem? The internet connection is 100 kbps. Useless.</p>

<p>The solution was to create completely offline digital learning environments using Kiwix. Today, that project has transformed education in 61 schools throughout Kenya, reaching over 70,000 children. They&#39;ve installed 164 microservers running Kiwix—probably one of the largest networks in the world.</p>

<p>The results? In primary schools where the Trust operates, national exam results improved from 8 to 12%. In special needs units, where absenteeism reached 50%, attendance now exceeds 90%.</p>

<p>Mary Mungai, principal of a school with special needs units, says: “All our children have benefited tremendously from the digital libraries. We have children who refused to attend classes but now do so faithfully, some who couldn&#39;t read or write but now do very well on computers.”</p>

<h4 id="ghana-the-kiwix4schools-project" id="ghana-the-kiwix4schools-project">Ghana: the Kiwix4Schools Project</h4>

<p>In 2019, four Ghanaian students from Ashesi University launched Kiwix4Schools with a simple goal: bring digital education to rural schools. They installed Kiwix on 15 Raspberry Pi devices, reaching 2,000 students in 15 schools.</p>

<p>The impact was immediate. Teachers reported students staying after school to explore content. Children who had never touched a computer were navigating Wikipedia articles. Science class changed completely when students could look up experiments, see diagrams, understand concepts beyond what the single available textbook offered.</p>

<h4 id="india-internet-blackouts-and-censorship" id="india-internet-blackouts-and-censorship">India: Internet blackouts and censorship</h4>

<p>In 2019-2020, the Indian government imposed internet blackouts in Kashmir—the longest in a democracy&#39;s history. For months, millions of people were cut off from the digital world. Hospitals, schools, businesses paralyzed.</p>

<p>But those who had Kiwix continued accessing medical information, educational content, technical documentation. It wasn&#39;t a complete solution, but it was a lifeline. It demonstrated that offline access isn&#39;t just for poor countries—it&#39;s a resilience tool even in developed nations with unstable political situations.</p>

<h3 id="the-zim-format-open-everything" id="the-zim-format-open-everything">The ZIM format: open everything</h3>

<p>The genius of Kiwix lies in the <a href="https://wiki.openzim.org">ZIM format</a>. It&#39;s not just a compression format—it&#39;s an open standard specifically designed for offline content distribution. Any developer can create ZIM files, any software can read them. There&#39;s no vendor lock-in, no proprietary license.</p>

<p>But ZIM isn&#39;t just for Wikipedia. Today ZIM archives exist for:</p>
<ul><li>Project Gutenberg (50,000+ public domain books)</li>
<li>Stack Exchange (all sites, all Q&amp;As)</li>
<li>TED Talks (thousands of videos with subtitles)</li>
<li>Khan Academy</li>
<li>Ubuntu documentation</li>
<li>Arch Wiki</li>
<li>WikiMed (medical encyclopedia, used by 100,000 doctors and students)</li></ul>

<p>The format is completely open, documented, and anyone can create ZIM archives of their content. It&#39;s the open source spirit in its purest form.</p>

<h3 id="everything-works" id="everything-works">Everything works</h3>

<p>In 2018, Kiwix formalized collaboration with the Wikimedia Foundation, receiving $275,000 to improve offline access. In 2023, came a $250,000 grant from the Wikimedia Endowment.</p>

<p>Stephane Coillet-Matillon, Kiwix CEO, in <a href="https://wikimediafoundation.org/news/2018/12/21/kiwix-is-connecting-the-unconnected/">December 2018</a> declared:</p>

<blockquote><p>Our hope is that one day everyone will have access to the internet, and eliminate the need for other offline methods of access to information. But we know that there are still serious gaps in internet access globally that require solutions today. Kiwix is a tool to start fixing things right now.</p></blockquote>

<p>Today, in 2025:</p>
<ul><li>Over 10 million users in more than 220 countries</li>
<li>More than 10,000 websites crawled regularly</li>
<li>Available on all platforms: Android, iOS, Windows, macOS, Linux</li>
<li>Browser extensions for Firefox, Chrome, Edge</li>
<li>Partnership with Orange Foundation to reach 500,000 children in West Africa</li></ul>

<p>You can explore the entire catalog at <a href="https://library.kiwix.org/">library.kiwix.org</a>.</p>

<h3 id="the-philosophy-behind-the-code" id="the-philosophy-behind-the-code">The philosophy behind the code</h3>

<p>Here we arrive at the heart of the matter. Why is Kiwix important? Not just because it works, not just because it&#39;s helped millions of people. But because it represents a way of thinking about technology.</p>

<p>Kiwix is:</p>
<ul><li><strong>Open Source</strong>: all code on GitHub, GPL license. Anyone can study it, modify it, improve it.</li>
<li><strong>Completely local</strong>: doesn&#39;t depend on central servers, cloud, or connections. Each installation is autonomous.</li>
<li><strong>Privacy-first</strong>: no tracking, no telemetry, no data sent to third parties. Impossible—it&#39;s offline.</li>
<li><strong>Community-driven</strong>: developed by volunteers, funded by donations.</li>
<li><strong>Accessible</strong>: designed to work even on old or limited hardware.</li></ul>

<p>It&#39;s the antithesis of the Big Tech model. There&#39;s no company controlling access, no centralized database of who reads what, no algorithms deciding which information to show you. It&#39;s technology as it should be: serving the user, before corporations transformed it into a machine for extracting data and selling advertising.</p>

<h3 id="a-dangerous-precedent" id="a-dangerous-precedent">A “dangerous” precedent</h3>

<p>There&#39;s an interesting paradox. Kiwix exists because the internet isn&#39;t accessible to everyone. But its success demonstrates that maybe we don&#39;t even need it to be—at least not the way we conceive it now.</p>

<p>Think about it: if I can have Wikipedia, Stack Exchange, Project Gutenberg, Khan Academy on a 128GB SD card, why should I depend on an always-on internet connection? If I can sync updates once a month when I pass by the library with WiFi, why should I pay 50 euros a month for a home connection?</p>

<p>Kiwix demonstrates that the “always connected, always online, always tracked” model isn&#39;t the only possible one. That an alternative exists where knowledge is local, accessible, controllable. The monopoly isn&#39;t inevitable.</p>

<p>And this, for Big Tech, is dangerous. Because if people realize they can access information without going through Google, without being tracked, without seeing ads... well, the entire business model collapses. It&#39;s also no secret that the entire streaming model—everything, no one excluded: Spotify, YouTube, Netflix, etc.—is ecologically unsustainable. Downloading once and playing a thousand times (locally) is less wasteful than downloading zero times and playing a thousand times (remotely). If it can be done for Wikipedia, TED Talks, and Project Gutenberg, it can be done for everything else.</p>

<p>But the biggest challenge remains the same: making Kiwix known. Because the software exists, works, is free. But how many people know they can have Wikipedia in their pocket without the internet? How many African schools know they can have a complete digital library for the cost of a Raspberry Pi?</p>

<h3 id="conclusions-what-i-learned" id="conclusions-what-i-learned">Conclusions: what I learned</h3>

<p>Innovation often doesn&#39;t come from Silicon Valley. It comes from a young French engineer working in Germany asking a simple question. It comes from developers scattered around the world contributing in their free time. It comes from the community, not corporations.</p>

<p>Open source works. Kiwix is almost twenty years old, has overcome technical crises that would have killed a proprietary project, has continued to grow with ridiculous budgets. Why? Because the community believes in it. Because the code is open. Because the mission is clear.</p>

<p>Technology is political. Deciding that knowledge must be accessible offline is a political choice. Deciding to use open source licenses is a political choice. Deciding not to track users is a political choice.</p>

<p>Kiwix shows us an alternative. That we don&#39;t have to choose between functionality and ethics. That another web is possible.</p>

<p>And now, if you&#39;ll excuse me, I&#39;m going to add a Python ZIM library to my Kiwix container, because I&#39;m studying it—or rather, “I have to study it”—for a bunch of small projects I have in mind. AI server included.</p>

<p><a href="https://jolek78.writeas.com/tag:Kiwix" class="hashtag"><span>#</span><span class="p-category">Kiwix</span></a> <a href="https://jolek78.writeas.com/tag:SmallWeb" class="hashtag"><span>#</span><span class="p-category">SmallWeb</span></a> <a href="https://jolek78.writeas.com/tag:DigitalSovereignty" class="hashtag"><span>#</span><span class="p-category">DigitalSovereignty</span></a> <a href="https://jolek78.writeas.com/tag:OpenSource" class="hashtag"><span>#</span><span class="p-category">OpenSource</span></a> <a href="https://jolek78.writeas.com/tag:Wikipedia" class="hashtag"><span>#</span><span class="p-category">Wikipedia</span></a> <a href="https://jolek78.writeas.com/tag:Offline" class="hashtag"><span>#</span><span class="p-category">Offline</span></a> <a href="https://jolek78.writeas.com/tag:Privacy" class="hashtag"><span>#</span><span class="p-category">Privacy</span></a> <a href="https://jolek78.writeas.com/tag:Education" class="hashtag"><span>#</span><span class="p-category">Education</span></a> <a href="https://jolek78.writeas.com/tag:Africa" class="hashtag"><span>#</span><span class="p-category">Africa</span></a></p>

<p><a href="https://remark.as/p/jolek78/kiwix-wikipedia-in-your-pocket">Discuss...</a></p>

<div class="center">
· 🦣 <a href="https://fosstodon.org/@jolek78">Mastodon</a> · 📸 <a href="https://pixelfed.social/jolek78">Pixelfed</a> ·  📬 <a href="mailto:jolek78@jolek78.dev">Email</a> ·
· ☕ <a href="https://liberapay.com/jolek78">Support this work on Liberapay</a>
</div>
]]></content:encoded>
      <guid>https://jolek78.writeas.com/kiwix-wikipedia-in-your-pocket</guid>
      <pubDate>Thu, 18 Dec 2025 14:46:00 +0000</pubDate>
    </item>
    <item>
      <title>A song, an algorithm, and the end of the analog world</title>
      <link>https://jolek78.writeas.com/a-song-an-algorithm-and-the-end-of-the-analog-world?pk_campaign=rss-feed</link>
      <description>&lt;![CDATA[There&#39;s a moment in the history of technology when everything changes. We don&#39;t always recognise it. Sometimes it takes years to understand that a small spark, an apparently insignificant detail, ignited a revolution that would forever change the way we live, communicate, and consume culture. In 1987, an American singer-songwriter named Suzanne Vega released a minimalist track called &#34;Tom&#39;s Diner&#34;. Two minutes and nine seconds of a cappella vocals, no instrumental accompaniment, no special effects. Just a voice telling the story of an ordinary morning in a New York diner. A song so essential, so pure in its simplicity, that someone on the other side of the world – a German engineer obsessed with #audio compression – would use it as a benchmark to create a technology that would shake the global music industry to its core. That technology was called #MP3. And that voice, that &#34;warm a cappella voice&#34; as Karlheinz Brandenburg would later describe it, would become the ultimate test to determine whether a compression algorithm actually worked or not. &#xA;!--more--&#xA;&#xA;This is the story – part documented reality, part urban legend – of how a folk song became the unwitting mother of the greatest revolution in music distribution since vinyl. A story that has always fascinated me because it contains all the contradictions of our digital age: innovation and destruction, democratization and loss of quality, openness and control. And yes, it&#39;s also because I&#39;ve always had a soft spot for stories that intertwine in unexpected ways. Perhaps because I too, during my years in radio, saw first-hand what it means to work with audio, manipulate it, compress it, broadcast it. Perhaps because, like many of us who lived through the transition from analog to digital, I still carry the memory of those first MP3 collections downloaded via a 56k modem (crimes do become time-barred after 20 years, right?). But above all, this story fascinates me because it reminds us that behind every technological innovation there&#39;s always a human element: a voice, an aesthetic choice, an obsession. And in the case of MP3, that human element was precisely Suzanne Vega&#39;s voice singing about coffee and rain on a November morning.&#xA;&#xA;Late 1980s: the race for compression&#xA;&#xA;To understand how &#34;Tom&#39;s Diner&#34; ended up in the laboratories of the #Fraunhofer Institute, we need to step back and understand what was happening in the world of digital audio in the late 1980s. The CD had arrived in 1981, bringing the promise of perfect audio quality, crystalline, immune to scratches and the wear of time. But there was a massive problem: digital audio files were enormous. A three-minute song, encoded in PCM (Pulse-Code Modulation) format at 44.1 kHz and 16 bits, occupied around 30-35 megabytes. An entire album? Over 600 megabytes.&#xA;&#xA;To put this in perspective: in the 1980s, the portable listening revolution was the Sony Walkman, which played analog cassettes. With the arrival of CDs, Sony launched the Discman, but these portable CD players were bulky, drained batteries, and skipped at the slightest movement. The idea of carrying an entire record collection was still science fiction.&#xA;&#xA;In an era when a 40MB hard drive was considered gigantic, these numbers were simply impractical. You couldn&#39;t think of transmitting music via the internet – which was still an academic and military network – nor of efficiently archiving it on home computers. A radical solution was needed: audio had to be compressed while maintaining acceptable quality. This is where the small city of Erlangen, in Bavaria, enters the scene. Not exactly Silicon Valley, but a German town with a long tradition of scientific excellence. Here was the headquarters of the Fraunhofer Institute for Integrated Circuits, a research centre that would forever change the way we listen to music. The team was led by a man named Dieter Seitzer, who had worked for years on psychoacoustics – that branch of science studying how humans perceive sounds. Seitzer had a vision: to find a way to transmit high-quality music through ISDN telephone lines. It seemed like science fiction, but his doctoral student, a young engineer named Karlheinz Brandenburg, was convinced it was possible. The underlying idea was elegant in its simplicity: the human ear isn&#39;t perfect. There are frequencies we don&#39;t hear, sounds that get &#34;masked&#34; by louder ones, sonic details that our brain simply discards. Why waste disk space for information we can&#39;t perceive anyway?&#xA;&#xA;The goal, therefore, was to create an algorithm that eliminated everything the human ear couldn&#39;t distinguish, reducing an audio file to a tenth of its original size without the average listener noticing the difference. But the competition was fierce. In 1989, when the Moving Picture Experts Group (MPEG) – the international standardisation organisation – issued a call for audio codec proposals, 14 candidates arrived from around the world. Among them were AT&amp;T Bell Labs in the United States, Thomson in France, Philips in the Netherlands, and naturally the Erlangen team with their algorithm called ASPEC (Adaptive Spectral Perceptual Entropy Coding). It was a race where whoever demonstrated the most efficient algorithm won: maximum compression, minimum perceptible quality loss. And to prove it, tests were needed. Many tests. Obsessive, maniacal tests, repeated hundreds, thousands of times. In other words, a reference song was needed. A song that would put the algorithm to the most ruthless test possible.&#xA;&#xA;Why that voice?&#xA;Several versions exist of how Brandenburg discovered &#34;Tom&#39;s Diner&#34;. In one interview, he tells of hearing it on the radio while walking down a corridor. In another, he says he read about this song in a hi-fi magazine that used it to test high-quality speakers. The stories change, overlap, contradict each other. Brandenburg himself has given different versions over the years. But one thing is certain: when he heard that voice, he immediately knew he had found his ultimate test.&#xA;&#xA;  &#34;I was ready to fine-tune my compression algorithm,&#34; Brandenburg recalls in a 2009 interview, &#34;and somewhere down the corridor a radio was playing Tom&#39;s Diner. I was electrified. I knew it would be nearly impossible to compress this warm a cappella voice.&#34;&#xA;&#xA;And it&#39;s precisely in that phrase – &#34;nearly impossible&#34; – that you understand the challenge. The human voice is the most difficult instrument to compress. Evolutionarily, our ears are optimised to recognise voices. We evolved to hear nuances, emotions, the micro-tonal variations that distinguish one person from another, that tell us if someone is happy or sad, sincere or lying. Voice is the primary interface of human communication, and our brain has developed sophisticated mechanisms to analyse it. For this reason, any artifact, any distortion introduced by compression, immediately jumps out when dealing with voice. If MP3 could faithfully reproduce Suzanne Vega&#39;s voice, then it could handle anything.&#xA;&#xA;But why &#34;Tom&#39;s Diner&#34; specifically? What made this song so special?&#xA;&#xA;First: it&#39;s an a cappella recording. There are no instruments to mask or distract. There&#39;s no powerful bass covering the low frequencies, no electric guitars filling the mid-range. It&#39;s just voice. Naked, exposed, with nowhere to hide. Second: it&#39;s an exceptionally high-quality recording. It was recorded at A&amp;M Records studio with professional equipment, meaning it captures all the nuances, all the breaths, all the details of Vega&#39;s performance. There&#39;s no background noise that might mask compression artifacts. Third: Suzanne Vega&#39;s voice has a particular timbre – warm, intimate, with that touch of huskiness that makes it instantly recognisable. It has an interesting dynamic range, with more whispered passages and more assertive ones. It is, in essence, an acoustically &#34;complex&#34; voice.&#xA;&#xA;Brandenburg began working obsessively on that song. He listened to it hundreds of times a day, modifying the algorithm, listening again, modifying again. It was an exhausting, maniacal process. Every time he made a change to the code, he had to listen again to verify whether the result was acceptable or not. The problem was that where instrumental music still sounded acceptable, the voice became a disaster.&#xA;&#xA;Brandenburg had to keep refining, optimising, adjusting the algorithm until that voice sounded good, until he managed to capture that warmth, that intimacy, that human quality that made &#34;Tom&#39;s Diner&#34; so special. To be fair, &#34;Tom&#39;s Diner&#34; wasn&#39;t the only song used in testing. Brandenburg and his team also used other tracks: &#34;Mountains O&#39; Things&#34; by Tracy Chapman, &#34;In All Languages&#34; by Ornette Coleman, &#34;Diamonds on the Soles of Her Shoes&#34; by Paul Simon. James Johnston, from the AT&amp;T team working on a competing algorithm, also used some of these tracks. But &#34;Tom&#39;s Diner&#34; became the symbol, the ultimate test, the benchmark. If the algorithm could reproduce that voice, it could reproduce anything.&#xA;&#xA;1992: the MPEG Audio Layer-3 Standard is born&#xA;The hard work paid off. In 1992, after years of comparative testing conducted by independent institutes, the MPEG committee approved the MPEG-1 Audio Layer-3 standard. Brandenburg&#39;s team had won the competition. Their algorithm had proven superior to the others, capable of compressing audio by a factor of 10-12 while maintaining quality that most listeners judged &#34;indistinguishable&#34; from the original. But no one, at that moment, could imagine what was about to happen. MPEG-1 included three audio encoding layers: Layer-1, Layer-2, and Layer-3. Layer-3 was the most complex and most efficient, but also the most computationally demanding. In the early 1990s, home computers were still too slow to encode audio in Layer-3 in real time. It was cutting-edge technology, but without immediate practical applications. Layer-2, simpler and less efficient, was adopted for Digital Audio Broadcasting (DAB) in Europe. It seemed that Layer-3 – what would later become MP3 – was destined for a marginal role, a technical curiosity for audiophiles with powerful computers.&#xA;&#xA;Brandenburg himself had already developed a successor called Advanced Audio Coding (AAC), which was even more efficient than MP3. It seemed Layer-3 was destined for oblivion before it even took off. And then 1995 arrived. Two things changed everything: the World Wide Web and Windows 95. The Web was exploding. Suddenly, millions of people had internet access and wanted to share things: images, texts, and naturally, music. But connections were incredibly slow – 28.8k modems, if you were lucky, that took hours to download files of just a few megabytes. A format was needed that allowed music sharing in reasonable sizes. Windows 95 brought increasingly powerful computers into millions of homes, with processors capable of decoding compressed audio in real time. And, crucially, Windows used three-character file extensions to identify file types. On 14 July 1995, with a simple internal email at the Fraunhofer Institute, Layer-3 got its definitive name: .mp3&#xA;&#xA;Date: Fri, 14 Jul 1995 12:29:49 +0200&#xA;Subject: File extension for Layer 3: .mp3&#xA;Hello, In light of the overwhelming consensus of the survey participants, &#xA;the file extension for ISO MPEG Audio Layer 3 is .mp3&#xA;&#xA;Three letters that would change the history of music.&#xA;&#xA;But MP3 still needed a catalyst to take off. That catalyst arrived in the form of software. Brandenburg and his team, perhaps sensing the possibilities, perhaps just to experiment, developed a software player for Windows. They released it for free. Other developers began creating MP3 encoders, some legal with Fraunhofer licenses, others less so. The format spread virally, completely beyond its creators&#39; control. And when #Napster arrived in 1999 – the peer-to-peer file sharing service – MP3 became the standard format for large-scale music piracy. The record industry, caught completely off guard, cried scandal. Metallica protested (anyone who remembers that period raise your hand...). But it was too late. The genie was out of the bottle.&#xA;&#xA;The Irony: A Lossy Technology to Democratise Music&#xA;There&#39;s a profound irony in all this. MP3 is a &#34;lossy&#34; technology – with loss of information. Every time you compress an audio file to MP3, data is lost. Permanently. It&#39;s not reversible. An MP3, technically speaking, is a degraded version of the original. Yet this &#34;imperfect&#34; technology democratised access to music in a way no one could have predicted. It made it possible to have an entire record collection in your pocket. It allowed millions of people to discover artists they would never have listened to otherwise. It gave independent artists the ability to distribute their music without needing record labels. Brandenburg himself always had mixed feelings about MP3&#39;s success. On one hand, he was proud that his technology had had such an enormous impact. On the other, he was frustrated that many people used low bitrates – 128 kbps or less – that produced obvious sonic artifacts.&#xA;&#xA;MP3 at 320 kbps sounded excellent, practically indistinguishable from the original for most listeners. But for reasons of space and download speed, many settled for lower quality. And then there was the piracy question. Brandenburg had never imagined his technology would be used primarily to violate copyright on an industrial scale. The Fraunhofer team had worked for years on copy protection systems, DRM, digital watermarking. But none of these technologies were ever effectively implemented in the MP3 ecosystem that developed in the wild (but beautiful) west of the internet at the end of the &#39;90s. In a 1994 interview, Ricky Adar – an Indo-British entrepreneur – said to Brandenburg: &#34;Do you know that you will destroy the music industry?&#34;&#xA;&#xA;Brandenburg, at the time, thought it was an exaggeration. It wasn&#39;t. MP3 didn&#39;t destroy the music industry in the literal sense – music still exists, artists continue to create, people continue to listen. But it radically transformed it. The business model based on selling physical albums collapsed. Record labels lost their power, only to reorganise and regain it in subsequent years. Distribution became democratised. And all this thanks to a mathematical formula that eliminated frequencies the human ear struggles to perceive.&#xA;&#xA;How MP3 compression actually works&#xA;Behind the &#34;magic&#34; of MP3 lies solid mathematics. The algorithm is based on four fundamental pillars:&#xA;&#xA;MDCT Transform&#xA;The audio signal is broken down into 576 samples per frame, transformed from the time domain to the frequency domain. Basically, instead of having a waveform, we get a spectrum.&#xA;&#xA;Psychoacoustics&#xA;The algorithm calculates which frequencies are &#34;masked&#34; by louder ones. Example: if there&#39;s a very powerful drum at 100 Hz, our ear won&#39;t hear a weak sound at 110 Hz. Why waste bits encoding it? The psychoacoustic model divides the spectrum into 32 critical bands that correspond to the frequency resolution of the human ear.&#xA;&#xA;Quantisation&#xA;The &#34;important&#34; frequencies (those we hear) are encoded with more bits. Those masked or barely audible are coarsely quantised or eliminated entirely. A sound at 15 kHz, almost at the limit of audibility, might be represented with 2-3 bits instead of 16.&#xA;&#xA;Huffman Coding&#xA;The already compressed data is further compressed with entropy coding. More frequent patterns get shorter codes.&#xA;&#xA;Numerical result:&#xA;PCM Audio: 44100 samples/sec × 16 bits × 2 channels = 1411.2 kbps&#xA;MP3 at 128 kbps: compression ratio 11:1&#xA;MP3 at 320 kbps: compression ratio 4.4:1&#xA;&#xA;Suzanne Vega discovers she&#39;s the mother of MP3s&#xA;For years, Suzanne Vega had no idea of the role her song had played in MP3 development. It was the year 2000. Vega, by then an established artist with a consolidated career, was taking her daughter to nursery school. A father approached and congratulated her on being &#34;the mother of the MP3&#34;. Vega had no idea what he was talking about. The man explained he had read an article – hyperbolically titled &#34;Ich Bin Ein Paradigm Shifter: The MP3 Format is a Product of Suzanne Vega&#39;s Voice and This Man&#39;s Ears&#34; – that recounted how Brandenburg had used &#34;Tom&#39;s Diner&#34; to develop the compression algorithm. Vega was astonished. Her song, that small intimate track she had written in the 1980s while attending Barnard College, had become a fundamental piece in the history of digital technology.&#xA;&#xA;In 2007, Vega was invited to the Fraunhofer Institute in Erlangen. Brandenburg and his team played her how &#34;Tom&#39;s Diner&#34; sounded in the early versions of the algorithm, before it was refined. It was, in Brandenburg&#39;s own words, &#34;horrible&#34;. The voice was distorted, full of artifacts, almost unrecognisable. They then showed her how they had worked for months, iteration after iteration, to capture that vocal quality that made the track special. They explained the psychoacoustics, the listening tests, the obsession with detail. Vega, who had always been attentive to the quality of her recordings, appreciated the irony: a song recorded with maniacal care had helped develop a compression technology that, in a sense, sacrificed part of that quality for practical reasons.&#xA;&#xA;And there&#39;s another irony in this story. In 2012, Vega was invited to the Thomas Edison National Historical Park in New Jersey. There, she sang &#34;Tom&#39;s Diner&#34; – the song that had become the symbol of the digital revolution – recording it onto an Edison cylinder, one of the oldest and most analog recording technologies in existence. It was a symbolic gesture: bringing the song back to its analog roots, recording it with technology that predated even vinyl by decades. And naturally, someone took that Edison cylinder recording and converted it to MP3, closing the circle in a way that only modern technology could allow. The Museum of Portable Sound made that MP3 file available – an analog wax recording of the track that defined digital audio compression – as a gift for enthusiasts. An act that symbolically connects the Edison era to the Spotify era.&#xA;&#xA;From Walkman to Spotify, via iPod&#xA;Before the iPod: for twenty years, from 1979, the Sony Walkman had dominated portable listening. First with cassettes, then with the Discman for CDs. But you always had a physical limit: one cassette, one CD at a time. Pre-iPod MP3 players – like the MPMan F10 of 1998 – promised to solve this problem, but with only 32MB of storage (about 8 songs at 128kbps) they were little more than technological curiosities.&#xA;&#xA;1999: Napster arrives. Shawn Fanning, a nineteen-year-old student, creates software that allows MP3 files to be shared directly between users, without central servers. Within months, millions of people are downloading music for free. The record industry panics. Lawsuits follow, court battles. Napster is shut down in 2001, but it&#39;s too late. The model has been established: music can circulate freely online.&#xA;&#xA;2001: Apple launches the iPod. &#34;1000 songs in your pocket&#34; is the slogan. The definitive MP3 player, elegant, with an intuitive interface. The iPod wasn&#39;t the first MP3 player – there were already dozens on the market – but it was the one that made the idea mainstream. Suddenly, having your entire music collection in your pocket wasn&#39;t a nerd&#39;s dream anymore, it was a consumer reality.&#xA;&#xA;2003: Apple launches iTunes. Finally, a legal way to buy digital music. 99 cents per song, reasonable quality, no invasive DRM. It doesn&#39;t solve the piracy problem, but it offers a valid alternative. Within a few years, iTunes becomes the world&#39;s largest music retailer.&#xA;&#xA;2008: Spotify launches in Sweden. A new model: streaming, not downloading. Unlimited access to millions of tracks for a monthly fee (or free with ads). The MP3 as a file to own slowly begins to become obsolete. Why have files on your hard drive when you can have instant access to everything?&#xA;&#xA;2017: MP3 patents expire. The Fraunhofer Institute officially announces the &#34;death&#34; of MP3 and focuses on more modern codecs like AAC and Opus. But it&#39;s a purely technical death: MP3 continues to be used everywhere, a legacy format that will probably never completely die.&#xA;&#xA;Throughout all these years, Fraunhofer earned hundreds of millions of euros in royalties from MP3 patents. That money was reinvested in research, creating new generations of ever more efficient audio codecs: AAC (used by Apple), MPEG-H (for immersive audio), EVS (for 5G calls). Brandenburg, who in 2000 received the prestigious &#34;Deutscher Zukunftspreis&#34; (the German innovation prize), never stopped. Today he leads Brandenburg Labs, a startup working on advanced audio technologies like immersive audio for headphones, trying to create sonic experiences indistinguishable from reality. The original Fraunhofer team – Brandenburg, Bernhard Grill, Jürgen Herre, Harald Popp, Ernst Eberlein – has been awarded prizes and recognition worldwide. They&#39;ve entered the Internet Hall of Fame. The CE Hall of Fame. The German Research Hall of Fame. But perhaps the most significant recognition is the simplest: go to any corner of the world, ask someone of any age what an &#34;MP3&#34; is, and they&#39;ll know. A format that defined an entire era of digital culture.&#xA;&#xA;FLAC, OGG, vinyl, and the return of quality&#xA;And here we arrive at one of the most interesting parts of this story. Because not everyone embraced MP3. Not everyone embraced streaming. Not everyone settled for convenience at the expense of freedom and control. In the 2000s, while MP3 dominated and Fraunhofer profited from patents, there was already a counterculture growing silently.&#xA;&#xA;OGG Vorbis – released in 2000 by the Xiph.Org Foundation – was the open source community&#39;s response to the MP3 monopoly. While Fraunhofer and Thomson required licenses and royalties for MP3 encoders, OGG was completely free, without patents, without restrictions. Not only that: at the same bitrate, OGG often offered quality superior to MP3. It was technically better and philosophically consistent with free software ethics. For those who believed in open source, for those who rejected the idea of paying royalties on an audio format, for those who wanted full control over their tools, OGG became the format of choice. It wasn&#39;t just a technical matter: it was a matter of principle. The same spirit that had animated the free software movement in the 1980s – the GPL, the Free Software Foundation, all of Stallman&#39;s work – now extended to the world of audio codecs.&#xA;&#xA;And then there were those who completely rejected lossy compression. #FLAC – Free Lossless Audio Codec, released in 2001 – offered compression without data loss. Larger files, sure, but bit-for-bit identical to the original. For the most uncompromising audiophiles, FLAC was the only acceptable choice. But it wasn&#39;t just about digital formats. Just as digital seemed to have won, vinyl records began making a comeback. Sales, which had collapsed in the &#39;90s and 2000s, started growing again. In 2020, for the first time in decades, vinyl sales surpassed CD sales.&#xA;&#xA;Nostalgia, certainly. The charm of the physical object, the large cover, the ritual of putting the record on the turntable, certainly. But there&#39;s also a &#34;visceral&#34; element: owning a vinyl, or a CD, means owning something real, tangible. Something that can&#39;t be deleted from a server, revoked by a streaming service, lost in a hard drive crash.&#xA;&#xA;I myself, for years, have decided to stay out of streaming services. I buy, physically, CDs (almost always used), rip them to OGG, tag them properly, and put them on my FreeBSD NAS with ZFS. And then my #Navidrome server, calling them via NFS, does the rest. I&#39;ve chosen to maintain control over my data, to privilege a free and open source format over proprietary convenience. It&#39;s a choice that requires time (and a few scattered curses...), hard drives to manage, docker compose files to update, backups to make, players to configure. But it&#39;s also a choice that gives me a sense of ownership, of control that streaming cannot provide.&#xA;&#xA;There&#39;s an irony in all this: the technology that &#34;Tom&#39;s Diner&#34; helped create – MP3, lossy compression, the idea that &#34;good enough&#34; is sufficient – triggered two types of resistance. Those who rejected it for quality reasons (audiophiles with FLAC), and those who rejected it for freedom reasons (the open source community with OGG). And often, these two souls overlapped.&#xA;&#xA;But this choice is only possible because hard drives have become enormous, internet connections fast, storage cheap. The same technologies that made MP3 obsolete have made it possible to collect OGG or FLAC without thinking twice. In a sense, MP3 created the conditions for its own obsolescence – and for the birth of freer and often better alternatives.&#xA;&#xA;Some Lessons to Take Away&#xA;This story has taught us several things. It taught us that convenience often beats perfection. It taught us that technologies developed for one purpose (professional transmission via ISDN) can end up being used in completely different ways (mass file sharing). It taught us that established industries can be disrupted by technologies that initially seem marginal or niche. But perhaps the most important lesson is this: technology is always, at its core, a human matter. MP3 isn&#39;t just a mathematical algorithm. It&#39;s Suzanne Vega&#39;s voice singing about coffee and rain.&#xA;&#xA;  I am sitting in the morning&#xA;  At the diner on the corner&#xA;  I am waiting at the counter&#xA;  For the man to pour the coffee&#xA;&#xA;It&#39;s Brandenburg&#39;s obsession with capturing that warm vocal tonality. We are living, in other words, the consequences of those thousands of repeated listens to &#34;Tom&#39;s Diner&#34;, of that obsession with detail, of that search for perfect compression.&#xA;&#xA;And if Suzanne Vega hadn&#39;t written that song? If Brandenburg had chosen another track for his tests? Probably MP3 would have been developed anyway. The technology was in the air, the problem of audio compression had to be solved. But perhaps it would have taken longer. Perhaps the algorithm would have been slightly different. Perhaps history would have taken a different turn.&#xA;&#xA;I like to think that technological progress is inevitable, deterministic, that it follows an unstoppable internal logic. But stories like this remind us how random it is, how much it depends on individual choices, on coincidences.&#xA;&#xA;And now, if you&#39;ll excuse me, I&#39;m going to update the latest release of Navidrome on my Proxmox server. With Docker, obviously.&#xA;&#xA;#MP3 #DigitalAudio #SuzanneVega #TomsDiner #Fraunhofer #MusicHistory #AudioCompression #OpenSource #FLAC #TechHistory&#xA;&#xA;a href=&#34;https://remark.as/p/jolek78/a-song-an-algorithm-and-the-end-of-the-analog-world&#34;Discuss.../a&#xA;&#xA;div class=&#34;center&#34;&#xD;&#xA;· 🦣 a href=&#34;https://fosstodon.org/@jolek78&#34;Mastodon/a · 📸 a href=&#34;https://pixelfed.social/jolek78&#34;Pixelfed/a ·  📬 a href=&#34;mailto:jolek78@jolek78.dev&#34;Email/a ·&#xD;&#xA;· ☕ a href=&#34;https://liberapay.com/jolek78&#34;Support this work on Liberapay/a&#xD;&#xA;/div]]&gt;</description>
      <content:encoded><![CDATA[<p>There&#39;s a moment in the history of technology when everything changes. We don&#39;t always recognise it. Sometimes it takes years to understand that a small spark, an apparently insignificant detail, ignited a revolution that would forever change the way we live, communicate, and consume culture. In 1987, an American singer-songwriter named Suzanne Vega released a minimalist track called “Tom&#39;s Diner”. Two minutes and nine seconds of a cappella vocals, no instrumental accompaniment, no special effects. Just a voice telling the story of an ordinary morning in a New York diner. A song so essential, so pure in its simplicity, that someone on the other side of the world – a German engineer obsessed with <a href="https://jolek78.writeas.com/tag:audio" class="hashtag"><span>#</span><span class="p-category">audio</span></a> compression – would use it as a benchmark to create a technology that would shake the global music industry to its core. That technology was called <a href="https://jolek78.writeas.com/tag:MP3" class="hashtag"><span>#</span><span class="p-category">MP3</span></a>. And that voice, that “warm a cappella voice” as Karlheinz Brandenburg would later describe it, would become the ultimate test to determine whether a compression algorithm actually worked or not.
</p>

<p>This is the story – part documented reality, part urban legend – of how a folk song became the unwitting mother of the greatest revolution in music distribution since vinyl. A story that has always fascinated me because it contains all the contradictions of our digital age: innovation and destruction, democratization and loss of quality, openness and control. And yes, it&#39;s also because I&#39;ve always had a soft spot for stories that intertwine in unexpected ways. Perhaps because I too, during my years in radio, saw first-hand what it means to work with audio, manipulate it, compress it, broadcast it. Perhaps because, like many of us who lived through the transition from analog to digital, I still carry the memory of those first MP3 collections downloaded via a 56k modem (crimes do become time-barred after 20 years, right?). But above all, this story fascinates me because it reminds us that behind every technological innovation there&#39;s always a human element: a voice, an aesthetic choice, an obsession. And in the case of MP3, that human element was precisely Suzanne Vega&#39;s voice singing about coffee and rain on a November morning.</p>

<h2 id="late-1980s-the-race-for-compression" id="late-1980s-the-race-for-compression">Late 1980s: the race for compression</h2>

<p>To understand how “Tom&#39;s Diner” ended up in the laboratories of the <a href="https://jolek78.writeas.com/tag:Fraunhofer" class="hashtag"><span>#</span><span class="p-category">Fraunhofer</span></a> Institute, we need to step back and understand what was happening in the world of digital audio in the late 1980s. The CD had arrived in 1981, bringing the promise of perfect audio quality, crystalline, immune to scratches and the wear of time. But there was a massive problem: digital audio files were enormous. A three-minute song, encoded in PCM (Pulse-Code Modulation) format at 44.1 kHz and 16 bits, occupied around 30-35 megabytes. An entire album? Over 600 megabytes.</p>

<p>To put this in perspective: in the 1980s, the portable listening revolution was the Sony Walkman, which played analog cassettes. With the arrival of CDs, Sony launched the Discman, but these portable CD players were bulky, drained batteries, and skipped at the slightest movement. The idea of carrying an entire record collection was still science fiction.</p>

<p>In an era when a 40MB hard drive was considered gigantic, these numbers were simply impractical. You couldn&#39;t think of transmitting music via the internet – which was still an academic and military network – nor of efficiently archiving it on home computers. A radical solution was needed: audio had to be compressed while maintaining acceptable quality. This is where the small city of Erlangen, in Bavaria, enters the scene. Not exactly Silicon Valley, but a German town with a long tradition of scientific excellence. Here was the headquarters of the Fraunhofer Institute for Integrated Circuits, a research centre that would forever change the way we listen to music. The team was led by a man named Dieter Seitzer, who had worked for years on psychoacoustics – that branch of science studying how humans perceive sounds. Seitzer had a vision: to find a way to transmit high-quality music through ISDN telephone lines. It seemed like science fiction, but his doctoral student, a young engineer named Karlheinz Brandenburg, was convinced it was possible. The underlying idea was elegant in its simplicity: the human ear isn&#39;t perfect. There are frequencies we don&#39;t hear, sounds that get “masked” by louder ones, sonic details that our brain simply discards. Why waste disk space for information we can&#39;t perceive anyway?</p>

<p>The goal, therefore, was to create an algorithm that eliminated everything the human ear couldn&#39;t distinguish, reducing an audio file to a tenth of its original size without the average listener noticing the difference. But the competition was fierce. In 1989, when the Moving Picture Experts Group (MPEG) – the international standardisation organisation – issued a call for audio codec proposals, 14 candidates arrived from around the world. Among them were AT&amp;T Bell Labs in the United States, Thomson in France, Philips in the Netherlands, and naturally the Erlangen team with their algorithm called ASPEC (Adaptive Spectral Perceptual Entropy Coding). It was a race where whoever demonstrated the most efficient algorithm won: maximum compression, minimum perceptible quality loss. And to prove it, tests were needed. Many tests. Obsessive, maniacal tests, repeated hundreds, thousands of times. In other words, a reference song was needed. A song that would put the algorithm to the most ruthless test possible.</p>

<h3 id="why-that-voice" id="why-that-voice">Why that voice?</h3>

<p>Several versions exist of how Brandenburg discovered “Tom&#39;s Diner”. In one interview, he tells of hearing it on the radio while walking down a corridor. In another, he says he read about this song in a hi-fi magazine that used it to test high-quality speakers. The stories change, overlap, contradict each other. Brandenburg himself has given different versions over the years. But one thing is certain: when he heard that voice, he immediately knew he had found his ultimate test.</p>

<blockquote><p>“I was ready to fine-tune my compression algorithm,” Brandenburg recalls in a 2009 interview, “and somewhere down the corridor a radio was playing Tom&#39;s Diner. I was electrified. I knew it would be nearly impossible to compress this warm a cappella voice.”</p></blockquote>

<p>And it&#39;s precisely in that phrase – “nearly impossible” – that you understand the challenge. The human voice is the most difficult instrument to compress. Evolutionarily, our ears are optimised to recognise voices. We evolved to hear nuances, emotions, the micro-tonal variations that distinguish one person from another, that tell us if someone is happy or sad, sincere or lying. Voice is the primary interface of human communication, and our brain has developed sophisticated mechanisms to analyse it. For this reason, any artifact, any distortion introduced by compression, immediately jumps out when dealing with voice. If MP3 could faithfully reproduce Suzanne Vega&#39;s voice, then it could handle anything.</p>

<p>But why “Tom&#39;s Diner” specifically? What made this song so special?</p>

<p>First: it&#39;s an a cappella recording. There are no instruments to mask or distract. There&#39;s no powerful bass covering the low frequencies, no electric guitars filling the mid-range. It&#39;s just voice. Naked, exposed, with nowhere to hide. Second: it&#39;s an exceptionally high-quality recording. It was recorded at A&amp;M Records studio with professional equipment, meaning it captures all the nuances, all the breaths, all the details of Vega&#39;s performance. There&#39;s no background noise that might mask compression artifacts. Third: Suzanne Vega&#39;s voice has a particular timbre – warm, intimate, with that touch of huskiness that makes it instantly recognisable. It has an interesting dynamic range, with more whispered passages and more assertive ones. It is, in essence, an acoustically “complex” voice.</p>

<p>Brandenburg began working obsessively on that song. He listened to it hundreds of times a day, modifying the algorithm, listening again, modifying again. It was an exhausting, maniacal process. Every time he made a change to the code, he had to listen again to verify whether the result was acceptable or not. The problem was that where instrumental music still sounded acceptable, the voice became a disaster.</p>

<p>Brandenburg had to keep refining, optimising, adjusting the algorithm until that voice sounded good, until he managed to capture that warmth, that intimacy, that human quality that made “Tom&#39;s Diner” so special. To be fair, “Tom&#39;s Diner” wasn&#39;t the only song used in testing. Brandenburg and his team also used other tracks: “Mountains O&#39; Things” by Tracy Chapman, “In All Languages” by Ornette Coleman, “Diamonds on the Soles of Her Shoes” by Paul Simon. James Johnston, from the AT&amp;T team working on a competing algorithm, also used some of these tracks. But “Tom&#39;s Diner” became the symbol, the ultimate test, the benchmark. If the algorithm could reproduce that voice, it could reproduce anything.</p>

<h3 id="1992-the-mpeg-audio-layer-3-standard-is-born" id="1992-the-mpeg-audio-layer-3-standard-is-born">1992: the MPEG Audio Layer-3 Standard is born</h3>

<p>The hard work paid off. In 1992, after years of comparative testing conducted by independent institutes, the MPEG committee approved the MPEG-1 Audio Layer-3 standard. Brandenburg&#39;s team had won the competition. Their algorithm had proven superior to the others, capable of compressing audio by a factor of 10-12 while maintaining quality that most listeners judged “indistinguishable” from the original. But no one, at that moment, could imagine what was about to happen. MPEG-1 included three audio encoding layers: Layer-1, Layer-2, and Layer-3. Layer-3 was the most complex and most efficient, but also the most computationally demanding. In the early 1990s, home computers were still too slow to encode audio in Layer-3 in real time. It was cutting-edge technology, but without immediate practical applications. Layer-2, simpler and less efficient, was adopted for Digital Audio Broadcasting (DAB) in Europe. It seemed that Layer-3 – what would later become MP3 – was destined for a marginal role, a technical curiosity for audiophiles with powerful computers.</p>

<p>Brandenburg himself had already developed a successor called Advanced Audio Coding (AAC), which was even more efficient than MP3. It seemed Layer-3 was destined for oblivion before it even took off. And then 1995 arrived. Two things changed everything: the World Wide Web and Windows 95. The Web was exploding. Suddenly, millions of people had internet access and wanted to share things: images, texts, and naturally, music. But connections were incredibly slow – 28.8k modems, if you were lucky, that took hours to download files of just a few megabytes. A format was needed that allowed music sharing in reasonable sizes. Windows 95 brought increasingly powerful computers into millions of homes, with processors capable of decoding compressed audio in real time. And, crucially, Windows used three-character file extensions to identify file types. On 14 July 1995, with a simple internal email at the Fraunhofer Institute, Layer-3 got its definitive name: .mp3</p>

<pre><code>Date: Fri, 14 Jul 1995 12:29:49 +0200
Subject: File extension for Layer 3: .mp3
Hello, In light of the overwhelming consensus of the survey participants, 
the file extension for ISO MPEG Audio Layer 3 is .mp3
</code></pre>

<p>Three letters that would change the history of music.</p>

<p>But MP3 still needed a catalyst to take off. That catalyst arrived in the form of software. Brandenburg and his team, perhaps sensing the possibilities, perhaps just to experiment, developed a software player for Windows. They released it for free. Other developers began creating MP3 encoders, some legal with Fraunhofer licenses, others less so. The format spread virally, completely beyond its creators&#39; control. And when <a href="https://jolek78.writeas.com/tag:Napster" class="hashtag"><span>#</span><span class="p-category">Napster</span></a> arrived in 1999 – the peer-to-peer file sharing service – MP3 became the standard format for large-scale music piracy. The record industry, caught completely off guard, cried scandal. Metallica protested (anyone who remembers that period raise your hand...). But it was too late. The genie was out of the bottle.</p>

<h3 id="the-irony-a-lossy-technology-to-democratise-music" id="the-irony-a-lossy-technology-to-democratise-music">The Irony: A Lossy Technology to Democratise Music</h3>

<p>There&#39;s a profound irony in all this. MP3 is a “lossy” technology – with loss of information. Every time you compress an audio file to MP3, data is lost. Permanently. It&#39;s not reversible. An MP3, technically speaking, is a degraded version of the original. Yet this “imperfect” technology democratised access to music in a way no one could have predicted. It made it possible to have an entire record collection in your pocket. It allowed millions of people to discover artists they would never have listened to otherwise. It gave independent artists the ability to distribute their music without needing record labels. Brandenburg himself always had mixed feelings about MP3&#39;s success. On one hand, he was proud that his technology had had such an enormous impact. On the other, he was frustrated that many people used low bitrates – 128 kbps or less – that produced obvious sonic artifacts.</p>

<p>MP3 at 320 kbps sounded excellent, practically indistinguishable from the original for most listeners. But for reasons of space and download speed, many settled for lower quality. And then there was the piracy question. Brandenburg had never imagined his technology would be used primarily to violate copyright on an industrial scale. The Fraunhofer team had worked for years on copy protection systems, DRM, digital watermarking. But none of these technologies were ever effectively implemented in the MP3 ecosystem that developed in the wild (but beautiful) west of the internet at the end of the &#39;90s. In a 1994 interview, Ricky Adar – an Indo-British entrepreneur – said to Brandenburg: “Do you know that you will destroy the music industry?”</p>

<p>Brandenburg, at the time, thought it was an exaggeration. It wasn&#39;t. MP3 didn&#39;t destroy the music industry in the literal sense – music still exists, artists continue to create, people continue to listen. But it radically transformed it. The business model based on selling physical albums collapsed. Record labels lost their power, only to reorganise and regain it in subsequent years. Distribution became democratised. And all this thanks to a mathematical formula that eliminated frequencies the human ear struggles to perceive.</p>

<h3 id="how-mp3-compression-actually-works" id="how-mp3-compression-actually-works">How MP3 compression actually works</h3>

<p>Behind the “magic” of MP3 lies solid mathematics. The algorithm is based on four fundamental pillars:</p>

<p><strong>MDCT Transform</strong>
The audio signal is broken down into 576 samples per frame, transformed from the time domain to the frequency domain. Basically, instead of having a waveform, we get a spectrum.</p>

<p><strong>Psychoacoustics</strong>
The algorithm calculates which frequencies are “masked” by louder ones. Example: if there&#39;s a very powerful drum at 100 Hz, our ear won&#39;t hear a weak sound at 110 Hz. Why waste bits encoding it? The psychoacoustic model divides the spectrum into 32 critical bands that correspond to the frequency resolution of the human ear.</p>

<p><strong>Quantisation</strong>
The “important” frequencies (those we hear) are encoded with more bits. Those masked or barely audible are coarsely quantised or eliminated entirely. A sound at 15 kHz, almost at the limit of audibility, might be represented with 2-3 bits instead of 16.</p>

<p><strong>Huffman Coding</strong>
The already compressed data is further compressed with entropy coding. More frequent patterns get shorter codes.</p>

<p><strong>Numerical result:</strong>
PCM Audio: 44100 samples/sec × 16 bits × 2 channels = 1411.2 kbps
MP3 at 128 kbps: compression ratio 11:1
MP3 at 320 kbps: compression ratio 4.4:1</p>

<h3 id="suzanne-vega-discovers-she-s-the-mother-of-mp3s" id="suzanne-vega-discovers-she-s-the-mother-of-mp3s">Suzanne Vega discovers she&#39;s the mother of MP3s</h3>

<p>For years, Suzanne Vega had no idea of the role her song had played in MP3 development. It was the year 2000. Vega, by then an established artist with a consolidated career, was taking her daughter to nursery school. A father approached and congratulated her on being “the mother of the MP3”. Vega had no idea what he was talking about. The man explained he had read an article – hyperbolically titled “Ich Bin Ein Paradigm Shifter: The MP3 Format is a Product of Suzanne Vega&#39;s Voice and This Man&#39;s Ears” – that recounted how Brandenburg had used “Tom&#39;s Diner” to develop the compression algorithm. Vega was astonished. Her song, that small intimate track she had written in the 1980s while attending Barnard College, had become a fundamental piece in the history of digital technology.</p>

<p>In 2007, Vega was invited to the Fraunhofer Institute in Erlangen. Brandenburg and his team played her how “Tom&#39;s Diner” sounded in the early versions of the algorithm, before it was refined. It was, in Brandenburg&#39;s own words, “horrible”. The voice was distorted, full of artifacts, almost unrecognisable. They then showed her how they had worked for months, iteration after iteration, to capture that vocal quality that made the track special. They explained the psychoacoustics, the listening tests, the obsession with detail. Vega, who had always been attentive to the quality of her recordings, appreciated the irony: a song recorded with maniacal care had helped develop a compression technology that, in a sense, sacrificed part of that quality for practical reasons.</p>

<p>And there&#39;s another irony in this story. In 2012, Vega was invited to the Thomas Edison National Historical Park in New Jersey. There, she sang “Tom&#39;s Diner” – the song that had become the symbol of the digital revolution – recording it onto an Edison cylinder, one of the oldest and most analog recording technologies in existence. It was a symbolic gesture: bringing the song back to its analog roots, recording it with technology that predated even vinyl by decades. And naturally, someone took that Edison cylinder recording and converted it to MP3, closing the circle in a way that only modern technology could allow. The Museum of Portable Sound made that MP3 file available – an analog wax recording of the track that defined digital audio compression – as a gift for enthusiasts. An act that symbolically connects the Edison era to the Spotify era.</p>

<h3 id="from-walkman-to-spotify-via-ipod" id="from-walkman-to-spotify-via-ipod">From Walkman to Spotify, via iPod</h3>

<p><strong>Before the iPod:</strong> for twenty years, from 1979, the Sony Walkman had dominated portable listening. First with cassettes, then with the Discman for CDs. But you always had a physical limit: one cassette, one CD at a time. Pre-iPod MP3 players – like the MPMan F10 of 1998 – promised to solve this problem, but with only 32MB of storage (about 8 songs at 128kbps) they were little more than technological curiosities.</p>

<p><strong>1999:</strong> Napster arrives. Shawn Fanning, a nineteen-year-old student, creates software that allows MP3 files to be shared directly between users, without central servers. Within months, millions of people are downloading music for free. The record industry panics. Lawsuits follow, court battles. Napster is shut down in 2001, but it&#39;s too late. The model has been established: music can circulate freely online.</p>

<p><strong>2001:</strong> Apple launches the iPod. “1000 songs in your pocket” is the slogan. The definitive MP3 player, elegant, with an intuitive interface. The iPod wasn&#39;t the first MP3 player – there were already dozens on the market – but it was the one that made the idea mainstream. Suddenly, having your entire music collection in your pocket wasn&#39;t a nerd&#39;s dream anymore, it was a consumer reality.</p>

<p><strong>2003:</strong> Apple launches iTunes. Finally, a legal way to buy digital music. 99 cents per song, reasonable quality, no invasive DRM. It doesn&#39;t solve the piracy problem, but it offers a valid alternative. Within a few years, iTunes becomes the world&#39;s largest music retailer.</p>

<p><strong>2008:</strong> Spotify launches in Sweden. A new model: streaming, not downloading. Unlimited access to millions of tracks for a monthly fee (or free with ads). The MP3 as a file to own slowly begins to become obsolete. Why have files on your hard drive when you can have instant access to everything?</p>

<p><strong>2017:</strong> MP3 patents expire. The Fraunhofer Institute officially announces the “death” of MP3 and focuses on more modern codecs like AAC and Opus. But it&#39;s a purely technical death: MP3 continues to be used everywhere, a legacy format that will probably never completely die.</p>

<p>Throughout all these years, Fraunhofer earned hundreds of millions of euros in royalties from MP3 patents. That money was reinvested in research, creating new generations of ever more efficient audio codecs: AAC (used by Apple), MPEG-H (for immersive audio), EVS (for 5G calls). Brandenburg, who in 2000 received the prestigious “Deutscher Zukunftspreis” (the German innovation prize), never stopped. Today he leads Brandenburg Labs, a startup working on advanced audio technologies like immersive audio for headphones, trying to create sonic experiences indistinguishable from reality. The original Fraunhofer team – Brandenburg, Bernhard Grill, Jürgen Herre, Harald Popp, Ernst Eberlein – has been awarded prizes and recognition worldwide. They&#39;ve entered the Internet Hall of Fame. The CE Hall of Fame. The German Research Hall of Fame. But perhaps the most significant recognition is the simplest: go to any corner of the world, ask someone of any age what an “MP3” is, and they&#39;ll know. A format that defined an entire era of digital culture.</p>

<h3 id="flac-ogg-vinyl-and-the-return-of-quality" id="flac-ogg-vinyl-and-the-return-of-quality">FLAC, OGG, vinyl, and the return of quality</h3>

<p>And here we arrive at one of the most interesting parts of this story. Because not everyone embraced MP3. Not everyone embraced streaming. Not everyone settled for convenience at the expense of freedom and control. In the 2000s, while MP3 dominated and Fraunhofer profited from patents, there was already a counterculture growing silently.</p>

<p><a href="https://jolek78.writeas.com/tag:OGG" class="hashtag"><span>#</span><span class="p-category">OGG</span></a> Vorbis – released in 2000 by the Xiph.Org Foundation – was the open source community&#39;s response to the MP3 monopoly. While Fraunhofer and Thomson required licenses and royalties for MP3 encoders, OGG was completely free, without patents, without restrictions. Not only that: at the same bitrate, OGG often offered quality superior to MP3. It was technically better and philosophically consistent with free software ethics. For those who believed in open source, for those who rejected the idea of paying royalties on an audio format, for those who wanted full control over their tools, OGG became the format of choice. It wasn&#39;t just a technical matter: it was a matter of principle. The same spirit that had animated the free software movement in the 1980s – the GPL, the Free Software Foundation, all of Stallman&#39;s work – now extended to the world of audio codecs.</p>

<p>And then there were those who completely rejected lossy compression. <a href="https://jolek78.writeas.com/tag:FLAC" class="hashtag"><span>#</span><span class="p-category">FLAC</span></a> – Free Lossless Audio Codec, released in 2001 – offered compression without data loss. Larger files, sure, but bit-for-bit identical to the original. For the most uncompromising audiophiles, FLAC was the only acceptable choice. But it wasn&#39;t just about digital formats. Just as digital seemed to have won, vinyl records began making a comeback. Sales, which had collapsed in the &#39;90s and 2000s, started growing again. In 2020, for the first time in decades, vinyl sales surpassed CD sales.</p>

<p>Nostalgia, certainly. The charm of the physical object, the large cover, the ritual of putting the record on the turntable, certainly. But there&#39;s also a “visceral” element: owning a vinyl, or a CD, means owning something real, tangible. Something that can&#39;t be deleted from a server, revoked by a streaming service, lost in a hard drive crash.</p>

<p>I myself, for years, have decided to stay out of streaming services. I buy, physically, CDs (almost always used), rip them to OGG, tag them properly, and put them on my FreeBSD NAS with ZFS. And then my <a href="https://jolek78.writeas.com/tag:Navidrome" class="hashtag"><span>#</span><span class="p-category">Navidrome</span></a> server, calling them via NFS, does the rest. I&#39;ve chosen to maintain control over my data, to privilege a free and open source format over proprietary convenience. It&#39;s a choice that requires time (and a few scattered curses...), hard drives to manage, docker compose files to update, backups to make, players to configure. But it&#39;s also a choice that gives me a sense of ownership, of control that streaming cannot provide.</p>

<p>There&#39;s an irony in all this: the technology that “Tom&#39;s Diner” helped create – MP3, lossy compression, the idea that “good enough” is sufficient – triggered two types of resistance. Those who rejected it for quality reasons (audiophiles with FLAC), and those who rejected it for freedom reasons (the open source community with OGG). And often, these two souls overlapped.</p>

<p>But this choice is only possible because hard drives have become enormous, internet connections fast, storage cheap. The same technologies that made MP3 obsolete have made it possible to collect OGG or FLAC without thinking twice. In a sense, MP3 created the conditions for its own obsolescence – and for the birth of freer and often better alternatives.</p>

<h3 id="some-lessons-to-take-away" id="some-lessons-to-take-away">Some Lessons to Take Away</h3>

<p>This story has taught us several things. It taught us that convenience often beats perfection. It taught us that technologies developed for one purpose (professional transmission via ISDN) can end up being used in completely different ways (mass file sharing). It taught us that established industries can be disrupted by technologies that initially seem marginal or niche. But perhaps the most important lesson is this: technology is always, at its core, a human matter. MP3 isn&#39;t just a mathematical algorithm. It&#39;s Suzanne Vega&#39;s voice singing about coffee and rain.</p>

<blockquote><p>I am sitting in the morning
At the diner on the corner
I am waiting at the counter
For the man to pour the coffee</p></blockquote>

<p>It&#39;s Brandenburg&#39;s obsession with capturing that warm vocal tonality. We are living, in other words, the consequences of those thousands of repeated listens to “Tom&#39;s Diner”, of that obsession with detail, of that search for perfect compression.</p>

<p>And if Suzanne Vega hadn&#39;t written that song? If Brandenburg had chosen another track for his tests? Probably MP3 would have been developed anyway. The technology was in the air, the problem of audio compression had to be solved. But perhaps it would have taken longer. Perhaps the algorithm would have been slightly different. Perhaps history would have taken a different turn.</p>

<p>I like to think that technological progress is inevitable, deterministic, that it follows an unstoppable internal logic. But stories like this remind us how random it is, how much it depends on individual choices, on coincidences.</p>

<p>And now, if you&#39;ll excuse me, I&#39;m going to update the latest release of Navidrome on my Proxmox server. With Docker, obviously.</p>

<p><a href="https://jolek78.writeas.com/tag:MP3" class="hashtag"><span>#</span><span class="p-category">MP3</span></a> <a href="https://jolek78.writeas.com/tag:DigitalAudio" class="hashtag"><span>#</span><span class="p-category">DigitalAudio</span></a> <a href="https://jolek78.writeas.com/tag:SuzanneVega" class="hashtag"><span>#</span><span class="p-category">SuzanneVega</span></a> <a href="https://jolek78.writeas.com/tag:TomsDiner" class="hashtag"><span>#</span><span class="p-category">TomsDiner</span></a> <a href="https://jolek78.writeas.com/tag:Fraunhofer" class="hashtag"><span>#</span><span class="p-category">Fraunhofer</span></a> <a href="https://jolek78.writeas.com/tag:MusicHistory" class="hashtag"><span>#</span><span class="p-category">MusicHistory</span></a> <a href="https://jolek78.writeas.com/tag:AudioCompression" class="hashtag"><span>#</span><span class="p-category">AudioCompression</span></a> <a href="https://jolek78.writeas.com/tag:OpenSource" class="hashtag"><span>#</span><span class="p-category">OpenSource</span></a> <a href="https://jolek78.writeas.com/tag:FLAC" class="hashtag"><span>#</span><span class="p-category">FLAC</span></a> <a href="https://jolek78.writeas.com/tag:TechHistory" class="hashtag"><span>#</span><span class="p-category">TechHistory</span></a></p>

<p><a href="https://remark.as/p/jolek78/a-song-an-algorithm-and-the-end-of-the-analog-world">Discuss...</a></p>

<div class="center">
· 🦣 <a href="https://fosstodon.org/@jolek78">Mastodon</a> · 📸 <a href="https://pixelfed.social/jolek78">Pixelfed</a> ·  📬 <a href="mailto:jolek78@jolek78.dev">Email</a> ·
· ☕ <a href="https://liberapay.com/jolek78">Support this work on Liberapay</a>
</div>
]]></content:encoded>
      <guid>https://jolek78.writeas.com/a-song-an-algorithm-and-the-end-of-the-analog-world</guid>
      <pubDate>Sun, 30 Nov 2025 23:21:53 +0000</pubDate>
    </item>
    <item>
      <title>ChatGPT didn&#39;t invent anything.</title>
      <link>https://jolek78.writeas.com/chatgpt-didnt-invent-anything?pk_campaign=rss-feed</link>
      <description>&lt;![CDATA[When the world woke up astonished in November 2022 to this &#34;magical&#34; chatbot, few realized that this magic was the result of decades of research. The history of artificial intelligence begins in 1943, when Warren McCulloch and Walter Pitts proposed the first mathematical model of an artificial neuron. In 1956, at the Dartmouth Conference, John McCarthy coined the term &#34;Artificial Intelligence&#34; and the discipline was officially born.&#xA;&#xA;The &#39;60s and &#39;70s were characterized by excessive optimism: people thought strong AI was just around the corner. Two &#34;AI winters&#34; followed – periods when funding disappeared and research slowed – because promises weren&#39;t materializing. But some continued working in the shadows. Geoffrey Hinton, Yann LeCun, Yoshua Bengio – those we now call the &#34;godfathers of deep learning&#34; – continued their studies on neural networks when no one believed in them anymore.&#xA;&#xA;!--more--&#xA;&#xA;The real breakthrough came with three converging factors: computational power (GPUs), enormous amounts of data, and better algorithms. In 2012, AlexNet won the ImageNet Challenge by an overwhelming margin, demonstrating that deep learning really worked. From there, an unstoppable acceleration.&#xA;&#xA;Once upon a time in the Carboniferous...&#xA;Before ChatGPT exploded, my only knowledge of AI came from science fiction books. Philip K. Dick and his reflections on what it means to be human. Cyberpunk in general, with its technological dystopias. Gibson&#39;s Sprawl trilogy, where AIs live in cyberspace like digital deities. Those pages were my only window to a future that seemed incredibly distant.&#xA;&#xA;When I hosted the podcast Caccia al Fotone (a nice thing, but now belonging to the Carboniferous period...), I delved deeper into the subject. I read several papers published on arXiv and dedicated two episodes to AI development. In 2019, during the pandemic period, I devoured &#34;Artificial Intelligence: A Guide for Thinking Humans&#34; by Melanie Mitchell – a book that also helped me write a &#34;thing&#34; (those who know, know; those who don&#39;t, never mind...) on the evolution of computer systems and surveillance capitalism.&#xA;&#xA;I thought I had a clear picture. I thought I was prepared.&#xA;&#xA;Mea culpa&#xA;Then ChatGPT arrived.&#xA;&#xA;November 2022. First approach: total amazement. I couldn&#39;t believe my eyes. I kept asking questions, and despite all the initial hallucinations I encountered, I continued to have that &#34;wow effect&#34; typical of a child finding the most beautiful shell on the seashore (forgive me Newton for stealing that phrase, but it&#39;s always too beautiful).&#xA;&#xA;And here&#39;s my mea culpa: I set aside all my protective filters that I generally have regarding privacy, open source, control over my data. I let myself go for hours of conversations on the most diverse topics. Until one night – one of many sleepless nights – I found myself discussing with that LLM about depression, various mental disorders, and how one or more abuses can influence a person&#39;s life.&#xA;&#xA;When I realized what was happening, I stopped abruptly. I deleted the conversation, canceled my OpenAI subscription and didn&#39;t touch any LLM for more than a month. I was entrusting my most intimate thoughts to a proprietary system controlled by a corporation. I was betraying every principle I believed in.&#xA;&#xA;But I work in IT. This is a huge revolution. I couldn&#39;t afford to fall behind, nor could I simply reject it on principle. I had to find an alternative. I began to study seriously.&#xA;&#xA;Local, always local&#xA;I encountered the first models I could test locally. I discovered Hugging Face, and it was like finding an oasis in the desert. I began studying transformers, the datasets developed by the community. And I was astounded.&#xA;&#xA;Transformers are the architecture that revolutionized AI. Presented in the 2017 paper &#34;Attention Is All You Need&#34;, they replaced old recurrent neural networks (RNNs) with a more elegant and efficient mechanism: the attention mechanism.&#xA;&#xA;In simple words: instead of processing text word by word in sequence, a transformer looks at all words simultaneously and calculates which ones are most relevant to the context. When you read &#34;The bank of the river was green,&#34; the attention mechanism understands that &#34;bank&#34; refers to the river and not the financial institution, because it evaluates the weight of each word relative to the others.&#xA;&#xA;This architecture made models like BERT, GPT, and all modern LLMs possible. It&#39;s scalable, parallelizable, and extremely powerful.&#xA;&#xA;Hugging Face and the Open Source revolution&#xA;Hugging Face is much more than a platform: it has become the Library of Alexandria of the artificial intelligence era. Founded in 2016, it now hosts over 500,000 pre-trained models, 250,000 datasets, and thousands of demo applications.&#xA;&#xA;Their transformers library has democratized access to AI. With a few lines of Python you can download and use models that would cost millions of dollars to train from scratch. Hugging Face isn&#39;t the only platform doing this – there are also Ollama, LM Studio, GPT4All – but it&#39;s certainly the most extensive and collaborative.&#xA;&#xA;Here, praise must be given to the developers: this community of people scattered around the world is doing extraordinary work. They release open source models, share knowledge, meticulously document everything. They&#39;re building a real alternative to Big Tech&#39;s monopoly on AI.&#xA;&#xA;History repeating&#xA;Watching this explosion of open models, global collaboration, shared code, I had a powerful déjà-vu. This is incredibly similar to the open source revolution that happened 30 years ago.&#xA;&#xA;In the &#39;90s, Linux and the free software movement challenged Microsoft&#39;s dominance and proprietary systems. Many said it was impossible, that free software would never work. Today Linux powers 96% of the world&#39;s servers, all Android smartphones, and much of the Internet infrastructure.&#xA;&#xA;Now the same thing is happening with AI. Llama, Mistral, Falcon, Mixtral – &#34;open weight/open source&#34; models that compete with (and often surpass) their proprietary counterparts. History repeats itself, and this time I know which side to be on.&#xA;&#xA;Another server in my homeLab&#xA;I resumed studying Python, a study I had left on standby years ago. I began experimenting with training local LLM models. I added old scripts to provide my writing style (yes, it seems incredible but every coder has their own style, and it says a lot about their personality). I used Llama 3 to improve my Bash coding.&#xA;&#xA;And when I was ready, I decided to make an important purchase: I bought a small server – to add to my homelab: Proxmox, pfSense, Nextcloud, WireGuard etc... – that I would transform into an OpenWebUI system.&#xA;&#xA;OpenWebUI is a self-hosted web interface for local language models. Like ChatGPT, but running entirely on local hardware, without sending a single byte to someone else&#39;s servers.&#xA;&#xA;For the nerds reading: the simplest way to install is obviously through Docker. Here&#39;s a basic example:&#xA;&#xA;docker run -d -p 3000:8080 \&#xA;  -v open-webui:/app/backend/data \&#xA;  --name open-webui \&#xA;  --restart always \&#xA;  ghcr.io/open-webui/open-webui:main&#xA;&#xA;Once installed, just connect OpenWebUI to Ollama (the runtime for local models), download your preferred models, and you&#39;re operational.&#xA;&#xA;GPU usage is fundamental: a medium-sized LLM requires a lot of RAM and computing power. A dedicated GPU (like an NVIDIA GTX of various types) makes an enormous difference. For those using AMD, there&#39;s ROCm. With 16GB of RAM and an 8GB GPU, you can comfortably run 7B parameter models quantized to 4-bit.&#xA;&#xA;My favorite combo? AMD, Debian, Docker, OpenWebUI, Ollama and Mistral.&#xA;&#xA;A revolution. and a choice to make&#xA;We&#39;re facing a revolution that we cannot avoid. AI is here, it&#39;s powerful, and it&#39;s evolving rapidly. There are two roads ahead of us.&#xA;&#xA;The first: avoid it now, close our eyes, hope it passes or that someone else deals with it. And then, in twenty years, find ourselves chasing an evolved AI, probably impossible to understand, completely in the hands of those who controlled it from the beginning. This is the path of least resistance, but also of maximum risk. It means ceding control, understanding, and ultimately power to whoever gets there first.&#xA;&#xA;The second: study it, analyze it, use it and understand it today to be able to handle it better tomorrow. Actively participate in its evolution. Contribute to the open source community, ensure that this technology remains accessible, understandable, in the hands of many instead of a few. This path requires effort, time, sometimes admitting we were wrong (as I did). But it&#39;s the only path that leads to actual agency over our technological future.&#xA;&#xA;The choice seems obvious when stated this way, but it&#39;s not easy in practice. It requires overcoming fear, investing time, challenging our assumptions. It means getting our hands dirty with code, running models locally, understanding how these systems actually work instead of treating them as black boxes.&#xA;&#xA;I made my choice that night when I deleted my ChatGPT conversation history. I chose not to be a passive consumer of AI technology controlled by corporations. I chose to understand, to build, to contribute to the alternative that&#39;s being constructed by thousands of developers around the world.&#xA;&#xA;The technology is already here. The question is: will it be controlled by a few companies optimizing for profit and control, or will it be a tool accessible to everyone, understandable, modifiable, improvable by the community?&#xA;&#xA;As I&#39;ve learned on this journey, choosing to understand – even when it&#39;s difficult, even when it means admitting you were wrong – is always better than passively submitting.&#xA;&#xA;AI is not magic. It&#39;s mathematics, code, hardware, and above all: it&#39;s made by people. And if it&#39;s made by people, it can be understood, modified and shaped by people. For the better, not for the worse.&#xA;&#xA;The revolution is happening. The only question is: are you participating, or are you watching?&#xA;&#xA;#AI #OpenSource #LocalLLM #Privacy #ChatGPT #HuggingFace #Ollama #SelfHosted #MachineLearning #DigitalSovereignty&#xA;&#xA;a href=&#34;https://remark.as/p/jolek78/chatgpt-didnt-invent-anything&#34;Discuss.../a&#xA;&#xA;div class=&#34;center&#34;&#xD;&#xA;· 🦣 a href=&#34;https://fosstodon.org/@jolek78&#34;Mastodon/a · 📸 a href=&#34;https://pixelfed.social/jolek78&#34;Pixelfed/a ·  📬 a href=&#34;mailto:jolek78@jolek78.dev&#34;Email/a ·&#xD;&#xA;· ☕ a href=&#34;https://liberapay.com/jolek78&#34;Support this work on Liberapay/a&#xD;&#xA;/div]]&gt;</description>
      <content:encoded><![CDATA[<p>When the world woke up astonished in November 2022 to this “magical” chatbot, few realized that this magic was the result of decades of research. The history of artificial intelligence begins in 1943, when Warren McCulloch and Walter Pitts proposed the first mathematical model of an artificial neuron. In 1956, at the Dartmouth Conference, John McCarthy coined the term “Artificial Intelligence” and the discipline was officially born.</p>

<p>The &#39;60s and &#39;70s were characterized by excessive optimism: people thought strong AI was just around the corner. Two “AI winters” followed – periods when funding disappeared and research slowed – because promises weren&#39;t materializing. But some continued working in the shadows. Geoffrey Hinton, Yann LeCun, Yoshua Bengio – those we now call the “godfathers of deep learning” – continued their studies on neural networks when no one believed in them anymore.</p>



<p>The real breakthrough came with three converging factors: computational power (GPUs), enormous amounts of data, and better algorithms. In 2012, AlexNet won the ImageNet Challenge by an overwhelming margin, demonstrating that deep learning really worked. From there, an unstoppable acceleration.</p>

<h3 id="once-upon-a-time-in-the-carboniferous" id="once-upon-a-time-in-the-carboniferous">Once upon a time in the Carboniferous...</h3>

<p>Before ChatGPT exploded, my only knowledge of AI came from science fiction books. Philip K. Dick and his reflections on what it means to be human. Cyberpunk in general, with its technological dystopias. Gibson&#39;s Sprawl trilogy, where AIs live in cyberspace like digital deities. Those pages were my only window to a future that seemed incredibly distant.</p>

<p>When I hosted the podcast Caccia al Fotone (a nice thing, but now belonging to the Carboniferous period...), I delved deeper into the subject. I read several papers published on arXiv and dedicated two episodes to AI development. In 2019, during the pandemic period, I devoured “Artificial Intelligence: A Guide for Thinking Humans” by Melanie Mitchell – a book that also helped me write a “thing” (those who know, know; those who don&#39;t, never mind...) on the evolution of computer systems and surveillance capitalism.</p>

<p>I thought I had a clear picture. I thought I was prepared.</p>

<h3 id="mea-culpa" id="mea-culpa">Mea culpa</h3>

<p>Then ChatGPT arrived.</p>

<p>November 2022. First approach: total amazement. I couldn&#39;t believe my eyes. I kept asking questions, and despite all the initial hallucinations I encountered, I continued to have that “wow effect” typical of a child finding the most beautiful shell on the seashore (forgive me Newton for stealing that phrase, but it&#39;s always too beautiful).</p>

<p>And here&#39;s my mea culpa: I set aside all my protective filters that I generally have regarding privacy, open source, control over my data. I let myself go for hours of conversations on the most diverse topics. Until one night – one of many sleepless nights – I found myself discussing with that LLM about depression, various mental disorders, and how one or more abuses can influence a person&#39;s life.</p>

<p>When I realized what was happening, I stopped abruptly. I deleted the conversation, canceled my OpenAI subscription and didn&#39;t touch any LLM for more than a month. I was entrusting my most intimate thoughts to a proprietary system controlled by a corporation. I was betraying every principle I believed in.</p>

<p>But I work in IT. This is a huge revolution. I couldn&#39;t afford to fall behind, nor could I simply reject it on principle. I had to find an alternative. I began to study seriously.</p>

<h3 id="local-always-local" id="local-always-local">Local, always local</h3>

<p>I encountered the first models I could test locally. I discovered <a href="https://huggingface.co">Hugging Face</a>, and it was like finding an oasis in the desert. I began studying transformers, the datasets developed by the community. And I was astounded.</p>

<p><strong>Transformers</strong> are the architecture that revolutionized AI. Presented in the 2017 paper <a href="https://arxiv.org/abs/1706.03762">“Attention Is All You Need”</a>, they replaced old recurrent neural networks (RNNs) with a more elegant and efficient mechanism: the attention mechanism.</p>

<p>In simple words: instead of processing text word by word in sequence, a transformer looks at all words simultaneously and calculates which ones are most relevant to the context. When you read “The bank of the river was green,” the attention mechanism understands that “bank” refers to the river and not the financial institution, because it evaluates the weight of each word relative to the others.</p>

<p>This architecture made models like BERT, GPT, and all modern LLMs possible. It&#39;s scalable, parallelizable, and extremely powerful.</p>

<h3 id="hugging-face-and-the-open-source-revolution" id="hugging-face-and-the-open-source-revolution">Hugging Face and the Open Source revolution</h3>

<p><a href="https://huggingface.co">Hugging Face</a> is much more than a platform: it has become the Library of Alexandria of the artificial intelligence era. Founded in 2016, it now hosts over 500,000 pre-trained models, 250,000 datasets, and thousands of demo applications.</p>

<p>Their <a href="https://github.com/huggingface/transformers">transformers library</a> has democratized access to AI. With a few lines of Python you can download and use models that would cost millions of dollars to train from scratch. Hugging Face isn&#39;t the only platform doing this – there are also <a href="https://ollama.com">Ollama</a>, <a href="https://lmstudio.ai">LM Studio</a>, <a href="https://gpt4all.io">GPT4All</a> – but it&#39;s certainly the most extensive and collaborative.</p>

<p>Here, praise must be given to the developers: this community of people scattered around the world is doing extraordinary work. They release open source models, share knowledge, meticulously document everything. They&#39;re building a real alternative to Big Tech&#39;s monopoly on AI.</p>

<h3 id="history-repeating" id="history-repeating">History repeating</h3>

<p>Watching this explosion of open models, global collaboration, shared code, I had a powerful déjà-vu. This is incredibly similar to the open source revolution that happened 30 years ago.</p>

<p>In the &#39;90s, Linux and the free software movement challenged Microsoft&#39;s dominance and proprietary systems. Many said it was impossible, that free software would never work. Today Linux powers 96% of the world&#39;s servers, all Android smartphones, and much of the Internet infrastructure.</p>

<p>Now the same thing is happening with AI. Llama, Mistral, Falcon, Mixtral – “open weight/open source” models that compete with (and often surpass) their proprietary counterparts. History repeats itself, and this time I know which side to be on.</p>

<h3 id="another-server-in-my-homelab" id="another-server-in-my-homelab">Another server in my homeLab</h3>

<p>I resumed studying Python, a study I had left on standby years ago. I began experimenting with training local LLM models. I added old scripts to provide my writing style (yes, it seems incredible but every coder has their own style, and it says a lot about their personality). I used Llama 3 to improve my Bash coding.</p>

<p>And when I was ready, I decided to make an important purchase: I bought a small server – to add to my homelab: Proxmox, pfSense, Nextcloud, WireGuard etc... – that I would transform into an <a href="https://openwebui.com">OpenWebUI</a> system.</p>

<p>OpenWebUI is a self-hosted web interface for local language models. Like ChatGPT, but running entirely on local hardware, without sending a single byte to someone else&#39;s servers.</p>

<p>For the nerds reading: the simplest way to install is obviously through Docker. Here&#39;s a basic example:</p>

<pre><code>docker run -d -p 3000:8080 \
  -v open-webui:/app/backend/data \
  --name open-webui \
  --restart always \
  ghcr.io/open-webui/open-webui:main
</code></pre>

<p>Once installed, just connect OpenWebUI to <a href="https://ollama.com">Ollama</a> (the runtime for local models), download your preferred models, and you&#39;re operational.</p>

<p>GPU usage is fundamental: a medium-sized LLM requires a lot of RAM and computing power. A dedicated GPU (like an NVIDIA GTX of various types) makes an enormous difference. For those using AMD, there&#39;s ROCm. With 16GB of RAM and an 8GB GPU, you can comfortably run 7B parameter models quantized to 4-bit.</p>

<p>My favorite combo? AMD, Debian, Docker, OpenWebUI, Ollama and Mistral.</p>

<h3 id="a-revolution-and-a-choice-to-make" id="a-revolution-and-a-choice-to-make">A revolution. and a choice to make</h3>

<p>We&#39;re facing a revolution that we cannot avoid. AI is here, it&#39;s powerful, and it&#39;s evolving rapidly. There are two roads ahead of us.</p>

<p><strong>The first:</strong> avoid it now, close our eyes, hope it passes or that someone else deals with it. And then, in twenty years, find ourselves chasing an evolved AI, probably impossible to understand, completely in the hands of those who controlled it from the beginning. This is the path of least resistance, but also of maximum risk. It means ceding control, understanding, and ultimately power to whoever gets there first.</p>

<p><strong>The second:</strong> study it, analyze it, use it and understand it today to be able to handle it better tomorrow. Actively participate in its evolution. Contribute to the open source community, ensure that this technology remains accessible, understandable, in the hands of many instead of a few. This path requires effort, time, sometimes admitting we were wrong (as I did). But it&#39;s the only path that leads to actual agency over our technological future.</p>

<p>The choice seems obvious when stated this way, but it&#39;s not easy in practice. It requires overcoming fear, investing time, challenging our assumptions. It means getting our hands dirty with code, running models locally, understanding how these systems actually work instead of treating them as black boxes.</p>

<p>I made my choice that night when I deleted my ChatGPT conversation history. I chose not to be a passive consumer of AI technology controlled by corporations. I chose to understand, to build, to contribute to the alternative that&#39;s being constructed by thousands of developers around the world.</p>

<p>The technology is already here. The question is: will it be controlled by a few companies optimizing for profit and control, or will it be a tool accessible to everyone, understandable, modifiable, improvable by the community?</p>

<p>As I&#39;ve learned on this journey, choosing to understand – even when it&#39;s difficult, even when it means admitting you were wrong – is always better than passively submitting.</p>

<p>AI is not magic. It&#39;s mathematics, code, hardware, and above all: it&#39;s made by people. And if it&#39;s made by people, it can be understood, modified and shaped by people. For the better, not for the worse.</p>

<p>The revolution is happening. The only question is: are you participating, or are you watching?</p>

<p><a href="https://jolek78.writeas.com/tag:AI" class="hashtag"><span>#</span><span class="p-category">AI</span></a> <a href="https://jolek78.writeas.com/tag:OpenSource" class="hashtag"><span>#</span><span class="p-category">OpenSource</span></a> <a href="https://jolek78.writeas.com/tag:LocalLLM" class="hashtag"><span>#</span><span class="p-category">LocalLLM</span></a> <a href="https://jolek78.writeas.com/tag:Privacy" class="hashtag"><span>#</span><span class="p-category">Privacy</span></a> <a href="https://jolek78.writeas.com/tag:ChatGPT" class="hashtag"><span>#</span><span class="p-category">ChatGPT</span></a> <a href="https://jolek78.writeas.com/tag:HuggingFace" class="hashtag"><span>#</span><span class="p-category">HuggingFace</span></a> <a href="https://jolek78.writeas.com/tag:Ollama" class="hashtag"><span>#</span><span class="p-category">Ollama</span></a> <a href="https://jolek78.writeas.com/tag:SelfHosted" class="hashtag"><span>#</span><span class="p-category">SelfHosted</span></a> <a href="https://jolek78.writeas.com/tag:MachineLearning" class="hashtag"><span>#</span><span class="p-category">MachineLearning</span></a> <a href="https://jolek78.writeas.com/tag:DigitalSovereignty" class="hashtag"><span>#</span><span class="p-category">DigitalSovereignty</span></a></p>

<p><a href="https://remark.as/p/jolek78/chatgpt-didnt-invent-anything">Discuss...</a></p>

<div class="center">
· 🦣 <a href="https://fosstodon.org/@jolek78">Mastodon</a> · 📸 <a href="https://pixelfed.social/jolek78">Pixelfed</a> ·  📬 <a href="mailto:jolek78@jolek78.dev">Email</a> ·
· ☕ <a href="https://liberapay.com/jolek78">Support this work on Liberapay</a>
</div>
]]></content:encoded>
      <guid>https://jolek78.writeas.com/chatgpt-didnt-invent-anything</guid>
      <pubDate>Tue, 28 Oct 2025 12:56:35 +0000</pubDate>
    </item>
  </channel>
</rss>