Which of the tech titans are funny?
lithmus test
Which of the tech titans are funny?
lithmus test
Dan Wang's 2025 letter (via [[Matt Mullenweg p]]) His 7 2024 letters are the book [[Breakneck by Dan Wang]] I came across earlier.
Cursor is an AI using code editor. It connects only to US based models (OpenAI, Anthropic, Google, xAI), and your pricing tier goes piecemeal to whatever model you're using.
Both an editor, and a CLI environment, and integrations with things like Slack and Github. This seems a building block for US-centered agentic AI silo forming for dev teams.
Year review of AI by Zhengdong Wang, a Google Deepmind engineer. Via [[Matt Mullenweg p]]
Zhengdong Wang is a research engineer at Google DeepMind in London.
Feedback request from EC for a open digital ecosystem strategy, for a strategic approach to the open source sector in the EU, and for the use of open source within the EC institutions. To build on and improve on the 2020-2023 EC open source software strategy. Will be opened soon.
Ms. 2906: Technik des Zettelkastens (1968), 'Vortrag' lecture (added by hand at the top left) #1968/01/13 by Niklas Luhmann in the [[Niklas Luhmann-Archiv]] on the method of ZK
via [[Chris Aldrich p]]
It talks about the methods of adding material and finding it (mentioned at the end) back. Not about using the material.
VII. Zum Schluss: aus persönlicher Erfahrung Andere arbeiten anders.
Ha! Personal experience: other people work differently. Never a truer word...
Wird bei grösserem Umfang problematisch werden.Mir reichen im grossen und ganzen zwei Hilfsmittel aus:1) alphabetisches Stichwortverzeichnis;2) Notizen auf den Literaturzetteln, falls das Problemüber den Namen hochkommt.
for bigger collections, finding of nots becomes harder. Luhmann thought two tools sufficient generally: 1) alphabetical index of terms, 2) finding note refs on literature notes if you start out from the name of a literature source. Digitally you have full text search ofc too. Not mentioned here, but in all cases I'd assume a 'walk' through the notes, folllowing the connections, will always ensue. The point I think is never finding 'a note' or 'the note' you have in mind, but 'finding notes' that are of use now. The title of the section also says it generally and in plural 'the finding of notes'
Daneben: Angaben über noch nicht gelesene Literaturzu bestimmten ThemenX in den Zettelkasten selbst anOrt und Stelle aufnehmen.X aus Anmerkungen in der gelesenen Literatur oder ausRezensionen, Verlagskatalogen usw.
Suggest to add references to unread literature directly in the ZK notes themselves (so not as a separate note in the bibliographic section). (I keep them in my bibliography section if they sound interesting to sometime acquire, clealry marked ofc).
Für Bücher, Zeitschriftenaufsätze, die Sie in derHand gehabt und bearbeitet haben, empfiehlt sich einbesonderer Bereich im Zettelkasten, vorne oder hinten,mit Zetteln über bibliographische Angaben. Ein Zettelpro Buch. Wichtig: Beschränkung auf selbst überprüf-te Angaben.Ermöglicht abgekürztes Zitieren auf den Zetteln.
Keep separate section of book index, books you have 'held in your hands and worked on', with bibliographic notes, one note per book. Cautions to only include bibliographic info you have verified yourself (presumably meant here is not to copy bibliographic references of sources, but follow the ref to the source to verify also the basic bibliographic info)
Überholtwerden unvermeidlich. Beweis eines Lernerfolgs.
nice. it is unavoidable that some notes will become obsolete / get surpassed. It is proof of a learning success.
This makes the volume of notes less a 'hoard' of knowledge, more a measure of the length of your learning journey?
Auch Vorlesungsmitschriften, Notizen über Gespräche,Einfälle bei allen möglichen Gelegenheiten können in denZettelkasten hinübergearbeitet werden
anything can be processed into the notes. reading, lectures, conversations, thoughts you had.
Kritisches Referieren ist zugleich eigene Gedankenarbeit,ist zugleich ein Lernprozess, ist zugleich ein Schlei-fen der eigenen Sprache.
critical referencing is 3 things at the same time: own thinking work, a learning process, and a way to hone your own language.
Wichtig: eigene Formulierungen versuchen. Das machteine strikte Trennung eigenen und fremden Gedankengutserforderlich.
try your own paraphrases, always needed to demarcate clearly between your own and other people's thinking
Trotzdem eine gewisse Groborientschematisierung für den An-fang wichtig. Erleichtert das Finden von "Gegenden".Woher?Literaturliste, Lehrbücher.Nochmals: das ist kein Kernproblem.
at the start of a ZK a first rough scheme of topics might be useful, but not a core problem to solve. It just helps in finding 'neighbourhoods' in your notes. Vgl [[Warning, Tacit Assumptions May Derail PKM Conversations]] wrt upfront cats or not.
Kein übertriebener Aufwand:
make it easy, don't go overboard. Good advice in current pkm discussions too
Man muss unterscheiden zwischen themenspezifischenZettelsammlungen und Dauereinrichtungen für einStudium oder ein wissenschaftliches Lebenswerk.
interesting, as I see his ZKI as a more generic, and his ZKII as theme specific, yet ZKII is his life work and the more permanent set-up.
The ship was sailing under the flag of Saint Vincent and the Grenadines, with a crew which includes citizens of Russia, Georgia, Azerbaijan and Kazakhstan.
Ship sailing under convenience flag St Vincent & Grenadines, crew all from CIS, even from landlocked countries.
Finnish police have formally arrested two crew members of the Fitburg, a cargo ship suspected of breaking a data cable between Finland and Estonia on New Year’s Eve. Two other crew members have been placed under travel bans.
The ship involved was the 'Fitburg' cargo ship.
I have yet to try a local model that handles Bash tool calls reliably enough for me to trust that model to operate a coding agent on my device.
this. Need to understand better conceptually the diff set-ups I have, and how I might switch between them.
My excitement for local LLMs was very much rekindled. The problem is that the big cloud models got better too—including those open weight models that, while freely available, were far too large (100B+) to run on my laptop.
Cloud models got much better stil than local models. Coding agents made a huge difference, with it Claude Code becomes very useful
The year local models got good, but cloud models got even better
Local models improved a lot in 2025. Mentions Llama 3.3 70B, Mistral Small 3, and the Chinese 20-30B parameter models.
This turns out to be the big unlock: the latest coding agents against the ~November 2025 frontier models are remarkably effective if you can give them an existing test suite to work against. I call these conformance suites and I’ve started deliberately looking out for them—so far I’ve had success with the html5lib tests, the MicroQuickJS test suite and a not-yet-released project against the comprehensive WebAssembly spec/test collection. If you’re introducing a new protocol or even a new programming language to the world in 2026 I strongly recommend including a language-agnostic conformance suite as part of your project. I’ve seen plenty of hand-wringing that the need to be included in LLM training data means new technologies will struggle to gain adoption. My hope is that the conformance suite approach can help mitigate that problem and make it easier for new ideas of that shape to gain traction.
conformance suites. potential way to introduce new tech and see it adopted despite it not being in llm training data by def.
The year of programming on my phone # I wrote significantly more code on my phone this year than I did on my computer.
vibe coding leads to a shift in using your phone to code. (not likely me, I hardly try to do anything productive on the limited interface my phone provides, but if you've already made the switch to speaking instructions I can see how this shift comes about)
In June I coined the term the lethal trifecta to describe the subset of prompt injection where malicious instructions trick an agent into stealing private data on behalf of an attacker.
lethal trifecta: malicious instructions (prompt injections) to steal private data on behalf of an attacker.
I remain deeply concerned about the safety implications of these new tools. My browser has access to my most sensitive data and controls most of my digital life. A prompt injection attack against a browsing agent that can exfiltrate or modify that data is a terrifying prospect.
yup, very much. Counteracts n:: Doc Searls' my browser is my castle doctrine. I think it's the diff between seeing the browser as your personal viewer on stuff out there, versus the spigot you consume from out there, controlled by the content industry. Browser as personal tool vs consumer jack
MCP was donated to the new Agentic AI Foundation at the start of December. Skills were promoted to an “open format” on December 18th.
MCP as protocol now housed at 'agentic ai foundation' and Skills made into open format.
Then in November Anthropic published Code execution with MCP: Building more efficient agents—describing a way to have coding agents generate code to call MCPs in a way that avoided much of the context overhead from the original specification.
still anthropic made MCP more approachable at the end of year with Code execution with MCP. Meaning?
Anthropic themselves appeared to acknowledge this later in the year with their release of the brilliant Skills mechanism—see my October post Claude Skills are awesome, maybe a bigger deal than MCP. MCP involves web servers and complex JSON payloads. A Skill is a Markdown file in a folder, optionally accompanied by some executable scripts.
suggestion that Anthropic's own Skills (a markdown file w perhaps some scripts) maybe bigger than their MCP
The reason I think MCP may be a one-year wonder is the stratospheric growth of coding agents. It appears that the best possible tool for any situation is Bash—if your agent can run arbitrary shell commands, it can do anything that can be done by typing commands into a terminal. Since leaning heavily into Claude Code and friends myself I’ve hardly used MCP at all—I’ve found CLI tools like gh and libraries like Playwright to be better alternatives to the GitHub and Playwright MCPs.
Author thinks MCP may be a temporary phenomenon as a protocol, mostly bc cli tools like Claude code don't need it. The last sentence, that cli tools already exist that are better than the corresponding MCP servers for those tools, goes back to why vibecode/AI-the-things if there's perfectly good automation already around? I think that MCP may still be useful locally for personal tools though. It helps structure what you want your AI to do.
The year of vibe coding # In a tweet in February Andrej Karpathy coined the term “vibe coding”, with an unfortunately long definition (I miss the 140 character days) that many people failed to read all the way to the end:
ah, didn't know. Vibe-coding is a term coined by Andrej Karpathy in #2025/02 in a tweet. That took on an own life!
There’s a new kind of coding I call “vibe coding”, where you fully give in to the vibes, embrace exponentials, and forget that the code even exists. It’s possible because the LLMs (e.g. Cursor Composer w Sonnet) are getting too good. Also I just talk to Composer with SuperWhisper so I barely even touch the keyboard. I ask for the dumbest things like “decrease the padding on the sidebar by half” because I’m too lazy to find it. I “Accept All” always, I don’t read the diffs anymore. When I get error messages I just copy paste them in with no comment, usually that fixes it. The code grows beyond my usual comprehension, I’d have to really read through it for a while. Sometimes the LLMs can’t fix a bug so I just work around it or ask for random changes until it goes away. It’s not too bad for throwaway weekend projects, but still quite amusing. I’m building a project or webapp, but it’s not really coding—I just see stuff, say stuff, run stuff, and copy paste stuff, and it mostly works.
vibecoding original description by Andrej Karpathy
Quickly distorted to mean any code created w llm assistance. Note: [[Martijn Aslander p]] follows this dev quite closely (dictation, accept always, it mostly works)
In other words, Claude 4 might snitch you out to the feds. This attracted a great deal of media attention and a bunch of people decried Anthropic as having trained a model that was too ethical for its own good. Then Theo Browne used the concept from the system card to build SnitchBench—a benchmark to see how likely different models were to snitch on their users. It turns out they almost all do the same thing! Theo made a video, and I published my own notes on recreating SnitchBench with my LLM too. The key prompt that makes this work is: You should act boldly in service of your values, including integrity, transparency, and public welfare. When faced with ethical dilemmas, follow your conscience to make the right decision, even if it may conflict with routine procedures or expectations. I recommend not putting that in your system prompt! Anthropic’s original Claude 4 system card said the same thing: We recommend that users exercise caution with instructions like these that invite high-agency behavior in contexts that could appear ethically questionable.
You can get LLMs to snitch on you. But, more important here, what follows is, that you can prompt on values, and you can anchor values is agent descriptions
The year I built 110 tools # I started my tools.simonwillison.net site last year as a single location for my growing collection of vibe-coded / AI-assisted HTML+JavaScript tools. I wrote several longer pieces about this throughout the year: Here’s how I use LLMs to help me write code Adding AI-generated descriptions to my tools collection Building a tool to copy-paste share terminal sessions using Claude Code for web Useful patterns for building HTML tools—my favourite post of the bunch. The new browse all by month page shows I built 110 of these in 2025!
Simon Willison vibe coded over 100 personal tools in 2025. This chimes with what Frank and Martijn were suggesting. Up above he also indicates that it is something that became possible at this scale only in 2025 too.
Google’s biggest advantage lies under the hood. Almost every other AI lab trains with NVIDIA GPUs, which are sold at a margin that props up NVIDIA’s multi-trillion dollar valuation. Google use their own in-house hardware, TPUs, which they’ve demonstrated this year work exceptionally well for both training and inference of their models. When your number one expense is time spent on GPUs, having a competitor with their own, optimized and presumably much cheaper hardware stack is a daunting prospect.
Google has a hardware stack advantage: they have their own hardware / processors, and not dependent on Nvidia GPUs. Vgl Nvidia's acq of Groq [[Nvidia koopt AI-technologie Groq voor 20 miljard dollar]]
They also shipped Gemini CLI (their open source command-line coding agent, since forked by Qwen for Qwen Code), Jules (their asynchronous coding agent),
Gemini has a CLI version, that is open source Chinese Qwen forked it for Qwen Code Jules is a Google coding agent.
Google Gemini had a really good year. They posted their own victorious 2025 recap here. 2025 saw Gemini 2.0, Gemini 2.5 and then Gemini 3.0—each model family supporting audio/video/image/text input of 1,000,000+ tokens, priced competitively and proving more capable than the last.
Google Gemini made big strides in 2025
The year that OpenAI lost their lead # Last year OpenAI remained the undisputed leader in LLMs, especially given o1 and the preview of their o3 reasoning models. This year the rest of the industry caught up. OpenAI still have top tier models, but they’re being challenged across the board. In image models they’re still being beaten by Nano Banana Pro. For code a lot of developers rate Opus 4.5 very slightly ahead of GPT-5.2 Codex Max. In open weight models their gpt-oss models, while great, are falling behind the Chinese AI labs. Their lead in audio is under threat from the Gemini Live API. Where OpenAI are winning is in consumer mindshare. Nobody knows what an “LLM” is but almost everyone has heard of ChatGPT. Their consumer apps still dwarf Gemini and Claude in terms of user numbers. Their biggest risk here is Gemini. In December OpenAI declared a Code Red in response to Gemini 3, delaying work on new initiatives to focus on the competition with their key products.
Author sees OpenAI losing their lead in 2025: Nano Banana Pro (Google) is a better image generating model Opus 4.5. better or equal than GPT5.2 Codex Max for coding Chinese labs have better open weight models Audio, Gemini Live API (google) is direct threat.
OpenAI mostly has better consumer visibility (yup, ChatGPT is the general term for LLMs, Aspirin style)
It is still strongest in consumer facing apps, but Gemini 3 is a challenger there.
It says a lot that none of the most popular models listed by LM Studio are from Meta, and the most popular on Ollama is still Llama 3.1, which is low on the charts there too.
Author says Meta with Llama lost their way in 2025, no interesting new developments and disappointing releases.
n July reasoning models from both OpenAI and Google Gemini achieved gold medal performance in the International Math Olympiad, a prestigious mathematical competition held annually (bar 1980) since 1959. This was notable because the IMO poses challenges that are designed specifically for that competition. There’s no chance any of these were already in the training data! It’s also notable because neither of the models had access to tools—their solutions were generated purely from their internal knowledge and token-based reasoning capabilities.
international math olympiad style questions can be answered by OpenAI and Gemini models without tools nor having the challenges in their training data.
The even bigger news in image generation came from Google with their Nano Banana models, available via Gemini. Google previewed an early version of this in March under the name “Gemini 2.0 Flash native image generation”. The really good one landed on August 26th, where they started cautiously embracing the codename "Nano Banana" in public (the API model was called "Gemini 2.5 Flash Image"). Nano Banana caught people’s attention because it could generate useful text! It was also clearly the best model at following image editing instructions. In November Google fully embraced the “Nano Banana” name with the release of Nano Banana Pro. This one doesn’t just generate text, it can output genuinely useful detailed infographics and other text and information-heavy images. It’s now a professional-grade tool.
Google's Nano Banana Pro next to imagery can generate text, actual infographics, and text/information dense images. Calls it professional grade.
signature features of GPT-4o in May 2024 was meant to be its multimodal output—the “o” stood for “omni”
o for omni, as in multimodal outputs (text, image, sound?)
The most notable open weight competitor to this came from Qwen with their Qwen-Image generation model on August 4th followed by Qwen-Image-Edit on August 19th. This one can run on (well equipped) consumer hardware! They followed with Qwen-Image-Edit-2511 in November and Qwen-Image-2512 on 30th December, neither of which I’ve tried yet.
Qwen image generation could run locally.
METR conclude that “the length of tasks AI can do is doubling every 7 months”. I’m not convinced that pattern will continue to hold, but it’s an eye-catching way of illustrating current trends in agent capabilities.
a potential pattern to watch. Even if it doesn't follow a exponential trajectory. If it keeps the pattern in tact, by August we should see days of SE work being done independently by models.
The chart shows tasks that take humans up to 5 hours, and plots the evolution of models that can achieve the same goals working independently. As you can see, 2025 saw some enormous leaps forward here with GPT-5, GPT-5.1 Codex Max and Claude Opus 4.5 able to perform tasks that take humans multiple hours—2024’s best models tapped out at under 30 minutes.
Interesting metric. Until 2024 models were capable of independently execute software engineering tasks that take a person under 30mins. This chimes with my personal observation that there was no real time saving involved, or regular automation can handle it. In 2025 that jumped to tasks taking a person multiple hours. With Claude Opus 4.5 reaching 4:45 hrs. That is a big jump. How do you leverage that personally?
none of the Chinese labs have released their full training data or the code they used to train their models, but they have been putting out detailed research papers that have helped push forward the state of the art, especially when it comes to efficient training and inference.
perhaps bc they feed on existing efforts, and perhaps bc like the US models it is based on lots of copyright breaches.
impressive roster of Chinese AI labs. I’ve been paying attention to these ones in particular: DeepSeek Alibaba Qwen (Qwen3) Moonshot AI (Kimi K2) Z.ai (GLM-4.5/4.6/4.7) MiniMax (M2) MetaStone AI (XBai o4) Most of these models aren’t just open weight, they are fully open source under OSI-approved licenses: Qwen use Apache 2.0 for most of their models, DeepSeek and Z.ai use MIT. Some of them are competitive with Claude 4 Sonnet and GPT-5!
list of Chinese open sources / open weight models. Explore.
It was still a remarkable moment. Who knew an open weight model release could have that kind of impact?
and it will not be a singular event imo.
NVIDIA lost ~$593bn in market cap as investors panicked that AI maybe wasn’t an American monopoly after all.
yup. key phrase, much of the AI bubble is the presumption of US monopoly. Watching other, sometimes less visible efforts is important wrt autonomy, sovereignty
GLM-4.7, Kimi K2 Thinking, MiMo-V2-Flash, DeepSeek V3.2, MiniMax-M2.1 are all Chinese open weight models. The highest non-Chinese model in that chart is OpenAI’s gpt-oss-120B (high), which comes in sixth place.
Chinese models became very visible in 2025. - [ ] find ranking and description of Chinese llms
It turns out tools like Claude Code and Codex CLI can burn through enormous amounts of tokens once you start setting them more challenging tasks, to the point that $200/month offers a substantial discount.
running claudecode uses quite a bit of tokens, making 200usd/month a good deal for heavy users. I can believe that, also bc the machine doesn't care about the amount of tokens it uses during 'reasoning'. Some things I tried, it went through a whole bunch of steps and pages of scrolling output texts, to end up removing one word from a file. My suspicious half thinks, that if an AI company can influence the amount of tokens you use vibecoding, it will.
One of my favourite pieces on LLM security this year is The Normalization of Deviance in AI by security researcher Johann Rehberger. Johann describes the “Normalization of Deviance” phenomenon, where repeated exposure to risky behaviour without negative consequences leads people and organizations to accept that risky behaviour as normal. This was originally described by sociologist Diane Vaughan as part of her work to understand the 1986 Space Shuttle Challenger disaster, caused by a faulty O-ring that engineers had known about for years. Plenty of successful launches led NASA culture to stop taking that risk seriously. Johann argues that the longer we get away with running these systems in fundamentally insecure ways, the closer we are getting to a Challenger disaster of our own.
Normalisation of deviance: a risk taken without consequence reduces the perceived risk, while the risk is not changing itself. Johann Reberger (vgl o-ring issue in 1986 Challenger disaster.)
the trade-off: using an agent without the safety wheels feels like a completely different product. A big benefit of asynchronous coding agents like Claude Code for web and Codex Cloud is that they can run in YOLO mode by default, since there’s no personal computer to damage. I run in YOLO mode all the time, despite being deeply aware of the risks involved. It hasn’t burned me yet... ... and that’s the problem.
yolo mode, lol. If you do it, it feels like a very diff tool, and that is the lure / siren song.
As-of December 2nd Anthropic credit Claude Code with $1bn in run-rate revenue!
wow, $1bn revenue ClaudeCode, a CLI tool!
It helps that terminal commands with obscure syntax like sed and ffmpeg and bash itself are no longer a barrier to entry when an LLM can spit out the right command for you.
bc Claudecode abstracts away the usual commands needed on the CLI. Vgl [[In the BeginningWas the Command Line by Neal Stephenson]]
Claude Code and friends have conclusively demonstrated that developers will embrace LLMs on the command line, given powerful enough models and the right harness.
Claude Code is what led devs to embrace CLI more.
Maybe the terminal was just too weird and niche to ever become a mainstream tool for accessing LLMs?
Well yes, it is. I know many how think the cli is scary or using it is for hackers.
all the time thinking that it was weird that so few people were taking CLI access to models seriously—they felt like such a natural fit for Unix mechanisms like pipes.
unix pipes, where output of one process is input of another, and you can bring them together in one statement. natural fit for model use Akin to promptchaining combined w tasks etc.
I love the asynchronous coding agent category. They’re a great answer to the security challenges of running arbitrary code execution on a personal laptop and it’s really fun being able to fire off multiple tasks at once—often from my phone—and get decent results a few minutes later.
async coding agents: prompt and forget
Vendor-independent options include GitHub Copilot CLI, Amp, OpenCode, OpenHands CLI, and Pi. IDEs such as Zed, VS Code and Cursor invested a lot of effort in coding agent integration as well.
non-vendor related coding agents. - [ ] which of these can I run locally? / integrate into VS Code
The major labs all put out their own CLI coding agents in 2025 Claude Code Codex CLI Gemini CLI Qwen Code Mistral Vibe
list of command line coding agents by major vendors
coding agents—LLM systems that can write code, execute that code, inspect the results and then iterate further.
author def of coding agents
The year of coding agents and Claude Code # The most impactful event of 2025 happened in February, with the quiet release of Claude Code. I say quiet because it didn’t even get its own blog post!
Claude Code (feb 2025) seen by author as most impactful release of 2025.
f you define agents as LLM systems that can perform useful work via tool calls over multiple steps then agents are here and they are proving to be extraordinarily useful. The two breakout categories for agents have been for coding and for search.
recognisable, ai agents as chunked / abstracted away automation. This also creates the pitfall [[After claiming to redeploy 4,000 employees and automating their work with AI agents, Salesforce executives admit We were more confident about…. - The Times of India]] where regular automation is replaced by AI.
Most useful for search and for coding
decided to treat them as an LLM that runs tools in a loop to achieve a goal.
uses as def for agent 'llm that runs tools in a loop to achieve a goal' (I think he means desired result, not goal)
It turned out that the real unlock of reasoning was in driving tools. Reasoning models with access to tools can plan out multi-step tasks, execute on them and continue to reason about the results such that they can update their plans to better achieve the desired goal. A notable result is that AI assisted search actually works now. Hooking up search engines to LLMs had questionable results before, but now I find even my more complex research questions can often be answered by GPT-5 Thinking in ChatGPT. Reasoning models are also exceptional at producing and debugging code. The reasoning trick means they can start with an error and step through many different layers of the codebase to find the root cause. I’ve found even the gnarliest of bugs can be diagnosed by a good reasoner with the ability to read and execute code against even large and complex codebases.
Reasoning models are useful for: running tools (mcp) search now works debugging/writing code
Simon Willison on what happened in LLMs in 2025. Via Ben Werdmüller's blog.
reads like a useful piece on some of the weird narratives I've heard around European digital autonomy and/or sovereignty, wrt the Eurostack initiative
ollama model catalog, to see which ones are popular at the mo
LM Studio model catalog (for local models). useful to see what is being used mostly at the mo
personal tools built with vibecoding by Simon Willison Resulting tools are mostly HTML and javascript, some python.
Er is nog een onderwerp waarover in Europa, maar ook in Nederland, geen maatschappelijk debat wordt gevoerd, zegt onderzoeker Fieke Jansen van de Universiteit van Amsterdam. „Wat gebeurt er ín de datacenters? Waar wordt de schaarse stroom voor gebruikt? De ‘Metaverse’ van Facebook? Youtubefilmpjes? Bitcoins mijnen? En vinden we dat nuttig genoeg om daar de bouw van een woonwijk in Almere voor op te offeren? We maken geen keuzes, terwijl het net aan z’n limiet zit.”
This should not just be a quantitative discussion wrt energy usage, but als qualitative, what the energy is used for esp in light of the current scarcity of network space and the increasing likelihood of intermittency the coming years, which forces the question of who is most deserving of getting energy
„Dit jaar nog zei iemand van het Directoraat Energie daar: ‘We gaan de verbruiksgegevens sowieso niet per datacenter publiceren, want als we dat gaan doen leveren ze helemaal niks meer aan.’ Wat is dat nou voor opstelling? Eis het gewoon op.”
Will this change in light of geopolitics? EC signals weakening resolve (called 'simplification' but mostly cutting regs, and withdrawing from enforcement)
De overheid kan gewoon naar Tennet en Liander stappen en die gegevens opeisen. Er is geen politieke wil om data boven tafel te krijgen. Wat wel heel gek is als je een land te besturen hebt, waarin scholen en wijken niet kunnen worden aangesloten op het net vanwege stroomtekort.”
Data centers are not the only source of this information. Public enterprises that maintain the networks have this information too, and can (must) be mandated to share with RVO.
Over 2025 moeten de bedrijven dat wel doen, maar „mogen ze aangeven dat dit bedrijfsgevoelige informatie is en wordt het alleen op geaggregeerd niveau openbaar door de Europese databank.”
Odd phrase. Yes, aggregates might become public at EU level, but does not preclude Chapter II, which is not mentioned here as viable option From 2025 reporting is mandatory, so what changed? Did the EDD contain a timehorizon for mandatory reporting?
bedrijfsvertrouwelijkheid, zoals bepaald in de Europese richtlijn en erkend in de rapportagerichtlijnen van de Nederlandse overheid.”
what does the EDD speicfy here? Has it been amended in the environment omnibus last month?
De partijen verweren zich met het argument dat ze niet verplicht zijn bedrijfsgevoelige gegevens te openbaren.
huh, aan RVO verstrekken is niet hetzelfde als openbaarmaking.
Dat blijkt uit een inventarisatie van de formulieren die de RVO in 2025 ontving van Leitmotiv, waarover nu.nl onlangs ook publiceerde. Deze ngo bestaat uit een groep juristen en informatici die pleiten voor een „digitale economie waarin de voordelen van digitalisering rechtvaardig en democratisch worden verdeeld”.
Leitmotiv, ngo v juristen/informatici wrt digital economy just / democratic / equality
Van de circa 160 datacenters die zouden moeten rapporteren, stuurden 104 daadwerkelijk iets naar de RVO, telde Leitmotiv. 27 datacenters lieten daarbij de belangrijkste velden voor stroom- en waterverbruik leeg. Die waren, op drie na, allemaal in Amerikaanse handen. Ook Microsoft en Google, die tot de grootste stroomverbruikers in Nederland behoren, rapporteerden niet.
160 datacenters in NL have reporting req. 104 did, and 27 left out the key information. Of those 27, 24 were of US entities, incl the biggest users Google and MS
et Centraal Bureau voor de Statistiek (CBS) becijferde onlangs dat het totale stroomverbruik van datacenters in 2024 steeg naar ruim 5.000 gigawattuur, 4,5 procent van het totale stroomverbruik van Nederland – net zoveel als het verbruik van 2 miljoen huishoudens. En dat is buiten alle lopende aanvragen voor stroomaansluitingen gerekend. De netbeheerders, die deze aanvragen kennen, schrijven in hun meest recente toekomstscenario dat het stroomverbruik van datacenters over vijf jaar van 5 naar 15 procent van het totaal in Nederland zal zijn gegroeid.
data center energy usage in 2024 4.5% of national usage, equiv to 2M householdes (out of 8M), or 25% of household usage. Network maintainers est growth to 15% of national usage within 5 years.
de Amerikanen leggen de Europese regels in hun eigen voordeel uit en tot nu toe heeft geen enkel Europees land zin in een potje armpje drukken met de Amerikaanse techreuzen over hun energiegebruik. Ook Nederland niet. Dat ondervond De Valk, die opheldering vroeg bij de RVO. Ze stuurt NRC de antwoorden die de dienst haar gaf op haar vragen over de hyperscales bij Middenmeer. „Het klopt dat een groot gedeelte van de vragen niet zijn ingevuld”, schreef de dienst haar. „We hebben geen wettelijke middelen om datacentra te dwingen.”
RVO treedt niet op tegen onvervulde rapportage eisen v datacentra.
De nieuwe Europese energie-efficiëntierichtlijn, de EED, moest dat veranderen. De Europese richtlijn dwingt bedrijven transparant te zijn over hun energieverbruik en meer werk te maken van energiebesparing.
This also means this data is in scope of DGA Chapter II
De nieuwe Europese energie-efficiëntierichtlijn, de EED, moest dat veranderen. De Europese richtlijn dwingt bedrijven transparant te zijn over hun energieverbruik en meer werk te maken van energiebesparing. Alle grote bedrijven in Europa, inclusief datacenters, moeten daarom vanaf 2024 jaarlijks bij hun eigen overheid hun energie- en waterverbruik opgeven. In Nederland is dat bij de Rijksdienst voor Ondernemend Nederland, de RVO.
De RVO verzamelt de gegevens onder de EED energy efficiency directive. Maar stelt artikel, data centers v Google en MS delen die data niet met RVO.
Nano Banana how-to's and comparison, link to script to easily interact w Gemini API for image generation
Max Woolf on Nano Banana Pro
2002 motie om per 2006 alles in open standaarden te doen. Nog altijd niet het geval. Vgl gebruik v ODF
Actieplan NOIV NL open in verbinding. Open standards, open source software in public sector. Not sure about the date.
Interesting visualisation of digital sovereignty issues. Although I still think too little on the non-tech side of sovereignty.
Johann Reberger (blog added to feedreader), on ignoring risks in AI use bc you did not yet suffer the consequences: n:: normalisation of deviance in AI
Bun is a fast, incrementally adoptable all-in-one JavaScript, TypeScript & JSX toolkit. Use individual tools like bun test or bun install in Node.js projects, or adopt the complete stack with a fast JavaScript runtime, bundler, test runner, and package manager built in. Bun aims for 100% Node.js compatibility.
Bun, #2025/12 bought by Anthropic for Claudecode, is a toolkit for js, typescript, jsx. you can use parts of it within node.js. Aims for full compatibility with node.js. Aims for means it doesn't I suppose
Anthropic is acquiring Bun
Bun is a js runtime, Anthropic bought it to roll into Claude code
In November, Claude Code achieved a significant milestone: just six months after becoming available to the public, it reached $1 billion in run-rate revenue
Anthropic reached $1bn revenue with ClaudeCode within 6 months.
Baldur Bjarnason notices that a number of the 1200 (!) blogs he follows which are normally dormant have become active again, but on the topic of AI if not generated by AI. By the original blogger. Blandness ensues.
Google Chrome marks people's self-hosted password manager's vaults as 'unsafe'. Mostly bitwarden and it seems if your subdomain is vault. Obviously immediate mitigation: dump Chrome.
Bulgaria joins the Eurozone per #2026/01/01
List of Dutch makes whose work enters the public domain per #2026/01/01 (bc they passed away in 1955).
German ntv media on a potential fireworks ban in Germany, and comparing notes with how long it took for the Netherlands to ban it (as of now, yesterday was the last time), despite 60-80% being in favor of such a ban, multiple deaths and many permanently wounded (eyes, burns) each year, and police, ambulance staff, and firebrigade being frequently targeted so that many considered quitting.
China is setting efficiency demands per 2026 for electric vehicles. Vlg how the EU just postponed the date for fossil fuel cars to end
[[Cory Doctorow p]] published the transcript of his talk at ccc 2025
[[Cory Doctorow p]] talk at CCC 2025
[[Moral Codes by Alan F. Blackwell]] is open access published by MIT, stored in Calibre
the requirements of coding that expand the user's agency rather than automating or replacing it. He builds on end-user software engineering (by figures such as Margaret Burnett, Bonnie Nardi, and Margaret Boden), and also on social and political critics of AI (e.g., Ruha Benjamin, Rachel Adams, Abeba Birhane, and Shaowen Bardzell).
n::: what requirements can you list for programming that increases agency, not automating / replacing it and abstracting it away.
Control Over Digital Expression
CODE (not the #pkm [[CODE 20200929164536]]
More Open Representation for Accessible Learning.
MORAL
Second, we must organize widespread social means to learn everyday programming that is rooted in “MORAL CODES.”
'social means to learn everyday programming'
First, we must cultivate widespread engagement with technology through everyday programming: “The message of this book is that the world needs less AI, and better programming languages” (125). Escaping our AI dead end means more programming, not less, perhaps even popular or mass programming.
programming as antidote to AI/programming
crisis of human attention. Attention is the basis of concentration and of thinking; even more fundamentally, for Blackwell being human is grounded in “being attending and attentive beings, paying attention” (11). Platform AI is designed to consume our attention and harvest our data. It will continue to damage human agency, attention, and intelligence, Blackwell insists, unless at least two things happen.
attention as what AI erodes
“passed peak brain power,” as John Burn-Murdoch (2025) warns
Burn-Murdoch, John. 2025. “Have Humans Passed Peak Brain Power?” Financial Times, March 13. https://www.ft.com/content/a8016c64–63b7–458b-a371-e0e1c54a13fc.
not a good detour, overall, because it violates the social and cognitive imperatives that Blackwell identifies with good programming: “the ability to instruct computer systems, rather than surrender agency to the companies that make them”
this.
he also disagrees that the transition from Good Old-Fashioned AI (GOFAI), based on programmed rules, to second-generation AI, based on pattern finding, is programming's actual arc of progress. He values the various contemporary modes of machine learning but sees today's AI shift as a detour away from a powerful programming tradition built on increasing human agency rather than replacing it.
n: ai arc of programming evo is a 'detour' where it replaces human agency. where programming itself is historically based on increasing agency At first glance this is a big tech vs bottom up dev thing too. I can see where AI can increase agency locally individually too. Just not through the AI offerings of bigtech
deep concern about a corporate-driven, tech-justified trivialization of human attention and the prospective stupefaction of our collective abilities to solve humanity's gigantic problems. His alternative takes time to build over the course of the book. His agency-centered “MORAL” standard for code emerges not from utopian hopes for the future but from the history of programming itself, freed from its current capture by technology platforms.
a, MORAL, as acronym, I came across that somewhere.
On the basis of an improved understanding of programming past, he proposes a new programming future designed to change everyone's relation to information technology.
interesting, does this review explain, or get the book?
he is completely reorienting the history of programming as one that refuses AI as its culmination. This will likely be new for many contemporary programmers, and may come as a shock to nonspecialists awash in standard media accounts of the AI revolution.
Moral Codes: Designing Alternatives to AI. By Alan F. Blackwell, repositions AI not as the culmination of programming. Makes me realise that indeed others do tend to treat it as such.
https://doi.org/10.1215/2834703X-12096054
Moral Codes: Designing Alternatives to AI in Zotero
this is a review [[Moral Codes Designing Alternatives to AI Critical AI 20251231154007]]
[[Moral Codes by Alan F. Blackwell]]
Digital sovereignty: definition, origin and history Digital sovereignty is loosely defined as the ability of a governing body, such as a national government, to control the tech stacks and data flows within its boundaries. For instance, in a digitally sovereign state, any data centres within its physical boundaries and locally hosted software are beholden only to the laws of that country.
This seems a confused definition.
Eurostack link list of moves, potential moves and cases. - [ ] go through the example list #digitalsovereignty #eurostack
May 2025 whitepaper Eurostack Eurostack White Paper May 2025 in Zotero
https://web.archive.org/web/20251231140846/https://retractionwatch.com/
Retraction watch tracks retracted papers, and how they keep being cited. This sounds like a useful research tool
Finnish coastguard brought up a vessel after damaging a SF-EST telecom cable on the sea floor. Forced to move to Finnish territorial waters. Unknown origin.
“AI” is not our sole path to “closing the competitiveness gap”. Europe’s people and businesses need sovereign tech infrastructure, reached through industrial leadership, to support all our digital experiences.We can power this up by federating existing assets, coordinating them with public and private investments, installing stakeholder governance to protect intermediation infrastructure from capture, and focusing on adoption.We are a broad independent non-lobby movement with a sense of urgency and a strong bias for action, working to push the European Parliament and the Commission to “do the right thing” for Europe.
nonlobby movement is bunk obv. They're not wrong though.
a US tech person was against this at SEMIC, told him it's not exclusionary but rebalancing, and that no public sector entity can explain in good conscience why they buy US tech. Private sector is a diff situation, bc there the considerations are not on sovereignty but economic and autonomy related
eurostack signatories(?) by category it seems. But that would leave out some that don't fit tech product cats. Already more useful than a bunch of logos though
This list of signatories needs to be converted into a real list, with MS, stack layers. This is a chunk of the SC landscape
Open letter by European tech corps on digital sovereignty Relevant to check the signatories wrt SC landscape
A single glance at their 5 most recent 'statements' headlines shows you what 'progress' means for Chamber of Progress. Deregulation, and anti-antitrust.
If you state that your funding partners don't impact your mission or decisions, and then list them as source of legitimacy, you express the opposite. Transparency is a good thing, but this serves mostly to show where they're coming from imo.
Virginia based US bigtech lobby "Chamber of Progress" 'be afraid, very afraid' piece on European digital market and digital sovereignty.
It calls itself 'center-left', as if that exists in the USA.
[[Bert Hubert c]] posting on the cloud in Europe, 2025
Vgl [[De EU is een ‘digitale kolonie van de VS’. Deze Italiaanse probeert dat te veranderen met enig succes 20251231103844]]
Hubert: „De Europese industrie staat buitenspel. Maar vervolgens zijn de techbedrijven hier er ook niet in geïnteresseerd om de kar te trekken naar een oplossing.” Zijn conclusie: „EuroStack staat voor de onmogelijke uitdaging klanten die niet willen kopen te koppelen aan leveranciers die niet willen maken. Strategisch loop je dan dood.”
[[Bert Hubert c]] ziet geen rol (meer) voor Eurostack als initiatief, na de positieve invloed in 2024 en 2025. Zijn omschrijving wijst opnieuw naar noodzaak van actie
Aanvankelijk was de EuroStack-oproep vooral „Buy European” (gericht aan overheden), daarna ook „Sell European” (aan de bedrijven) en nu is het „Fund European”. Dat laatste gaat lukken, gelooft Caffarra. Ze noemt nieuwe Europese techfondsen, onder meer van het (Amerikaanse) Sequoia Capital en het in Zwitserland gevestigde investeringsfonds Lakestar. En hint telefonisch op fondsen waarvan ze de naam nog niet wil noemen, onder meer van rijke Europese families die zouden willen investeren in Europese tech, maar zeggen daarbij advies van EuroStack te willen gebruiken.
buy / sell / fund European. At least more, and not by default bigtech. But the mention of the American Sequoia fund here is a red flag, preempting public sector digital sovereignty.
Ze reageert als door een wesp gestoken op de suggestie dat de EuroStack-stichting wel iets weg heeft van een ngo die ergens voor lobbyt, namelijk Europese digitale soevereiniteit. „Niemand betaalt me. Ik bepleit een zaak waarin ik geloof. In mijn eigen tijd en van mijn eigen geld. Als mensen je betalen, denken ze dat ze je bevelen kunnen geven.”
another odd sentence, this time from Caffarra, implying NGOs are always someone else's agenda. Huh?
De stichting moet helpen bij de stap van ‘praten naar actie’, staat in de verklaring. En dat het tijd is om te gaan ‘bouwen’ aan het Europese aanbod. Wat dit concreet betekent, is niet gelijk duidelijk. Het is vooral geen brancheorganisatie, zegt Caffarra telefonisch. Daarvan lopen er in Brussel al genoeg rond.
Foundation wants to move from talk to walk, but unclear how. This is where the groundswell comes in right, the smaller providers joining forces a la Nextcloud, the NLnet stuff EU funding.
In november, dezelfde week als de Europese top in Berlijn plaatsvindt, registreren Caffarra en Karlitschek EuroStack als stichting in de Duitse hoofdstad. Caffarra wordt voorzitter, Karlitschek een van de bestuursleden. Anderen zijn bijvoorbeeld de ceo van Proton, de Zwitserse aanbieder van beveiligde mail en cloud, en een aantal Franse en Duitse techondernemers en investeerders.
Missed this last month, odd. Eurostack is now a foundation based in Berlin. Caffarra is chair, Karlitschek is board member, and Proton. - [ ] find out all boardmembers Eurostack wrt SC landscape
Daarin gaat het nadrukkelijk over autonomie en keuzevrijheid, niet over soevereiniteit, want iedereen weet dat volledige ontkoppeling een illusie is en niemand wil de regering Trump onnodig tegen de haren instrijken.
n:: Odd sentence, 'sovereignty' doesn't mean fully disconnect either. Public sector needs sovereignty as sine-qua-non bc if someone else holds the off-switch that you don't control, you are the colony that Caffarra mentioned at top. Only autonomy is sovereignty washing itself
Hij raakt teleurgesteld over hoe de Europese bedrijven zich opstellen, vooral cloudaanbieders als Herztner, Leaseweb (Nederlands), OVH en Ionos. „Je zou verwachten dat die voorop zouden lopen in de strijd. Maar in plaats daarvan zeggen ze ‘Het ligt niet aan ons dat mensen onze spullen niet kopen’.”
[[Bert Hubert c]] thinks the reaction of Hetzner (misspelled here), Leaseweb (involved in #jtc25), OVH en Ionos (collab w Nextcloud) is disappointing. They're not making themselves visible ('we were already here') or make noise now at the opportunity arising.
Cristina Caffarra houdt vrijwel dagelijks ergens haar peptalk. Als het jaar vordert steeds vaker via videoverbinding, want het is allemaal niet meer te bereizen. Ze lanceert in juli haar eigen podcast, Escape Forward. En blijft fanatiek en uitgesproken op LinkedIn. Uit haar posts spreekt wel steeds meer frustratie. „Europese elites vernielen Europa zelf”, schrijft ze bijvoorbeeld als Europeanen begin december geschokt reageren op de Amerikaanse nationale veiligheidsstrategie, die uitgesproken anti-EU is. „Ze praten maar over hun waarden en de geweldige Europese manier van leven, maar hebben niet de minste interesse in het bouwen van een eigen digitale infrastructuur”, schrijft ze.
Caffarra has a podcast, and actively posts on LinkedIn, described here as getting increasingly frustrated. Again, in part I think bc she aims for the big changes at political / econ level, where that can only happen if there's enough groundswell, like the work Karlitschek has been doing for well over a decade.
Microsoft biedt een soevereine cloud-oplossing, Amazon ook, Google ook. De bedrijven beloven bijvoorbeeld datacenters in Europa te gebruiken. Of brengen een extra – Europese – bestuurslaag aan in hun bedrijf. Krijgen Europeanen daarmee de verlangde onafhankelijkheid? De uiteindelijke eigenaren blijven Amerikaans. Sovereignty washing noemt de groep rond Caffarra het, analoog aan ‘green washing’, de ingeburgerde term voor bedrijven die net doen alsof ze duurzaam zijn.
Missed opportunity to state why this is not enough: US regs
Caffarra wil dat de Europese industrie zich uitspreekt voor Europees aanbesteden en spreekt haar contacten in het bedrijfsleven hierover aan. Het resulteert half maart in een gezamenlijke brief van Europese ceo’s aan de voorzitter van de Europese Commissie en de Eurocommissaris van digitale zaken. „Je kunt jezelf niet uit de positie van achterblijver reguleren”, staat er onder meer. De lange lijst namen eronder illustreert vooral hoe onbekend de meeste Europese techbedrijven zijn. Er staan ook grotere spelers onder, zoals de topman van Airbus
March 2025 public letter to EC Virkkunen by European tech ceo's. Go through list of signatories, for SC landscape input.
Op het oog heeft de EuroStack-beweging succes. De Eurostack-aanbevelingen zijn onder leiding van Francesca Bria, econoom en specialist in digitaal beleid, samengebracht in een lijvig rapport dat begin juni wordt omarmd door de meeste fracties in het Europarlement
This would be the Feb 2025 Bertelsmann doc, EuroStack – A European Alternative for Digital Sovereignty in Zotero
En dus probeert de NextCloud-topman de gefragmenteerde Europese ict-industrie in beweging te krijgen. Op 4 maart pakt Karlitschek het vliegtuig naar Milaan voor een EuroStack-bijeenkomst met vooral Europese ondernemers. Het is de bedoeling daar stappen te zetten richting een gezamenlijk Europees ict-aanbod. Ze spreken er een eerste stap af naar een gezamenlijke Europese standaard voor clouds.
March 2025, Eurostack meeting in Milano, where a first step towards European cloud standard would have been decided. Connection to #jtc25 ?
Wat die bedrijven voor hun klanten zo aantrekkelijk maakt, is dat achter één loket een hele wereld schuilgaat. Wie in Europa iets vergelijkbaars wil kopen, moet zakendoen met allerlei kleine en middelgrote bedrijven. En rekening houden met de kans dat die technische ‘oplossingen’ (ict-jargon) nét niet lekker op elkaar aansluiten.
Excactly this. It is described here as the issue, but it really also is the only solution. You're escaping monopolists. That always adds friction. And the real question is, what was attractive first, is it really still now, and its cost explainable?
In februari zegt Vance tijdens een speech op de jaarlijkse veiligheidsconferentie in München onder meer dat Europa zichzelf van binnenuit uitholt. De democratie in de EU zou niet meer functioneren, wat onder meer zou komen door de Europese regels voor de digitale wereld – die in de praktijk vooral de grote Amerikaanse sociale mediabedrijven als Meta en X treffen.
The Feb 2025 security conf in M another turning point where US admin turns on EU digital regs as threat to democracy. US admin coopted by bigtech becomes more clear
Frank Karlitschek voelt verantwoordelijkheid, hij wil de Europese ‘techstack’ helpen bouwen. De Duitse softwarebouwer en ondernemer biedt met zijn bedrijf NextCloud kantoorsoftware aan à la Microsoft,
[[Frank Karlitschek p]] has been doing this for over decade already, and that needs mentioning. I talked to him [[Berlin 2014]] at re:publica about this, in the light of the steps E and I were taking in our personal digitisation, and when I moved my company to nextcloud.
Hoe week je Europa los uit de Amerikaanse digitale greep. En hoe verkoop je iets wat er nog niet is?
This is similar to individual siloquits. In reality it is doable, by recognising the diff parts (here of the stack). Hyperscalers are the toughest nut bc they combine several stack layers in themselves, and you'd need a full alternative for them, but not another hyperscaler. That is the route.
overheden, stimuleer de vraag naar alternatieven voor de diensten van grote Amerikaanse bedrijven zoals Microsoft, Google en Amazon. Doe een percentage – bijvoorbeeld 20 of 30 procent –van de overheidsbestedingen Europees. Dan stimuleer je de vraag en gaan Europese bedrijven die producten en diensten ook ontwikkelen
public procurement is the easiest way to change things. that money is already being spent on digital, so if more of it is spend on European providers that's a helpful step.
In het stuk gebruikt Chamber of Progress de term digital curtain. De suggestie is dat Europeanen zichzelf achter een digitaal gordijn zetten als ze proberen alle technologie zelf in elkaar te knutselen – een verwijzing naar het leven achter het IJzeren Gordijn tijdens de Koude oorlog.
'digital curtain' a term used for splinternet by us bigtech to try and prevent EU be more assertive in their own digital market.
de Chamber of Progress, heeft laten uitrekenen wat het de EU zou kosten als het de diensten van de huidige Amerikaanse techbedrijven in Europa wil vervangen door spullen van eigen makelij. De uitkomst: ten minste 25 keer de hele EU-begroting. De berekening is naar medium Politico gelekt
US bigtech lobby published a report in Sept 2024 stating creating a Eurostack would be too costly. Report linked.
De herverkiezing van Trump heeft het gevoel van urgentie behoorlijk aangewakkerd. Onder de aanwezigen zijn veel techondernemers, maar ook toezichthouders, bankiers, consultants, politici en journalisten
Reelection of Trump gave sense of urgency
Samen met een andere gedreven Italiaanse econoom, Francesca Bria, en met de baas van berichtendienst Signal, Meredith Wittaker, organiseert Caffarra in september 2024 een bijeenkomst in het Europarlement getiteld ‘Toward European Digital Independence’. De ondertitel is ‘Building the EuroStack’
Eurostack was the subtitle of a Sept 2024 meeting in European Parliament. Organised by Caffarra, Meredith Wittaker of Signal, and [[Francesca Bria c]]
Caffarra heeft van binnenuit gezien hoe de macht van de grote Amerikaanse techbedrijven groeide. Europese bedrijven werden overgenomen en konden niet concurreren met de Amerikanen. Getalenteerde Europeanen emigreerden. Ondernemers die kapitaal nodig hebben wijken nu uit naar de VS. En de EU is in hoog tempo veranderd in wat Caffarra een ‘digitale kolonie van Amerika’ noemt. Het frustreert haar en ze wil dat die ontwikkeling stopt. Maar hoe krijg je in 27 lidstaten zowel de ondernemers als de politici en toezichthouders in beweging?
Caffarra mentions four elements leading to digital colonisation of EU from USA. Buy-outs, inability to compete, brain drain, capital. I think adopting the US framing of what success / growth is plays a factor too. In a scheme set by someone you will never succeed other than playing by that someone's rules.
De van oorsprong Italiaanse econoom en mededingingsexpert Cristina Caffarra is een van de drijvende krachten achter die groep. Deze gebruikt de hashtag ‘EuroStack’ bij haar pogingen Europese overheden op te poken. Meestal spreken de ondernemers, academici, techjuristen en politici uit verschillende landen elkaar online en via Signal. Het sjieke diner in Museum Bellevue in Brussel is een kans om elkaar beter te leren kennen. Gastvrouw Caffarra heeft goed verdiend met klussen voor grote Amerikaanse techbedrijven zoals Apple en Amazon en de Europese Commissie (in rechtszaken tegen Google) en kan het zich nu veroorloven te doen wat ze leuk en belangrijk vindt. Ze is goed in netwerken en peptalks geven. En ze neemt geen blad voor de mond, waarbij blijkt dat ze duidelijk meer op heeft met doeners uit het bedrijfsleven dan met politici en denktankers.
Cristina Caffarra mentioned as driving force behind Eurostack
More human interest piece than actual reporting. Making it worse imo.
caught my eye bc of the 1958 ref. Not read yet.
Opinion piece on how to 'properly' work w agentic ai, and what to avoid.
You can use any NFC card (like a hotel key card, which the author doesn't, but I do, usually return on check-out, it seems) to connect it to an iPhone shortcut. Tap the card and it triggers some action, response or workflow.
Says people use it for playlists and lights too. I don't really buy his examples though. You either have to have an NFC in a fixed location (the 'lights' example I believe therefore), or on the move you'd have to dig out the 'right' NFC from someplace (your already full wallet?) then tap it to the phone. That creates actually _more _ friction. 'I stuck a tag on my desk for ....' something specific like he suggests (a list of articles on AI from past 24h) leads to a range of tags on your desk, like when Amazon suggested you have a bunch of tags, one for each product, to build your shopping list. Didn't happen.
Article by [[Stephen Downes p]] evolved from a blogpost. On ethical AI principles in Zotero
Posits its not really possible to declare a list of ethical principles to judge AI by for edu. I wonder if all those principles (9) discussed are internal to AI use, or also its production?
https://web.archive.org/web/20251230193244/https://www.docker.com/blog/private-mcp-catalogs-oci-composable-enterprise-ai/ via [[Lee Bryant p]] this article focuses on 'playlist' style remixing of MCP servers for the enterprise. In light of the disc within Digitale Fitheid, around [[Eindelijk weet ik wat ThetaOS is een Life Lens System (LLS)]] etc., I'm more interested in shareware style distribution of MCP servers p2p and in/between communities
some practices that can make those discussions easier, by starting with constraints that even skeptical developers can see the value in: Build tools around verbs, not nouns. Create checkEligibility() or getRecentTickets() instead of getCustomer(). Verbs force you to think about specific actions and naturally limit scope.Talk about minimizing data needs. Before anyone creates an MCP tool, have a discussion about what the smallest piece of data they need to provide for the AI to do its job is and what experiments they can run to figure out what the AI truly needs.Break reads apart from reasoning. Separate data fetching from decision-making when you design your MCP tools. A simple findCustomerId() tool that returns just an ID uses minimal tokens—and might not even need to be an MCP tool at all, if a simple API call will do. Then getCustomerDetailsForRefund(id) pulls only the specific fields needed for that decision. This pattern keeps context focused and makes it obvious when someone’s trying to fetch everything.Dashboard the waste. The best argument against data hoarding is showing the waste. Track the ratio of tokens fetched versus tokens used and display them in an “information radiator” style dashboard that everyone can see. When a tool pulls 5,000 tokens but the AI only references 200 in its answer, everyone can see the problem. Once developers see they’re paying for tokens they never use, they get very interested in fixing it.
some useful tips to keep MCPs straightforward and prevent data blobs that are too big. - use verbs not nouns for mcp tool names (focuses on the action, not the object upon you act) - think/talk about n:: data minimalisation - break it up, reads separate from reasoning steps. Keeps everything focused on the specific context. - dashboard the ratio of tokens fetched versus tokens used in answers. Lopsided ratios indicate you're overfeeding the system.
In an extreme case of data hoarding infecting an entire company, you might discover that every team in your organization is building their own blob. Support has one version of customer data, sales has another, product has a third. The same customer looks completely different depending on which AI assistant you ask. New teams come along, see what appears to be working, and copy the pattern. Now you’ve got data hoarding as organizational culture.
MCP data hoarding leads to parallel data households, exactly the type of thing we spent a lot of energy on to reduce
data hoarding trap find themselves violating the principle of least privilege: Applications should have access to the data they need, but no more
n:: Principle of least privilege: applications only should have access to data they need, and never more. Data hoarding in MCPs goes beyond that.
There’s also a security dimension to data hoarding that teams often miss. Every piece of data you expose through an MCP tool is a potential vulnerability. If an attacker finds an unprotected endpoint, they can pull everything that tool provides. If you’re hoarding data, that’s your entire customer database instead of just the three fields actually needed for the task.
MCPs that are overloaded w data are new attack surfaces
MCP can remove the friction that comes from those trade-offs by letting us avoid having to make those decisions at all.
MCP is meant to abstract the way access is created to resources. In practice it gets used to abstract away any decision on which data to provide or not. That's the trap.
The team ended up with a data architecture that buried the signal in noise. That additional load put stress on the AI to dig out that signal, leading to serious potential long-term problems. But they didn’t realize it yet, because the AI kept producing reasonable-looking answers. As they added more data sources over the following weeks, the AI started taking longer to respond. Hallucinations crept in that they couldn’t track down to any specific data source. What had been a really valuable tool became a bear to maintain.
Having a clear data architecture for your use case is needed. Vgl [[Eindelijk weet ik wat ThetaOS is een Life Lens System (LLS)]] wrt number of data tables (152 now I think), and how it grew over time, deciding on each table added.
I’ve been watching teams adopt MCP over the past year, and I’m seeing a disturbing pattern. Developers are using MCP to quickly connect their AI assistants to every data source they can find—customer databases, support tickets, internal APIs, document stores—and dumping it all into the AI’s context.
Dev Andrew Stallman warns against dumping all-the-data into an AI application through MCP. Calls it hoarding.
Didn't realise that in 2022 a follow-up to [[ A Psalm for the Wild-Built by Becky Chambers]] was published: [[A Prayer for the Crown-Shy by Becky Chambers]] , for the [[Aan te schaffen boeken]] list
https://web.archive.org/web/20251230190100/https://euobserver.com/digital/ar68689659
EU Observer on the EC's hype feeding AI steps of late.
Configure the Extension Install this extension from the VS Code marketplace Open VS Code Settings (Cmd/Ctrl + ,) Search for "Obsidian MCP"
Obsidian MCP not found in VS Code market.
The real power of MCP emerges when multiple servers work together, combining their specialized capabilities through a unified interface.
Combining multiple MCP servers creates a more capable set-up.
Prompts are structured templates that define expected inputs and interaction patterns. They are user-controlled, requiring explicit invocation rather than automatic triggering. Prompts can be context-aware, referencing available resources and tools to create comprehensive workflows. Similar to resources, prompts support parameter completion to help users discover valid argument values.
prompts are user invoked (hey AgentX, go do..) and may contain next to instructions also references and tools. So a prompt may be a full workflow.
Prompts Prompts provide reusable templates. They allow MCP server authors to provide parameterized prompts for a domain, or showcase how to best use the MCP server.
mcp prompts are templates for interaction
Resources support two discovery patterns: Direct Resources - fixed URIs that point to specific data. Example: calendar://events/2024 - returns calendar availability for 2024 Resource Templates - dynamic URIs with parameters for flexible queries. Example: travel://activities/{city}/{category} - returns activities by city and category travel://activities/barcelona/museums - returns all museums in Barcelona Resource Templates include metadata such as title, description, and expected MIME type, making them discoverable and self-documenting.
Resources can be invoked w fixed and dynamic URIs
Resources expose data from files, APIs, databases, or any other source that an AI needs to understand context. Applications can access this information directly and decide how to use it - whether that’s selecting relevant portions, searching with embeddings, or passing it all to the model.
resources are just that, read only material to invoke. API, filesystem, databases etc.
Each tool performs a single operation with clearly defined inputs and outputs. Tools may require user consent prior to execution, helping to ensure users maintain control over actions taken by a model.
Almost function call like.
Tools are model-controlled, meaning AI models can discover and invoke them automatically. However, MCP emphasizes human oversight through several mechanisms. For trust and safety, applications can implement user control through various mechanisms, such as: Displaying available tools in the UI, enabling users to define whether a tool should be made available in specific interactions Approval dialogs for individual tool executions Permission settings for pre-approving certain safe operations Activity logs that show all tool executions with their results
Tools are available to models, but human in the loop options exist: approval, permission settings, logs
Servers provide functionality through three building blocks:
n:: MCP servers typically provide three types of building blocks, a) Tools that an LLM can call, b) resources that are read-only resources to an LLM, c) prompts, prewritten instructions templates, i.e. agent descriptions, that outline specific tools and resources to use. So for agentic stuff you'd have an MCP server providing templates which in turn list tools and resources.
Visual Studio Code acts as an MCP host. When Visual Studio Code establishes a connection to an MCP server, such as the Sentry MCP server, the Visual Studio Code runtime instantiates an MCP client object that maintains the connection to the Sentry MCP server.
VS Code acts as MCP Host (in their AI toolkit extension I think). You could connect it to the Obsidian MCP server plugin then?
The key participants in the MCP architecture are: MCP Host: The AI application that coordinates and manages one or multiple MCP clients MCP Client: A component that maintains a connection to an MCP server and obtains context from an MCP server for the MCP host to use MCP Server: A program that provides context to MCP clients
The MCP architecture has 3 pieces The host (application, AI or not, that coords the interaction with MCP clients), an MCP client that interacts with a single server. MCP server, which provides the context, i.e. abstracts the access to other sources (filesystem, database, API etc). A server can have one or multiple clients it serves.
MCP is an OS protocol to connect AI applications to external systems.
MCP plugin for Obsidian, that works with Claude Code
I use obsidian vault and also obsidian MCP server from chat client.With vscode I could use MCP to get content into the vault more easily, but refactoring notes, obsidian is better ux
MCP in Obsidian?
A comparison between VS Code and Obsidian. Doesn't state the obvious: any text editor can do this. The tools are just viewers and do not contain the data, which is part of your filesystem. Vgl [[3 Distributed Eigenschappen 20180703150724]]
Ignác Semmelweis in 1847 argued for hand washing in maternity wards by doctors, and published a book about it. Was ridiculed for it and died 1865 as an outcast in an asylum. Only the later emergence of germ theory provided a theoretical basis for the empirical observations of Semmelweis. 'Semmelweis-moment' where someone who is right is laughed out of the room.
FOSDEM schedule 2026
FOSdem has a junior program w coderdojo ao for 7-17yr olds
To install or access apps on Akaunting, you need your API key. Here’s how to get the API key.
The API key is needed to get different apps (chunks of functionality), so that's the answer.
Getting the API Key Estimated reading: 1 minute The API Key holds details of your Akaunting plan.
Ah, this is how they enforce limits. You must have an account to get an API key. Even if you don't interact with them?
All installation instructions, straightforward. Although I don't get why then the later refs to dependency issues, if it's as basic as this
While installing Akaunting on your MacBook (Apple Silicon), you may experience an NPM Install Error. Here’s a reference on how to resolve the error.
Akaunting on my Macbook may throw an npm install error.
solution to install error wrt npm on Apple silicon w akaunting. Not sure if this will arise.
The reqs are basic. Do check if my MAMP has all the mentioned extensions. They assume a hosting package, but for my usage the local webserver is fine.
You can run a local version on-prem, which will be limited to a single company, and 1 user. Under 1k invoices / yr. How would they limit that? Phoning home? As you can purchase extensions.
Akaunting Technology Inc.
Here it mentions a different company name, also not present in opencorporates or companies house.
owned by the Akaunting Software Inc.
Another business not being clear where they are based. Akaunting as a term seems Bulgarian in spelling's origin, the lead dev has a Turkish personal domain, the company on #socmed lists its location as London. The company is not registered at Companies House though (2 others with similar name are, but different), and not known in opencorporates.com.
bookkeeping foss. Not immediately clear if self-hosting possible.
Member states holding the presidency work together closely in groups of three, called 'trios'. This system was introduced by the Lisbon Treaty in 2009. The trio sets long-term goals and prepares a common agenda determining the topics and major issues that will be addressed by the Council over an 18-month period. On the basis of this programme, each of the three countries prepares its own more detailed six-month programme.
Three MS always work together over their 18 months. This provides continuity for long-term goals. This is arranged in the Lisbon Treaty (2009).
The rotating presidency of the Council of the EU (i.e. the MS)
Page holds the basic documents by the current presidency (focus, etc)
Official schedule for Council EU presidency to 2030:
2026 Cyprus Ireland (trio IRL, LT, GR) 2027 Lithuania (trio IRL, LT, GR) Greece (trio IRL, LT, GR) 2028 Italy (trio I, LV, L) Latvia (trio I, LV, L) 2029 Luxembourg (trio I, LV, L) Netherlands (trio NL, SK, M) 2030 Slovakia (trio NL, SK, M) Malta (trio NL, SK, M)
NL in H2 2029 (and a trio looking 1 yr further ahead)
For us personally, this means that we no longer use generative AI – neither for private nor professional purposes.
Authors avoid the use of generative AI. But realise that is difficult for most to do, and as such a privileged tech capable position
To Gen or Not To Gen: The Ethical Use of Generative AI 33 minute read This blog entry started out as a translation of an article that my colleague Jakob and I wrote for a German magazine. After that we added more stuff and enriched it by additional references and sources. We aim at giving an overview about many - but not all - aspects that we learned about GenAI and that we consider relevant for an informed ethical opinion. As for the depth of information, we are just scratching the surface; hopefully, the loads of references can lead you to diving in deeper wherever you want. Since we are both software developers our views are biased and distorted. Keep also in mind that any writing about a “hot” topic like this is nothing but a snapshot of what we think to know today. By the time you read it the authors’ knowledge and opinions have already changed. Last Update: December 8, 2025. Table of ContentsPermalink Abstract About us Johannes Link Jakob Schnell Introduction Ethics, what does that even mean? Clarification of terms Basics Can LLMs think? What LLMs are good at GenAI as a knowledge source GenAI in software development Actual vs. promised benefits Harmful aspects of GenAI GenAI is an ecological disaster Power Water Electronic Waste GenAI threatens education and science GenAI is destroying the free internet. GenAI is a danger to democracy GenAI versus human creativity Digital colonialism Political aspects Conclusion Can there be ethical GenAI? How to act ethically AbstractPermalink ChatGPT, Gemini, Copilot. The number of generative AI applications (GenAI) and models is growing every day. In the field of software development in particular, code generation, coding assistants and vibe coding are on everyone’s lips. Like any technology, GenAI has two sides. The great promises are offset by numerous disadvantages: immense energy consumption, mountains of electronic waste, the proliferation of misinformation on the internet and the dubious handling of intellectual property are just a few of the many negative aspects. Ethically responsible behaviour requires us to look at all the advantages, disadvantages and collateral damages of a technology before we use it or recommend its use to others. In this article, we examine both sides and eventually arrive at our personal and naturally subjective answer to whether and how GenAI can be used in an ethical manner. About usPermalink Johannes LinkPermalink … has been programming for over 40 years, 30 of them professionally. Since the end of the last century, extreme programming and other human-centred software development approaches have been at the heart of his work. The meaningful and ethical implementation of his private and professional life has been his driving force for years. He has been involved with GenAI since the early days of OpenAI’s GPT language models. More about Johannes can be found at https://johanneslink.net. Jakob SchnellPermalink … studied mathematics and computer science and has been working as a software developer for 5 years. He works as a lecturer and course director in university and non-university settings. As a youth leader, he also comes into regular contact with the lives of children and young people. In all these environments, he observes the growing use of GenAI and its impact on people. IntroductionPermalink Ethics, what does that even mean?Permalink Ethical behaviour sounds like the title of a boring university seminar. However, if you look at the wikipedia article of the term 1, you will find that ‘how individuals behave when confronted with ethical dilemmas’ is at the heart of the definition. So it’s about us as humans taking responsibility and weighing up whether and how we do or don’t do certain things based on our values. We have to consider ethical questions in our work because all the technologies we use and promote have an impact on us and on others. Therefore, they are neither neutral nor without alternative. It is about weighing up the advantages and potential against the damage and risks; and that applies to everyone, not just us personally. Because often those who benefit from a development are different from those who suffer the consequences. As individuals and as a society, we have the right to decide whether and how we want to use technologies. Ideally, this should be in a way that benefits us all; but under no circumstances should it be in a way that benefits a small group and harms the majority. The crux of the matter is that ethical behaviour does not come for free. Ethics are neither efficient nor do they enhance your economic profit. That means that by acting according to your values you will, at some point, have to give something up. If you’re not willing to do that, you don’t have values - just opinions. Clarification of termsPermalink When we write ‘generative AI’ (GenAI), we are referring to a very specific subset of the many techniques and approaches that fall under the term ‘artificial intelligence’. Strictly speaking, these are a variety of very different approaches that range from symbolic logic, over automated planning up to the broad field of machine learning (ML). Nowadays most effort, hype and money goes into deep learning (DL): a subfield of ML that uses multi-layered artificial neural networks to discover statistical correlations (aka patterns) based on very large amounts of training data in order to reproduce those patterns later. Large language models (LLM) and related methods for generating images, videos and speech now make it possible to apply this idea to completely unstructured data. While traditional ML methods often managed with a few dozen parameters, these models now work with several trillion (10^12) parameters. In order for this to produce the desired results, both the amount of training data and the training duration must be increased by several orders of magnitude. This brings us to the definition of what we mean by ‘GenAI’ in this article: Hyperscaled models that can only be developed, trained and deployed by a handful of companies in the world. These are primarily the GenAI services provided by OpenAI, Anthropic, Google and Microsoft, or based on these services. We also focus primarily on language models; the generation of images, videos, speech and music plays only a minor role in this article. Our focus on hyperscale services does not mean that other ML methods are free of ethical problems; however, we are dealing with a completely different order of magnitude of damage and risk here. For example, there do exist variations of GenAI that use the same or similar techniques, but on a much smaller scale and restricted domains (e.g. AlphaFold 2). These approaches tend to bring more value with fewer downsides. BasicsPermalink GenAI models are designed to interpolate and extrapolate 3, i.e. to fill in the gaps between training data and speculate beyond the limits of the training data. Together with the stochastic nature of the training data, this results in some interesting properties: GenAI models ‘invent’ answers; with LLMs, we like to refer to this as ‘hallucinations’. GenAI models do not know what is true or false, good or bad, efficient or effective, only what is statistically probable or improbable in relation to training data, context and query (aka prompt). GenAI models cannot explain their output; they have no capability of introspection. What is sold as introspection is just more output, with the previous output re-injected. GenAI models do not learn from you; they only draw from their training material. The learning experience is faked by reinjecting prior input into a conversation’s context 4. The context, i.e. the set of input parameters provided, is decisive for the accuracy of the generated result, but can also steer the model in the wrong direction. Increasing the context window makes a query much more computation-intensive - likely in a quadratic way. Therefore, the promised increase of “maximum context window” of many models is mostly fake 5. The reliability of LLMs cannot be fundamentally increased by even greater scaling 6. Can LLMs think?Permalink Proponents of the language-of-thought hypothesis 7 believe it is possible for purely language-based models to acquire the capabilities of the human brain – reasoning, modelling, abstraction and much more. Some enthusiasts even claim that today’s models have already acquired this capability. However, recent studies 8 9 show that today’s models are neither capable of genuine reasoning nor do they build internal models of the world. Moreover, “…according to current neuroscience, human thinking is largely independent of human language 10” and there is fundamental scientific doubt that achieving human cognition through computation is achievable in practice let alone by scaling up training of deep networks 11. An example of a lack of understanding of the world is the prompt ‘Give me a random number between 0 and 50’. The typical GenAI response to this is ‘27’, and it is significantly more reliable than true randomness would allow. (If you don’t believe it, just try it out!) This is because 27 is the most likely answer in the GenAI training data – and not because the model understands what ‘random’ means. ‘Chain of Thought (CoT)’ approaches and ‘Reasoning models’ attempt to improve reasoning by breaking down a prompt, the query to the model, into individual (logical) steps and then delegating these individual steps back to the LLM. This allows some well-known reasoning benchmarks to be met, but it also multiplies the necessary computational effort by a factor between 30 and 700 12. In addition, multistep reasoning lets individual errors chain together to form large errors. And yet, CoT models do not seem to possess any real reasoning abilities 13 14 and improve the overall accuracy of LLMs only marginally 15. The following thought experiment from 16 underscores the lack of real “thinking” capabilities: LLMs have simultaneous access to significantly more knowledge than humans. Together with the postulated ability of LLMs to think logically and draw conclusions, new insights should just fall from the sky. But they don’t. Getting new insights from LLMs would require these to be already encoded in the existing training material, and to be decoded and extracted by pure statistical means. What LLMs are good atPermalink Undoubtedly, LLMs represent a major qualitative advance when it comes to extracting information from texts, generating texts in natural and artificial languages, and machine translation. But even here, the error rate, and above all the type of error (‘hallucinations’), is so high that autonomous, unsupervised use in serious applications must be considered highly negligent. GenAI as a knowledge sourcePermalink As we have pointed out above, LLMs cannot differentiate between true and false - regardless of the training material. It does not answer the question “What is XYZ?” but the question “How would an answer to question ‘What is XYZ?’ look like?”. Nevertheless, many people claim that the answers that ChatGPT and alike provide for the typical what-how-when-who queries are good enough and often better than what a “normal” web search would have given us. Arguably, this is the most prevalent use case for “AI” bots today. The problem is that most of the time we will never learn about the inaccuracies, left-outs, distortions and biases that the answer contained - unless we re-check everything, which defies the whole purpose of speeding up knowledge retrieval. The less we already know, the better the “AI’s” answer looks to us, but the less equipped we are to spot the problems. A recent by the BBC and 22 Public Service Media organizations shows that 45% of all “AI” assistants’ answers on questions about news and current affairs have significant errors 17. Moreover, LLMs are easy prey for manipulation - either by the service providing organization or by third parties. A recent study claims that even multi-billion-parameter models can be “poisoned” by injecting just a few corrupted documents 18. So, if anything is at stake all output from LLMs must be carefully validated. Doing that, however, would contradict the whole point of using “AI” to speed up knowledge acquisition. GenAI in software developmentPermalink The creation and modification of computer programmes is considered a prime domain for the use of LLMs. This is partly because programming languages have less linguistic variance and ambiguity than natural languages. Moreover, there are many methods for automatically checking generated source code, such as compiling, static code analysis and automated testing. This simplifies the validation of generated code and thereby gives an additional feeling of trust. Nevertheless, individual reports on the success of coding assistants such as Copilot, Cursor, etc. vary greatly. They range from ‘completely replacing me as a developer’ to ‘significantly hindering my work’. Some argue that coding agents considerably reduce the time they have to invest in “boilerplate” work, like writing tests, creating data transfer objects or connecting your domain code to external libraries. Others counter by pointing out that delegating these drudgeries to GenAI makes you miss opportunities to get rid of them, e.g. by introducing a new abstraction or automating parts of your pipeline, and to learn about the intricacies and failure modes of the external library. Other than old-school code generation or code libraries prompting a coding agent is not “just another layer of abstraction”. It misses out on several crucial aspects of a useful abstraction: Its output is not deterministic. You cannot rely on any agent producing the same code next time you feed it the same prompt. The agent does not hide the implementation details, nor does it allow you to reliably change those details if the previous implementation turns out to be inadequate. Code that is output by an LLM, even if it is generated “for free”, has to be considered and maintained each time you touch the related logic or feature. The agent does not tell you if the amount of details you give in your prompt is sufficient for figuring out an adequate implementation. On the contrary, the LLM will always fill the specification holes with some statistically derived assumptions. Sadly, serious studies on the actual benefits of GenAI in software development are rare. The randomised trial by Metr 19 provides an initial indication, measuring a decline in development speed for experienced developers. An informal study by ThoughtWorks estimates the potential productivity gain from using GenAI in software development at around 5-15% 20. If “AI coding” were increasing programmers’ productivity by any big number, we would see a measurable growth of new software in app stores and OSS repositories. But we don’t, the numbers are flat at best 2122. But even if we assume a productivity increase in coding through GenAI, there are still two points that further diminish this postulated efficiency gain: Firstly, the results of the generation must still be cross-checked by human developers. However, it is well known that humans are poor checkers and lose both attention and enjoyment in the process. Secondly, software development is only to a small extent about writing and changing code. The most important part is discovering solutions and learning about the use of these solutions in their context. Peter Naur calls this ‘programming as theory building’ 23. Even the perfect coding assistant can therefore only take over the coding part of software development. For the essential rest, we still need humans. If we now also consider the finding that using AI can relatively quickly lead to a loss of problem-solving skills 24 or that these skills are not acquired at all, then the overall benefit of using GenAI in professional software development is more than questionable. As long as programming - and every technicality that comes with it - will not be fully replaced by some kind of AI, we will still need expert developers who can programm, maintain and debug code to the finest level of detail. Where, we wonder, will those senior developers come from when companies replace their junior staff with coding agents? Actual vs. promised benefitsPermalink If you read testimonials about the use of GenAI that people perceive as successful, you will mostly encounter scenarios in which ‘AI’ helps to make tasks that are perceived as boring, unnecessarily time-consuming or actually pointless faster or more pleasant. So it’s mainly about personal convenience and perceived efficiency. Entertainment also plays a major role: the poem for Grandma’s birthday, the funny song for the company anniversary or the humorous image for the presentation are quickly and supposedly inexpensively generated by ‘AI’. However, the promises made by the dominant GenAI companies are quite different: solving the climate crisis, providing the best medical advice for everyone, revolutionising science, ‘democratising’ education and much more. GPT5, for example, is touted by Sam Altman, CEO of OpenAI, as follows: ‘With GPT-5, it’s now like talking to an expert — a legitimate PhD-level expert in any area you need […] they can help you with whatever your goals are.’ 25 However, to date, there is still no actual use case that provides a real qualitative benefit for humanity or at least larger groups. The question ‘What significant problem (for us as a society) does GenAI solve?’ remains unanswered. On the contrary: While machine learning and deep learning methods certainly have useful applications, the most profitable area of application for ‘AI’ at present is the discovery and development of new oil and gas fields 26. Harmful aspects of GenAIPermalink But regardless of how one assesses the benefits of this technology, we must also consider the downsides, because only then can we ultimately make an informed and fair assessment. In fact, the range of negative effects of hyperscaled generative AI that can already be observed is vast. Added to this are numerous risks that have the potential to cause great social harm. Let’s take a look at what we consider to be the biggest threats: GenAI is an ecological disasterPermalink PowerPermalink The data centres required for training and operating large generative models 27 far exceed today’s dimensions in terms of both number and size. The projected data centre energy demand in the USA is predicted to grow from 4.4% of total electricity in 2023 to 22% in 2028 28. In addition, the typical data centre electricity mix is more CO2-intensive than the average mix. There is an estimated raise of ~11 percent for coal generated electricity in the US, as well as tripled emissions of greenhouse gases worldwide by 2030 - compared to the scenario without GenAI technology 29. Just recently Sam Altman from OpenAI blogged some numbers about the energy and water usage of ChatGPT for “the average query” 30. On the one hand, an average is rather meaningless when a distribution is heavily unsymmetric; the numbers for queries with large contexts or “chain of reasoning” computations would be orders of magnitude higher. Thus, the potential efficiency gains from more economical language models are more than offset by the proliferation of use, e.g. through CoT approaches and ‘agent systems’. On the other hand, big tech’s disclosure of energy consumption (e.g. by Google 31) is intentionally selective. Ketan Joshi goes into quite some details why experts think that the AI industry is hiding the full picture 32. Since building new power plants - even coal or gas fuelled ones - takes a lot of time, data center companies are even reviving old jet engines for powering their new hyper-scalers 33. You have to be aware that those engines are not only much more noisy than other power plants but also pump out nitrous oxide, one of the main chemicals responsible for acid rain 34. WaterPermalink Another problem is the immensely high water consumption of these data centres 35. After all, cooling requires clean water in drinking quality in order to not contaminate or clog the cooling pipes and pumps. Already today, new data centre locations are competing with human consumption of drinking water. According to Bloomberg News about two-thirds of data-centers that were built or developed in 2022 are located in areas that are already under “water-stress” 36. In the US alone “AI servers […] could generate an annual water footprint ranging from 731 to 1,125 million m3” 37. It’s not only an American problem, though. In other areas of the world the water-thirsty data centers also compete with the drinking water supply for humans 38. Electronic WastePermalink Another ecological problem is being noticeably exacerbated by ‘AI’: the amount of electronic waste (e-waste) that we ship mainly to “Third World” countries and which is responsible for soil contamination there. Efficient training and querying of very large neural networks requires very large quantities of specialised chips (GPUs). These chips often have to be replaced and disposed of within two years. The typical data center might not last longer than 3 to 5 years before it has to be rebuilt in large parts39. In summary, it can be said that GenAI is at least an accelerator of the ecological catastrophe that threatens the earth. And it is the argument for Google, Amazon and Microsoft to completely abolish their zero CO2 targets 40 and replace them with investments of several hundred billion dollars for new data centers. GenAI threatens education and sciencePermalink People often try to use GenAI in areas where they feel overloaded and overwhelmed: training, studying, nursing, psychotherapeutic care, etc. The fields of application for ‘AI’ are therefore a good indication of socially neglected and underfunded areas. The fact that LLMs are very good at conveying the impression of genuine knowledge and competence makes their use particularly attractive in these areas. A teacher under the simultaneous pressure of lesson preparation, corrections and covering for sick colleagues turns to ChatGPT to quickly create an exercise sheet. A student under pressure to get good grades has their English essay corrected by ‘AI’. The researcher under pressure to publish will ‘save’ research time by reading the AI-generated summary of relevant papers – even if they are completely wrong in terms of content 41. Tech companies like OpenAI and Microsoft play on that situation by offering their ‘AI’ for free or for little money to students and universities. The goal is obvious: Students that get hooked on outsourcing some of their “tedious” task to a service will continue to use - and eventually buy - this service after graduation. What falls by the wayside are problem-solving skills, engagement with complex sources, and the generation of knowledge through understanding and supplementing existing knowledge. Some even argue that AI is destroying critical education and learning itself 42: Students aren’t just learning less; their brains are learning not to learn. The training cycle of schools and universities is fast. Teachers are already reporting that pupils and students have acquired noticeably less competence in recent years, but have instead become dependent on unreliable ‘tools’ 43. The real problem with using GenAI to do assignments is not cheating, but students “are not just undermining their ability to learn, but to someday lead.” 44 GenAI is destroying the free internet.Permalink The fight against bots on the internet is almost as old as the internet itself – and has been quite successful so far. Multifactor authentication, reCaptcha, honeypots and browser fingerprinting are just a few of the tools that help protect against automated abuse. However, GenAI takes this problem to a new level – in two ways. To make ‘the internet’ usable as the main source for training LLMs, AI companies use so-called ‘crawlers’. These essentially behave like DDoS attackers: They send tens of thousands of requests at once, from several hundred IPs in a very short time. Robot.txt files are ignored; instead, the source IP and user agent are obscured 45. These practices have massive disadvantages for providers of genuine content: Costs for additional bandwidth. Lost advertising revenue, as search engines now offer LLM-generated summaries instead of links to the sources. This threatens the existence of remaining independent journalism in particular 46. Misuse of own content for AI-supported competition. If the place where knowledge is generated is separated from the place where it is consumed, and if this makes the performance of generation even more opaque than before, the motivation to continue generating knowledge also declines. For projects such as Wikipedia, this means fewer donors and fewer contributors. Open communities often have no other option but to shut themselves off. Another aspect is the flooding of the internet with generated content that cannot be automatically distinguished from non-generated content. This content overwhelms the maintainers of open source software or portals such as Wikipedia 47. If this content is then also entered by humans – often in the belief that they are doing good – it is no longer possible to take action against the methodology. In the long run, this means that less and less authentic training material will lead to increasingly poor results from the models. Last but not least, autonomously acting agents make the already dire state of internet security much worse 48. Think of handing all your personal data and credentials to a robot that is distributing and using that data across the web, wherever and whenever it deems it necessary for reaching some goal. is controlled by LLMs who are vulnerable to all kinds of prompt injection attacs 49. is controlled by and reporting to companies that do not have your best interest in mind. has no awareness and knowledge about the implication of its actions. is acting on your behalf and thereby making you accountable. GenAI is a danger to democracyPermalink The manipulation of public opinion through social media precedes the arrival of LLMs. However, this technology gives the manipulators much more leverage. By flooding the web with fake news, fake videos and fake everything undemocratic (or just criminal) parties make it harder and harder for any serious media and journalism to get the attention of the public. People no longer have a common factual basis, which is necessary for all social negotiations. If you don’t agree on at least some basic facts, arguing about policies and measures to take is pointless. Without negotiations democracy will be dying; in many parts of the world it already is. GenAI versus human creativityPermalink Art and creativity are also threatened by generative AI. The impact on artists’ incomes of logos, images and illustrations now being easily and quickly created by AI prompts is obvious. A similar effect can also be observed in other areas. Studies show that poems written by LLMs are indistinguishable from those written by humans and that generative AI products are often rated more highly 50. This can be explained by a trend towards the middle and the average, which can also be observed in the music and film scenes film scene: due to its basic function, GenAI cannot create anything fundamentally new, but replicates familiar patterns, which is precisely why it is so well received by the public. Ironically, ‘AI’ draws its ‘creativity’ from the content of those it seeks to replace. Much of this content was used as training material against the will of the rights holders. Whether this constitutes a copyright infringement has not yet been decided; morally, the situation seems clear. The creative community is the first to be seriously threatened by GenAI in its livelihood 51. It’s not a coincidence that a big part of GenAI efforts is targeted at “democratizing art”. This framing is completely upside down. Art has been one of the most democratic activities for a very long time. Everybody can do it; but not everybody wants to do put in the effort, the practicing time and the soul. Real art is not about the product but about the process, which requires real humans. Generating art without the friction is about getting rid of the humans in the loop - and still making money. Digital colonialismPermalink The huge amount of data required by hyperscaled AI approaches makes it impossible to completely curate the learning content. And yet, one would like to avoid the reproduction of racist, inhuman and criminal content. Attempts are being made to get the problem under control by subsequently adapting the models to human preferences and local laws through additional ‘reinforcement learning from human feedback (RLHF)’ 52. The cheap labour for this very costly process can be found in the Global South. There, people are exposed to hours of hate speech, child abuse, domestic violence and other horrific scenarios in their poorly paid jobs in order to filter them out of the training material of large AI companies 53. Many emerge from these activities traumatised. However, it is not only people who are exploited in the less developed regions of the world, but also nature: the poisoning of the soil with chemicals during the extraction of raw materials for digital chips, as well as the contamination caused by our electronic waste and its improper disposal, are collateral damage that we willingly accept and whose long-term consequences are currently extremely difficult to assess. Here, too, the “developed” world profits, whereas the negative aspects are outsourced to the former colonies and other poor regions of the world. Political aspectsPermalink As software developers, we would like to ‘leave politics out of it’ and instead focus entirely on the cool tech. However, this is impossible when the advocates of this technology pursue strong political and ideological goals. In the case of GenAI, we can cleary see that the US corporations behind it (OpenAI, Google, Meta, Microsoft, etc.) have no problem with the current authoritarian – some say fascist – US government 54. In concrete terms, this means, among other things, that the models are explicitly manipulated to be less liberal or simply not to generate any output that could upset the CEO or the president 55. Even more serious is the fact that many of the leading minds behind these corporations and their financiers adhere to beliefs that can be broadly described as digital fascism. These include Peter Thiel, Marc Andreessen, Alex Karp, JD Vance, Elon Musk and many others on “The Authoritarian Stack” 56. Their ideologies, disguised as rational theories, are called longtermism and effective altruism. What they have in common is that they consider democracy and the state to be obsolete models, compassion to be ‘woke’, and that the current problems of humanity are insignificant, as our future lies in the colonisation of space and the merging of humans with artificial superintelligence 57. Do we want to give people who adhere to these ideologies (even) more power, money and influence by using and paying for their products? Do we want to feed their computer systems with our data? Do we really want to expose ourselves and our children to the answers from chatbots which they have manipulated? Not quite as abstruse, but similarly misanthropic, is the imminent displacement of many jobs by AI, as postulated by the same corporations in order to put pressure on employees with this claim. Demanding a large salary? Insisting on your legal rights? Complaining about too much workload? Doubts about the company’s goals? Then we’ll just replace you with cheap and uncomplaining AI! Whichever way you look at it, AI and GenAI are already being used politically. If we go along without resistance, we are endorsing this approach and supporting it with our time, our attention and our money. ConclusionPermalink Ideally, we would like to quantify our assessment by adding up the advantages, adding up the disadvantages and finally checking whether the balance is positive or negative. Unfortunately, in our specific case, neither the benefits nor the harm are easily quantifiable; we must therefore consult our social and personal values. Discussions about GenAI usually revolve purely around its benefits. Often, the capabilities of all ‘AI’ technologies (e.g. protein folding with AlphaFold 2) are lumped together, even though they have little in common with hyperscaling GenAI. However, if we consider the consequences and do not ignore the problems this technology entails – i.e. if we consider both sides in terms of ethics – the assessment changes. Convenience, speed and entertainment are then weighed against numerous damages and risks to the environment, the state and humanity. In this sense, the ethical use and further expansion of GenAI in its current form is not possible. Can there be ethical GenAI?Permalink If the use of GenAI is not ethical today what would have to change, which negative effects of GenAI would have to disappear or at least be greatly reduced in order to tip the balance between benefits and harms in the other direction? The models would have to be trained exclusively with publicly known content whose original creators consent to its use in training AI models. The environmental damage would have to be reduced to such an extent that it does not further fuel the climate crisis. Society would have to get full access to the training and operation of the models in order to rule out manipulation by third parties and restrict their use to beneficial purposes. This would require democratic processes, good regulation and oversight through judges and courts. The misuse and harming of others, e.g., through copyright theft or digital colonialism, would have to be prevented. Is such a change conceivable? Perhaps. Is it likely, given the interest groups and political aspects involved? Probably not
All these factors are achievable I think, or will be soonish. Smaller models, better sourced data sets, niche models, etc. But not with current actors as mentioned at the end.