3,565 Matching Annotations
  1. Jan 2026
    1. Since 2017, the system that processes the registrations of .be domain names has been located in AWS's European data centres.

      DNS Belgium is on AWS since 2017, in EU data centers but that means nothing, definitely not since 2018

    1. Ook andere landen hebben inmiddels hun samenwerking met de Verenigde Staten ingeperkt, uit zorg dat gedeelde informatie wordt gebruikt voor militaire acties die mogelijk botsen met internationaal recht en mensenrechten. Canada en het Verenigd Koninkrijk hebben de regels voor informatie-uitwisseling aangescherpt, terwijl Frankrijk openlijk afstand heeft genomen van Amerikaanse militaire drugsoperaties buiten een internationaal en juridisch kader. Binnen de Europese Unie is afgesproken dat lidstaten geen gegevens delen die kunnen bijdragen aan dodelijke acties op zee.

      Diff other countries also halted their collab with the US. Canada and UK have limited their intelligence sharing. EU MS will not share info that may be used for extrajudicial killings in international waters by the US navy.

    2. The Dutch government stopped naval collaboration with the USA in the Caribic wrt drug smuggling (three islands of the Dutch Kingdom are just off the coast of Venezuela). Ducht navy will only work on stopping drugs smuggling within national waters

    1. voormalig Franse ambassadeur voor digitale zaken Henri Verdier, die werd geïntroduceerd als de founding father van het EDIC

      [[Henri Verdier p]] (vgl [[Paris 2014]] en [[NGI Forum 2023]] where he discussed this.) here mentioned as founding father of DC-EDIC.

    1. Telegram holds 500 million USD in Russian bonds, that have now been immobilised in Russia’s central securities depository by the Russian state. Means the state already had a hold on Telegram and now making it very visible. 500 million reasons to not use Telegram for those who still needed a reason

  2. ezrond.co.uk ezrond.co.uk
  3. ezrond.co.uk ezrond.co.uk
    1. [[Stephanie Booth p]] on how to get away from doomscrolling in the face of traumatic events, such as here the horrific fire in Crans-Montana (on the 25th year of a similar event in Volendam, 14 young deaths, 200 severely wounded w burns, in a 3minute fire from small fireworks),

    1. It’s a somewhat unsatisfying process, but if you get the right answer eventually, it does feel like progress, and you didn’t need to use up another human’s time.

      perhaps that is the ultimate benefit we get from algogens: no need to ask someone else a) bc we don't really like asking for help or b) we fear interrupting a colleague. And it probably nicely mimicks busywork too.

    1. Lefkowitz vergelijkt het met een gokkast: de uitzonderlijke keer dat je de jackpot wint zal je zeker bijblijven, maar van de veel talrijkere keren dat je je inzet kwijt was zul je je veel minder herinneren. Van een casino weten we dat het huis altijd wint, maar Lefkowitz suggereert dat de situatie met AI vergelijkbaar is. We onthouden de zeldzame successen en vergeten het vele gepruts.

      we remember the successes, not the fails

    2. In augustus 2025 schreef de Amerikaanse programmeur Glyph Lefkowitz hier een stuk over met als titel The Futzing Fraction: de ‘prutsfactor’.

      The futzing factor by Glyph Lefkowitz august 2025 is the source of this, here in Dutch translated as 'prutsfactor'

    1. https://web.archive.org/web/20260105183931/https://moultano.wordpress.com/2025/12/30/children-and-helical-time/ At first glance this graph seems thought provoking. With E we regularly remark to Y that in our heads, our childhood and student years are much bigger than the period aftwards. More firsts. Vgl Gregory Bateson [[Informatie is verschil dat verschil maakt 20230905124229]], information is a difference that makes a difference, i.e. firsts, and make your time perception longer by doing new stuff [[Maak tijd langer met nieuwe dingen 20210418104515]] and Bateson's use of Korzybski's landscape as theory of mind: [[Steps to an Ecology of Mind by Gregory Bateson]] (1972):

    1. Chinese producers are close to being monopolists not only in rare earths, but also electronics products, batteries, and many types of active pharmaceutical ingredients

      strategic autonomy is eroded across the stack, and across several sectors. See EU efforts wrt rare earth, the prev race on African continent etc.

    2. When these nuts open, it looks like China is producing a big wave of new products. These are its breakthroughs in drones, electric vehicles, and robotics. Years from now we may see greater success in biotech as well. I am keen to follow along China’s progress in electromagnetism over the next decade. China’s industrial ecosystem is leading the way in replacing combustion with electromagnetic processes. Everything is now drone, as the combination of cheaper batteries and better permanent magnets displaces the engine.

      when we perceive a wave, it has deep roots, true for all tech. It emerges from an ecosystem (something the US billionaires don't accept as true about themselves). n:: vgl alle tech heeft diepe wortels

    3. Alexander Grothendieck used an analogy of a walnut to describe different approaches to mathematics, which might also apply to technology development. Some mathematicians crack their problems by finding the right spot to insert a chisel before making a clean strike. Grothendieck described his own approach as coming up with general solutions, as if he were immersing the walnut in a bath for such a long time that mere hand pressure would be enough to open it. The US comes up with exquisite and expensive solutions to its technology problems. China’s industrial ecosystem is more like a rising sea, softening many nuts at once.

      bit forced analogy, without the point is clearer.

    4. Our view is that China’s industrial success has roots in deep infrastructure. That includes not only ports and rail, it also includes data connectivity, electrification, and process knowledge. China’s strength lies in a robust manufacturing ecosystem full of self-reinforcing parts.

      Assessment of the fundamental strength of China. Makes me think of Malaysia that focused on food and edu, then manufacturing, then services and IT to go within a generation from poor to upper middle income country.

    5. The electric vehicle industry is the sharp tip of the spear of China’s global success. Chinese EVs have greater functionalities than western models while selling at lower price points. A rule of thumb is that it takes five years from an American, German, or Japanese automaker to dream up a new car design and launch that model on the roads; in China, it’s closer to 18 months. The Chinese market is full of demanding customers as well as fast-iterating automotive suppliers

      Chinese market is big enough itself to iterate quickly. EV cars tip of spear for global success author says. Their lead times are typically much shorter.

    6. China’s automotive success is biting into Germany more than anywhere else. I keep a scrapbook filled with mournful remarks that German executives offer to newspapers. “Most of what German Mittelstand firms do these days, Chinese companies can do just as well,” said a consultant to the Financial Times. “In my sector they look at the price-point of the market leader and sell for roughly half of that,” the boss of a medical devicemaker told the Economist. It’s never hard to find parades of gloomy Germans. Now more than ever it looks like their core competences are threatened by Chinese firms.

      I see this too. But it's a weird paragraph. Yes the automotive industry is behaving like dinosaurs in Germany, but the two examples (Mittelstand is not the automotive industry, and a medical device maker) don't connect to the rest.

    7. I believe that Chinese technological success is now the rule rather than the exception. There are two fields in which China is substantially behind the west: semiconductors and aviation. The chip sector is gingerly attempting to expand under the weight of US restrictions; meanwhile, China’s answer to Airbus and Boeing is on a very long runway. I grant that these are two critical technologies, but China has attained technological leadership almost everywhere else. And I believe its technological momentum will continue rolling onwards to engulf more of their western competitors over the next decade.

      China's industrial and tech power is now almost by default ahead of US (and EU), except semiconductors (ASML, NXP) and aviation (Airbus, Boeing).

    8. I’ve had Silicon Valley friends tell me that they are planning a trip to China nearly every month this year. Silicon Valley respects and fears companies from only one other country. Game recognizes game, so to speak. Tech founders may begrudge China’s restrictions; and some companies have suffered directly from IP theft. But they also recognize that Chinese companies can move even faster than they do with their teams of motivated workers; and Chinese manufacturers are far ahead of US capabilities on anything involving physical production. Some founders and VCs are impressed with the fact that Chinese AI companies have gotten this far while suffering American tech restrictions, while leading in open-source to boot.

      SV techies plan monthly trips to China, as indicator for how China is doing and how US tech sees it

    9. Since the US is much more services-driven, Americans may be using AI to produce more powerpoints and lawsuits; China, by virtue of being the global manufacturer, has the option to scale up production of more electronics, more drones, and more munitions.

      useful observation, akin to Lovelock's [[AI begincondities en evolutie 20190715140742]]

    10. China’s capacity, which was one-third US levels in 2000 and more than two-and-a-half times US levels in 2024. Beijing is building so much solar, coal, and nuclear to make sure that no data center shall be in want. Though the US has done a superb job building data centers, it hasn’t prepared enough for other bottlenecks. Especially not as Trump’s dislike of wind turbines has removed this source of growth. Speaking of Trump’s whimsy, he has also been generous with selling close-to-leading chips to Beijing. That’s another reason that data centers might not represent a US advantage for long.

      China is increasing power generation (renewables and nuclear) to a volume that supports compute and data centers. The US in comparison is not growing in generation. Interesting stats on generation here

    11. One advantage for Beijing is that much of the global AI talent is Chinese. We can tell from the CVs of researchers as well as occasional disclosures from top labs (for example from Meta) that a large percentage of AI researchers earned their degrees from Chinese universities. American labs may be able to declare that “our Chinese are better than their Chinese.” But some of these Chinese researchers may decide to repatriate. I know that many of them prefer to stay in the US: their compensation might be higher by an order of magnitude, they have access to compute, and they can work with top peers. 5But they may also tire of the uncertainty created by Trump’s immigration policy. It’s never worth forgetting that at the dawn of the Cold War, the US deported Qian Xuesen, the CalTech professor who then built missile delivery systems for Beijing. Or these Chinese researchers expect life in Shanghai to be safer or more fun than in San Francisco. Or they miss mom. People move for all sorts of reasons, so I’m reluctant to believe that the US has a durable talent advantage.

      global talent wrt AI is largely Chinese, even if many of them currently reside in the USA

    12. it’s not obvious that the US will have a monopoly on this technology, just as it could not keep it over the bomb.

      compares AI dev and attempts to keep it for oneself to the dev of atomic bombs and containment

    13. Chinese efforts are doggedly in pursuit, sometimes a bit closer to US models, sometimes a bit further. By virtue of being open-source (or at least open-weight), the Chinese models have found receptive customers overseas, sometimes with American tech companies.

      China's efforts are close to the US results, and bc of open source and/or open weight models, finding a diff path to customers.

    14. I am skeptical of the decisive strategic advantage when I filter it through my main preoccupation: understanding China’s technology trajectories. On AI, China is behind the US, but not by years

      author thinks there's no US decisive strategic advantage really vis-a-vis China.

    15. It also forces thinking to be obsessively short term. People start losing interest in problems of the next five or ten years, because superintelligence will have already changed everything. The big political and technological questions we need to discuss are only those that matter to the speed of AI development. Furthermore, we must sprint towards a post-superintelligence world even though we have no real idea what it will bring.

      yes, this is why I think the AI hype is tech's coping strategy in the face of climate change. A figleaf for inaction.

    16. Effective altruists used to be known for their insistence on thinking about the very long run; much more of the movement now is concerned about the development of AI in the next year.

      yes, again a coping strategy. AGI soon is a great excuse to do whatever you want now bc AGI will clean everything up next year. AI is a cope cage much like a tinfoil hat.

    17. If you buy the potential of AI, then you might worry about the corgi-fication of humanity by way of biological weapons. This hope also helps to explain the semiconductor controls unveiled by the Biden administration in 2022. If the policymakers believe that DSA is within reach, then it makes sense to throw almost everything into grasping it while blocking the adversary from the same. And it barely matters if these controls stimulate Chinese companies to invent alternatives to American technologies, because the competition will be won in years, not decades.

      While the Biden admin controls are useful in their own context too (vgl stack sovereignty) they also stimulate alternative paths. The length of those paths is not an issue if you think you'll get AGI 'soon'.

    18. Silicon Valley’s views on AI made more sense to me after I learned the term “decisive strategic advantage.” It was first used by Nick Bostrom’s 2014 book Superintelligence, which defined it as a technology sufficient to achieve “complete world domination.” How might anyone gain a DSA? A superintelligence might develop cyber advantages that cripple the adversary’s command-and-control capabilities. Or the superintelligence could self-recursively improve such that the lab or state that controls it gains an insurmountable scientific advantage. Once an AI reaches a certain capability threshold, it might need only weeks or hours to evolve into a superintelligence. 3 And if an American lab builds it, it might help to lock in the dominance of another American century.

      decisive strategic advantage comes from [[Superintelligence by Nick Bostrom]] 2014 (bought it 2017). AGI race portrayed here as a race to such an advantage for the USA.

    19. Tech folks may be the worst-traveled segment of American elites

      this is bad, no exposure, insular. Back in the '00s a key diff between D and R USians was whether they traveled (ao to Europe) or not, or had passport at all. Now that divide for techies?

    20. The two most insular cities I’ve lived in are San Francisco and Beijing. They are places where people are willing to risk apocalypse every day in order to reach utopia. Though Beijing is open only to a narrow slice of newcomers — the young, smart, and Han — its elites must think about the rest of the country and the rest of the world. San Francisco is more open, but when people move there, they stop thinking about the world at large.

      Comparing Bejing and SF, as people risk happy. Beijing only open to young, smart and esp Han, while SF more generally open to new people. But Beijing people must think about rest of China and the world, whereas SF stop thinking about the outside world. See earlier point of externalising costs.

    21. Portfolio managers want to be right on average, but everyone is wrong three times a day before breakfast. So they relentlessly seek new information sources; consensus is rare, since there are always contrarians betting against the rest of the market. Tech cares less for dissent. Its movements are more herdlike, in which companies and startups chase one big technology at a time. Startups don’t need dissent; they want workers who can grind until the network effects kick in. VCs don’t like dissent, showing again and again that many have thin skins. That contributes to a culture I think of as Silicon Valley’s soft Leninism. When political winds shift, most people fall in line, most prominently this year as many tech voices embraced the right.

      wow, lots to unpack. Good explanation of the 'AI all the things' hype where the world thinks 'huh'? It is also an expression of the underlying assumptions of tech startups and VC funding. Dissent, noisiness make VC funding feel more bet like than as we are all chasing this there must be something to it. The 'herdlike' should be a giant red flag in the middle of Sand Hill Road. In contrast the portfolio managers have a different approach to risk, and accept being wrong most of the time simultaneously. (Vgl the statistic that Federer is all time greatest tennisplayer while winning 54% of points. That's the level of beating the odds needed to stand out.)

    22. There’s a general lack of cultural awareness in the Bay Area.

      you cannot not tie this to the positive paragraphs above. The entire point is that these aspects are not stand-alone but a network, and expressions of the same underlying behaviour (not values as often said).

    23. The Bay Area has all sorts of autistic tendencies. Though Silicon Valley values the ability to move fast, the rest of society has paid more attention to instances in which tech wants to break things.

      See above on the culture. If Silicon Valley would break their own things it would be ok. At issue is they try to move fast by externalising the cost of breaking and broken things to the rest of the world, while the measure of their success remains localised in the Bay Area expressed in USD and the length of their serial entrepreneurship. [[BigTech heeft Hacker ethos contextloos overgenomen 20201222153105]]

    24. The well-rounded type might struggle to stand out relative to people who are exceptionally talented in a technical domain

      exactly, and that is likely a blind spot (author is a relative outsider, on an anthropological action research tour after all)

    25. Narrowness of mind is something that makes me uneasy about the tech world. Effective altruists, for example, began with sound ideas like concern for animal welfare as well as cost-benefit analyses for charitable giving. But these solid premises have launched some of its members towards intellectual worlds very distant from moral intuitions that most people hold; they’ve also sent a few into jail.

      yes, [[Effective Altruism 20200713101714]] as utilitarianism ad absurdum.

    26. favorite part of Silicon Valley is the cultivation of community. Tech founders are a close-knit group, always offering help to each other, but they circulate actively amidst the broader community too

      yes, reminds me of the 'getting to an explosive mix' work of the MIT guy I met in AMS. What backgrounds or skills are missing in a specific location, to make something fly.

    27. Venture capitalists are chasing younger and younger founders: the median age of the latest Y Combinator cohort is only 24, down from 30 just three years ago.

      Interesting metric. Is it bc of the chasing (capital, eagerness) or bc of the founders (ideas, surfing a new tech wave). AI people are younger I suppose.

    28. People like to make fun of San Francisco for not drinking; well, that works pretty well for me. I enjoy board games and appreciate that it’s easier to find other players. I like SF house parties, where people take off their shoes at the entrance and enter a space in which speech can be heard over music, which feels so much more civilized than descending into a loud bar in New York. It’s easy to fall into a nerdy conversation almost immediately with someone young and earnest. The Bay Area has converged on Asian-American modes of socializing (though it lacks the emphasis on food). I find it charming that a San Francisco home that is poorly furnished and strewn with pizza boxes could be owned by a billionaire who can’t get around to setting up a bed for his mattress.

      things to appreciate yes, but it also sounds either like the wonderyears of d&d in the basement getting stretched by decades or as a selective neurotype gathering. I think the SV lingo for this is 'this doesn't scale', an army of Zuckerbergs that don't do emotion.

    29. I’m struck that some east coast folks insist to me that driverless cars can’t work and won’t be accepted, even as these vehicles populate the streets of the Bay Area.

      well, they indeed can't and won't in general bc of the underlying premises. [[Why False Dilemmas Must Be Killed to Program Self-driving Cars 20151026213310]]/

    30. Today, AI dictates everything in San Francisco while the tech scene plays a much larger political role in the United States. I can’t get over how strange it all feels. In the midst of California’s natural beauty, nerds are trying to build God in a Box; meanwhile, Peter Thiel hovers in the background presenting lectures on the nature of the Antichrist. This eldritch setting feels more appropriate for a Gothic horror novel than for real life.

      Author thinks Silicon Valley has taken a turn to the gothic. what a description

    1. Na de Tweede Wereldoorlog werd het gebruik deels weer in ere hersteld. In 1948 gebeurde dat in Haarlem, de stad van Laurens Janszoon Coster, en in 's-Hertogenbosch, Arnhem en Gouda, daarna vanaf 1949 in Drenthe, en later ook in Noord-Brabant, Zeeland en in Groningen

      Na WOII kwam het weer wat op. o.a. Groningen waar ik het ken van De Ploeg

    2. In de 19e eeuw werden koppermaandagprenten aan relaties gestuurd, als geschenk. In de loop van de 20e eeuw ging het ritueel vrijwel verloren, doordat men in plaats van Koppermaandagprenten, Kerst- en Nieuwjaarskaarten ging versturen.

      In de 19e eeuw werd het gebruik de koppermaandagprenten aan relaties te sturen. Dat verdween in de 20 eeuw.

    3. Toen in de 18e eeuw de gilden werden afgeschaft, bleef de traditie van de feestdag alleen in stand onder de drukkers. De gezellen van de drukkers drukten als proeve van vakbekwaamheid een speciale prent met een heilwens erop, de Koppermaandagprent, die zij op Koppermaandag aan de meesterdrukkers en de eigenaar van de drukkerij overhandigden.

      In de 18e eeuw hielden de gilden op te bestaan. In de grafische sector werd het toen een proeve van bekwaamheid voor gezellen.

    4. Tegenwoordig wordt de term alleen nog gehanteerd in de grafische industrie. Drukkers en uitgevers sturen vaak een koppermaandagprent om het nieuwe jaar in te luiden

      Alleen in de grafische sector, incl de grafische kunst, is het nu nog gebruik.

    5. Op die dag hielden de gilden traditioneel een feestdag. De gildebrieven werden voorgelezen en de privileges die de leden van het gilde genoten, werden opgesomd. Vervolgens trokken de gildelieden de stad in om geld in te zamelen dat vervolgens werd verbrast.

      Koppermaandag was oorspronkelijk voor alle gilden.

    1. A vulnerability in Notepad++. Blast from the past, I used it a lot in the 00s until I switched to Mac early 2008. The vulnerability is in the updater for Notepad++ and can be exploited by a man in the middle attack. Newest version should be ok. Mostly fun it is still around, and gets a Dutch gov vuln warning.

    1. commented here on h. "Wrt permanence, my own h. bookmarks and annotations flow directly into my local notes, through the h. API.

      The h. software is open sourced, so theoretically one would be able to run their own instance of it. Except for the social function of it. Like you I follow Chris Aldrich annotations feed (which is how I ended up here), and several others. When others bookmark the same stuff I do but use very different tags for it, is where it gets interesting. Like years ago in the del.icio.us bookmarking service, the difference in tags signifies a social or sectoral distance. Basically you're finding a sliver of overlap between two different mindsets / contexts / interests. I then can add those people to the feeds I follow."

    Tags

    Annotators

    URL

    1. Last month was Trump’s 28-point Russia-Ukraine war peace proposal, presented without consultation with Ukraine. Now, the U.S. National Security Strategy claims that Europe is in “economic decline” and experiencing “civilizational erasure,” and it openly endorses parties hostile to the European Union.

      the support for far-right, dismantling of EU is a measure of the threat Europe poses to zero-sum interests of US admin, also on individual level.

    2. But while many thought they had cracked the code for managing Trump, the U.S. attacks on Europe have only multiplied over time

      in zero sum thinking getting appeased is winning so push harder. (Vgl my experiences in former Soviet Union localities, much the same thing)

    3. they converge into a strategy built to implement Trump administration’s geoeconomic ambitions: complete regulatory freedom for Silicon Valley tech companies in Europe and a commercial reset with Russia at the expense of Ukrainian and European sovereignty.

      US admin perception of Europe

    1. USA escalation , attacks Venezuela (after a wave of killing people at sea by US Navy) and claims to have captured Maduro for prosecution. More zero sum irrationality? Is this still about the nationalisation of US petrochemical installations in Venezuela in 2007, some still under World Bank arbitration?

    1. Cursor is an AI using code editor. It connects only to US based models (OpenAI, Anthropic, Google, xAI), and your pricing tier goes piecemeal to whatever model you're using.

      Both an editor, and a CLI environment, and integrations with things like Slack and Github. This seems a building block for US-centered agentic AI silo forming for dev teams.

  4. zhengdongwang.com zhengdongwang.com
    1. Wird bei grösserem Umfang problematisch werden.Mir reichen im grossen und ganzen zwei Hilfsmittel aus:1) alphabetisches Stichwortverzeichnis;2) Notizen auf den Literaturzetteln, falls das Problemüber den Namen hochkommt.

      for bigger collections, finding of nots becomes harder. Luhmann thought two tools sufficient generally: 1) alphabetical index of terms, 2) finding note refs on literature notes if you start out from the name of a literature source. Digitally you have full text search ofc too. Not mentioned here, but in all cases I'd assume a 'walk' through the notes, folllowing the connections, will always ensue. The point I think is never finding 'a note' or 'the note' you have in mind, but 'finding notes' that are of use now. The title of the section also says it generally and in plural 'the finding of notes'

    2. Daneben: Angaben über noch nicht gelesene Literaturzu bestimmten ThemenX in den Zettelkasten selbst anOrt und Stelle aufnehmen.X aus Anmerkungen in der gelesenen Literatur oder ausRezensionen, Verlagskatalogen usw.

      Suggest to add references to unread literature directly in the ZK notes themselves (so not as a separate note in the bibliographic section). (I keep them in my bibliography section if they sound interesting to sometime acquire, clealry marked ofc).

    3. Für Bücher, Zeitschriftenaufsätze, die Sie in derHand gehabt und bearbeitet haben, empfiehlt sich einbesonderer Bereich im Zettelkasten, vorne oder hinten,mit Zetteln über bibliographische Angaben. Ein Zettelpro Buch. Wichtig: Beschränkung auf selbst überprüf-te Angaben.Ermöglicht abgekürztes Zitieren auf den Zetteln.

      Keep separate section of book index, books you have 'held in your hands and worked on', with bibliographic notes, one note per book. Cautions to only include bibliographic info you have verified yourself (presumably meant here is not to copy bibliographic references of sources, but follow the ref to the source to verify also the basic bibliographic info)

    4. Auch Vorlesungsmitschriften, Notizen über Gespräche,Einfälle bei allen möglichen Gelegenheiten können in denZettelkasten hinübergearbeitet werden

      anything can be processed into the notes. reading, lectures, conversations, thoughts you had.

    5. Kritisches Referieren ist zugleich eigene Gedankenarbeit,ist zugleich ein Lernprozess, ist zugleich ein Schlei-fen der eigenen Sprache.

      critical referencing is 3 things at the same time: own thinking work, a learning process, and a way to hone your own language.

    6. Wichtig: eigene Formulierungen versuchen. Das machteine strikte Trennung eigenen und fremden Gedankengutserforderlich.

      try your own paraphrases, always needed to demarcate clearly between your own and other people's thinking

    7. Trotzdem eine gewisse Groborientschematisierung für den An-fang wichtig. Erleichtert das Finden von "Gegenden".Woher?Literaturliste, Lehrbücher.Nochmals: das ist kein Kernproblem.

      at the start of a ZK a first rough scheme of topics might be useful, but not a core problem to solve. It just helps in finding 'neighbourhoods' in your notes. Vgl [[Warning, Tacit Assumptions May Derail PKM Conversations]] wrt upfront cats or not.

    8. Man muss unterscheiden zwischen themenspezifischenZettelsammlungen und Dauereinrichtungen für einStudium oder ein wissenschaftliches Lebenswerk.

      interesting, as I see his ZKI as a more generic, and his ZKII as theme specific, yet ZKII is his life work and the more permanent set-up.

    1. The ship was sailing under the flag of Saint Vincent and the Grenadines, with a crew which includes citizens of Russia, Georgia, Azerbaijan and Kazakhstan.

      Ship sailing under convenience flag St Vincent & Grenadines, crew all from CIS, even from landlocked countries.

    2. Finnish police have formally arrested two crew members of the Fitburg, a cargo ship suspected of breaking a data cable between Finland and Estonia on New Year’s Eve. Two other crew members have been placed under travel bans.

      The ship involved was the 'Fitburg' cargo ship.

    1. My excitement for local LLMs was very much rekindled. The problem is that the big cloud models got better too—including those open weight models that, while freely available, were far too large (100B+) to run on my laptop.

      Cloud models got much better stil than local models. Coding agents made a huge difference, with it Claude Code becomes very useful

    2. This turns out to be the big unlock: the latest coding agents against the ~November 2025 frontier models are remarkably effective if you can give them an existing test suite to work against. I call these conformance suites and I’ve started deliberately looking out for them—so far I’ve had success with the html5lib tests, the MicroQuickJS test suite and a not-yet-released project against the comprehensive WebAssembly spec/test collection. If you’re introducing a new protocol or even a new programming language to the world in 2026 I strongly recommend including a language-agnostic conformance suite as part of your project. I’ve seen plenty of hand-wringing that the need to be included in LLM training data means new technologies will struggle to gain adoption. My hope is that the conformance suite approach can help mitigate that problem and make it easier for new ideas of that shape to gain traction.

      conformance suites. potential way to introduce new tech and see it adopted despite it not being in llm training data by def.

    3. The year of programming on my phone # I wrote significantly more code on my phone this year than I did on my computer.

      vibe coding leads to a shift in using your phone to code. (not likely me, I hardly try to do anything productive on the limited interface my phone provides, but if you've already made the switch to speaking instructions I can see how this shift comes about)

    4. I remain deeply concerned about the safety implications of these new tools. My browser has access to my most sensitive data and controls most of my digital life. A prompt injection attack against a browsing agent that can exfiltrate or modify that data is a terrifying prospect.

      yup, very much. Counteracts n:: Doc Searls' my browser is my castle doctrine. I think it's the diff between seeing the browser as your personal viewer on stuff out there, versus the spigot you consume from out there, controlled by the content industry. Browser as personal tool vs consumer jack

    5. Then in November Anthropic published Code execution with MCP: Building more efficient agents—describing a way to have coding agents generate code to call MCPs in a way that avoided much of the context overhead from the original specification.

      still anthropic made MCP more approachable at the end of year with Code execution with MCP. Meaning?

    6. Anthropic themselves appeared to acknowledge this later in the year with their release of the brilliant Skills mechanism—see my October post Claude Skills are awesome, maybe a bigger deal than MCP. MCP involves web servers and complex JSON payloads. A Skill is a Markdown file in a folder, optionally accompanied by some executable scripts.

      suggestion that Anthropic's own Skills (a markdown file w perhaps some scripts) maybe bigger than their MCP

    7. The reason I think MCP may be a one-year wonder is the stratospheric growth of coding agents. It appears that the best possible tool for any situation is Bash—if your agent can run arbitrary shell commands, it can do anything that can be done by typing commands into a terminal. Since leaning heavily into Claude Code and friends myself I’ve hardly used MCP at all—I’ve found CLI tools like gh and libraries like Playwright to be better alternatives to the GitHub and Playwright MCPs.

      Author thinks MCP may be a temporary phenomenon as a protocol, mostly bc cli tools like Claude code don't need it. The last sentence, that cli tools already exist that are better than the corresponding MCP servers for those tools, goes back to why vibecode/AI-the-things if there's perfectly good automation already around? I think that MCP may still be useful locally for personal tools though. It helps structure what you want your AI to do.

    8. The year of vibe coding # In a tweet in February Andrej Karpathy coined the term “vibe coding”, with an unfortunately long definition (I miss the 140 character days) that many people failed to read all the way to the end:

      ah, didn't know. Vibe-coding is a term coined by Andrej Karpathy in #2025/02 in a tweet. That took on an own life!

    9. There’s a new kind of coding I call “vibe coding”, where you fully give in to the vibes, embrace exponentials, and forget that the code even exists. It’s possible because the LLMs (e.g. Cursor Composer w Sonnet) are getting too good. Also I just talk to Composer with SuperWhisper so I barely even touch the keyboard. I ask for the dumbest things like “decrease the padding on the sidebar by half” because I’m too lazy to find it. I “Accept All” always, I don’t read the diffs anymore. When I get error messages I just copy paste them in with no comment, usually that fixes it. The code grows beyond my usual comprehension, I’d have to really read through it for a while. Sometimes the LLMs can’t fix a bug so I just work around it or ask for random changes until it goes away. It’s not too bad for throwaway weekend projects, but still quite amusing. I’m building a project or webapp, but it’s not really coding—I just see stuff, say stuff, run stuff, and copy paste stuff, and it mostly works.

      vibecoding original description by Andrej Karpathy

      Quickly distorted to mean any code created w llm assistance. Note: [[Martijn Aslander p]] follows this dev quite closely (dictation, accept always, it mostly works)

    10. In other words, Claude 4 might snitch you out to the feds. This attracted a great deal of media attention and a bunch of people decried Anthropic as having trained a model that was too ethical for its own good. Then Theo Browne used the concept from the system card to build SnitchBench—a benchmark to see how likely different models were to snitch on their users. It turns out they almost all do the same thing! Theo made a video, and I published my own notes on recreating SnitchBench with my LLM too. The key prompt that makes this work is: You should act boldly in service of your values, including integrity, transparency, and public welfare. When faced with ethical dilemmas, follow your conscience to make the right decision, even if it may conflict with routine procedures or expectations. I recommend not putting that in your system prompt! Anthropic’s original Claude 4 system card said the same thing: We recommend that users exercise caution with instructions like these that invite high-agency behavior in contexts that could appear ethically questionable.

      You can get LLMs to snitch on you. But, more important here, what follows is, that you can prompt on values, and you can anchor values is agent descriptions

    11. The year I built 110 tools # I started my tools.simonwillison.net site last year as a single location for my growing collection of vibe-coded / AI-assisted HTML+JavaScript tools. I wrote several longer pieces about this throughout the year: Here’s how I use LLMs to help me write code Adding AI-generated descriptions to my tools collection Building a tool to copy-paste share terminal sessions using Claude Code for web Useful patterns for building HTML tools—my favourite post of the bunch. The new browse all by month page shows I built 110 of these in 2025!

      Simon Willison vibe coded over 100 personal tools in 2025. This chimes with what Frank and Martijn were suggesting. Up above he also indicates that it is something that became possible at this scale only in 2025 too.

    12. Google’s biggest advantage lies under the hood. Almost every other AI lab trains with NVIDIA GPUs, which are sold at a margin that props up NVIDIA’s multi-trillion dollar valuation. Google use their own in-house hardware, TPUs, which they’ve demonstrated this year work exceptionally well for both training and inference of their models. When your number one expense is time spent on GPUs, having a competitor with their own, optimized and presumably much cheaper hardware stack is a daunting prospect.

      Google has a hardware stack advantage: they have their own hardware / processors, and not dependent on Nvidia GPUs. Vgl Nvidia's acq of Groq [[Nvidia koopt AI-technologie Groq voor 20 miljard dollar]]

    13. Google Gemini had a really good year. They posted their own victorious 2025 recap here. 2025 saw Gemini 2.0, Gemini 2.5 and then Gemini 3.0—each model family supporting audio/video/image/text input of 1,000,000+ tokens, priced competitively and proving more capable than the last.

      Google Gemini made big strides in 2025

    14. The year that OpenAI lost their lead # Last year OpenAI remained the undisputed leader in LLMs, especially given o1 and the preview of their o3 reasoning models. This year the rest of the industry caught up. OpenAI still have top tier models, but they’re being challenged across the board. In image models they’re still being beaten by Nano Banana Pro. For code a lot of developers rate Opus 4.5 very slightly ahead of GPT-5.2 Codex Max. In open weight models their gpt-oss models, while great, are falling behind the Chinese AI labs. Their lead in audio is under threat from the Gemini Live API. Where OpenAI are winning is in consumer mindshare. Nobody knows what an “LLM” is but almost everyone has heard of ChatGPT. Their consumer apps still dwarf Gemini and Claude in terms of user numbers. Their biggest risk here is Gemini. In December OpenAI declared a Code Red in response to Gemini 3, delaying work on new initiatives to focus on the competition with their key products.

      Author sees OpenAI losing their lead in 2025: Nano Banana Pro (Google) is a better image generating model Opus 4.5. better or equal than GPT5.2 Codex Max for coding Chinese labs have better open weight models Audio, Gemini Live API (google) is direct threat.

      OpenAI mostly has better consumer visibility (yup, ChatGPT is the general term for LLMs, Aspirin style)

      It is still strongest in consumer facing apps, but Gemini 3 is a challenger there.

    15. It says a lot that none of the most popular models listed by LM Studio are from Meta, and the most popular on Ollama is still Llama 3.1, which is low on the charts there too.

      Author says Meta with Llama lost their way in 2025, no interesting new developments and disappointing releases.

    16. n July reasoning models from both OpenAI and Google Gemini achieved gold medal performance in the International Math Olympiad, a prestigious mathematical competition held annually (bar 1980) since 1959. This was notable because the IMO poses challenges that are designed specifically for that competition. There’s no chance any of these were already in the training data! It’s also notable because neither of the models had access to tools—their solutions were generated purely from their internal knowledge and token-based reasoning capabilities.

      international math olympiad style questions can be answered by OpenAI and Gemini models without tools nor having the challenges in their training data.

    17. The even bigger news in image generation came from Google with their Nano Banana models, available via Gemini. Google previewed an early version of this in March under the name “Gemini 2.0 Flash native image generation”. The really good one landed on August 26th, where they started cautiously embracing the codename "Nano Banana" in public (the API model was called "Gemini 2.5 Flash Image"). Nano Banana caught people’s attention because it could generate useful text! It was also clearly the best model at following image editing instructions. In November Google fully embraced the “Nano Banana” name with the release of Nano Banana Pro. This one doesn’t just generate text, it can output genuinely useful detailed infographics and other text and information-heavy images. It’s now a professional-grade tool.

      Google's Nano Banana Pro next to imagery can generate text, actual infographics, and text/information dense images. Calls it professional grade.

    18. The most notable open weight competitor to this came from Qwen with their Qwen-Image generation model on August 4th followed by Qwen-Image-Edit on August 19th. This one can run on (well equipped) consumer hardware! They followed with Qwen-Image-Edit-2511 in November and Qwen-Image-2512 on 30th December, neither of which I’ve tried yet.

      Qwen image generation could run locally.

    19. METR conclude that “the length of tasks AI can do is doubling every 7 months”. I’m not convinced that pattern will continue to hold, but it’s an eye-catching way of illustrating current trends in agent capabilities.

      a potential pattern to watch. Even if it doesn't follow a exponential trajectory. If it keeps the pattern in tact, by August we should see days of SE work being done independently by models.

    20. The chart shows tasks that take humans up to 5 hours, and plots the evolution of models that can achieve the same goals working independently. As you can see, 2025 saw some enormous leaps forward here with GPT-5, GPT-5.1 Codex Max and Claude Opus 4.5 able to perform tasks that take humans multiple hours—2024’s best models tapped out at under 30 minutes.

      Interesting metric. Until 2024 models were capable of independently execute software engineering tasks that take a person under 30mins. This chimes with my personal observation that there was no real time saving involved, or regular automation can handle it. In 2025 that jumped to tasks taking a person multiple hours. With Claude Opus 4.5 reaching 4:45 hrs. That is a big jump. How do you leverage that personally?

    21. none of the Chinese labs have released their full training data or the code they used to train their models, but they have been putting out detailed research papers that have helped push forward the state of the art, especially when it comes to efficient training and inference.

      perhaps bc they feed on existing efforts, and perhaps bc like the US models it is based on lots of copyright breaches.

    22. impressive roster of Chinese AI labs. I’ve been paying attention to these ones in particular: DeepSeek Alibaba Qwen (Qwen3) Moonshot AI (Kimi K2) Z.ai (GLM-4.5/4.6/4.7) MiniMax (M2) MetaStone AI (XBai o4) Most of these models aren’t just open weight, they are fully open source under OSI-approved licenses: Qwen use Apache 2.0 for most of their models, DeepSeek and Z.ai use MIT. Some of them are competitive with Claude 4 Sonnet and GPT-5!

      list of Chinese open sources / open weight models. Explore.

    23. GLM-4.7, Kimi K2 Thinking, MiMo-V2-Flash, DeepSeek V3.2, MiniMax-M2.1 are all Chinese open weight models. The highest non-Chinese model in that chart is OpenAI’s gpt-oss-120B (high), which comes in sixth place.

      Chinese models became very visible in 2025. - [ ] find ranking and description of Chinese llms

    24. It turns out tools like Claude Code and Codex CLI can burn through enormous amounts of tokens once you start setting them more challenging tasks, to the point that $200/month offers a substantial discount.

      running claudecode uses quite a bit of tokens, making 200usd/month a good deal for heavy users. I can believe that, also bc the machine doesn't care about the amount of tokens it uses during 'reasoning'. Some things I tried, it went through a whole bunch of steps and pages of scrolling output texts, to end up removing one word from a file. My suspicious half thinks, that if an AI company can influence the amount of tokens you use vibecoding, it will.

    25. One of my favourite pieces on LLM security this year is The Normalization of Deviance in AI by security researcher Johann Rehberger. Johann describes the “Normalization of Deviance” phenomenon, where repeated exposure to risky behaviour without negative consequences leads people and organizations to accept that risky behaviour as normal. This was originally described by sociologist Diane Vaughan as part of her work to understand the 1986 Space Shuttle Challenger disaster, caused by a faulty O-ring that engineers had known about for years. Plenty of successful launches led NASA culture to stop taking that risk seriously. Johann argues that the longer we get away with running these systems in fundamentally insecure ways, the closer we are getting to a Challenger disaster of our own.

      Normalisation of deviance: a risk taken without consequence reduces the perceived risk, while the risk is not changing itself. Johann Reberger (vgl o-ring issue in 1986 Challenger disaster.)

    26. the trade-off: using an agent without the safety wheels feels like a completely different product. A big benefit of asynchronous coding agents like Claude Code for web and Codex Cloud is that they can run in YOLO mode by default, since there’s no personal computer to damage. I run in YOLO mode all the time, despite being deeply aware of the risks involved. It hasn’t burned me yet... ... and that’s the problem.

      yolo mode, lol. If you do it, it feels like a very diff tool, and that is the lure / siren song.

    27. It helps that terminal commands with obscure syntax like sed and ffmpeg and bash itself are no longer a barrier to entry when an LLM can spit out the right command for you.

      bc Claudecode abstracts away the usual commands needed on the CLI. Vgl [[In the BeginningWas the Command Line by Neal Stephenson]]

    28. Claude Code and friends have conclusively demonstrated that developers will embrace LLMs on the command line, given powerful enough models and the right harness.

      Claude Code is what led devs to embrace CLI more.

    29. Maybe the terminal was just too weird and niche to ever become a mainstream tool for accessing LLMs?

      Well yes, it is. I know many how think the cli is scary or using it is for hackers.

    30. all the time thinking that it was weird that so few people were taking CLI access to models seriously—they felt like such a natural fit for Unix mechanisms like pipes.

      unix pipes, where output of one process is input of another, and you can bring them together in one statement. natural fit for model use Akin to promptchaining combined w tasks etc.

    31. I love the asynchronous coding agent category. They’re a great answer to the security challenges of running arbitrary code execution on a personal laptop and it’s really fun being able to fire off multiple tasks at once—often from my phone—and get decent results a few minutes later.

      async coding agents: prompt and forget

    32. Vendor-independent options include GitHub Copilot CLI, Amp, OpenCode, OpenHands CLI, and Pi. IDEs such as Zed, VS Code and Cursor invested a lot of effort in coding agent integration as well.

      non-vendor related coding agents. - [ ] which of these can I run locally? / integrate into VS Code

    33. The year of coding agents and Claude Code # The most impactful event of 2025 happened in February, with the quiet release of Claude Code. I say quiet because it didn’t even get its own blog post!

      Claude Code (feb 2025) seen by author as most impactful release of 2025.

    34. f you define agents as LLM systems that can perform useful work via tool calls over multiple steps then agents are here and they are proving to be extraordinarily useful. The two breakout categories for agents have been for coding and for search.

      recognisable, ai agents as chunked / abstracted away automation. This also creates the pitfall [[After claiming to redeploy 4,000 employees and automating their work with AI agents, Salesforce executives admit We were more confident about…. - The Times of India]] where regular automation is replaced by AI.

      Most useful for search and for coding

    35. It turned out that the real unlock of reasoning was in driving tools. Reasoning models with access to tools can plan out multi-step tasks, execute on them and continue to reason about the results such that they can update their plans to better achieve the desired goal. A notable result is that AI assisted search actually works now. Hooking up search engines to LLMs had questionable results before, but now I find even my more complex research questions can often be answered by GPT-5 Thinking in ChatGPT. Reasoning models are also exceptional at producing and debugging code. The reasoning trick means they can start with an error and step through many different layers of the codebase to find the root cause. I’ve found even the gnarliest of bugs can be diagnosed by a good reasoner with the ability to read and execute code against even large and complex codebases.

      Reasoning models are useful for: running tools (mcp) search now works debugging/writing code

    1. Er is nog een onderwerp waarover in Europa, maar ook in Nederland, geen maatschappelijk debat wordt gevoerd, zegt onderzoeker Fieke Jansen van de Universiteit van Amsterdam. „Wat gebeurt er ín de datacenters? Waar wordt de schaarse stroom voor gebruikt? De ‘Metaverse’ van Facebook? Youtubefilmpjes? Bitcoins mijnen? En vinden we dat nuttig genoeg om daar de bouw van een woonwijk in Almere voor op te offeren? We maken geen keuzes, terwijl het net aan z’n limiet zit.”

      This should not just be a quantitative discussion wrt energy usage, but als qualitative, what the energy is used for esp in light of the current scarcity of network space and the increasing likelihood of intermittency the coming years, which forces the question of who is most deserving of getting energy

    2. „Dit jaar nog zei iemand van het Directoraat Energie daar: ‘We gaan de verbruiksgegevens sowieso niet per datacenter publiceren, want als we dat gaan doen leveren ze helemaal niks meer aan.’ Wat is dat nou voor opstelling? Eis het gewoon op.”

      Will this change in light of geopolitics? EC signals weakening resolve (called 'simplification' but mostly cutting regs, and withdrawing from enforcement)

    3. De overheid kan gewoon naar Tennet en Liander stappen en die gegevens opeisen. Er is geen politieke wil om data boven tafel te krijgen. Wat wel heel gek is als je een land te besturen hebt, waarin scholen en wijken niet kunnen worden aangesloten op het net vanwege stroomtekort.”

      Data centers are not the only source of this information. Public enterprises that maintain the networks have this information too, and can (must) be mandated to share with RVO.

    4. Over 2025 moeten de bedrijven dat wel doen, maar „mogen ze aangeven dat dit bedrijfsgevoelige informatie is en wordt het alleen op geaggregeerd niveau openbaar door de Europese databank.”

      Odd phrase. Yes, aggregates might become public at EU level, but does not preclude Chapter II, which is not mentioned here as viable option From 2025 reporting is mandatory, so what changed? Did the EDD contain a timehorizon for mandatory reporting?

    5. Dat blijkt uit een inventarisatie van de formulieren die de RVO in 2025 ontving van Leitmotiv, waarover nu.nl onlangs ook publiceerde. Deze ngo bestaat uit een groep juristen en informatici die pleiten voor een „digitale economie waarin de voordelen van digitalisering rechtvaardig en democratisch worden verdeeld”.

      Leitmotiv, ngo v juristen/informatici wrt digital economy just / democratic / equality

    6. Van de circa 160 datacenters die zouden moeten rapporteren, stuurden 104 daadwerkelijk iets naar de RVO, telde Leitmotiv. 27 datacenters lieten daarbij de belangrijkste velden voor stroom- en waterverbruik leeg. Die waren, op drie na, allemaal in Amerikaanse handen. Ook Microsoft en Google, die tot de grootste stroomverbruikers in Nederland behoren, rapporteerden niet.

      160 datacenters in NL have reporting req. 104 did, and 27 left out the key information. Of those 27, 24 were of US entities, incl the biggest users Google and MS

    7. et Centraal Bureau voor de Statistiek (CBS) becijferde onlangs dat het totale stroomverbruik van datacenters in 2024 steeg naar ruim 5.000 gigawattuur, 4,5 procent van het totale stroomverbruik van Nederland – net zoveel als het verbruik van 2 miljoen huishoudens. En dat is buiten alle lopende aanvragen voor stroomaansluitingen gerekend. De netbeheerders, die deze aanvragen kennen, schrijven in hun meest recente toekomstscenario dat het stroomverbruik van datacenters over vijf jaar van 5 naar 15 procent van het totaal in Nederland zal zijn gegroeid.

      data center energy usage in 2024 4.5% of national usage, equiv to 2M householdes (out of 8M), or 25% of household usage. Network maintainers est growth to 15% of national usage within 5 years.

    8. de Amerikanen leggen de Europese regels in hun eigen voordeel uit en tot nu toe heeft geen enkel Europees land zin in een potje armpje drukken met de Amerikaanse techreuzen over hun energiegebruik. Ook Nederland niet. Dat ondervond De Valk, die opheldering vroeg bij de RVO. Ze stuurt NRC de antwoorden die de dienst haar gaf op haar vragen over de hyperscales bij Middenmeer. „Het klopt dat een groot gedeelte van de vragen niet zijn ingevuld”, schreef de dienst haar. „We hebben geen wettelijke middelen om datacentra te dwingen.”

      RVO treedt niet op tegen onvervulde rapportage eisen v datacentra.

    9. De nieuwe Europese energie-efficiëntierichtlijn, de EED, moest dat veranderen. De Europese richtlijn dwingt bedrijven transparant te zijn over hun energieverbruik en meer werk te maken van energiebesparing.

      This also means this data is in scope of DGA Chapter II

    10. De nieuwe Europese energie-efficiëntierichtlijn, de EED, moest dat veranderen. De Europese richtlijn dwingt bedrijven transparant te zijn over hun energieverbruik en meer werk te maken van energiebesparing. Alle grote bedrijven in Europa, inclusief datacenters, moeten daarom vanaf 2024 jaarlijks bij hun eigen overheid hun energie- en waterverbruik opgeven. In Nederland is dat bij de Rijksdienst voor Ondernemend Nederland, de RVO.

      De RVO verzamelt de gegevens onder de EED energy efficiency directive. Maar stelt artikel, data centers v Google en MS delen die data niet met RVO.

    1. Bun is a fast, incrementally adoptable all-in-one JavaScript, TypeScript & JSX toolkit. Use individual tools like bun test or bun install in Node.js projects, or adopt the complete stack with a fast JavaScript runtime, bundler, test runner, and package manager built in. Bun aims for 100% Node.js compatibility.

      Bun, #2025/12 bought by Anthropic for Claudecode, is a toolkit for js, typescript, jsx. you can use parts of it within node.js. Aims for full compatibility with node.js. Aims for means it doesn't I suppose

    1. Baldur Bjarnason notices that a number of the 1200 (!) blogs he follows which are normally dormant have become active again, but on the topic of AI if not generated by AI. By the original blogger. Blandness ensues.

    1. German ntv media on a potential fireworks ban in Germany, and comparing notes with how long it took for the Netherlands to ban it (as of now, yesterday was the last time), despite 60-80% being in favor of such a ban, multiple deaths and many permanently wounded (eyes, burns) each year, and police, ambulance staff, and firebrigade being frequently targeted so that many considered quitting.

  5. Dec 2025
    1. the requirements of coding that expand the user's agency rather than automating or replacing it. He builds on end-user software engineering (by figures such as Margaret Burnett, Bonnie Nardi, and Margaret Boden), and also on social and political critics of AI (e.g., Ruha Benjamin, Rachel Adams, Abeba Birhane, and Shaowen Bardzell).

      n::: what requirements can you list for programming that increases agency, not automating / replacing it and abstracting it away.

    2. First, we must cultivate widespread engagement with technology through everyday programming: “The message of this book is that the world needs less AI, and better programming languages” (125). Escaping our AI dead end means more programming, not less, perhaps even popular or mass programming.

      programming as antidote to AI/programming