3,565 Matching Annotations
  1. Dec 2025
    1. crisis of human attention. Attention is the basis of concentration and of thinking; even more fundamentally, for Blackwell being human is grounded in “being attending and attentive beings, paying attention” (11). Platform AI is designed to consume our attention and harvest our data. It will continue to damage human agency, attention, and intelligence, Blackwell insists, unless at least two things happen.

      attention as what AI erodes

    2. he also disagrees that the transition from Good Old-Fashioned AI (GOFAI), based on programmed rules, to second-generation AI, based on pattern finding, is programming's actual arc of progress. He values the various contemporary modes of machine learning but sees today's AI shift as a detour away from a powerful programming tradition built on increasing human agency rather than replacing it.

      n: ai arc of programming evo is a 'detour' where it replaces human agency. where programming itself is historically based on increasing agency At first glance this is a big tech vs bottom up dev thing too. I can see where AI can increase agency locally individually too. Just not through the AI offerings of bigtech

    3. deep concern about a corporate-driven, tech-justified trivialization of human attention and the prospective stupefaction of our collective abilities to solve humanity's gigantic problems. His alternative takes time to build over the course of the book. His agency-centered “MORAL” standard for code emerges not from utopian hopes for the future but from the history of programming itself, freed from its current capture by technology platforms.

      a, MORAL, as acronym, I came across that somewhere.

    4. he is completely reorienting the history of programming as one that refuses AI as its culmination. This will likely be new for many contemporary programmers, and may come as a shock to nonspecialists awash in standard media accounts of the AI revolution.

      Moral Codes: Designing Alternatives to AI. By Alan F. Blackwell, repositions AI not as the culmination of programming. Makes me realise that indeed others do tend to treat it as such.

    1. Digital sovereignty: definition, origin and history Digital sovereignty is loosely defined as the ability of a governing body, such as a national government, to control the tech stacks and data flows within its boundaries. For instance, in a digitally sovereign state, any data centres within its physical boundaries and locally hosted software are beholden only to the laws of that country.

      This seems a confused definition.

    1. “AI” is not our sole path to “closing the competitiveness gap”. Europe’s people and businesses need sovereign tech infrastructure, reached through industrial leadership, to support all our digital experiences.We can power this up by federating existing assets, coordinating them with public and private investments, installing stakeholder governance to protect intermediation infrastructure from capture, and focusing on adoption.We are a broad independent non-lobby movement with a sense of urgency and a strong bias for action, working to push the European Parliament and the Commission to “do the right thing” for Europe.

      nonlobby movement is bunk obv. They're not wrong though.

    1. If you state that your funding partners don't impact your mission or decisions, and then list them as source of legitimacy, you express the opposite. Transparency is a good thing, but this serves mostly to show where they're coming from imo.

    1. Hubert: „De Europese industrie staat buitenspel. Maar vervolgens zijn de techbedrijven hier er ook niet in geïnteresseerd om de kar te trekken naar een oplossing.” Zijn conclusie: „EuroStack staat voor de onmogelijke uitdaging klanten die niet willen kopen te koppelen aan leveranciers die niet willen maken. Strategisch loop je dan dood.”

      [[Bert Hubert c]] ziet geen rol (meer) voor Eurostack als initiatief, na de positieve invloed in 2024 en 2025. Zijn omschrijving wijst opnieuw naar noodzaak van actie

    2. Aanvankelijk was de EuroStack-oproep vooral „Buy European” (gericht aan overheden), daarna ook „Sell European” (aan de bedrijven) en nu is het „Fund European”. Dat laatste gaat lukken, gelooft Caffarra. Ze noemt nieuwe Europese techfondsen, onder meer van het (Amerikaanse) Sequoia Capital en het in Zwitserland gevestigde investeringsfonds Lakestar. En hint telefonisch op fondsen waarvan ze de naam nog niet wil noemen, onder meer van rijke Europese families die zouden willen investeren in Europese tech, maar zeggen daarbij advies van EuroStack te willen gebruiken.

      buy / sell / fund European. At least more, and not by default bigtech. But the mention of the American Sequoia fund here is a red flag, preempting public sector digital sovereignty.

    3. Ze reageert als door een wesp gestoken op de suggestie dat de EuroStack-stichting wel iets weg heeft van een ngo die ergens voor lobbyt, namelijk Europese digitale soevereiniteit. „Niemand betaalt me. Ik bepleit een zaak waarin ik geloof. In mijn eigen tijd en van mijn eigen geld. Als mensen je betalen, denken ze dat ze je bevelen kunnen geven.”

      another odd sentence, this time from Caffarra, implying NGOs are always someone else's agenda. Huh?

    4. De stichting moet helpen bij de stap van ‘praten naar actie’, staat in de verklaring. En dat het tijd is om te gaan ‘bouwen’ aan het Europese aanbod. Wat dit concreet betekent, is niet gelijk duidelijk. Het is vooral geen brancheorganisatie, zegt Caffarra telefonisch. Daarvan lopen er in Brussel al genoeg rond.

      Foundation wants to move from talk to walk, but unclear how. This is where the groundswell comes in right, the smaller providers joining forces a la Nextcloud, the NLnet stuff EU funding.

    5. In november, dezelfde week als de Europese top in Berlijn plaatsvindt, registreren Caffarra en Karlitschek EuroStack als stichting in de Duitse hoofdstad. Caffarra wordt voorzitter, Karlitschek een van de bestuursleden. Anderen zijn bijvoorbeeld de ceo van Proton, de Zwitserse aanbieder van beveiligde mail en cloud, en een aantal Franse en Duitse techondernemers en investeerders.

      Missed this last month, odd. Eurostack is now a foundation based in Berlin. Caffarra is chair, Karlitschek is board member, and Proton. - [ ] find out all boardmembers Eurostack wrt SC landscape

    6. Daarin gaat het nadrukkelijk over autonomie en keuzevrijheid, niet over soevereiniteit, want iedereen weet dat volledige ontkoppeling een illusie is en niemand wil de regering Trump onnodig tegen de haren instrijken.

      n:: Odd sentence, 'sovereignty' doesn't mean fully disconnect either. Public sector needs sovereignty as sine-qua-non bc if someone else holds the off-switch that you don't control, you are the colony that Caffarra mentioned at top. Only autonomy is sovereignty washing itself

    7. Hij raakt teleurgesteld over hoe de Europese bedrijven zich opstellen, vooral cloudaanbieders als Herztner, Leaseweb (Nederlands), OVH en Ionos. „Je zou verwachten dat die voorop zouden lopen in de strijd. Maar in plaats daarvan zeggen ze ‘Het ligt niet aan ons dat mensen onze spullen niet kopen’.”

      [[Bert Hubert c]] thinks the reaction of Hetzner (misspelled here), Leaseweb (involved in #jtc25), OVH en Ionos (collab w Nextcloud) is disappointing. They're not making themselves visible ('we were already here') or make noise now at the opportunity arising.

    8. Cristina Caffarra houdt vrijwel dagelijks ergens haar peptalk. Als het jaar vordert steeds vaker via videoverbinding, want het is allemaal niet meer te bereizen. Ze lanceert in juli haar eigen podcast, Escape Forward. En blijft fanatiek en uitgesproken op LinkedIn. Uit haar posts spreekt wel steeds meer frustratie. „Europese elites vernielen Europa zelf”, schrijft ze bijvoorbeeld als Europeanen begin december geschokt reageren op de Amerikaanse nationale veiligheidsstrategie, die uitgesproken anti-EU is. „Ze praten maar over hun waarden en de geweldige Europese manier van leven, maar hebben niet de minste interesse in het bouwen van een eigen digitale infrastructuur”, schrijft ze.

      Caffarra has a podcast, and actively posts on LinkedIn, described here as getting increasingly frustrated. Again, in part I think bc she aims for the big changes at political / econ level, where that can only happen if there's enough groundswell, like the work Karlitschek has been doing for well over a decade.

    9. Microsoft biedt een soevereine cloud-oplossing, Amazon ook, Google ook. De bedrijven beloven bijvoorbeeld datacenters in Europa te gebruiken. Of brengen een extra – Europese – bestuurslaag aan in hun bedrijf. Krijgen Europeanen daarmee de verlangde onafhankelijkheid? De uiteindelijke eigenaren blijven Amerikaans. Sovereignty washing noemt de groep rond Caffarra het, analoog aan ‘green washing’, de ingeburgerde term voor bedrijven die net doen alsof ze duurzaam zijn.

      Missed opportunity to state why this is not enough: US regs

    10. Caffarra wil dat de Europese industrie zich uitspreekt voor Europees aanbesteden en spreekt haar contacten in het bedrijfsleven hierover aan. Het resulteert half maart in een gezamenlijke brief van Europese ceo’s aan de voorzitter van de Europese Commissie en de Eurocommissaris van digitale zaken. „Je kunt jezelf niet uit de positie van achterblijver reguleren”, staat er onder meer. De lange lijst namen eronder illustreert vooral hoe onbekend de meeste Europese techbedrijven zijn. Er staan ook grotere spelers onder, zoals de topman van Airbus

      March 2025 public letter to EC Virkkunen by European tech ceo's. Go through list of signatories, for SC landscape input.

    11. En dus probeert de NextCloud-topman de gefragmenteerde Europese ict-industrie in beweging te krijgen. Op 4 maart pakt Karlitschek het vliegtuig naar Milaan voor een EuroStack-bijeenkomst met vooral Europese ondernemers. Het is de bedoeling daar stappen te zetten richting een gezamenlijk Europees ict-aanbod. Ze spreken er een eerste stap af naar een gezamenlijke Europese standaard voor clouds.

      March 2025, Eurostack meeting in Milano, where a first step towards European cloud standard would have been decided. Connection to #jtc25 ?

    12. Wat die bedrijven voor hun klanten zo aantrekkelijk maakt, is dat achter één loket een hele wereld schuilgaat. Wie in Europa iets vergelijkbaars wil kopen, moet zakendoen met allerlei kleine en middelgrote bedrijven. En rekening houden met de kans dat die technische ‘oplossingen’ (ict-jargon) nét niet lekker op elkaar aansluiten.

      Excactly this. It is described here as the issue, but it really also is the only solution. You're escaping monopolists. That always adds friction. And the real question is, what was attractive first, is it really still now, and its cost explainable?

    13. In februari zegt Vance tijdens een speech op de jaarlijkse veiligheidsconferentie in München onder meer dat Europa zichzelf van binnenuit uitholt. De democratie in de EU zou niet meer functioneren, wat onder meer zou komen door de Europese regels voor de digitale wereld – die in de praktijk vooral de grote Amerikaanse sociale mediabedrijven als Meta en X treffen.

      The Feb 2025 security conf in M another turning point where US admin turns on EU digital regs as threat to democracy. US admin coopted by bigtech becomes more clear

    14. Frank Karlitschek voelt verantwoordelijkheid, hij wil de Europese ‘techstack’ helpen bouwen. De Duitse softwarebouwer en ondernemer biedt met zijn bedrijf NextCloud kantoorsoftware aan à la Microsoft,

      [[Frank Karlitschek p]] has been doing this for over decade already, and that needs mentioning. I talked to him [[Berlin 2014]] at re:publica about this, in the light of the steps E and I were taking in our personal digitisation, and when I moved my company to nextcloud.

    15. Hoe week je Europa los uit de Amerikaanse digitale greep. En hoe verkoop je iets wat er nog niet is?

      This is similar to individual siloquits. In reality it is doable, by recognising the diff parts (here of the stack). Hyperscalers are the toughest nut bc they combine several stack layers in themselves, and you'd need a full alternative for them, but not another hyperscaler. That is the route.

    16. overheden, stimuleer de vraag naar alternatieven voor de diensten van grote Amerikaanse bedrijven zoals Microsoft, Google en Amazon. Doe een percentage – bijvoorbeeld 20 of 30 procent –van de overheidsbestedingen Europees. Dan stimuleer je de vraag en gaan Europese bedrijven die producten en diensten ook ontwikkelen

      public procurement is the easiest way to change things. that money is already being spent on digital, so if more of it is spend on European providers that's a helpful step.

    17. In het stuk gebruikt Chamber of Progress de term digital curtain. De suggestie is dat Europeanen zichzelf achter een digitaal gordijn zetten als ze proberen alle technologie zelf in elkaar te knutselen – een verwijzing naar het leven achter het IJzeren Gordijn tijdens de Koude oorlog.

      'digital curtain' a term used for splinternet by us bigtech to try and prevent EU be more assertive in their own digital market.

    18. de Chamber of Progress, heeft laten uitrekenen wat het de EU zou kosten als het de diensten van de huidige Amerikaanse techbedrijven in Europa wil vervangen door spullen van eigen makelij. De uitkomst: ten minste 25 keer de hele EU-begroting. De berekening is naar medium Politico gelekt

      US bigtech lobby published a report in Sept 2024 stating creating a Eurostack would be too costly. Report linked.

    19. Samen met een andere gedreven Italiaanse econoom, Francesca Bria, en met de baas van berichtendienst Signal, Meredith Wittaker, organiseert Caffarra in september 2024 een bijeenkomst in het Europarlement getiteld ‘Toward European Digital Independence’. De ondertitel is ‘Building the EuroStack’

      Eurostack was the subtitle of a Sept 2024 meeting in European Parliament. Organised by Caffarra, Meredith Wittaker of Signal, and [[Francesca Bria c]]

    20. Caffarra heeft van binnenuit gezien hoe de macht van de grote Amerikaanse techbedrijven groeide. Europese bedrijven werden overgenomen en konden niet concurreren met de Amerikanen. Getalenteerde Europeanen emigreerden. Ondernemers die kapitaal nodig hebben wijken nu uit naar de VS. En de EU is in hoog tempo veranderd in wat Caffarra een ‘digitale kolonie van Amerika’ noemt. Het frustreert haar en ze wil dat die ontwikkeling stopt. Maar hoe krijg je in 27 lidstaten zowel de ondernemers als de politici en toezichthouders in beweging?

      Caffarra mentions four elements leading to digital colonisation of EU from USA. Buy-outs, inability to compete, brain drain, capital. I think adopting the US framing of what success / growth is plays a factor too. In a scheme set by someone you will never succeed other than playing by that someone's rules.

    21. De van oorsprong Italiaanse econoom en mededingingsexpert Cristina Caffarra is een van de drijvende krachten achter die groep. Deze gebruikt de hashtag ‘EuroStack’ bij haar pogingen Europese overheden op te poken. Meestal spreken de ondernemers, academici, techjuristen en politici uit verschillende landen elkaar online en via Signal. Het sjieke diner in Museum Bellevue in Brussel is een kans om elkaar beter te leren kennen. Gastvrouw Caffarra heeft goed verdiend met klussen voor grote Amerikaanse techbedrijven zoals Apple en Amazon en de Europese Commissie (in rechtszaken tegen Google) en kan het zich nu veroorloven te doen wat ze leuk en belangrijk vindt. Ze is goed in netwerken en peptalks geven. En ze neemt geen blad voor de mond, waarbij blijkt dat ze duidelijk meer op heeft met doeners uit het bedrijfsleven dan met politici en denktankers.

      Cristina Caffarra mentioned as driving force behind Eurostack

    1. https://web.archive.org/web/20251230194055/https://www.tomsguide.com/ai/i-turned-a-hotel-key-card-into-a-one-tap-shortcut-for-chatgpt-and-now-i-use-it-every-day

      You can use any NFC card (like a hotel key card, which the author doesn't, but I do, usually return on check-out, it seems) to connect it to an iPhone shortcut. Tap the card and it triggers some action, response or workflow.

      Says people use it for playlists and lights too. I don't really buy his examples though. You either have to have an NFC in a fixed location (the 'lights' example I believe therefore), or on the move you'd have to dig out the 'right' NFC from someplace (your already full wallet?) then tap it to the phone. That creates actually _more _ friction. 'I stuck a tag on my desk for ....' something specific like he suggests (a list of articles on AI from past 24h) leads to a range of tags on your desk, like when Amazon suggested you have a bunch of tags, one for each product, to build your shopping list. Didn't happen.

    1. some practices that can make those discussions easier, by starting with constraints that even skeptical developers can see the value in: Build tools around verbs, not nouns. Create checkEligibility() or getRecentTickets() instead of getCustomer(). Verbs force you to think about specific actions and naturally limit scope.Talk about minimizing data needs. Before anyone creates an MCP tool, have a discussion about what the smallest piece of data they need to provide for the AI to do its job is and what experiments they can run to figure out what the AI truly needs.Break reads apart from reasoning. Separate data fetching from decision-making when you design your MCP tools. A simple findCustomerId() tool that returns just an ID uses minimal tokens—and might not even need to be an MCP tool at all, if a simple API call will do. Then getCustomerDetailsForRefund(id) pulls only the specific fields needed for that decision. This pattern keeps context focused and makes it obvious when someone’s trying to fetch everything.Dashboard the waste. The best argument against data hoarding is showing the waste. Track the ratio of tokens fetched versus tokens used and display them in an “information radiator” style dashboard that everyone can see. When a tool pulls 5,000 tokens but the AI only references 200 in its answer, everyone can see the problem. Once developers see they’re paying for tokens they never use, they get very interested in fixing it.

      some useful tips to keep MCPs straightforward and prevent data blobs that are too big. - use verbs not nouns for mcp tool names (focuses on the action, not the object upon you act) - think/talk about n:: data minimalisation - break it up, reads separate from reasoning steps. Keeps everything focused on the specific context. - dashboard the ratio of tokens fetched versus tokens used in answers. Lopsided ratios indicate you're overfeeding the system.

    2. In an extreme case of data hoarding infecting an entire company, you might discover that every team in your organization is building their own blob. Support has one version of customer data, sales has another, product has a third. The same customer looks completely different depending on which AI assistant you ask. New teams come along, see what appears to be working, and copy the pattern. Now you’ve got data hoarding as organizational culture.

      MCP data hoarding leads to parallel data households, exactly the type of thing we spent a lot of energy on to reduce

    3. data hoarding trap find themselves violating the principle of least privilege: Applications should have access to the data they need, but no more

      n:: Principle of least privilege: applications only should have access to data they need, and never more. Data hoarding in MCPs goes beyond that.

    4. There’s also a security dimension to data hoarding that teams often miss. Every piece of data you expose through an MCP tool is a potential vulnerability. If an attacker finds an unprotected endpoint, they can pull everything that tool provides. If you’re hoarding data, that’s your entire customer database instead of just the three fields actually needed for the task.

      MCPs that are overloaded w data are new attack surfaces

    5. MCP can remove the friction that comes from those trade-offs by letting us avoid having to make those decisions at all.

      MCP is meant to abstract the way access is created to resources. In practice it gets used to abstract away any decision on which data to provide or not. That's the trap.

    6. The team ended up with a data architecture that buried the signal in noise. That additional load put stress on the AI to dig out that signal, leading to serious potential long-term problems. But they didn’t realize it yet, because the AI kept producing reasonable-looking answers. As they added more data sources over the following weeks, the AI started taking longer to respond. Hallucinations crept in that they couldn’t track down to any specific data source. What had been a really valuable tool became a bear to maintain.

      Having a clear data architecture for your use case is needed. Vgl [[Eindelijk weet ik wat ThetaOS is een Life Lens System (LLS)]] wrt number of data tables (152 now I think), and how it grew over time, deciding on each table added.

    7. I’ve been watching teams adopt MCP over the past year, and I’m seeing a disturbing pattern. Developers are using MCP to quickly connect their AI assistants to every data source they can find—customer databases, support tickets, internal APIs, document stores—and dumping it all into the AI’s context.

      Dev Andrew Stallman warns against dumping all-the-data into an AI application through MCP. Calls it hoarding.

    1. Didn't realise that in 2022 a follow-up to [[ A Psalm for the Wild-Built by Becky Chambers]] was published: [[A Prayer for the Crown-Shy by Becky Chambers]] , for the [[Aan te schaffen boeken]] list

    1. The real power of MCP emerges when multiple servers work together, combining their specialized capabilities through a unified interface.

      Combining multiple MCP servers creates a more capable set-up.

    2. Prompts are structured templates that define expected inputs and interaction patterns. They are user-controlled, requiring explicit invocation rather than automatic triggering. Prompts can be context-aware, referencing available resources and tools to create comprehensive workflows. Similar to resources, prompts support parameter completion to help users discover valid argument values.

      prompts are user invoked (hey AgentX, go do..) and may contain next to instructions also references and tools. So a prompt may be a full workflow.

    3. Prompts Prompts provide reusable templates. They allow MCP server authors to provide parameterized prompts for a domain, or showcase how to best use the MCP server. ​

      mcp prompts are templates for interaction

    4. Resources support two discovery patterns: Direct Resources - fixed URIs that point to specific data. Example: calendar://events/2024 - returns calendar availability for 2024 Resource Templates - dynamic URIs with parameters for flexible queries. Example: travel://activities/{city}/{category} - returns activities by city and category travel://activities/barcelona/museums - returns all museums in Barcelona Resource Templates include metadata such as title, description, and expected MIME type, making them discoverable and self-documenting.

      Resources can be invoked w fixed and dynamic URIs

    5. Resources expose data from files, APIs, databases, or any other source that an AI needs to understand context. Applications can access this information directly and decide how to use it - whether that’s selecting relevant portions, searching with embeddings, or passing it all to the model.

      resources are just that, read only material to invoke. API, filesystem, databases etc.

    6. Each tool performs a single operation with clearly defined inputs and outputs. Tools may require user consent prior to execution, helping to ensure users maintain control over actions taken by a model.

      Almost function call like.

    7. Tools are model-controlled, meaning AI models can discover and invoke them automatically. However, MCP emphasizes human oversight through several mechanisms. For trust and safety, applications can implement user control through various mechanisms, such as: Displaying available tools in the UI, enabling users to define whether a tool should be made available in specific interactions Approval dialogs for individual tool executions Permission settings for pre-approving certain safe operations Activity logs that show all tool executions with their results

      Tools are available to models, but human in the loop options exist: approval, permission settings, logs

    8. Servers provide functionality through three building blocks:

      n:: MCP servers typically provide three types of building blocks, a) Tools that an LLM can call, b) resources that are read-only resources to an LLM, c) prompts, prewritten instructions templates, i.e. agent descriptions, that outline specific tools and resources to use. So for agentic stuff you'd have an MCP server providing templates which in turn list tools and resources.

    9. Visual Studio Code acts as an MCP host. When Visual Studio Code establishes a connection to an MCP server, such as the Sentry MCP server, the Visual Studio Code runtime instantiates an MCP client object that maintains the connection to the Sentry MCP server.

      VS Code acts as MCP Host (in their AI toolkit extension I think). You could connect it to the Obsidian MCP server plugin then?

    10. The key participants in the MCP architecture are: MCP Host: The AI application that coordinates and manages one or multiple MCP clients MCP Client: A component that maintains a connection to an MCP server and obtains context from an MCP server for the MCP host to use MCP Server: A program that provides context to MCP clients

      The MCP architecture has 3 pieces The host (application, AI or not, that coords the interaction with MCP clients), an MCP client that interacts with a single server. MCP server, which provides the context, i.e. abstracts the access to other sources (filesystem, database, API etc). A server can have one or multiple clients it serves.

    1. A comparison between VS Code and Obsidian. Doesn't state the obvious: any text editor can do this. The tools are just viewers and do not contain the data, which is part of your filesystem. Vgl [[3 Distributed Eigenschappen 20180703150724]]

    1. Ignác Semmelweis in 1847 argued for hand washing in maternity wards by doctors, and published a book about it. Was ridiculed for it and died 1865 as an outcast in an asylum. Only the later emergence of germ theory provided a theoretical basis for the empirical observations of Semmelweis. 'Semmelweis-moment' where someone who is right is laughed out of the room.

    1. owned by the Akaunting Software Inc.

      Another business not being clear where they are based. Akaunting as a term seems Bulgarian in spelling's origin, the lead dev has a Turkish personal domain, the company on #socmed lists its location as London. The company is not registered at Companies House though (2 others with similar name are, but different), and not known in opencorporates.com.

    1. Member states holding the presidency work together closely in groups of three, called 'trios'. This system was introduced by the Lisbon Treaty in 2009. The trio sets long-term goals and prepares a common agenda determining the topics and major issues that will be addressed by the Council over an 18-month period. On the basis of this programme, each of the three countries prepares its own more detailed six-month programme.

      Three MS always work together over their 18 months. This provides continuity for long-term goals. This is arranged in the Lisbon Treaty (2009).

    1. Official schedule for Council EU presidency to 2030:

      2026 Cyprus Ireland (trio IRL, LT, GR) 2027 Lithuania (trio IRL, LT, GR) Greece (trio IRL, LT, GR) 2028 Italy (trio I, LV, L) Latvia (trio I, LV, L) 2029 Luxembourg (trio I, LV, L) Netherlands (trio NL, SK, M) 2030 Slovakia (trio NL, SK, M) Malta (trio NL, SK, M)

      NL in H2 2029 (and a trio looking 1 yr further ahead)

    1. For us personally, this means that we no longer use generative AI – neither for private nor professional purposes.

      Authors avoid the use of generative AI. But realise that is difficult for most to do, and as such a privileged tech capable position

    2. To Gen or Not To Gen: The Ethical Use of Generative AI 33 minute read This blog entry started out as a translation of an article that my colleague Jakob and I wrote for a German magazine. After that we added more stuff and enriched it by additional references and sources. We aim at giving an overview about many - but not all - aspects that we learned about GenAI and that we consider relevant for an informed ethical opinion. As for the depth of information, we are just scratching the surface; hopefully, the loads of references can lead you to diving in deeper wherever you want. Since we are both software developers our views are biased and distorted. Keep also in mind that any writing about a “hot” topic like this is nothing but a snapshot of what we think to know today. By the time you read it the authors’ knowledge and opinions have already changed. Last Update: December 8, 2025. Table of ContentsPermalink Abstract About us Johannes Link Jakob Schnell Introduction Ethics, what does that even mean? Clarification of terms Basics Can LLMs think? What LLMs are good at GenAI as a knowledge source GenAI in software development Actual vs. promised benefits Harmful aspects of GenAI GenAI is an ecological disaster Power Water Electronic Waste GenAI threatens education and science GenAI is destroying the free internet. GenAI is a danger to democracy GenAI versus human creativity Digital colonialism Political aspects Conclusion Can there be ethical GenAI? How to act ethically AbstractPermalink ChatGPT, Gemini, Copilot. The number of generative AI applications (GenAI) and models is growing every day. In the field of software development in particular, code generation, coding assistants and vibe coding are on everyone’s lips. Like any technology, GenAI has two sides. The great promises are offset by numerous disadvantages: immense energy consumption, mountains of electronic waste, the proliferation of misinformation on the internet and the dubious handling of intellectual property are just a few of the many negative aspects. Ethically responsible behaviour requires us to look at all the advantages, disadvantages and collateral damages of a technology before we use it or recommend its use to others. In this article, we examine both sides and eventually arrive at our personal and naturally subjective answer to whether and how GenAI can be used in an ethical manner. About usPermalink Johannes LinkPermalink … has been programming for over 40 years, 30 of them professionally. Since the end of the last century, extreme programming and other human-centred software development approaches have been at the heart of his work. The meaningful and ethical implementation of his private and professional life has been his driving force for years. He has been involved with GenAI since the early days of OpenAI’s GPT language models. More about Johannes can be found at https://johanneslink.net. Jakob SchnellPermalink … studied mathematics and computer science and has been working as a software developer for 5 years. He works as a lecturer and course director in university and non-university settings. As a youth leader, he also comes into regular contact with the lives of children and young people. In all these environments, he observes the growing use of GenAI and its impact on people. IntroductionPermalink Ethics, what does that even mean?Permalink Ethical behaviour sounds like the title of a boring university seminar. However, if you look at the wikipedia article of the term 1, you will find that ‘how individuals behave when confronted with ethical dilemmas’ is at the heart of the definition. So it’s about us as humans taking responsibility and weighing up whether and how we do or don’t do certain things based on our values. We have to consider ethical questions in our work because all the technologies we use and promote have an impact on us and on others. Therefore, they are neither neutral nor without alternative. It is about weighing up the advantages and potential against the damage and risks; and that applies to everyone, not just us personally. Because often those who benefit from a development are different from those who suffer the consequences. As individuals and as a society, we have the right to decide whether and how we want to use technologies. Ideally, this should be in a way that benefits us all; but under no circumstances should it be in a way that benefits a small group and harms the majority. The crux of the matter is that ethical behaviour does not come for free. Ethics are neither efficient nor do they enhance your economic profit. That means that by acting according to your values you will, at some point, have to give something up. If you’re not willing to do that, you don’t have values - just opinions. Clarification of termsPermalink When we write ‘generative AI’ (GenAI), we are referring to a very specific subset of the many techniques and approaches that fall under the term ‘artificial intelligence’. Strictly speaking, these are a variety of very different approaches that range from symbolic logic, over automated planning up to the broad field of machine learning (ML). Nowadays most effort, hype and money goes into deep learning (DL): a subfield of ML that uses multi-layered artificial neural networks to discover statistical correlations (aka patterns) based on very large amounts of training data in order to reproduce those patterns later. Large language models (LLM) and related methods for generating images, videos and speech now make it possible to apply this idea to completely unstructured data. While traditional ML methods often managed with a few dozen parameters, these models now work with several trillion (10^12) parameters. In order for this to produce the desired results, both the amount of training data and the training duration must be increased by several orders of magnitude. This brings us to the definition of what we mean by ‘GenAI’ in this article: Hyperscaled models that can only be developed, trained and deployed by a handful of companies in the world. These are primarily the GenAI services provided by OpenAI, Anthropic, Google and Microsoft, or based on these services. We also focus primarily on language models; the generation of images, videos, speech and music plays only a minor role in this article. Our focus on hyperscale services does not mean that other ML methods are free of ethical problems; however, we are dealing with a completely different order of magnitude of damage and risk here. For example, there do exist variations of GenAI that use the same or similar techniques, but on a much smaller scale and restricted domains (e.g. AlphaFold 2). These approaches tend to bring more value with fewer downsides. BasicsPermalink GenAI models are designed to interpolate and extrapolate 3, i.e. to fill in the gaps between training data and speculate beyond the limits of the training data. Together with the stochastic nature of the training data, this results in some interesting properties: GenAI models ‘invent’ answers; with LLMs, we like to refer to this as ‘hallucinations’. GenAI models do not know what is true or false, good or bad, efficient or effective, only what is statistically probable or improbable in relation to training data, context and query (aka prompt). GenAI models cannot explain their output; they have no capability of introspection. What is sold as introspection is just more output, with the previous output re-injected. GenAI models do not learn from you; they only draw from their training material. The learning experience is faked by reinjecting prior input into a conversation’s context 4. The context, i.e. the set of input parameters provided, is decisive for the accuracy of the generated result, but can also steer the model in the wrong direction. Increasing the context window makes a query much more computation-intensive - likely in a quadratic way. Therefore, the promised increase of “maximum context window” of many models is mostly fake 5. The reliability of LLMs cannot be fundamentally increased by even greater scaling 6. Can LLMs think?Permalink Proponents of the language-of-thought hypothesis 7 believe it is possible for purely language-based models to acquire the capabilities of the human brain – reasoning, modelling, abstraction and much more. Some enthusiasts even claim that today’s models have already acquired this capability. However, recent studies 8 9 show that today’s models are neither capable of genuine reasoning nor do they build internal models of the world. Moreover, “…according to current neuroscience, human thinking is largely independent of human language 10” and there is fundamental scientific doubt that achieving human cognition through computation is achievable in practice let alone by scaling up training of deep networks 11. An example of a lack of understanding of the world is the prompt ‘Give me a random number between 0 and 50’. The typical GenAI response to this is ‘27’, and it is significantly more reliable than true randomness would allow. (If you don’t believe it, just try it out!) This is because 27 is the most likely answer in the GenAI training data – and not because the model understands what ‘random’ means. ‘Chain of Thought (CoT)’ approaches and ‘Reasoning models’ attempt to improve reasoning by breaking down a prompt, the query to the model, into individual (logical) steps and then delegating these individual steps back to the LLM. This allows some well-known reasoning benchmarks to be met, but it also multiplies the necessary computational effort by a factor between 30 and 700 12. In addition, multistep reasoning lets individual errors chain together to form large errors. And yet, CoT models do not seem to possess any real reasoning abilities 13 14 and improve the overall accuracy of LLMs only marginally 15. The following thought experiment from 16 underscores the lack of real “thinking” capabilities: LLMs have simultaneous access to significantly more knowledge than humans. Together with the postulated ability of LLMs to think logically and draw conclusions, new insights should just fall from the sky. But they don’t. Getting new insights from LLMs would require these to be already encoded in the existing training material, and to be decoded and extracted by pure statistical means. What LLMs are good atPermalink Undoubtedly, LLMs represent a major qualitative advance when it comes to extracting information from texts, generating texts in natural and artificial languages, and machine translation. But even here, the error rate, and above all the type of error (‘hallucinations’), is so high that autonomous, unsupervised use in serious applications must be considered highly negligent. GenAI as a knowledge sourcePermalink As we have pointed out above, LLMs cannot differentiate between true and false - regardless of the training material. It does not answer the question “What is XYZ?” but the question “How would an answer to question ‘What is XYZ?’ look like?”. Nevertheless, many people claim that the answers that ChatGPT and alike provide for the typical what-how-when-who queries are good enough and often better than what a “normal” web search would have given us. Arguably, this is the most prevalent use case for “AI” bots today. The problem is that most of the time we will never learn about the inaccuracies, left-outs, distortions and biases that the answer contained - unless we re-check everything, which defies the whole purpose of speeding up knowledge retrieval. The less we already know, the better the “AI’s” answer looks to us, but the less equipped we are to spot the problems. A recent by the BBC and 22 Public Service Media organizations shows that 45% of all “AI” assistants’ answers on questions about news and current affairs have significant errors 17. Moreover, LLMs are easy prey for manipulation - either by the service providing organization or by third parties. A recent study claims that even multi-billion-parameter models can be “poisoned” by injecting just a few corrupted documents 18. So, if anything is at stake all output from LLMs must be carefully validated. Doing that, however, would contradict the whole point of using “AI” to speed up knowledge acquisition. GenAI in software developmentPermalink The creation and modification of computer programmes is considered a prime domain for the use of LLMs. This is partly because programming languages have less linguistic variance and ambiguity than natural languages. Moreover, there are many methods for automatically checking generated source code, such as compiling, static code analysis and automated testing. This simplifies the validation of generated code and thereby gives an additional feeling of trust. Nevertheless, individual reports on the success of coding assistants such as Copilot, Cursor, etc. vary greatly. They range from ‘completely replacing me as a developer’ to ‘significantly hindering my work’. Some argue that coding agents considerably reduce the time they have to invest in “boilerplate” work, like writing tests, creating data transfer objects or connecting your domain code to external libraries. Others counter by pointing out that delegating these drudgeries to GenAI makes you miss opportunities to get rid of them, e.g. by introducing a new abstraction or automating parts of your pipeline, and to learn about the intricacies and failure modes of the external library. Other than old-school code generation or code libraries prompting a coding agent is not “just another layer of abstraction”. It misses out on several crucial aspects of a useful abstraction: Its output is not deterministic. You cannot rely on any agent producing the same code next time you feed it the same prompt. The agent does not hide the implementation details, nor does it allow you to reliably change those details if the previous implementation turns out to be inadequate. Code that is output by an LLM, even if it is generated “for free”, has to be considered and maintained each time you touch the related logic or feature. The agent does not tell you if the amount of details you give in your prompt is sufficient for figuring out an adequate implementation. On the contrary, the LLM will always fill the specification holes with some statistically derived assumptions. Sadly, serious studies on the actual benefits of GenAI in software development are rare. The randomised trial by Metr 19 provides an initial indication, measuring a decline in development speed for experienced developers. An informal study by ThoughtWorks estimates the potential productivity gain from using GenAI in software development at around 5-15% 20. If “AI coding” were increasing programmers’ productivity by any big number, we would see a measurable growth of new software in app stores and OSS repositories. But we don’t, the numbers are flat at best 2122. But even if we assume a productivity increase in coding through GenAI, there are still two points that further diminish this postulated efficiency gain: Firstly, the results of the generation must still be cross-checked by human developers. However, it is well known that humans are poor checkers and lose both attention and enjoyment in the process. Secondly, software development is only to a small extent about writing and changing code. The most important part is discovering solutions and learning about the use of these solutions in their context. Peter Naur calls this ‘programming as theory building’ 23. Even the perfect coding assistant can therefore only take over the coding part of software development. For the essential rest, we still need humans. If we now also consider the finding that using AI can relatively quickly lead to a loss of problem-solving skills 24 or that these skills are not acquired at all, then the overall benefit of using GenAI in professional software development is more than questionable. As long as programming - and every technicality that comes with it - will not be fully replaced by some kind of AI, we will still need expert developers who can programm, maintain and debug code to the finest level of detail. Where, we wonder, will those senior developers come from when companies replace their junior staff with coding agents? Actual vs. promised benefitsPermalink If you read testimonials about the use of GenAI that people perceive as successful, you will mostly encounter scenarios in which ‘AI’ helps to make tasks that are perceived as boring, unnecessarily time-consuming or actually pointless faster or more pleasant. So it’s mainly about personal convenience and perceived efficiency. Entertainment also plays a major role: the poem for Grandma’s birthday, the funny song for the company anniversary or the humorous image for the presentation are quickly and supposedly inexpensively generated by ‘AI’. However, the promises made by the dominant GenAI companies are quite different: solving the climate crisis, providing the best medical advice for everyone, revolutionising science, ‘democratising’ education and much more. GPT5, for example, is touted by Sam Altman, CEO of OpenAI, as follows: ‘With GPT-5, it’s now like talking to an expert — a legitimate PhD-level expert in any area you need […] they can help you with whatever your goals are.’ 25 However, to date, there is still no actual use case that provides a real qualitative benefit for humanity or at least larger groups. The question ‘What significant problem (for us as a society) does GenAI solve?’ remains unanswered. On the contrary: While machine learning and deep learning methods certainly have useful applications, the most profitable area of application for ‘AI’ at present is the discovery and development of new oil and gas fields 26. Harmful aspects of GenAIPermalink But regardless of how one assesses the benefits of this technology, we must also consider the downsides, because only then can we ultimately make an informed and fair assessment. In fact, the range of negative effects of hyperscaled generative AI that can already be observed is vast. Added to this are numerous risks that have the potential to cause great social harm. Let’s take a look at what we consider to be the biggest threats: GenAI is an ecological disasterPermalink PowerPermalink The data centres required for training and operating large generative models 27 far exceed today’s dimensions in terms of both number and size. The projected data centre energy demand in the USA is predicted to grow from 4.4% of total electricity in 2023 to 22% in 2028 28. In addition, the typical data centre electricity mix is more CO2-intensive than the average mix. There is an estimated raise of ~11 percent for coal generated electricity in the US, as well as tripled emissions of greenhouse gases worldwide by 2030 - compared to the scenario without GenAI technology 29. Just recently Sam Altman from OpenAI blogged some numbers about the energy and water usage of ChatGPT for “the average query” 30. On the one hand, an average is rather meaningless when a distribution is heavily unsymmetric; the numbers for queries with large contexts or “chain of reasoning” computations would be orders of magnitude higher. Thus, the potential efficiency gains from more economical language models are more than offset by the proliferation of use, e.g. through CoT approaches and ‘agent systems’. On the other hand, big tech’s disclosure of energy consumption (e.g. by Google 31) is intentionally selective. Ketan Joshi goes into quite some details why experts think that the AI industry is hiding the full picture 32. Since building new power plants - even coal or gas fuelled ones - takes a lot of time, data center companies are even reviving old jet engines for powering their new hyper-scalers 33. You have to be aware that those engines are not only much more noisy than other power plants but also pump out nitrous oxide, one of the main chemicals responsible for acid rain 34. WaterPermalink Another problem is the immensely high water consumption of these data centres 35. After all, cooling requires clean water in drinking quality in order to not contaminate or clog the cooling pipes and pumps. Already today, new data centre locations are competing with human consumption of drinking water. According to Bloomberg News about two-thirds of data-centers that were built or developed in 2022 are located in areas that are already under “water-stress” 36. In the US alone “AI servers […] could generate an annual water footprint ranging from 731 to 1,125 million m3” 37. It’s not only an American problem, though. In other areas of the world the water-thirsty data centers also compete with the drinking water supply for humans 38. Electronic WastePermalink Another ecological problem is being noticeably exacerbated by ‘AI’: the amount of electronic waste (e-waste) that we ship mainly to “Third World” countries and which is responsible for soil contamination there. Efficient training and querying of very large neural networks requires very large quantities of specialised chips (GPUs). These chips often have to be replaced and disposed of within two years. The typical data center might not last longer than 3 to 5 years before it has to be rebuilt in large parts39. In summary, it can be said that GenAI is at least an accelerator of the ecological catastrophe that threatens the earth. And it is the argument for Google, Amazon and Microsoft to completely abolish their zero CO2 targets 40 and replace them with investments of several hundred billion dollars for new data centers. GenAI threatens education and sciencePermalink People often try to use GenAI in areas where they feel overloaded and overwhelmed: training, studying, nursing, psychotherapeutic care, etc. The fields of application for ‘AI’ are therefore a good indication of socially neglected and underfunded areas. The fact that LLMs are very good at conveying the impression of genuine knowledge and competence makes their use particularly attractive in these areas. A teacher under the simultaneous pressure of lesson preparation, corrections and covering for sick colleagues turns to ChatGPT to quickly create an exercise sheet. A student under pressure to get good grades has their English essay corrected by ‘AI’. The researcher under pressure to publish will ‘save’ research time by reading the AI-generated summary of relevant papers – even if they are completely wrong in terms of content 41. Tech companies like OpenAI and Microsoft play on that situation by offering their ‘AI’ for free or for little money to students and universities. The goal is obvious: Students that get hooked on outsourcing some of their “tedious” task to a service will continue to use - and eventually buy - this service after graduation. What falls by the wayside are problem-solving skills, engagement with complex sources, and the generation of knowledge through understanding and supplementing existing knowledge. Some even argue that AI is destroying critical education and learning itself 42: Students aren’t just learning less; their brains are learning not to learn. The training cycle of schools and universities is fast. Teachers are already reporting that pupils and students have acquired noticeably less competence in recent years, but have instead become dependent on unreliable ‘tools’ 43. The real problem with using GenAI to do assignments is not cheating, but students “are not just undermining their ability to learn, but to someday lead.” 44 GenAI is destroying the free internet.Permalink The fight against bots on the internet is almost as old as the internet itself – and has been quite successful so far. Multifactor authentication, reCaptcha, honeypots and browser fingerprinting are just a few of the tools that help protect against automated abuse. However, GenAI takes this problem to a new level – in two ways. To make ‘the internet’ usable as the main source for training LLMs, AI companies use so-called ‘crawlers’. These essentially behave like DDoS attackers: They send tens of thousands of requests at once, from several hundred IPs in a very short time. Robot.txt files are ignored; instead, the source IP and user agent are obscured 45. These practices have massive disadvantages for providers of genuine content: Costs for additional bandwidth. Lost advertising revenue, as search engines now offer LLM-generated summaries instead of links to the sources. This threatens the existence of remaining independent journalism in particular 46. Misuse of own content for AI-supported competition. If the place where knowledge is generated is separated from the place where it is consumed, and if this makes the performance of generation even more opaque than before, the motivation to continue generating knowledge also declines. For projects such as Wikipedia, this means fewer donors and fewer contributors. Open communities often have no other option but to shut themselves off. Another aspect is the flooding of the internet with generated content that cannot be automatically distinguished from non-generated content. This content overwhelms the maintainers of open source software or portals such as Wikipedia 47. If this content is then also entered by humans – often in the belief that they are doing good – it is no longer possible to take action against the methodology. In the long run, this means that less and less authentic training material will lead to increasingly poor results from the models. Last but not least, autonomously acting agents make the already dire state of internet security much worse 48. Think of handing all your personal data and credentials to a robot that is distributing and using that data across the web, wherever and whenever it deems it necessary for reaching some goal. is controlled by LLMs who are vulnerable to all kinds of prompt injection attacs 49. is controlled by and reporting to companies that do not have your best interest in mind. has no awareness and knowledge about the implication of its actions. is acting on your behalf and thereby making you accountable. GenAI is a danger to democracyPermalink The manipulation of public opinion through social media precedes the arrival of LLMs. However, this technology gives the manipulators much more leverage. By flooding the web with fake news, fake videos and fake everything undemocratic (or just criminal) parties make it harder and harder for any serious media and journalism to get the attention of the public. People no longer have a common factual basis, which is necessary for all social negotiations. If you don’t agree on at least some basic facts, arguing about policies and measures to take is pointless. Without negotiations democracy will be dying; in many parts of the world it already is. GenAI versus human creativityPermalink Art and creativity are also threatened by generative AI. The impact on artists’ incomes of logos, images and illustrations now being easily and quickly created by AI prompts is obvious. A similar effect can also be observed in other areas. Studies show that poems written by LLMs are indistinguishable from those written by humans and that generative AI products are often rated more highly 50. This can be explained by a trend towards the middle and the average, which can also be observed in the music and film scenes film scene: due to its basic function, GenAI cannot create anything fundamentally new, but replicates familiar patterns, which is precisely why it is so well received by the public. Ironically, ‘AI’ draws its ‘creativity’ from the content of those it seeks to replace. Much of this content was used as training material against the will of the rights holders. Whether this constitutes a copyright infringement has not yet been decided; morally, the situation seems clear. The creative community is the first to be seriously threatened by GenAI in its livelihood 51. It’s not a coincidence that a big part of GenAI efforts is targeted at “democratizing art”. This framing is completely upside down. Art has been one of the most democratic activities for a very long time. Everybody can do it; but not everybody wants to do put in the effort, the practicing time and the soul. Real art is not about the product but about the process, which requires real humans. Generating art without the friction is about getting rid of the humans in the loop - and still making money. Digital colonialismPermalink The huge amount of data required by hyperscaled AI approaches makes it impossible to completely curate the learning content. And yet, one would like to avoid the reproduction of racist, inhuman and criminal content. Attempts are being made to get the problem under control by subsequently adapting the models to human preferences and local laws through additional ‘reinforcement learning from human feedback (RLHF)’ 52. The cheap labour for this very costly process can be found in the Global South. There, people are exposed to hours of hate speech, child abuse, domestic violence and other horrific scenarios in their poorly paid jobs in order to filter them out of the training material of large AI companies 53. Many emerge from these activities traumatised. However, it is not only people who are exploited in the less developed regions of the world, but also nature: the poisoning of the soil with chemicals during the extraction of raw materials for digital chips, as well as the contamination caused by our electronic waste and its improper disposal, are collateral damage that we willingly accept and whose long-term consequences are currently extremely difficult to assess. Here, too, the “developed” world profits, whereas the negative aspects are outsourced to the former colonies and other poor regions of the world. Political aspectsPermalink As software developers, we would like to ‘leave politics out of it’ and instead focus entirely on the cool tech. However, this is impossible when the advocates of this technology pursue strong political and ideological goals. In the case of GenAI, we can cleary see that the US corporations behind it (OpenAI, Google, Meta, Microsoft, etc.) have no problem with the current authoritarian – some say fascist – US government 54. In concrete terms, this means, among other things, that the models are explicitly manipulated to be less liberal or simply not to generate any output that could upset the CEO or the president 55. Even more serious is the fact that many of the leading minds behind these corporations and their financiers adhere to beliefs that can be broadly described as digital fascism. These include Peter Thiel, Marc Andreessen, Alex Karp, JD Vance, Elon Musk and many others on “The Authoritarian Stack” 56. Their ideologies, disguised as rational theories, are called longtermism and effective altruism. What they have in common is that they consider democracy and the state to be obsolete models, compassion to be ‘woke’, and that the current problems of humanity are insignificant, as our future lies in the colonisation of space and the merging of humans with artificial superintelligence 57. Do we want to give people who adhere to these ideologies (even) more power, money and influence by using and paying for their products? Do we want to feed their computer systems with our data? Do we really want to expose ourselves and our children to the answers from chatbots which they have manipulated? Not quite as abstruse, but similarly misanthropic, is the imminent displacement of many jobs by AI, as postulated by the same corporations in order to put pressure on employees with this claim. Demanding a large salary? Insisting on your legal rights? Complaining about too much workload? Doubts about the company’s goals? Then we’ll just replace you with cheap and uncomplaining AI! Whichever way you look at it, AI and GenAI are already being used politically. If we go along without resistance, we are endorsing this approach and supporting it with our time, our attention and our money. ConclusionPermalink Ideally, we would like to quantify our assessment by adding up the advantages, adding up the disadvantages and finally checking whether the balance is positive or negative. Unfortunately, in our specific case, neither the benefits nor the harm are easily quantifiable; we must therefore consult our social and personal values. Discussions about GenAI usually revolve purely around its benefits. Often, the capabilities of all ‘AI’ technologies (e.g. protein folding with AlphaFold 2) are lumped together, even though they have little in common with hyperscaling GenAI. However, if we consider the consequences and do not ignore the problems this technology entails – i.e. if we consider both sides in terms of ethics – the assessment changes. Convenience, speed and entertainment are then weighed against numerous damages and risks to the environment, the state and humanity. In this sense, the ethical use and further expansion of GenAI in its current form is not possible. Can there be ethical GenAI?Permalink If the use of GenAI is not ethical today what would have to change, which negative effects of GenAI would have to disappear or at least be greatly reduced in order to tip the balance between benefits and harms in the other direction? The models would have to be trained exclusively with publicly known content whose original creators consent to its use in training AI models. The environmental damage would have to be reduced to such an extent that it does not further fuel the climate crisis. Society would have to get full access to the training and operation of the models in order to rule out manipulation by third parties and restrict their use to beneficial purposes. This would require democratic processes, good regulation and oversight through judges and courts. The misuse and harming of others, e.g., through copyright theft or digital colonialism, would have to be prevented. Is such a change conceivable? Perhaps. Is it likely, given the interest groups and political aspects involved? Probably not

      All these factors are achievable I think, or will be soonish. Smaller models, better sourced data sets, niche models, etc. But not with current actors as mentioned at the end.

    1. Irish gov will use their presidency (2nd half of 2026 I think, coming half year is Cyprus, no?) to look into ID-verified social media in the EU. On Mastodon this was posted w t question how it relates to Fediverse. Presumably this will be based on the Digital Services Act, DSA (and GDPR). My current assessment is that DSA hardly applies to fediverse, esp not if there's plenty of federation vs centralisation on a handful of instances. The DSA legislates platforms, not social features, and those features are possible without platforms.

    1. CEPS report on Fediverse wrt DSA, from early 2025. I mistook this one for the IVIR one from 2024 IVIR DSA en Fediverse 2024 in Zotero, which I have remarks about. I have seen this doc, but not sure if I actually took it in.

      Was pointed out to me wrt ID reqs for social media, and if they would apply to fediverse (my first approx is no, if you federate much more agressively, yes if very large instances)

      Relevant wrt DSA too, size and revenue type.

      CEPS DSA en Fediverse 2025 in Zotero

    1. I'm not advocating that everyone should self-host everything. But the pendulum has swung too far toward managed services. There's a large sweet spot where self-hosting makes perfect sense, and more teams should seriously consider it. Start small. If you're paying more than $200/month for RDS, spin up a test server and migrate a non-critical database. You might be surprised by how straightforward it is. The future of infrastructure is almost certainly more hybrid than it's been recently: managed services where they add genuine value, self-hosted where they're just expensive abstractions. Postgres often falls into the latter category. Footnotes They're either just hosting a vanilla postgres instance that's tied to the deployed hardware config, or doing something opaque with edge deploys and sharding. In the latter case they near guarantee your DB will stay highly available but costs can quickly spiral out of control. ↩ Maybe up to billions at this point. ↩ Even on otherwise absolutely snail speed hardware. ↩ This was Jeff Bezos's favorite phrase during the early AWS days, and it stuck. ↩ Similar options include OVH, Hetzner dedicated instances, or even bare metal from providers like Equinix. ↩ AWS RDS & S3 has had several major outages over the years. The most memorable was the 2017 US-East-1 outage that took down half the internet. ↩

      Cloud hosting can become an expensive abstraction layer quickly. I also think there's an entire generation of coders / engineers who treat silo'd cloudhosting as a given, without considering other options and their benefits. Large window for selfhosting in which postgres almost always falls

    2. Storage Tuning: NVMe SSDs make having content on disk less harmful than conventional spinning hard drives, so you'll want to pay attention to the disk type that you're hosted on:

      storage tuning is a selfhosting postgres concern too

    3. Making fresh connections in postgres has pretty expensive overhead, so you almost always want to put a load balancer on front of it. I'm using pgbouncer on all my projects by default - even when load might not call for it. Python asyncio applications just work better with a centralized connection pooler.

      Postgres parallel connections is something you want to stay on top of. load balancing needed

    4. Memory Configuration: This is where most people mess up. Pulling the standard postgres docker image won't cut it. You have to configure memory bounds with static limits that correspond to hardware. I've automated some of these configurations. But whether you do it manually or use some auto-config, tweaking these params is a must.

      Selfhosting Postgres requires to set static limits wrt memory.

    5. When self-hosting doesn't make sense I'd argue self-hosting is the right choice for basically everyone, with the few exceptions at both ends of the extreme: If you're just starting out in software & want to get something working quickly with vibe coding, it's easier to treat Postgres as just another remote API that you can call from your single deployed app If you're a really big company and are reaching the scale where you need trained database engineers to just work on your stack, you might get economies of scale by just outsourcing that work to a cloud company that has guaranteed talent in that area. The second full freight salaries come into play, outsourcing looks a bit cheaper. Regulated workloads (PCI-DSS, FedRAMP, HIPAA, etc.) sometimes require a managed platform with signed BAAs or explicit compliance attestations.

      Sees use for silo'd postgres hosting on the extremes of the spectrum: when you start without knowledge and are vibecoding, so you can treat the database as just another API, and when you are megacorp (outsourcing looks cheaper quickly if you have to otherwise pay multiple FTE salaries otherwise), or/and have to prove regulatory compliance.

    6. The real operational complexity

      Good overview of tasks wrt selfhosting (or rather dedicated server in a non silo datacenter), weekly, monthly, quarterly. Amounts to half a day per quarter at most. You do need to time risky updates etc better, if you run stuff others depend upon, so that you can do incident response

    7. I helped prove this to myself when I migrated off RDS. I took a pg_dump of my RDS instance, restored it to a self-hosted server with identical specs, and ran my application's test suite. Performance was identical. In some cases, it was actually better because I could tune parameters that RDS locks down.

      This reads like what I did wrt Gmail 2014 and Amazon ebooks 2025. [[Leaving a walled garden or silo 20160820203833]] is listing the bits it does and originally improved for you, then rebuild outside the silo based on same components. Not difficult but needs bit of focus.

    8. The value proposition is operational: they handle the monitoring, alerting, backup verification, and incident response. It's also a production ready configuration at minute zero of your first deployment.

      The value proposition is twofold: 1. operational handling (monitoring, backups all that) 2. production ready config out of the box (which feels like fast progress, but also locks you in ofc)

    9. For the most part managed database services aren't running some magical proprietary technology. They're just running the same open-source Postgres you can download with some operational tooling wrapped around it. Take AWS RDS. Under the hood, it's: Standard Postgres compiled with some AWS-specific monitoring hooks A custom backup system using EBS snapshots Automated configuration management via Chef/Puppet/Ansible Load balancers and connection pooling (PgBouncer) Monitoring integration with CloudWatch Automated failover scripting

      AWS RDS is not much else as open source postgres with operational tooling that is not complex itself.

    10. Fast forward to 2025 and I hope the pendulum might be swinging back. RDS pricing has grown considerably more aggressive. A db.r6g.xlarge instance (4 vCPUs, 32GB RAM) now costs $328/month before you add storage, backups, or multi-AZ deployment. For that price, you could rent a dedicated server with 32 cores and 256GB of RAM.

      Amazon database servers have become much more expensive since 2015. You could run a dedicated full server for similar money.

    11. The real shift happened around 2015 when cloud adoption accelerated. Companies started to view any infrastructure management as "undifferentiated heavy lifting"4. Running your own database became associated with legacy thinking. A new orthodoxy emerged of focusing on your application logic and letting AWS handle the infrastructure.

      This post places the shift to hyperscaler dependency in 2015, when (I presume: software) companies began to view any involvement in digital infrastructure management as hassle.

    1. * "Digital platforms are used for hybrid campaigns."* "EU can't compete with US tech ON THEIR TERMS."* "Post-reality US is what happens when tech is unregulated."* "Ireland is a Trojan Horse for Big Tech."* "The Digital Omnibus is sabotage."

      Quotes van [[Defend Democracy o]] event w DK EU presidency cohosting. All convey an aspect of where work is needed. On each I see one could define [[Handelen 20040327155224]] as [[SC landscape van EU Dataspace]] interventions and broader.

      the last one pertains to the AI / GDPR omnibus, not the data one, I think.

    1. Web2.0 seems to have been a mistake and is being rolled back entirely. The whole concept of having your own web property that other people can write to (leave comments and other things) has gone away. That means also there’s no need for a dynamic website with database anymore.

      The second follows from the first, but the first not necessarily true. The commenting thing is bigger than ever, but coopted in the silos to create ad revenue. I have more interaction on my blog again than I had in years, due to the non-silo'd social platforms. I wonder, interaction is only ever needed on newish stuff. I like the speed of static content. So what if I had a mix, latest x posts are dynamic, the rest served statically on the same urls as before. Vgl [[Providing Blog Posts in Plain Text – Interdependent Thoughts 20251229121518]]

    2. Reading back all those old posts and weeknotes I have here is super nice and reminds me:Keeping a record of things is really valuable. Just write and trust that it will come in handy at some point.I used to do so many things in a given week. Compared to what I’m doing now, my life was insanely eventful.I was consistently (too) early on a lot of things. For instance: I read myself complaining about restaurants and food in Amsterdam, something which is mostly solved now.

      Like myself Alper is his own most frequent reader of his blog. Mentions realising how much he did in a week earlier in his life. Same for me, things that now might be a big thing in a week, were Tuesday afternoon 15yrs ago. It's not just age I suspect, but also an overall attenuation that Covid brought?

    1. Our backup nodes are located in Switzerland, and we use AWS (Amazon Web Services).

      Anytype uses Swiss data centers for backup nodes, and uses AWS to get stuff there. Everything is encrypted so in that sense not problematic, but still it means Amazon holds the off-switch.

    1. Media files are not directly downloaded in overall syncing to save bandwidth. Instead, when that file is requested, it is streamed to your device from the backup node or your devices on the network. For example, if you have a 4K Video, it will be streamed from the backup node or P2P devices to your device. So when you open an object with an image, it downloads. When you press play on video & audio, it begins to download. After that, this file will be stored in the application cache.

      media files may not be locally available, and require a internet connection to be streamed/downloaded on demand. Generally excluded from syncing to save bandwidth. Doesn't this also mean that media files aren't backed-up, in the sense that people will treat sync as back-ups.

    1. Pricing for Self-HostersSelf-hosters can manage the limits of Viewers/Editors they invite to their Spaces themselves. Those who would like to purchase a name in the Anytype naming system or access priority support, can purchase a membership at the same price as other beta testers.

      The free tier is tied to Anytype hosting the syncing stuff. Self-hosters do not have limitations. You do not have a name in the Anytype naming system (needed for IPFS, they have a 'private' IPFS network set-up). This is your lock-in right there. Why would you opt-in to that?

    1. Frontmatter (the --- metadata at the top) → becomes object properties.

      Obsidian frontmatter will be changed into properties. This also means that any other data markers, not as frontmatter, will be ignored and treated as regular text.

    2. Step 1: Export from ObsidianClean up your notesConvert embeds like ![[Note Title]] into simple links [[Note Title]]Remove or convert plugin-specific syntax, such as Dataview queries, Templater fields, etc.

      One can import their Obsidian folder into Anytype, but you will lose some information. Embeds don't convert, nor scripting language. I think this is externalising friction. I get that it doesn't convert as functionality, but the convertor should be able to recognise these elements and ignore or adapt them. Most people won't know how to even find all their inclusions e.g.

    1. Pick the plan that works for you now, add more teammates and additional storage anytime you need. Switch or cancel anytime.

      Anytype has paid tiers (4-20/month depending on volume of remote storage, collaborative channels Free tier has 100mb remote storage, 10 collab channels. You can use it locally fully though, w local storage.

    1. Furthermore, the choice of incorporating as an Association rather than an LLC or other legal structure distinguishes ANY as a company managed in accordance with commercial practices, but which exists for the greater good. We consider ourselves a remote team, but we maintain one hub in Berlin (where our team is employed by our local GmbH, a subsidiary of the Swiss Association).

      ANY is a Swiss association as it provides more IP protection and signals a community affinity. All work remote, but have an office in Berlin. The staff in Berlin is employed in a German GmbH, as a subsidiary. So based in B, not CH.

    1. Single objects, infinite possibilities Visualise connections using graph & database views

      Provides several views, as everything is an object, the graphs can make any combi/selection. Graph, table, kanban, gallery. That is however just two basic views (graph and table), unlike e.g. Tinderbox (treemap, network graph, outline, timeline, ecology, landscape)

    2. Nobody can see what’s in your vault, except for you

      locally and on device encrypted. Means you can't access the same material through something other than Anytype? Iow it's not a viewer but a gatekeeper? Runs foul of [[3 Distributed Eigenschappen 20180703150724]] req.

    1. A: The Calibre dedrm plugin only removes the Readium 1.0 drm. As for Readium 2.5, it employs the tougher drm, it cannot be handled with Calibre drdrm plugin.

      Calibre DeDRM plugin can only work w Readium 1 lcpl, current version 2.5 needs diff path.

    2. The DeDRM plugin i

      Calibre DeDRM plugin, but an older version needed to remove Lcpl drm, you do need the passphrase still. For reading in Calibre no need to de-drm but needed if you want to move it to a device that does not support lcp.

    3. All you need to open (or decrypt) LCP eBooks is the account passphrase given to you by the eBook provider - the very same passphrase you'd have to enter into your eBook reader device (once) to read LCP-encrypted books. Sometime, the passphrase can be the passwords.

      .lcpl LCP drmd books have a passphrase. Typically this will be the one for your account from the platform where you bought the book

    1. https://web.archive.org/web/20251228132740/https://www.conferencesthatwork.com/index.php/event-design/2017/01/stories-have-dark-side/ A good story is not per def a true or helpful story. Dark stories can evoke dark passions in audiences. Call to as event organisers to not just platform stories bc they're compelling, but determine if you want to provide a megaphone. Duh. Does lead me to think, e.g. like during epsiplatform, and IndieWeb meetups, can you put some of that down upfront (rather than 'feel' it) as a charter internally or externally (if it doesn't turn the charter into the story). Platforming or not, is choosing to pay attention and [[Aandacht is een morele keuze 20201217074345]]

    1. https://web.archive.org/web/20251228115234/https://www.downes.ca/post/78645

      [[Stephen Downes p]] on an article by Erin Kissane on collective and individual knowledge. Sees parallel with [[Ludwig Wittgenstein h]] 'riverbed propositions' analogy „Man könnte sich vorstellen, daß gewisse Sätze von der Form der Erfahrungssätze erstarrt wären und als Leitung für die nicht erstarrten, flüssigen Erfahrungssätze funktionierten; und daß sich dieses Verhältnis mit der Zeit änderte, indem flüssige Sätze erstarrten und feste flüssig würden. // Die Mythologie [i. e. die unbezweifelten Sätze] kann wieder in Fluß geraten, das Flußbett der Gedanken sich verschieben. Aber ich unterscheide zwischen der Bewegung des Wassers im Flußbett und der Verschiebung dieses; obwohl es eine scharfe Trennung der beiden nicht gibt.“ (Über die Gewißheit § 96 und 97) which distinguishes between the river flow in fixed boundaries, and the shifting of the river itself over time. Vgl [[Waarheid en kennis kent historische periodes 20250914161603]] wrt epistemic periods (Foucault)

    1. drei Erkenntnisse

      Three insights / results from doing it a month 1) small patterns emerge, after 2 weeks. Repeated observations bring them to the fore, and no longer done a way with as incident. The weekly review was key in this. 2) more aware of positive moments (a pattern in itself imo), again weekly review key Vgl #microsucces 3) reflection changed his practice. Small feedback loop was doing something slightly diff the next session, based on the action formulated the previous one. The review served to see the impact of micro-interventions. The reflection provided agency as professional.

    2. Zusätzlich habe ich mir jeden Freitag eine halbe Stunde Zeit genommen für einen wöchentlichen Rückblick. Dabei habe ich die Einträge durchgeblättert, Muster markiert und die wichtigsten Episoden noch einmal verdichtet. Dieser Rückblick half mir besonders, nicht nur auf einzelne Stunden zu reagieren, sondern auch langfristige Entwicklungen zu erkennen, z. B. wiederkehrende Probleme bei Gruppenarbeiten oder Fortschritte in der Diskussionskultur.

      Used a weekly review of 30mins to see patterns across individual entries, and recurring issues in teaching (or rather recurring observations, not the same thing)

    3. Der Zeitaufwand war überschaubar. In der Regel war ich in rund 10 Minuten fertig. Entscheidend war die Routine: Das Journal hatte in meinem Ablauf denselben Stellenwert wie das Aufräumen des Kursraums.

      Vgl [[Gewoonte maak het makkelijk 20201008140324]] mbt template en tijdsbesteding. Anchored journaling in the same frame as clearing the classroom afterwards.

    4. Ich habe mein Journal in einem einfachen A5-Notizbuch geführt. Jede Doppelseite gehörte einer Unterrichtseinheit. Links: die Events in Stichpunkten. Rechts: Episode und Analyse. Der Ablauf sah so aus: Direkt nach der Lektion: fünf Minuten für die Events – die wichtigsten Beobachtungen notieren. Episode auswählen: eine kleine Szene beschreiben, sodass ich sie auch in drei Monaten noch klar vor Augen habe. Analysis: Reflexion über Ursachen und Konsequenzen, verbunden mit einem konkreten nächsten Schritt.

      Did it by hand on paper. A5 note book, two pages for each teaching session. Left for events, right for an single episodic event and reflection on it, ending with a tangible next step

    5. Einfache Struktur: Events – Episode – Analysis Beim Start meiner Challenge habe ich bewusst eine schlanke Form gewählt. In Richards & Farrell bin ich auf ein Modell gestossen, das mir sofort einleuchtete: Events – Episode – Analysis. Events: kurze, stichwortartige Notizen zu den wichtigsten Geschehnissen der Lektion. Episode: eine kleine Szene, die ich erzählerisch ausformuliere: vielleicht eine gute Diskussion, vielleicht ein technisches Problem. Analysis: meine Gedanken dazu: Warum war diese Episode wichtig? Was hat dazu geführt? Und was nehme ich mir für die nächste Einheit vor?

      Example of a simple format that guides you through the reflection, Vgl prompting questions in sensemaking, of voor blogposts.

    6. ein Arbeitsinstrument. Es schafft einen Raum, um nach dem Unterricht innezuhalten, zentrale Ereignisse zu notieren, kleine Episoden festzuhalten und darüber nachzudenken, was gelungen ist – und was nicht.

      A learningjournal prestend here as work tool, a knowledge tool, like an [[AAR after action reviews 20030913131201]].

    7. Donald Schön hat den Begriff „reflection-on-action“ geprägt: Wir lernen, indem wir nach der Handlung darüber nachdenken, warum wir etwas getan haben, und welche Alternativen möglich gewesen wären [2].

      Reflection-on-action, also on other alternatives that would have been possible. Vgl [[Action Research is vraag-reflectief leven 20031215142900]]

      The ref is to D. A. Schön, The Reflective Practitioner: How Professionals Think in Action. New York: Basic Books, 1983. - [ ] zoek boek [[The Reflective Practitioner by Donald A. Schön ]]1983. #pkm

    8. Pädagogische Forschung und Praxis empfehlen es seit Jahren als Werkzeug, um den eigenen Unterricht zu reflektieren und weiterzuentwickeln. Jack Richards und Thomas Farrell definieren es schlicht als „a teacher’s written record of classroom events and of the teacher’s own thoughts and reflections about teaching“ [1].

      Learning journals are seen as useful both in research/theory as in practice. A basic def by Jack Richards and Thomas Farrell is a teacher's record of classroom events, and connected thoughts and reflection on teaching. The ref is to J. C. Richards und T. S. C. Farrell, Professional Development for Language Teachers: Strategies for Teacher Learning. Cambridge: Cambridge University Press, 2005, Kap. 5.

      Would this be useful for mentoring, or even change mgmt consultancy?

    9. 30 Tage lang wollte ich nach jedem Unterricht ein handschriftliches Lehrjournal führen. Kein grosses Projekt, einfach ein Notizbuch und der feste Vorsatz, nach der Lektion zehn bis fünfzehn Minuten zu schreiben. Am Anfang war ich skeptisch, ob sich dieser zusätzliche Aufwand lohnen würde. Heute weiss ich: Es war eine der produktivsten Routinen, die ich mir als Dozent angewöhnt habe.

      As a teacher he kept a journal in a note book to write about a lesson he taught immediately afterwards.

    1. the Language Processing Unit (LPU), a new category of processor. Groq created and built the LPU from the ground up to meet the unique needs of AI. LPUs run Large Language Models (LLMs) and other leading models at substantially faster speeds and, on an architectural level, up to 10x more efficiently from an energy perspective compared to GPUs.

      Groq self describes their core tech as a new cat of processor that run LLMs. Faster speeds. And saves energy an order of magnitude on 'an architectural level' compared to gpu's. That deliberate phrasing suggests there's a trade-off somewhere else.

      Also contrast with FPGAs and their use for on-device ai applications like in [[Vydar wil het Europese bolwerk worden voor navigatie zonder GPS]]

    1. Long existing German NGOs get their bank accounts closed down by German banks based on Trump admin irrational sanctions wrt 'antifa'. bc there's an 'automatic' propagation of such things into the EU. However that is based on a shared notion of what to sanction, which no longer applies. Meaning the outcome is unexplainable in European context. If one end of a negotiated relationship goes off the deep end, you need to realign your processes, which didn't happen here. Can't treat the US as rational actor currently.

    1. Benioff had recently told Business Insider that he's drafting the company's annual strategic document with data foundations—not AI models—as the top priority, explicitly citing concerns about "hallucinations" without proper data context.

      The annual strategic document now puts data foundations in focus, not AI models. Well duh. How even get to the notion that you can AI-all the things, it implies an uncritical belief in the promises of vendors, or magical thinking. How do you get to be CEO if you fall for that. Vibe-leading iow, the wizard behind the curtain.

    2. Phil Mui described as AI "drift" in an October blog post. When users ask irrelevant questions, AI agents lose focus on their primary objectives. For instance, a chatbot designed to guide form completion may become distracted when customers ask unrelated questions.

      ha, you can distract chatbots, as we've seen from the start. This is the classic 'it's not for me but for my mom' train ticket sales automation hangup in response to 'to which destination would you like a ticket', and then 'unknown railway station 'for my mom' in a new guise. And they didn't even expect that to happen? It's an attack service!

    3. Home security company Vivint, which uses Agentforce to handle customer support for 2.5 million customers, experienced these reliability problems firsthand. Despite providing clear instructions to send satisfaction surveys after each customer interaction, The Information reported that Agentforce sometimes failed to send surveys for unexplained reasons. Vivint worked with Salesforce to implement "deterministic triggers" to ensure consistent survey delivery.

      wtf? Why ever use AI to send out a survey, something you probably already had fully automated beforehand. 'deterministic triggers' is a euphemism for regular scripted automation like 'clicking done on a ticket triggers an e-mail for feedback', which we've had for decades.

    4. Chief Technology Officer of Agentforce, pointed out that when given more than eight instructions, the models begin omitting directives—a serious flaw for precision-dependent business tasks.

      Whut? AI-so-human! Vgl 8-bits-schuifregister metafoor. [[Korte termijngeheugen 7 dingen 30 secs 20250630104247]] Is there a chunking style work-around? Where does this originate, token limit, bite sizes?

    5. The company is now emphasizing that Agentforce can help "eliminate the inherent randomness of large models," marking a significant departure from the AI-first messaging that dominated the industry just months ago.

      meaning? probabilities isn't random and isn't perfect. Dial down the temp on models and what do you get?

    6. All of us were more confident about large language models a year ago," Parulekar stated, revealing the company's strategic shift away from generative AI toward more predictable "deterministic" automation in its flagship product, Agentforce.

      Salesforce moving back from fully embracing llms, towards regular automation. I think this is symptomatic in diy enthusiasm too: there is likely an existing 'regular' automation that helps more.

    1. would take seriously the fact that intelligence is now being scaled and distributed through organizations long before it is unified or fully understood

      there's no other way, understanding comes from using it, and having stuff go wrong. The scandals around algos are important in this. Scale and distribution are different beasts. Distribution does not need scale (but a network effect helps) in order to work. The need for scale in digital is an outcome of the financing structure and chosen business model, and is the power grab essentially. #openvraag hoe zet je meer focus op distributie als tegenkracht tegen de schalingshonger van actoren?

    2. examine power as an emergent consequence of deployment and incentives, not intent.

      Intent def is there too though, much of this is entrenching, and much of it is a power grab (esp US tech at the mo), to get from capital/tech concentration to coopting governance structures

      AI is a tech where by design it is not lowering a participation threshold, it positions itself as bigger-than-us, like nuclear reactors, not just anyone can run with it. That only after 3 years we see a budding diy / individual agency angle shows as much. It was only designed to create and entrench power (or transform it to another form), other digital techs originate as challenge to power, this one clearly the opposite. The companies involved fight against things that push towards smaller than us ai tech, like local offline first. E.g. DMA/DSA

    3. Empirical grounding. In 2015, scaling laws, emergent capabilities, and deployment‑driven feedback loops were speculative. Today, they are measurable. That shift changes the nature of responsibility, governance, and urgency in ways that were difficult to justify rigorously at the time.

      States that, in contrast to a decade ago, now we can measure scaling, emergent capabilities, feedback loops. Interesting. - [ ] #30mins #ai-ethics werk dit uit in meer detail. Wat meet je dan, hoe kan dat er uit zien? Hoe vergelijkt dat met div beoordelingsmechanismen?

    4. Political economy and power. The book largely brackets capital concentration, platform dynamics, and geopolitical competition. Today, these are central to any serious discussion of AI, not because the technology changed direction, but because it scaled fast enough to collide with real institutions and entrenched interests.

      geopolitics, whether in shape of capital, tech or politics has become key, which he overlooked in 2015/8

    5. Alignment as an operational problem. The book assumes that sufficiently advanced intelligences would recognize the value of cooperation, pluralism, and shared goals. A decade of observing misaligned incentives in human institutions amplified by algorithmic systems makes it clear that this assumption requires far more rigorous treatment. Alignment is not a philosophical preference. It is an engineering, economic, and institutional problem.

      The book did not address alignment, assumed it would sort itself out (in contrast to [[AI begincondities en evolutie 20190715140742]] how starting conditions might influence that. David recognises how algo's are also used to make diffs worse.

    6. what it feels like to live through an intelligence transition that does not arrive as a single rupture, but as a rolling transformation, unevenly distributed across institutions, regions, and social strata.

      More detailed formulation of Gibson future is already here but distributed. Add sectors/domains. There's more here to tease out wrt my change management work. - [ ] #30mins #ai-ethics vul in met concretere voorbeelden hoe deze quote vorm krijgt

    7. As a result, the debate shifted. The central question is no longer “Can we build this?” but “What does this do to power, incentives, legitimacy, and trust?”

      David posits questions that are all on the application side, what is the impact of using ai. There are also questions on the design side, how do we shape the tools wrt those concepts. Vgl [[AI begincondities en evolutie 20190715140742]] e.g. diff outcomes if you start from military ai params or civil aviation (much stricter), in ref to [[Novacene by James Lovelock]]

    8. The book’s central argument was not about timelines or machines outperforming humans at specific tasks. It was about scale. Artificial intelligence, I argued, should not be understood at the level of an individual mind, but at the level of civilization. Technology does not merely support humanity. It shapes what humanity is. If AI crossed certain thresholds, it would not just automate tasks, but it would reconfigure social coordination, knowledge production, and agency itself. That framing has aged better than I expected, not because any particular prediction came true, but because the underlying question turned out to be the right one.

      The premise of the book that scale mattered wrt AI (SU vibes). AI to be understood at societal level, not from an individual perspective, as tech and society mutually shape eachother (basic WWTS premise). Given certain thresholds it would impact coordination, knowledge and agency.

    1. This article is an example from USA State Dept Legal Office about what [[Matt Gurney We will never fucking trust you again]] mentions wrt loyalists coming and likely staying for a long time, eroding the institution and its credibility

    1. America’s former role is gone. And I think that Americans themselves are having the hardest time of all coming to terms with what that might actually mean in the long run.

      USians will have hard time coming to terms with this. Vgl Bush years where tourists claimed to be Canadian. The ugly American in the Whitehouse etc. With Bush the Lesser there was a return to normal (bc institutions were kept in place), with Trump that road is cut off.

    2. The officer then said that even a swift return of America to its former role won’t matter. Because “we will never fucking trust you again.”The Americans at the table seemed somewhat startled by the heat of that pronouncement. I agreed with it entirely. So, it seemed to me, did most of the non-Americans. This wasn’t the only such moment at the forum this year, but it was, to me, the most interesting. And it was still being talked about the next day. “Thank God,” one allied official said to me. “Someone had to tell them.”

      Whatever happens in the USA in the coming 3 yrs: "We will never trust you again". This has very deep reaching impacts.

    3. But before I could worry about it too much, a senior military officer from a major (non-American) allied nation drove a stake right through the heart of the matter.America has blown 80 years of accumulated goodwill and trust among its allies, our American moderator was told. A rock-steady assumption of allied defence and security planning for literally generations has been that America would act in its own interests, sure, but that those interests would be rational, and would still generally value the institutions that America itself worked so hard to build after the Second World War. America’s recent actions have destroyed the ability of any ally to continue to have faith in America to act even within its own strategic self-interest, let alone that of any ally.

      8 decades of softpower squandered, rationality gone and institutionalised governance dismantled. In short the US cannot even be assumed to act within its own self-interest

    4. And the damage to America’s soft power — the shutting down of aid programs and things like Voice of America — can’t be undone rapidly no matter who wins the midterms. U.S. troops that are pulled out of bases where the U.S. no longer sees a strategic reason for their presence aren’t likely to come back.And, this is the critical part, wouldn’t necessarily be welcomed even if they did.

      Undo many decisions will be impossible

    5. The damage is already done. I know firsthand that a great many Americans who really do believe in the post-1945 global order, and of America’s prior role in the world and the value of that role to America and Americans, are still inside the U.S. government. But I also know that many of them are retiring, or seeking early retirement, or switching to consulting gigs. They can’t stomach what U.S. foreign policy is becoming, and they won’t be a part of it.Good for them. But every single person who departs is being replaced by someone who is totally fine with the new U.S. foreign policy. And sometimes is actually quite enthusiastic about it. That will accelerate the process that’s already underway. And those new people are going to have long careers, shaping things both in public and behind the scenes.

      author calls bs on back-to-normal hopes. Many officials are leaving and get replaced by younger ones who buy into the new US policy and will shape it for decades.

    6. The session, over dinner, was a small group. It was about America’s moral leadership in the world. Our moderator was a now-former American official. She was pretty frank and clear-eyed about how America’s allies currently view the country’s place in the world, but also expressed some hope that after the midterms next year or maybe the next presidential election, things would start to get more back to normal. We were assured that a lot of people in America are still with us. Some of the other Americans present nodded their heads.

      Pre-Trump officials in the US think there's a road back to where the US was before

    7. Amazon or any other U.S.-based company will then make the decision that best promotes and protects long-term shareholder value. And that decision will be, in every case, to submit and comply. Everyone in the room knew that. America is different, now. It’s inescapable.

      klept Gleichschaltung

    8. There were two fascinating things about that exchange (it starts around the 17-minute mark of that video). The first was the question itself; it alone was a signal of how much things have changed. The second interesting thing is that Zapolsky’s answer was, with respect, bullshit. I can see why he’s a legal officer! He gave an answer that was legally correct — the only way that the U.S. government can officially bar Amazon from providing cloud services for a foreign military, for example, would be by sanctions or some comparable legislation.

      The evasive answer is bs bc it isn't how it would go in reality

    9. What would happen is that someone senior at Amazon, maybe Jeff Bezos himself, would get a call from some golf partner or drinking buddy in the administration, and the message will be simple: “Stop, or you won’t get contracts. We’ll arrange some hearings into your operations. Your little spaceflight company will find itself under way more levels of regulatory review than your Musk-owned competitor. This is what the boss wants. Make it happen.”

      mobster governance. Klept

    10. Tarabay dropped a humdinger of a question on Zapolsky. Here’s the quote (slightly cleaned up for clarity): “We’re in an age where there’s a government that puts pressure on companies [and] people for [Trump’s] own gain. You have been so steadfast in your support for Ukraine. What will Amazon do if your government says ‘Stop’?”Zapolsky replied that the company has contracts with foreign governments and NATO allies and said that Amazon would only change those relationships if it was legally forced to do so via something like a sanction.

      Amazon when asked said they would change relationship if legally forced to via e.g. sanctions. Vgl parallel w the same bland avoidance Dutch Stas gave wrt Cloud Act.

    11. Kucher said this to Shaheen: “We’ve talked about allyship. What should the allies, who uphold democratic values, in the reality that the United States has walked away from them … what should the allies do?”

      question by Canadian senator to US senator at Halifax security forum. Premise: USA walked away from democracy and allies

    1. Hypothetically, under 12333 the NSA could target a single foreigner abroad. And hypothetically if, while targeting that single person, they happened to collect every single Gmail and every single Facebook message on the company servers not just from the one person who is the target, but from everyone—then the NSA could keep and use the data from those three billion other people. That’s called 'incidental collection.'

      Example of how EO12333 'can' be used: take all bigtech data as 'incidental' data around a legal foreign intelligence target.

    2. Executive Order 12333 has been regarded by the American intelligence community as a fundamental document authorizing the expansion of data collection activities.[9] The document has been employed by the National Security Agency as legal authorization for its collection of unencrypted information flowing through the data centers of internet communications giants Google and Yahoo!.[9]

      US intelligence see EO12333 as the primary ground for their data collection activities, such as collecting any unencrypted data that flows through bigtech data centers

    3. Part 2.3 permits collection, retention and dissemination of the following types of information along with several others. .mw-parser-output .templatequote{overflow:hidden;margin:1em 0;padding:0 32px}.mw-parser-output .templatequotecite{line-height:1.5em;text-align:left;margin-top:0}@media(min-width:500px){.mw-parser-output .templatequotecite{padding-left:1.6em}}(c) Information obtained in the course of lawful foreign intelligence, counterintelligence, international narcotics or international terrorism investigation ... (i) Incidentally obtained information that may indicate involvement in activities that may violate federal, state, local or foreign laws[1]

      EO12333 in part 2.3 permits the ability for collection / retention and sharing of any data obtained during lawful intelligence / international law enforcement

      and any other data that may indicate violate a law

    1. Amerikaanse wetgeving zoals de CLOUD Act (Clarifying Lawful Overseas Use of Data Act), de Foreign Intelligence Surveillance Act (FISA)5 en Executive Order 12333 komt te vallen?
      • [ ] Naast Cloud Act ook FISA sect 702 en EO 12333 even expliciet opslaan. #digitalsovereignty #geonovumtb