37 Matching Annotations
  1. Last 7 days
    1. Takeaways This list isn’t comprehensive. I’m still experimenting and would love to learn from your experiments as well.

      I don't feel convinced by specfically the naming of these roles it seems, and also don't per se find them very amanuensis like. The amanuensis / assistant frame is a useful one as such (not just for AI, but also for thinking up new [[Personal Software]] for [[Mijn personal tools list]].

    2. 9. Reflector This final role is different. Whereas the others took as the object of inquiry a particular work — e.g., a novel or a movie — this last one takes as the object your knowledge garden itself. That is, you point the LLM to a series of notes to analyze patterns over time and suggest improvements. Example: I fed all 52 weekly posts from my humanities crash course to Claude Code, and asked it to identify the various roles in which I used AI for learning throughout the year. Its answers — with some curation from me — are the roles you just read. Suggested prompt: Here are my notes from [X weeks/months] of reading on [TOPIC]. What patterns do you notice in what I pay attention to? What do I seem to find most interesting, and what do I seem to avoid or underweight?

      Role 9 Reflector, give it a bunch of your own notes to analyze patterns. Not sure it differs much of the Connector/Analyst roles other than the object of inquiry being your own notes. I thought of doing this for my blog in one of the earlier roles just now.

    3. 8. Mapper This one’s a bit more esoteric. Some people — me included — are primarily visual: diagrams and drawings aid our understanding. Concept maps can be especially helpful. I’ve built an Agent Skill to allow LLMs like Claude draw concept maps. (Download it from Github.) Example: I used this mapping skill to generate a concept map of Virginia Woolf’s To the Lighthouse. It’s not especially insightful, but more of a proof point of using LLMs in a more visual modality. Suggested prompt: (Note: install my LLMapper Skill before issuing this prompt) Generate a concept map for [WORK] centered on the question: “How does the novel’s treatment of [THEME] illuminate [BROADER QUESTION]?”

      Role 8 Mapper. Interesting role, though I wonder if the friction in making concept maps is actually the work to be done here by yourself. Getting a mapping exercise ready (elements that likely need to be on the map, feeding it my [[Systems Convening by Etienne and Beverly Wenger-Trayner]] mapping elements library) I think would be useful, and apply my Excalidraw template to it e.g. More amanuensis like too, I think.

    4. 7. Analyst This role will also help you appreciate a work from a different perspective. It’s easy: you ask for the LLM to apply a specific critical lens to a reading. Common lenses include Freudian, Marxist, feminist, Girardian, etc. Example: The same week I read Freud, my son and I watched Predator, the 1980s sci fi film starring Arnold Schwarzenegger. For fun, I asked ChatGPT to analyze the film through a Freudian lens. The result was both enlightening and hilarious. Suggested prompt: Apply a [Marxist / feminist / postcolonial / Jungian] reading to [WORK]. What does this lens reveal that a neutral summary would miss?

      role 7 analyst. The description is not analysis in the data/argument sense, but interpretative more like. Vgl [[Filosofische stromingen als gereedschap 20030212105451]] taking a different perspectives on a question to bring thinking further.

    5. 6. Adversary Here’s a fun role: asking for an LLM to push back on your position or steelman the opposing point of view. The idea is to expand your understanding by bringing your assumptions to the surface and challenging them. Example: After watching Modern Times, I asked ChatGPT to correct my understanding of the movie as a work of Marxist propaganda. The LLM convinced me that the film is in fact more of a humanist statement than a political one. As a result of this interaction, I changed my mind on Chaplin’s work. Suggested prompt: Here are my notes on [TOPIC]. Please help me see it through the lens of someone who might be sympathetic to [OPPOSING POSITION] without fully realizing it. What could I improve? Where is my argument weakest? [paste notes]

      Role 6 Adversary. To challenge assumptions, better understand opposing views. This is a very interesting role. Having a debater, not as performance, but to deepen knowledge

    6. 5. Recommender This is a useful role for deepening your understanding of a subject: asking for related works that reflect similar themes. It’s also a use case where I noticed considerable improvements in LLM performance over 2025. Example: Early in 2025, I read Confucius’s Analects. Perplexity was ahead in web-backed interactions at the time, so I asked it for a list of classic Chinese movies that reflected Confucian values. It responded with five suggestions, some of which it hallucinated. But one of them, Spring in a Small Town, was a bona fide classic — and I likely wouldn’t have learned of it without an LLM. (Later in the year, other chatbots gained this ability and hallucinations dropped across the board.) Suggested prompt: I just finished [WORK]. Recommend three films that explore similar themes or ideas. Prioritize films with strong critical reputations — I’d rather have one great recommendation than five mediocre ones.

      Role 5 recommender, described as recommending works to deepen one's understanding. The example to me is more about finding more superficial things to see content in a different shape again (here films, podcasts before), a broadening. Perhaps to get a more emotional tie in with a concept, bringing it into scope of one's perception of beauty, next to K as such?

    7. 4. Orienter This role is something of an inversion of the validator. Instead of asking for feedback on your notes after reading a text, here you ask the AI for guidance before reading. You’re looking for framing, historical context, high level outlines, etc. — ideally, without spoilers. Example: Before reading Nietzsche’s Beyond Good and Evil and Tolstoy’s The Death of Ivan Illych, I uploaded both books to NotebookLM, which created a podcast for me that explained their thematic contexts. Listening to this podcast in my daily walk helped me better understand the readings. Suggested prompt: I’m about to read [WORK] for the first time. Give me enough context to make sense of it — historical background, key arguments, things to watch for — but don’t spoil the experience of discovering it myself.

      Role 4 Orientor, asking about works' meaning upfront as prep for one's own reading. As inversion of the validator in role 2. The example is about giving something a different form for consumption (comparison of works as podcast). NotebookLM used.

    8. 3. Connector Here’s yet another role you can easily do via chat: identifying thematic, philosophical, or narrative parallels between works. Note I wrote “works” — it’s fun and illuminating to ask for connections across media, genre, time, etc. Example: I watched Francis Ford Coppola’s The Conversation on the same week I read Oedipus Rex. For fun, I asked ChatGPT for possible parallels between the two works. Its reply was enlightening: it pointed out how the protagonists of both stories undertook an obsessive investigation that uncovered terrible knowledge. Suggested prompt: I’ve been reading [WORK A] and [WORK B]. What philosophical or thematic threads connect them? I’m looking for non-obvious resonances, not surface similarities.

      Role 3 connector, also chat based. Connector seems a generic term (and in general, wrt [[Netwerkleren Connectivism 20100421081941]] a own brain effort), but the example is more about syntopic readng vgl [[Gebruik AI om podcasts syntopisch samen te vatten 20260306123338]]

    9. 2. Validator Another basic role for AI is validating your understanding. To do this, you ask it to review your notes for errors or gaps, do basic fact checking, or critique your reasoning. Again, you can do this via the chat interface, but I also experimented with passing my notes in Obsidian using the Copilot plugin and in Emacs using gptel. Example: After reading The Epic of Gilgamesh, I wrote a note in Obsidian summarizing its plot. When I asked ChatGPT to critique my summary, it pointed out that I’d given the central character a redemption arc that isn’t present in the text. I’m so accustomed to the standard hero’s journey, that I projected it onto the book — and an LLM helped me correct this ‘hallucination.’ Suggested prompt: Here are my notes on [WORK]. What important ideas did I miss or underemphasize? Don’t rewrite my notes — just flag the gaps.

      Role 2 validator of one's understanding, also seen as basic. Might be a good complement to e.g. turning some of my notes into [[Anki]] card decks or combine in another way w spaced repetition. [[Spaced repetition 20201012201559]] [[Connecting my PKM to Anki]]

    10. 1. Tutor The simplest role for AI is as a tutor. You ask it to explain a difficult concept, clarify a confusing passage, translate jargon, etc. I mostly did this via the standard chat UI (although I created a ChatGPT project to preserve context for the course.) Example: While reading Freud’s The Interpretation of Dreams, I came across three unfamiliar German terms: es, ich, and über-ich. ChatGPT helpfully explained these are more commonly known as id, ego, and superego — three terms I already understood. Suggested prompt: I just read [PASSAGE]. I understand [X] but I’m confused about [Y]. Can you explain [Y] in plain terms, without assuming I have background in [FIELD]?

      Role 1 as Tutor, simplest role. Ask a chatbot for clarification. I think this skips a bit of exploration (wikipedia as jumping off point e.g.), but it is also much more contextual and specific. Includes translation of concepts. You could run this locally I think, and as Jorge states, create a bit of persistent context for it.

    11. It was a messy process. That’s what you do in a garden! And the outcome wasn’t an enthusiastic endorsement of AI. Instead, I landed at a map of roles and modalities for how AI can help at different points in the spectrum. Let’s look at nine of these roles.

      there are more than 9 it seems. Perhaps check his blog over the year to see what else. Says process was messy, bc yes garden, and implies mixed results.

      Quick glance at the 9 roles I don't see all of them as fitting the amanuensis metaphor imo

    12. Robots in the garden

      Arango tried it out on major texts (reminiscent of the original version of [[How to Read a Book The Ultimate Guide by Mortimer Adler]], not the 2nd edition. ) Over a year he came to define 9 roles for the robots in his garden.

  2. Mar 2026
  3. Feb 2026
  4. Jan 2026
    1. why asynchronous agents deserve more attention than they currently receive, provides practical guidelines for working with them effectively, and shares real-world experience using multiple agents to refactor a production codebase.

      3 things in this article: - why async agents deserve more attention - practical guidelines for effective deployment - real world examples

    1. While the initial results fall short, the AI field has a history of blowing through challenging benchmarks. Now that the APEX-Agents test is public, it’s an open challenge for AI labs that believe they can do better — something Foody fully expects in the months to come.

      expectation that models will get trained against the tests they currently fail.

    2. “The way we do our jobs isn’t with one individual giving us all the context in one place. In real life, you’re operating across Slack and Google Drive and all these other tools.” For many agentic AI models, that kind of multi-domain reasoning is still hit or miss.

      I understand this para but the phrasing is off. slack and google drive is not 'multi-domain' but tools. Seems like two arguments joined up: multitool / multidomain, meaning ai agents can't switch. (In practice I see people build small agents for each facet and then chain / join them)

    3. The new research looks at how leading AI models hold up doing actual white-collar work tasks, drawn from consulting, investment banking, and law. The result is a new benchmark called APEX-Agents — and so far, every AI lab is getting a failing grade. Faced with queries from real professionals, even the best models struggled to get more than a quarter of the questions right. The vast majority of the time, the model came back with a wrong answer or no answer at all.

      In consulting, investment banking, law, ai agents had 18-24% score or worse (and in real life circumstances you don't know which is which so you need to check all output)

  5. Dec 2025

    Tags

    Annotators

    URL

    1. this type of thing sounds like what I thought wrt annotation of [[AI agents als virtueel team]]. The example prompts of questions make me think of [[Filosofische stromingen als gereedschap 20030212105451]] die al per stroming een vraagstramien bevat. Making persona's of diff thinking styles, lines of questioning. Idem for reviews, or starting a project etc.

    1. Het zijn markdown bestanden met een persoonlijkheid, frameworks, en output templates. Die heb ik niet zelf geschreven - ik heb Claude gevraagd om ze te maken. “Maak een Product Owner agent die goed is in prioriteren en impact/effort analyses kan doen.” Claude schrijft dan het volledige bestand, inclusief werkwijze en voorbeelden.Als ik vervolgens zeg “vraag dit aan Tessa”, laadt Claude dat bestand en wordt Tessa.

      Seems like these agent .md files contain description of a role that is then included in a prompt.

    1. In mijn werkmap heb ik een verzameling “agents” - tekstbestanden die Claude vertellen hoe hij zich moet gedragen. Tessa is er één van. Als ik haar “laad”, denkt Claude vanuit het perspectief van een product owner.

      Author has .md files that describe separate 'agents' she involves in her coding work, for each of the roles in a dev team. Would something like that work for K-work? #openvraag E.g. for project management roles, or for facets you're less fond of yourself?

  6. Nov 2025
    1. AI checking AI inherits vulnerabilities, Hays warned. "Transparency gaps, prompt injection vulnerabilities and a decision-making chain becomes harder to trace with each layer you add." Her research at Salesforce revealed that 55% of IT security leaders lack confidence that they have appropriate guardrails to deploy agents safely.

      abstracting away responsibilities is a dead-end. Over half of IT security think now no way to deploy agentic AI safely.

  7. Jun 2025
    1. https://web.archive.org/web/20250630134724/https://www.theregister.com/2025/06/29/ai_agents_fail_a_lot/

      'agent washing' Agentic AI underperforms, getting at most 30% tasks right (Gemini 2.5-Pro) but mostly under 10%.

      Article contains examples of what I think we should agentic hallucination, where not finding a solution, it takes steps to alter reality to fit the solution (e.g. renaming a user so it was the right user to send a message to, as the right user could not be found). Meredith Witthaker is mentioned, but from her statement I saw a key element is missing: most of that access will be in clear text, as models can't do encryption. Meaning not just the model, but the fact of access existing is a major vulnerability.