16 Matching Annotations
  1. Last 7 days
    1. statistical knowledge is still required in order toformulate the correct prompts and to ensure that the AIdoes not leave out any step of the analysis.

      rhetoric: author presents a prescriptive claim that AI needs humans with competent knowledge (in this case, statistics) to create prompts and ensure that the AI does not leave out any steps of the analysis. He positions domain knowledge not as a tool for using AI for statistical analysis, but a prerequisite for management of the AI and auditing the output.

      inference: In addition to policing and correcting the AI outputs, the deep domain knowledge is what allows the AI to do complex data analysis without mistakes, hallucinated results, or mathematically false outcomes. This is basically the job description of a human with "Augmented Human Wisdom". The human's value is no longer in doing math, but in possessing the vertical expertise (flesh/wisdom) to know exact what math needs to be done and ultimately auditing the assistant machine's work.

    2. ChatGPT Data Analyst clearly produced a false resulthere, precisely because the application assumptions for theANOVA were not checked.

      rhetoric: Schwarz employs cause-and-effect reasoning here based on empirical testing. He links a specific technical failure (not checking assumptions) to a definitive unwanted outcome (a false result).

      inference: the "Data Analyst" function of ChatGPT hallucinated a result during the use of it's core function! This is the best evidence so far of the 'Crisis of Truth' and the dangers of the 'Headless Automatons' in my essay. If a generalist with no deep knowledge uses AI, they are at great risk of blindly accepting mathematically false conclusions. Synthetic syntax without competent human validation is a liability.

    3. The results show that generative AI canfacilitate data analysis for individuals with minimal knowledge of statistics,mainly by generating appropriate code, but only partly by following standardprocedures.

      rhetoric: author uses comparative, objective statement (logos) to establish the main boundary of the technology's capability/capacity -- it excels at technical generation (things like coding) but fails at standard procedures (methodological adherence to SOPs).

      inference: the proves the 'Raising the Floor' concept. AI completely automates the entry-level syntax (the "Word"), meaning that the Generalist coder is obsolete! However, because it fails at standard procedures, it requires a human architect to guide it to outputs that are valuable in the real world.

    1. Perseverations that are input into the system are essentially mag-nified by the system’s suggested sentences,

      rhetoric: authors explain an unintended consequence of using the AI tool: it scales the errors or the emptiness of the human prompt.

      inference: this is an excellent metaphor for the 'manager fallacy'. If the human user in incompetent (or provides empty or incomplete input), the AI does not magically create wisdom -- it just amplifies the user's incompetence in a a highly articulate synthetic thought.

    2. Participant 2 stated the age of her daughters (“Name1 is 18, Name2 is21”), Aphasia-GPT transformed it as “Name1 is 18 and 21”, which is an impossible, butrelated, hallucination

      rhetoric: researchers use a specific, clinical observation of an error to demonstrate the model's inability to comprehend logical reality despite the human relaying a perfectly structured sentence.

      inference: this shows that AI is amoral and lacks the lived experience necessary to make logical judgments that work in the real world. It can format a sentence beautifully, but it does not/will not always understand that a single human cannot be two ages at once. This is why it is very important/necessary for the "flesh" to text the output against reality

    Tags

    Annotators

    1. The cost of the time that it takes fix "workslop" could add up too, with a $186 monthly cost per employee on average, according to a survey of desk workers by BetterUp in partnership with the Stanford Social Media Lab. Forty percent of the workers surveyed said they received "workslop" in the last month and that it took an average of two hours to resolve each incident.

      $186/per employee/per month!

      10 employees = ($22,320) 25 employees = ($55,800) 50 employees = ($111,600) 100 employees = ($223,200) 250 employees = ($558,000) 500 employees = ($1,116,000) 1000 employees = ($2,232,000)

    2. “Younger workers aren’t necessarily more careless, but they’re often using AI more frequently and earlier in their workflows," Dennison said. "There is also a training gap. Organizations often assume younger employees intuitively understand AI, yet provide little guidance on verification, risk, or appropriate use cases. As a result, AI may be treated as an answer engine rather than a support tool."

      this is another great quote, which helps to establish how orgs treat younger generations, and how they tend to overtrust their understanding of AI.

    3. 58% said direct reports submitted work that contained factual inaccuracies generated by AI tools, while fewer reported that AI failed to account for critical contextual factors. Other issues cited include low-quality content, poor recommendations and inappropriate messaging.

      from reporting managers, 58% of them said that employees were submitting work that contained factual inaccuracies in the work that was generated by AI, and that fewer of them reported that AI failed to account for "critical contextual factors", implying that the writing was generic and not directly applicable to the context that the writing was written in. Other issues were: low quality content, poor recommendations and inappropriate messaging.

    4. 59% of managers saying that they had to invest additional time to correct or redo work created by AI. Similarly, 53% said their direct reports had to take on extra work, while 45% said they had to bring in co-workers to help fix the mistake.

      Extra time and money spent to repair errors made by AI but not caught by the human in the middle. 59% is almost 2/3 (closer to 3/5) needed to correct or redo the work created by AI without a human auditing it. 53% claim extra work is needed to repair the AI mistakes, and 45% also needed to bring in a (perhaps more senior) co-worker to help fix the mistake. I can imagine workers needing to work on a mistake the hits production code, and all of the thousands (or more) mistakes that would need to be later repaired and rolled back. very expensive and costly.

    5. While 18% of managers said they did not suffer any financial losses from the mistakes, and 20% said those losses were less than $1,000, a significant number reported bigger losses. Twelve percent said those losses were more than $25,000, while 11% said between $10,000 and $24,999. Another 27% placed the value of those losses above $1,000 but below $10,000.

      great stats for the cost of using AI without human auditing.

    6. “AI is reliable when used as an assistant, not a decision-maker," Dennison said. "Without human judgment and clear processes, speed becomes a risk, and efficiency gains can turn into costly mistakes,”

      great quote. directly mentions my concept of requiring human judgement, and how not having a human in the loop can make work move faster, but can also lead to very costly mistakes.

    7. “Employees treat AI outputs as finished work rather than as a starting point. Current AI tools are very good at generating fluent content, but they don’t understand context, business nuance, risk, or consequences. That gap shows up in factual errors, missing constraints, poor judgment calls, and tone misalignment.”

      another great quote -- ties into the abdicating human agency to a robot, and the full quote even illustrates the dangers of doing so.