343 Matching Annotations
  1. Feb 2023
    1. However, the article does not take a clear stance on the matter and does not offer a conclusion on whether the capitalization of the word "black" is a good or bad thing.

      This implies that it should take a stand but doesn't say why. Note that the New York Times article is not an editorial.

      This summary also misses the stand implied by the choice to end on a quote from scholar Crystal Fleming explaining why she capitalizes "Black":

      "'Frankly,” she said, “because I want to. That’s also something that I think we need to see more of in every field — black people doing what we want to do.'"

    2. long overdue

      Why? What value do these people see in the capitalization? This summary ignores the idea of power, which is central in the article. Eligon writes that capitalizing "Black" "confers a sense of power and respect to black people, who have often been relegated to the lowest rungs of society through racist systems, black scholars say."

    3. John Eligon, a New York Times columnist, writes about the debate surrounding the capitalization of the word "black" to refer to African-Americans. The move to capitalize the word has gained widespread acceptance in recent weeks and has sparked a debate over identity, race, and power. Many news organizations have changed their style to capitalize the word, but not all have embraced it, including The New York Times and The Washington Post, who are still grappling with the decision. The capitalization of black is part of a long-standing struggle to refer to people of African descent in the most respectful and accurate way.

      Here's a sample ChatGPT critical assessment of the NY Times article at https://www.nytimes.com/2020/06/26/us/black-african-american-style-debate.html

      For contrast, see this human-written sample essay from the textbook How Arguments Work: A Guide to Writing and Analyzing Texts in College: https://human.libretexts.org/Bookshelves/Composition/Advanced_Composition/Book%3A_How_Arguments_Work_-A_Guide_to_Writing_and_Analyzing_Texts_in_College(Mills)/04%3A_Assessing_the_Strength_of_an_Argument/4.11%3A_Sample_Assessment_Essays/4.11.02%3A_Sample_Assessment-_Typography_and_Identity

  2. platform.openai.com platform.openai.com
    1. Educator considerations for ChatGPT<br /> https://platform.openai.com/docs/chatgpt-education

      <small><cite class='h-cite via'> <span class='p-author h-card'>Hypothesis</span> in Liquid Margins 38: The rise of ChatGPT and how to work with and around it : Hypothesis (<time class='dt-published'>02/09/2023 16:11:54</time>)</cite></small>

    2. upskilling activities in areas like writing and coding (debugging code, revising writing, asking for explanations)

      I'm concerned people will see this and remember it without thinking of all the errors that are described later on in this document.

    3. ChatGPT use in Bibtex format as shown below:

      Glad they are addressing this, and I hope they will continue to offer such suggestions. I don't think ChatGPT should be classed as a journal. We really need a new way to acknowledge its use that doesn't imply that it was written with intention or that a person stands behind what it says.

    4. will continue to broaden as we learn.

      Since there is a concern about the bias of the tool toward English and developed nations, it would be great if they could include global educators from the start.

    5. As part of this effort, we invite educators and others to share any feedback they have on our feedback form as well as any resources that they are developing or have found helpful (e.g. course guidelines, honor code and policy updates, interactive tools, AI literacy programs, etc).

      I wonder how this information will be shared back so that other educators can benefit from it. I maintain a resource list for educators at https://wac.colostate.edu/repository/collections/ai-text-generators-and-teaching-writing-starting-points-for-inquiry/

    6. one factor out of many when used as a part of an investigation determining a piece of content’s source and making a holistic assessment of academic dishonesty or plagiarism.

      It's still not clear to me how they can be used as evidence at of academic dishonesty at all, even in combination with other factors, when they have so many false positives and false negatives. I can see them used to initiate a conversation with a student and possibly a rewrite of a paper. This is tricky.

    7. Ultimately, we believe it will be necessary for students to learn how to navigate a world where tools like ChatGPT are commonplace. This includes potentially learning new kinds of skills, like how to effectively use a language model, as well as about the general limitations and failure modes that these models exhibit.

      I agree, though I think we should emphasize teaching about the limitations before teaching how to use the models. Critical AI literacy must become part of digital literacy.

    8. Some of this is STEM education, but much of it also draws on students’ understanding of ethics, media literacy, ability to verify information from different sources, and other skills from the arts, social sciences, and humanities.

      Glad they mention this since I am skeptical of claims that students need to learn prompt engineering. The rhetorical skills I use to prompt ChatGPT are mainly learned by writing and editing without it.

    9. While tools like ChatGPT can often generate answers that sound reasonable, they can not be relied upon to be accurate consistently or across every domain. Sometimes the model will offer an argument that doesn't make sense or is wrong. Other times it may fabricate source names, direct quotations, citations, and other details. Additionally, across some topics the model may distort the truth – for example, by asserting there is one answer when there isn't or by misrepresenting the relative strength of two opposing arguments.

      If we teach about ChatGPT, we might do well to showcase examples of these kinds of problems in output so that students develop an eye for them and an intuitive understanding that the model isn't thinking or reasoning or checking what it says.

    10. While the model may appear to give confident and reasonable sounding answers,

      This is a bigger problem when we use ChatGPT in education than in other arenas because students are coming in without expertise, seeking to learn from experts. They are especially susceptible to considering plausible ChatGPT outputs to be authoritative.

    11. . Web browsing capabilities and improving factual accuracy are an open research area that you can learn more in our blog post on WebGPT.

      Try PerplexityAI for an example of this. Google's Bard should be another example when released.

    12. subtle ways.

      Glad they mention this in the first line. People will see the various safeguards and assume that ChatGPT is safe because work has been done on this, but there are so many ways these biases can still surface, and since they are baked into the training data, there's not much prospect of eliminating them.

    13. Verifying AI recommendations often requires a high degree of expertise,

      This is a central idea that I would wish were foregrounded. If we are trying to use auto-generated text in a situation in with truth matters, we need to be quite knowledgeable and also invest time in evaluating what that text says. Sometimes that takes more time than writing something ourselves.

    14. students may need to develop more skepticism of information sources, given the potential for AI to assist in the spread of inaccurate content.

      It strikes me that OpenAI itself is warning of a coming flood of misinformation from language models. I'm glad they are doing so, and I hope they keep investing in improving their AI text classifier so we have some ways to distinguish human writing from machine-generated text.

    15. Educators should also disclose the use of ChatGPT in generating learning materials, and ask students to do so when they incorporate the use of ChatGPT in assignments or activities.

      Yes! We must begin to cultivate an ethic of transparency around synthetic text. We can acknowledge to students that we might sometimes be tempted to autogenerate a document and not acknowledge the role of ChatGPT (I have certainly felt this temptation).

    16. export their ChatGPT use and share it with educators. Currently students can do this with third-party browser extensions.

      This would be wonderful. Currently we can use the ShareGPT extension for this.

    17. they and their educators should understand the limitations of the tools outlined below.

      I appreciate these cautions, but I'm still concerned that by foregrounding the bulleted list of enticing possibilities, this document will mainly have the effect of encouraging experimentation with only lip service to the cautions.

    18. custom tutoring tools

      I'm concerned that any use of ChatGPT for tutoring would fall under the "overreliance" category as defined below. Students who need tutoring do not usually have the expertise or the time to critically assess or double check everything the tutor tells them. ChatGPT already comes off as more authoritative than it is. It will come across as even more authoritative if teachers are recommending it as a tutor.

    1. ChatGPT could be used as a writing prompt for writers to leverage for their work in much the same way that [[Benjamin Franklin]] rewrote existing works or the major plot point in the movie [[Finding Forrester]] in which Jamal used William's work as a springboard for his own.

      Link to: https://hypothes.is/a/HPQLinKXEemyqafW9xlIFQ.

    1. Scaling neural network models—making them bigger—has made their faux writing more and more authoritative-sounding, but not more and more truthful.

      Yes -- distinguishing the more realistic from more truthful. That's where the conversation should be.

    1. Why are people so quick to be impressed by the output of large language models (LLMs)?

      This take-down isn't actually address this question. It's using it as a dismissal.

      It is a good question though and one not to be dismissed as its causes might interrogated.

      I am impressed (while also skeptical of ChatGPT). Does that make me dumb?

    1. So will AI text generation tools revolutionize or kill college writing? Both! Neither! For sure! Probably! Eventually! Somewhat! It’s…complicated.

      Nice summary of the discourse on ChatGPT!

    2. e-Literate isn’t about what I know. It’s about what I’m learning.

      There's an interesting point to be made about process here. Can the same be said for course work: that writing for a class isn't about what you know it's about what you are learning.

    3. Particularly if used judiciously as part of the writing curriculum rather than the whole thing, it could be quite useful.

      Very sensible.

    4. students are heavily influenced by whether they believe their teacher cares about their learning.

      Making writing more of a process rather than a product, a process in which the teacher gives regular feedback to the student, would help build that relationship.

    5. Then I would have edited the output

      Interesting. Collaborating with the bot in composition. It gets you started, but you are still needed.

    1. ChatGPT doesn’t mark the end of high school English class, but it can mark the end of formulaic, mediocre writing performance as a goal for students and teachers. That end is long overdue, and if ChatGPT hastens that end, then that is good news.

      Provocative argument: ironically, it's the standardization of learning that is killed by AI writing platforms.

    2. Both started with a version of “Work A and Work B have many similarities and many differences,” an opening sentence that I would have rejected from a live student

      So what's the point, ChatGPT isn't really all that sophisticated in its analysis? Relies on cliched structures? Either way or both, I kind of buy it. It's not a creative writer. It' utilitarian.

      There's also an interesting point to be made here in terms of the prompts teachers provide students for essays. They too need to be sophisticated rather than simply compare and contrast these two books.

    3. If they put a great degree of thought into designing a prompt, would that not mean that they were doing something involving real learning?

      Yes!

    4. I suspect that test runs with ChatGPT depend in part on the richness of the prompt given,

      Writing good prompts could be something we teach students.

    5. And the algorithm cannot manage supporting its points with quotes from the works, a pretty fundamental part of writing about literature.

      ChatGPT not good at integration of quotes, a key piece of writing from evidence.

    1. He said it was “very naive” to think it would be possible to impose restrictions on internet platforms, particularly with Microsoft primed to integrate AI into its search engine, Bing.“Are you going to ban Google and Bing?”

      Fair point.

    1. At the same time, we need to continue building activities and assessments to make classroom work more specific and experiential.

      Yes! Not sure that means banning AI as a tool which this essay ends up arguing.

    2. Pedagogically speaking, focusing on the grunt work of trying out ideas—watching them develop, wither, and cede ground to better ones—is the most valuable time we can spend with our students. We surrender that time to Silicon Valley and the messy database that is the internet at the peril of our students.

      This turns into a very traditional argument of the don't use Wikipedia variety.

    3. digital utopians might claim that students and teachers will have more opportunities for critical thinking because generating ideas—the grunt work of writing—isn’t taking up any of our time. Along this line of thinking, ChatGPT is just another calculator, but for language instead of numerical calculation.

      I'm still compelled by this idea TBH...

    1. Analysis of recent events not in the training data for the system.

      Wouldn't analysis and commentary on recent events be readily available on the Internet?

    2. Note that ChatGPT can produce outputs that take the form of  “brainstorms,” outlines, and drafts. It can also provide commentary in the style of peer review or self-analysis. Nonetheless, students would need to coordinate multiple submissions of automated work in order to complete this type of assignment with a text generator.

      Interesting. It almost takes MORE work to use ChatGPT in the context of such heavily scaffolded writing process,

    3. get a better sense of their thinking

      And if we're reading more of their writing through social annotation or other "steps" in the process, we also become familiar with their thinking.

    4. a process that empowers critical thinking

      Yes, I've never felt I was simply teaching writing when I taught composition. Writing was a visible end product of a lot of other work (reading, thinking, and non-summative pre-writing activities) that I was training students in.

    5. students who feel connected to their writing will be less interested in outsourcing their work to an automated process.

      Love this idea. Teaching students to own and enjoy their writing.

    6. skip the learning and thinking around which their writing assignments are designed.

      Or does it focus the learning? Just as I don't really care if my students know how to spell as long as they use spell check, what does writing with ChatGPT open up in terms of enabling students and instructors to focus on different aspects of writing.

    1. Augmenting teachers, not replacing them

      Amen!

    2. There’s a line somewhere between using ChatGPT in collaboration, and getting it to do all the work.

      Important point.

    3. ChatGPT is not an original thinker, but you are.

      This is important to remind students of too. And maybe a key area for teachers to focus on what students could contribute to a writing process that includes ChatGPT.

    4. using the model’s suggestions as a starting point

      Perhaps the same with students. Not using ChatGPT to write the essay, but perhaps in the brainstorming process.

    5. Right now, one of the most powerful things you can learn about ChatGPT is how to write quality prompts.

      Interesting. Writing instructors could start to train students in writing prompts for AI. The rubrics below are not dissimilar from what we traditionally ask student to do in their writing. So maybe ChatGPT isn't the death of the essay!

    6. Beyond the media hype about cheating,

      I think it's important to move past the plagiarism aspect of the debates around ChatGPT, but don't think it's just "hype." Teachers are concerned.

    1. "I would much rather have ChatGPT teach me about something than go read a textbook."

      What about accuracy? Textbooks go through a rigorous process of composition and editing to ensure accuracy. Most of what exists to be scraped on the internet does not. I realize this is an old Web 2.0 "problem."

      (Would textbooks even be available for scraping by ChatGPT? What does it have access to?)

    2. the company has also heard from them that the chat bot can be "an unbelievable personal tutor for each kid," Altman said.

      ChatGPT as a tutor. Perhaps with the same guardrails in place so that tutors don't do the work for the students.

    3. "We adapted to calculators and changed what we tested for in math class, I imagine.

      What are the implications here for the writing instructor? What "computational" equivalent to basic calculation would then be no longer central to teaching writing?

    1. , an arithmetic operator in Python is not a function. An arithmetic operator is a symbol that performs

      test. I am trying to see if we can cite chapgpt through hypothes.is annotations.

    1. This framing means that as educators we need to be clear not only about what we hope our students are learning but also about how and why.

      This seems to point to process over product and more formative assessment or scaffolding as part of instruction.

    2. The main goal of transparent teaching is simple: to promote students’ conscious understanding of how they learn.

      So metacognition?

    3. The educational issues surrounding ChatGPT are similar in kind to those we've seen with the growing power of the web

      Yeah, is this even a new thing? It this the same debate we've always had?

    1. Note that students will not be able to cite ChatGPT using a link to their generated response;instead, ask students to repeat the exact language of their search query in the footnotes in lieu of a link

      Actually citation is possible with this extension.

    2. formulaic syntax

      Interesting. So creativity is not it's strength. It's imitative.

    3. These tools, along with a range of other practices,

      Yes, the practices are key! I doubt the battle of algorithms can be won by either side.

    1. I've been using ChatGPT pretty consistently during the workday and have found it useful for open ended programming questions, "cleaning up" rough bullet points into a coherent paragraph of text, etc. $20/month useful is questionable though, especially with all the filters. My "in between" solution has been to configure BetterTouchTool (Mac App) with a hotkey for "Transform & Replace Selection with Javascript". This is intended for doing text transforms, but putting an API call instead seems to work fine. I highlight some text, usually just an open ended "prompt" I typed in the IDE, or Notes app, or an email body, hit the hotkey, and ~1s later it adds the answer underneath. This works...surprisingly well. It feels almost native to the OS. And it's cheaper than $20/month, assuming you aren't feeding it massive documents worth of text or expecting paragraphs in response. I've been averaging like 2-10c a day, depending on use.Here is the javascript if anyone wants to do something similar. I don't know JS really, so I'm sure it could be improved. But it seems to work fine. You can add your own hard coded prompt if you want even. async (clipboardContentString) => { try { const response = await fetch("https://api.openai.com/v1/completions", { method: "POST", headers: { "Content-Type": "application/json", "Authorization": "Bearer YOUR API KEY HERE" }, body: JSON.stringify({ model: "text-davinci-003", prompt: `${clipboardContentString}.`, temperature: 0, max_tokens: 256 }) }); const data = await response.json(); const text = data.choices[0].text; return `${clipboardContentString} ${text}`; } catch (error) { return "Error" } }

      .

    1. create assessments that “take into consideration the processes and experiences of learning.”

      Annotation!

    2. Ask students to engage in metacognitive reflection that has them articulate what they have learned, how they have learned it, and why the knowledge is valuable.

      Students annotating their own writing?

    1. Is this moment more like the invention of the calculator, saving me from the tedium of long division, or more like the invention of the player piano, robbing us of what can be communicated only through human emotion?

      Great question!

    2. The question isn’t “How will we get around this?” but rather “Is this still worth doing?”

      Somewhat defeatist. Quit rather than evolve?

    3. The rudiments of writing will be considered a given, and every student will have direct access to the finer aspects of the enterprise.

      I wonder if there are analogs in math.

      The graphic calculator, for example, must have changed how math was taught, removing the need for that lower-order computation in math.

    4. Last night, I received an essay draft from a student. I passed it along to OpenAI’s bots. “Can you fix this essay up and make it better?” Turns out, it could. It kept the student’s words intact but employed them more gracefully; it removed the clutter so the ideas were able to shine through. It was like magic.

      This is probably scariest of all. ChatGBT as editor rather than author.

    5. nor does it successfully integrate quotations from the original texts

      Interesting. Probably easy for AI develop this skill rather than a limit of the technology.

      But, for now, maybe a good indicator of more sophisticated writing.

    6. What GPT can produce right now is better than the large majority of writing seen by your average teacher or professor.

      Wow, that's a provocative statement! What is meant by better here?

      On some level, I've always felt that a poorly-written, but original essay is better than a well-written, well-analyzed but plagiarized one.

    1. methods of assessment that take into consideration the processes and experiences of learning, rather than simply relying on a single artifact like an essay or exam. The evidence of learning comes in a little of different packages

      How about Hypothesis social annotation throughout a course and throughout the process of essay composition.

    2. The fact that the AI writes in fully fluent, error-free English with clear structure virtually guarantees it a high score on an AP exam

      Yikes!

    3. ChatGPT may be a threat to some of the things students are asked to do in school contexts, but it is not a threat to anything truly important when it comes to student learning.

      Great line, powerful claim.

    4. an opportunity to re-examine our practices and make sure how and what we teach is in line with our purported pedagogical values.

      Love this.

    5. Rather than letting students explore the messy and fraught process of learning how to write, we have instead incentivized them to behave like algorithms, creating simulations that pass surface-level muster

      Annotation shows that messy process.

  3. Jan 2023
    1. In The New Laws of Robotics, legal scholar Frank Pasquale argues for guidance from professional organizations about whether and how to use data-driven statistical models in domains such as education or health care.

      Very interesting. Hypothesis, in its small way, can perhaps help some educators...

    2. we need collaborative processes to seek clarity.

      Indeed!

      And the reminder that writing (and knowledge production more generally) is always collaborative, has an audience, both potentially elided by relying on ChatGPT to generate prose/ideas.

    3. slow thinking,

      Love it! Social annotation certainly help slow reading IMO.

    4. Should I ask students to prompt a language model and then critique its output?

      Great assignment idea!

    5. preferences of data scraped from internet sites hardly renowned for their wisdom or objectivity.

      Something else we try to teach our students, right?

    6. “mathy math,” a model of language sequences built by “scraping” the internet and then, with massive computing, “training” the model to predict the sequence of words most likely to follow a user’s prompt

      A kind of plagiarism in and of itself?

    7. What a contrast to the masochistic persistence I had practiced for so many years and preached to my struggling students.

      So true. Writing is hard, isn't it? ChatGPT sometimes makes it look easy. What will students make of that!?

    1. Back in the early 2000s, I used to demonstrate to students how EasyBib often gets it wrong when it comes to MLA formatting.

      This is a great analogy. I remember feeling the same way about EasyBib when teaching comp.

    2. having students socially annotate the paper, practicing their editing and fact-checking skills.

      Yes! Would love to see an example of such an assignment.

    3. The text is being generated on behalf of the student and is being substituted for the student’s self-generated text. This use of AI is inherently dishonest.

      Could one still argue that it's a component piece of the text/writing that is generated? Just like spelling, grammar, and citation are?

      No doubt it's a lot MORE of the text that is generated and COULD be handed in completely as is in many cases. But could it nonetheless be seen as a kind of starting point for students to then focus on other work, other skills? Like the editing processes mentioned above.

    4. Teaching students to be good critical readers takes time and requires instructors develop activities, such as social annotation assignments, that draw students’ attention to the details of a well-written text.

      Yes! And they ARE writing when they read and annotate, so they can still practice and instructors can still evaluate that skill. It's just a very different writing assignment than a final paper.

    5. So, while effective editors may or may not be exceptional writers, they must be great critical readers.

      I have often wondered (when I was an English teacher), am I teaching writing or reading? Obviously the answer is both.

      The product of so much English courses is paper writing, but that's also meant to be an assessment of a student's reading, right?

      So maybe there's a shift to focus more on reading as a formative assessment that is needed?

    1. I could instead present students with ChatGPT’s response alongside some marking instructions and ask them to provide a critique on what grade the automated response deserves and why.

      What a great assignment idea (and Hypothesis could be used). Would really help students reflect on what writing is and what techniques/skills are needed to be an effective writer, sone modeled by ChatGPT, some not.

    2. do we really need all students to be writing the same essays and responding to the same questions?

      Hmm

    3. an opportunity to improve the way we assess

      Twist!

    4. articulate its inability to fully replicate the expertise and real-world experience that human teachers bring to the classroom

      Learning from the discourse over the past 6 weeks?

    5. If ChatGPT is used to grade assignments or exams,

      Cheating for teachers?

    6. making it capable of engaging in natural language conversations.

      Is it conversation?

    1. Figure 3. The average drop in log probability (perturbation discrep-ancy) after rephrasing a passage is consistently higher for model-generated passages than for human-written passages. Each plotshows the distribution of the perturbation discrepancy d (x, pθ , q)for human-written news articles and machine-generated arti-cles; of equal word length from models GPT-2 (1.5B), GPT-Neo-2.7B (Black et al., 2021), GPT-J (6B; Wang & Komatsuzaki (2021))and GPT-NeoX (20B; Black et al. (2022)). Human-written arti-cles are a sample of 500 XSum articles; machine-generated textis generated by prompting each model with the first 30 tokens ofeach XSum article, sampling from the raw conditional distribution.Discrepancies are estimated with 100 T5-3B samples.

      quite striking here is the fact that more powerful/larger models are more capable of generating unusual or "human-like" responses - looking at the overlap in log likelihoods

    2. if we apply small perturbations to a passagex ∼ pθ , producing ̃x, the quantity log pθ (x) − log pθ ( ̃x)should be relatively large on average for machine-generatedsamples compared to human-written text.

      By applying small changes to text sample x, we should be able to find the log probs of x and the perturbed example and there should be a fairly big delta for machine generated examples.

    3. As in prior work, we study a ‘white box’ setting (Gehrmannet al., 2019) in which the detector may evaluate the log prob-ability of a sample log pθ (x). The white box setting doesnot assume access to the model architecture or parameters.While most public APIs for LLMs (such as GPT-3) enablescoring text, some exceptions exist

      The authors assume white-box access to the log probability of a sample \(log p_{\Theta}(x)\) but do not require access to the model's actual architecture or weights.

    4. Empirically, we find predictive entropy to be positively cor-related with passage fake-ness more often that not; there-fore, this baseline uses high average entropy in the model’spredictive distribution as a signal that a passage is machine-generated.

      this makes sense and aligns with the gltr - humans add more entropy to sentences by making unusual choices in vocabulary that a model would not.

    5. We find that supervised detectors can provide similardetection performance to DetectGPT on in-distribution datalike English news, but perform significantly worse than zero-shot methods in the case of English scientific writing andfail altogether for German writing. T

      supervised detection methods fail on out of domain examples whereas detectgpt seems to be robust to changes in domain.

    6. ex-tending DetectGPT to use ensembles of models for scoring,rather than a single model, may improve detection in theblack box setting

      DetectGPT could be extended to use ensembles of models allowing iot to work in black box settings where the log probs are unknown

    7. hile in this work, we use off-the-shelfmask-filling models such as T5 and mT5 (for non-Englishlanguages), some domains may see reduced performanceif existing mask-filling models do not well represent thespace of meaningful rephrases, reducing the quality of thecurvature estimate.

      The approach requires access to language models that can meaningfully and accurately rephrase (perturbate) the outputs from the model under evaluation. If these things do not align then it may not work well.

    8. For models be-hind APIs that do provide probabilities (such as GPT-3),evaluating probabilities nonetheless costs money.

      This does cost money to do for paid APIs and requires that log probs are made available.

    9. We simulate human re-vision by replacing 5 word spans of the text with samplesfrom T5-3B until r% of the text has been replaced, andreport performance as r varies.

      I question the trustworthiness of this simulation - human edits are probably going to be more sporadic and random.

    10. Figure 5. We simulate human edits to machine-generated text byreplacing varying fractions of model samples with T5-3B gener-ated text (masking out random five word spans until r% of text ismasked to simulate human edits to machine-generated text). Thefour top-performing methods all generally degrade in performancewith heavier revision, but DetectGPT is consistently most accurate.Experiment is conducted on the XSum dataset

      DetectGPT shows 95% AUROC for texts that have been modified by about 10% and this drops off to about 85% when text is changed up to 24%.

    11. DetectGPT’s performancein particular is mostly unaffected by the change in languagefrom English to Germa

      Performance of this method is robust against changes between languages (e.g. English to German)

    12. ecause the GPT-3 API does not provideaccess to the complete conditional distribution for each to-ken, we cannot compare to the rank, log rank, and entropy-based prior methods

      GPT-3 api does not expose the cond probs for each token so we can't compare to some of the prior methods. That seems to suggest that this method can be used with limited knowledge about the probabilities.

    13. improving detection offake news articles generated by 20B parameterGPT-NeoX

      The authors test their approach on GPT-NeoX. The question would be whether we can get hold of the log probs from ChatGPT to do the same

    14. his approach, which we call DetectGPT,does not require training a separate classifier, col-lecting a dataset of real or generated passages, orexplicitly watermarking generated text. It usesonly log probabilities computed by the model ofinterest and random perturbations of the passagefrom another generic pre-trained language model(e.g, T5)

      The novelty of this approach is that it is cheap to set up as long as you have the log probabilities generated by the model of interest.

    15. See ericmitchell.ai/detectgptfor code, data, and other project information.

      Code and data available at https://ericmitchell.ai/detectgpt

    1. The real danger is not to people who are experts in their fields. Super experts in every field will continue to do what they have always done. All of us, however, are novices in almost everything we do. Most of us will never be experts in anything. The vast majority of the human experience of learning about something is done at the novice level. That experience is about to be autotuned.

      This. And connected to the perverse incentives of views and engagement, the flood of autotuned, adequate enough to pass through filters, songs is upon us.

    2. Starting this year, we’re going to be returned a mishmash of all the information that is available on the Internet, sorted by mysterious practices

      We're already there, really. The question isn't whether ChatGPT output is seen as, or mistaken as, human to a degree convincing to everyone (though it eventually will be), but that it is already indistinguishable from the lower-tier content, the shite that clogs up search already.

    1. but an opinion is different from a grounded understanding.

      Preach! Here maybe we're approaching at the limits of AI writing chatbots and the horizons of where we need to push student writing.

    2. Humanities departments judge their undergraduate students on the basis of their essays. They give Ph.D.s on the basis of a dissertation’s composition. What happens when both processes can be significantly automated?

      Scary!

    1. It can summarize things you’ve said to it in new language that helps you look at yourself in a different light and reframe situations more effectively. 

      This IS fascinating. Is something lost here, though?

      I keep thinking about the journey versus the destination. There's not doubt a car gets you places faster and more efficiently than a bike. But riding a bike does open about physical and geographic awareness less accessible in an automobile.

    2. Journaling in GPT-3 feels more like a conversation, so you don’t have to stare at a blank page or feel silly because you don’t know what to say.

      Is this part of the generative (and sometimes frustrating) part of journaling?

      In general, this article seems rather utilitarian in its understanding of journaling. But I don't journal regularly so maybe I'm not one to talk.

    3. If you know how to use it correctly and you want to use it for this purpose, GPT-3 is pretty close, in a lot of ways, to being at the level of an empathic friend

      Interesting. In other contexts, AI has been aligned with the unfeeling?

    1. more and more jobs involve the use of generative AI for everything from discovering new drug molecules to developing ad copy,

      Working with ChatGBT is preparing students for the workplace.

    2. more creative assessments that require students to demonstrate application of knowledge rather than simply the ability to produce information.

      More creative and more formative.

    3. It forces us to reconsider what is distinctly human about intelligence if a machine can generate human language complete with analysis.

      Really situates this moment in history.

    4. I fully believe that the fact that the essay was written by AI and not a live person would be undetectable for many college admissions committees.

      Yikes!

    5. synthesize

      Is AI really not synthesizing?

    1. deeper humanistic questions like, what is truth? What is. beauty? How do we know what we know?

      Like the calculator opened up other areas of math education. Let chat bots do some of the grunt work?

    2. I'm happy to say good riddance to the college essay and other "skills" that we've come to see as proving the value of the humanities.

      Again, with the with the throwing things away...

    3. one that demands students find out something about themselves and tell it to you in a voice that is their own.

      I do think student voice is an interesting place to focus attention on in this debate.

    1. One high school teacher told me that he used ChatGPT to evaluate a few of his students’ papers, and that the app had provided more detailed and useful feedback on them than he would have, in a tiny fraction of the time.

      Interesting concern: AI writing chat bot replacing teacher. Like concerns over Perusall's algo-grading.

    2. They’ll need to know their way around these tools — their strengths and weaknesses, their hallmarks and blind spots — in order to work alongside them. To be good citizens,

      And good workers!

    1. If my goal as a teacher is to help students learn, then should I withhold access to the same information until the timing is convenient for me? Some teachers might look at this and ask if they’re even relevant anymore. I would suggest that the answer is yes, but our focus will shift on helping students develop even deeper critical thinking skills in English. Less motivated teachers who just want to teach grammar may be weeded out over time, but those who are ready to take deeper dives will get the opportunity to mentor students on their path to becoming confident and autonomous English users.
    2. ChatGPT can be used as a writing assistance tool. ELLs can use ChatGPT to generate ideas and receive feedback on their writing. ChatGPT can also provide grammar and spelling assistance, which can be particularly helpful for ELLs who are still learning the rules of the English language.
    3. you can use pretty much any device with a microphone to transcribe your spoken English into the prompt box. Once the responses are generated, students can use screen reading software to verbalize the response.
    1. In OOP, objects are used to represent real-world entities, and the methods (i.e., functions) and attributes (i.e., v

      this is a test to see if hypothes.is works on ChatGPt

  4. Dec 2022
    1. At the end of the day, Copilot is supposed to be a tool to help developers write code faster, while ChatGPT is a general purpose chatbot, yet it still can streamline the development process, but GitHub Copilot wins hands down when the task is coding focused!

      GitHub Copilot is better at generating code than ChatGPT