GPT-5 Pro ranks higher; purple (right) = Claude Opus ranks higher.
what if both Llms rank it higher than the human? Do we mean 'highest'?
GPT-5 Pro ranks higher; purple (right) = Claude Opus ranks higher.
what if both Llms rank it higher than the human? Do we mean 'highest'?
Each column shows rank order (1 = highest rated) within that source;
needs better explanation?
r n_papers_slope-
computation not rendering here
Rank comparison. Figure 2.2 places human rankings
this is cool. would be much better if it enabled hover labels
matched sample
By "matched sample" this means the ones that matched with the LLM evaluations - this should be something we can improve on soon
(explaining the slightly different ρ values).
The difference between the value in the diagram and in the table. I didn't understand what difference was being referred to at first... So this should say, "Explaining the slightly different ... values between the table and the figure."
scale relative to humans, producing narrower spread
Give some actual statistics backing this up
ge but diverge more at the extremes.
Is this my eyeballing of it or is this something that has actually been measured?
evaluator pairs is no tighter than the LLM-human scatter in panel
The vague statement. I mean it's not obviously tighter but you can't eyeball it and say that it's no tighter. Okay it's less tight if we use the Spearman measure. That should be made a bit clearer. I didn't see that you gave us the Spearmans.
But I still think this gets back to the question of whether it's fair to compare the human-human individual evaluator correlations to the correlation between the LLM and the average of humans. Given both signal and noise, I imagine that the average of 2 measures tends to be more reliably predicted than by a measure with signal and noise than one individual measure predicts a second measure.
Compare panels (b) and (c) directly to see whether LLM-human scatter is tighter than human-human scatter.
Probably put at least one correlation metric in each plot because it's really hard to eyeball this
evaluator pair (papers with 3 raters contribute 3 points); uses all r n_matched
Not being computed here, coding error
Figure 2.1:
Paper Labels are missing for the human-human level comparison
Results CodeShow All CodeHide All CodeView Source Setup and libraries
"Hide all code" button does not seem to be working
) ratings for each paper, revealing inter-rater variability—CIs often span 20–40~points.
The statement is confusing and not fully explained. What CIs are we talking about here? Note that we ask each rater to explicitly provide 90% credible intervals for each rating. Is that what this is referring to? But that's a different thing than inter-rater variability
In most cases both LLMs fall within the range of human opinions, though several papers show substantial divergence.
This might have been my own language question? In any case we should have some numbers to back this up - it's not clear to me that this statement is in fact justified using the diagram. I seem to see a lot of cases where the LLM ratings fall outside the humans or at least more than a few
Per-paper overview and model comparison. Figure 2.1 presents three complementary views of overall (0–100 percentile) ratings. Panel (a) displays individual human evaluator ratings alongside GPT-5 Pro (orange diamo
I guess this is ordered from highest human rating to lowest average human rating? Check this and explain it in the diagram or the discussion
evaluator ratings
human ratings "in green circles" -- note this here
Per-paper overview and model comparison. Figure 2.1 presents three
Diagrams are too small. I can barely see them. At least in this version of it. If these are meant to be printed out or whatever, there's no way people will be able to see it. In an online hosting you could have them zoom in of course. But for a sort of printable version you'd need to make these a lot bigger. And no one can read the names either
We evaluate 6 frontier LLMs against human expert reviews from The Unjournal.
This seems repetitive of what we said in the first section... to the extent it needs to repeat, please take on board the hypothesis comments there
Results CodeShow All CodeHide All CodeView Source
Putting results before methods might be the norm in computer science or something but in economics I think we usually see the methods and discussion come first (although people often mention the results in the introduction, )
criterion-level ceiling.
I don't know why they use the word "ceiling." It's not really a ceiling here. Maybe it's a point of comparison, but there's nothing that statistically or mathematically bounds the others to be below this. And in fact sometimes the models do better at matching humans at least in the stats that I've seen by this measure, than humans do.
If two human evaluators agree at Spearman ρ = 0.55, an LLM achieving ρ = 0.57 against the human mean is performing within human inter-rater range.
Not sure I completely understand the claim here and what is meant by "performing within human interrater range."
severity, topic familiarity, interpretation of the scale)
Perhaps also mention that we're asking them to provide percentiles relative to papers in this area that they read in the last two years, and different evaluators may have read different selections of research. There should be a link here to the actual guidelines that we gave the humans (https://globalimpact.gitbook.io/the-unjournal-project-and-communication-space/policies-projects-evaluation-workflow/evaluation/guidelines-for-evaluators)
33 of these were also evaluated by Claude Opus 4.6
We want to make sure this is either dynamically coded or that the LLM is called "updated" as we should increase this
six frontier LLMs
These are not all frontier, I would say. Or am I wrong here? Does the term "frontier" include faster but less deeply thinking models?
Funding for The Unjournal has been provided by the Survival and Flourishing Fund, the Long Term Future Fund, and EA Funds.
do we need to mention Unjournal funding here?
The Unjournal setting is particularly well suited for this comparison. It commissions paid expert evaluations using a structured rubric covering seven percentile criteria with 90% credible intervals plus journal-tier predictions, and publishes the resulting packages openly
A bit more context on The Unjournal would probably be helpful here, mentioning our prioritization, etc.
Claude added this comment (on an earlier version?) Claude: Selection bias: Unjournal selects papers from NBER/top working paper series. This is not a random sample of research. LLM performance on pre-screened quality papers may differ from performance on the full distribution (including poor papers). Explicitly note: "Our sample is pre-selected for quality; results may not generalize to evaluating lower-quality submissions."
strain
citation needed
Our headline finding is that the best-performing model (GPT-5 Pro) matches or exceeds pairwise human inter-rater rank agreement on overall quality,
I don't want to be seen as cherry-picking here. When we report this we should also report the other important statistics like Krippendorfer and at least mention which metrics the LLM performs worse on in terms of matching this
while the journal-tier predictions provide an external reference point2
By the language here, the predictions are not an external reference point. The publication outcomes and perhaps citation outcomes are an external reference point even though as we say, this is not a precise measure of the "quality" of the paper.
reducing classic gatekeeping motives and increasing reviewer effort.
Not sure what they mean by "reducing classic gatekeeping motives." We argue that it does lead to a high level of reviewer effort for a few reasons, but this is not fully justified here. The case we make is that we manage it carefully, that the reviews (we call them "evaluations") will be made public. So people may want to set a better standard, and some people leave their names (aka sign their reviews) so there's a reputation motive. We offer compensation as well as prizes for the strongest work. So there's an incentive the direct financial incentive although our compensation is fairly modest.
structured measurement schemas (Asirvatham, Mokski, and Shleifer 2026), iterative quality-checking workflows (Zhang and Abernethy 2025), or the kind of prompt-robustness engineering motivated by specification-search concerns (Asher et al. 2026)—should improve further.
Of course we want to look at these carefully before we praise them. I'm not super familiar with what each of these things are. And I'm not sure that I would state it so strongly that it will necessarily improve on this. There may be countervailing constraints and limitations. ... Taking a look at the AMS abstract, I don't quite see that this is the same sort of thing we're trying to do.
can achieve
I'd add the word "currently" here.... can "currently" achieve
with no iteration, retrieval augmentation, chain-of-thought scaffolding, or multi-step agentic loop.
Rephrase as "We do not do any iteration..." As a separate sentence otherwise, it's a little confusing what we're saying we are doing versus not doing.
deliberately
Repeats what we just said above
deliberately
I would take out the word "deliberately" here
strong case that frontier LLMs can serve as additional expert raters in structured evaluation pipelines,
I think saying "strong case" is probably too strong
Our headline finding is that the best-performing model (GPT-5 Pro) matches or exceeds pairwise human inter-rater rank agreement on overall quality,
Just need some clarification. It's meeting or exceeding this if we compare it to the average of the human ratings. Is that a fair comparison? Double-check or compare it to how it would compare with the individual human raters.
against expert evaluations for 60 economics and social-science working papers
I don't think this number should be 60 - I thought we only have 57, at least 57 that are publicly released.
multi-dimensional
It's not clear what the advantage of our evaluations being "multi-dimensional" here is. At least this paragraph doesn't make it clear. The paper should make it clear that we also ask for overall judgments in comparison to familiar journal tiers. I would say the advantage of the multi-dimensional is it gives us a sense of the aspects of the research that the LLM tools tend to agree or disagree with the humans on... Something like an understanding of tastes and prioritization.
These developments make the evidentiary gap salient: funders, editors, and policymakers need to know when AI evaluation outputs are trustworthy enough to use, and when they are unstable, biased, or manipulable. Recent work highlights all three concerns. First, reproducibility can be “jagged”: repeated runs of the same models on the same corpus over time can be highly consistent for some tasks and models, but much less so for others (Thomas, Romasanta, and Pujol Priego 2026); robustness may require separating scientific judgment from computational execution (Xu and Yang 2026); and even without overt adversarial intent, subtle reframings of the same task can induce systematic shifts in outputs—a form of LLM “specification search”—raising concerns about frame-sensitive biases when models serve as measurement instruments (Asher et al. 2026). Second, adversarial manipulation is not hypothetical: invisible-text “prompt injection” can substantially inflate LLM-assigned review scores and acceptance recommendations in simulated peer review (Choi et al. 2026), and prompt-injection vulnerabilities are also documented in other high-stakes advice settings (Lee et al. 2025). Third, even when outputs look fluent and plausible, it remains unclear whether AI models approximate expert judgment: AI-generated reviews tend to cover more surface-level sections while being less thematically diverse and less focused on interpretation, originality, and applicability than human reviews (Rajakumar et al. 2026); LLMs used as manuscript quality checkers identify only a small fraction of confirmed critical errors even with the strongest reasoning models (Zhang and Abernethy 2025); and LLM scoring exhibits systematic range restriction and halo effects that can distort agreement metrics (Wang et al. 2025).
This seems too long. This isn't really coming from us, so we might mention some of these things, but I tend to make this a lot shorter. Perhaps some things can be put in footnotes. Obviously we need to check these carefully to see if we agree with them.
I think I mentioned before I'm not sure our work really speaks to the prompt injection issue. The set of work we're putting the LLMs and humans to evaluate would seem to be rather unlikely to have such prompt injection, so we can't really test that (unless we modified the work being fed in, but I don't think that's in our wheelhouse right now. )
Meanwhile, publishers are formalizing policies that treat manuscripts and reviews as confidential and prohibit reviewers from uploading them into general-purpose generative AI tools
I'm just checking some of these references to see if there's hallucination going on - this one seems to check out ... from the cited policy " Reviewers should not upload a submitted manuscript or any part of it into a generative AI tool as this may violate the authors’ confidentiality and proprietary rights and, where the paper contains personally identifiable information, may breach data privacy right".
under strain
citation needed
how much do plant-based products actually substitute for animal products?
maybe 'replace' rather than 'substitute for'?
irr::kripp.alpha(M, method = "interval")$value }, error = function(e) NA_real_)
This is the interval version of K's alpha which penalizes the square of the distances. I guess this means that it particularly penalizes cases where the raters are very far apart and a larger number of small differences won't matter as much. I'm not sure if this is appropriate to ask something to think about. Perhaps we also want to provide the ordinal version for comparison or something else. I believe we've thought about this but I can't remember what we came up with. We'll have to re-consult the notes.
0.07
Wow this is nearly zero agreement among humans but I wonder if something is coming up here because of the way we changed the category/introducing new criteria? I think the claims might've been something we introduced later in the process, at the same point that I coalesced the two things related to global relevance. (I could check this. We have documentation.)
α_HL (GPT-5.2 Pro)
We should put this one as the second from the left? As it seems to be performing nearly the best if not the best
Table A.3: Krippendorff’s αHH
I suggest we should also include the agreement measures for the journal tears? It should be comparable to the others at least if the measure is fairly unit-less
At 20 kTA reference scale
Still needs more explanation. I don't know why you're using this reference scale. I don't know why we're talking about pharma grade, etc.
CAPEX scaling
But what equation does it actually appear in? I guess it's the scale exponent r?
Scalable GF technology 50% Switches to “cheap” GF prices Pivotal uncertaint
As mentioned elsewhere, this needs a lot more explanation or discussion. What is the major factor switching us between cheap and expensive growth factors here? How much does this affect the outcomes? What are the different price distributions for the cheap versus expensive ones? I'm not actually seeing growth factors or any of these p's in any of the equations you give. At least not in a way that allows me to unpack each element. ... Okay, now it's partially explained above, but I still don't see what the different price distributions are and where they come from for the cheap versus expensive
(e.g., breakthrough growth factor technology but prohibitively expensive financing).
Explain a little bit more why these two things should be correlated. You're saying that if there are breakthroughs in growth factors, the industry will be less risky, so financing will be cheaper?
yments:
You should provide an explanation in a folding box for the CRF formula in intuitive terms. I suppose it depends on the interest rate r and n, the number of years, but then explain how that works and why it equals this complicated formula.
The slider complements this by letting users explore “what if progress is partial?” scenarios.
That seems to be underexplained and seems to contradict what you just said.
If any one of these succeeds at commercial scale, the “cheap” price regime applies
That makes sense, but then what determines how you model the price in the 'cheap' price regime?
$0.20 – $1.20
What is the source of these price numbers? Everything should be referenced / linked
Consider correlated scenarios via the maturity slider
We probably want to unpack this more. One could imagine some forms of technical development going together and others less so.
Use for relative comparisons rather than absolute predictions
What do you mean by relative comparisons?
Equipment depreciation period
Which equipment here? Again, I want links directly to the equation.
Weighted average cost of capital
That seems rather high - what are references for this? Why should it be so expensive? Here, is this comparable to some benchmarks?
And again, I want to be able to look up each of these elements within an equation somewhere - I don't see where that equation is. Make the links clearer.
Breakthrough technologies that could trigger the “cheap” scenario: - Autocrine cell lines (cells produce own FGF2) - Plant molecular farming ($1-10/g target) - Precision fermentation at scale - Polyphenol substitution (reduces GF requirements by 80%)
Okay, you got to my question here that I asked above, although it still seems underexplained. Wouldn't each of these things have independent effects on the cost of growth factors? So why is it just a zero-one switch?
30-200 g/L Final biomass at harvest Cycle time 0.5-5 days Time per production batch Media turnover 1-10 ratio 1=batch, >1=perfusion
Interesting, but it should be more clear how this maps into the ultimate cost equation. Everything should be linked back in some way to a total cost formula. I'd like to be able to open and close and unpack the different elements.
Food-grade micros
Give a link or a hover footnote for what each of these things are
Distribution
Why were these distributions chosen? What's the justification? Were these used in other related models you could reference ?
Why correlate? In “good worlds” for cultured chicken: - Technologies are more likely adopted (higher P) - Custom reactors are more common (lower CAPEX) - Financing is cheaper (lower WACC) This prevents unrealistic scenarios where technology succeeds but financing remains prohibitively expensive.
this is really not well explained. I don't see how the discussion relates to the equations here
The model uses a latent maturity factor (0–1) to correlate technology adoption, reactor costs, and financing: Padopted=bound(Pbase+k⋅(m−0.5),0,1) What does “bound” mean? bound(x, 0, 1) ensures the result stays between 0 and 1. Also c
what are k and m here? define and explain
The GF progress slider interpolates between current and target prices: PGF=Pcurrent×(0.01)progress At 0% progress: current prices ($5,000–500,000/g) At 100% progress: target prices ($1–100/g for cheap scenario)
The equation doesn't seem to be correct/displayed correctly here. Explain more but also I don't understand what "0.01^(progress)" means
Example calculation: - Cell density: 50 g/L → need 1000/50 = 20 L per kg - Media turnover: 3× (perfusion system) → 20 × 3 = 60 L/kg - Media price: $0.50/L (hydrolysates) → 60 × 0.50 = $30/kg
is the 'per liter' meaningful though? Doesn't the density depends strongly on the contents used?
Media turnover
what is 'media turnover"?
the dashes are meant to be bullet points -- not rendering right
uld the resulting adjustments change the cost-effectiven
Adjust this to 'meaningfully change' (as defined in the 'resolution criteria')
WELLBY is a reasonably useful measure in this context
Clarify: the WELLBY with linear aggregation as in the definition above
Given the available collected data [...], how should [funders] measure the impact on wellbeing? [...] What measures of well-being should charities, NGOs, and RCTs collect for impact analysis?
Let's split up the answer boxes within this question to ask separately about the best use of currently collected data for these cases, and also ask what data should be collected in the future.
How reliable is the WELLBY measure [...] relative to other available measures in the 'wellbeing space'? How much insight is lost by using WELLBY and when will it steer us wrong?
signpost more that we are talking about the very simple use of the WELLBY measure
More detailed questions on WELLBY reliability
Should be 'on WELLBY reliability and wellbeing measures' ... but also the folding box is still not ideal here -- better for this to link out to another page/subpage (open in new window)
"Meaningful change" = at least one intervention currently in the top 5 moves out of the top 5, OR the #1 ranked intervention changes. This assumes future RCTs incorporate these methods and Founders Pledge updates their CEA accordingly.
This one is nice -- is it the same in the PQ table?
If you propose a measure other than linear WELLBY in your answer above, how much more would it cost to achieve the same welfare improvement using linear WELLBY instead?
Make it clear that 'speculative' is OK here
Currently, standard practice (used by HLI and Founders Pledge) treats SDs on different mental health instruments as interconvertible with WELLBY SDs on a roughly 1:1 basis.
I don't know if that's standard practice -- these are 2 EA-linked or adjacent groups -- moderate
calibration questions can partially adjust for it.
"Adjust for it" is too vague, use more precise language
Key assumptions
Skip the 'key assumptions' part -- I don't think this gets it correctly anyways. E.g., comparisons of SD units shouldn't require linearity per se, that would be sufficient but not necessary. I'm not sure if interpersonal comparability in levels is necessary either -- if we had linearity and cardinality the measured changes wouldn't depend on the starting points
If it's unreliable or systematically misleading, billions of dollars in funding decisions could be poorly directed.
This point seems obvious. Maybe skip it
Is the WELLBY (linear, 0–10 life satisfaction) a useful and reliable measure for comparing interventions—particularly those involving mental health, consumption, and life-saving—in the context that organizations like Founders Pledge use it?
Claude rephrased these in simpler but less rigorous and imo less useful ways. This page should use the actual wording of the key questions on the page https://coda.io/d/Unjournal-Public-Pages_ddIEzDONWdb/Wellbeing-PQ_suPg8sEH#_luVrD0mE -- use quotes where possible, ellipse where necessary, and link or fold the details
I find 'useful and reliable' a bit too vague perhaps
Benjamin et al. show that calibration questions
"Provide evidence suggesting" -- not "show"
effective altruism-aligned
Effective Altruism in caps
How the workshop is structured
This is one proposed agenda -- it has not been finalized yet. Make this known.
3. Could calibration questions improve things?
I would add something like 'or other more in-depth approaches'
"calibration questions"
no need for scare quotes. But note the calibration questions is one of the 2 aspects of this question.
If "7 out of 10" means something very different to different people, that's a fundamental challenge for the WELLBY as a tool for comparing interventions.
This is a bit too simple. Note WELLBY, as used in the simplest approaches, mainly requires differences to be comparable-- and even linear -- across individuals. Moving from 1 to 3 is equally valued as moving from 4 to 6 or 8 to 10, and gets twice the value in this measure as moving 2 people from 3 to 4.
This is one of the most important recent papers
"Is ... most important" ... seems a bit presumptious ... moderate this
Two measures dominate these analyses. The DALY (disability-adjusted life year) comes from health economics and captures years of healthy life lost to disease or disability.
Not sure 'dominate' is accurate. Is DALY used more than QALY? Is WELLBY on par with either of these? Are there other heavily used measures
Founders Pledge, GiveWell, and Open Philanthropy
As well as government ODA agencies and mainstream charities and NGOs like Oxfam and UNICEF
Our highest priority will be to avoid wasting author time. We’re very cognizant from first-hand experience that poor conversion quality, perhaps requiring back-and-forth with the author, is very unpleasant and a huge time suck.
I suspect this is much less of an issue in the days of Claude
Following acceptance, authors may pass their manuscript to the journal in any reasonable format (LaTeX or markdown preferred; Word and PDF acceptable).The document will be published in a “web-first” format, such as the Distill version of R Markdown.This allows reflowable text and mobile readability.We currently do not plan to support interactive content, as we do not think the large effort is worth the modest benefit.
You don't have to host -- why not just evaluate and curte?
Or you can have a compromise -- a 'traditional summary' in the journal, linking to the interactive version created by the author, the latter being the canonical one
NB, I think interactive content is high value, but the authors can produce it, especially given Claude code etc
The review process will be done using a manuscript in PDF format, which can be generated by the authors using whatever software they prefer (e.g., LaTeX). This avoids wasting the time of authors of papers that are later rejected.
Not sure you even need pdf -- markdown should be acceptable, for example
The journal Alignment will be a fast and rigorous venue for theoretical AI alignment—research on agency, understanding, and asymptotic behavior of advanced and potentially self-modifying synthetic agents
Definitely theoretical alignment, not AI governance?
Many potential criticisms of papers are “NP” (can be checked easily), so credentials of reviewer should be irrelevant
I see that as a reasonable sterile manning of what PREreview is doing. In contrast The Unjournal we look for legible signs of expertise when we source and commission evaluators, although we do also encourage in a separate "independent evaluation" mode (which has had very little takeup)
If confidential: Massive reviewer effort (the report) boiled down to a single bit (!)
Yeah, that's the most obvious limitation of the journal system. That's why we say "publicly evaluate and rate, don't accept/reject"
. Our bet is that we can could unify and expand the field of alignment by establishing a legitimate academic journal with an unorthodox review pipeline.
There are some costs here -- maybe you can have your thing ALSO be an overlay journal Publish-review-curate thing, at least for those interested
Since they are regarded as informal by institutional academia, time spent on such outputs is dead time, from the perspective of institutional research performances indicators and career progressio
This is big -- and something Unjournal is also hoping to remedy in a sense
Our experimental solution to address this problem is to publish each accepted paper with a “reviewer abstract”. Its main goal is to help a potential reader decide — on the paper’s merits — if the paper is worth reading.
I like this idea. We ask for "abstracts" too but I particularly like the way you have phrased it, targeted at a potential reader
Reviewer Compensation
We have a promptness bonus and an additional incentive and prize based reward system that follows up later
We think it’s very reasonable to spend an average of ~$3k per paper on reviewer payments.
We spend less on the 'evaluators' but something of this magnitude including eval manager time and my own time etc.
We intend to experiment with LLM recommendations to surface candidates that might not be salient to the editors.
We have experience with this and I can sugggest some good tools
Author identity known to reviewers
This would not work well in situations I'm familiar with. Need to provide the opportunity for single-blind review, especially if there is some meaningful rating or filtering. Otherwise you just get back-slapping
If a submission is published in the journal, the AF post is updated to reflect this, and the reviewer abstract is added. The reviewer abstract can be upvoted on AF, with the reviewers with AF accounts who sign the abstract receiving karma as appropriate.
Why filter rather than just rate and sort? ... and let users choose how to filter?
Journal Not Conference
OK, your situation is rather different than for Unjournal -- I guess you are trying to build credibility and institutional structure for a new and fledgling tield
We are tentatively planning on making the journal archival, meaning that publication there constitutes the “version of record”, in contrast to a workshop publication. (Preprints of course are allowed.)
Bad idea IMO ... although these concerns may vary by field
Public review avoids this, but introduces additional problems due to lack of confidentiality: less honest, more combative and defensive conversations between authors and reviewers. Public review also produces an artifact that is poorly suited to a reader because the c
anonymous public review exists
Paper-by-Paper Comparison
Fix margins -- way too narrow to see
Full LLM evaluations, human adjudication, new human evaluations
This iss likely to occur over a longer time. There will be overlap between these phases.
Research Goals and Questions (overview)
Let's make this some sort of offset box
Table 3.1: Token usage and estimated cost per model
@Valentin some confusion here. In some cases we have pro on 3 papers. In some cases we have pro on 40 or so papers
To assess whether different LLMs produce systematically different evaluations, we collected ratings from multiple providers: OpenAI (GPT-4o-mini), Anthropic (Claude Sonnet 4), and Google (Gemini 2.0 Flash)
We should note that none of these are frontier models other than GPT pro (which version?) @Valentin
?tbl-llm-token-cost-summary
@Valentin missing reference/link
We first use the earlier GPT‑5 Pro evaluation run that covered all papers in our Unjournal sample with a simpler JSON‑schema prompt
@Valentin which 'simpler prompt' was this? We should link it
temporal “data leakage” between the multi‑year regrowth label and contemporaneous predictors;
Did it really show this? OK we can check this in the appendix -- but it might be good to present a bit more side-by-side comparison.
OK I checked the appendix and I couldn't find any mention of the temporal data leakage issue. It mentioned other issues that I interpreted as more about fitting a model on one time period and expecting it to pertain to another period, but that's not 'leakage'.
We start by examining selected evaluations in detail. In the next step we will juxtapose these LLM assessments with the human evaluators’ written reports.
Let's put a table of the relative ratings here (human vs AI, for each category etc), especially for this subset
sensitivity of return‑on‑investment calculations to assumptions about donor lifetime value and unobserved costs.
iirc this overlaps the human evaluation
potential spillovers and spatial correlation across postal codes
this was raised by the authors themselves, and they had an approach to accounting for it
Case study: Williams et al. (2024)
Valentin: Did an LLM write this comparison or did you? It's so very detailed that I am wondering how you had time to dig in this much.
not fully propagate uncertainty
Did the model really mention this? If so, awesome. (Although I'm a bit concerned about whether our evaluations entered it's corpus because I on't think it did so in the last version)
additionality and opportunity costs in the policy framing,
you didn't mention this one for the LLM
the treatment of “biophysical potential” versus business‑as‑usual regrowth,
is this the same as something the LLM identified? Maybe better to present the exact language in parallel (for human vs llm)?
edictors; incomplete or optimistic treatment of uncertainty around the headline 215 Mha estimate; a broad and permissive definition of land “available for natural regeneration”; limitations of the carbon overlay and permanence assumptions; and only partial openness of code and workflows, which increases barriers to full replication.
I would want to look at this correspondence between human and LLM critiques more closely. (Can also ask LLMs to check that)
relatively
Relative to other papers or relative to other ratings categories? I don't think we showed summary statistics or plots for the overall set of evaluations to compare this to know if it is 'relatively high'.
Or maybe you meant 'relative to the human evaluations'?
This October 2025 run asked the model only for numeric ratings and journal‑tier scores (no diagnostic summary or reasoning trace);
I thought we asked it for 'reasoning for each rating'?
To understand what GPT‑5 Pro is actually responding to, we re‑ran the model on four focal papers (Adena and Hager 2024; Peterman et al. 2024; Williams et al. 2024; Green, Smith, and Mathur 2025) using a refined prompt (as shown in the previous section).
It's not clear to me how the prompt use here is different from the prompt used on the rest of the papers.
temporal leakage from contemporaneous predictors,
I actually didn't see this in the discussion! Looked through it in some detail
ourth, when sufficient information to compute a standardized mean difference (SMD) was lacking and the text reported a “null,” outcomes were set to an “unspecified null” of 0.01. This imputation is transparent but ad hoc; it could bias pooled estimates upward (relative to zero) and may not reflect the true variance of those effects. The manuscript would benefit from sensitivity checks setting these to 0, excluding them, or modeling them with conservative variances.
IIRC this echoes the human evaluation (although one of the evaluators had a particular detailed suggestion for this)
Model assessment summary
Paragraph breaks missing here ... makes it hard to read.
For “tier_will,” given its status as a WZB discussion paper and the need to disregard actual publication knowledge, I might predict it will land around 3.2 to 4.0.
This suggests/confirms that the model does not have access to the latest 'news' about the publication (in Management Science)
clearer pre-analysis plan deviation trackin
OK it gets at this a bit
Heterogeneity analyses suggest stronger effects in urban areas and in PLZs with higher employment, more children, and more Catholics, and with higher predicted giving potential. These patterns can guide targeting but also indicate that the ITT estimates average over meaningful heterogeneity.
Seems to miss the issue of MHT, and some very surprising heterogeneity suggests spurious estimates.
Also, divergence from the PAP, although I'm not sure it had access to the PAP
Adena and Hager 2024
Maybe make these either folded or one paper per chapter in the Quarto?
Data construction choices appear reasonable but introduce some judgment calls. Winsorizing PLZ-day donations at €1,000 reduces variance from heavy tails; the authors show that results are directionally robust, but precision trades off.
Good it noted the Winsorizing -- something Reiley emphasized.
The most important methodological limitations concern exposure heterogeneity and spillovers. Treatment is assigned at the PLZ level, but impressions are probabilistic and sparse (roughly one in ten Facebook users in treated PLZs received at least one impression), so the estimates are ITT and likely attenuated relative to the effect of actually seeing the ad; the TOT is not estimated. The allocation strategy partly allows Facebook to endogenously concentrate impressions, creating within-treatment variation in exposure that is not exploited for causal TOT analysis (e.g., using randomized budgets as an instrument in a dose–response framework). Spillovers across PLZs are plausible (algorithmic leakage of geotargeting and social diffusion). The authors document positive “share of treated neighbors” effects and argue the main estimates are lower bounds, but the neighbor-treatment share is not itself randomized, and spatial correlation or common shocks could inflate these coefficients; the spillover analysis should be interpreted cautiously. Robustness to spatial correlation in errors is only partly addressed by robust standard errors and randomization inference; alternative SEs (e.g., spatial HAC or clustering at larger administrative units) and placebo geographies would further strengthen inference.
At a first look (and from my memory) this seems like an extremely useful and plausible report!
claims & evidence, methods, logic & communication, open science, global relevance, and an overall
remove bold font
For each of 47 such papers,
we have more now -- is this '47' hard-coded?
In this project, we test whether current large language models (LLMs) can generate research evaluations that are comparable, in structure and content, to expert human reviews.
This is only a part of the project though
Furthermore, a key promise of AI is to directly improve science and research.
rephrase using language from grant application
works
"works" seems too simple
a high‑stakes, policy‑relevant domain, and as the first step toward a broader benchmark and set of tools for comparing and combining human and AI research evaluations.
Last sentence seems relevant to the grant application language
model reliably identifies many of the same methodological and interpretive issues
"reliably identifies" feels too strong ... or at least I haven't seen the evidence yet.
and to produce a narrative assessment anchored in the PDF of each paper.
I don't understand 'anchored in the PDF of each paper' -- maybe LLM wrote this?
Comparing LLM and human reviews of social science research using data from Unjournal.org
No link to slides anymore?
Strong pantropical mapping, but several methodological and interpretive risks remain. Training data on natural regrowth include substantial omission error in humid biomes; pseudo-absences and misclassification may bias the model. Random forests were trained on class-balanced points and probabilities are treated as calibrated; no prevalence correction or probability calibration is shown, yet expected areas/carbon rely on these values. Validation is not fully spatial; accuracy likely inflated by autocorrelation (declines at greater distances); no formal spatial block cross-validation or predictive uncertainty mapping. Reported “confidence intervals” for total area/carbon are effectively deterministic sums, not uncertainty; overall uncertainty is understated. Predictions at 30 m depend on several coarser predictors (300 m–1 km), so effective map resolution is coarser and may mislead fine-scale planning. Final maps omit socioeconomic predictors (despite similar accuracy), assuming stationarity from 2000–2016 to 2030 and potentially overstating practical feasibility. Carbon estimates exclude permanence/leakage dynamics and use coarse downscaled inputs. Data products are open, but code is only “on request,” limiting full reproducibility.
This new one seems to show some potential to be reflecting the key concerns but I need to check this in more detail as it could just be credible sounding garbage. It still doesn't seem to pick up the key 'data leakage' concern.
But actually I'm a bit puzzled as to what data is being piped in there because if I recall the latest version we had didn't ask for rationales for specific categories. So where is it getting this from?
78
Much higher than the humans, but this category is rated lowest or second-lowest for both ... so perhaps close to a monotonic transformation of sorts.
Show cod
@Valentin These forest plots are really hard to read, it's so dense without spacing. Let's work together on some ways of making it more informative.
I'm also puzzled as to why so many papers are shown only one type of rating and not the other. I know that some of our evaluators did not give ratings like this, and in some cases, we didn't even encourage it. But why is it missing for some of the llms? Did it just not have time to finish processing it?
Maybe it's a display issue? It seems that the papers that were rated highest in terms of these tiers by the human raters did not get rated by the LLMs. Or maybe it just didn't show up in the graph?
(only partially available)
@Valentin why/how 'partially available'?
Inter-rater reliability (Cohen’s κ): We treated the AI as one “reviewer” and the average human as another, and asked: how often do they effectively give similar ratings, beyond what chance alignment would predict? For overall scores, Cohen’s κ (with quadratic weighting for partial agreement) came out around 0.25–0.30. This would be characterized as “fair” agreement at best – a significant gap remains between AI and human judgments. For most specific criteria, the weighted κ was lower, often in the 0.1–0.2 range (and for one category, effectively 0). A κ of 0 would mean no more agreement than random chance (given the distribution of scores), so some categories are bordering on that. In contrast, typical inter-human agreement on these kinds of scoring tasks can also be low, but usually one would hope for κ in the 0.3–0.6 range among trained reviewers on well-defined criteria. Our finding of κ < 0.3 in all cases suggests that the AI’s ratings are not interchangeable with a human reviewer’s – at least not without further calibration.
@Valentin I want to focus on interpreting statistics like these. I think we could make info-theoretic statements about things like "how often would the LLM's relative ranking of 2 randomly chosen papers agree with a human's rannking". Is that kappa? Let's dig in
Correlation (Pearson’s r) between the AI’s and human scores across papers: This tells us, for example, if a paper that humans gave a high score also tended to get a high score from AI (regardless of absolute difference). For the Overall scores, Pearson r ≈ 0.30, indicating a weak-to-moderate positive correlation.
@Valentin Important -- let's put these measures in context. How does this correlation compare with what other papers found, and with inter-human ratings, for example.
We might also find a way to introduce information-theoretic measures that say something in an intuitive absolute sense.
r ~ 0
@Valentin this should be stated more precisely, I think. (And with soft coding)
Table 3.1 show agreement metrics across rating criteria. To quantify the agreements and differences observed, we calculated several statistics comparing LLM scores to the human scores, aggregated by criterion:
@Valentin perhaps we should start with the measures of agreement and then get into unpacking the discrepancies?
has a distinct “taste,” elevating some work and devaluing other work differently than human referees.
I don't think we can call this 'taste' yet. It might be random noise or perhaps ~bias (to top institutions, authors, journals, etc.).
or example, Aghion et al. 2017 was among the top few for human reviewers, but the LLM overall score put it notably lower relative to others, hence a downward green curve.
@Valentin I don't think that was the greatest discrepancy -- should we identify some with a greater discrepancy here? (Ideally, we even soft-code it, as this is likely to change as we adjust the prompts, anonymize, etc.)
high quality due to its robust modeling and accessible methodologies, making it relevant for climate mitigation and biodiversity
This is also not particular coherent. Strange. If we just asked the llm to evaluate the paper it tends to give a much better response.
Comparing its midpoint percentiles to a reference group of serious research from the last three years will help inform this.
@Valentin Yikes -- this sentence doesn't really make sense
Shared notes
@Valentin Let's replace this link with a link to the Coda doc
on the right, the papers are ordered by the AI’s overall score (rank 1 = highest rated by AI)
labeling-wise, would it be better to add the paper authors on the right as well? @valentin.klotzbuecher@posteo.de
For example, in the Logic & Communication column, we see many light-orange cells – the AI often thought papers were a bit clearer or better argued (by its judgment) than the human evaluators did.
I wonder if we should normalize this in a few ways, at least as an alternative measure.
I suspect the AI's distribution of ratings may have different than the human distribution of ratings overall and, the "bias" may also differ by category.
Actually, that might be something to do first -- compare the distributions of (middle -- later more sophisticated) ratings for humans and for LLMs in an overall sense.
One possible normalization would be to state these as percentiles relative to the other stated percentiles within that group (humans, LLMs), or even within categories of paper/field/cause area (I suspect there's some major difference between the more applied and niche-EA work and the standard academic work (the latter is also probably concentrated in GH&D and environmental econ). On the other hand, the systematic differences between LLM and human ratings on average might also tell us something interesting. So I wouldn't want to only use normalized measures.
I think a more sophisticated version of this normalization just becomes a statistical (random effects?) model where you allow components of variation along several margins.
It's true the ranks thing gets at this issue to some extent, as I guess Spearman also does? But I don't think it fully captures it.
deed, GPT often noted lack of code or data sharing in papers and penalized for it, whereas some human reviewers may have been more forgiving or did not emphasize open-science practices as strongly (especially if they focused more on content quality). As a result, for many papers the AI’s Open Science score is 5–10 points below the human average.
This is interesting. The human evaluators may have had low expectations because they don't expect the open code and data to be provided until the paper has been published in a peer-reviewed journal. Here I would agree more with the LLM. "What should be" sense.
Figure 3.2: Relative ranking (overall) by LLM and Human evaluators
A quick impression: the LLMs tend to rank the papers from prominent academic authors particularly high?
he brief rationales clarify what evidence in the paper drove each score.
I don't think we can be confident that whatever it puts here accurately reflects the reasoning or process that determined the rating. (added discussion in text).
Also, is there some way to extract the 'thinking steps' from the process ... that reasoning thread the models show you (which May or may not also reflect its true reasoning )
- tier_should = where the paper deserves to publish if quality-only decides. - tier_will = realistic prediction given status/noise/connections.
This seems like an LLM abbreviation of our instructinos. @valik can you put the actual instructions back in?
Round all scores to the nearest 0.5
We use a single‑step call with a reasoning model that supports file input. One step avoids hand‑offs and summary loss from a separate “ingestion” stage. The model reads the whole PDF and produces the JSON defined above. We do not retrieve external sources or cross‑paper material for these scores; the evaluation is anchored in the manuscript itself.
We should probably give a citation for this point.
But is this the same point you made above?
• Default = arithmetic mean of the other six midpoints (rounded).
@Valentin I'm not sure why this should be the default. Note that in an earlier version of our evaluation framework, we used a weighted scheme, which we dropped
I suspect it would be better for you to ask the question the same way we asking the evaluators here ... Which would simply mean getting rid of this bullet point. I think.
The second bullet point is interesting though. I would be curious to hear how it considered the relevance of each metric in its overall. ... although I expect that asking that question might alter the overall score
@valik Maybe GPT suggested this default averaging? It also seems so optimal because we're asking to get something that we could easily compute ourselves.
plus a short rationale
I don't see where in the prompt you explained to it what the rationale was supposed to mean. I only see some discussion on how to use that. It should be stated if they override the simple mean for aggregating the other categories to get the overall.
(And see my other comment on why I think we should remove the request to do simple means.)
The credible intervals communicate uncertainty rather than false precision
@Valentin That's the intention, sure, but this paragraph makes it seem like the JSON schema somehow ensure that this is used. I'll try to adjust.
# Environment setup
Note to self -- this chunk is set to 'eval=false', it doesn't need re-running every time (and has trouble playing in the Quarto/Rstudio environment)
Direct ingestion preserves tables, figures, equations, and sectioning, which ad‑hoc text scraping can mangle. It also avoids silent trimming or segmentation choices that would bias what the model sees.
Useful to add a citation here.
he sample includes r dim(research_done)[1] papers
@Valentin the number comes out in my Rstudio but not on render. Wondering what I did wrong here. I think it may be syntax, but also it starts again with every chapter, so we need to carry the data in first for this discussion.
49
@Valentin Let's soft-code these numbers in based on the data.
Something like
The sample includes `r dim(research_done)[1]` papers spanning
per criterion (and noting the range of individual scores).
@valentin we should probably do something more sophisticated at the next pass ... Either using each evaluation as a separate observation in the analysis or imputing the median taking the statedcredible intervals into account with some Bayesian procedure.
Please turn on the hypothes.is plugin, and view public annotations to see Latex math representations
What if I annotate this in hypothes.is -- is it preserved usefully?\(CRF(r,n) = \frac{r(1+r)^n}{(1+r)^n - 1} \)
Yes, it seems to stay in the same place in the notebook even if the notebook is edited.
Wait now it's an orphan?
model for now)
DR @ VK: Should we use R inline code for things that will change, like the model number, and centralize these?
Figure 3.4: Human uncertainty (CI width) vs |LLM − Human|. Spearman correlation reported.
This one is rather intricate, it might need some more analysis and talking through.
Table 3.3: CI coverage: does one interval contain the other’s point estimate?
I think DHK also looked at 'human in human' -- Might as well add this as a comparator?
coverage
Not sure if we want to use the formal statistical term 'coverage' here either (unless it's in scare quotes)
Table 3.1: Agreement metrics by metric (Pearson, Spearman, mean/median bias, κ unweighted/weighted).
I assume this is referring to the correlation of (the average of?) the human ratings and the LLM ratings?
Because we could also look at the agreement between humans (something David HJ was working on ... Might be useful to share this with him for his thoughts if he's interested at all.)
metric
Is it worth adding confidence intervals or "significance stars" for these metrics?
mean_bias
are the units "percentage points" or something else?
metric
maybe 'by rating category' To avoid repeating the word "metric",
Calibration
Not actually calibration unless we asked for a specific prediction.
calibration
Maybe we should put in a footnote explaining that these aren't really measures of calibration.
As another indicator of agreement
Exactly. It's sort of a measure of agreement, although I'm not sure quite how to interpret it. It's not a measure of calibration per se because we were asking them to rate these things, not to predict them.
Although we could ask the LLM to do this as a prediction exercise, with or without training data, that might be interesting, And then, of course, the calibration would be meaningful.
monotonic relationship
I'm not sure what the "monotonic relationship" in parentheses means here?
he horizontal radius covers the union of all human evidence for that paper — combining individual raters’ point scores and their CIs
It's not fully clear to me what this does. I guess you center it (horizontally) at the midpoints of the human ratings.
Are you using the outer bounds of both raters' CIs?
Probably the best thing to do ultimately would be to impute a distribution over each individual ratings and CI these and then do some sort of belief aggregation. -- Horizontal linear aggregation actually feels the most intuitive to me from what I've read. ... And then give the implied CIs for that
For each paper (selected metric), the ellipse is centered at the pair of midpoints (Human, LLM).
This is pretty clear but it took me a second to get this, so maybe mention human = horizontal, LLM = vertical in the text.
(unless a point lies outside its own stated CI, which is flagged in a diagnostic check).
Did that ever happen?
robust CI ellipses
Robust in what sense? Also might be worth mentioning that we are talking about 90% credible intervals (at least that's what we asked the humans to give us).
Uncertainty fields. Where available, we carry lower/upper bounds for both LLM and humans; these are used in optional uncertainty checks but do not affect the mid‑point comparisons below.
I suspect we ultimately should do something more sophisticated with this ... like some Bayesian updating/averaging. It's also not entirely clear what you mean by "we carry".
But of course, the "right" way to do this will depend on what precisely is the question that we're asking. Something we should have some nice chats about.
optional
Instead of 'optional' maybe you mean 'additional analyses' or something?
we fold these into the single LLM metric global_relevance.
Do you average them?
Human criteria are recoded to the LLM schema (e.g., claims → claims_evidence, adv_knowledge → advancing_knowledge, etc.)
This is just about coding the variables, right? Not really about the content?
Sources. LLM ratings come from results/metrics_long.csv (rendered in the previous chapter). Human ratings are imported from your hand‑coded spreadsheet and mapped to LLM paper IDs via UJ_map.csv.
Okay, I see these are basically notes to ourselves here.
compute paired scores per (paper, metric)
I'm not sure what is meant by "compute paired scores."
We (i) harmonize the two sources to a common set of metrics and paper IDs
This is fine for our own notes for now, but it's not something that outsiders need to read, I guess.
compares LLM‑generated ratings
Link the ~prompt used so people can see exactly what we asked the LLM
human evaluations across the Unjournal criteria.
"across The Unjournal's criteria"
Limited International Scope: The focus on US data and markets (acknowledged in footnote 20) may limit generalizability, particularly given different regulatory environments and consumer preferences globally.
Perhaps we should footnote something more about how we're focusing on US data because of its availability and thus focusing the question around the US, but will encourage making reasonable extrapolations to global consumption. #todo? Although we already have a footnote about it