Discussion: Questions, metrics, tests

Include global setup and parameters
source("setup_params.R")

See Coda here for now (will integrate back into this Qmd once we have converged on reasonable measures and tests.)

Relying on insights from GPT-pro (see here and here).

*Note: the (*) indicates deferred questions and measures, requiring additional data collection, etc.

Comparing models: prompts and environment engineering for effective LLM evaluation

Which ‘model’, i.e., which prompts, parameters, model, and contexts do best for:

Predicting human ratings (#PREDICT_HUMAN_RATINGS)

Which best produces a vector of ratings that is the closest to human research ratings, in meaningful dimensions? (Beware overfitting; we need clean testing data.)

For each of these we should also compare how well one human rater predicts another human rater. This is high-value for The Unjournal and provides an interesting benchmark for the LLMs.

Simplest comparisons among models

[~GPT] For face-value sanity checks, report and compare models ‘predictions’ of human ratings by: Mean absolute error (MAE), Root mean squared error (RMSE), Correlation (Pearson and Spearman)

Possibly also: Landing within ±5 points, ±10 points, of each human rating (or +/- 1 tier for the 0.0-5.0 measure), (Pseudo)-calibration of LLM CIs: How often does the human point estimate fall within the LLM CI?

Desirable metrics for comparing point estimates

1

Intraclass correlation:2

[GPT] measures absolute agreement across raters viewing the same targets; report CIs and the specific ICC form. Use Koo & Li’s guideline to choose/report ICC form and interpret magnitudes. irr::icc(…).

R: irr::icc() (set model="twoway", type="agreement", unit="single" for single‑rater agreement; average if you aggregate multiple LLM runs).

Krippendorff’s \(\alpha\): “handles missingness, multiple raters, and different scale types (nominal vs ratio). irr::kripp.alpha()

We have missingness and multiple raters; and this also could allow us to combine across scale types (e.g. 0-100 and 0-5). But for ‘combinations’, should we weigh all measures equally, or use an ad hoc weighting for how we see the importance of the measure?

Maybe also estimate

[GPT] Concordance of rank‑orders across papers: Spearman \(\rho\) or Kendall \(\tau\) if you also collect overall ‘rankings’ of paper quality. cor(…, method = “spearman”|“kendall”); DescTools::KendallW() for many raters.

We don’t ask evaluators to do rankings, but we do make these comparisons, and we could see ‘human evaluators’ as a class. And the rank-sensitivity seems compelling. But why choose Kendall vs Spearman?

Probably skip:3

Desirable for comparing distributions (imputed from CIs)

Distributional similarity / information theory

Jensen–Shannon divergence (JSD) — symmetric, bounded in bits; intuitive “distance between distributions”. philentropy::JSD()

K-L divergence: Directional: focus on how well the LLM models predict the humans, i.e., how ‘surprising’ the human distributions are to the LLMs.

Probably skip:4

Desirable: Hierarchical models for identifying systematic patterns (with caveats)

[GPT] Use hierarchical ordinal models: brms::brm(rating ~ method + (1|paper) + (1|rater), family = cumulative("logit")) to estimate the average method effect (LLM vs human) on latent rating and to quantify uncertainty; for frequentist, ordinal::clmm. Report posterior contrasts / CIs (Bayesian) or Wald/likelihood‑ratio tests (frequentist).

We could add parameters for paper characteristics, especially field/cause area, method, and source. However, these should not be treated as causal effects, as they may be confounded with paper quality and rater characteristics. Even the implied ‘interactions’ with ‘method’ (LLM vs human) may be confounded. But it seems descriptively useful and suggestive.

Predicting bibliometrics (*) (#PREDICT_BIBLIOMETRICS)

Making the most accurate predictions for publication outcomes and bibliometrics (citation counts etc.)

Contamination is a substantial concern here (also for human ratings), so we’d need to focus on papers currently in early stages.

Possible outcomes: Top-tier journal (binary), journal REPEC-index, (weighted) citation counts 2y/5y, time to publication in tier above threshold.

Metrics (return to this, we are not likely to have this data for a while).

Why are these questions useful?:

Refining LLM evaluation practice: The research and practice community (as well as The Unjournal) may want to use LLMs to do research evaluation for the reasons described in our outline (which we can elaborate on). Thus, we want to learn and refine the best approaches. Error-checking: Similarly, for ‘checking for errors’ as a complement to peer review.

(Detecting LLM reviews — who cares?): If we want to be able to detect LLM-generated reviews, we want to know how close it can get to human reviews and in what ways it characteristically differs. But do we care?

How will we measure and test each of these?

To be fleshed out in the sections above.

GPT-pro guidance, following my prompt:

Measurements that are useful, intuitive, valid (incl. information‑theoretic), with R recipes. 5

Design backbone (applies throughout):

Unit of analysis = paper × rater/model × metric.6

Use held‑out test sets and time‑based splits (train on past, test on future) to avoid leakage/contamination the proposal warns about.7

Prefer proper scoring rules, agreement coefficients, and hierarchical (mixed) models with paper and rater random effects for inference and uncertainty.8

Adding some ‘pro tips’ that seemed useful:

Report both an agreement coefficient and a simple % agreement so readers can sanity-check chance-correction effects.

Pair coefficients with hierarchical models for inference: e.g., an ordinal mixed model rating ~ method + (1|paper) + (1|rater) to test if LLM differs from humans while the coefficient gives you the practical agreement level.

How do human and LLM evaluations compare? Which is ‘better’, and by how much…?

Better review discussions (#BETTER_REVIEW_SUBJECTIVE, #BETTER_REVIEWS_LLM_JUDGES) (/*)

[BETTER_REVIEW_SUBJECTIVE] Which review discussion is better?

  • According to the subjective judgment of outside expert readers;
  • ideally blinded as to which is AI, but that’s a challenge to implement
  • Divide this up by categories; readability, insight into econometrics, understanding of context, etc.
  • Metrics for ‘how much’ could refer to a percentile distribution of human evaluators.

[BETTER_REVIEWS_LLM_JUDGES] According to the judgment of other LLMs (perhaps those with system prompts that humans rate as useful and good at judging the strength of evaluations — overambitious? Consider — could judging the quality be easier than producing the review?)

Better ratings (#BETTER_RATINGS_SUBJECTIVE)

[BETTER_RATINGS_SUBJECTIVE] Which set of ratings is better (same considerations as for the review discussion)

[FEEDBACK_TO_AUTHORS] In terms of the authors’ ratings for useful feedback?

[PREDICT_BIBLIOMETRICS_COMPARISON] At making predictions for journal, bibliometric, measurable impact, and replication outcomes?

[ERROR_FINDING] At finding errors of particular types? (Flesh these out; simple math errors, mischaracterizing results, unrealistic assumptions, use of unjustified and inferior methods, etc. )

Why are these questions useful?:

  • Should we use LLMs to do evaluation? (And related tasks like pre-screening, grants)

  • Or, how should we use it as a complement?

  • (application) Value of The Unjournal reports

  • Insights into research and insight capabilities of LLMs (e.g., do they rate certain things particularly well?)

  • LLM agreement as a measure of consensus is in an academic sub-field

How will we measure and test these?

Research taste, strengths and limitations

This might need to be qualitative/informal. I’m not sure how to make it quantitative.

  • Which of our rating categories or rating frameworks does its overall assessment tend to correlate with more? (E.g., credibility, innovation, comms)

  • Specific opinions on methodological divides? That seems pretty unlikely to be inherent.

  • Which types of econometric issues and critiques does it tend to favor or diagnose correctly vs. incorrectly? (E.g., causal identification, statistical inference, experimental design…)

  • Characterization of results — “importance hacking”

How ‘malleable’ is it’s taste based on prompts?

What are some deeper but still concrete definitions of different types of either ‘taste’ or ‘research skills’ or more generally ‘cognitive skills.’

Prompt and context variation

These could be assembled in a sheet or data frame. There are probably too many variations here for a full factorial design, or for designs reporting and comparing all ratings across all designs (at least this seems overwhelming).

I suggest 1. Focus on 2-3 ‘likely best setups’, supported by previous work and intuition, for full comparisons 2. For 1-2 of these ‘best setups’ varying other dimensions one at a time, comparing key ratings only.

“(+)” Indicates the most promising options to try first.

Dimensions of variation

Which baseline model

  • GPT5 pro (+) (or o3 until it’s available)
  • Deep research (GPT5) – (comes with web search natively?, may be compatible with supplementary materials)
  • Other providers (*)

Prompt variations

  • “Evaluate as an economist” (+)

  • vs “predict what an Unjournal evaluator would say”

  • Exact guidelines for evaluators (+)

  • LLM-specific tailoring

  • Don’t mention The Unjournal (+)

  • Mention The Unjournal

Paper+ context variations

  • Abstract only (if we have time to strip these)
  • Paper only (+)
  • Paper + supplementary materials + preregistration etc. (When we can)
  • Paper + human evaluations (contamination risk?)

Ordering and separation of prompts

  • A single rating (tier) first

  • Ratings first, then discussion

  • Discussion first, then ratings

  • Explicitly asked to consider discussion in ratings

  • Ratings only

TMZ: This probably doesn’t matter because it’s a reasoning model

  • Split each rating and task into a separate prompt (*)

De-identification and debiasing

[Note: remember to use non-journal versions of PDFs where possible.]

  • No de-identification, no debiasing instructions (+)

  • No de-identification, strong de-biasing instructions (e.g. “ignore author names, institutions, funding sources, and any other information that might bias your evaluation. Focus solely on the content and quality of the research itself.”) (+)

  • Remove author names, institutions, acknowledgements, funding sources (needs to be done with text to pdf conversion first processing)

To do: Check if it still guesses the authors etc.

  • Replace author names with fake prestigious / non-prestigious names (*)

  • Replace institutions with fake prestigious / non-prestigious institutions (*)

Parameter variations

Skip for now?

  • Temperature (Maybe this is fixed at 1 for reasoning models?)

  • Top-p (Rem: distribution of possible tokens to sample from)

  • Number of runs (for aggregation) (Just do it 1x for now)

Between prompt-stability measures

Within‑model repeatability: run k independent generations per prompt; report within‑LLM variance of ratings and text similarity across runs (cosine similarity of embeddings; quanteda.textstats::textstat_simil() or text::textSimilarity())


  1. DR: I think most of the following should work for both the continuous 0-100 ratings and the 0-5 journal tier ratings. I believe these are agreement coefficients, which correct for chance agreement.↩︎

  2. DR: As I understand it, the share of the variance attributable of the paper rather than the identity of the rater.↩︎

  3. Quadratic‑weighted \(\kappa\) (Cohen’s weighted kappa) “penalizes big disagreements more than small ones” intuitive 0–1 scale. DR: I don’t think we want to penalize big disagreements nonlinearly.↩︎

  4. “Mutual information between human and LLM ratings — ‘how many nats/bits of information one gives about the other’.” – I don’t think our measures are reasonably expressed in bits?↩︎

  5. DR: As per my request↩︎

  6. DR: OK, but we want to enable some aggregation and some higher-level effects.↩︎

  7. DR: Aiming to do this using the prospective evaluations. There are two issues here: 1. The usual overfitting issues 2. The possibility of actual model contamination↩︎

  8. DR: I was unsure about the “proper scoring rules” as I thought that was about incentives for human forecasters, but it applies here as well. These have good properties, including “ensures that models are ranked by true predictive accuracy, not by quirks of scale, variance, or overconfidence.”↩︎