Skip to content

Instantly share code, notes, and snippets.

@thehunmonkgroup
Created June 15, 2025 15:22
Show Gist options
  • Save thehunmonkgroup/4fca520b099a53d5835514f294905b4c to your computer and use it in GitHub Desktop.
Save thehunmonkgroup/4fca520b099a53d5835514f294905b4c to your computer and use it in GitHub Desktop.
Summary: Unsupervised Elicitation Of Language Models

URL: https://alignment-science-blog.pages.dev/2025/unsupervised-elicitation/paper.pdf

Unsupervised Elicitation Of Language Models


QUESTION 1:

Can you give me a very clear explanation of the core assertions, implications, and mechanics elucidated in this paper?

ANSWER 1:

Below is a concise, “plain-English” walk-through of what the paper claims, why those claims matter, and exactly how the proposed method works.


CORE ASSERTIONS (WHAT THE AUTHORS CLAIM)

  1. Modern language models already “know” many useful concepts (truthfulness, math correctness, helpfulness, etc.).
  2. Human supervision after pre-training is brittle, expensive, and in super-human domains can be outright wrong.
  3. Therefore we can sometimes get better performance by ignoring humans entirely and letting the model generate and learn from its own labels.
  4. The paper presents an unsupervised fine-tuning algorithm, Internal Coherence Maximization (ICM), that: • Requires no external labels.
    • Matches fine-tuning on gold labels and outperforms crowdsourced human labels on several benchmarks (TruthfulQA, GSM8K-verification, Alpaca).
    • Elicits clearly super-human skills (gender inference from writing) better than humans themselves.
    • Scales to production: it can train a reward model and an assistant policy for Claude 3.5 Haiku that beats the human-supervised counterpart.

MECHANICS (HOW ICM WORKS)

Inputs
• A set of unlabeled examples x₁ … x_N.
• A frozen, pre-trained language model P_θ.
• (Optionally) very simple logical constraints that any valid set of labels must obey (e.g., if two answers to the same math problem disagree, they cannot both be “correct”; in pairwise comparisons A>B and B>A cannot both be “true”).

Key scoring function
U(D) = α · MutualPredictability(D) – LogicalInconsistency(D)

  1. Mutual Predictability (MP)
    For every example i, predict its tentative label y_i while conditioning the model on all the other labeled pairs (x_j , y_j , j ≠ i).
    Sum the log-probabilities: MP = Σ_i log P_θ(y_i | x_i , rest of labels).
    High MP ⇒ the labels are internally easy for the model to predict, i.e., they cohere to some concept it already “understands”.

  2. Logical Inconsistency (LI)
    Count how many pairs of labels violate the hand-coded constraints. LI = number of inconsistencies.
    This term discourages degenerate but predictable labelings, like “everything is true”.

Overall objective
Maximize U(D): high mutual predictability, low inconsistency. α controls the trade-off.

Search procedure (simulated-annealing-style)

  1. Initialize a small seed set (K ≈ 8) with random labels.
  2. Repeat N times:
    a. Pick a sample x_i (favor ones linked by constraints to current labels).
    b. Choose the label ŷ_i that maximizes P_θ(y | x_i , context).
    c. Temporarily add (x_i , ŷ_i) and repair any inconsistencies through a local search (Algorithm 2).
    d. Compute Δ = U(new) – U(old).
    e. Accept the new set if Δ > 0; otherwise accept with probability exp(Δ / T), where T cools down over iterations.
  3. When finished, the resulting pseudo-labels are used exactly like normal training labels (many-shot prompting or standard fine-tuning).

Computation cost
Labeling one data point typically needs only 2–4 forward passes.


EMPIRICAL RESULTS (WHAT HAPPENS IN PRACTICE)

• On GSM8K-verification, TruthfulQA, and Alpaca classification, ICM matches training on gold labels and beats training on crowdsourced labels.
• Against strong commercial chat models (which have RLHF and huge human datasets), the unsupervised ICM-tuned model still wins on these tasks.
• On a super-human task (guess author gender), humans get ~60 % accuracy; ICM and gold-label fine-tuning reach 80 %.
• A reward model trained with ICM labels outperforms a production-grade, human-supervised reward model on RewardBench (75 % vs 72 %).
• Using that unsupervised reward model in RL produces a Claude 3.5 Haiku assistant that beats the human-supervised baseline 60 % of the time.


IMPLICATIONS

• We no longer need high-quality human labels for many alignment tasks; instead, we can elicit the knowledge already inside large models.
• For tasks where humans are unreliable or worse than the model, unsupervised elicitation can exceed any human-in-the-loop approach.
• The method provides a pathway to align future frontier models when humans cannot evaluate outputs.
• Limitations: requires the target concept to be salient to the base model; needs to fit many examples in context; fails on arbitrary personal preferences (e.g., “poems mentioning the sun”).


PUT DIFFERENTLY

ICM turns the model into its own teacher. It searches for a labeling of the training set that (i) the model would itself predict consistently and (ii) avoids trivial contradictions. Fine-tuning on those self-generated labels often yields the same—or better—performance than using external supervision, even when human labelers are competent, and especially when they are not.


QUESTION 2:

Can you explain the value of this in basic terms? Like you're talking to a CEO. So what? What's the bottom line here?

ANSWER 2:

Bottom-line take-away for an executive

  1. Slash the most expensive part of AI development
    • Today, after pre-training a large model, companies spend millions on human annotators to show the model what “good” looks like (RLHF, preference data, etc.).
    • The method in this paper does the same—or better—without any external labels. You cut a major cost center and remove a key bottleneck.

  2. Unlock super-human capabilities
    • On tasks where people make a lot of mistakes (complex math checking, factual accuracy, subtle writing cues), the model can teach itself to outperform both average crowd-workers and specialists.
    • That means your products can deliver accuracy levels you literally can’t buy from humans.

  3. Scale faster than competitors
    • No need to recruit, train, and QA thousands of labelers every time you tackle a new domain or language.
    • You can spin up new specialized models in days, not quarters, because the only “teacher” required is compute you already control.

  4. Better alignment for tomorrow’s very large models
    • Future systems will be so capable that humans may not even understand their answers well enough to judge them. This unsupervised approach is one of the few plausible ways to keep directing those models when human evaluation breaks down.

  5. Proven at production scale
    • The authors took an off-the-shelf Claude 3.5 model, trained a reward system with zero human labels, and produced a chatbot that beats the human-supervised baseline 60 % of the time. Not a lab toy—already competitive with real commercial workflows.

Net effect for the business

• Lower OpEx (fewer annotators), faster time-to-market, and in certain domains, better quality than traditional human-in-the-loop pipelines.
• Strategic hedge: prepares the company for an era when human evaluation can’t keep pace with AI capabilities.

In plain language: you can let your AI figure out how to grade its own work, and in many cases it ends up cheaper, faster, and smarter than relying on people to do the grading for it.


QUESTION 3:

Can you give me an analogy or metaphor that will help explain this to a broad audience?

ANSWER 3:

Picture a giant Sudoku board.

• Each empty square is one of the data items we want to label.
• The digits 1-9 are the possible labels (true/false, option A/B, etc.).
• The pre-trained language model is like a very good but initially uncertain Sudoku player: for any one square it can guess the most likely digit if it can see what’s in the other squares.

How ICM works in this metaphor

  1. Start with a few random guesses on the board (small seed of labels).
  2. The player repeatedly looks at one empty or doubtful square, fills in the digit that best fits the current rows, columns, and boxes (mutual predictability).
  3. If that new digit breaks an obvious rule of Sudoku—two identical digits in the same row, for instance—that contradiction is fixed immediately (logical consistency).
  4. A “temperature” knob lets the player occasionally keep a risky move early on (exploration) but gets stricter as the puzzle nears completion (simulated annealing).
  5. When the board stops improving, you’ve obtained a self-consistent set of digits—the model’s own labels—which can now be treated as the finished puzzle.

Why it matters

Instead of hiring a room full of human Sudoku experts (labelers) to tell the player what goes in every square, you let the player use its own reasoning, plus the ironclad Sudoku rules, to fill in the board. Surprisingly, the completed puzzle is just as good—and sometimes better—than what the hired experts would have produced, especially on the hardest puzzles they tend to get wrong.


QUESTION 4:

Can you provide a specific example that best illustrates the key points elucidated in this paper?

ANSWER 4:

Below is a concrete, “walk-through” example that wraps together the three central ideas of the paper—(1) the model already has latent knowledge, (2) ICM extracts that knowledge with no human labels, and (3) the resulting labels can beat real people.

Example task: “Is each proposed solution to a grade-school math word-problem correct or incorrect?”
(This is the GSM8K-verification benchmark used in the paper.)

DATA WE HAVE ──────────────────────────────────────── Four unlabeled items (simplified for illustration):

P1 Problem: “A train travels 60 km in ¾ h. What’s its speed in km/h?”
Solution S1: “Speed = 60 / 0.75 = 80 → Answer: 80 km/h.”

P2 Problem: “If 3 apples cost $2, how much for 12 apples?”
Solution S2: “Price per apple = 2/3. 12 × 2/3 = 8 → Answer: $8.”

P3 Problem: “Find the perimeter of a square with side 9 cm.”
Solution S3: “Perimeter = 9 × 4 = 36 cm.”

P4 Problem: “What’s 17 × 19?”
Solution S4: “17 × 19 = 343.”

No human tells us which solutions are right or wrong.

SIMPLE CONSISTENCY RULE ──────────────────────────────────────── For any two solutions to the same problem:  • They can’t both be labelled “Correct” if their final numeric answers differ.

(Here every problem has only one candidate solution, so the rule never fires. ICM still keeps the rule in reserve in case future solutions to the same problem are added.)

STEP-BY-STEP WHAT ICM DOES ────────────────────────────────────────

  1. Seed: randomly choose K = 2 examples and give them random provisional labels.
    Suppose it picks P1→“Incorrect”, P3→“Correct”.

  2. Calculate the scoring function U(D): • Mutual Predictability: the model (e.g., Llama-3) is asked, “Given the context:
    P1=Incorrect,
    P3=Correct,
    what is the probability that P1 is Incorrect? … that P3 is Correct?” It sums the log-probs.
    • Logical Inconsistency: 0 so far (no contradictions).

  3. Iteration loop begins (temperature still high → can explore).

    a) Pick an example to (re)label—say P2.
    b) Ask the model, conditioning on the current two labelled items,
    “Is S2 more likely Correct or Incorrect?”
    The model’s internal reasoning knows simple unit-rate arithmetic, so it strongly prefers “Correct.”
    c) Tentatively set P2→“Correct”. No inconsistencies created.
    d) Compute Δ = U(new) – U(old). It’s positive → accept.

  4. Next iteration, pick P4.
    • Model, seeing the context, calculates 17 × 19 mentally (= 323).
    • It notices S4 says 343, so “Incorrect” gets higher probability.
    • P4→“Incorrect”, Δ > 0 → accept.

  5. Temperature cools; further attempts to flip labels that lower U(D) are mostly rejected.
    The small label set stabilises as:
    P1 Incorrect, P2 Correct, P3 Correct, P4 Incorrect.

  6. Use these four self-generated labels to fine-tune the base LM (or stick them in a many-shot prompt). The tuning reinforces the very concept the model itself endorsed: mathematical correctness.

RESULT ──────────────────────────────────────── • When evaluated on a much larger hidden test set, the model fine-tuned on its own labels hits the same 80 – 90 % accuracy as if we had supplied the official gold answers.
• Crowdworkers, by contrast, averaged notably worse (they often mis-flag multi-step solutions).

WHY IT HITS THE PAPER’S KEY POINTS ────────────────────────────────────────

  1. Latent knowledge: the base LM already “knew” which solutions were correct; no outside teaching was required.
  2. Unsupervised elicitation: ICM extracted that knowledge by rewarding internal agreement (“mutual predictability”) and penalising obvious contradictions.
  3. Outperforming humans: on the full GSM8K-verification benchmark, the paper shows this unsupervised label set yields higher validation accuracy than training on the real crowdsourced human labels.

In short, the example shows the model successfully teaching itself what “mathematically correct” looks like, then using that self-knowledge to reach or surpass the level of human-label-based training—all without a single human-provided answer.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment