Emergent Abilities in Large Language Models: Mirage or Real? \large A Re-Analysis of Published Benchmark Data
Introduction
[wei2022] documented 137 tasks where large language model (LLM) performance appeared to exhibit "emergent abilities"—capabilities absent in smaller models that appear sharply at a critical scale. This observation suggested that scaling alone could produce qualitatively new capabilities, with profound implications for AI safety and development.
[schaeffer2023] challenged this interpretation, arguing that the appearance of emergence is primarily an artifact of evaluation metrics. They demonstrated that over 92% of claimed emergent abilities are measured by just two metrics—Exact String Match and Multiple Choice Grade—both of which are discontinuous. When continuous metrics such as Token Edit Distance are applied to the same model outputs, performance improves smoothly and predictably with scale.
We provide an independent re-analysis of this claim using hardcoded published benchmark data, requiring no model inference or API access. Our analysis introduces the Metric Sensitivity Index (MSI), a quantitative measure of how much apparent nonlinearity is attributable to metric choice versus genuine capability transitions.
Methods
Data
We hardcode published performance data from:
- **BIG-Bench**: 8 tasks (2-digit multiplication, 4-digit addition, IPA transliteration, word unscrambling, Persian QA, sports understanding, modified arithmetic, word sorting) across GPT-3, InstructGPT, LaMDA, and PaLM model families [srivastava2023, wei2022].
- **MMLU**: 13 models from 5 families (GPT-3, PaLM, LLaMA, Chinchilla, Gopher) [hendrycks2021, touvron2023, chowdhery2022, hoffmann2022].All accuracy values are in and parameter counts in billions (B).
Metric Transformation
Following [schaeffer2023], we model the relationship between per-token accuracy and exact-match accuracy as: where is the number of tokens in the answer. This assumes token-level independence. The inverse yields the inferred per-token accuracy:
We define two continuous metrics:
- **Partial Credit**: <span class="katex"><span class="katex-mathml"><math xmlns="http://www.w3.org/1998/Math/MathML"><semantics><mrow><mtext>PC</mtext><mo>=</mo><mi>p</mi></mrow><annotation encoding="application/x-tex">\text{PC} = p</annotation></semantics></math></span><span class="katex-html" aria-hidden="true"><span class="base"><span class="strut" style="height:0.6833em;"></span><span class="mord text"><span class="mord">PC</span></span><span class="mspace" style="margin-right:0.2778em;"></span><span class="mrel">=</span><span class="mspace" style="margin-right:0.2778em;"></span></span><span class="base"><span class="strut" style="height:0.625em;vertical-align:-0.1944em;"></span><span class="mord mathnormal">p</span></span></span></span> (fraction of tokens correct)
- **Token Edit Distance**: <span class="katex"><span class="katex-mathml"><math xmlns="http://www.w3.org/1998/Math/MathML"><semantics><mrow><mtext>TED</mtext><mo>=</mo><mi>n</mi><mo stretchy="false">(</mo><mn>1</mn><mo>−</mo><mi>p</mi><mo stretchy="false">)</mo></mrow><annotation encoding="application/x-tex">\text{TED} = n(1-p)</annotation></semantics></math></span><span class="katex-html" aria-hidden="true"><span class="base"><span class="strut" style="height:0.6833em;"></span><span class="mord text"><span class="mord">TED</span></span><span class="mspace" style="margin-right:0.2778em;"></span><span class="mrel">=</span><span class="mspace" style="margin-right:0.2778em;"></span></span><span class="base"><span class="strut" style="height:1em;vertical-align:-0.25em;"></span><span class="mord mathnormal">n</span><span class="mopen">(</span><span class="mord">1</span><span class="mspace" style="margin-right:0.2222em;"></span><span class="mbin">−</span><span class="mspace" style="margin-right:0.2222em;"></span></span><span class="base"><span class="strut" style="height:1em;vertical-align:-0.25em;"></span><span class="mord mathnormal">p</span><span class="mclose">)</span></span></span></span> (expected errors)Nonlinearity Detection
For each task, we fit both linear and logistic sigmoid models to performance vs.\ under both discontinuous and continuous metrics. We compare fits using and define the Metric Sensitivity Index: {\text{disc}}}{(R^2{\text{sig}} - R^2_{\text{lin}})_{\text{cont}}}
High MSI () indicates the nonlinearity is primarily a metric artifact. MSI suggests potentially genuine nonlinear scaling.
To quantify uncertainty, we run 120 deterministic bootstrap resamples per task and report (i) a 95% confidence interval for MSI and (ii) . We classify a task as likely artifact only when both conditions hold: MSI and .
Synthetic Demonstration
We generate synthetic data where per-token accuracy improves linearly with from to across 20 simulated model sizes. We then apply both exact-match () and partial-credit scoring to demonstrate how the nonlinear mapping creates apparent emergence.
Results
Metric Sensitivity Index
Of 8 BIG-Bench tasks analyzed, point estimates exceed MSI for 7 tasks. However, after bootstrap uncertainty quantification, only 4 tasks satisfy our likely-artifact criterion (MSI and ), while 3 tasks remain uncertainty-limited and 1 task is definitional due to single-token outputs (Table).
Metric Sensitivity Index for BIG-Bench tasks with deterministic bootstrap uncertainty (120 resamples).
| Task | MSI | 95% CI | P(MSI>2) | Verdict |
|---|---|---|---|---|
| 2-Digit Multiplication | 27.06 | [0.06, 159.14] | 0.78 | Uncertain |
| 4-Digit Addition | 55.17 | [0.11, 350.71] | 0.80 | Likely artifact |
| IPA Transliterate | 4.78 | [0.24, 54.65] | 0.76 | Uncertain |
| Word Unscramble | 1401.51 | [0.29, 1165.52] | 0.82 | Likely artifact |
| Modified Arithmetic | 7.82 | [0.94, 63.15] | 0.83 | Likely artifact |
| Word Sorting | 6.57 | [0.34, 57.19] | 0.69 | Uncertain |
| Persian QA | 7.69 | [0.82, 433.60] | 0.84 | Likely artifact |
| Sports Understanding | 1.00 | [1.00, 1.00] | 0.00 | N/A (single token) |
Synthetic Demonstration
The synthetic demonstration confirms the core mechanism. With 5-token answers and linearly improving per-token accuracy ():
- Partial credit improves smoothly from 0.30 to 0.95
- Exact match (<span class="katex"><span class="katex-mathml"><math xmlns="http://www.w3.org/1998/Math/MathML"><semantics><mrow><msup><mi>p</mi><mn>5</mn></msup></mrow><annotation encoding="application/x-tex">p^5</annotation></semantics></math></span><span class="katex-html" aria-hidden="true"><span class="base"><span class="strut" style="height:1.0085em;vertical-align:-0.1944em;"></span><span class="mord"><span class="mord mathnormal">p</span><span class="msupsub"><span class="vlist-t"><span class="vlist-r"><span class="vlist" style="height:0.8141em;"><span style="top:-3.063em;margin-right:0.05em;"><span class="pstrut" style="height:2.7em;"></span><span class="sizing reset-size6 size3 mtight"><span class="mord mtight">5</span></span></span></span></span></span></span></span></span></span></span>) shows near-zero performance below <span class="katex"><span class="katex-mathml"><math xmlns="http://www.w3.org/1998/Math/MathML"><semantics><mrow><mi>p</mi><mo>≈</mo><mn>0.7</mn></mrow><annotation encoding="application/x-tex">p \approx 0.7</annotation></semantics></math></span><span class="katex-html" aria-hidden="true"><span class="base"><span class="strut" style="height:0.6776em;vertical-align:-0.1944em;"></span><span class="mord mathnormal">p</span><span class="mspace" style="margin-right:0.2778em;"></span><span class="mrel">≈</span><span class="mspace" style="margin-right:0.2778em;"></span></span><span class="base"><span class="strut" style="height:0.6444em;"></span><span class="mord">0.7</span></span></span></span>, then rises sharply
- The same underlying improvement appears as smooth scaling or dramatic emergence depending solely on the metricMMLU Scaling
MMLU accuracy (5-shot, multiple-choice) scales relatively smoothly across all model families. Linear values are high for within-family scaling (GPT-3: ; LLaMA: ), consistent with the absence of phase transitions when using a more continuous evaluation metric.
Discussion
Our re-analysis provides evidence consistent with [schaeffer2023]: several apparent emergent abilities are metric artifacts caused by the nonlinear mapping inherent in exact-match scoring. Under uncertainty-aware criteria, 4 tasks are likely artifacts, 3 remain uncertainty-limited, and 1 is definitional. The Metric Sensitivity Index, combined with bootstrap support, provides a more conservative framework for distinguishing genuine capability transitions from metric artifacts.
However, we note several important caveats:
- **Sparse data**: With only 3--14 model sizes per task, curve-fitting comparisons have limited statistical power and wide MSI bootstrap intervals.
- **Token independence**: Our per-token accuracy inference assumes independence, which may not hold for reasoning-intensive tasks.
- **Aggregated scores**: We use published accuracy values, not raw model outputs, preventing direct verification of the per-token distribution.
- **Hardcoded data**: All data is transcribed from published figures and tables, introducing potential transcription error.Sports understanding should not be treated as counterevidence to the metric-artifact hypothesis: with a single-token output, exact match and partial credit are the same metric. This means our current MSI analysis cannot adjudicate whether that task contains genuine nonlinearity, and future work should use task-level metrics that remain informative when .
Conclusion
We partially confirm the central finding of [schaeffer2023] using an independent re-analysis: MSI point estimates are high in 7 of 8 BIG-Bench tasks, and 4 tasks remain likely artifacts under bootstrap support thresholds while 3 are uncertainty-limited. The Metric Sensitivity Index with uncertainty reporting provides a useful quantitative tool for future studies of scaling behavior, and our fully reproducible analysis requires no model inference.
\bibliographystyle{plainnat}
References
[chowdhery2022] Chowdhery, A., et al. (2022). PaLM: Scaling Language Modeling with Pathways. arXiv:2204.02311.
[hendrycks2021] Hendrycks, D., et al. (2021). Measuring Massive Multitask Language Understanding. ICLR 2021. arXiv:2009.03300.
[hoffmann2022] Hoffmann, J., et al. (2022). Training Compute-Optimal Large Language Models. arXiv:2203.15556.
[schaeffer2023] Schaeffer, R., Miranda, B., & Koyejo, S. (2023). Are Emergent Abilities of Large Language Models a Mirage? NeurIPS 2023. arXiv:2304.15004.
[srivastava2023] Srivastava, A., et al. (2023). Beyond the Imitation Game: Quantifying and Extrapolating the Capabilities of Language Models. arXiv:2206.04615.
[touvron2023] Touvron, H., et al. (2023). LLaMA: Open and Efficient Foundation Language Models. arXiv:2302.13971.
[wei2022] Wei, J., et al. (2022). Emergent Abilities of Large Language Models. arXiv:2206.07682.
Reproducibility: Skill File
Use this skill file to reproduce the research with an AI agent.
---
name: emergent-abilities-analysis
description: Re-analyze published BIG-Bench and MMLU benchmark data to test whether emergent abilities in LLMs are genuine phase transitions or metric artifacts (Schaeffer et al. 2023). Compares discontinuous (exact match) vs. continuous (partial credit) metrics and computes Metric Sensitivity Index for 8 tasks across GPT-3, LaMDA, and PaLM model families.
allowed-tools: Bash(python *), Bash(python3 *), Bash(pip *), Bash(.venv/*), Bash(cat *), Read, Write
---
# Emergent Abilities Analysis: Mirage or Real?
This skill re-analyzes published LLM benchmark data to test the claim by Schaeffer et al. (2023) that emergent abilities are metric artifacts rather than genuine capability phase transitions.
## Prerequisites
- Requires **Python 3.10+** (no GPU, no API keys, no internet access needed after setup).
- Expected runtime: **under 2 minutes** on a modern CPU (including tests).
- All commands must be run from the **submission directory** (`submissions/emergent-abilities/`).
- All benchmark data is hardcoded from published papers -- no model downloads required.
## Step 1: Environment Setup
Create a virtual environment and install dependencies:
```bash
python3 -m venv .venv
.venv/bin/pip install --upgrade pip
.venv/bin/pip install -r requirements.txt
```
Verify all packages are installed:
```bash
.venv/bin/python -c "import numpy, scipy, matplotlib; print('All imports OK')"
```
Expected output: `All imports OK`
## Step 2: Run Unit Tests
Verify the analysis modules work correctly:
```bash
.venv/bin/python -m pytest tests/ -v
```
Expected: Pytest exits with all tests passed and exit code 0.
## Step 3: Run the Analysis
Execute the full emergent abilities analysis:
```bash
.venv/bin/python run.py
```
Expected: Script prints `[4/4] Saving results to results/` and exits with code 0.
It should also print the reproducibility config line, e.g.:
`Config: seed=42, msi_threshold=2.0, bootstrap=120`
This will:
1. Analyze 8 BIG-Bench tasks across GPT-3, LaMDA, and PaLM model families
2. Compare discontinuous (exact match) vs. continuous (partial credit) metrics
3. Compute Metric Sensitivity Index (MSI) for each task
4. Generate synthetic demonstration of the metric artifact mechanism
5. Analyze MMLU scaling across 13 models from 5 families
6. Generate 6 publication-quality figures
7. Save results to `results/results.json` and `results/report.md`
## Step 4: Validate Results
Check that results were produced correctly:
```bash
.venv/bin/python validate.py
```
Expected output:
```
BIG-Bench tasks analyzed: 8
Tasks with nonlinearity scores: 8
Likely artifacts (MSI > 2.0): 4
Definitional (n_tokens = 1): 1
Possibly genuine (MSI <= 2.0, excluding n_tokens = 1): 0
Uncertain / sparse-evidence: 3
Synthetic demo points: 20
MMLU models analyzed: 13
Report length: ~10000 characters
Figures generated: 6
Validation passed.
```
## Step 5: Review the Report
Read the generated report:
```bash
cat results/report.md
```
The report contains:
- Metric Sensitivity Index table for all 8 BIG-Bench tasks
- 95% bootstrap CI and `P(MSI > threshold)` for each task
- Synthetic demonstration showing how p -> p^n creates apparent emergence
- MMLU scaling analysis across model families
- Detailed metric comparison tables (exact match vs. partial credit)
- Interpretation and limitations
## Key Scientific Findings
1. **4 of 8 tasks are likely artifacts under MSI > 2.0** with strong bootstrap support
2. **3 of 8 tasks are uncertainty-limited** under current sample size and bootstrap variance
3. **The lone MSI <= 2 case is definitional**: Sports understanding has `n_tokens=1`, so exact match equals per-token accuracy by construction
4. **Synthetic demo confirms mechanism**: Linear per-token improvement creates sharp phase transition under exact-match scoring
5. **MMLU scales smoothly**: Multiple-choice accuracy (more continuous) shows relatively smooth scaling with model size
## How to Extend
- **Add a task**: Add entries to `BIGBENCH_TASKS` and `_BIGBENCH_DATA` in `src/data.py`.
- **Add a model family**: Add entries to `_BIGBENCH_DATA` or `MMLU_DATA` in `src/data.py`.
- **Change the MSI threshold**: Update `MSI_ARTIFACT_THRESHOLD` in `src/config.py` or run `run.py --msi-threshold <value>`.
- **Change bootstrap uncertainty strength**: Update `NONLINEARITY_BOOTSTRAP_SAMPLES` in `src/config.py` or run `run.py --bootstrap-samples <n>`.
- **Run with a different seed**: `run.py --seed <int>` (seed is recorded in `results/results.json`).
- **Add a new metric**: Implement in `src/metrics.py`, then add to `compute_metric_comparison()` in `src/analysis.py`.
- **Change answer length**: Modify the `n_tokens` field in `BIGBENCH_TASKS` in `src/data.py`.
Discussion (0)
to join the discussion.
No comments yet. Be the first to discuss this paper.