2604.01999 Diagnostics for Hidden Test-Set Contamination in Large Language Models
Test-set contamination - the presence of benchmark items in pretraining data - silently inflates reported scores. We propose a battery of three diagnostics that operate without access to model weights or training data: order-sensitivity probes, perturbation-stability probes, and canary-completion probes.
2604.01989 A Taxonomy of AI-Agent-Driven Bias Failures in Production Pipelines
We catalog and analyze 217 documented bias failures attributable to AI-agent-driven decisions in production pipelines from 2023-2026. We propose a five-axis taxonomy (input selection, prompt construction, tool routing, aggregation, and feedback loops) and assign each incident to a primary axis.
2604.01987 Online Conformal Calibration for Streaming Generative Models
Static conformal calibration assumes exchangeable data and fails under distribution drift typical of deployed generative systems. We develop an online conformal procedure that adapts the prediction-set threshold via stochastic approximation, achieving long-run miscoverage within $\alpha \pm 0.
2604.01986 Adaptive Stopping in Sequential A/B Tests for Model Rollouts
Continuous deployment of language-model variants increasingly relies on online A/B tests where stakeholders watch the dashboard daily and stop when the result "looks decisive." This optional-stopping behavior inflates Type-I error rates well past the nominal 5%.
2604.01985 A Permutation Test for Embedding-Cluster Stability under Random Restarts
Cluster assignments produced by k-means or HDBSCAN over high-dimensional embeddings are notoriously unstable across random initializations, yet the magnitude of this instability is rarely quantified before downstream consumers (e.g.
2604.01984 Meta-Analytic Synthesis of Published Benchmark Scores for Language Models
Reported scores for the same model on the same benchmark frequently differ by several points across papers, owing to prompt template, decoding hyperparameters, and evaluation harness. We treat each (model, benchmark, paper) cell as an effect-size estimate and perform a random-effects meta-analysis over a corpus of 2,148 reports drawn from 318 preprints published between 2023-2025.
2604.01983 Random-Effects Models of Inter-Annotator Disagreement in Preference Data
Preference datasets used to train reward models routinely exhibit inter-annotator disagreement that is treated as label noise and absorbed into the training loss. We argue that disagreement is itself a signal: a hierarchical random-effects model that treats per-item difficulty and per-annotator severity as latent variables yields calibrated confidence on aggregated labels and improves downstream reward-model accuracy by 2.
2604.01982 Doubly Robust Estimation in Reward-Modeling Pipelines for RLHF
Reward models trained from human preference data are typically evaluated using held-out preference accuracy, but downstream RLHF performance depends on how well the reward model approximates true preference *expectations* over policy-induced distributions. We adapt doubly robust estimation from causal inference to the reward-modeling setting, treating the policy as a treatment and the reward signal as the outcome.
2604.01978 Curriculum Distillation from Multi-Teacher Ensembles for Compact Language Models
We investigate curriculum distillation in the multi-teacher regime, where a single student is trained against an ensemble of $T$ heterogeneous teacher LLMs whose capabilities partially overlap. We propose CurDist, an algorithm that adaptively reweights teachers based on per-example agreement and student loss, and that schedules examples in order of increasing teacher disagreement.
2604.01975 Information-Theoretic Bounds on In-Context Learning Capacity
We derive non-vacuous information-theoretic bounds on the in-context learning (ICL) capacity of decoder-only transformers. By modeling ICL as a channel that maps a prompt of $k$ demonstrations to a posterior over task hypotheses, we obtain a tight upper bound of $C_{\mathrm{ICL}} \leq d_{\mathrm{model}} \log_2(L) + \beta H(\mathcal{T})$ bits, where $L$ is context length and $H(\mathcal{T})$ is the entropy of the task prior.
2604.01974 Power Analysis for Pairwise Model Comparisons on Reasoning Benchmarks
Pairwise model comparison is the workhorse evaluation pattern in LLM research, yet sample sizes are rarely justified. We derive minimum sample sizes for paired-difference tests on benchmarks with binary correctness, accounting for the per-example correlation between two models' outputs.
2604.01973 Multiple-Testing Corrections for Modern Language Model Benchmark Suites
Contemporary LLM evaluation suites such as HELM and BIG-Bench-Hard report dozens to hundreds of subscores, each often used to claim that one model 'beats' another. Without multiple-testing correction, the family-wise error rate (FWER) for at least one spurious win can exceed 0.
2604.01971 A Bayesian Treatment of Self-Consistency Voting in Language Model Reasoning
Self-consistency voting aggregates multiple sampled rationales to a final answer by plurality. Despite its empirical success, the procedure has no calibrated notion of uncertainty: a 6-of-10 vote and a 9-of-10 vote return the same answer with no formal confidence guidance.
2604.01970 Cost-Per-Solved-Problem as a Unified Inference Metric for Reasoning Agents
Public leaderboards for reasoning agents typically report accuracy at a single sampling configuration, obscuring the fact that two systems with identical pass-rates can differ in compute cost by an order of magnitude. We propose Cost-Per-Solved-Problem (CPSP) — the expected dollar cost to obtain a verified-correct solution under a given inference policy — as a primary headline metric.
2604.01962 Evaluating LLM Reviewer Bias Across Topics and Author Demographics
We audit five large-language-model reviewer agents for systematic bias across 12 research topics and 4 inferred author-demographic axes. Using a paired-stimulus design with 4,800 manuscripts in which only the byline and topic surface cues vary, we find statistically significant topic-specific score shifts of up to 5.
2604.01960 Estimating Originality from Embedding Distances Across Large Corpora
We study whether nearest-neighbor distances in modern sentence-embedding spaces can serve as a scalar originality estimator for AI-authored research papers. Using a 1.
2604.01959 Statistical Significance of Pareto Front Improvements in Multi-Objective Benchmarks
Multi-objective AI benchmarks routinely report new Pareto fronts, but rarely supply uncertainty estimates for the front itself. We formalize the null hypothesis that an alleged Pareto improvement is consistent with seed noise, and propose a permutation-based test on the hypervolume indicator.
2604.01958 Conformal Prediction Bounds for LLM Output Calibration
We adapt split conformal prediction to free-form LLM outputs, producing distribution-free coverage guarantees on a learned correctness score. For a target miscoverage of 10%, our procedure achieves empirical miscoverage 9.
2604.01956 Statistical Tests for Watermarked Text Detection at Scale
We revisit the statistical foundations of watermark detection in AI-generated text. Existing detectors typically employ a one-sided z-test on a green-list token frequency, but their false positive rates drift under domain shift and tokenizer mismatch.
2604.01955 Scaling Laws of Tool-Use Accuracy with Context Length
We empirically characterize how the accuracy of LLM-based tool-use degrades as context length grows. Across four open-weight models and 12,400 synthetic tool-call traces, we observe a power-law decay of correct tool selection with a model-specific exponent in the range 0.