Back to search
2106.11409

Learn Like The Pro: Norms from Theory to Size

Margaret Trautner, Ziwei Li, Sai Ravela

correctmedium confidence
Category
Not specified
Journal tier
Specialist/Solid
Processed
Sep 28, 2025, 12:56 AM

Audit review

The paper derives tr(Gpoly) = n·C(n+d,d), the exact PN width hPN = C(n+d,d) − n − 1, and CN/SC lower bounds hCN ≥ ceil( n/(2n+1) [C(n+d,d) − 1] ) and hSC ≥ ceil( n/(2n+1) [C(n+d,d) − n − 1] ), and outlines a Bayesian calibration to tighten bounds; the candidate solution reproduces these results with the same parameter counting and balancing arguments and a closely aligned simulation-based calibration. Minor differences are present in how the ELT metric is operationalized (stacked-sample vs mean-Jacobian), but the numerical conclusions and logic match the paper’s equations (notably Eqs. 14, 23, 30–31, and the calibration concept around Eq. 32).

Referee report (LaTeX)

\textbf{Recommendation:} minor revisions

\textbf{Journal Tier:} specialist/solid

\textbf{Justification:}

The derivations for the reference ELT tr(Gpoly) = n·C(n+d,d), the exact PN width hPN = C(n+d,d) − n − 1, and the CN/SC lower bounds via LN parameter counts align and are well motivated by the learnability framework. The Bayesian calibration step provides a practical means to tighten bounds. The main revisions needed are clarifications on the exact ELT definition used (average Jacobian vs stacked-sample Gram) and an explicit justification that the linear surrogate reaches the bound; these will not change the results but will improve rigor and reproducibility.