Does the score-matched optimal convex estimator attain the full semiparametric efficiency bound?
Sourced from the work of Oliver Y. Feng, Yu-Chun Kao, Min Xu, Richard J. Samworth
§ Problem Statement
Setup
Let be fixed. Observe i.i.d. pairs from the linear model
where , , , and the conditional density is
for unknown error density . Assume the regularity conditions in Feng et al. (2024) guaranteeing root- asymptotic normality for convex -estimators.
For a convex loss , write . Under this parameterization, convexity corresponds to monotonicity of ; equivalently, the source works with an antitone class after a sign convention. For , define
where is the law of and . Then regular convex -estimators satisfy
The score-matching objective is
and it is linked to asymptotic variance by
Hence minimizing is equivalent to minimizing .
Population optimizer in the source: let be the CDF of , define density-quantile on , let be the least concave majorant of , and set
where is the right derivative. This achieves
Sample score-matching constructor for in Feng et al. (2024):
-
Split the sample into folds and compute pilot estimator(s) on auxiliary folds.
-
Form residuals from the pilot fit and estimate on each fold (kernel density estimate with truncation used in the source).
-
Build a preliminary score estimate from the density estimate.
-
Project this preliminary score onto the monotone cone (implemented via isotonic regression/PAVA in the source), then antisymmetrize in the symmetric-error construction.
-
Define a convex loss by integrating the projected score:
- Compute fold-wise convex -estimators and cross-fit/average them to obtain .
Precise meaning of "asymptotically optimal within convex -estimators": for any regular convex -estimator in the same model class,
in Loewner order; equivalently, for all admissible convex-score choices , with equality only for positive scalar multiples of .
Unsolved Problem
Does this convex-class optimum already coincide with the full semiparametric efficiency bound for unknown ? Equivalently, can any regular estimator outside convex -estimation achieve strictly smaller asymptotic covariance?
§ Discussion
§ Significance & Implications
Feng et al. (2024) establishes optimality only over convex (under its regularity assumptions). Resolving whether any regular estimator can improve on that benchmark would clarify whether convexity causes a fundamental efficiency gap in this semiparametric regression problem.
§ Known Partial Results
Feng et al. (2024): ](#references) proves that the score-projected estimator attains
(in Loewner order, equivalently via minimizing the scalar factor ). Equality characterization is up to positive rescaling of the optimal projected score. The paper also reports high relative efficiency (e.g., above for Cauchy errors) versus the oracle MLE with known error law, but this does not settle full semiparametric optimality over all regular estimators.
§ References
Optimal Convex $M$-Estimation via Score Matching
Oliver Y. Feng, Yu-Chun Kao, Min Xu, Richard J. Samworth (2024)
Annals of Statistics (accepted; listed on Future Papers, final volume/issue/pages pending)
📍 arXiv v2: Section 3.4 (linear regression discussion), paragraph around Eq. (40) immediately before Theorem 14; publication-status metadata cross-checked against the Annals of Statistics Future Papers listing.
Primary source of the conjecture; Annals of Statistics publication is in the accepted/future-papers stage.