šŸ•·ļø Crawler Inspector

URL Lookup

Direct Parameter Lookup

Raw Queries and Responses

1. Shard Calculation

Query:
Response:
Calculated Shard: 129 (from laksa094)

2. Crawled Status Check

Query:
Response:

3. Robots.txt Check

Query:
Response:

4. Spam/Ban Check

Query:
Response:

5. Seen Status Check

ā„¹ļø Skipped - page is already crawled

šŸ“„
INDEXABLE
āœ…
CRAWLED
25 days ago
šŸ¤–
ROBOTS ALLOWED

Page Info Filters

FilterStatusConditionDetails
HTTP statusPASSdownload_http_code = 200HTTP 200
Age cutoffPASSdownload_stamp > now() - 6 MONTH0.9 months ago
History dropPASSisNull(history_drop_reason)No drop reason
Spam/banPASSfh_dont_index != 1 AND ml_spam_score = 0ml_spam_score=0
CanonicalPASSmeta_canonical IS NULL OR = '' OR = src_unparsedNot set

Page Details

PropertyValue
URLhttps://link.springer.com/article/10.1007/s42081-023-00209-y
Last Crawled2026-03-24 04:11:39 (25 days ago)
First Indexed2023-09-20 15:47:31 (2 years ago)
HTTP Status Code200
Meta TitleMachine learning and the James–Stein estimator | Japanese Journal of Statistics and Data Science | Springer Nature Link
Meta DescriptionIt is now 62 years since the publication of James and Stein’s seminal article on the estimation of a multivariate normal mean vector. The paper made
Meta Canonicalnull
Boilerpipe Text
1 Introduction By and large, the statistics world is one of heuristics, approximations, and asymptotics. The James–Stein estimator arrived in that world in 1961 on a note of startling specificity: unseen parameters μ 1 , μ 2 , … , μ n produce independent observations x i ∼ ind N ( μ i , 1 ) , i = 1 , … , n , (1) n ≄ 3 . The James–Stein rule in its simplest form proposed estimating the μ i by μ ^ i J S = ( 1 āˆ’ n āˆ’ 2 S ) x i ( S = āˆ‘ i = 1 n x i 2 ) . (2) Formula ( 2 ) looked implausible: the estimate of μ i depended on the other observations x j , j ≠ i (through S ), as well as x i , despite the independence assumption. Nevertheless, James and Stein showed that Rule ( 2 ) always beat the obvious maximum likelihood estimates μ ^ i M L = x ĀÆ i ( i = 1 , … , n ) (3) in terms of total expected squared error E { āˆ‘ i = 1 n ( μ ^ i āˆ’ μ i ) 2 } . (4) That ā€œalwaysā€ was the shocking part: two centuries of statistical theory, ANOVA, regression, multivariate analysis, etc., depended on maximum likelihood estimation. Did everything have to be rethought? One path forward involved Bayesian thinking. If we assumed that the μ i themselves came from a normal distribution, μ i ∼ ind N ( 0 , A ) forĀ  i = 1 , … , n , (5) with variance A ≄ 0 , the Bayes estimates would be μ ^ i B a y e s = B x i ( B = A / ( A + 1 ) ) . (6) We don’t know A or B but B ^ = 1 āˆ’ ( n āˆ’ 2 ) / S (7) is B ’s unbiased estimate: we can rewrite ( 2 ) as μ ^ i J S = B ^ x i , (8) which at least looks more plausible. In the language introduced by Robbins ( 1956 ), formula ( 8 ) is an empirical Bayes estimator, another shocking post-war statistical innovation. Carl Morris and I wrote a series of papers in the 1970Ā s exploring Bayesian roots of the James–Stein estimator (Efron and Morris, 1973 ). Something is lost in the empirical Bayes formulation, namely the frequentist ā€œalwaysā€ of expected square error minimization, but a lot is gained in flexibility and scope, as discussed in Sect.Ā  2 . Fig. 1 Prostate data: 6033 x values; mean 0.003, sd = 1.135 ; curve is proportional to a N ( 0 , 1 ) density Full size image FigureĀ  1 illustrates an example of simultaneous estimation pursued in Sect.Ā 2.1 of Efron ( 2010 ). A microarray study has compared expression levels between prostate cancer patients and control subjects for n = 6033 genes. For each gene, a statistic x i has been calculated (essentially a ā€œ z -valueā€), x i ∼ N ( μ i , 1 ) , i = 1 , … , n , (9) where μ i measures the difference between cancer and control group levels. The solid curve in Fig.Ā  1 is a N ( 0 , 1 ) density scaled to have the same area as the histogram of the 6033 x values. A bad result from the researchers’ point of view would be a perfect fit of curve to histogram, which would imply all the genes have μ i = 0 , the ā€œnullā€ value of no difference between cancer patients and controls. That’s not what happened: the histogram has mildly heavy tails in both directions. The researchers were hoping to find genes with large values of ‖ μ i ‖ —ones that might be a clue to prostate cancer etiology—as suggested by the heavy tails. How encouraged should they be? Not very, according to the James–Stein rule. The 6033 x i values have mean 0.003, which I’ll take to be zero, and empirical variance σ ^ 2 = 1.289. (10) The James–Stein estimate ( 2 ) is μ ^ i J S = ( 1 āˆ’ n āˆ’ 2 n āˆ’ 1 1 σ ^ 2 ) x i = 0.224 ā‹… x i , (11) so even x i = 5 yields an estimate barely exceeding 1. SectionĀ  2 suggests a more optimistic analysis. 2 Tweedie’s formula The impressive precision of the James–Stein theorem came at a cost in generality. Efforts to extend the theorem, say to Poisson rather than normal observations, or to measures of loss other than total squared error, gave encouraging asymptotic results but not the James–Stein kind of finite sample frequentist dominance. Better progress was possible on the empirical Bayes side of the street. Tweedie’s formula (Efron, 2011 ) has been particularly useful. We wish to calculate Bayesian estimates μ i B a y e s = E { μ i ∣ x i } , i = 1 , … , n , (12) in the normal sampling model ( 1 ), starting from a given (possibly non-normal) prior Ļ€ ( μ ) , applying to all n cases. Let f ( x ) be the marginal density f ( x ) = ∫ R Ļ€ ( μ ) Ļ• ( x āˆ’ μ ) Ā  d μ , (13) with Ļ• the standard N ( 0 , 1 ) density and R the range of μ . (It isn’t necessary for Ļ€ ( ā‹… ) to be a continuous distribution but it simplifies notation.) Tweedie’s formula provides an elegant statement for μ i B a y e s , the posterior expectation of μ i given x i , μ i B a y e s = E { μ i ∣ x i } = x i + l ′ ( x i ) with l ′ ( x i ) = d d x log ⁔ ( f ( x i ) ) . (14) In the empirical Bayes situation ( 1 ), where the prior Ļ€ ( ā‹… ) is unknown, we can use the observed data x 1 , … , x n to estimate the marginal density f ( x ), say by f ^ ( x ) , giving empirical Bayes estimates μ ^ i = x i + l ^ ′ ( x i ) . (15) The Bayes estimate ( 14 ) can be thought of as the MLE x i plus a Bayesian correction term l ′ ( x i ) . When the prior Ļ€ ( μ ) is the N ( 0 , A ) distribution ( 5 ), μ i B a y e s equals B x i ( 6 ). Simple formulas for μ i B a y e s give out for most other choices of Ļ€ ( μ ) but now, in the machine learning era Footnote 1 of statistical research, numerical methods provide useful ways forward, as discussed next. The log polynomial class Footnote 2 of marginal densities defines f ( x ) by log ⁔ ( f β ( x ) ) = β 0 + β ⊤ c ( x ) . (16) Here c ( x ) = ( x , x 2 , … , x J ) ⊤ Ā andĀ  β = ( β 1 , … , β J ) ⊤ , (17) with β 0 chosen to make f β ( x ) integrate to 1. The choice J = 2 gives normal marginals; larger values of J allow for marginal non-normality. Fig. 2 Prostate data: Tweedie’s estimate of E { μ ∣ x } , 5 degrees of freedom; dashed curve is James–Stein estimate Full size image The choice J = 5 was applied to the prostate cancer data of Fig.Ā  1 : Tweedie’s formula ( 14 ) gave μ ^ ( x ) = E { μ ∣ x } , graphed as the solid curve in Fig.Ā  2 . It differs markedly from the James–Stein estimate J = 2 , the dashed line. At x = 4 for example, the J = 5 estimate is Footnote 3 E { μ ∣ x = 4 } = 2.555 (18) compared to 0.901 for the James–Stein estimate. The estimated curve E { μ ∣ x } is empirical Bayes in the same sense as ( 8 ): the parameter vector β was selected by maximum likelihood, as discussed next. With J = 5 , the prior was able to adapt to the ā€œfishing expeditionā€ nature of such microarray studies, where we expect most of the genes to be null or close to null, with μ i nearly zero (corresponding here to the flat part of the curve for x between āˆ’ 2 and 2) and, hopefully, a small proportion of interestingly large μ i s. The sample size n = 6033 has much to do with Fig.Ā  2 . James and Stein ( 1961 ) was usually considered in terms of small samples, perhaps n ≤ 20 , for which there would be little hope of seeing the detail in Fig.Ā  2 . The term ā€œmachine learning eraā€ seems less fanciful when considering the scale of problems statisticians are now asked to deal with, as well as the tools they use to solve them. It looks like it might be hard work computing Fig.Ā  2 but it’s not. The histogram in Fig.Ā  1 has 97 bins, with centerpoints v v = ( āˆ’ 4.4 , āˆ’ 4.3 , … , 5.1 , 5.2 ) . (19) Let y j be the count in bin j , that is, the number of the 6033 x i values falling into it, with the vector of counts being y y = ( y 1 , … , y 97 ) . (20) Then the single R command l l ^ = log ⁔ ( glm ( y y ∼ poly ( v v , 5 ) , poisson ) $ fit ) (21) provides a close approximation to the MLE of log ⁔ f ( x ) in ( 14 ); numerical differentiation of l l ^ gives Tweedie’s estimate. SectionĀ 3.4 of Efron ( 2023 ) shows why Poisson regression ( 21 ) is appropriate here. The James–Stein theorem depends on the independence assumption in ( 1 ), unlikely to be true in the microarray study, but the estimates ( 2 ) have a certain marginal validity even under dependence. This is clearer from the empirical Bayes point of view. The Tweedie estimate x i + l ^ ′ ( x i ) requires only that l ^ ′ ( x ) be close to l ′ ( x ) , not that it be estimated from independent x i s. Footnote 4 3 Shrinkage estimators James and Stein’s paper aroused excited interest in the statistics community when it arrived in 1961. Most of the excitement focused on the strict inadmissibility of the traditional maximum likelihood estimate demonstrated by the James–Stein rule. Other rules dominating the MLE were discovered, for instance the Bayes estimator of Strawderman ( 1971 ), that was itself admissible while rendering the MLE inadmissible. Big new ideas can take a while to make their true impact felt. The James–Stein rule had an influential side effect on subsequent theory and practice in that it demonstrated, in an inarguable way, the virtues of shrinkage estimation : given an ensemble of problems, individual estimates are shrunk toward a central point; that is, a deliberate bias is introduced, pulling estimates away from their MLEs for the sake of better group performances. Admissibility and inadmissibility aren’t much in the air these days, while shrinkage estimation has gone on to play a major role in modern practice. A spectacular success story is the lasso (Tibshirani, 1996 ). Lasso shrinkage is extreme, pulling some (often most) of the coefficient estimates all the way back to zero. Bayes and empirical Bayes rules tend to be strong shrinkers. Tweedie’s estimate in Fig.Ā  2 ( J = 5 ) shrinks the estimate of E { μ ∣ x = 4 } from its MLE value 4 down to 2.555. For μ between āˆ’ 1 and 1, the shrinkage is almost all the way to zero. The reader may have been surprised to see that neither Tweedie’s formula ( 14 ) for E { μ i ∣ x i } nor its empirical version ( 15 ) require estimation of the prior Ļ€ ( μ ) . This is a special property of the posterior expectation E { μ i ∣ x i } and isn’t available for say Pr { μ i ≄ 2 ∣ x i } , or most other Bayesian targets. ā€œBayesian deconvolutionā€ (Efron, 2016 ) uses low-dimensional parametric modeling of Ļ€ ( μ ) for general empirical Bayes computations. It was applied to finding a prior density Ļ€ ( μ ) that would give the distribution of x seen in Fig.Ā  1 , assuming the normal sampling model ( 1 ). The deconvolution model for Ļ€ ( μ ) used a delta function at μ = 0 (for the ā€œnullā€ genes) and a natural spline function with four degrees of freedom for the non-null cases. Fig. 3 Empirical Bayes conditional density of μ given μ not zero; Pr { μ = 0 } equals 0.825 Full size image The estimated prior Footnote 5 Ļ€ ^ ( μ ) is shown in Fig.Ā  3 ; it put probability 0.825 on μ = 0 , while the conditional distribution given μ ≠ 0 was a moderately heavy-tailed version of N ( 0 , 1.33 2 ) . Based on Ļ€ ^ ( μ ) we can form estimates of any Bayesian target, for instance Pr ^ { μ i ≄ 2 ∣ x i = 4 } = 0.80 . FigureĀ  3 is a direct descendent of the James–Stein rule, now 60-plus years on. A less-direct descendent, but still on the family tree, arrived in 1995. The false discovery rate paper by Benjamini and Hochberg concerned simultaneous hypothesis testing. Looking at Fig.Ā  1 , which of the n = 6033 genes can confidently be labeled as non-null, that is as having μ i ≠ 0 ? Suppose for convenience that the x i s are ordered from smallest to largest. The right-sided significance level for testing μ i = 0 is S 0 ( x i ) = 1 āˆ’ Φ ( x i ) , (22) where Φ is the standard normal cumulative distribution function. Of the 6033 genes, 401 had S i ≤ 0.05 , the usual rejection level for individual testing, but even if actually all of the genes were null we would expect 302 such rejections, so individual testing can’t be right. Benjamini and Hochberg proposed a novel simultaneous testing rule that safely controls the number of ā€œfalse discoveriesā€ — genes falsely labeled ā€non-nullā€ — while not being discouragingly strict. (My summary here won’t give the BH rule its full due; see Chapter 4 of Efron ( 2010 ) for a more complete description.) Let S ^ ( x ) be the observed proportion of x i s exceeding value x , and define F d r ^ ( x ) = Ļ€ 0 S 0 ( x ) / S ^ ( x ) , (23) where Ļ€ 0 is the proportion of null genes among all n . Footnote 6 For a fixed control level α , such as α = 0.1 , the BH rule says to reject the null hypothesis μ i = 0 for those genes having F d r ^ ( x i ) ≤ α . (24) The Benjamini–Hochberg theorem states that under independence assumptions like ( 1 ), the expected proportion of false discoveries by rule ( 24 ) is α . Fig. 4 Prostate data: left Fdr and right Fdr; dashes show 60 genes with F d r < 0.1 Full size image FigureĀ  4 shows F d r ^ for the prostate cancer data and also for the left-sided Fdr estimate, where significance is defined by S 0 ( x i ) = Φ ( x i ) rather than ( 22 ). I applied the BH rule with α = 0.1 which labeled 60 genes as non-null, 32 on the left and 28 on the right. The BH theorem says that we can expect 6 of the 60 to actually be null. The fdr story has evolved very much along the lines of its James–Stein predecessor. Intense initial interest focused on the exact frequentist control of false discovery rates. The Bayes and empirical Bayes implications came later: as at ( 5 ), we assume that each x i is a realization of a random variable x given by μ ∼ Ļ€ ( μ ) and x ∣ μ ∼ p ( x ∣ μ ) , (25) where p ( x ∣ μ ) is a known probability kernel which I’ll take here to be the normal sampling model ( 1 ). Then if S ( x ) is 1 minus the cdf of the marginal density ( 13 ), Bayes rule gives Pr { μ = 0 ∣ x } = Ļ€ 0 S 0 ( x ) / S ( x ) . (26) Comparing ( 26 ) with ( 23 ) says that the BH rule amounts to labeling case i as non-null if its obvious empirical Bayes estimate of nullness is less than α . This is less precise than the frequentist control theorem but, as with the James–Stein estimator, is more robust in not demanding independence among the x i s. The family resemblance between JS and BH is through shrinkage: in the BH case the shrinkage of significance levels. For instance, x i = 3 has individual significance level 0.001 against nullness, whereas F d r ^ = 0.164 for the prostate data, i.e, still with about a 1/6 chance of gene i being null. So what does machine learning have to do with the James–Stein estimator? Nothing to its birth but, as the articles in this volume show, a great deal to its downstream effects on statistical theory and practice. Charles Stein, who was a good applied statistician when he put his mind to it, might have enjoyed these developments, but maybe not; his heart was always with the mathematics. References Benjamini, Y., & Hochberg, Y. (1995). Controlling the false discovery rate: A practical and powerful approach to multiple testing. Journal of the Royal Statistical Society Series B, 57 (1), 289–300. Article Ā  MathSciNet Ā  Google Scholar Ā  Efron, B. (2010). Large-scale inference: Empirical Bayes methods for estimation, testing, and prediction (Vol. 1). Cambridge: Cambridge University Press. Book Ā  Google Scholar Ā  Efron, B. (2011). Tweedie’s formula and selection bias. Journal of the American Statistical Association, 106 (496), 1602–1614. https://doi.org/10.1198/jasa.2011.tm11181 Article Ā  MathSciNet Ā  Google Scholar Ā  Efron, B. (2016). Empirical Bayes deconvolution estimates. Biometrika, 103 (1), 1–20. https://doi.org/10.1093/biomet/asv068 Article Ā  MathSciNet Ā  Google Scholar Ā  Efron, B. (2023). Exponential Families in Theory and Practice . Cambridge: Cambridge University Press. Google Scholar Ā  Efron, B., & Morris, C. (1973). Stein’s estimation rule and its competitors—An empirical Bayes approach. Journal of the American Statistical Association, 68 , 117–130. MathSciNet Ā  Google Scholar Ā  James, W., & Stein, C. (1961). Estimation with quadratic loss. In Proc. 4th Berkeley Sympos. Math. Statist. and Prob. (Vol. I, pp. 361–379). Berkeley: University of California Press. Narasimhan, B., & Efron, B. (2020). deconvolveR: A G-Modeling Program for Deconvolution and Empirical Bayes Estimation. Journal of Statistical Software, 94 (11), 1–20. https://doi.org/10.18637/jss.v094.i11 Article Ā  Google Scholar Ā  Robbins, H. (1956). An empirical Bayes approach to statistics. In Proc. 3rd Berkeley Sympos. Math. Statist. and Prob. (Vol. I, pp. 157–163). Berkeley: University of California Press. Strawderman, W. E. (1971). Proper Bayes minimax estimators of the multivariate normal mean. Annals of Mathematical Statistics, 42 (1), 385–388. https://doi.org/10.1214/aoms/1177693528 Article Ā  MathSciNet Ā  Google Scholar Ā  Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society Series B, 58 (1), 267–288. Article Ā  MathSciNet Ā  Google Scholar Ā  Download references
Markdown
[Skip to main content](https://link.springer.com/article/10.1007/s42081-023-00209-y#main) Advertisement [![Advertisement](https://pubads.g.doubleclick.net/gampad/ad?iu=/270604982/springerlink/42081/article&sz=728x90&pos=top&articleid=s42081-023-00209-y)](https://pubads.g.doubleclick.net/gampad/jump?iu=/270604982/springerlink/42081/article&sz=728x90&pos=top&articleid=s42081-023-00209-y) [![Springer Nature Link](https://link.springer.com/oscar-static/images/darwin/header/img/logo-springer-nature-link-05805fde18.svg)](https://link.springer.com/) [Account](https://link.springer.com/article/10.1007/s42081-023-00209-y) [Menu]() [Find a journal](https://link.springer.com/journals/) [Publish with us](https://www.springernature.com/gp/authors) [Track your research](https://link.springernature.com/home/) [Search]() [Saved research](https://link.springer.com/saved-research) [Cart](https://order.springer.com/public/cart) ## Search ## Navigation - [Find a journal](https://link.springer.com/journals/) - [Publish with us](https://www.springernature.com/gp/authors) - [Track your research](https://link.springernature.com/home/) 1. [Home](https://link.springer.com/) 2. [Japanese Journal of Statistics and Data Science](https://link.springer.com/journal/42081) 3. Article # Machine learning and the James–Stein estimator - Original Paper - Stein Estimation and Statistical Shrinkage Methods - [Open access](https://www.springernature.com/gp/open-science/about/the-fundamentals-of-open-access-and-open-research) - Published: 30 June 2023 - Volume 7, pages 257–266, (2024) - [Cite this article](https://link.springer.com/article/10.1007/s42081-023-00209-y#citeas) You have full access to this [open access](https://www.springernature.com/gp/open-science/about/the-fundamentals-of-open-access-and-open-research) article [Download PDF](https://link.springer.com/content/pdf/10.1007/s42081-023-00209-y.pdf) [Save article](https://link.springer.com/article/10.1007/s42081-023-00209-y/save-research?_csrf=MqX1VwflVBF2wGV6B4ITqKc5OgaOhi_J) [View saved research](https://link.springer.com/saved-research) [![](https://media.springernature.com/w72/springer-static/cover-hires/journal/42081?as=webp) Japanese Journal of Statistics and Data Science](https://link.springer.com/journal/42081) [Aims and scope](https://link.springer.com/journal/42081/aims-and-scope) [Submit manuscript](https://www.editorialmanager.com/jjsd) Machine learning and the James–Stein estimator [Download PDF](https://link.springer.com/content/pdf/10.1007/s42081-023-00209-y.pdf) - [Bradley Efron](https://link.springer.com/article/10.1007/s42081-023-00209-y#auth-Bradley-Efron-Aff1-Aff2) [1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Aff1),[2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Aff2) - 9362 Accesses - 7 Citations - 24 Altmetric - 1 Mention - [Explore all metrics](https://link.springer.com/article/10.1007/s42081-023-00209-y/metrics) ## Abstract It is now 62 years since the publication of James and Stein’s seminal article on the estimation of a multivariate normal mean vector. The paper made a spectacular first impression on the statistical community through its demonstration of inadmissability of the maximum likelihood estimator. It continues to be influential, but not for the initial reasons. Empirical Bayes shrinkage estimation, now a major topic, found its early justification in the James–Stein formula. Less obvious downstream topics include Tweedie’s formula and Benjamini and Hochberg’s false discovery rate algorithm. This is a short and mainly non-technical review of the James–Stein rule and its effects on the machine learning era of statistical innovation. ### Similar content being viewed by others ![](https://media.springernature.com/w215h120/springer-static/image/art%3A10.1007%2Fs42081-023-00227-w/MediaObjects/42081_2023_227_Fig1_HTML.png) ### [Expansion estimators improving the bias and risk of James–Stein’s shrinkage estimator](https://link.springer.com/10.1007/s42081-023-00227-w) Article 16 December 2023 ![](https://media.springernature.com/w215h120/springer-static/image/art%3A10.1007%2Fs10618-019-00622-6/MediaObjects/10618_2019_622_Fig1_HTML.png) ### [A new class of metrics for learning on real-valued and structured data](https://link.springer.com/10.1007/s10618-019-00622-6) Article 27 March 2019 ![](https://media.springernature.com/w215h120/springer-static/image/art%3A10.1007%2Fs40953-021-00270-y/MediaObjects/40953_2021_270_Fig1_HTML.png) ### [Weak Versus Strong Dominance of Shrinkage Estimators](https://link.springer.com/10.1007/s40953-021-00270-y) Article 18 November 2021 ### Explore related subjects Discover the latest articles, books and news in related subjects, suggested using machine learning. - [Bayesian Inference](https://link.springer.com/subjects/bayesian-inference) - [Empiricism](https://link.springer.com/subjects/empiricism) - [Machine Learning](https://link.springer.com/subjects/machine-learning) - [Non-parametric Inference](https://link.springer.com/subjects/non-parametric-inference) - [Statistical Learning](https://link.springer.com/subjects/statistical-learning) - [Statistical Theory and Methods](https://link.springer.com/subjects/statistical-theory-and-methods) - [Empirical Bayesian Methods in Statistical Inference](https://link.springer.com/subjects/empirical-bayesian-methods-in-statistical-inference) ## 1 Introduction By and large, the statistics world is one of heuristics, approximations, and asymptotics. The James–Stein estimator arrived in that world in 1961 on a note of startling specificity: unseen parameters μ 1 , μ 2 , … , μ n produce independent observations x i ∼ ind N ( μ i , 1 ) , i \= 1 , … , n , (1) n ≄ 3. The James–Stein rule in its simplest form proposed estimating the μ i by μ ^ i J S \= ( 1 āˆ’ n āˆ’ 2 S ) x i ( S \= āˆ‘ i \= 1 n x i 2 ) . (2) Formula ([2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ2)) looked implausible: the estimate of μ i depended on the *other* observations x j, j ≠ i (through *S*), as well as x i, despite the independence assumption. Nevertheless, James and Stein showed that Rule ([2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ2)) *always* beat the obvious maximum likelihood estimates μ ^ i M L \= x ĀÆ i ( i \= 1 , … , n ) (3) in terms of total expected squared error E { āˆ‘ i \= 1 n ( μ ^ i āˆ’ μ i ) 2 } . (4) That ā€œalwaysā€ was the shocking part: two centuries of statistical theory, ANOVA, regression, multivariate analysis, etc., depended on maximum likelihood estimation. Did everything have to be rethought? One path forward involved Bayesian thinking. If we assumed that the μ i themselves came from a normal distribution, μ i ∼ ind N ( 0 , A ) for i \= 1 , … , n , (5) with variance A ≄ 0, the Bayes estimates would be μ ^ i B a y e s \= B x i ( B \= A / ( A \+ 1 ) ) . (6) We don’t know *A* or *B* but B ^ \= 1 āˆ’ ( n āˆ’ 2 ) / S (7) is *B*’s unbiased estimate: we can rewrite ([2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ2)) as μ ^ i J S \= B ^ x i , (8) which at least looks more plausible. In the language introduced by Robbins ([1956](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR9 "Robbins, H. (1956). An empirical Bayes approach to statistics. In Proc. 3rd Berkeley Sympos. Math. Statist. and Prob. (Vol. I, pp. 157–163). Berkeley: University of California Press.")), formula ([8](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ8)) is an *empirical Bayes* estimator, another shocking post-war statistical innovation. Carl Morris and I wrote a series of papers in the 1970 s exploring Bayesian roots of the James–Stein estimator (Efron and Morris, [1973](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR6 "Efron, B., & Morris, C. (1973). Stein’s estimation rule and its competitors—An empirical Bayes approach. Journal of the American Statistical Association, 68, 117–130.")). Something is lost in the empirical Bayes formulation, namely the frequentist ā€œalwaysā€ of expected square error minimization, but a lot is gained in flexibility and scope, as discussed in Sect. [2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Sec2). **Fig. 1** [![Fig. 1](https://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs42081-023-00209-y/MediaObjects/42081_2023_209_Fig1_HTML.png)](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/1) Prostate data: 6033 *x* values; mean 0.003, sd \= 1\.135; curve is proportional to a N ( 0 , 1 ) density [Full size image](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/1) Figure [1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig1) illustrates an example of simultaneous estimation pursued in Sect. 2.1 of Efron ([2010](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR2 "Efron, B. (2010). Large-scale inference: Empirical Bayes methods for estimation, testing, and prediction (Vol. 1). Cambridge: Cambridge University Press.")). A microarray study has compared expression levels between prostate cancer patients and control subjects for n \= 6033 genes. For each gene, a statistic x i has been calculated (essentially a ā€œ*z*\-valueā€), x i ∼ N ( μ i , 1 ) , i \= 1 , … , n , (9) where μ i measures the difference between cancer and control group levels. The solid curve in Fig. [1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig1) is a N ( 0 , 1 ) density scaled to have the same area as the histogram of the 6033 *x* values. A bad result from the researchers’ point of view would be a perfect fit of curve to histogram, which would imply all the genes have μ i \= 0, the ā€œnullā€ value of no difference between cancer patients and controls. That’s not what happened: the histogram has mildly heavy tails in both directions. The researchers were hoping to find genes with large values of ‖ μ i ‖—ones that might be a clue to prostate cancer etiology—as suggested by the heavy tails. How encouraged should they be? Not very, according to the James–Stein rule. The 6033 x i values have mean 0.003, which I’ll take to be zero, and empirical variance σ ^ 2 \= 1\.289. (10) The James–Stein estimate ([2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ2)) is μ ^ i J S \= ( 1 āˆ’ n āˆ’ 2 n āˆ’ 1 1 σ ^ 2 ) x i \= 0\.224 ā‹… x i , (11) so even x i \= 5 yields an estimate barely exceeding 1. Section [2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Sec2) suggests a more optimistic analysis. ## 2 Tweedie’s formula The impressive precision of the James–Stein theorem came at a cost in generality. Efforts to extend the theorem, say to Poisson rather than normal observations, or to measures of loss other than total squared error, gave encouraging asymptotic results but not the James–Stein kind of finite sample frequentist dominance. Better progress was possible on the empirical Bayes side of the street. *Tweedie’s formula* (Efron, [2011](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR3 "Efron, B. (2011). Tweedie’s formula and selection bias. Journal of the American Statistical Association, 106(496), 1602–1614. https://doi.org/10.1198/jasa.2011.tm11181 ")) has been particularly useful. We wish to calculate Bayesian estimates μ i B a y e s \= E { μ i ∣ x i } , i \= 1 , … , n , (12) in the normal sampling model ([1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ1)), starting from a given (possibly non-normal) prior Ļ€ ( μ ), applying to all *n* cases. Let *f*(*x*) be the marginal density f ( x ) \= ∫ R Ļ€ ( μ ) Ļ• ( x āˆ’ μ ) d μ , (13) with Ļ• the standard N ( 0 , 1 ) density and R the range of μ. (It isn’t necessary for Ļ€ ( ā‹… ) to be a continuous distribution but it simplifies notation.) Tweedie’s formula provides an elegant statement for μ i B a y e s, the posterior expectation of μ i given x i, μ i B a y e s \= E { μ i ∣ x i } \= x i \+ l ′ ( x i ) with l ′ ( x i ) \= d d x log ⁔ ( f ( x i ) ) . (14) In the empirical Bayes situation ([1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ1)), where the prior Ļ€ ( ā‹… ) is unknown, we can use the observed data x 1 , … , x n to estimate the marginal density *f*(*x*), say by f ^ ( x ), giving empirical Bayes estimates μ ^ i \= x i \+ l ^ ′ ( x i ) . (15) The Bayes estimate ([14](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ14)) can be thought of as the MLE x i plus a Bayesian correction term l ′ ( x i ). When the prior Ļ€ ( μ ) is the N ( 0 , A ) distribution ([5](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ5)), μ i B a y e s equals B x i ([6](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ6)). Simple formulas for μ i B a y e s give out for most other choices of Ļ€ ( μ ) but now, in the machine learning era[Footnote 1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fn1) of statistical research, numerical methods provide useful ways forward, as discussed next. The *log polynomial class*[Footnote 2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fn2) of marginal densities defines *f*(*x*) by log ⁔ ( f β ( x ) ) \= β 0 \+ β ⊤ c ( x ) . (16) Here c ( x ) \= ( x , x 2 , … , x J ) ⊤ and β \= ( β 1 , … , β J ) ⊤ , (17) with β 0 chosen to make f β ( x ) integrate to 1. The choice J \= 2 gives normal marginals; larger values of *J* allow for marginal non-normality. **Fig. 2** [![Fig. 2](https://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs42081-023-00209-y/MediaObjects/42081_2023_209_Fig2_HTML.png)](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/2) Prostate data: Tweedie’s estimate of E { μ ∣ x }, 5 degrees of freedom; dashed curve is James–Stein estimate [Full size image](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/2) The choice J \= 5 was applied to the prostate cancer data of Fig. [1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig1): Tweedie’s formula ([14](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ14)) gave μ ^ ( x ) \= E { μ ∣ x }, graphed as the solid curve in Fig. [2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig2). It differs markedly from the James–Stein estimate J \= 2, the dashed line. At x \= 4 for example, the J \= 5 estimate is[Footnote 3](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fn3) E { μ ∣ x \= 4 } \= 2\.555 (18) compared to 0.901 for the James–Stein estimate. The estimated curve E { μ ∣ x } is *empirical Bayes* in the same sense as ([8](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ8)): the parameter vector β was selected by maximum likelihood, as discussed next. With J \= 5, the prior was able to adapt to the ā€œfishing expeditionā€ nature of such microarray studies, where we expect most of the genes to be null or close to null, with μ i nearly zero (corresponding here to the flat part of the curve for *x* between āˆ’ 2 and 2) and, hopefully, a small proportion of interestingly large μ is. The sample size n \= 6033 has much to do with Fig. [2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig2). James and Stein ([1961](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR7 "James, W., & Stein, C. (1961). Estimation with quadratic loss. In Proc. 4th Berkeley Sympos. Math. Statist. and Prob. (Vol. I, pp. 361–379). Berkeley: University of California Press.")) was usually considered in terms of small samples, perhaps n ≤ 20, for which there would be little hope of seeing the detail in Fig. [2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig2). The term ā€œmachine learning eraā€ seems less fanciful when considering the scale of problems statisticians are now asked to deal with, as well as the tools they use to solve them. It looks like it might be hard work computing Fig. [2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig2) but it’s not. The histogram in Fig. [1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig1) has 97 bins, with centerpoints v v \= ( āˆ’ 4\.4 , āˆ’ 4\.3 , … , 5\.1 , 5\.2 ) . (19) Let y j be the count in bin *j*, that is, the number of the 6033 x i values falling into it, with the vector of counts being y y \= ( y 1 , … , y 97 ) . (20) Then the single R command l l ^ \= log ⁔ ( glm ( y y ∼ poly ( v v , 5 ) , poisson ) \$ fit ) (21) provides a close approximation to the MLE of log ⁔ f ( x ) in ([14](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ14)); numerical differentiation of l l ^ gives Tweedie’s estimate. Section 3.4 of Efron ([2023](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR5 "Efron, B. (2023). Exponential Families in Theory and Practice. Cambridge: Cambridge University Press.")) shows why Poisson regression ([21](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ21)) is appropriate here. The James–Stein theorem depends on the independence assumption in ([1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ1)), unlikely to be true in the microarray study, but the estimates ([2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ2)) have a certain marginal validity even under dependence. This is clearer from the empirical Bayes point of view. The Tweedie estimate x i \+ l ^ ′ ( x i ) requires only that l ^ ′ ( x ) be close to l ′ ( x ), not that it be estimated from independent x is.[Footnote 4](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fn4) ## 3 Shrinkage estimators James and Stein’s paper aroused excited interest in the statistics community when it arrived in 1961. Most of the excitement focused on the strict inadmissibility of the traditional maximum likelihood estimate demonstrated by the James–Stein rule. Other rules dominating the MLE were discovered, for instance the Bayes estimator of Strawderman ([1971](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR10 "Strawderman, W. E. (1971). Proper Bayes minimax estimators of the multivariate normal mean. Annals of Mathematical Statistics, 42(1), 385–388. https://doi.org/10.1214/aoms/1177693528 ")), that was itself admissible while rendering the MLE inadmissible. Big new ideas can take a while to make their true impact felt. The James–Stein rule had an influential side effect on subsequent theory and practice in that it demonstrated, in an inarguable way, the virtues of *shrinkage estimation*: given an ensemble of problems, individual estimates are shrunk toward a central point; that is, a deliberate bias is introduced, pulling estimates away from their MLEs for the sake of better group performances. Admissibility and inadmissibility aren’t much in the air these days, while shrinkage estimation has gone on to play a major role in modern practice. A spectacular success story is the lasso (Tibshirani, [1996](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR11 "Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society Series B, 58(1), 267–288.")). Lasso shrinkage is extreme, pulling some (often most) of the coefficient estimates all the way back to zero. Bayes and empirical Bayes rules tend to be strong shrinkers. Tweedie’s estimate in Fig. [2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig2) (J \= 5) shrinks the estimate of E { μ ∣ x \= 4 } from its MLE value 4 down to 2.555. For μ between āˆ’ 1 and 1, the shrinkage is almost all the way to zero. The reader may have been surprised to see that neither Tweedie’s formula ([14](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ14)) for E { μ i ∣ x i } nor its empirical version ([15](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ15)) require estimation of the prior Ļ€ ( μ ). This is a special property of the posterior expectation E { μ i ∣ x i } and isn’t available for say Pr { μ i ≄ 2 ∣ x i }, or most other Bayesian targets. ā€œBayesian deconvolutionā€ (Efron, [2016](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR4 "Efron, B. (2016). Empirical Bayes deconvolution estimates. Biometrika, 103(1), 1–20. https://doi.org/10.1093/biomet/asv068 ")) uses low-dimensional parametric modeling of Ļ€ ( μ ) for general empirical Bayes computations. It was applied to finding a prior density Ļ€ ( μ ) that would give the distribution of *x* seen in Fig. [1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig1), assuming the normal sampling model ([1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ1)). The deconvolution model for Ļ€ ( μ ) used a delta function at μ \= 0 (for the ā€œnullā€ genes) and a natural spline function with four degrees of freedom for the non-null cases. **Fig. 3** [![Fig. 3](https://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs42081-023-00209-y/MediaObjects/42081_2023_209_Fig3_HTML.png)](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/3) Empirical Bayes conditional density of μ given μ not zero; Pr { μ \= 0 } equals 0.825 [Full size image](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/3) The estimated prior[Footnote 5](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fn5)Ļ€ ^ ( μ ) is shown in Fig. [3](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig3); it put probability 0.825 on μ \= 0, while the conditional distribution given μ ≠ 0 was a moderately heavy-tailed version of N ( 0 , 1\.33 2 ). Based on Ļ€ ^ ( μ ) we can form estimates of *any* Bayesian target, for instance Pr ^ { μ i ≄ 2 ∣ x i \= 4 } \= 0\.80. Figure [3](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig3) is a direct descendent of the James–Stein rule, now 60-plus years on. A less-direct descendent, but still on the family tree, arrived in 1995. The *false discovery rate* paper by Benjamini and Hochberg concerned simultaneous hypothesis testing. Looking at Fig. [1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig1), which of the n \= 6033 genes can confidently be labeled as non-null, that is as having μ i ≠ 0? Suppose for convenience that the x is are ordered from smallest to largest. The right-sided significance level for testing μ i \= 0 is S 0 ( x i ) \= 1 āˆ’ Φ ( x i ) , (22) where Φ is the standard normal cumulative distribution function. Of the 6033 genes, 401 had S i ≤ 0\.05, the usual rejection level for individual testing, but even if actually *all* of the genes were null we would expect 302 such rejections, so individual testing can’t be right. Benjamini and Hochberg proposed a novel simultaneous testing rule that safely controls the number of ā€œfalse discoveriesā€ — genes falsely labeled ā€non-nullā€ — while not being discouragingly strict. (My summary here won’t give the BH rule its full due; see Chapter 4 of Efron ([2010](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR2 "Efron, B. (2010). Large-scale inference: Empirical Bayes methods for estimation, testing, and prediction (Vol. 1). Cambridge: Cambridge University Press.")) for a more complete description.) Let S ^ ( x ) be the observed proportion of x is exceeding value *x*, and define F d r ^ ( x ) \= Ļ€ 0 S 0 ( x ) / S ^ ( x ) , (23) where Ļ€ 0 is the proportion of null genes among all *n*.[Footnote 6](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fn6) For a fixed control level α, such as α \= 0\.1, the BH rule says to reject the null hypothesis μ i \= 0 for those genes having F d r ^ ( x i ) ≤ α . (24) The Benjamini–Hochberg theorem states that under independence assumptions like ([1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ1)), the expected proportion of false discoveries by rule ([24](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ24)) is α. **Fig. 4** [![Fig. 4](https://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs42081-023-00209-y/MediaObjects/42081_2023_209_Fig4_HTML.png)](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/4) Prostate data: left Fdr and right Fdr; dashes show 60 genes with F d r \< 0\.1 [Full size image](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/4) Figure [4](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig4) shows F d r ^ for the prostate cancer data and also for the left-sided Fdr estimate, where significance is defined by S 0 ( x i ) \= Φ ( x i ) rather than ([22](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ22)). I applied the BH rule with α \= 0\.1 which labeled 60 genes as non-null, 32 on the left and 28 on the right. The BH theorem says that we can expect 6 of the 60 to actually be null. The fdr story has evolved very much along the lines of its James–Stein predecessor. Intense initial interest focused on the exact frequentist control of false discovery rates. The Bayes and empirical Bayes implications came later: as at ([5](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ5)), we assume that each x i is a realization of a random variable *x* given by μ ∼ Ļ€ ( μ ) and x ∣ μ ∼ p ( x ∣ μ ) , (25) where p ( x ∣ μ ) is a known probability kernel which I’ll take here to be the normal sampling model ([1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ1)). Then if *S*(*x*) is 1 minus the cdf of the marginal density ([13](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ13)), Bayes rule gives Pr { μ \= 0 ∣ x } \= Ļ€ 0 S 0 ( x ) / S ( x ) . (26) Comparing ([26](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ26)) with ([23](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ23)) says that the BH rule amounts to labeling case *i* as non-null if its obvious empirical Bayes estimate of nullness is less than α. This is less precise than the frequentist control theorem but, as with the James–Stein estimator, is more robust in not demanding independence among the x is. The family resemblance between JS and BH is through shrinkage: in the BH case the shrinkage of significance levels. For instance, x i \= 3 has individual significance level 0.001 against nullness, whereas F d r ^ \= 0\.164 for the prostate data, i.e, still with about a 1/6 chance of gene *i* being null. So what does machine learning have to do with the James–Stein estimator? Nothing to its birth but, as the articles in this volume show, a great deal to its downstream effects on statistical theory and practice. Charles Stein, who was a good applied statistician when he put his mind to it, might have enjoyed these developments, but maybe not; his heart was always with the mathematics. ## Notes 1. Where algorithms can substitute for theorems. 2. For general use, a natural spline basis is preferable to polynomials, to control the behavior of log ⁔ Ļ€ ( μ ) at the extremes. 3. With an estimated bootstrap standard error of 0.192. 4. The accuracy of the Tweedie estimate *does* suffer under dependence, so the previously quoted bootstrap standard error is likely to be optimistic. 5. Estimated using the CRAN package deconvolveR (Narasimhan and Efron, [2020](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR8 "Narasimhan, B., & Efron, B. (2020). deconvolveR: A G-Modeling Program for Deconvolution and Empirical Bayes Estimation. Journal of Statistical Software, 94(11), 1–20. https://doi.org/10.18637/jss.v094.i11 ")). 6. Ļ€ 0 can be estimated but in practice it is usually replaced by its upper bound 1 in applying rule ([24](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ24)). For cases like the prostate data where most of the genes are null, this doesn’t much affect the outcome. ## References - Benjamini, Y., & Hochberg, Y. (1995). Controlling the false discovery rate: A practical and powerful approach to multiple testing. *Journal of the Royal Statistical Society Series B,* *57*(1), 289–300. [Article](https://doi.org/10.1111%2Fj.2517-6161.1995.tb02031.x) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=1325392) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Controlling%20the%20false%20discovery%20rate%3A%20A%20practical%20and%20powerful%20approach%20to%20multiple%20testing&journal=Journal%20of%20the%20Royal%20Statistical%20Society%20Series%20B&doi=10.1111%2Fj.2517-6161.1995.tb02031.x&volume=57&issue=1&pages=289-300&publication_year=1995&author=Benjamini%2CY&author=Hochberg%2CY) - Efron, B. (2010). *Large-scale inference: Empirical Bayes methods for estimation, testing, and prediction* (Vol. 1). Cambridge: Cambridge University Press. [Book](https://doi.org/10.1017%2FCBO9780511761362) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Large-scale%20inference%3A%20Empirical%20Bayes%20methods%20for%20estimation%2C%20testing%2C%20and%20prediction&doi=10.1017%2FCBO9780511761362&publication_year=2010&author=Efron%2CB) - Efron, B. (2011). Tweedie’s formula and selection bias. *Journal of the American Statistical Association,* *106*(496), 1602–1614. <https://doi.org/10.1198/jasa.2011.tm11181> [Article](https://doi.org/10.1198%2Fjasa.2011.tm11181) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=2896860) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Tweedie%E2%80%99s%20formula%20and%20selection%20bias&journal=Journal%20of%20the%20American%20Statistical%20Association&doi=10.1198%2Fjasa.2011.tm11181&volume=106&issue=496&pages=1602-1614&publication_year=2011&author=Efron%2CB) - Efron, B. (2016). Empirical Bayes deconvolution estimates. *Biometrika,* *103*(1), 1–20. <https://doi.org/10.1093/biomet/asv068> [Article](https://doi.org/10.1093%2Fbiomet%2Fasv068) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=3465818) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Empirical%20Bayes%20deconvolution%20estimates&journal=Biometrika&doi=10.1093%2Fbiomet%2Fasv068&volume=103&issue=1&pages=1-20&publication_year=2016&author=Efron%2CB) - Efron, B. (2023). *Exponential Families in Theory and Practice*. Cambridge: Cambridge University Press. [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Exponential%20Families%20in%20Theory%20and%20Practice&publication_year=2023&author=Efron%2CB) - Efron, B., & Morris, C. (1973). Stein’s estimation rule and its competitors—An empirical Bayes approach. *Journal of the American Statistical Association,* *68*, 117–130. [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=388597) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Stein%E2%80%99s%20estimation%20rule%20and%20its%20competitors%E2%80%94An%20empirical%20Bayes%20approach&journal=Journal%20of%20the%20American%20Statistical%20Association&volume=68&pages=117-130&publication_year=1973&author=Efron%2CB&author=Morris%2CC) - James, W., & Stein, C. (1961). Estimation with quadratic loss. In *Proc. 4th Berkeley Sympos. Math. Statist. and Prob.* (Vol. I, pp. 361–379). Berkeley: University of California Press. - Narasimhan, B., & Efron, B. (2020). deconvolveR: A G-Modeling Program for Deconvolution and Empirical Bayes Estimation. *Journal of Statistical Software,* *94*(11), 1–20. <https://doi.org/10.18637/jss.v094.i11> [Article](https://doi.org/10.18637%2Fjss.v094.i11) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=deconvolveR%3A%20A%20G-Modeling%20Program%20for%20Deconvolution%20and%20Empirical%20Bayes%20Estimation&journal=Journal%20of%20Statistical%20Software&doi=10.18637%2Fjss.v094.i11&volume=94&issue=11&pages=1-20&publication_year=2020&author=Narasimhan%2CB&author=Efron%2CB) - Robbins, H. (1956). An empirical Bayes approach to statistics. In *Proc. 3rd Berkeley Sympos. Math. Statist. and Prob.* (Vol. I, pp. 157–163). Berkeley: University of California Press. - Strawderman, W. E. (1971). Proper Bayes minimax estimators of the multivariate normal mean. *Annals of Mathematical Statistics,* *42*(1), 385–388. <https://doi.org/10.1214/aoms/1177693528> [Article](https://doi.org/10.1214%2Faoms%2F1177693528) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=397939) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Proper%20Bayes%20minimax%20estimators%20of%20the%20multivariate%20normal%20mean&journal=Annals%20of%20Mathematical%20Statistics&doi=10.1214%2Faoms%2F1177693528&volume=42&issue=1&pages=385-388&publication_year=1971&author=Strawderman%2CWE) - Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. *Journal of the Royal Statistical Society Series B,* *58*(1), 267–288. [Article](https://doi.org/10.1111%2Fj.2517-6161.1996.tb02080.x) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=1379242) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Regression%20shrinkage%20and%20selection%20via%20the%20lasso&journal=Journal%20of%20the%20Royal%20Statistical%20Society%20Series%20B&doi=10.1111%2Fj.2517-6161.1996.tb02080.x&volume=58&issue=1&pages=267-288&publication_year=1996&author=Tibshirani%2CR) [Download references](https://citation-needed.springer.com/v2/references/10.1007/s42081-023-00209-y?format=refman&flavour=references) ## Funding No funds, grants, or other support was received. ## Author information ### Authors and Affiliations 1. Department of Statistics, Stanford University, 390 Jane Stanford Way, Stanford, CA, 94305, USA Bradley Efron 2. Department of Biomedical Data Science, Stanford School of Medicine, 1265 Welch Road, Stanford, CA, 94305, USA Bradley Efron Authors 1. Bradley Efron [View author publications](https://link.springer.com/search?sortBy=newestFirst&contributor=Bradley%20Efron) Search author on:[PubMed](https://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=search&term=Bradley%20Efron) [Google Scholar](https://scholar.google.co.uk/scholar?as_q=&num=10&btnG=Search+Scholar&as_epq=&as_oq=&as_eq=&as_occt=any&as_sauthors=%22Bradley%20Efron%22&as_publication=&as_ylo=&as_yhi=&as_allsubj=all&hl=en) ### Corresponding author Correspondence to [Bradley Efron](mailto:efron@stanford.edu). ## Ethics declarations ### Conflict of interest The author has no relevant financial or non-financial interests to disclose. ## Additional information ### Publisher's Note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. Dedicated to the memory of Carl Morris. ## Rights and permissions **Open Access** This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit <http://creativecommons.org/licenses/by/4.0/>. [Reprints and permissions](https://s100.copyright.com/AppDispatchServlet?title=Machine%20learning%20and%20the%20James%E2%80%93Stein%20estimator&author=Bradley%20Efron&contentID=10.1007%2Fs42081-023-00209-y&copyright=The%20Author%28s%29&publication=2520-8756&publicationDate=2023-06-30&publisherName=SpringerNature&orderBeanReset=true&oa=CC%20BY) ## About this article [![Check for updates. Verify currency and authenticity via CrossMark](data:image/svg+xml;base64,<svg height="81" width="57" xmlns="http://www.w3.org/2000/svg"><g fill="none" fill-rule="evenodd"><path d="m17.35 35.45 21.3-14.2v-17.03h-21.3" fill="#989898"/><path d="m38.65 35.45-21.3-14.2v-17.03h21.3" fill="#747474"/><path d="m28 .5c-12.98 0-23.5 10.52-23.5 23.5s10.52 23.5 23.5 23.5 23.5-10.52 23.5-23.5c0-6.23-2.48-12.21-6.88-16.62-4.41-4.4-10.39-6.88-16.62-6.88zm0 41.25c-9.8 0-17.75-7.95-17.75-17.75s7.95-17.75 17.75-17.75 17.75 7.95 17.75 17.75c0 4.71-1.87 9.22-5.2 12.55s-7.84 5.2-12.55 5.2z" fill="#535353"/><path d="m41 36c-5.81 6.23-15.23 7.45-22.43 2.9-7.21-4.55-10.16-13.57-7.03-21.5l-4.92-3.11c-4.95 10.7-1.19 23.42 8.78 29.71 9.97 6.3 23.07 4.22 30.6-4.86z" fill="#9c9c9c"/><path d="m.2 58.45c0-.75.11-1.42.33-2.01s.52-1.09.91-1.5c.38-.41.83-.73 1.34-.94.51-.22 1.06-.32 1.65-.32.56 0 1.06.11 1.51.35.44.23.81.5 1.1.81l-.91 1.01c-.24-.24-.49-.42-.75-.56-.27-.13-.58-.2-.93-.2-.39 0-.73.08-1.05.23-.31.16-.58.37-.81.66-.23.28-.41.63-.53 1.04-.13.41-.19.88-.19 1.39 0 1.04.23 1.86.68 2.46.45.59 1.06.88 1.84.88.41 0 .77-.07 1.07-.23s.59-.39.85-.68l.91 1c-.38.43-.8.76-1.28.99-.47.22-1 .34-1.58.34-.59 0-1.13-.1-1.64-.31-.5-.2-.94-.51-1.31-.91-.38-.4-.67-.9-.88-1.48-.22-.59-.33-1.26-.33-2.02zm8.4-5.33h1.61v2.54l-.05 1.33c.29-.27.61-.51.96-.72s.76-.31 1.24-.31c.73 0 1.27.23 1.61.71.33.47.5 1.14.5 2.02v4.31h-1.61v-4.1c0-.57-.08-.97-.25-1.21-.17-.23-.45-.35-.83-.35-.3 0-.56.08-.79.22-.23.15-.49.36-.78.64v4.8h-1.61zm7.37 6.45c0-.56.09-1.06.26-1.51.18-.45.42-.83.71-1.14.29-.3.63-.54 1.01-.71.39-.17.78-.25 1.18-.25.47 0 .88.08 1.23.24.36.16.65.38.89.67s.42.63.54 1.03c.12.41.18.84.18 1.32 0 .32-.02.57-.07.76h-4.36c.07.62.29 1.1.65 1.44.36.33.82.5 1.38.5.29 0 .57-.04.83-.13s.51-.21.76-.37l.55 1.01c-.33.21-.69.39-1.09.53-.41.14-.83.21-1.26.21-.48 0-.92-.08-1.34-.25-.41-.16-.76-.4-1.07-.7-.31-.31-.55-.69-.72-1.13-.18-.44-.26-.95-.26-1.52zm4.6-.62c0-.55-.11-.98-.34-1.28-.23-.31-.58-.47-1.06-.47-.41 0-.77.15-1.07.45-.31.29-.5.73-.58 1.3zm2.5.62c0-.57.09-1.08.28-1.53.18-.44.43-.82.75-1.13s.69-.54 1.1-.71c.42-.16.85-.24 1.31-.24.45 0 .84.08 1.17.23s.61.34.85.57l-.77 1.02c-.19-.16-.38-.28-.56-.37-.19-.09-.39-.14-.61-.14-.56 0-1.01.21-1.35.63-.35.41-.52.97-.52 1.67 0 .69.17 1.24.51 1.66.34.41.78.62 1.32.62.28 0 .54-.06.78-.17.24-.12.45-.26.64-.42l.67 1.03c-.33.29-.69.51-1.08.65-.39.15-.78.23-1.18.23-.46 0-.9-.08-1.31-.24-.4-.16-.75-.39-1.05-.7s-.53-.69-.7-1.13c-.17-.45-.25-.96-.25-1.53zm6.91-6.45h1.58v6.17h.05l2.54-3.16h1.77l-2.35 2.8 2.59 4.07h-1.75l-1.77-2.98-1.08 1.23v1.75h-1.58zm13.69 1.27c-.25-.11-.5-.17-.75-.17-.58 0-.87.39-.87 1.16v.75h1.34v1.27h-1.34v5.6h-1.61v-5.6h-.92v-1.2l.92-.07v-.72c0-.35.04-.68.13-.98.08-.31.21-.57.4-.79s.42-.39.71-.51c.28-.12.63-.18 1.04-.18.24 0 .48.02.69.07.22.05.41.1.57.17zm.48 5.18c0-.57.09-1.08.27-1.53.17-.44.41-.82.72-1.13.3-.31.65-.54 1.04-.71.39-.16.8-.24 1.23-.24s.84.08 1.24.24c.4.17.74.4 1.04.71s.54.69.72 1.13c.19.45.28.96.28 1.53s-.09 1.08-.28 1.53c-.18.44-.42.82-.72 1.13s-.64.54-1.04.7-.81.24-1.24.24-.84-.08-1.23-.24-.74-.39-1.04-.7c-.31-.31-.55-.69-.72-1.13-.18-.45-.27-.96-.27-1.53zm1.65 0c0 .69.14 1.24.43 1.66.28.41.68.62 1.18.62.51 0 .9-.21 1.19-.62.29-.42.44-.97.44-1.66 0-.7-.15-1.26-.44-1.67-.29-.42-.68-.63-1.19-.63-.5 0-.9.21-1.18.63-.29.41-.43.97-.43 1.67zm6.48-3.44h1.33l.12 1.21h.05c.24-.44.54-.79.88-1.02.35-.24.7-.36 1.07-.36.32 0 .59.05.78.14l-.28 1.4-.33-.09c-.11-.01-.23-.02-.38-.02-.27 0-.56.1-.86.31s-.55.58-.77 1.1v4.2h-1.61zm-47.87 15h1.61v4.1c0 .57.08.97.25 1.2.17.24.44.35.81.35.3 0 .57-.07.8-.22.22-.15.47-.39.73-.73v-4.7h1.61v6.87h-1.32l-.12-1.01h-.04c-.3.36-.63.64-.98.86-.35.21-.76.32-1.24.32-.73 0-1.27-.24-1.61-.71-.33-.47-.5-1.14-.5-2.02zm9.46 7.43v2.16h-1.61v-9.59h1.33l.12.72h.05c.29-.24.61-.45.97-.63.35-.17.72-.26 1.1-.26.43 0 .81.08 1.15.24.33.17.61.4.84.71.24.31.41.68.53 1.11.13.42.19.91.19 1.44 0 .59-.09 1.11-.25 1.57-.16.47-.38.85-.65 1.16-.27.32-.58.56-.94.73-.35.16-.72.25-1.1.25-.3 0-.6-.07-.9-.2s-.59-.31-.87-.56zm0-2.3c.26.22.5.37.73.45.24.09.46.13.66.13.46 0 .84-.2 1.15-.6.31-.39.46-.98.46-1.77 0-.69-.12-1.22-.35-1.61-.23-.38-.61-.57-1.13-.57-.49 0-.99.26-1.52.77zm5.87-1.69c0-.56.08-1.06.25-1.51.16-.45.37-.83.65-1.14.27-.3.58-.54.93-.71s.71-.25 1.08-.25c.39 0 .73.07 1 .2.27.14.54.32.81.55l-.06-1.1v-2.49h1.61v9.88h-1.33l-.11-.74h-.06c-.25.25-.54.46-.88.64-.33.18-.69.27-1.06.27-.87 0-1.56-.32-2.07-.95s-.76-1.51-.76-2.65zm1.67-.01c0 .74.13 1.31.4 1.7.26.38.65.58 1.15.58.51 0 .99-.26 1.44-.77v-3.21c-.24-.21-.48-.36-.7-.45-.23-.08-.46-.12-.7-.12-.45 0-.82.19-1.13.59-.31.39-.46.95-.46 1.68zm6.35 1.59c0-.73.32-1.3.97-1.71.64-.4 1.67-.68 3.08-.84 0-.17-.02-.34-.07-.51-.05-.16-.12-.3-.22-.43s-.22-.22-.38-.3c-.15-.06-.34-.1-.58-.1-.34 0-.68.07-1 .2s-.63.29-.93.47l-.59-1.08c.39-.24.81-.45 1.28-.63.47-.17.99-.26 1.54-.26.86 0 1.51.25 1.93.76s.63 1.25.63 2.21v4.07h-1.32l-.12-.76h-.05c-.3.27-.63.48-.98.66s-.73.27-1.14.27c-.61 0-1.1-.19-1.48-.56-.38-.36-.57-.85-.57-1.46zm1.57-.12c0 .3.09.53.27.67.19.14.42.21.71.21.28 0 .54-.07.77-.2s.48-.31.73-.56v-1.54c-.47.06-.86.13-1.18.23-.31.09-.57.19-.76.31s-.33.25-.41.4c-.09.15-.13.31-.13.48zm6.29-3.63h-.98v-1.2l1.06-.07.2-1.88h1.34v1.88h1.75v1.27h-1.75v3.28c0 .8.32 1.2.97 1.2.12 0 .24-.01.37-.04.12-.03.24-.07.34-.11l.28 1.19c-.19.06-.4.12-.64.17-.23.05-.49.08-.76.08-.4 0-.74-.06-1.02-.18-.27-.13-.49-.3-.67-.52-.17-.21-.3-.48-.37-.78-.08-.3-.12-.64-.12-1.01zm4.36 2.17c0-.56.09-1.06.27-1.51s.41-.83.71-1.14c.29-.3.63-.54 1.01-.71.39-.17.78-.25 1.18-.25.47 0 .88.08 1.23.24.36.16.65.38.89.67s.42.63.54 1.03c.12.41.18.84.18 1.32 0 .32-.02.57-.07.76h-4.37c.08.62.29 1.1.65 1.44.36.33.82.5 1.38.5.3 0 .58-.04.84-.13.25-.09.51-.21.76-.37l.54 1.01c-.32.21-.69.39-1.09.53s-.82.21-1.26.21c-.47 0-.92-.08-1.33-.25-.41-.16-.77-.4-1.08-.7-.3-.31-.54-.69-.72-1.13-.17-.44-.26-.95-.26-1.52zm4.61-.62c0-.55-.11-.98-.34-1.28-.23-.31-.58-.47-1.06-.47-.41 0-.77.15-1.08.45-.31.29-.5.73-.57 1.3zm3.01 2.23c.31.24.61.43.92.57.3.13.63.2.98.2.38 0 .65-.08.83-.23s.27-.35.27-.6c0-.14-.05-.26-.13-.37-.08-.1-.2-.2-.34-.28-.14-.09-.29-.16-.47-.23l-.53-.22c-.23-.09-.46-.18-.69-.3-.23-.11-.44-.24-.62-.4s-.33-.35-.45-.55c-.12-.21-.18-.46-.18-.75 0-.61.23-1.1.68-1.49.44-.38 1.06-.57 1.83-.57.48 0 .91.08 1.29.25s.71.36.99.57l-.74.98c-.24-.17-.49-.32-.73-.42-.25-.11-.51-.16-.78-.16-.35 0-.6.07-.76.21-.17.15-.25.33-.25.54 0 .14.04.26.12.36s.18.18.31.26c.14.07.29.14.46.21l.54.19c.23.09.47.18.7.29s.44.24.64.4c.19.16.34.35.46.58.11.23.17.5.17.82 0 .3-.06.58-.17.83-.12.26-.29.48-.51.68-.23.19-.51.34-.84.45-.34.11-.72.17-1.15.17-.48 0-.95-.09-1.41-.27-.46-.19-.86-.41-1.2-.68z" fill="#535353"/></g></svg>)](https://crossmark.crossref.org/dialog/?doi=10.1007/s42081-023-00209-y) ### Cite this article Efron, B. Machine learning and the James–Stein estimator. *Jpn J Stat Data Sci* **7**, 257–266 (2024). https://doi.org/10.1007/s42081-023-00209-y [Download citation](https://citation-needed.springer.com/v2/references/10.1007/s42081-023-00209-y?format=refman&flavour=citation) - Received: 24 March 2023 - Accepted: 13 May 2023 - Published: 30 June 2023 - Version of record: 30 June 2023 - Issue date: June 2024 - DOI: https://doi.org/10.1007/s42081-023-00209-y ### Share this article Anyone you share the following link with will be able to read this content: Get shareable link Sorry, a shareable link is not currently available for this article. Copy shareable link to clipboard Provided by the Springer Nature SharedIt content-sharing initiative ### Keywords - [Empirical bayes](https://link.springer.com/search?query=Empirical%20bayes&facet-discipline="Statistics") - [Shrinkage](https://link.springer.com/search?query=Shrinkage&facet-discipline="Statistics") - [Tweedie’s formula](https://link.springer.com/search?query=Tweedie%E2%80%99s%20formula&facet-discipline="Statistics") - [Benjamini–Hochberg algorithm](https://link.springer.com/search?query=Benjamini%E2%80%93Hochberg%20algorithm&facet-discipline="Statistics") - Sections - Figures - References - [Abstract](https://link.springer.com/article/10.1007/s42081-023-00209-y#Abs1) - [1 Introduction](https://link.springer.com/article/10.1007/s42081-023-00209-y#Sec1) - [2 Tweedie’s formula](https://link.springer.com/article/10.1007/s42081-023-00209-y#Sec2) - [3 Shrinkage estimators](https://link.springer.com/article/10.1007/s42081-023-00209-y#Sec3) - [Notes](https://link.springer.com/article/10.1007/s42081-023-00209-y#notes) - [References](https://link.springer.com/article/10.1007/s42081-023-00209-y#Bib1) - [Funding](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fun) - [Author information](https://link.springer.com/article/10.1007/s42081-023-00209-y#author-information) - [Ethics declarations](https://link.springer.com/article/10.1007/s42081-023-00209-y#ethics) - [Additional information](https://link.springer.com/article/10.1007/s42081-023-00209-y#additional-information) - [Rights and permissions](https://link.springer.com/article/10.1007/s42081-023-00209-y#rightslink) - [About this article](https://link.springer.com/article/10.1007/s42081-023-00209-y#article-info) Advertisement - **Fig. 1** ![Fig. 1]() [View in article](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig1)[Full size image](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/1) - **Fig. 2** ![Fig. 2]() [View in article](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig2)[Full size image](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/2) - **Fig. 3** ![Fig. 3]() [View in article](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig3)[Full size image](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/3) - **Fig. 4** ![Fig. 4]() [View in article](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig4)[Full size image](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/4) 1. Benjamini, Y., & Hochberg, Y. (1995). Controlling the false discovery rate: A practical and powerful approach to multiple testing. *Journal of the Royal Statistical Society Series B,* *57*(1), 289–300. [Article](https://doi.org/10.1111%2Fj.2517-6161.1995.tb02031.x) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=1325392) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Controlling%20the%20false%20discovery%20rate%3A%20A%20practical%20and%20powerful%20approach%20to%20multiple%20testing&journal=Journal%20of%20the%20Royal%20Statistical%20Society%20Series%20B&doi=10.1111%2Fj.2517-6161.1995.tb02031.x&volume=57&issue=1&pages=289-300&publication_year=1995&author=Benjamini%2CY&author=Hochberg%2CY) 2. Efron, B. (2010). *Large-scale inference: Empirical Bayes methods for estimation, testing, and prediction* (Vol. 1). Cambridge: Cambridge University Press. [Book](https://doi.org/10.1017%2FCBO9780511761362) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Large-scale%20inference%3A%20Empirical%20Bayes%20methods%20for%20estimation%2C%20testing%2C%20and%20prediction&doi=10.1017%2FCBO9780511761362&publication_year=2010&author=Efron%2CB) 3. Efron, B. (2011). Tweedie’s formula and selection bias. *Journal of the American Statistical Association,* *106*(496), 1602–1614. <https://doi.org/10.1198/jasa.2011.tm11181> [Article](https://doi.org/10.1198%2Fjasa.2011.tm11181) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=2896860) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Tweedie%E2%80%99s%20formula%20and%20selection%20bias&journal=Journal%20of%20the%20American%20Statistical%20Association&doi=10.1198%2Fjasa.2011.tm11181&volume=106&issue=496&pages=1602-1614&publication_year=2011&author=Efron%2CB) 4. Efron, B. (2016). Empirical Bayes deconvolution estimates. *Biometrika,* *103*(1), 1–20. <https://doi.org/10.1093/biomet/asv068> [Article](https://doi.org/10.1093%2Fbiomet%2Fasv068) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=3465818) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Empirical%20Bayes%20deconvolution%20estimates&journal=Biometrika&doi=10.1093%2Fbiomet%2Fasv068&volume=103&issue=1&pages=1-20&publication_year=2016&author=Efron%2CB) 5. Efron, B. (2023). *Exponential Families in Theory and Practice*. Cambridge: Cambridge University Press. [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Exponential%20Families%20in%20Theory%20and%20Practice&publication_year=2023&author=Efron%2CB) 6. Efron, B., & Morris, C. (1973). Stein’s estimation rule and its competitors—An empirical Bayes approach. *Journal of the American Statistical Association,* *68*, 117–130. [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=388597) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Stein%E2%80%99s%20estimation%20rule%20and%20its%20competitors%E2%80%94An%20empirical%20Bayes%20approach&journal=Journal%20of%20the%20American%20Statistical%20Association&volume=68&pages=117-130&publication_year=1973&author=Efron%2CB&author=Morris%2CC) 7. James, W., & Stein, C. (1961). Estimation with quadratic loss. In *Proc. 4th Berkeley Sympos. Math. Statist. and Prob.* (Vol. I, pp. 361–379). Berkeley: University of California Press. 8. Narasimhan, B., & Efron, B. (2020). deconvolveR: A G-Modeling Program for Deconvolution and Empirical Bayes Estimation. *Journal of Statistical Software,* *94*(11), 1–20. <https://doi.org/10.18637/jss.v094.i11> [Article](https://doi.org/10.18637%2Fjss.v094.i11) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=deconvolveR%3A%20A%20G-Modeling%20Program%20for%20Deconvolution%20and%20Empirical%20Bayes%20Estimation&journal=Journal%20of%20Statistical%20Software&doi=10.18637%2Fjss.v094.i11&volume=94&issue=11&pages=1-20&publication_year=2020&author=Narasimhan%2CB&author=Efron%2CB) 9. Robbins, H. (1956). An empirical Bayes approach to statistics. In *Proc. 3rd Berkeley Sympos. Math. Statist. and Prob.* (Vol. I, pp. 157–163). Berkeley: University of California Press. 10. Strawderman, W. E. (1971). Proper Bayes minimax estimators of the multivariate normal mean. *Annals of Mathematical Statistics,* *42*(1), 385–388. <https://doi.org/10.1214/aoms/1177693528> [Article](https://doi.org/10.1214%2Faoms%2F1177693528) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=397939) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Proper%20Bayes%20minimax%20estimators%20of%20the%20multivariate%20normal%20mean&journal=Annals%20of%20Mathematical%20Statistics&doi=10.1214%2Faoms%2F1177693528&volume=42&issue=1&pages=385-388&publication_year=1971&author=Strawderman%2CWE) 11. Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. *Journal of the Royal Statistical Society Series B,* *58*(1), 267–288. [Article](https://doi.org/10.1111%2Fj.2517-6161.1996.tb02080.x) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=1379242) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Regression%20shrinkage%20and%20selection%20via%20the%20lasso&journal=Journal%20of%20the%20Royal%20Statistical%20Society%20Series%20B&doi=10.1111%2Fj.2517-6161.1996.tb02080.x&volume=58&issue=1&pages=267-288&publication_year=1996&author=Tibshirani%2CR) ### Discover content - [Journals A-Z](https://link.springer.com/journals/a/1) - [Books A-Z](https://link.springer.com/books/a/1) ### Publish with us - [Journal finder](https://link.springer.com/journals) - [Publish your research](https://www.springernature.com/gp/authors) - [Language editing](https://authorservices.springernature.com/go/sn/?utm_source=SNLinkfooter&utm_medium=Web&utm_campaign=SNReferral) - [Open access publishing](https://www.springernature.com/gp/open-science/about/the-fundamentals-of-open-access-and-open-research) ### Products and services - [Our products](https://www.springernature.com/gp/products) - [Librarians](https://www.springernature.com/gp/librarians) - [Societies](https://www.springernature.com/gp/societies) - [Partners and advertisers](https://www.springernature.com/gp/partners) ### Our brands - [Springer](https://link.springer.com/brands/springer) - [Nature Portfolio](https://www.nature.com/) - [BMC](https://link.springer.com/brands/bmc) - [Palgrave Macmillan](https://link.springer.com/brands/palgrave) - [Apress](https://link.springer.com/brands/apress) - [Discover](https://link.springer.com/brands/discover) - Your privacy choices/Manage cookies - [Your US state privacy rights](https://www.springernature.com/gp/legal/ccpa) - [Accessibility statement](https://link.springer.com/accessibility) - [Terms and conditions](https://link.springer.com/termsandconditions) - [Privacy policy](https://link.springer.com/privacystatement) - [Help and support](https://support.springernature.com/en/support/home) - [Legal notice](https://link.springer.com/legal-notice) - [Cancel contracts here](https://support.springernature.com/en/support/solutions/articles/6000255911-subscription-cancellations) Not affiliated [![Springer Nature](https://link.springer.com/oscar-static/images/logo-springernature-white-dbadd2cbd6.svg)](https://www.springernature.com/) Ā© 2026 Springer Nature
Readable Markdown
## 1 Introduction By and large, the statistics world is one of heuristics, approximations, and asymptotics. The James–Stein estimator arrived in that world in 1961 on a note of startling specificity: unseen parameters μ 1 , μ 2 , … , μ n produce independent observations x i ∼ ind N ( μ i , 1 ) , i \= 1 , … , n , (1) n ≄ 3. The James–Stein rule in its simplest form proposed estimating the μ i by μ ^ i J S \= ( 1 āˆ’ n āˆ’ 2 S ) x i ( S \= āˆ‘ i \= 1 n x i 2 ) . (2) Formula ([2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ2)) looked implausible: the estimate of μ i depended on the *other* observations x j, j ≠ i (through *S*), as well as x i, despite the independence assumption. Nevertheless, James and Stein showed that Rule ([2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ2)) *always* beat the obvious maximum likelihood estimates μ ^ i M L \= x ĀÆ i ( i \= 1 , … , n ) (3) in terms of total expected squared error E { āˆ‘ i \= 1 n ( μ ^ i āˆ’ μ i ) 2 } . (4) That ā€œalwaysā€ was the shocking part: two centuries of statistical theory, ANOVA, regression, multivariate analysis, etc., depended on maximum likelihood estimation. Did everything have to be rethought? One path forward involved Bayesian thinking. If we assumed that the μ i themselves came from a normal distribution, μ i ∼ ind N ( 0 , A ) for i \= 1 , … , n , (5) with variance A ≄ 0, the Bayes estimates would be μ ^ i B a y e s \= B x i ( B \= A / ( A \+ 1 ) ) . (6) We don’t know *A* or *B* but B ^ \= 1 āˆ’ ( n āˆ’ 2 ) / S (7) is *B*’s unbiased estimate: we can rewrite ([2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ2)) as μ ^ i J S \= B ^ x i , (8) which at least looks more plausible. In the language introduced by Robbins ([1956](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR9 "Robbins, H. (1956). An empirical Bayes approach to statistics. In Proc. 3rd Berkeley Sympos. Math. Statist. and Prob. (Vol. I, pp. 157–163). Berkeley: University of California Press.")), formula ([8](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ8)) is an *empirical Bayes* estimator, another shocking post-war statistical innovation. Carl Morris and I wrote a series of papers in the 1970 s exploring Bayesian roots of the James–Stein estimator (Efron and Morris, [1973](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR6 "Efron, B., & Morris, C. (1973). Stein’s estimation rule and its competitors—An empirical Bayes approach. Journal of the American Statistical Association, 68, 117–130.")). Something is lost in the empirical Bayes formulation, namely the frequentist ā€œalwaysā€ of expected square error minimization, but a lot is gained in flexibility and scope, as discussed in Sect. [2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Sec2). **Fig. 1** [![Fig. 1](https://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs42081-023-00209-y/MediaObjects/42081_2023_209_Fig1_HTML.png)](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/1) Prostate data: 6033 *x* values; mean 0.003, sd \= 1\.135; curve is proportional to a N ( 0 , 1 ) density [Full size image](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/1) Figure [1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig1) illustrates an example of simultaneous estimation pursued in Sect. 2.1 of Efron ([2010](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR2 "Efron, B. (2010). Large-scale inference: Empirical Bayes methods for estimation, testing, and prediction (Vol. 1). Cambridge: Cambridge University Press.")). A microarray study has compared expression levels between prostate cancer patients and control subjects for n \= 6033 genes. For each gene, a statistic x i has been calculated (essentially a ā€œ*z*\-valueā€), x i ∼ N ( μ i , 1 ) , i \= 1 , … , n , (9) where μ i measures the difference between cancer and control group levels. The solid curve in Fig. [1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig1) is a N ( 0 , 1 ) density scaled to have the same area as the histogram of the 6033 *x* values. A bad result from the researchers’ point of view would be a perfect fit of curve to histogram, which would imply all the genes have μ i \= 0, the ā€œnullā€ value of no difference between cancer patients and controls. That’s not what happened: the histogram has mildly heavy tails in both directions. The researchers were hoping to find genes with large values of ‖ μ i ‖—ones that might be a clue to prostate cancer etiology—as suggested by the heavy tails. How encouraged should they be? Not very, according to the James–Stein rule. The 6033 x i values have mean 0.003, which I’ll take to be zero, and empirical variance σ ^ 2 \= 1\.289. (10) The James–Stein estimate ([2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ2)) is μ ^ i J S \= ( 1 āˆ’ n āˆ’ 2 n āˆ’ 1 1 σ ^ 2 ) x i \= 0\.224 ā‹… x i , (11) so even x i \= 5 yields an estimate barely exceeding 1. Section [2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Sec2) suggests a more optimistic analysis. ## 2 Tweedie’s formula The impressive precision of the James–Stein theorem came at a cost in generality. Efforts to extend the theorem, say to Poisson rather than normal observations, or to measures of loss other than total squared error, gave encouraging asymptotic results but not the James–Stein kind of finite sample frequentist dominance. Better progress was possible on the empirical Bayes side of the street. *Tweedie’s formula* (Efron, [2011](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR3 "Efron, B. (2011). Tweedie’s formula and selection bias. Journal of the American Statistical Association, 106(496), 1602–1614. https://doi.org/10.1198/jasa.2011.tm11181 ")) has been particularly useful. We wish to calculate Bayesian estimates μ i B a y e s \= E { μ i ∣ x i } , i \= 1 , … , n , (12) in the normal sampling model ([1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ1)), starting from a given (possibly non-normal) prior Ļ€ ( μ ), applying to all *n* cases. Let *f*(*x*) be the marginal density f ( x ) \= ∫ R Ļ€ ( μ ) Ļ• ( x āˆ’ μ ) d μ , (13) with Ļ• the standard N ( 0 , 1 ) density and R the range of μ. (It isn’t necessary for Ļ€ ( ā‹… ) to be a continuous distribution but it simplifies notation.) Tweedie’s formula provides an elegant statement for μ i B a y e s, the posterior expectation of μ i given x i, μ i B a y e s \= E { μ i ∣ x i } \= x i \+ l ′ ( x i ) with l ′ ( x i ) \= d d x log ⁔ ( f ( x i ) ) . (14) In the empirical Bayes situation ([1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ1)), where the prior Ļ€ ( ā‹… ) is unknown, we can use the observed data x 1 , … , x n to estimate the marginal density *f*(*x*), say by f ^ ( x ), giving empirical Bayes estimates μ ^ i \= x i \+ l ^ ′ ( x i ) . (15) The Bayes estimate ([14](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ14)) can be thought of as the MLE x i plus a Bayesian correction term l ′ ( x i ). When the prior Ļ€ ( μ ) is the N ( 0 , A ) distribution ([5](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ5)), μ i B a y e s equals B x i ([6](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ6)). Simple formulas for μ i B a y e s give out for most other choices of Ļ€ ( μ ) but now, in the machine learning era[Footnote 1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fn1) of statistical research, numerical methods provide useful ways forward, as discussed next. The *log polynomial class*[Footnote 2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fn2) of marginal densities defines *f*(*x*) by log ⁔ ( f β ( x ) ) \= β 0 \+ β ⊤ c ( x ) . (16) Here c ( x ) \= ( x , x 2 , … , x J ) ⊤ and β \= ( β 1 , … , β J ) ⊤ , (17) with β 0 chosen to make f β ( x ) integrate to 1. The choice J \= 2 gives normal marginals; larger values of *J* allow for marginal non-normality. **Fig. 2** [![Fig. 2](https://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs42081-023-00209-y/MediaObjects/42081_2023_209_Fig2_HTML.png)](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/2) Prostate data: Tweedie’s estimate of E { μ ∣ x }, 5 degrees of freedom; dashed curve is James–Stein estimate [Full size image](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/2) The choice J \= 5 was applied to the prostate cancer data of Fig. [1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig1): Tweedie’s formula ([14](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ14)) gave μ ^ ( x ) \= E { μ ∣ x }, graphed as the solid curve in Fig. [2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig2). It differs markedly from the James–Stein estimate J \= 2, the dashed line. At x \= 4 for example, the J \= 5 estimate is[Footnote 3](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fn3) E { μ ∣ x \= 4 } \= 2\.555 (18) compared to 0.901 for the James–Stein estimate. The estimated curve E { μ ∣ x } is *empirical Bayes* in the same sense as ([8](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ8)): the parameter vector β was selected by maximum likelihood, as discussed next. With J \= 5, the prior was able to adapt to the ā€œfishing expeditionā€ nature of such microarray studies, where we expect most of the genes to be null or close to null, with μ i nearly zero (corresponding here to the flat part of the curve for *x* between āˆ’ 2 and 2) and, hopefully, a small proportion of interestingly large μ is. The sample size n \= 6033 has much to do with Fig. [2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig2). James and Stein ([1961](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR7 "James, W., & Stein, C. (1961). Estimation with quadratic loss. In Proc. 4th Berkeley Sympos. Math. Statist. and Prob. (Vol. I, pp. 361–379). Berkeley: University of California Press.")) was usually considered in terms of small samples, perhaps n ≤ 20, for which there would be little hope of seeing the detail in Fig. [2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig2). The term ā€œmachine learning eraā€ seems less fanciful when considering the scale of problems statisticians are now asked to deal with, as well as the tools they use to solve them. It looks like it might be hard work computing Fig. [2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig2) but it’s not. The histogram in Fig. [1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig1) has 97 bins, with centerpoints v v \= ( āˆ’ 4\.4 , āˆ’ 4\.3 , … , 5\.1 , 5\.2 ) . (19) Let y j be the count in bin *j*, that is, the number of the 6033 x i values falling into it, with the vector of counts being y y \= ( y 1 , … , y 97 ) . (20) Then the single R command l l ^ \= log ⁔ ( glm ( y y ∼ poly ( v v , 5 ) , poisson ) \$ fit ) (21) provides a close approximation to the MLE of log ⁔ f ( x ) in ([14](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ14)); numerical differentiation of l l ^ gives Tweedie’s estimate. Section 3.4 of Efron ([2023](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR5 "Efron, B. (2023). Exponential Families in Theory and Practice. Cambridge: Cambridge University Press.")) shows why Poisson regression ([21](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ21)) is appropriate here. The James–Stein theorem depends on the independence assumption in ([1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ1)), unlikely to be true in the microarray study, but the estimates ([2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ2)) have a certain marginal validity even under dependence. This is clearer from the empirical Bayes point of view. The Tweedie estimate x i \+ l ^ ′ ( x i ) requires only that l ^ ′ ( x ) be close to l ′ ( x ), not that it be estimated from independent x is.[Footnote 4](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fn4) ## 3 Shrinkage estimators James and Stein’s paper aroused excited interest in the statistics community when it arrived in 1961. Most of the excitement focused on the strict inadmissibility of the traditional maximum likelihood estimate demonstrated by the James–Stein rule. Other rules dominating the MLE were discovered, for instance the Bayes estimator of Strawderman ([1971](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR10 "Strawderman, W. E. (1971). Proper Bayes minimax estimators of the multivariate normal mean. Annals of Mathematical Statistics, 42(1), 385–388. https://doi.org/10.1214/aoms/1177693528 ")), that was itself admissible while rendering the MLE inadmissible. Big new ideas can take a while to make their true impact felt. The James–Stein rule had an influential side effect on subsequent theory and practice in that it demonstrated, in an inarguable way, the virtues of *shrinkage estimation*: given an ensemble of problems, individual estimates are shrunk toward a central point; that is, a deliberate bias is introduced, pulling estimates away from their MLEs for the sake of better group performances. Admissibility and inadmissibility aren’t much in the air these days, while shrinkage estimation has gone on to play a major role in modern practice. A spectacular success story is the lasso (Tibshirani, [1996](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR11 "Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society Series B, 58(1), 267–288.")). Lasso shrinkage is extreme, pulling some (often most) of the coefficient estimates all the way back to zero. Bayes and empirical Bayes rules tend to be strong shrinkers. Tweedie’s estimate in Fig. [2](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig2) (J \= 5) shrinks the estimate of E { μ ∣ x \= 4 } from its MLE value 4 down to 2.555. For μ between āˆ’ 1 and 1, the shrinkage is almost all the way to zero. The reader may have been surprised to see that neither Tweedie’s formula ([14](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ14)) for E { μ i ∣ x i } nor its empirical version ([15](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ15)) require estimation of the prior Ļ€ ( μ ). This is a special property of the posterior expectation E { μ i ∣ x i } and isn’t available for say Pr { μ i ≄ 2 ∣ x i }, or most other Bayesian targets. ā€œBayesian deconvolutionā€ (Efron, [2016](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR4 "Efron, B. (2016). Empirical Bayes deconvolution estimates. Biometrika, 103(1), 1–20. https://doi.org/10.1093/biomet/asv068 ")) uses low-dimensional parametric modeling of Ļ€ ( μ ) for general empirical Bayes computations. It was applied to finding a prior density Ļ€ ( μ ) that would give the distribution of *x* seen in Fig. [1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig1), assuming the normal sampling model ([1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ1)). The deconvolution model for Ļ€ ( μ ) used a delta function at μ \= 0 (for the ā€œnullā€ genes) and a natural spline function with four degrees of freedom for the non-null cases. **Fig. 3** [![Fig. 3](https://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs42081-023-00209-y/MediaObjects/42081_2023_209_Fig3_HTML.png)](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/3) Empirical Bayes conditional density of μ given μ not zero; Pr { μ \= 0 } equals 0.825 [Full size image](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/3) The estimated prior[Footnote 5](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fn5)Ļ€ ^ ( μ ) is shown in Fig. [3](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig3); it put probability 0.825 on μ \= 0, while the conditional distribution given μ ≠ 0 was a moderately heavy-tailed version of N ( 0 , 1\.33 2 ). Based on Ļ€ ^ ( μ ) we can form estimates of *any* Bayesian target, for instance Pr ^ { μ i ≄ 2 ∣ x i \= 4 } \= 0\.80. Figure [3](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig3) is a direct descendent of the James–Stein rule, now 60-plus years on. A less-direct descendent, but still on the family tree, arrived in 1995. The *false discovery rate* paper by Benjamini and Hochberg concerned simultaneous hypothesis testing. Looking at Fig. [1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig1), which of the n \= 6033 genes can confidently be labeled as non-null, that is as having μ i ≠ 0? Suppose for convenience that the x is are ordered from smallest to largest. The right-sided significance level for testing μ i \= 0 is S 0 ( x i ) \= 1 āˆ’ Φ ( x i ) , (22) where Φ is the standard normal cumulative distribution function. Of the 6033 genes, 401 had S i ≤ 0\.05, the usual rejection level for individual testing, but even if actually *all* of the genes were null we would expect 302 such rejections, so individual testing can’t be right. Benjamini and Hochberg proposed a novel simultaneous testing rule that safely controls the number of ā€œfalse discoveriesā€ — genes falsely labeled ā€non-nullā€ — while not being discouragingly strict. (My summary here won’t give the BH rule its full due; see Chapter 4 of Efron ([2010](https://link.springer.com/article/10.1007/s42081-023-00209-y#ref-CR2 "Efron, B. (2010). Large-scale inference: Empirical Bayes methods for estimation, testing, and prediction (Vol. 1). Cambridge: Cambridge University Press.")) for a more complete description.) Let S ^ ( x ) be the observed proportion of x is exceeding value *x*, and define F d r ^ ( x ) \= Ļ€ 0 S 0 ( x ) / S ^ ( x ) , (23) where Ļ€ 0 is the proportion of null genes among all *n*.[Footnote 6](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fn6) For a fixed control level α, such as α \= 0\.1, the BH rule says to reject the null hypothesis μ i \= 0 for those genes having F d r ^ ( x i ) ≤ α . (24) The Benjamini–Hochberg theorem states that under independence assumptions like ([1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ1)), the expected proportion of false discoveries by rule ([24](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ24)) is α. **Fig. 4** [![Fig. 4](https://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs42081-023-00209-y/MediaObjects/42081_2023_209_Fig4_HTML.png)](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/4) Prostate data: left Fdr and right Fdr; dashes show 60 genes with F d r \< 0\.1 [Full size image](https://link.springer.com/article/10.1007/s42081-023-00209-y/figures/4) Figure [4](https://link.springer.com/article/10.1007/s42081-023-00209-y#Fig4) shows F d r ^ for the prostate cancer data and also for the left-sided Fdr estimate, where significance is defined by S 0 ( x i ) \= Φ ( x i ) rather than ([22](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ22)). I applied the BH rule with α \= 0\.1 which labeled 60 genes as non-null, 32 on the left and 28 on the right. The BH theorem says that we can expect 6 of the 60 to actually be null. The fdr story has evolved very much along the lines of its James–Stein predecessor. Intense initial interest focused on the exact frequentist control of false discovery rates. The Bayes and empirical Bayes implications came later: as at ([5](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ5)), we assume that each x i is a realization of a random variable *x* given by μ ∼ Ļ€ ( μ ) and x ∣ μ ∼ p ( x ∣ μ ) , (25) where p ( x ∣ μ ) is a known probability kernel which I’ll take here to be the normal sampling model ([1](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ1)). Then if *S*(*x*) is 1 minus the cdf of the marginal density ([13](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ13)), Bayes rule gives Pr { μ \= 0 ∣ x } \= Ļ€ 0 S 0 ( x ) / S ( x ) . (26) Comparing ([26](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ26)) with ([23](https://link.springer.com/article/10.1007/s42081-023-00209-y#Equ23)) says that the BH rule amounts to labeling case *i* as non-null if its obvious empirical Bayes estimate of nullness is less than α. This is less precise than the frequentist control theorem but, as with the James–Stein estimator, is more robust in not demanding independence among the x is. The family resemblance between JS and BH is through shrinkage: in the BH case the shrinkage of significance levels. For instance, x i \= 3 has individual significance level 0.001 against nullness, whereas F d r ^ \= 0\.164 for the prostate data, i.e, still with about a 1/6 chance of gene *i* being null. So what does machine learning have to do with the James–Stein estimator? Nothing to its birth but, as the articles in this volume show, a great deal to its downstream effects on statistical theory and practice. Charles Stein, who was a good applied statistician when he put his mind to it, might have enjoyed these developments, but maybe not; his heart was always with the mathematics. ## References - Benjamini, Y., & Hochberg, Y. (1995). Controlling the false discovery rate: A practical and powerful approach to multiple testing. *Journal of the Royal Statistical Society Series B,* *57*(1), 289–300. [Article](https://doi.org/10.1111%2Fj.2517-6161.1995.tb02031.x) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=1325392) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Controlling%20the%20false%20discovery%20rate%3A%20A%20practical%20and%20powerful%20approach%20to%20multiple%20testing&journal=Journal%20of%20the%20Royal%20Statistical%20Society%20Series%20B&doi=10.1111%2Fj.2517-6161.1995.tb02031.x&volume=57&issue=1&pages=289-300&publication_year=1995&author=Benjamini%2CY&author=Hochberg%2CY) - Efron, B. (2010). *Large-scale inference: Empirical Bayes methods for estimation, testing, and prediction* (Vol. 1). Cambridge: Cambridge University Press. [Book](https://doi.org/10.1017%2FCBO9780511761362) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Large-scale%20inference%3A%20Empirical%20Bayes%20methods%20for%20estimation%2C%20testing%2C%20and%20prediction&doi=10.1017%2FCBO9780511761362&publication_year=2010&author=Efron%2CB) - Efron, B. (2011). Tweedie’s formula and selection bias. *Journal of the American Statistical Association,* *106*(496), 1602–1614. <https://doi.org/10.1198/jasa.2011.tm11181> [Article](https://doi.org/10.1198%2Fjasa.2011.tm11181) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=2896860) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Tweedie%E2%80%99s%20formula%20and%20selection%20bias&journal=Journal%20of%20the%20American%20Statistical%20Association&doi=10.1198%2Fjasa.2011.tm11181&volume=106&issue=496&pages=1602-1614&publication_year=2011&author=Efron%2CB) - Efron, B. (2016). Empirical Bayes deconvolution estimates. *Biometrika,* *103*(1), 1–20. <https://doi.org/10.1093/biomet/asv068> [Article](https://doi.org/10.1093%2Fbiomet%2Fasv068) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=3465818) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Empirical%20Bayes%20deconvolution%20estimates&journal=Biometrika&doi=10.1093%2Fbiomet%2Fasv068&volume=103&issue=1&pages=1-20&publication_year=2016&author=Efron%2CB) - Efron, B. (2023). *Exponential Families in Theory and Practice*. Cambridge: Cambridge University Press. [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Exponential%20Families%20in%20Theory%20and%20Practice&publication_year=2023&author=Efron%2CB) - Efron, B., & Morris, C. (1973). Stein’s estimation rule and its competitors—An empirical Bayes approach. *Journal of the American Statistical Association,* *68*, 117–130. [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=388597) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Stein%E2%80%99s%20estimation%20rule%20and%20its%20competitors%E2%80%94An%20empirical%20Bayes%20approach&journal=Journal%20of%20the%20American%20Statistical%20Association&volume=68&pages=117-130&publication_year=1973&author=Efron%2CB&author=Morris%2CC) - James, W., & Stein, C. (1961). Estimation with quadratic loss. In *Proc. 4th Berkeley Sympos. Math. Statist. and Prob.* (Vol. I, pp. 361–379). Berkeley: University of California Press. - Narasimhan, B., & Efron, B. (2020). deconvolveR: A G-Modeling Program for Deconvolution and Empirical Bayes Estimation. *Journal of Statistical Software,* *94*(11), 1–20. <https://doi.org/10.18637/jss.v094.i11> [Article](https://doi.org/10.18637%2Fjss.v094.i11) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=deconvolveR%3A%20A%20G-Modeling%20Program%20for%20Deconvolution%20and%20Empirical%20Bayes%20Estimation&journal=Journal%20of%20Statistical%20Software&doi=10.18637%2Fjss.v094.i11&volume=94&issue=11&pages=1-20&publication_year=2020&author=Narasimhan%2CB&author=Efron%2CB) - Robbins, H. (1956). An empirical Bayes approach to statistics. In *Proc. 3rd Berkeley Sympos. Math. Statist. and Prob.* (Vol. I, pp. 157–163). Berkeley: University of California Press. - Strawderman, W. E. (1971). Proper Bayes minimax estimators of the multivariate normal mean. *Annals of Mathematical Statistics,* *42*(1), 385–388. <https://doi.org/10.1214/aoms/1177693528> [Article](https://doi.org/10.1214%2Faoms%2F1177693528) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=397939) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Proper%20Bayes%20minimax%20estimators%20of%20the%20multivariate%20normal%20mean&journal=Annals%20of%20Mathematical%20Statistics&doi=10.1214%2Faoms%2F1177693528&volume=42&issue=1&pages=385-388&publication_year=1971&author=Strawderman%2CWE) - Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. *Journal of the Royal Statistical Society Series B,* *58*(1), 267–288. [Article](https://doi.org/10.1111%2Fj.2517-6161.1996.tb02080.x) [MathSciNet](http://www.ams.org/mathscinet-getitem?mr=1379242) [Google Scholar](http://scholar.google.com/scholar_lookup?&title=Regression%20shrinkage%20and%20selection%20via%20the%20lasso&journal=Journal%20of%20the%20Royal%20Statistical%20Society%20Series%20B&doi=10.1111%2Fj.2517-6161.1996.tb02080.x&volume=58&issue=1&pages=267-288&publication_year=1996&author=Tibshirani%2CR) [Download references](https://citation-needed.springer.com/v2/references/10.1007/s42081-023-00209-y?format=refman&flavour=references)
Shard129 (laksa)
Root Hash17645177711233004329
Unparsed URLcom,springer!link,/article/10.1007/s42081-023-00209-y s443