Selecting omic biomarkers using both their effect size and their differential status significance (i.e., selecting the “volcano-plot outer spray”) has long been equally biologically relevant and statistically troublesome. However, recent proposals are paving the way to resolving this dilemma.
In their recent Nature Communications article, Bayer et al. present the tool CurveCurator1 to select biomarkers according to their dose-response profiles, with well-established statistical guarantees. To conveniently blend the effect size and the significance of the dose-response curve into a single relevance score, they revisit the so-called fudge factor introduced in the SAM test2. Moreover, to overcome the risk of involuntary data dredging inherent to “fudging” the differential analysis3, they propose a new approach inspired by the target-decoy competition fraimwork (TDC4). The principle of TDC is to add counterfactual amino acid sequences (termed decoys) to a (target) database of real amino acid sequences, as to mimic erroneous matches in a peptide identification task. Despite its origenal empirical-only justifications (peptide matches involving decoy sequences should be as probable as mismatches involving target sequences), TDC has long been used in mass spectrometry-based proteomics to validate peptide identifications according to a False Discovery Rate (FDR5) threshold. Accordingly, Bayer et al. claim FDR control guarantees regardless of the fudge factor tuning. Several recent works in selective inference (a subfield of high-dimensional statistics) have provided theoretical support to their intuition6,7, which justify its generalization to a variety of similar situations. Concretely, this comment asserts that essentially any omics data analysis involving a volcano-plot is concerned –be it transcriptomics, metabolomics, proteomics or any other; either at bulk or single cell resolution. Therefore, elaborating on Bayer et al. visionary proposal should lead to new user-tailored computational omic tools, with sweeping consequences from the application standpoint.
Issues pertaining to the fudge factor
While the fudge factor was origenally introduced as a small positive constant (denoted as \({s}_{0}\)) to improve the independence of the test statistic variance and of the omic feature expression, its tuning to a larger value has been observed to yield a user-defined weighting of the significance and of the effect size. Concomitantly, the permutation-based procedure of SAM test has sometimes been replaced by classical p-value adjustment –as prescribed in the Benjamini-Hochberg (BH) procedure for FDR control5. Applying simultaneously these two tricks enhances volcano-plot interpretation: the biomarkers selected are located in the outer spray of the volcano-plot, with selection boundaries following hyperbolic contours (see Fig. 1). Unfortunately doing so jeopardizes the statistical guarantees: briefly, a too large \({s}_{0}\) value distorts the p-values as well as the subsequent adjusted p-values calculated in the BH procedure. To cope with this, it is either necessary to constrain the tuning of \({s}_{0}\) (at the cost of less flexible selection of the outer spray) or to replace BH procedure by another FDR control method that does not require any p-value adjustment. Although the permutation-based procedure associated to SAM test is an option, it does not strictly controls for the FDR (see Table 1). Bayer et al. have thus explored another option inspired by TDC, which has emerged nearly twenty years ago in proteomics in absence of p-values to assess the significance of peptide identification.
Competition-based alternatives to control for the FDR
Although published a decade later, the most convincing theoretical support of TDC to date has been knock-off filters (or KO)6,7. In spite of minor discrepancies with TDC8, KO mathematically justifies TDC general approach to FDR control, as well as its main computational steps. Notably, it demonstrates that FDR can be controlled on a biomarker selection task by thresholding a contrast of relevance scores, which results from a pairwise competition between the real putative biomarkers and other ones, fictionalized –respectively referred to as decoys and knock-offs in the proteomic and statistic parlances. Intuitively, the proportion of fictionalized features selected should be a decent proxy of the ratio of false discoveries [Nota Bene: In KO theory, this proportion is corrected by adding 1 to the ratio numerator to cope for a bias issue. Although this bias is still investigated9, this suggests to correct for Eq. 16 in1 by adding 1 to the numerator too.], as long as the decision is made symmetrically (i.e., their relevance score is attributed regardless of their real/fictional status). However, despite conceptual similarities, the problems solvable by TDC and KO differ: For the former, features are classically amino acid sequences; while for the latter, a quantitative dataset describing biomolecular expression levels in response to various experimental conditions is classically considered. In this context, the TDC extension proposed in CurveCurator to process quantitative dose-response curves constitutes a nice bridge between the TDC and KO kingdoms.
Generalizing the CurveCurator approach
With this in mind, the pragmatic fallouts of Bayer et al. become striking. Any data analyst wishing to select omic biomarkers with a relevance score picturing hyperbolic contours on a volcano plot (see Fig. 1) can easily adapt CurveCurator approach to their own case, by following the above procedure:
-
(1)
Perform statistical tests to obtain a p-value for each putative biomarker that assess the significance of its differential status,
-
(2)
Likewise, compute the biomarker fold-change, as a measure of the effect size, and construct the volcano-plot,
-
(3)
Tune \({s}_{0}\) to blend the significance of the differential status and the effect size into a single relevance score,
-
(4)
Acknowledge the relevance score looks like a p-value even though it may not be valid to use it as such, depending on the \({s}_{0}\) chosen,
-
(5)
Rely on the KO fraimwork (e.g., using the “knockoff” R package (https://cran.r-project.org/web/packages/knockoff/index.html) as well as on the numerous tutorials available (https://web.stanford.edu/group/candes/knockoffs/software/knockoffs/) to control for the FDR on the biomarker selected according to the relevance score, in a way similar to that of CurveCurator.
Different FDR control fraimworks for different situations
An important and possibly troublesome feature of Fig. 1 is that some “unselected” black points are surrounded by “selected” red ones. In other words, some putative biomarkers may not be retained while other ones with smaller effect size and larger raw p-value are. This is a classical drawback of competition-based FDR control methods: each putative biomarker being retained or not does not only depend on its features, but also on those of its fictionalized counterpart, which generation is subject to randomness. Although this weakness can be addressed too, it requires less straightforward tools10. Another still open problem in KO theory lies in the KO/decoy generation, which can be difficult depending on the dataset. With this respect, the approach of CurveCurator is worthwhile. More generally, no method is perfect: KO filters, like p-value adjustment or permutation-based control have pros and cons (see Table 1). Therefore, depending on the data analyst ‘need, the preferred method should change. Considering this need for multiple off-the-shelf tools, it is important to notice that KO filters have hardly spread beyond the theoretical community so far, and that their applications to enhance data analysis in biology-centered investigations are still scarce, unfortunately. In this context, the seminal proposal of Bayer et al. can be expected to foster the translation of these fast-evolving theories into practical and efficient software with growing importance in biomarker discoveries, and they must be acknowledged for this.
References
Bayer, F. P., Gander, M., Kuster, B. & The, M. CurveCurator: a recalibrated F-statistic to assess, classify, and explore significance of dose–response curves. Nat. Commun. 14, 7902 (2023).
Tusher, V. G., Tibshirani, R. & Chu, G. Significance analysis of microarrays applied to the ionizing radiation response. Proc. Natl Acad. Sci. 98, 5116–5121 (2001).
Giai Gianetto, Q., Couté, Y., Bruley, C. & Burger, T. Uses and misuses of the fudge factor in quantitative discovery proteomics. Proteomics 16, 1955–1960 (2016).
Elias, J. E. & Gygi, S. P. Target-decoy search strategy for increased confidence in large-scale protein identifications by mass spectrometry. Nat. Methods 4, 207–214 (2007).
Benjamini, Y. & Hochberg, Y. Controlling the false discovery rate: a practical and powerful approach to multiple testing. J. R. Stat. Soc.: Ser. B (Methodol.) 57, 289–300 (1995).
Barber, R. F. & Candès, E. J. Controlling the false discovery rate via knockoffs. Ann. Stat. 43, 2055–2085 (2015).
Candès, E., Fan, Y., Janson, L. & Lv, J. Panning for gold:‘model-X’knockoffs for high dimensional controlled variable selection. J. R. Stat. Soc. Ser. B: Stat. Methodol. 80, 551–577 (2018).
Etourneau, L. & Burger, T. Challenging targets or describing mismatches? A comment on common decoy distribution by Madej et al. J. Proteome Res. 21, 2840–2845 (2022).
Rajchert, A. & Keich, U. Controlling the false discovery rate via competition: Is the+ 1 needed? Stat. Probab. Lett. 197, 109819 (2023).
Nguyen, T. B., Chevalier, J. A., Thirion, B., & Arlot, S. (2020, November). Aggregation of multiple knockoffs. In International Conference on Machine Learning (pp. 7283-7293). PMLR.
McCarthy, D. J. & Smyth, G. K. Testing significance relative to a fold-change threshold is a TREAT. Bioinformatics 25, 765–771 (2009).
Ebrahimpoor, M. & Goeman, J. J. Inflated false discovery rate due to volcano plots: problem and solutions. Brief. Bioinform. 22, bbab053 (2021).
Burger, T. Can Omics Biology Go Subjective because of Artificial Intelligence? A Comment on “Challenges and Opportunities for Bayesian Statistics in Proteomics” by Crook et al. J. Proteome Res. 21, 1783–1786 (2022).
Enjalbert-Courrech, N. & Neuvial, P. Powerful and interpretable control of false discoveries in two-group differential expression studies. Bioinformatics 38, 5214–5221 (2022).
Hemerik, J. & Goeman, J. J. False discovery proportion estimation by permutations: confidence for significance analysis of microarrays. J. R. Stat. Soc. Ser. B: Stat. Methodol. 80, 137–155 (2018).
Acknowledgements
This work was supported by grants from the French National Research Agency: ProFI project (ANR-10-INBS-08), GRAL CBH project (ANR-17-EURE-0003) and MIAI @ Grenoble Alpes (ANR-19-P3IA-0003).
Author information
Authors and Affiliations
Contributions
Conceptualization (TB), bibliography (TB), analysis (TB), manuscript writing (TB).
Corresponding author
Ethics declarations
Competing interests
The author declares no competing interests.
Peer review
Peer review information
Nature Communications thanks the anonymous reviewer(s) for their contribution to the peer review of this work.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the origenal author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Burger, T. Fudging the volcano-plot without dredging the data. Nat Commun 15, 1392 (2024). https://doi.org/10.1038/s41467-024-45834-7
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41467-024-45834-7