Welcome to E2P Simulator! This guide will help you understand what it does,
why it is needed, and how to use it.
E2P Simulator (Effect-to-Prediction Simulator) allows researchers to interactively and quantitatively explore the relationship between effect sizes (e.g., Cohen's d, Odds Ratio, Pearson's r) and their predictive utility (e.g., AUC, PR-AUC, MCC), while accounting for measurement reliability and outcome base rates.
In other words, E2P Simulator is a tool for performing predictive utility analysis - estimating how effect sizes will translate into real-world prediction or what effect sizes are needed to achieve a desired predictive performance. Much like how power analysis tools (such as G*Power) help researchers plan for statistical significance, E2P Simulator helps plan for practical significance.
E2P Simulator has several key applications:
Many research areas such as biomedical, behavioral, education, and sports sciences, are increasingly studying individual differences in order to build predictive models to personalize treatments, learning, and training. Identifying reliable biomarkers and other predictors is central to these efforts. Yet, several entrenched research practices continue to undermine the search for predictors:
Together, these issues undermine the quality and impact of academic research, resulting in an abundance of statistically significant but practically negligible findings. Conflating statistical significance with practical significance, researchers develop unrealistic expectations about the potential impact of their findings, leading to inefficient study planning, resource misallocation, and considerable waste of time and funding. For example, collecting large datasets or building complex predictive models is pointless if the effect sizes of individual predictors are known to be small (which is usually the case).
E2P Simulator is designed to address these fundamental challenges by placing effect sizes, measurement reliability, and outcome base rates at the center of study planning and interpretation. It helps researchers understand how these factors jointly shape predictive utility, and guides them in making more informed decisions about their studies.
E2P Simulator is designed to be intuitive and interactive. You can explore different scenarios by adjusting effect sizes, measurement reliability, base rates, and decision threshold, and immediately see how these changes impact predictive performance through various visualizations and metrics. Still, in this section we will highlight and clarify some of the key features and assumptions of the simulator.
The image above provides an overview of all E2P Simulator's interactive components.
E2P Simulator provides two analysis modes that cover the two most common research scenarios:
Measurement reliability attenuates observed effect sizes, and in turn attenuates predictive performance. The simulator allows you to specify reliability using the Intraclass Correlation Coefficient (ICC) for continuous measurements and Cohen's kappa (κ) for binary classifications. These typically correspond to test-retest reliability and inter-rater reliability, respectively. You can toggle between "true" effect sizes (what would be observed with perfect measurement) and "observed" effect sizes (what we actually see given imperfect reliability).
Note that the simulator does not account for sample size limitations, which can introduce additional uncertainty around the true effect size through sampling error.
The base rate refers to how common an outcome is within a specific population. For instance, if an effect size measures the difference between a group with a disorder and a healthy control group, the base rate would be the disorder's prevalence in the general population. However, if you're focusing on a pre-identified high-risk group, the relevant base rate becomes the disorder's prevalence within this specific high-risk cohort. In Bayesian statistics, this base rate is analogous to the prior probability of the outcome, established before considering any particular predictor.
The base rate directly affects PPV and NVP, and indirectly the metrics that depend on it: PR-AUC, MCC, and F-1 score (see Understanding Predictive Metrics).
Both binary and continuous outcomes analysis modes include calculators that help explore how multiple predictors can be combined to achieve stronger effects:
The calculators can help approximate the expected performance of multivariate models without having to train the full models and thus help with research planning and model development. More specifically, they illustrate:
Both calculators correspond to fundamental predictive models in statistics: the Mahalanobis D Calculator approximates Linear Discriminant Analysis and logistic regression, while the Multivariate R² Calculator aligns with multiple linear regression.
Like the statistical methods they approximate, the calculators operate under several simplifying assumptions:
Despite these limitations, these calculators serve as valuable tools for building intuition about how multiple predictors combine to achieve stronger effects. Even though real-world predictors will vary in their individual strengths and collinearity, the overall patterns demonstrated (such as diminishing returns and the impact of shared variance) remain informative for understanding multivariate relationships.
When using a predictor to classify cases into two groups, there are four possible outcomes: True Positives (TP), False Positives (FP), False Negatives (FN), and True Negatives (TN). These form the basis for all predictive metrics.
The image above illustrates how these four outcomes relate to various classification metrics. On the left, you can see how negative (e.g., controls) and positive (e.g., cases) distributions overlap and how a classification threshold (red line) creates these four outcomes. On the right, you'll find the confusion matrix and the formulas for key metrics derived from it. Note how some metrics have multiple names (e.g., Sensitivity/Recall/TPR, Precision/PPV) - this reflects how the same concepts are referred to differently across fields like medicine, cognitive science, and machine learning.
Depending on your research context, certain metrics may be more relevant than others:
Some metrics evaluate performance across all possible thresholds and can serve as a better summary of the overall model performance. These include:
Both ROC-AUC and PR-AUC represent areas under their respective curves and are mathematically expressed as integrals:
Where P is Precision (PPV) and R is Recall (Sensitivity). These integrals are computed using trapezoidal numerical integration.
Decision Curve Analysis (Vickers & Elkin, 2006) evaluates the clinical utility of a predictive model or a single predictor by plotting net benefit across different threshold probabilities. Unlike ROC and PR curves that focus on discrimination, DCA incorporates the relative costs of false positives and false negatives, helping determine whether using a prediction model to guide treatment decisions provides more benefit than simple strategies.
A DCA plot typically includes three key curves:
The net benefit formula accounts for both the benefits of true positives and the costs of false positives:
Where Pt is the threshold probability, representing the odds at which a patient would be willing to accept treatment. In practical terms, Pt reflects how many unnecessary interventions one is willing to accept to prevent one adverse outcome. The ratio Pt/(1-Pt) in the net benefit formula captures the relative weighting of false positives compared to true positives. Pt itself can be calculated as:
Where CFP is the cost of a false positive (unnecessary treatment) and CFN is the cost of a false negative (missed case). Pt can also be estimated through clinician surveys, patient preferences, or clinical guidelines. The simulator allows you to explore a range of Pt values to identify when your predictor provides meaningful clinical benefit.
A-NBC (Area Under the Net Benefit Curve) summarizes the overall clinical utility of a model across a range of threshold probabilities (Talluri & Shete, 2016). It is calculated using trapezoidal integration:
Where the integration is performed over a clinically relevant range of threshold probabilities (adjustable using the movable gray bars in the simulator). A larger A-NBC indicates greater overall clinical utility within that range.
ΔA-NBC (Delta A-NBC) represents the additional clinical utility that a predictive model provides compared to the best available simple strategy (either treating all patients or treating none). Rather than comparing only against treating no one, ΔA-NBC is calculated as:
At each threshold probability, the model's net benefit is compared against whichever simple strategy performs better at that threshold. This provides a more conservative and clinically meaningful assessment of the model's added value. A positive ΔA-NBC indicates that the predictive model offers genuine improvement over the best simple strategy, while values near zero suggest that simple strategies may be equally effective.
Note: For computational simplicity, the ΔA-NBC calculation in this simulator assumes a uniform distribution of threshold probabilities across the selected range. In practice, the distribution of clinically relevant threshold probabilities may be non-uniform, which could affect the relative weighting of different threshold ranges in the overall utility assessment (Talluri & Shete, 2016).
DCA is particularly valuable because it:
For more information about DCA, visit https://mskcc-epi-bio.github.io/decisioncurveanalysis.
Can we predict depression diagnosis using a cognitive biomarker?
Can we predict who will respond to antidepressant treatment using task-based brain activity measures?
Can we predict who will attempt suicide using a combination of risk factors?
Here we will rely mostly on Borges et al., 2011, who analyzed the data of 108,705 adults from 21 countries, and using a logistic regression model found that combining a range of risk factors (socio-demographics, parent psychopathology, childhood adversities, DSM-IV disorders, and history of suicidal behavior) led to AUC = 0.74-0.8 discrimination performance between those who did and did not attempt suicide.
E2P Simulator is an open-source project - feedback, bug reports, and suggestions for improvement are welcome. The easiest way to do so is through the GitHub Issues page.
You can view the source code, track development, and contribute directly at the project's GitHub repository.
For other inquiries, you can find my contact information here.