Loading Icon

Regression analysis

In statistical modeling, regression analysis is a set of statistical processes for estimating the relationships between a dependent variable (often called the 'outcome variable') and one or more independent variables (often called 'predictors', 'covariates', or 'features'). The most common form of regression analysis is linear regression, in which one finds the line (or a more complex linear combination) that most closely fits the data according to a specific mathematical criterion. For example, the method of ordinary least squares computes the unique line (or hyperplane) that minimizes the sum of squared differences between the true data and that line (or hyperplane). For specific mathematical reasons (see linear regression), this allows the researcher to estimate the conditional expectation (or population average value) of the dependent variable when the independent variables take on a given set of values. Less common forms of regression use slightly different procedures to estimate alternative location parameters (e.g., quantile regression or Necessary Condition Analysis) or estimate the conditional expectation across a broader collection of non-linear models (e.g., nonparametric regression).

Metrics Summary

Total Publications
Lifetime
33,509
Prior Five Years
10,291
Total Citations
Lifetime
637,820
Prior Five Years
65,803
Total Scholars
Lifetime
50,903
Prior Five Years
42,335

Publications and Citation History

Publications based on Disciplines

Scholars based on Disciplines

Publications based on Fields

Scholars based on Fields

Highly Ranked Scholars™

Lifetime
Prior Five Years

Highly Cited Publications

Lifetime