🕷️ Crawler Inspector

URL Lookup

Direct Parameter Lookup

Raw Queries and Responses

1. Shard Calculation

Query:
Response:
Calculated Shard: 169 (from laksa078)

2. Crawled Status Check

Query:
Response:

3. Robots.txt Check

Query:
Response:

4. Spam/Ban Check

Query:
Response:

5. Seen Status Check

ℹ️ Skipped - page is already crawled

đź“„
INDEXABLE
âś…
CRAWLED
1 day ago
🤖
ROBOTS ALLOWED

Page Info Filters

FilterStatusConditionDetails
HTTP statusPASSdownload_http_code = 200HTTP 200
Age cutoffPASSdownload_stamp > now() - 6 MONTH0.1 months ago
History dropPASSisNull(history_drop_reason)No drop reason
Spam/banPASSfh_dont_index != 1 AND ml_spam_score = 0ml_spam_score=0
CanonicalPASSmeta_canonical IS NULL OR = '' OR = src_unparsedNot set

Page Details

PropertyValue
URLhttps://builtin.com/data-science/ols-regression
Last Crawled2026-04-11 03:16:09 (1 day ago)
First Indexed2023-02-14 19:21:42 (3 years ago)
HTTP Status Code200
Meta TitleUnderstanding Ordinary Least Squares (OLS) Regression | Built In
Meta DescriptionOrdinary least squares (OLS) regression is an optimization technique applied to linear regression models to minimize the sum of squared differences between observed and predicted values. It obtains a straight line as close as possible to data points.
Meta Canonicalnull
Boilerpipe Text
Ordinary least squares (OLS) regression is an optimization strategy that helps you find a straight line as close as possible to your data points in a linear regression model (known as a best-fit line). It works to minimize the sum of squared differences between the observed and predicted values in the model, with the best-fit line representing this minimization. OLS is considered one of the most useful optimization strategies for linear regression models as it can help you find unbiased real value estimates for your alpha (α) and beta (β) parameters. What Is Ordinary Least Squares (OLS) Regression? Ordinary least squares (OLS) regression is an optimization strategy that allows you to find a straight line that’s as close as possible to your data points in a linear regression model.   Why is that? It’s helpful to first understand linear regression algorithms . An introduction to ordinary least squares (OLS) in linear regression. | Video: XLSAT How OLS Applies to Linear Regression Linear regression is a statistical method employed in supervised machine learning tasks to model the linear relationship between one dependent variable and one or more independent variables. In particular, it works to predict the value of a dependent variable based on the value of an independent variable. Since supervised machine learning tasks are normally divided into classification and regression tasks, we can allocate linear regression algorithms into the regression category. It differs from classification because of the nature of the target variable. In classification, the target is a categorical value (“yes/no,” “red/blue/green,” “spam/not spam,” etc.). Regression involves numerical, continuous values as a target. As a result, the algorithm will be asked to predict a continuous number rather than a class or category. Imagine that you want to predict the price of a house based on some relative features , the output of your model will be the price, hence, a continuous number. Regression tasks can be divided into two main groups: simple or linear regression (which uses only one independent variable to predict the target variable), and multiple regression (which uses more than one independent variable to predict the target variable).  To give you an example, let’s consider the house task above. If you want to predict a house’s price only based on its squared meters, you will fall into the first situation (one feature), but if you are going to predict the price based on its squared meters, its position and the liveability of the surrounding environment, you are going to fall into the second group for multiple features. In the first scenario, you are likely to employ a simple linear regression algorithm, which we’ll explore more later in this article. On the other hand, whenever you’re facing more than one feature to explain the target variable, you are likely to employ a multiple linear regression. Simple linear regression is a statistical model widely used in machine learning regression tasks. It’s based on the idea that the relationship between two variables can be explained by the following formula: Simple linear regression formula. | Image: Valentina Alto Where εi is the error term, and α, β are the true (but unobserved) parameters of the regression. The parameter β represents the variation of the dependent variable when the independent variable has a unitary variation. If my parameter is equal to 0.75, when my x increases by one, my dependent variable will increase by 0.75. On the other hand, the parameter α represents the value of our dependent variable when the independent one is equal to zero. Let’s visualize it graphically: Simple linear regression graph. | Image: Valentina Alto More on Machine Learning: What Is Unsupervised Learning? How to Find OLS in a Linear Regression Model The goal of simple linear regression is to find the parameters α and β for which the error term is minimized. To be more precise, the model will minimize the squared errors. Indeed, we don’t want our positive errors to be compensated for by the negative ones, since they are equally penalizing our model. Ordinary least squares equation. | Image: Valentina Alto This procedure is called ordinary least squares (OLS) error. Let’s demonstrate those optimization problems step-by-step. If we reframe our squared error sum as follows: Equation for reframing the squared error sum. | Image: Valentina Alto. We can set our optimization problem as follows: Optimization problem equation. | Image: Valentina Alto So, let’s start with β: Solving for beta in the equation. | Image: Valentina Alto Knowing that the sample covariance between two variables is given by: Equation for finding the covariance between two variables. | Image: Valentina Alto And that the sample correlation coefficient between two variables is equal to: Sample correlation coefficient equation. | Image: Valentina Alto We can reframe the above expression as follows: Reframing the expression. | Image: Valentina Alto The same reasoning holds for our α: Equation to solve for alpha. | Image: Valentina Alto Once obtained, those values of α and β, which minimize the squared errors, our model’s equation will look like this: Linear regression model equation. | Image: Valentina Alto More on Machine Learning: Multiclass Classification With an Imbalanced Data Set Advantages of OLS Regression Think of OLS as an optimization strategy to obtain a straight line from your model that is as close as possible to your data points.  Even though OLS is not the only linear regression optimization strategy, it’s  commonly used for these tasks since the outputs of the regression (coefficients) are unbiased estimators of the real values of alpha and beta. OLS regression’s simplicity and computational efficiency as a linear regression optimization technique also lends to its popularity. Indeed, according to the Gauss-Markov Theorem, under some assumptions of the linear regression model (linearity in parameters, random sampling of observations, conditional mean equal to zero, absence of multicollinearity and homoscedasticity of errors), the OLS estimators α and β are the best linear unbiased estimators (BLUE) of the real values of α and β​. Limitations of OLS Regression Despite its popular applications in linear regression models, OLS regression can have some drawbacks. For one, OLS regression is sensitive to outliers in data, such as extremely large or small values for the dependent variable in comparison to the rest of the data set. Since OLS focuses on minimizing the sum of squared errors, outliers can disproportionately affect model results. OLS also assumes linearity in data and attempts to fit data to a straight line, though this may not always reflect the complexities of relationships between values in real life. For example, OLS can attempt to apply a best-fit line to curved or non-linear data points, leading to inaccurate model results. Additionally, the OLS algorithm can become less effective as more features (independent variables) are added to the model. When a large number of independent variables are added — where the number of features is more than the number of data points — errors are more likely to occur and the OLS solution may not be unique. Ordinary least squares (OLS) regression is an optimization strategy used in linear regression models that finds a straight line that fits as close as possible to the data points, in order to help estimate the relationship between a dependent variable and one or more independent variables. OLS determines this line by minimizing the sum of squared errors (or sum of squared differences) between the observed and predicted values in the regression model, providing real value estimates for the regression parameters alpha (α) and beta (β). Linear regression is a statistical method used to estimate the linear relationship between a dependent variable and one or more independent variables. Ordinary least squares (OLS) regression is an optimization technique applied to linear regression models to minimize the sum of squared differences between observed and predicted values. Ordinary least squares (OLS) regression is an optimization technique used for linear regression models to minimize the sum of squared differences between observed and predicted values. Multiple regression (or multiple linear regression) is a type of linear regression used to estimate the linear relationship between a dependent variable and two or more independent variables. Ordinary less squares (OLS) regression minimizes the sum of squared errors between observed and predicted values, while least absolute deviations (LAD) regression minimizes the sum of absolute errors between observed and predicted values. LAD gives equal weight to all observed data values and is based on the median of the data, making it less sensitive to outliers in data than OLS, which is based on the mean of the data.
Markdown
[![Built In Logo](https://static.builtin.com/dist/images/bi-header-logo.svg) ![Built In Logo](https://static.builtin.com/dist/images/bi-header-logo.svg)](https://builtin.com/) [In Jobs](https://builtin.com/jobs?search=) [View All Jobs](https://builtin.com/jobs) [For Employers](https://employers.builtin.com/?utm_medium=BIReferral&utm_source=foremployers) [Join](https://builtin.com/auth/signup?destination=%2Fdata-science%2Fols-regression) [Log In](https://builtin.com/auth/login?destination=%2Fdata-science%2Fols-regression) - [Jobs](https://builtin.com/jobs) - [Companies](https://builtin.com/companies) - [Remote](https://builtin.com/jobs/remote) - [Articles](https://builtin.com/tech-topics) - [Best Places To Work](https://builtin.com/awards/us/2026/best-places-to-work) - [Job Application Tracker](https://builtin.com/auth/login?destination=%2Fhome%23application-tracker-section) - [Artificial Intelligence](https://builtin.com/tag/artificial-intelligence "Artificial Intelligence") - [Data Science](https://builtin.com/tag/data-science "Data Science") - [Expert Contributors](https://builtin.com/tag/expert-contributors "Expert Contributors") - \+1 - [Machine Learning](https://builtin.com/tag/machine-learning "Machine Learning") - \+3 # Understanding Ordinary Least Squares (OLS) Regression Ordinary Least Squares (OLS) regression is a technique used in linear regression to minimize the sum of squared differences between observed and predicted values, and obtain a straight line as close as possible to your data points. ![Valentina Alto](https://cdn.builtin.com/cdn-cgi/image/f=auto,w=96,h=96,q=100/sites/www.builtin.com/files/2023-02/Valentina%20Alto.jpeg) Written by [Valentina Alto](https://builtin.com/authors/valentina-alto) ![Blocks with the letters OSL on them and a small graph.](https://cdn.builtin.com/cdn-cgi/image/f=auto,fit=cover,w=320,h=200,q=80/sites/www.builtin.com/files/2023-02/ols-regression_0.jpg) Image: Shutterstock / Built In ![Brand Studio Logo](https://static.builtin.com/dist/images/expert-badge.svg) UPDATED BY [Brennan Whitfield](https://builtin.com/authors/brennan-whitfield) \| Mar 12, 2025 Ordinary least squares (OLS) [regression](https://builtin.com/data-science/regression-machine-learning) is an optimization strategy that helps you find a straight line as close as possible to your data points in a [linear regression](https://builtin.com/data-science/linear-regression) model (known as a best-fit line). It works to minimize the sum of squared differences between the observed and predicted values in the model, with the best-fit line representing this minimization. OLS is considered one of the most useful optimization strategies for linear regression models as it can help you find [unbiased](https://builtin.com/data-science/auditing-algorithms-data-science-bias) real value estimates for your alpha (α) and beta (β) parameters. ## What Is Ordinary Least Squares (OLS) Regression? Ordinary least squares (OLS) regression is an optimization strategy that allows you to find a straight line that’s as close as possible to your data points in a linear regression model. Why is that? It’s helpful to first understand linear regression [algorithms](https://builtin.com/data-science/tour-top-10-algorithms-machine-learning-newbies). An introduction to ordinary least squares (OLS) in linear regression. \| Video: XLSAT ## How OLS Applies to Linear Regression Linear regression is a statistical method employed in [supervised machine learning](https://builtin.com/machine-learning/supervised-learning) tasks to model the linear relationship between one dependent variable and one or more independent variables. In particular, it works to predict the value of a dependent variable based on the value of an independent variable. Since [supervised machine learning](https://builtin.com/data-science/supervised-machine-learning-classification) tasks are normally divided into [classification](https://builtin.com/machine-learning/classification-machine-learning) and [regression](https://builtin.com/data-science/multiple-regression) tasks, we can allocate linear regression algorithms into the regression category. It differs from classification because of the nature of the target variable. In classification, the target is a categorical value (“yes/no,” “red/blue/green,” “spam/not spam,” etc.). Regression involves [numerical, continuous values](https://builtin.com/data-science/data-types-statistics) as a target. As a result, the algorithm will be asked to predict a continuous number rather than a class or category. Imagine that you want to predict the price of a house based on some relative [features](https://builtin.com/data-science/feature-importance), the output of your model will be the price, hence, a continuous number. Regression tasks can be divided into two main groups: simple or linear regression (which uses only one independent variable to predict the target variable), and multiple regression (which uses more than one independent variable to predict the target variable). To give you an example, let’s consider the house task above. If you want to predict a house’s price only based on its squared meters, you will fall into the first situation (one feature), but if you are going to predict the price based on its squared meters, its position and the liveability of the surrounding environment, you are going to fall into the second group for multiple features. In the first scenario, you are likely to employ a simple linear regression algorithm, which we’ll explore more later in this article. On the other hand, whenever you’re facing more than one feature to explain the target variable, you are likely to employ a multiple linear regression. Simple linear regression is a [statistical model](https://builtin.com/data-science/statistical-analysis) widely used in [machine learning](https://builtin.com/machine-learning) regression tasks. It’s based on the idea that the relationship between two variables can be explained by the following formula: ![Simple linear regression formula.](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/1_ols-regression.jpeg) Simple linear regression formula. \| Image: Valentina Alto Where εi is the error term, and α, β are the true (but unobserved) parameters of the regression. The parameter β represents the variation of the dependent variable when the independent variable has a unitary variation. If my parameter is equal to 0.75, when my *x* increases by one, my dependent variable will increase by 0.75. On the other hand, the parameter α represents the value of our dependent variable when the independent one is equal to zero. Let’s visualize it graphically: ![Simple linear regression graph.](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/2_ols-regression.jpeg) Simple linear regression graph. \| Image: Valentina Alto More on Machine Learning: [What Is Unsupervised Learning?](https://builtin.com/machine-learning/unsupervised-learning) ## How to Find OLS in a Linear Regression Model The goal of simple linear regression is to find the parameters α and β for which the error term is minimized. To be more precise, the model will minimize the squared errors. Indeed, we don’t want our positive errors to be compensated for by the negative ones, since they are equally penalizing our model. ![Ordinary least squares equation. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/3_ols-regression.jpeg) Ordinary least squares equation. \| Image: Valentina Alto This procedure is called ordinary least squares (OLS) error. Let’s demonstrate those optimization problems step-by-step. If we reframe our squared error sum as follows: ![Equation for reframing the squared error sum. \| Image: Valentina Alto.](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/4_ols-regression.jpeg) Equation for reframing the squared error sum. \| Image: Valentina Alto. We can set our optimization problem as follows: ![Optimization problem equation. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/5_ols-regression.jpeg) Optimization problem equation. \| Image: Valentina Alto So, let’s start with β: ![Solving for beta in the equation. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/6_ols-regression.jpeg) Solving for beta in the equation. \| Image: Valentina Alto Knowing that the sample [covariance](https://builtin.com/data-science/covariance-vs-correlation) between two variables is given by: ![Equation for finding the covariance between two variables. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/7_ols-regression.jpeg) Equation for finding the covariance between two variables. \| Image: Valentina Alto And that the sample correlation coefficient between two variables is equal to: ![Sample correlation coefficient equation. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/8_ols-regression.jpeg) Sample correlation coefficient equation. \| Image: Valentina Alto We can reframe the above expression as follows: ![Reframing the expression. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/9_ols-regression.jpeg) Reframing the expression. \| Image: Valentina Alto The same reasoning holds for our α: ![Equation to solve for alpha. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/10_ols-regression.jpeg) Equation to solve for alpha. \| Image: Valentina Alto Once obtained, those values of α and β, which minimize the squared errors, our model’s equation will look like this: ![Linear regression model equation. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/11_ols-regression.jpeg) Linear regression model equation. \| Image: Valentina Alto More on Machine Learning: [Multiclass Classification With an Imbalanced Data Set](https://builtin.com/machine-learning/multiclass-classification) ## Advantages of OLS Regression Think of OLS as an optimization strategy to obtain a straight line from your model that is as close as possible to your data points. Even though OLS is not the only linear regression optimization strategy, it’s commonly used for these tasks since the outputs of the regression (coefficients) are unbiased estimators of the real values of alpha and beta. OLS regression’s simplicity and computational efficiency as a linear regression optimization technique also lends to its popularity. Indeed, according to the Gauss-Markov Theorem, under some assumptions of the linear regression model (linearity in parameters, [random sampling](https://builtin.com/data-science/types-of-random-sampling) of observations, conditional [mean](https://builtin.com/data-science/intro-descriptive-statistics) equal to zero, absence of multicollinearity and homoscedasticity of errors), the OLS estimators α and β are the best linear unbiased estimators (BLUE) of the real values of α and β​. ## Limitations of OLS Regression Despite its popular applications in linear regression models, OLS regression can have some drawbacks. For one, OLS regression is sensitive to outliers in data, such as extremely large or small values for the dependent variable in comparison to the rest of the data set. Since OLS focuses on minimizing the sum of squared errors, outliers can disproportionately affect model results. OLS also assumes linearity in data and attempts to fit data to a straight line, though this may not always reflect the complexities of relationships between values in real life. For example, OLS can attempt to apply a best-fit line to curved or non-linear data points, leading to inaccurate model results. Additionally, the OLS algorithm can become less effective as more features (independent variables) are added to the model. When a large number of independent variables are added — where the number of features is more than the number of data points — errors are more likely to occur and the OLS solution may not be unique. ## Frequently Asked Questions ### What is OLS regression used for? Ordinary least squares (OLS) regression is an optimization strategy used in linear regression models that finds a straight line that fits as close as possible to the data points, in order to help estimate the relationship between a dependent variable and one or more independent variables. OLS determines this line by minimizing the sum of squared errors (or sum of squared differences) between the observed and predicted values in the regression model, providing real value estimates for the regression parameters alpha (α) and beta (β). ### What is the difference between linear regression and OLS regression? Linear regression is a statistical method used to estimate the linear relationship between a dependent variable and one or more independent variables. Ordinary least squares (OLS) regression is an optimization technique applied to linear regression models to minimize the sum of squared differences between observed and predicted values. ### What is the difference between OLS and multiple regression? Ordinary least squares (OLS) regression is an optimization technique used for linear regression models to minimize the sum of squared differences between observed and predicted values. Multiple regression (or multiple linear regression) is a type of linear regression used to estimate the linear relationship between a dependent variable and two or more independent variables. ### What is the difference between OLS and LAD regression? Ordinary less squares (OLS) regression minimizes the sum of squared errors between observed and predicted values, while least absolute deviations (LAD) regression minimizes the sum of absolute errors between observed and predicted values. LAD gives equal weight to all observed data values and is based on the median of the data, making it less sensitive to outliers in data than OLS, which is based on the mean of the data. ### Recent Artificial Intelligence Articles [![What Is Master Data Management (MDM)? (Definition, Tools)](https://cdn.builtin.com/cdn-cgi/image/f=auto,fit=contain,w=120,h=70,q=80/sites/www.builtin.com/files/2026-03/master-data-management.png) What Is Master Data Management (MDM)? (Definition, Tools)](https://builtin.com/articles/master-data-management) [![AI in Real Estate: 20 Companies Defining the Industry](https://cdn.builtin.com/cdn-cgi/image/f=auto,fit=contain,w=120,h=70,q=80/sites/www.builtin.com/files/header-ai-real-estate.jpg) AI in Real Estate: 20 Companies Defining the Industry](https://builtin.com/artificial-intelligence/ai-real-estate) [![Don’t Mistake AI Hype for a Bubble](https://cdn.builtin.com/cdn-cgi/image/f=auto,fit=contain,w=120,h=70,q=80/sites/www.builtin.com/files/2026-04/ai-hype-isnt-bubble.png) Don’t Mistake AI Hype for a Bubble](https://builtin.com/articles/ai-hype-isnt-bubble) Explore Job Matches. Job Title or Keyword Clear search Location Fully Remote, Hybrid, On Site Fully Remote Hybrid On Site Clear Apply See Jobs - [Jobs](https://builtin.com/jobs) - [Companies](https://builtin.com/companies) - [Articles](https://builtin.com/tech-topics) - [Tracker](https://builtin.com/auth/login?destination=%2Fhome%23application-tracker-section) - More ![Built In](https://static.builtin.com/dist/images/midnight_9.svg) [Join](https://builtin.com/auth/signup?destination=%2Fdata-science%2Fols-regression) [Log In](https://builtin.com/auth/login?destination=%2Fdata-science%2Fols-regression) - [Tech Jobs](https://builtin.com/jobs) - [Companies](https://builtin.com/companies) - [Articles](https://builtin.com/tech-topics) - [Remote](https://builtin.com/jobs/remote) - [Best Places To Work](https://builtin.com/awards/us/2026/best-places-to-work) - [Tech Hubs](https://builtin.com/tech-hubs) [Post Job](https://employers.builtin.com/membership?utm_medium=BIReferral&utm_source=foremployers) [![BuiltIn](https://static.builtin.com/dist/images/builtin-logo.svg)](https://builtin.com/) ![United We Tech](https://static.builtin.com/dist/images/united-we-tech.svg) Built In is the online community for startups and tech companies. Find startup jobs, tech news and events. About [Our Story](https://builtin.com/our-story) [Careers](https://employers.builtin.com/careers/) [Our Staff Writers](https://builtin.com/our-staff) [Content Descriptions](https://builtin.com/content-descriptions) *** Get Involved [Recruit With Built In](https://employers.builtin.com/membership?utm_medium=BIReferral&utm_source=foremployers) [Become an Expert Contributor](https://builtin.com/expert-contributors) *** Resources [Customer Support](https://knowledgebase.builtin.com/s/) [Share Feedback](https://form.jotform.com/223044927257054) [Report a Bug](https://knowledgebase.builtin.com/s/contactsupport) [Tech Job Tools + Career Resources](https://builtin.com/articles/grow-your-career) [Browse Jobs](https://builtin.com/browse-jobs) [Tech A-Z](https://builtin.com/tech-dictionary) *** Tech Hubs [Our Sites](https://builtin.com/our-sites) *** [Learning Lab User Agreement](https://builtin.com/learning-lab-user-agreement) [Accessibility Statement](https://builtin.com/accessibility-statement) [Copyright Policy](https://builtin.com/copyright-policy) [Privacy Policy](https://builtin.com/privacy-policy) [Terms of Use](https://builtin.com/community-terms-of-use) [Your Privacy Choices/Cookie Settings](https://builtin.com/california-do-not-sell-my-information) [CA Notice of Collection](https://builtin.com/ca-notice-collection) © Built In 2026
Readable Markdown
Ordinary least squares (OLS) [regression](https://builtin.com/data-science/regression-machine-learning) is an optimization strategy that helps you find a straight line as close as possible to your data points in a [linear regression](https://builtin.com/data-science/linear-regression) model (known as a best-fit line). It works to minimize the sum of squared differences between the observed and predicted values in the model, with the best-fit line representing this minimization. OLS is considered one of the most useful optimization strategies for linear regression models as it can help you find [unbiased](https://builtin.com/data-science/auditing-algorithms-data-science-bias) real value estimates for your alpha (α) and beta (β) parameters. ## What Is Ordinary Least Squares (OLS) Regression? Ordinary least squares (OLS) regression is an optimization strategy that allows you to find a straight line that’s as close as possible to your data points in a linear regression model. Why is that? It’s helpful to first understand linear regression [algorithms](https://builtin.com/data-science/tour-top-10-algorithms-machine-learning-newbies). An introduction to ordinary least squares (OLS) in linear regression. \| Video: XLSAT ## How OLS Applies to Linear Regression Linear regression is a statistical method employed in [supervised machine learning](https://builtin.com/machine-learning/supervised-learning) tasks to model the linear relationship between one dependent variable and one or more independent variables. In particular, it works to predict the value of a dependent variable based on the value of an independent variable. Since [supervised machine learning](https://builtin.com/data-science/supervised-machine-learning-classification) tasks are normally divided into [classification](https://builtin.com/machine-learning/classification-machine-learning) and [regression](https://builtin.com/data-science/multiple-regression) tasks, we can allocate linear regression algorithms into the regression category. It differs from classification because of the nature of the target variable. In classification, the target is a categorical value (“yes/no,” “red/blue/green,” “spam/not spam,” etc.). Regression involves [numerical, continuous values](https://builtin.com/data-science/data-types-statistics) as a target. As a result, the algorithm will be asked to predict a continuous number rather than a class or category. Imagine that you want to predict the price of a house based on some relative [features](https://builtin.com/data-science/feature-importance), the output of your model will be the price, hence, a continuous number. Regression tasks can be divided into two main groups: simple or linear regression (which uses only one independent variable to predict the target variable), and multiple regression (which uses more than one independent variable to predict the target variable). To give you an example, let’s consider the house task above. If you want to predict a house’s price only based on its squared meters, you will fall into the first situation (one feature), but if you are going to predict the price based on its squared meters, its position and the liveability of the surrounding environment, you are going to fall into the second group for multiple features. In the first scenario, you are likely to employ a simple linear regression algorithm, which we’ll explore more later in this article. On the other hand, whenever you’re facing more than one feature to explain the target variable, you are likely to employ a multiple linear regression. Simple linear regression is a [statistical model](https://builtin.com/data-science/statistical-analysis) widely used in [machine learning](https://builtin.com/machine-learning) regression tasks. It’s based on the idea that the relationship between two variables can be explained by the following formula: ![Simple linear regression formula.](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/1_ols-regression.jpeg) Simple linear regression formula. \| Image: Valentina Alto Where εi is the error term, and α, β are the true (but unobserved) parameters of the regression. The parameter β represents the variation of the dependent variable when the independent variable has a unitary variation. If my parameter is equal to 0.75, when my *x* increases by one, my dependent variable will increase by 0.75. On the other hand, the parameter α represents the value of our dependent variable when the independent one is equal to zero. Let’s visualize it graphically: ![Simple linear regression graph.](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/2_ols-regression.jpeg) Simple linear regression graph. \| Image: Valentina Alto More on Machine Learning: [What Is Unsupervised Learning?](https://builtin.com/machine-learning/unsupervised-learning) ## How to Find OLS in a Linear Regression Model The goal of simple linear regression is to find the parameters α and β for which the error term is minimized. To be more precise, the model will minimize the squared errors. Indeed, we don’t want our positive errors to be compensated for by the negative ones, since they are equally penalizing our model. ![Ordinary least squares equation. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/3_ols-regression.jpeg) Ordinary least squares equation. \| Image: Valentina Alto This procedure is called ordinary least squares (OLS) error. Let’s demonstrate those optimization problems step-by-step. If we reframe our squared error sum as follows: ![Equation for reframing the squared error sum. \| Image: Valentina Alto.](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/4_ols-regression.jpeg) Equation for reframing the squared error sum. \| Image: Valentina Alto. We can set our optimization problem as follows: ![Optimization problem equation. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/5_ols-regression.jpeg) Optimization problem equation. \| Image: Valentina Alto So, let’s start with β: ![Solving for beta in the equation. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/6_ols-regression.jpeg) Solving for beta in the equation. \| Image: Valentina Alto Knowing that the sample [covariance](https://builtin.com/data-science/covariance-vs-correlation) between two variables is given by: ![Equation for finding the covariance between two variables. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/7_ols-regression.jpeg) Equation for finding the covariance between two variables. \| Image: Valentina Alto And that the sample correlation coefficient between two variables is equal to: ![Sample correlation coefficient equation. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/8_ols-regression.jpeg) Sample correlation coefficient equation. \| Image: Valentina Alto We can reframe the above expression as follows: ![Reframing the expression. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/9_ols-regression.jpeg) Reframing the expression. \| Image: Valentina Alto The same reasoning holds for our α: ![Equation to solve for alpha. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/10_ols-regression.jpeg) Equation to solve for alpha. \| Image: Valentina Alto Once obtained, those values of α and β, which minimize the squared errors, our model’s equation will look like this: ![Linear regression model equation. ](https://builtin.com/sites/www.builtin.com/files/styles/ckeditor_optimize/public/inline-images/11_ols-regression.jpeg) Linear regression model equation. \| Image: Valentina Alto More on Machine Learning: [Multiclass Classification With an Imbalanced Data Set](https://builtin.com/machine-learning/multiclass-classification) ## Advantages of OLS Regression Think of OLS as an optimization strategy to obtain a straight line from your model that is as close as possible to your data points. Even though OLS is not the only linear regression optimization strategy, it’s commonly used for these tasks since the outputs of the regression (coefficients) are unbiased estimators of the real values of alpha and beta. OLS regression’s simplicity and computational efficiency as a linear regression optimization technique also lends to its popularity. Indeed, according to the Gauss-Markov Theorem, under some assumptions of the linear regression model (linearity in parameters, [random sampling](https://builtin.com/data-science/types-of-random-sampling) of observations, conditional [mean](https://builtin.com/data-science/intro-descriptive-statistics) equal to zero, absence of multicollinearity and homoscedasticity of errors), the OLS estimators α and β are the best linear unbiased estimators (BLUE) of the real values of α and β​. ## Limitations of OLS Regression Despite its popular applications in linear regression models, OLS regression can have some drawbacks. For one, OLS regression is sensitive to outliers in data, such as extremely large or small values for the dependent variable in comparison to the rest of the data set. Since OLS focuses on minimizing the sum of squared errors, outliers can disproportionately affect model results. OLS also assumes linearity in data and attempts to fit data to a straight line, though this may not always reflect the complexities of relationships between values in real life. For example, OLS can attempt to apply a best-fit line to curved or non-linear data points, leading to inaccurate model results. Additionally, the OLS algorithm can become less effective as more features (independent variables) are added to the model. When a large number of independent variables are added — where the number of features is more than the number of data points — errors are more likely to occur and the OLS solution may not be unique. Ordinary least squares (OLS) regression is an optimization strategy used in linear regression models that finds a straight line that fits as close as possible to the data points, in order to help estimate the relationship between a dependent variable and one or more independent variables. OLS determines this line by minimizing the sum of squared errors (or sum of squared differences) between the observed and predicted values in the regression model, providing real value estimates for the regression parameters alpha (α) and beta (β). Linear regression is a statistical method used to estimate the linear relationship between a dependent variable and one or more independent variables. Ordinary least squares (OLS) regression is an optimization technique applied to linear regression models to minimize the sum of squared differences between observed and predicted values. Ordinary least squares (OLS) regression is an optimization technique used for linear regression models to minimize the sum of squared differences between observed and predicted values. Multiple regression (or multiple linear regression) is a type of linear regression used to estimate the linear relationship between a dependent variable and two or more independent variables. Ordinary less squares (OLS) regression minimizes the sum of squared errors between observed and predicted values, while least absolute deviations (LAD) regression minimizes the sum of absolute errors between observed and predicted values. LAD gives equal weight to all observed data values and is based on the median of the data, making it less sensitive to outliers in data than OLS, which is based on the mean of the data.
Shard169 (laksa)
Root Hash7607033694470393769
Unparsed URLcom,builtin!/data-science/ols-regression s443