Generalized Linear Models (GLMs): The Supercharged Linear Regression

The limitations of linear regression and how GLMs solve them.

Linear regression is pretty restricted in terms of the kind of data it can model.

For instance, its assumed data generation process looks like this:

  • Firstly, we assume that the conditional distribution of Y given X is a Gaussian.

  • Next, linear regression assumes a very specific form for the mean of the above Gaussian. It says that the mean should always be a linear combination of the features (or predictors).

  • Lastly, it assumes a constant variance for the conditional distribution P(Y|X) across all levels of X.

A graphical way of illustrating this is as follows:

This often restricts its applicability to data situations that do not obey the above assumptions.

In other words, nothing stops real-world datasets from violating these assumptions.

In fact, in many scenarios, the data might exhibit complex relationships, heteroscedasticity (varying variance), or even follow entirely different distributions altogether.

Yet, if we intend to build linear models, we should formulate better algorithms that can handle these peculiarities.

Generalized linear models (GLMs) precisely do that.

As the name suggests, they are a generalization of linear regression models.

They relax the assumptions of linear regression to make linear models more adaptable to real-world datasets.

More specifically, they consider the following:

  • What if the distribution isn’t normal but some other distribution?

  • What if X has a more sophisticated relationship with the mean?

  • What if the variance varies with X?

The effectiveness of a specific GLM — Poisson regression over linear regression is evident from the image below:

  • Linear regression assumes the data is drawn from a Gaussian, when in reality, it isn’t. Hence, it underperforms.

  • Poisson regression adapts its regression fit to a non-Gaussian distribution. Hence, it performs significantly better.

But why are GLMs so powerful?

And how exactly do they extend the typical linear regression model to handle different data situations?

What are the kinds of data they can model?

If you are curious by now, then this is precisely what we are learning in today’s machine learning deep dive:

A linear regression model is undeniably an extremely powerful model, in my opinion.

However, it does make some strict assumptions about the type of data it can model.

So, if linear regression is the only linear model you know, it is difficult to expect good modeling results because nothing stops real-world datasets from violating these assumptions.

That is why being aware of its extensions is immensely important.

To help you learn GLMs, today’s article covers:

  • How does linear regression model data?

  • The limitations of linear regression.

  • What are GLMs?

  • What are the core components of GLMs?

  • How do they relax the assumptions of linear regression?

  • What are the common types of GLMs?

  • What are the assumptions of these GLMs?

  • How do we use maximum likelihood estimates with these GLMs?

  • How to build a custom GLM for your own data?

  • Best practices and takeaways.

Hope you will learn something new today :)

👉 If you liked this post, don’t forget to leave a like ❤️. It helps more people discover this newsletter on Substack and tells me that you appreciate reading these daily insights. The button is located towards the bottom of this email.

Thanks for reading :)

Reply

or to participate.