Machine Learning: Regression

  • 4.8
Approx. 22 hours to complete

Course Summary

Learn how to use regression techniques to analyze and predict outcomes in this machine learning course.

Key Learning Points

  • Understand the fundamentals of regression analysis
  • Learn how to implement regression models in Python
  • Discover how to evaluate and fine-tune regression models

Related Topics for further study


Learning Outcomes

  • Ability to implement regression models in Python
  • Understanding of model evaluation techniques
  • Knowledge of how to fine-tune regression models

Prerequisites or good to have knowledge before taking this course

  • Basic knowledge of Python programming
  • Familiarity with basic statistics

Course Difficulty Level

Intermediate

Course Format

  • Self-paced
  • Online
  • Video Lectures
  • Programming Assignments

Similar Courses

  • Applied Data Science: Machine Learning
  • Applied Machine Learning

Related Education Paths


Related Books

Description

Case Study - Predicting Housing Prices

Outline

  • Welcome
  • Welcome!
  • What is the course about?
  • Outlining the first half of the course
  • Outlining the second half of the course
  • Assumed background
  • Important Update regarding the Machine Learning Specialization
  • Slides presented in this module
  • Reading: Software tools you'll need
  • Simple Linear Regression
  • A case study in predicting house prices
  • Regression fundamentals: data & model
  • Regression fundamentals: the task
  • Regression ML block diagram
  • The simple linear regression model
  • The cost of using a given line
  • Using the fitted line
  • Interpreting the fitted line
  • Defining our least squares optimization objective
  • Finding maxima or minima analytically
  • Maximizing a 1d function: a worked example
  • Finding the max via hill climbing
  • Finding the min via hill descent
  • Choosing stepsize and convergence criteria
  • Gradients: derivatives in multiple dimensions
  • Gradient descent: multidimensional hill descent
  • Computing the gradient of RSS
  • Approach 1: closed-form solution
  • Approach 2: gradient descent
  • Comparing the approaches
  • Influence of high leverage points: exploring the data
  • Influence of high leverage points: removing Center City
  • Influence of high leverage points: removing high-end towns
  • Asymmetric cost functions
  • A brief recap
  • Slides presented in this module
  • Optional reading: worked-out example for closed-form solution
  • Optional reading: worked-out example for gradient descent
  • Download notebooks to follow along
  • Fitting a simple linear regression model on housing data
  • Simple Linear Regression
  • Fitting a simple linear regression model on housing data
  • Multiple Regression
  • Multiple regression intro
  • Polynomial regression
  • Modeling seasonality
  • Where we see seasonality
  • Regression with general features of 1 input
  • Motivating the use of multiple inputs
  • Defining notation
  • Regression with features of multiple inputs
  • Interpreting the multiple regression fit
  • Rewriting the single observation model in vector notation
  • Rewriting the model for all observations in matrix notation
  • Computing the cost of a D-dimensional curve
  • Computing the gradient of RSS
  • Approach 1: closed-form solution
  • Discussing the closed-form solution
  • Approach 2: gradient descent
  • Feature-by-feature update
  • Algorithmic summary of gradient descent approach
  • A brief recap
  • Slides presented in this module
  • Optional reading: review of matrix algebra
  • Exploring different multiple regression models for house price prediction
  • Numpy tutorial
  • Implementing gradient descent for multiple regression
  • Multiple Regression
  • Exploring different multiple regression models for house price prediction
  • Implementing gradient descent for multiple regression
  • Assessing Performance
  • Assessing performance intro
  • What do we mean by "loss"?
  • Training error: assessing loss on the training set
  • Generalization error: what we really want
  • Test error: what we can actually compute
  • Defining overfitting
  • Training/test split
  • Irreducible error and bias
  • Variance and the bias-variance tradeoff
  • Error vs. amount of data
  • Formally defining the 3 sources of error
  • Formally deriving why 3 sources of error
  • Training/validation/test split for model selection, fitting, and assessment
  • A brief recap
  • Slides presented in this module
  • Polynomial Regression
  • Assessing Performance
  • Exploring the bias-variance tradeoff
  • Ridge Regression
  • Symptoms of overfitting in polynomial regression
  • Overfitting demo
  • Overfitting for more general multiple regression models
  • Balancing fit and magnitude of coefficients
  • The resulting ridge objective and its extreme solutions
  • How ridge regression balances bias and variance
  • Ridge regression demo
  • The ridge coefficient path
  • Computing the gradient of the ridge objective
  • Approach 1: closed-form solution
  • Discussing the closed-form solution
  • Approach 2: gradient descent
  • Selecting tuning parameters via cross validation
  • K-fold cross validation
  • How to handle the intercept
  • A brief recap
  • Slides presented in this module
  • Download the notebook and follow along
  • Download the notebook and follow along
  • Observing effects of L2 penalty in polynomial regression
  • Implementing ridge regression via gradient descent
  • Ridge Regression
  • Observing effects of L2 penalty in polynomial regression
  • Implementing ridge regression via gradient descent
  • Feature Selection & Lasso
  • The feature selection task
  • All subsets
  • Complexity of all subsets
  • Greedy algorithms
  • Complexity of the greedy forward stepwise algorithm
  • Can we use regularization for feature selection?
  • Thresholding ridge coefficients?
  • The lasso objective and its coefficient path
  • Visualizing the ridge cost
  • Visualizing the ridge solution
  • Visualizing the lasso cost and solution
  • Lasso demo
  • What makes the lasso objective different
  • Coordinate descent
  • Normalizing features
  • Coordinate descent for least squares regression (normalized features)
  • Coordinate descent for lasso (normalized features)
  • Assessing convergence and other lasso solvers
  • Coordinate descent for lasso (unnormalized features)
  • Deriving the lasso coordinate descent update
  • Choosing the penalty strength and other practical issues with lasso
  • A brief recap
  • Slides presented in this module
  • Download the notebook and follow along
  • Using LASSO to select features
  • Implementing LASSO using coordinate descent
  • Feature Selection and Lasso
  • Using LASSO to select features
  • Implementing LASSO using coordinate descent
  • Nearest Neighbors & Kernel Regression
  • Limitations of parametric regression
  • 1-Nearest neighbor regression approach
  • Distance metrics
  • 1-Nearest neighbor algorithm
  • k-Nearest neighbors regression
  • k-Nearest neighbors in practice
  • Weighted k-nearest neighbors
  • From weighted k-NN to kernel regression
  • Global fits of parametric models vs. local fits of kernel regression
  • Performance of NN as amount of data grows
  • Issues with high-dimensions, data scarcity, and computational complexity
  • k-NN for classification
  • A brief recap
  • Slides presented in this module
  • Predicting house prices using k-nearest neighbors regression
  • Nearest Neighbors & Kernel Regression
  • Predicting house prices using k-nearest neighbors regression
  • Closing Remarks
  • Simple and multiple regression
  • Assessing performance and ridge regression
  • Feature selection, lasso, and nearest neighbor regression
  • What we covered and what we didn't cover
  • Thank you!
  • Slides presented in this module

Summary of User Reviews

This course on Machine Learning Regression is highly rated and recommended by many users. It is designed to provide an in-depth understanding of regression models and their applications in the real world.

Key Aspect Users Liked About This Course

The course is well-structured and provides clear explanations of complex concepts.

Pros from User Reviews

  • The course content is comprehensive and covers a wide range of topics.
  • The instructors are knowledgeable and provide clear explanations.
  • The assignments are challenging but rewarding.
  • The course includes hands-on projects that allow users to apply the concepts learned.
  • The course is suitable for both beginners and advanced learners.

Cons from User Reviews

  • Some users found the pace of the course to be too fast.
  • Some users felt that the course could benefit from more interactive elements.
  • Some users found the course to be too theoretical and lacking in practical applications.
English
Available now
Approx. 22 hours to complete
Emily Fox, Carlos Guestrin
University of Washington
Coursera

Instructor

Emily Fox

  • 4.8 Raiting
Share
Saved Course list
Cancel
Get Course Update
Computer Courses