Course Summary
Learn how to use regression techniques to analyze and predict outcomes in this machine learning course.Key Learning Points
- Understand the fundamentals of regression analysis
- Learn how to implement regression models in Python
- Discover how to evaluate and fine-tune regression models
Related Topics for further study
Learning Outcomes
- Ability to implement regression models in Python
- Understanding of model evaluation techniques
- Knowledge of how to fine-tune regression models
Prerequisites or good to have knowledge before taking this course
- Basic knowledge of Python programming
- Familiarity with basic statistics
Course Difficulty Level
IntermediateCourse Format
- Self-paced
- Online
- Video Lectures
- Programming Assignments
Similar Courses
- Applied Data Science: Machine Learning
- Applied Machine Learning
Related Education Paths
Related Books
Description
Case Study - Predicting Housing Prices
Outline
- Welcome
- Welcome!
- What is the course about?
- Outlining the first half of the course
- Outlining the second half of the course
- Assumed background
- Important Update regarding the Machine Learning Specialization
- Slides presented in this module
- Reading: Software tools you'll need
- Simple Linear Regression
- A case study in predicting house prices
- Regression fundamentals: data & model
- Regression fundamentals: the task
- Regression ML block diagram
- The simple linear regression model
- The cost of using a given line
- Using the fitted line
- Interpreting the fitted line
- Defining our least squares optimization objective
- Finding maxima or minima analytically
- Maximizing a 1d function: a worked example
- Finding the max via hill climbing
- Finding the min via hill descent
- Choosing stepsize and convergence criteria
- Gradients: derivatives in multiple dimensions
- Gradient descent: multidimensional hill descent
- Computing the gradient of RSS
- Approach 1: closed-form solution
- Approach 2: gradient descent
- Comparing the approaches
- Influence of high leverage points: exploring the data
- Influence of high leverage points: removing Center City
- Influence of high leverage points: removing high-end towns
- Asymmetric cost functions
- A brief recap
- Slides presented in this module
- Optional reading: worked-out example for closed-form solution
- Optional reading: worked-out example for gradient descent
- Download notebooks to follow along
- Fitting a simple linear regression model on housing data
- Simple Linear Regression
- Fitting a simple linear regression model on housing data
- Multiple Regression
- Multiple regression intro
- Polynomial regression
- Modeling seasonality
- Where we see seasonality
- Regression with general features of 1 input
- Motivating the use of multiple inputs
- Defining notation
- Regression with features of multiple inputs
- Interpreting the multiple regression fit
- Rewriting the single observation model in vector notation
- Rewriting the model for all observations in matrix notation
- Computing the cost of a D-dimensional curve
- Computing the gradient of RSS
- Approach 1: closed-form solution
- Discussing the closed-form solution
- Approach 2: gradient descent
- Feature-by-feature update
- Algorithmic summary of gradient descent approach
- A brief recap
- Slides presented in this module
- Optional reading: review of matrix algebra
- Exploring different multiple regression models for house price prediction
- Numpy tutorial
- Implementing gradient descent for multiple regression
- Multiple Regression
- Exploring different multiple regression models for house price prediction
- Implementing gradient descent for multiple regression
- Assessing Performance
- Assessing performance intro
- What do we mean by "loss"?
- Training error: assessing loss on the training set
- Generalization error: what we really want
- Test error: what we can actually compute
- Defining overfitting
- Training/test split
- Irreducible error and bias
- Variance and the bias-variance tradeoff
- Error vs. amount of data
- Formally defining the 3 sources of error
- Formally deriving why 3 sources of error
- Training/validation/test split for model selection, fitting, and assessment
- A brief recap
- Slides presented in this module
- Polynomial Regression
- Assessing Performance
- Exploring the bias-variance tradeoff
- Ridge Regression
- Symptoms of overfitting in polynomial regression
- Overfitting demo
- Overfitting for more general multiple regression models
- Balancing fit and magnitude of coefficients
- The resulting ridge objective and its extreme solutions
- How ridge regression balances bias and variance
- Ridge regression demo
- The ridge coefficient path
- Computing the gradient of the ridge objective
- Approach 1: closed-form solution
- Discussing the closed-form solution
- Approach 2: gradient descent
- Selecting tuning parameters via cross validation
- K-fold cross validation
- How to handle the intercept
- A brief recap
- Slides presented in this module
- Download the notebook and follow along
- Download the notebook and follow along
- Observing effects of L2 penalty in polynomial regression
- Implementing ridge regression via gradient descent
- Ridge Regression
- Observing effects of L2 penalty in polynomial regression
- Implementing ridge regression via gradient descent
- Feature Selection & Lasso
- The feature selection task
- All subsets
- Complexity of all subsets
- Greedy algorithms
- Complexity of the greedy forward stepwise algorithm
- Can we use regularization for feature selection?
- Thresholding ridge coefficients?
- The lasso objective and its coefficient path
- Visualizing the ridge cost
- Visualizing the ridge solution
- Visualizing the lasso cost and solution
- Lasso demo
- What makes the lasso objective different
- Coordinate descent
- Normalizing features
- Coordinate descent for least squares regression (normalized features)
- Coordinate descent for lasso (normalized features)
- Assessing convergence and other lasso solvers
- Coordinate descent for lasso (unnormalized features)
- Deriving the lasso coordinate descent update
- Choosing the penalty strength and other practical issues with lasso
- A brief recap
- Slides presented in this module
- Download the notebook and follow along
- Using LASSO to select features
- Implementing LASSO using coordinate descent
- Feature Selection and Lasso
- Using LASSO to select features
- Implementing LASSO using coordinate descent
- Nearest Neighbors & Kernel Regression
- Limitations of parametric regression
- 1-Nearest neighbor regression approach
- Distance metrics
- 1-Nearest neighbor algorithm
- k-Nearest neighbors regression
- k-Nearest neighbors in practice
- Weighted k-nearest neighbors
- From weighted k-NN to kernel regression
- Global fits of parametric models vs. local fits of kernel regression
- Performance of NN as amount of data grows
- Issues with high-dimensions, data scarcity, and computational complexity
- k-NN for classification
- A brief recap
- Slides presented in this module
- Predicting house prices using k-nearest neighbors regression
- Nearest Neighbors & Kernel Regression
- Predicting house prices using k-nearest neighbors regression
- Closing Remarks
- Simple and multiple regression
- Assessing performance and ridge regression
- Feature selection, lasso, and nearest neighbor regression
- What we covered and what we didn't cover
- Thank you!
- Slides presented in this module
Summary of User Reviews
This course on Machine Learning Regression is highly rated and recommended by many users. It is designed to provide an in-depth understanding of regression models and their applications in the real world.Key Aspect Users Liked About This Course
The course is well-structured and provides clear explanations of complex concepts.Pros from User Reviews
- The course content is comprehensive and covers a wide range of topics.
- The instructors are knowledgeable and provide clear explanations.
- The assignments are challenging but rewarding.
- The course includes hands-on projects that allow users to apply the concepts learned.
- The course is suitable for both beginners and advanced learners.
Cons from User Reviews
- Some users found the pace of the course to be too fast.
- Some users felt that the course could benefit from more interactive elements.
- Some users found the course to be too theoretical and lacking in practical applications.