You can get video resuorces from here: https://pan.baidu.com/wap/link?surl=1o8Frmno
Implemented linear regression with one variable to predict profits for a food truck in this part of this exercise.
Implemented linear regression with multiple variables to predict profits for a food truck In this part of this exercise. The following pictuer is a curve of cost function.
This exercise will program logstic regression to build a classification model that estimates an applicant’s probability of admission based the scores from those two exams.
This exercise will implement regularized logistic regression to predict whether microchips from a fabrication plant passes quality assurance (QA).
This part of the exercise extended the previous implemention of logistic regression and applied it to one-vs-all classification on recognizition of handwritten digits.
This part of the exercise implemented a neural network to recognize handwritten digits using the same training set as before. The neural network can be able to represent complex models that form non-linear hypotheses.
The accuracy rate of neural network on training set is :
showed predictions dynamicly (animate.py):
Implemented and tested the regularized cost function and feedforward propagation.
Selected 100 pictures to predict and showed results dynamicly (animate.py):
Implemented the backpropagation algorithm to compute the gradient for the neural network cost function. And then performed gradient checking to validate backpropagation.
This the the result of gradient checking:
At last we get parameters and the accuracy rate with these parameters:
Implemented regularized linear regression to predict the amount of water flowing out of a dam using the change of water level in a reservoir.
An important concept in machine learning is the bias-variance tradeoff. Models with high bias are not complex enough for the data and tend to underfit, while models with high variance overfit to the training data. In this part of the exercise, I ploted training and test errors on a learning curve to diagnose bias-variance problems.
The problem with our linear model was that it was too simple for the data and resulted in underfitting (high bias). In this part of the exercise, I addressed this problem by adding more features using polynomial regression.
Polynomial fit, λ = 1 (It did't fit well):
Polynomial learning curve, λ = 0 (overfitting (high-variance)):
Selecting λ using a cross validation set: (A good choice of λ=1 can provide a good fit to the data.)
Implemented support vector machines (SVMs) and used it with various example 2D datasets. Experimenting with these datasets helped me gain an intuition of how SVMs work and how to use a Gaussian kernel with SVMs.
Different values of the C parameter with linear kernel SVMs.
Used Gaussian Kernel SVMs to do non-linear classification.
Used the cross validation set Xval, yval to determine the best C and σ parameter to use:
Used the best C and σ parameter to plot the boudary
Many email services today provide spam filters that are able to classify emails into spam and non-spam email with high accuracy. In this part of the exercise, I used SVMs to build my own spam filter.
The accuracy of model on training dataset and test dataset repsectively.
Implemented the K-means algorithm and use it for image compression.
Runned the K-means algorithm on a toy 2D dataset
Applied K-means to image compression
Implemented principal component analysis (PCA) and used it to perform dimensionality reduction.
Experimented with an example 2D dataset to get intuition on how PCA works:
Runned PCA on face images to see how it can be used in practice for dimension reduction.
Implemented an anomaly detection algorithm to detect anomalous behavior in server computers.
Estimated parameters for a Gaussian and selected the threshold:
The Gaussian distribution contours of the distribution fit to the dataset:
Runned the implemented anomaly detection algorithm on a more realistic dataset:
Implemented the collaborative filtering learning algorithm and applied it to a dataset of movie ratings. This dataset consists of ratings on a scale of 1 to 5. The dataset has nu = 943 users, and nm = 1682 movies.