Nptel Introduction to Machine Learning Week 2 Assignment Answer

ABOUT THE COURSE :
With the increased availability of data from varied sources there has been increasing attention paid to the various data driven disciplines such as analytics and machine learning. In this course we intend to introduce some of the basic concepts of machine learning from a mathematically well motivated perspective. We will cover the different learning paradigms and some of the more popular algorithms and architectures used in each of these paradigms.
INTENDED AUDIENCE : This is an elective course. Intended for senior UG/PG students. BE/ME/MS/PhD
PREREQUISITES : We will assume that the students know programming for some of the assignments.If the students have done introductory courses on probability theory and linear algebra it would be helpful. We will review some of the basic topics in the first two weeks as well.
INDUSTRY SUPPORT : Any company in the data analytics/data science/big data domain would value this course.

Nptel Introduction to Machine Learning Week 2 Assignment Answer

Course layout

Week 0: Probability Theory, Linear Algebra, Convex Optimization – (Recap)
Week 1: Introduction: Statistical Decision Theory – Regression, Classification, Bias Variance
Week 2: Linear Regression, Multivariate Regression, Subset Selection, Shrinkage Methods, Principal Component Regression, Partial Least squares
Week 3: Linear Classification, Logistic Regression, Linear Discriminant Analysis
Week 4: Perceptron, Support Vector Machines
Week 5: Neural Networks – Introduction, Early Models, Perceptron Learning, Backpropagation, Initialization, Training & Validation, Parameter Estimation – MLE, MAP, Bayesian Estimation
Week 6: Decision Trees, Regression Trees, Stopping Criterion & Pruning loss functions, Categorical Attributes, Multiway Splits, Missing Values, Decision Trees – Instability Evaluation Measures
Week 7: Bootstrapping & Cross Validation, Class Evaluation Measures, ROC curve, MDL, Ensemble Methods – Bagging, Committee Machines and Stacking, Boosting
Week 8: Gradient Boosting, Random Forests, Multi-class Classification, Naive Bayes, Bayesian Networks
Week 9: Undirected Graphical Models, HMM, Variable Elimination, Belief Propagation
Week 10: Partitional Clustering, Hierarchical Clustering, Birch Algorithm, CURE Algorithm, Density-based Clustering
Week 11: Gaussian Mixture Models, Expectation Maximization
Week 12: Learning Theory, Introduction to Reinforcement Learning, Optional videos (RL framework, TD learning, Solution Methods, Applications)

Nptel Introduction to Machine Learning Week 2 Assignment Answer

Week 2 : Assignment 2

Due date: 2025-02-05, 23:59 IST.
Assignment not submitted
1 point

In a linear regression model y=θ0+θ1x1+θ2x2+...+θpxpy=θ0+θ1×1+θ2×2+…+θpxp, what is the purpose of adding an intercept term (θ0)(θ0)?

 
 
 
 
1 point
Which of the following is true about the cost function (objective function) used in linear regression?
 

 
 
1 point
Which of these would most likely indicate that Lasso regression is a better choice than Ridge regression?
 
 
 
 
1 point
Which of the following conditions must hold for the least squares estimator in linear regression to be unbiased?
 
 
 
 
1 point
When performing linear regression, which of the following is most likely to cause overfitting?
 
 
 
 
1 point

You have trained a complex regression model on a dataset. To reduce its complexity, you decide to apply Ridge regression, using a regularization parameterλλ. How does the relationship between bias and variance change asλλ becomes very large? Select the correct option

 
 
 
 
1 point
Given a training data set of 10,000 instances, with each input instance having 12 dimensions and each output instance having 3 dimensions, the dimensions of the design matrix used in applying linear regression to this data is
 
 
 
 
1 point

The linear regression model y=a0+a1x1+a2x2+...+apxpy=a0+a1x1+a2x2+…+apxp is to be fitted to a set of NN training data points having P attributes each. Let XX be NN x (p+1)(p+1) vectors of input values (augmented by 1‘s), YY be NN x 11 vector of target values, and θθ be (p+1)×1(p+1)×1 vector of parameter values (a0,a1,a2,...,ap)(a0,a1,a2,…,ap). If the sum squared error is minimized for obtaining the optimal regression model, which of the following equation holds?

1 point
Which of the following scenarios is most appropriate for using Partial Least Squares (PLS) regression instead of ordinary least squares (OLS)?
 
 
 
 
1 point
Consider forward selection, backward selection and best subset selection with respect to the same data set. Which of the following is true?
 
 
 
 
 

Related Posts