Question 1:
This is supervised learning model. It is example of regression. We are interested in prediction. We predict the most promising spot to dig. Here number of observations n = 80 and number of predictors p = 24.
1: (b)
This is supervised learning model. It is example of classification. Here we are interested in prediction. We predict the whether to display advertisement A or advertisement B to each customer. Here number of observations n = 300 and number of predictors p = 3 (age, zip code, and gender).
1: (c)
This is supervised learning model. It is example of regression. Here we are interested in inference. We are interested in discovering factors that are associated with the unemployment rate across different U.S. cities. Here number of observations n = 400 and number of predictors p = 6 (the population, state, average income, crime rate, percentage of students who graduate high school and unemployment level.).
1: (d)
This is unsupervised learning model. For the each students we don’t have responses (different subtypes) of students in the application pool.
1 (e):
This is supervised learning model. It is example of classification. Here we are interested in prediction. We predict the type of cells based on a few measurements. Here number of observations n = 68 and number of predictors p = 3 (the number of branch points, the number of active processes, and the average process length).
Question 2:
2 (a):
We preferred inflexible regression model as number of predictors (number of genes) p is extremely large, and the number of observations (number of patients) n is small.
2 (b):
We preferred flexible regression model as number of predictors (math, science and history grades in the 7th grade) p is small, and the number of observations (number of students) n is extremely large.
2 (c):
We preferred inflexible regression model as we variation in data is more.
2 (d):
We preferred flexible regression model as we variation in data is less. Flexible model will perform better to find non-linear effect also.
Question 3:
3 (a):
Flexible model performs better. A flexible models fits the data well with larger sample size and performs better than inflexible model.
3 (b):
Flexible model performs worse. A flexible models overfit small number of observation.
3 (c):
Flexible model performs worse. A flexible method would fit to the noise in the error terms and increase variance.
3 (d):
Flexible model performs better. A flexible models gets more degrees of freedom fits the data well.
3 (e):
Flexible model performs worse as it would fit to the noise in the error terms and increase variance.
Question 4:
Solution is obtained from R studio. Here we have given the some output of the program.
> summary(College) Private Apps Accept Enroll Top10perc No :212 Min. : 81 Min. : 72 Min. : 35 Min. : 1.00 Yes:565 1st Qu.: 776 1st Qu.: 604 1st Qu.: 242 1st Qu.:15.00 Median : 1558 Median : 1110 Median : 434 Median :23.00 Mean : 3002 Mean : 2019 Mean : 780 Mean :27.56 3rd Qu.: 3624 3rd Qu.: 2424 3rd Qu.: 902 3rd Qu.:35.00 Max. :48094 Max. :26330 Max. :6392 Max. :96.00 Top25perc F.Undergrad P.Undergrad Outstate Room.Board Min. : 9.0 Min. : 139 Min. : 1.0 Min. : 2340 Min. :1780 1st Qu.: 41.0 1st Qu.: 992 1st Qu.: 95.0 1st Qu.: 7320 1st Qu.:3597 Median : 54.0 Median : 1707 Median : 353.0 Median : 9990 Median :4200 Mean : 55.8 Mean : 3700 Mean : 855.3 Mean :10441 Mean :4358 3rd Qu.: 69.0 3rd Qu.: 4005 3rd Qu.: 967.0 3rd Qu.:12925 3rd Qu.:5050 Max. :100.0 Max. :31643 Max. :21836.0 Max. :21700 Max. :8124 Books Personal PhD Terminal S.F.Ratio Min. : 96.0 Min. : 250 Min. : 8.00 Min. : 24.0 Min. : 2.50 1st Qu.: 470.0 1st Qu.: 850 1st Qu.: 62.00 1st Qu.: 71.0 1st Qu.:11.50 Median : 500.0 Median :1200 Median : 75.00 Median : 82.0 Median :13.60 Mean : 549.4 Mean :1341 Mean : 72.66 Mean : 79.7 Mean :14.09 3rd Qu.: 600.0 3rd Qu.:1700 3rd Qu.: 85.00 3rd Qu.: 92.0 3rd Qu.:16.50 Max. :2340.0 Max. :6800 Max. :103.00 Max. :100.0 Max. :39.80 perc.alumni Expend Grad.Rate Min. : 0.00 Min. : 3186 Min. : 10.00 1st Qu.:13.00 1st Qu.: 6751 1st Qu.: 53.00 Median :21.00 Median : 8377 Median : 65.00 Mean :22.74 Mean : 9660 Mean : 65.46 3rd Qu.:31.00 3rd Qu.:10830 3rd Qu.: 78.00 Max. :64.00 Max. :56233 Max. :118.00 |
> length(which(College$Private==”Yes”))
[1] 565
Out of 777, 565 colleges are private whereas 212 are non-private.
> length(which(College$Elite==”Yes”))
[1] 78
There are 78 elite universities.
R Code:
# Set the working directory
# Read the data
College=read.csv(‘College.csv’,header=1)
#To view the data
View(College)
#For removing first column
rownames(College)= College[,1]
College=College[,-1]
View(College )
#Summary of Data
summary(College)
#scatterplot of the column PhD versus the column Grad.Rate.
plot(College$Grad.Rate,College$PhD,xlab=”Grad.Rate”,ylab=”PhD”,main=”Scatter Plot”)
#Number of Private Colleges
length(which(College$Private==”Yes”))
#(g)
Elite=rep(“No”,nrow (College))
Elite[College$Top10perc>50]=”Yes”
Elite=as.factor(Elite)
College=data.frame(College,Elite)
#How many elite universities are there?
length(which(College$Elite==”Yes”))
#Box Plot
plot(College$Elite, College$Outstate, xlab = “Elite University”, ylab =”Out of State tuition in USD”, main = “Outstate Tuition Plot”)
# (h) Histogram
par(mfrow=c(2,2))
hist(College$Top10perc, col = 4, xlab = “Top 10%”, ylab = “Count”,main=””)
hist(College$Top25perc, col = 6, xlab = “Top 25%”, ylab = “Count”,main=””)
hist(College$Books, col = 2, xlab = “Books”, ylab = “Count”,main=””)
hist(College$PhD, col = 3, xlab = “PhD”, ylab = “Count”,main=””)
Question 5:
One can see that variable is negatively skewed.
One can study the relation between different predictors by observing the above scatter plots of predictors.
(e)
> coef(summary(lm.model))[1:20,1:4]
Estimate Std. Error
(Intercept) 5.2523261623 1.005248e+01
Onset.Delta -0.0004375411 4.516149e-05
Symptom.Speech -0.0225458435 9.120369e-02
Symptom.WEAKNESS -0.1030724618 8.140383e-02
Site.of.Onset.Onset..Bulbar -0.3672390249 2.492628e-01
Site.of.Onset.Onset..Limb -0.2722652619 2.516364e-01
Race…Caucasian -0.1472308983 9.490391e-02
Age -0.0005163468 1.814937e-03
Sex.Female -0.0605416162 9.573716e-02
Sex.Male 0.0195376288 8.868793e-02
Mother -0.0462694066 7.581925e-02
Family 0.0067961369 5.629768e-02
Study.Arm.PLACEBO -3.1549354316 1.918113e+00
Study.Arm.ACTIVE -3.0519454033 1.917047e+00
max.alsfrs.score 0.0622267218 7.830124e-02
min.alsfrs.score -0.1666353650 8.465840e-02
last.alsfrs.score 0.4090457993 2.074172e-01
mean.alsfrs.score -0.1208903623 3.538618e-01
num.alsfrs.score.visits -0.0058420080 7.139877e-01
sum.alsfrs.score -0.0778763045 7.899171e-02
t value Pr(>|t|)
(Intercept) 0.522490777 6.014605e-01
Onset.Delta -9.688368476 3.700894e-21
Symptom.Speech -0.247203188 8.048088e-01
Symptom.WEAKNESS -1.266186858 2.057820e-01
Site.of.Onset.Onset..Bulbar -1.473300789 1.410284e-01
Site.of.Onset.Onset..Limb -1.081979013 2.795589e-01
Race…Caucasian -1.551368083 1.211739e-01
Age -0.284498398 7.760955e-01
Sex.Female -0.632373219 5.273077e-01
Sex.Male 0.220296371 8.256916e-01
Mother -0.610259353 5.418479e-01
Family 0.120717895 9.039421e-01
Study.Arm.PLACEBO -1.644811989 1.003665e-01
Study.Arm.ACTIVE -1.592003115 1.117439e-01
max.alsfrs.score 0.794709319 4.269974e-01
min.alsfrs.score -1.968326306 4.934477e-02
last.alsfrs.score 1.972092254 4.891257e-02
mean.alsfrs.score -0.341631529 7.327100e-01
num.alsfrs.score.visits -0.008182225 9.934735e-01
sum.alsfrs.score -0.985879479 3.244639e-01
We observed that R2 is 0.46 suggest that fitting is not too good.
We observed that RMSE is 0.4138632.
The error rate produced by using a simple linear regression on this data is much too high. It could be useful for variables selection. Bias–variance tradeoff shows that predictor with less bias has larger variance.
R Code:
#(a)
a=load(‘als.rData’)
#(b)
length(train.X)
length(train.y)
length(test.X)
length(test.y)
#(c)
summary(train.y) #Summary of train.y
hist(train.y,breaks = 40) #Histogram
#(d)
colnames(train.X)[1:20]
pairs(train.X[,21:25])
#Fitting of Regression model
lm.model=lm(train.y ~., data = data.frame(train.y,train.X))
#First 20 coefficients
coef(summary(lm.model))[1:20,1:4]
#R squared
summary(lm.model)$r.squared
pred=predict(lm.model)
#RMSE
RMSE=sqrt(mean((pred-train.y)^2))
RMSE
Reference:
James, G., Witten, D., Hastie, T. and Tibshirani, R., 2013. An introduction to statistical learning (Vol. 112). New York: springer.
Montgomery, D.C., Peck, E.A. and Vining, G.G., 2012. Introduction to linear regression analysis (Vol. 821). John Wiley & Sons.
Essay Writing Service Features
Our Experience
No matter how complex your assignment is, we can find the right professional for your specific task. Contact Essay is an essay writing company that hires only the smartest minds to help you with your projects. Our expertise allows us to provide students with high-quality academic writing, editing & proofreading services.Free Features
Free revision policy
$10Free bibliography & reference
$8Free title page
$8Free formatting
$8How Our Essay Writing Service Works
First, you will need to complete an order form. It's not difficult but, in case there is anything you find not to be clear, you may always call us so that we can guide you through it. On the order form, you will need to include some basic information concerning your order: subject, topic, number of pages, etc. We also encourage our clients to upload any relevant information or sources that will help.
Complete the order formOnce we have all the information and instructions that we need, we select the most suitable writer for your assignment. While everything seems to be clear, the writer, who has complete knowledge of the subject, may need clarification from you. It is at that point that you would receive a call or email from us.
Writer’s assignmentAs soon as the writer has finished, it will be delivered both to the website and to your email address so that you will not miss it. If your deadline is close at hand, we will place a call to you to make sure that you receive the paper on time.
Completing the order and download