0 ratings0% found this document useful (0 votes) 51 views2 pagesCS60050 Machine Learning EA 2015
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, 
claim it here.
Available Formats
Download as PDF or read online on Scribd
Department of Computer Science & Engineering
Indian Institute of Technology, Kharagpur
End-semester Examination, Autumn 2015
$6000: Machine Learning
Full Marks: 100
 
ime: 3 Hrs
Answer ALL questions. You can use calculators.
QI. Which of the following is true or false. Explain with reasons (max 2-3 sentences):
a. Rule post pruning in C4.5 can sometimes produce rules which do not correspond to a decision
tree [S marks]
b. SMO algorithm can be used to solve any quadratic programming problem. [5 marks}
©. In SGD, the objective functicn value decreases monotonically with the iterations. [5 marks]
4d. All conditional independences in encoded in the Bayesian network are preserved in the MRF
 
 
obtained after moralization, [5 marks]
e. The perceptron learning algorithm is a special case SGD. [5 marks}
f. The Nesterov and Vial’s proof for convergence of SGD applies to perceptron algorithm as
well. [5 marks}
2. Answer the following:
a. Write a stochastic gradient descent algorithm, for optimizing a function L(@) =
May Eker Un Yn, 9). [5 marks]
'b. Define the expected loss function, clearly mentioning the random variables. Write the formula
for convergence in expectation, clearly defining each variable. {5 marks]
©. Derive a condition on the step sizes which guarantees convergence in expectation. [5 marks]
4. Consider a situation where a biased coin with probability of head (p) is tossed N times and
the outcomes (head = 1,tail = 0) are recorded in random variables, xp, = 1,...,N.
Derive the log likelihood function for estimating parameter, p. Derive an SGD algorithm for
solving the Max-likelihood problem derived above. [10 marks}
 
Q3. Consider the following dataset for predicting outcome of a tennis match between Federer and
Nadale:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Time Mateh type | Court Best Effort | Outcome
surface
Moming Master Grass. 1 F
‘Afternoon Grand slam | Cla 1 F
Night Friendly Hard 0 F
‘Afternoon Friendly Mixed 0 N
‘Afternoon Master Clay a N
‘Afternoon Grand slam | Grass i F
‘Afternoon ‘Grand slam | Hard i F
‘Afternoon Grand slam | Hard 1 F
Morning. Master Grass 7 F
Afternoon Grand slam | Clay 7 N
Night Friendly Hard 0 F
Night Master Mixed 1 N
‘Afternoon Master Clay 1 N
‘Afternoon Master Grass 1 F
‘Afternoon Grand slam | Hard 1 F
Afternoon Grand slam [ Cla i FWrite the formula for information gain for an attribute. Compute information gain for all
attributes. Which is selected as a root node? [10 marks}
Compute the structure of the decision tree computed by ID3 algorithm. [5 marks}
Q4. Show 5 updates of perceptron algorithm for the following data, starting at w = (0,0)
Qs.
(6-4), HED), B,3) ELS OMS.29,
 
-2)} [10 marks]
Define d-separation in Bayesian Networks stating all the conditions. How does d-separation
help in determining conditional independence? [5 marks}
Consider the following Bayesian network connection various factors related to chest
diseases: [10 marks}
 
 
Broncrs?
 
 
 
 
roves? Postve
ore | om
i, Are Bronchitis and Tuberculosis independent when nothing is observed?
ii, Are they independent after observing x-ray?
iii, Are they independent after observing both x-ray and smoking?
Q6. Consider the following Markov Random Field with the corresponding potentials:
OQ OOO
(Xi, Xess) = 1 — XiXi4s for odd é, ard WeigaXi,Xisa) = 1 + XiXiva for even é; where X; €
{+1,—1). Find P(X,) and x3 = maxP(X3]X2 = 1) using an appropriate message-passing algorithm,
[20 marks}