Midterm 2 will be this Friday, April 12. The exam will focus on these topics.
Make sure you understand the basics of how layers of a neural network are defined by combining an affine linear function (where is a weight matrix and is a bias vector) and an activation function like or the sigmoid function . You should know how to draw a computation graph and how to do the backpropagation algorithm, and understand what backpropagation is used for. There will be questions on the example similar to the questions on this workshop:
You should understand the basic idea of how image convolution with a kernel works. If I give you an image matrix like
then you should be able to compute the convolution with a simple kernel . For the pixels on the edge of the original image matrix, you can assume that the pixels just off the edge are all zeros, so for example the top left pixel is 0 and so are all of its neighbors (above, below, left, and right).
I might also ask you to predict what a convolution matrix does, for example will calculate how different the center pixel is from its immediate neighbors, so it will tend to detect the edges of an image.
I could ask you to do one step of the k-means clustering algorithm with some simple data. For example, if I had 5 points in and I wanted to find clusters. Suppose that the five points are and I start with and as my representative points. What are the clusters and representative points after 1 iteration of the k-means algorithm?
I won’t ask you to calculate the principal components of a large matrix, but you should understand what principal components are and why PCA is useful. You should also know what the covariance matrix is: for a data matrix with each row representing one data point, the covariance matrix is where is a matrix with every entry in column equal to the average of the entries in column of .
Then the principal components are the orthogonal columns of a matrix such that where is a diagonal matrix. If is the matrix with the most important columns of , then you can compress the data in the original data matrix by computing Then to recover the original data (approximately), You do not need to memorize any of these formulas, but you should be able to calculate simple examples. For example, you could be asked to calculate the covariance matrix for something like this data matrix:
You should know this algorithm. I might ask you to apply this algorithm with pencil and paper on a simple example. For example, using the -norm () to measure distance, what are the 3 closest neighbors of the point in the set ?
Given a Markov chain with transition matrix and reward vector , you should know that the expected value vector is the solution of the recursive formula and that you can use the value iteration algorithm to find the solution. You should also understand that expected value is the theoretical average of the total reward for each possible starting state.
Make sure you know and can explain the following terminology & concepts.