cs231n - assignment1- k-Nearest Neighbor Classifier 梯度推导

时间:2021-04-16 21:22:08

assignment web page

k-Nearest Neighbor (kNN) exercise

Complete and hand in this completed worksheet (including its outputs and any supporting code outside of the worksheet) with your assignment submission. For more details see the assignments page on the course website.

The kNN classifier consists of two stages:

  • During training, the classifier takes the training data and simply remembers it
  • During testing, kNN classifies every test image by comparing to all training images and transfering the labels of the k most similar training examples
  • The value of k is cross-validated

In this exercise you will implement these steps and understand the basic Image Classification pipeline, cross-validation, and gain proficiency in writing efficient, vectorized code.

import numpy as np

class KNearestNeighbor(object):
  """ a kNN classifier with L2 distance """

  def __init__(self):
    pass

  def train(self, X, y):
    """ Train the classifier. For k-nearest neighbors this is just memorizing the training data. Inputs: - X: A numpy array of shape (num_train, D) containing the training data consisting of num_train samples each of dimension D. - y: A numpy array of shape (N,) containing the training labels, where y[i] is the label for X[i]. """
    self.X_train = X
    self.y_train = y

  def predict(self, X, k=1, num_loops=0):
    """ Predict labels for test data using this classifier. Inputs: - X: A numpy array of shape (num_test, D) containing test data consisting of num_test samples each of dimension D. - k: The number of nearest neighbors that vote for the predicted labels. - num_loops: Determines which implementation to use to compute distances between training points and testing points. Returns: - y: A numpy array of shape (num_test,) containing predicted labels for the test data, where y[i] is the predicted label for the test point X[i]. """
    if num_loops == 0:
      dists = self.compute_distances_no_loops(X)
    elif num_loops == 1:
      dists = self.compute_distances_one_loop(X)
    elif num_loops == 2:
      dists = self.compute_distances_two_loops(X)
    else:
      raise ValueError('Invalid value %d for num_loops' % num_loops)

    return self.predict_labels(dists, k=k)

  def compute_distances_two_loops(self, X):
    """ Compute the distance between each test point in X and each training point in self.X_train using a nested loop over both the training data and the test data. Inputs: - X: A numpy array of shape (num_test, D) containing test data. Returns: - dists: A numpy array of shape (num_test, num_train) where dists[i, j] is the Euclidean distance between the ith test point and the jth training point. """
    num_test = X.shape[0]
    num_train = self.X_train.shape[0]
    dists = np.zeros((num_test, num_train))
    for i in xrange(num_test):
      for j in xrange(num_train):
        #####################################################################
        # TODO: #
        # Compute the l2 distance between the ith test point and the jth #
        # training point, and store the result in dists[i, j]. You should #
        # not use a loop over dimension. #
        #####################################################################
        dists[i, j] = np.linalg.norm((X[i,:]-self.X_train[j,:]),ord = 2)
        #####################################################################
        # END OF YOUR CODE #
        #####################################################################
    return dists

  def compute_distances_one_loop(self, X):
    """ Compute the distance between each test point in X and each training point in self.X_train using a single loop over the test data. Input / Output: Same as compute_distances_two_loops """
    num_test = X.shape[0]
    num_train = self.X_train.shape[0]
    dists = np.zeros((num_test, num_train))
    for i in xrange(num_test):
      #######################################################################
      # TODO: #
      # Compute the l2 distance between the ith test point and all training #
      # points, and store the result in dists[i, :]. #
      #######################################################################
      dists[i, :] = np.linalg.norm((X[i,:]-self.X_train[:,:]), ord = 2, axis = 1)
      #######################################################################
      # END OF YOUR CODE #
      #######################################################################
    return dists

  def compute_distances_no_loops(self, X):
    """ Compute the distance between each test point in X and each training point in self.X_train using no explicit loops. Input / Output: Same as compute_distances_two_loops """
    num_test = X.shape[0]
    num_train = self.X_train.shape[0]
    dists = np.zeros((num_test, num_train)) 
    #########################################################################
    # TODO: #
    # Compute the l2 distance between all test points and all training #
    # points without using any explicit loops, and store the result in #
    # dists. #
    # #
    # You should implement this function using only basic array operations; #
    # in particular you should not use functions from scipy. #
    # #
    # HINT: Try to formulate the l2 distance using matrix multiplication #
    # and two broadcast sums. #
    #########################################################################
    X_vec_squared = np.diag( np.dot(X, X.transpose()) )

    X_train_vec_squared = np.diag( np.dot(self.X_train, self.X_train.transpose()) )

    X_X_train = -2*np.dot(self.X_train, X.transpose())

    dists = np.sqrt( (X_X_train + X_vec_squared).transpose() + X_train_vec_squared )

    #########################################################################
    # END OF YOUR CODE #
    #########################################################################
    return dists

  def predict_labels(self, dists, k=1):
    """ Given a matrix of distances between test points and training points, predict a label for each test point. Inputs: - dists: A numpy array of shape (num_test, num_train) where dists[i, j] gives the distance betwen the ith test point and the jth training point. Returns: - y: A numpy array of shape (num_test,) containing predicted labels for the test data, where y[i] is the predicted label for the test point X[i]. """
    #print 'dists shape: %d %d' % dists.shape
    num_test = dists.shape[0]
    y_pred = np.zeros(num_test)
    for i in xrange(num_test):
      # A list of length k storing the labels of the k nearest neighbors to
      # the ith test point.
      closest_y = []
      #########################################################################
      # TODO: #
      # Use the distance matrix to find the k nearest neighbors of the ith #
      # testing point, and use self.y_train to find the labels of these #
      # neighbors. Store these labels in closest_y. #
      # Hint: Look up the function numpy.argsort. #
      #########################################################################
      closest_y = self.y_train[ np.argsort(dists[i,:])[:k] ] 
      #########################################################################
      # TODO: #
      # Now that you have found the labels of the k nearest neighbors, you #
      # need to find the most common label in the list closest_y of labels. #
      # Store this label in y_pred[i]. Break ties by choosing the smaller #
      # label. #
      #########################################################################
      y_pred[i] = np.argmax( np.bincount(closest_y) ) 
      #########################################################################
      # END OF YOUR CODE # 
      #########################################################################

    return y_pred