6. Machine Learning Classifier (Partner Assignment) | COS 126 Spring'22 Search COS 126 Spring'22 COS 126 Spring'22 Syllabus Schedule Assignments Project Resources Exams People Light Dark Automatic 6. Machine Learning Classifier (Partner Assignment) Goals To learn about object-oriented programming concepts. To learn about machine learning (ML). To implement the perceptron algorithm - a simple and beautiful example of ML in action. To write a client program that classifies images. Getting Started Read/scan this entire project description (including through Possible Progress Steps below) before starting to code. This will provide a big picture before diving into the assignment! Download the project zip for this assignment, which contains the files you will need for this assignment. Read Sections 3.1 and 3.2 of the textbook to learn the basics of object-oriented programming and how to use the Picture and Color data types. This is a partner assignment. (See instructions on Ed for creating your TigerFile group.) Read the COS 126 Style Guide to ensure your code follows our conventions. Style is an important component of writing code, and not following guidelines will result in deductions. Background Classification is one of the central problems in ML, a discipline that is transforming 21st century computing. As a familiar example, consider the problem of classifying handwritten digits using this web application Reset Derived from https://github.com/DenseInL2/webcnn ML algorithms like this are widely used to classify handwritten digits (e.g., to recognize postal ZIP codes, process bank checks, and parse income tax forms). The full power of ML derives from its amazing versatility. ML algorithms rely upon data to learn to make predictions, without being explicitly programmed for the task. For example, the same code you will write to classify handwritten digits extends to classifying other types of images, simply by training the algorithm with different data. Moreover, ML techniques apply not only to images but also to numerical, text, audio, and video data. Modern applications of ML span science, engineering, and commerce: from autonomous vehicles, medical diagnostics, and video surveillance to product recommendations, voice recognition, and language translation. Implementation Tasks Implement three classes: - Perceptron.java - MultiPerceptron.java - ImageClassifier.java Submit a completed readme.txt file. Approach Supervised learning. To classify images, we will use a supervised learning algorithm. Supervised learning is divided into two phases — training and testing. Training. In the training phase, the algorithm learns a function that maps an input to an output (or a label) using training data consisting of known input–output pairs. For the handwritten digit application, the training data comprise 60,000 grayscale images (inputs) and associated digits (labels). Here is a small subset: In the binary classification problem, we seek to classify images into one of two classes (e.g., either the digit 6 or some digit other than 6). By convention, we use the labels +1 (positive) and -1 (negative) to denote the two classes. In the multiclass classification problem, we allow for \( m \) classes and label them \( 0,1,…,m-1 \). For our handwritten digit application, there are \( m=10 \) classes, with class \( i \) corresponding to digit \( i \). Testing. In the testing phase, the algorithm uses the learned function to predict labels for unseen inputs. Typically, the algorithm makes some prediction errors (e.g., predicts 9 when the handwritten digit is 6). An important quality metric is the test error rate — the fraction of testing inputs that the algorithm misclassifies. It measures how well the learning algorithm generalizes from the training data to new data. Perceptrons. A perceptron is a simplified model of a biological neuron. It is a function that takes a vector \( x = x_0, x_1, \ldots, x_{n-1} \) of \( n \) real numbers as input and outputs (or predicts) either +1 or −1. A perceptron is characterized by a vector \( w = w_0, w_1, \ldots, w_{n-1} \) of \( n \) real numbers known as the weight vector. The perceptron computes the weighted sum $$ S = w_0 \times x_0 +w_1 \times x_1 + \ldots + w_{n-1} \times x_{n-1} $$ and outputs the sign of that number. For the handwritten digit application, a perceptron will be trained to predict +1 for images that correspond to a target digit and −1 otherwise; the input vector \( x \) holds the grayscale values of each pixel in an image; and the weight vector \( w \) is pre-computed by a process described in the next paragraph. Perceptron algorithm. How do we determine the values of the weight vector \( w \) so that the perceptron makes accurate predictions? The core idea is to use the training data of known input–output pairs to incrementally refine the weights. Specifically, we initialize all the weights to 0 and then process the labeled inputs one at a time. When we process a labeled input, there are three possibilities: Correct prediction: The perceptron predicts the correct label (+1 or -1) for the given input vector \( x \). In this case, we leave \( w \) unchanged. False positive: The given input vector \( x \) is labeled -1 but the perceptron predicts +1. In this case, we adjust \( w \) as follows - for each \(j\): $$ w^\prime_j=w_j-x_j $$ False negative: The given input vector \( x \) is labeled +1 but the perceptron predicts -1. In this case, we adjust \( w \) as follows - for each \(j\): $$ w^\prime_j=w_j+ x_j $$ Example. Here is an example trace of the perceptron algorithm using four (4) labeled inputs, each of length \( n=3 \). In this example, an input \( x \) has a positive label +1 if and only if the following is true: \( x_0 \le x_1 \le x_2 \) ; otherwise it has a negative label -1. Perceptron data type Create a data type that represents a perceptron by implementing the following API: public class Perceptron {
// Creates a perceptron with n inputs. It should create an array
// of n weights and initialize them all to 0.
public Perceptron(int n)
// Returns the number of inputs n.
public int numberOfInputs()
// Returns the weighted sum of the weight vector and x.
public double weightedSum(double[] x)
// Predicts the label (+1 or -1) of input x. It returns +1
// if the weighted sum is positive and -1 if it is negative (or zero).
public int predict(double[] x)
// Trains this perceptron on the labeled (+1 or -1) input x.
// The weights vector is updated accordingly.
public void train(double[] x, int label)
// Returns a String representation of the weight vector, with the
// individual weights separated by commas and enclosed in parentheses.
// Example: (2.0, 1.0, -1.0, 5.0, 3.0)
public String toString()
// Tests this class by directly calling all instance methods.
public static void main(String[] args)
}
Corner cases. You may assume that the arguments to the constructor and instance methods are valid. For example, you may assume that any label is either +1 or -1, any input vector \( x \) is of length \( n \), and \( n \ge 1 \). Multiclass classification In the previous section, we used a single perceptron for the binary classification problem. For a multiclass classification problem with \( m \) classes, we create an array of \( m \) perceptrons, each solving its own binary classification problem. For our handwritten digit application, each perceptron \( i \) solves a binary classification problem: does the image correspond to the digit \( i \) We train each perceptron independently and make predictions by distilling the results from the \( m \) perceptrons. Multiclass training. Initialize the weight vector of each of the \( m \) perceptrons to be zero and process the labeled training inputs one at a time. To train the perceptrons on an input vector \( x \) with multiclass label \( i \) (0 to \( m-1 \)): Train perceptron \( i \) on input vector \( x \) with the label +1 Train the other \( m-1 \) perceptrons on input vector \( x \) with the label -1 That is, when training perceptron \( i \), we treat an input vector labeled \( i \) as a positive example and an input vector with any other label as a negative example. Multiclass prediction. To make a prediction for an input vector \( x \), compute the weighted sum for each of the \( m \) perceptrons on that input. The multiclass prediction is the index of the perceptron with the largest weighted sum. Intuitively, each perceptron with a positive weighted sum predicts that \( x \) is a positive example for its class, but we need to pick only one. Note that the perceptron with the largest weighted sum makes that prediction with the most intensity, so we use that perceptron for the prediction (regardless of whether it is positive or negative). This one-vs-all strategy decomposes a multiclass classification task into \(m\) binary classification tasks. In computer science, this decomposition is known as a reduction; this particular kind of reduction is used all over ML. Example. Here is an example of a multi-perceptron with \(m\) = 2 classes and \(n =\) 3 inputs. MultiPerceptron data type Create a data type that represents a multi-perceptron by implementing the following API: public class MultiPerceptron {
// Creates a multi-perceptron object with m classes and n inputs.
// It creates an array of m perceptrons, each with n inputs.
public MultiPerceptron(int m, int n)
// Returns the number of classes m.
public int numberOfClasses()
// Returns the number of inputs n (length of the feature vector).
public int numberOfInputs()
// Returns the predicted label (between 0 and m-1) for the given input.
public int predictMulti(double[] x)
// Trains this multi-perceptron on the labeled (between 0 and m-1) input.
public void trainMulti(double[] x, int label)
// Returns a String representation of this MultiPerceptron, with
// the string representations of the perceptrons separated by commas
// and enclosed in parentheses.
// Example with m = 2 and n = 3: ((2.0, 0.0, -2.0), (3.0, 4.0, 5.0))
public String toString()
// Tests this class by directly calling all instance methods.
public static void main(String[] args)
}
Ties. If two (or more) perceptrons tie for the largest weighted sum in predictMulti(), return the index of any such perceptron (between 0 and \( m-1 \)). Corner cases. You may assume that the arguments to the constructor and instance methods are valid. For example, you may assume that any label is an integer between 0 and \( m-1 \), any input vector \( x \) is of length \( n \), and \( m \ge 1 \) and \( n \ge 1 \). ImageClassifier client Your final task is to write a client program ImageClassifier.java that classifies images using the MultiPerceptron data type described in the previous section by: Training it using the input–output pairs specified in a training data file. Testing the predictions using the input–output pairs specified in a testing data file. Printing a list of misclassified images and the test error rate. Organize your client according to the following API: public class ImageClassifier {
// Creates a feature vector (1D array) from the given picture.
public static double[] extractFeatures(Picture picture)
// See below.
public static void main(String[] args)
}
Here are some details about the API: Input file format. A training data file consists of a sequence of lines: the first line contains the number of classes \(m\); the second line contains the width and height, respectively, of the images; each remaining line contains the name of an image file (e.g., corresponding to a handwritten digit) followed by an integer label (e.g., identifying the correct digit), separated by whitespace. The file format for testing data files is the same (but you will use the integer labels in the testing data files only to check the accuracy of your predictions). Input files. We provide a variety of sample input files in the specified format, including handwritten digits, fashion articles from Zalando, Hirigana characters, and doodles of fruit, animals, and musical instruments. Data files Description Examples Source digits.jar handwritten digits MNIST (Click Cancel if prompted to sign in) fashion.jar fashion articles Fashion MNIST Kuzushiji.jar Hirigana characters Kuzushiji MNIST fruit.jar fruit doodles Quick, Draw! animals.jar animal doodles Quick, Draw! music.jar musical instrument doodles Quick, Draw! Feature extraction. The extractFeatures() method converts a grayscale image into a one-dimensional array suitable for use with the MultiPerceptron data type. In one pass over the pixels of the image: extract the grayscale values from each pixel and rearrange them into a single 1D array (vector). Notes: Recall that a shade of gray has its red, green, and blue components all equal The one-dimensional array must be of length width x height In one pass, you must iterate over the image RGB pixel values in row-major order, extracting the grayscale value and setting the appropriate element in the 1D array (vector). Main. The main() method takes two command-line arguments: The name of a file that contains the training data. The name of a file that contains the testing data. It creates a MultiPerceptron object with \( m \) classes and \( n = width \times height \) inputs; trains it using the images and labels from the training data file; and classifies images from the testing data file, producing as output the following information: For each misclassified image, print the image’s filename, its true label, and the incorrectly predicted digit in the format shown below. The test error rate (the fraction of test images that the algorithm misclassified). Here are some sample executions: > java-introcs ImageClassifier digits-training20.txt digits-testing10.txt
digits/testing/1/46.png, label = 1, predict = 3
digits/testing/7/36.png, label = 7, predict = 2
digits/testing/7/80.png, label = 7, predict = 2
digits/testing/1/40.png, label = 1, predict = 2
digits/testing/1/39.png, label = 1, predict = 3
digits/testing/7/79.png, label = 7, predict = 2
digits/testing/9/20.png, label = 9, predict = 2
digits/testing/9/58.png, label = 9, predict = 4
test error rate = 0.8
> java-introcs ImageClassifier digits-training60K.txt digits-testing10K.txt
jar:file:digits.jar!/testing/5/9428.png, label = 5, predict = 3
jar:file:digits.jar!/testing/6/4814.png, label = 6, predict = 0
jar:file:digits.jar!/testing/5/4915.png, label = 5, predict = 8
...
jar:file:digits.jar!/testing/5/7870.png, label = 5, predict = 4
jar:file:digits.jar!/testing/4/1751.png, label = 4, predict = 3
jar:file:digits.jar!/testing/5/6043.png, label = 5, predict = 3
test error rate = 0.136
> java-introcs ImageClassifier fashion-training60K.txt fashion-testing10K.txt
...
> java-introcs ImageClassifier Kuzushiji-training60K.txt Kuzushiji-testing10K.txt
...
Possible Progress Steps These are purely suggestions for how you might make progress. You do not have to follow these steps. Perceptron.java Implement the Perceptron() constructor and the method numberOfInputs(). Initialize any instance variables. We recommend using two instance variables: an integer n and an array of floating-point numbers weights[]. Test: In the main() method, instantiate a few Perceptron objects and print the number of inputs for each object. Implement the toString() method. Test: In the main() method, print the various Perceptron objects. What should the output be for a newly instantiated Perceptron object? Implement the weightedSum() method. Test: In the main() method, print the result of invoking the weightedSum() method on the various Perceptron objects (using, of course, appropriately sized arrays). Implement the predict() method. Implement the train() method. Note: train() should call predict(). You can test by using the code in the Testing section (below) and submit to TigerFile. Do not move onto MultiPerceptron until Perceptron is working properly. MultiPerceptron.java Implement the MultiPerceptron() constructor and the methods numberOfClasses() and numberOfInputs(). Initialize any instance variables; we recommend using three instance variables: an integer m (for the number of classes), an integer n (for the length of the input feature vector), and an array of Perceptron objects perceptrons[]. Test: In the main() method, instantiate a few MultiPerceptron objects and print the number of classes and inputs for each object. Implement the toString() method. Test: In the main() method, print the various MultiPerceptron objects. What should the output be for a newly instantiated MultiPerceptron object? Implement the predictMulti() method. Implement the trainMulti() method. You can test by using the code in the Testing section (below) and submit to TigerFile. ImageClassifier.java Step 1. Feature extraction. Review Section 3.1 of the textbook, especially Program 3.1.4 (Grayscale.java) for using the Picture and Color data types. Note that the images are already grayscale, so you don’t need to use Luminance.java. In particular, the red, green, and blue components are equal, so you can use any of getRed(), getGreen(), or getBlue() to get the grayscale value. Create a Picture object for the image 49785.png (in the project folder) and display it in a window. (Remove this code after you have successfully displayed the image.) Create a Picture object for the image image3-by-3.png (in the project folder) corresponding to the 3-by-3 example given below. Extract its width and height and print the values to standard output. Then, extract the grayscale values of the pixels and print. If it’s not already in row-major order, adjust your code so that it prints the values in the specified order. If you are using IntelliJ, do not type the import java.awt.Color; statement that is normally needed to access Java’s Color data type. IntelliJ is pre-configured to automatically add import statements when needed (and remove them when not needed). Create a one-dimensional array of length width x height and copy the grayscale values to the array. Print the values of this array to confirm you can create a vector (row-major order) of values from a Picture object. Using the code from steps (3) and (4) as a guide, implement the static method extractFeatures() that takes a Picture as an argument and returns the grayscale values as a double[] in row-major order. Write a main() method that tests extractFeatures(). Testing Hint! Using image3-by-3.png, your main() method should print the values returned by extractFeatures() as shown in the above figure. Parsing input files. Review the In data type from Section 3.1, which is an object-oriented version of StdIn. You need the object-oriented version in this program since you will be reading from two different files in the same program. Create an In object for the training file digits-training20.txt. Read the integers m, width, and height, and print them to standard output. Read pairs of filenames (strings) and labels (integers) and print them to standard output. Note - printing is solely for checking progress; comment out the print statements once you have confidence it is working properly. For each image, create a new Picture object and display it in its own window. (Note: the image displays may overlap!) Note - displaying the images is solely for checking progress; comment out the display statements once you have confidence it is working properly. Repeat the previous two steps for the testing file digits-testing10.txt. Note - even though the testing file contains the same values as the training file for integers m, width, and height, you must still read these values from the testing file. Modify your program to take the names of the testing and training files as command-line arguments. Step 2. Classifying images. Classifier. After reading m, width, and height from the training file, create a MultiPerceptron object of the correct dimensions. Training. For each training image, train the classifier using the corresponding label. Testing. For each testing image, predict its class. Print each misclassified image to standard output and tabulate statistics on the number of misclassified images. Test error rate. After training and testing, print the test error rate to standard output. Use StdOut.println to print this value. Test. Test your program on some small data files, such as digits-training20.txt and digits-testing10.txt. Testing Perceptron.java. Here is Java code that trains a Perceptron on four input vectors (of length 3) from the assignment specification: int n = 3;
double[] training1 = { 3.0, 4.0, 5.0 }; // yes
double[] training2 = { 2.0, 0.0, -2.0 }; // no
double[] training3 = { -2.0, 0.0, 2.0 }; // yes
double[] training4 = { 5.0, 4.0, 3.0 }; // no
Perceptron perceptron = new Perceptron(n);
StdOut.println(perceptron);
perceptron.train(training1, +1);
StdOut.println(perceptron);
perceptron.train(training2, -1);
StdOut.println(perceptron);
perceptron.train(training3, +1);
StdOut.println(perceptron);
perceptron.train(training4, -1);
StdOut.println(perceptron);
And the desired output: (0.0, 0.0, 0.0)
(3.0, 4.0, 5.0)
(3.0, 4.0, 5.0)
(3.0, 4.0, 5.0)
(-2.0, 0.0, 2.0)
MultiPerceptron.java. Here is Java code that trains a MultiPerceptron on four input vectors (of length 3) from the assignment specification: int m = 2;
int n = 3;
double[] training1 = { 3.0, 4.0, 5.0 }; // class 1
double[] training2 = { 2.0, 0.0, -2.0 }; // class 0
double[] training3 = { -2.0, 0.0, 2.0 }; // class 1
double[] training4 = { 5.0, 4.0, 3.0 }; // class 0
MultiPerceptron perceptron = new MultiPerceptron(m, n);
StdOut.println(perceptron);
perceptron.trainMulti(training1, 1);
StdOut.println(perceptron);
perceptron.trainMulti(training2, 0);
StdOut.println(perceptron);
perceptron.trainMulti(training3, 1);
StdOut.println(perceptron);
perceptron.trainMulti(training4, 0);
StdOut.println(perceptron);
And the desired output: ((0.0, 0.0, 0.0), (0.0, 0.0, 0.0))
((0.0, 0.0, 0.0), (3.0, 4.0, 5.0))
((2.0, 0.0, -2.0), (3.0, 4.0, 5.0))
((2.0, 0.0, -2.0), (3.0, 4.0, 5.0))
((2.0, 0.0, -2.0), (-2.0, 0.0, 2.0))
Here is Java code that tests a MultiPerceptron on two input vectors (of length 3) from the assignment specification, based on the trained Multiperceptron object: double[] testing1 = { -1.0, -2.0, 3.0 };
double[] testing2 = { 2.0, -5.0, 1.0 };
StdOut.println(perceptron.predictMulti(testing1));
StdOut.println(perceptron);
StdOut.println(perceptron.predictMulti(testing2));
StdOut.println(perceptron);
And the desired output: 1
((2.0, 0.0, -2.0), (-2.0, 0.0, 2.0))
0
((2.0, 0.0, -2.0), (-2.0, 0.0, 2.0))
ImageClassifier.java. Now, the fun part. Use large training and testing input files. Be prepared to wait for one (1) minute (or more) while your program processes 60,000 images. > java-introcs ImageClassifier digits-training60K.txt digits-testing10K.txt
jar:file:digits.jar!/testing/5/9428.png, label = 5, predict = 3
jar:file:digits.jar!/testing/6/4814.png, label = 6, predict = 0
jar:file:digits.jar!/testing/5/4915.png, label = 5, predict = 8
...
jar:file:digits.jar!/testing/5/7870.png, label = 5, predict = 4
jar:file:digits.jar!/testing/4/1751.png, label = 4, predict = 3
jar:file:digits.jar!/testing/5/6043.png, label = 5, predict = 3
test error rate = 0.136
Don’t worry about the odd looking filenames. It’s just a verbose way to specify the location to a specific image file in a JAR (Java ARchive) file. Modern operating systems are not so adept at manipulating hundreds of thousands of individual image files, so this makes training more efficient. In this case, jar:file:digits.jar identifies the JAR file digits.jar, and /training/7/4545.png identifies a file named 4545.png, which is located in the subdirectory /training/7/ of the JAR file. readme.txt - Analysis Part 1: Some people (especially in Europe and Latin America) write a 7 with a line through the middle, while others (especially in Japan and Korea) make the top line crooked. Our data set 7 with line through it 7 with top line crooked Suppose that the training data consists solely of samples that do not use any of these conventions. How well do you think the algorithm will perform when you test it on different populations? What are the possible consequences? Now suppose that you are using a supervised learning algorithm to diagnose cancer. Suppose the training data consists of examples solely on individuals from population X but you use it on individuals from population Y. What are the possible consequences? Part 2: Run the following experiment (you may wish to redirect the standard output to a file): > java-introcs ImageClassifier digits-training60K.txt digits-testing1K.txt
What digit is misclassified the most frequently? For this digit, what are the top two digits that your MultiPerceptron incorrectly predicts? Examine some of these misclassified images. Provide an explanation of what might have caused these misclassifications. Provide your answers in your readme.txt file. Submission Before submitting the assignment, please ensure you have read the COS 126 Style Guide and that your code follows our conventions. Style is an important component of writing code, and not following guidelines will result in deductions. Submit Perceptron.java, MultiPerceptron.java, ImageClassifier.java, and a completed readme.txt file to TigerFile Enrichment (Click on the to expand and the to collapse the answer.) When was the perceptron algorithm first discovered? It was proposed in 1958 by Frank Rosenblatt, an American psychologist. It is one of the earliest machine-learning algorithms. At the time, researchers were overly optimistic about its short-term potential and grossly underestimated the formidable challenge of building intelligent machines. What is the significance of the perceptron algorithm? It is a really simple, beautiful algorithm that, nevertheless, can do something interesting. The perceptron algorithm is one of the most fundamental algorithms in an area of ML called online learning (learning from samples one at a time). The perceptron algorithm is closely related to the support-vector machine algorithm, another fundamental ML algorithm. The perceptron algorithm has some beautiful theoretical properties. For example, if the training data set is linearly separable (i.e., there exists some weight vector that correctly classifies all of the training examples), and if we cycle through the training data repeatedly, the perceptron algorithm will eventually (and provably) find such a weight vector. Perceptrons are a building block in neural networks. In neural networks, many perceptrons (or other artificial neurons) are connected in a network architecture, in which the outputs of some neurons are used as the inputs to other neurons. Training multi-layer neural networks requires a more sophisticated algorithm to adjust the weights, such as gradient descent. Does the perceptron algorithm produce a different weight vector depending on the order in which it processes the training examples? Yes. We’ve randomized the training data. It would be a bad idea, for example, to process all of the handwritten 0s before all of the handwritten 6s. Any intuition for why the perceptron algorithm works? Geometrically, you can view each input vector \( x \) as a point in \( R^n \) and the weight vector \( w \) as a hyperplane through the origin. The goal of the perceptron algorithm is to find a hyperplane that separates the positive examples from the negative examples. Using vector terminology, the weighted sum is known as the dot product; its sign determines on which side of the hyperplane the point lies. During training, when we encounter a point \( x \) that is on the wrong side of the hyperplane (i.e., a false positive or negative), we update the weight vector, thereby rotating the hyperplane slightly. After the rotation, x is either on the correct side of the hyperplane or, if not, at least a bit closer to the hyperplane (in terms of Euclidean distance). How can I improve accuracy of the perceptron algorithm? Here are three simple ideas: Multiclass perceptron. Instead of training all m perceptrons on each input vector, when there is a prediction error (multiclass perceptron predicts \( i \) but correct label is \( k \)), train only two perceptrons: train perceptron \( i \) (with label −1) and perceptron \( k \) (with label +1). Adjust the weights with a fraction of +1 or -1 for correct or incorrect predictions (this helps with a smoother convergence) and iterate over the training step multiple times, each time training the perceptron with the same set of training data (randomized in order). Normalize the Features array data to have values between 0 and 1 (divide the values with 255) and initialize the perceptron weights to random values (with uniform random or Gaussian random to be less than 1 and on average 0). Averaged perceptron. Instead of using the last weight vector, take the average of the weight vectors that are computed along the way. Incorporate more features. Instead of using the feature vector \( x_0, x_1,\ldots, x_{n-1} \), create additional features. In particular, for each pair of features \( x_i \) and \( x_k \), create a new feature \(x_{ik} = x_i * x_k \). You could also keep going, adding not just pairs of features, but also triples, etc. This can significantly improve accuracy, but it becomes prohibitively expensive in terms of computation. See this paper for additional ideas, including the kernel trick and the voted-perceptron algorithm. I noticed that the weights are always integral throughout the perceptron algorithm. Why is this? Adding or subtracting an integer to an integer yields an integer. The weights are always adjusted by either adding or subtracting the input vector; for the image classification problems we consider in this assignment, the input vector elements are integers between 0 and 255 (grayscale values). As a perceptron is trained with more and more data, the weights increase in absolute value. Is this to be expected? Yes. This implies that the algorithm makes smaller and smaller relative changes to the weights as it learns from more and more input–output pairs. Why not use the testing data for training? We seek a model that can make good predictions for unseen input–output pairs. If we use the testing data when training, the algorithm could “memorize” the input–output pairs. While this might achieve higher accuracy on those inputs, the results may not generalize to unseen inputs. What is the difference between supervised and unsupervised learning? In supervised learning, the algorithm has access to training data for which we know the correct labels. These labels act as a teacher supervising the learning process. In unsupervised learning, the training data is unlabeled, so there are no correct answers; instead, the goal may be to cluster similar inputs together. How can I make it recognize my own handwritten digits? Use the same technique, but you must be careful to size-normalize and center the images, as is done in the MNIST training data. Can ML algorithms be trained to classify alphabetic and mathematical symbols? Yes. Here are web apps that let you draw and find the most similar Unicode characters or LaTeX symbols. Any games based on classifying hand-drawn images? Yes. Try Quick, Draw!, which uses neural networks and the world’s largest doodling data set. What is the best performing algorithm for classifying handwritten digits using the MNIST dataset? The current champion uses convolution neural networks and achieves a 99.79% accuracy rate on the MNIST testing database consisting of 10,000 images. Here are the 21 incorrect predictions: There is not much room for improvement; indeed, some of the errors appear to be due to incorrectly labeled (or ambiguous) inputs. This assignment was developed by Robert DeLuca, Ruth Fong, Maia Ginsburg, Alan Kaplan, Kevin Jeon, and Kevin Wayne. Copyright © 2021-2022, Kevin Wayne. Published with Wowchemy — the free, open source website builder that empowers creators. Cite × Copy Download