정리 카테고리
Lecture 2 formalizes the problem of image classification. We discuss the inherent difficulties of image classification, and introduce data-driven approaches. We discuss two simple data-driven image classification algorithms: K-Nearest Neighbors and Linear Classifiers, and introduce the concepts of hyperparameters and cross-validation.
Keywords: Image classification, K-Nearest Neighbor, distance metrics, hyperparameters, cross-validation, linear classifiers
slides:
http://cs231n.stanford.edu/slides/2017/cs231n_2017_lecture2.pdf
system input: images, predermined category labels
what computer really seas: just a grid
our classification algorithm should be robust at different kinds of transforms.
Choose hyperparameters that work best on the data (Don't do this)
-> K=1 always works perfectly on training data
Split data into train
and test
; choose hyperparameters that work best on test data (Don't do this)
-> No idea how algorithm will perform on new data
Split data into train
and val
, and test
; choose hyperparameters on val and evaluate on test (Better!)
Cross-Validation: Split data in folds, try each fold as validation and average the results
Q. training set vs validation set
validation set.
only for checking
how well algorithm is doing.