* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download ImageNet Classification with Deep Convolutional Neural Networks
Neural coding wikipedia , lookup
Optogenetics wikipedia , lookup
Central pattern generator wikipedia , lookup
Neuropsychopharmacology wikipedia , lookup
Catastrophic interference wikipedia , lookup
Metastability in the brain wikipedia , lookup
Nervous system network models wikipedia , lookup
Neural binding wikipedia , lookup
Development of the nervous system wikipedia , lookup
Neural engineering wikipedia , lookup
Artificial neural network wikipedia , lookup
Recurrent neural network wikipedia , lookup
ImageNet Classification with Deep Convolutional Neural Networks Presenter: Weicong Chen Deep Convolutional Neural Networks • Led by Geoffrey Hinton, University of Toronto • Published in 2013 • Based on the datasets from ImageNet LSVRC-2010 Contest • Using graphic cards to train the neural network ImageNet LSVRC-2010 Contest • 1.2 million high-resolution images • 1,000 different classes • 50,000 validation images, 150,000 testing images • Top-1 error 47.1% best in contest, 45.7% best published • Top-5 errors 28.2% best in contest, 25.7% best published Convolutional Neural Networks (CNN) Architecture Novel/Unusual features in architecture • ReLU (Rectified Linear Units ) Nonlinearity Standard way: f(x) = tanh(x) or f(x) = (1 + e-x)-1 (logistic function) ReLU: f(x) = max(0, x) 6 times faster than hyperbolic function • Multiple GPUs Larger memory Parallel computing Control of communication Overfitting • Occurs when a statistical model describes random error or noise instead of the underlying relationship • Exaggerate minor fluctuations in the data • Will generally have poor predictive performance Reducing overfitting • Data Augmentation 1.Image translation and horizontal reflection Randomly extracting patches Four corner and one center patches with reflection for testing 2.Altering the intensities of the RGB channels in training images Approximately captures an important property of natural images reduces the top-1 error rate by over 1% Reducing Overfitting • Dropout Zero the output of each hidden neuron with probability 0.5. No longer contribute to forward pass and backward propagation Neural network samples a different architecture every time Reduce complex co-adaptations of neurons Used in two fully-connected layers Result • Two NVIDIA GTX 580 3GB GPUs • 6 days of training • 90 cycles • 60 million parameters • 37.5% top-1 error (45.7% second best) • 17% top-5 error (28.2% second best) Questions?