Download ImageNet Classification with Deep Convolutional Neural Networks

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Neural coding wikipedia , lookup

Optogenetics wikipedia , lookup

Central pattern generator wikipedia , lookup

Neuropsychopharmacology wikipedia , lookup

Catastrophic interference wikipedia , lookup

Metastability in the brain wikipedia , lookup

Nervous system network models wikipedia , lookup

Neural binding wikipedia , lookup

Development of the nervous system wikipedia , lookup

Neural engineering wikipedia , lookup

Artificial neural network wikipedia , lookup

Recurrent neural network wikipedia , lookup

Types of artificial neural networks wikipedia , lookup

Convolutional neural network wikipedia , lookup

Transcript
ImageNet Classification with Deep
Convolutional Neural Networks
Presenter: Weicong Chen
Deep Convolutional Neural Networks
• Led by Geoffrey Hinton, University of Toronto
• Published in 2013
• Based on the datasets from ImageNet LSVRC-2010 Contest
• Using graphic cards to train the neural network
ImageNet LSVRC-2010 Contest
• 1.2 million high-resolution images
• 1,000 different classes
• 50,000 validation images, 150,000 testing images
• Top-1 error 47.1% best in contest, 45.7% best published
• Top-5 errors 28.2% best in contest, 25.7% best published
Convolutional Neural Networks (CNN)
Architecture
Novel/Unusual features in architecture
• ReLU (Rectified Linear Units ) Nonlinearity
Standard way: f(x) = tanh(x) or f(x) = (1 + e-x)-1 (logistic function)
ReLU: f(x) = max(0, x)
6 times faster than hyperbolic function
• Multiple GPUs
Larger memory
Parallel computing
Control of communication
Overfitting
• Occurs when a statistical model describes random error or noise
instead of the underlying relationship
• Exaggerate minor fluctuations in the data
• Will generally have poor predictive performance
Reducing overfitting
• Data Augmentation
1.Image translation and horizontal reflection
Randomly extracting patches
Four corner and one center patches with reflection for testing
2.Altering the intensities of the RGB channels in training images
Approximately captures an important property of natural images
reduces the top-1 error rate by over 1%
Reducing Overfitting
• Dropout
Zero the output of each hidden neuron with probability 0.5.
No longer contribute to forward pass and backward propagation
Neural network samples a different architecture every time
Reduce complex co-adaptations of neurons
Used in two fully-connected layers
Result
• Two NVIDIA GTX 580 3GB GPUs
• 6 days of training
• 90 cycles
• 60 million parameters
• 37.5% top-1 error (45.7% second best)
• 17% top-5 error (28.2% second best)
Questions?