Christoph Körner discusses the evolution and applications of deep learning in computer vision, detailing advancements from neural networks to various architectures like AlexNet and ResNet. The document highlights deep learning's superiority over traditional methods and human performance, emphasizing its effectiveness in tasks such as classification, segmentation, and object detection. The conclusion asserts that deep learning's power lies in its ability to learn from data, with a focus on the importance of data quality and quantity.
Outline
1) Introduction toNeural Networks
2) Deep Learning
3) Applications in Computer Vision
4) Conclusion
3.
Why Deep Learning?
●
Winsevery computer vision challenge
(classification, segmentation, etc.)
●
Can be applied in various domains (speech
recognition, game prediction, computer vision,
etc.)
●
Beats human accuracy
●
Big communities and resources
●
Hardware for Deep Learning
What happened until2011?
●
Better Initialization
●
Better Non-linearities: ReLU
●
1000 times more training data
●
More computing power
●
Factor 1 million speedup in training time through
parallelization on GPUs
10.
Deep Learning
●
Conv-, Pool-and Fully-Connected Layers
●
ReLU activations
●
Deep nested models with many parameters
●
New layer types and structures
●
New techniques to reduce overfitting
●
Loads of training data and compute power
●
10.000.000 images
●
Weeks of training on multi-GPU machines
Conclusion
●
Powerful, learn fromdata instead of hand-crafted
feature extraction
●
Better than humans
●
Deeper is always better
●
Overfitting
●
More data is always better
●
Data quality
●
Ground truth