Convolutional Neural Networks
0
10
20
30
40
50
60
70
80
90
100
2004-01
2004-03
2004-05
2004-07
2004-09
2004-11
2005-01
2005-03
2005-05
2005-07
2005-09
2005-11
2006-01
2006-03
2006-05
2006-07
2006-09
2006-11
2007-01
2007-03
2007-05
2007-07
2007-09
2007-11
2008-01
2008-03
2008-05
2008-07
2008-09
2008-11
2009-01
2009-03
2009-05
2009-07
2009-09
2009-11
2010-01
2010-03
2010-05
2010-07
2010-09
2010-11
2011-01
2011-03
2011-05
2011-07
2011-09
2011-11
2012-01
2012-03
2012-05
2012-07
2012-09
2012-11
2013-01
2013-03
2013-05
2013-07
2013-09
2013-11
2014-01
2014-03
2014-05
2014-07
2014-09
2014-11
2015-01
2015-03
2015-05
2015-07
2015-09
2015-11
2016-01
2016-03
2016-05
2016-07
2016-09
2016-11
2017-01
2017-03
Google Trends
Deep learning Machine learning neural network
 Part of the machine learning field of learning representations
of data.
 hierarchy of multiple layers that mimic the neural networks
of our brain
 If you provide the system tons of information, it begins to
understand it and respond in useful ways.
 SuperIntelligent Devices
 Best Solution for
image recognition
speech recognition
natural language processing
Big Data
Geoffrey Hinton: University of Toronto & Google
Yann LeCun: New York University & Facebook
Andrew Ng: Stanford & Baidu
 Deep learning (DL) is a hierarchical structure network which
through simulates the human brain’s structure to extract the
internal and external input data’s features
Large data set with good quality
Measurable and describable goals
Enough computing power
Neural Network (Brain of Human)
Deep Neural Networks
Deep Belief Networks
Convolutional Neural Networks
Deep Boltzmann Machines
Deep Stacking Networks
 Convolution Neural Networks (CNN) is supervised learning and a
family of multi-layer neural networks particularly designed for use on
two dimensional data, such as images and videos.
 A CNN consists of a number of layers:
 Convolutional layers.
 Pooling Layers.
 Fully-Connected Layers.
 Convolutional layer acts as a feature extractor that extracts
features of the inputs such as edges, corners , endpoints.
 The pooling layer reduces the resolution of the image that
reduce the precision of the translation (shift and distortion)
effect.
 fully connected layer have full connections to all activations in
the previous layer.
 Fully connect layer act as classifier.
LeNet :The first successful applications of CNN
AlexNet: The ILSVRC 2012 winner
ZFNet: The ILSVRC 2013 winner
GoogLeNet: The ILSVRC 2014 winner
VGGNet: The runner-up in ILSVRC 2014
ResNet: The winner of ILSVRC 2015
MNIST Handwritten digits – 60000 Training + 10000 Test Data
Google House Numbers from street view - 600,000 digit images
CIFAR-10 60000 32x32 colour images in 10 classes
IMAGENET 1.2 million images, >150 GB
Tiny Images 80 Million tiny images
Flickr Data 100 Million Yahoo dataset
The ImageNet Large Scale Visual Recognition
Challenge (ILSVRC) evaluates algorithms for
object detection and image classification at large
scale.
26.2
15.3 14.8
7.3 6.7
3.6
0
5
10
15
20
25
30
Before 2012 AlexNet 2012 ZFNet 2013 VGGNet 2014 GoogleNet 2014 ResNet 2015
 MNIST is a large database of
handwritten digits.
 MNIST contains 60,000 training
images and 10,000 testing
images
 CNN on MNIST Dataset
 CIFAR-10 dataset consists
of 60000 32x32 colour
images in 10 classes
 CIFAR-10 contains 50000
training images and 10000
test images
 CNN on CIFAR-10 Dataset
 Overfitting Problem
 Larger network have a lots of
weights this lead to high model
complexity
 Network do excellent on training
data but very bad on validation
data
 CNN Optimization used to reduce the overfitting problem in CNN by:
1) Dropout
2) L2 Regularization
3) Mini-batch
4) Gradient descent algorithm
5) Early stopping
6) Data augmentation
 Dropout is a technique of reducing overfitting in CNN.
 L2 Regularization: Adding a regularization term for the weights
to the loss function is a way to reduce overfitting.
 where w is the weight vector, λ is the regularization factor
(coefficient), and the regularization function, Ω(w) is:
 Mini-batch is to divide the dataset into small batches of
examples, compute the gradient using a single batch, make an
update, then move to the next batch.
 The gradient descent algorithm updates the coefficients (weights
and biases) so as to minimize the error function by taking small steps
in the direction of the negative gradient of the loss function
 where i stands for the iteration number, α > 0 is the learning rate, P is
the parameter vector, and E(Pi) is the loss function.
 Early stopping
monitoring the deep
learning process of the
network from overfitting.
 If there is no more
improvement, or worse, the
performance on the test set
degrades, then the learning
process is aborted
 Data augmentation means increasing the number of dataset.
Today NVidia Support my work with
NVIDIA TITAN X
THE MOST ADVANCED GPU EVER BUILT
TITAN X Specifications
GPU Architecture Pascal
Standard Memory Config 12 GB GDDR5X
Memory Speed 10 Gbps
Boost Clock 1531 MHz
NVIDIA CUDA® Cores 3584
Transistors 12,000 million
TITAN X In Research
Deep Learning Augmented Reality
Machine Learning Image Recognition
Computer Vision Data Science
 Deep learning is a class of machine learning algorithms.
 Harder problems such as video understanding, image
understanding , natural language processing and Big data
will be successfully tackled by deep learning algorithms.
Each student select one of Benchmark Dataset from
https://goo.gl/DNQmtj
 Download the paper that describe the method used by the
authors.
 Make one page using word to summarize your selected paper.
Use MS Word
Send me e-mail to mloey@live.com with email subject “
Advanced Topics in CS2 – Task3 “
Put your Arabic name on word and email body
Finally, press Send
Deadline Next Lecture
facebook.com/mloey
mohamedloey@gmail.com
twitter.com/mloey
linkedin.com/in/mloey
mloey@fci.bu.edu.eg
mloey.github.io
www.YourCompany.com
© 2020 Companyname PowerPoint Business Theme. All Rights Reserved.
THANKS FOR
YOUR TIME

Lecture 3: Convolutional Neural Networks

Editor's Notes

  • #52 http://rodrigob.github.io/are_we_there_yet/build/classification_datasets_results.html