deep learning u3
deep learning u3
• Definition: CNNs are a class of deep learning neural networks commonly used for
analyzing visual imagery. They are specifically designed to process pixel data and
recognize patterns.
• Use Cases: Image classification, object detection, facial recognition, and more.
2. CNN Terminologies
• Stride: Refers to the number of pixels by which the filter slides over the input feature
map. A larger stride reduces the spatial dimensions of the output.
• Padding: Adding extra pixels around the border of the input feature map to control the
spatial dimensions of the output. Common types are 'valid' (no padding) and 'same'
(padding to keep output dimensions same as input).
• Convolutional Kernels: Small, learnable filters that slide over the input to detect
features like edges, textures, and patterns.
3. Types of Layers
• Pooling Layer: Reduces the spatial dimensions of the feature map, usually after a
convolutional layer.
• Fully Connected Layer: Connects every neuron in one layer to every neuron in the next
layer, used at the end of the network to make predictions.
4. Visualizing CNN
• Tools and techniques like saliency maps, gradient-weighted class activation mapping
(Grad-CAM), and activation maximization are used to understand what CNNs learn and
how they make decisions.
5. CNN Examples
• LeNet: One of the earliest CNN architectures, primarily used for handwritten digit
recognition.
• AlexNet: Introduced deeper layers and ReLU activation, won the ImageNet challenge in
2012.
• VGGNet: Known for its deep but simple architecture with 16-19 layers, using very small
(3x3) convolution filters.
• GoogLeNet (Inception): Introduced the Inception module that uses multiple filter sizes
to capture different features at various scales.
• RCNN: A series of networks for object detection that combine region proposals with
CNNs for accurate detection.
6. Deep Dream
7. Deep Art
• Uses neural networks to apply artistic styles to images, creating artwork that mimics
famous styles (e.g., Van Gogh, Picasso).
8. Regularization Techniques
• Dropout: Randomly sets a fraction of input units to zero during training to prevent
overfitting.
• Drop Connect: Similar to Dropout but drops connections between neurons rather than
the neurons themselves.
• Unit Pruning: Removes unimportant neurons to reduce the complexity of the model.
• Stochastic Pooling: Uses a probabilistic approach to pooling, selecting values based on
their magnitude.
• Injecting Noise in Input: Adds random noise to input data to make the model more
robust.
• Early Stopping: Stops training when the model's performance on a validation set starts
to degrade.
• Weight Decay: Adds a penalty term to the loss function proportional to the magnitude
of the weights to prevent large weights.