Chapter1. Introduction To Deep Learning
Chapter1. Introduction To Deep Learning
HCM
VIỆN CÔNG NGHỆ THÔNG TIN, ĐIỆN, ĐIỆN TỬ
Chapter1.
INTRODUCTION TO
DEEP LEARNING
● Resurgence (1980s-1990s):
○ The development of the backpropagation algorithm in the 1980s was a crucial breakthrough, enabling the training of
multi-layer neural networks.
○ Yann LeCun's work on Convolutional Neural Networks (CNNs) for handwritten digit recognition (LeNet) demonstrated
the practical potential of deep learning.
Key Considerations:
● Data Availability
● Computational Resources
● "Black Box" Nature: Deep learning models can be
difficult to interpret, making it challenging to
understand how they arrive at their predictions.
The main problems addressed by deep learning
Deep learning tackles a wide range of complex problems, often categorized as follows:
1. Classification:
● Description:
○ Classification aims to assign an object to one of a set of predefined categories.
○ Examples: image classification (cats, dogs), text classification (spam, non-spam).
● Common techniques:
○ Convolutional Neural Networks (CNNs) for image classification.
○ Recurrent Neural Networks (RNNs) and Transformers for text classification.
2. Regression:
● Description:
○ Regression predicts a continuous value.
○ Examples: predicting house prices, predicting temperature.
● Common techniques:
○ Feedforward Neural Networks.
3. Object Detection:
● Description:
○ Object detection not only classifies objects but also locates them within an image or video.
○ Examples: detecting pedestrians in surveillance video, detecting traffic signs.
● Common techniques:
○ R-CNN (Region-based Convolutional Neural Networks), YOLO (You Only Look Once), SSD (Single Shot MultiBox
Detector).
The main problems addressed by deep learning
4. Image Segmentation:
● Description:
○ Image segmentation divides an image into meaningful regions, with each pixel assigned a label.
○ Examples: medical image segmentation to identify organs, satellite image segmentation for land analysis.
● Common techniques:
○ U-Net, FCN (Fully Convolutional Networks).
6. Generative Models:
● Description:
○ Generative models learn to create new data similar to the training data.
○ Examples: generating realistic images, generating music, generating text.
● Common techniques:
○ GAN (Generative Adversarial Networks), VAE (Variational Autoencoders).
Challenges in Deep Learning
1. Data Dependence:
● Requires massive, high-quality datasets, which are often
costly and difficult to obtain.
● Data biases lead to biased models.
2. Computational Demands:
● Training deep networks is computationally intensive,
needing powerful hardware.
● Long training times.
3. Model Complexity:
● "Black box" nature: lack of interpretability.
● Difficult hyperparameter tuning.
● Overfitting and underfitting issues.
4. Ethical Concerns:
● Potential for bias and unfairness.
● Privacy risks.
● Security risks from adversarial attacks.
5. Deployment difficulties:
● Hardware limitations on edge devices.
● Scalability problems.
Where to Find Deep Learning Exercises
● Google Developers Machine Learning Crash Course:
○ This course provides interactive exercises that allow you to experiment with neural network configurations.
○ It's a fantastic resource for beginners.
○ Here is the link: Neural networks: Interactive exercises | Machine Learning - Google for Developers
● Kaggle:
○ Kaggle is a platform that hosts data science competitions and provides datasets and notebooks.
○ You can find numerous deep learning exercises and tutorials on Kaggle.
○ Here is a good starting point: Exercise: Deep Neural Networks - Kaggle
● GitHub:
○ GitHub is a treasure trove of deep learning projects and exercises.
○ Search for repositories related to deep learning tutorials or specific tasks.
○ Here is an example: qiyuangong/Deep_Learning_Exercise: Deep Learning Exercise and Notebook - GitHub
Dataset
Where to Find Datasets:
Dataset is a collection of data. This data can come in various
● Kaggle: A popular platform with a wide variety of
forms, such as: Numbers, Text, Images, Audio, Video. The key
datasets.
purpose of a dataset is to provide the raw material that machine
● UCI Machine Learning Repository: A classic source of
learning algorithms use to learn patterns and make predictions.
datasets for machine learning research.
● Google Dataset Search: A search engine for datasets.
Key Characteristics:
● Hugging Face Datasets: A library and hub for easily
● Structure. Datasets can be structured (like tables in a
accessing and sharing datasets.
database), semi-structured (like JSON or XML files), or
● Government open data portals (e.g., data.gov).
unstructured (like raw text or images).
● Size. Datasets can range from small collections of a few data
points to massive repositories containing billions of records.
● Purpose. Datasets are used for various purposes, including:
Training machine learning models, testing model
performance, analyzing data to gain insights
●
Importance in Deep Learning. Deep learning models, in particular,
require large datasets to achieve high accuracy. The more data a
model has, the better it can learn complex patterns.
Datasets are often divided into three subsets:
● Training set: Used to train the model.
● Validation set: Used to fine-tune the model's
hyperparameters.
● Testing set: Used to evaluate the model's final performance.
Dataset
Some common datasets perfect for beginners diving into deep
learning
1. MNIST (Modified National Institute of Standards and
Technology database):
● Description: A dataset of handwritten digits (0-9).
● Use Case: Image classification.
● Why it's good:
○ Very simple and well-understood.
○ Excellent for learning basic CNNs.
○ Abundant tutorials and examples.
● Where to find it: Built into many deep learning libraries
(TensorFlow, PyTorch).
2. Fashion-MNIST:
● Description: Similar to MNIST, but with images of
fashion items (e.g., shoes, shirts).
● Use Case: Image classification.
● Why it's good:
○ Slightly more challenging than MNIST.
○ Still relatively simple and easy to work with.
○ Replaces the original MNIST dataset for a more
complex use case.
● Where to find it: Built into many deep learning libraries
(TensorFlow, PyTorch).
Dataset
3. CIFAR-10 (Canadian Institute For Advanced Research): 5. Boston Housing Dataset:
● Description: A dataset of 60,000 32x32 color images in 10 ● Description: Data about housing values in the Boston
classes (e.g., airplane, dog, cat). area, including features like crime rate and number of
● Use Case: Image classification. rooms.
● Why it's good: ● Use Case: Regression (predicting house prices).
○ A step up in complexity from MNIST. ● Why it's good:
○ Color images introduce more challenges. ○ Simple numerical dataset.
○ Very popular. ○ Good for learning basic regression models.
● Where to find it: Built into many deep learning libraries ○ Easy to understand.
(TensorFlow, PyTorch). ● Where to find it: Built into scikit-learn.
https://atcold.github.io/NYU-DLSP20/
https://www.youtube.com/watch?v=0bMe_vCZo30&list=PL80I41oVxglKcAHllsU0txr3OuTTaWX2v