Machine Learning &deep Learning in Python &R

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 48

_MACHINE LEARNING &DEEP

LEARNING IN PYTHON &R


(UDEMY)

A Report submitted in partial fulfilment of the requirement for the


award of degree of
Bachelor of Technology
In
Electronics and Communication Engineering

Submitted
By
SWATI
(03196302818)

MAHARAJA SURAJMAL INSTITUTE OF TECHNOLOGY


C-4, Janakpuri, New Delhi-58
Affiliated to Guru Gobind Singh Indraprastha University, Delhi
APRIL, 2021

Page 1 of 49
DECLARATION

I ,Swati, hereby declare that I have completed my two weeks winter training at
Udemy (one of the world’s leading online certification training providers) from 10
March 2021 to 23 March 2021 under the guidance of Vivek Sridhar. I have declared
that I have worked with full dedication during these two weeks of training and my
learning outcomes fulfill the requirements of training for the award of degree of
Bachelor of Technology (B.Tech.) in ECE , Maharaja Surajmal Institute of
Technology, New Delhi.

Name of Student: - Swati

Enrollment No: -03196302818

Page 2 of 49
CERTIFICATE

Page 3 of 49
ACKNOWLEDGEMENT

A research work owes its success from commencement to completion , do the people with
researchers at various stages . Let me in this page express my gratitude to all those who help me at
various stages of this study. First I would like to express my sincere gratitude and indebtedness to
Dr. Pradeep Sangwan (HOD, Department of Electronics and Communication Engineering ,
Maharaja Surajmal Institute of technology, New Delhi) unknown me to undergo the winter
training of 3 weeks at Udemy .

I am very grateful to our guide Mrs.Neetu Sehrawat, for the help provided in completion of the
report which was assigned to me. Without is friendly help and guidance it was difficult to develop
this report.

I pay my sincere thanks and gratitude to all the staff members of udemy for their support and for
making our training valuable and fruitful.

Name of Student - Swati

Page 4 of 49
ABSTRACT
Artificial Intelligence, Machine Learning and Deep Learning are the buzzwords that have been
able to grasp the interest of many researchers since various numbers of years.

Machine learning is an application of artificial intelligence (AI) that provides systems the
ability to automatically learn and improve from experience without being explicitly
programmed. Machine learning focuses on the development of computer programs that can
access data and use it learn for themselves

Future of Machine Learning


 Machine Learning can be a competitive advantage to any company be
it a top MNC or a startup as things that are currently being done
manually will be done tomorrow by machines.
 Machine Learning revolution will stay with us for long and so will be
the future of Machine Learning.
CONTENTS
1. Declaration………………………………………………………………………………..02
2. Certificate………………………………………………………………………………………………………………………………….03
3. Acknowledgement…………………………………………………………………………………………………………………….04.
4. Abstract………………………………………………………………………………………………………………………………… …05

5. Introduction to Machine Learning.....................................................................................07


5.1. A Taste of Machine Learning.....................................................................................07
5.2. Relation to Data Mining..............................................................................................07
5.3. Relation to Optimization.............................................................................................07
5.4. Relation to Statistics....................................................................................................08

6. Technology Learnt.............................................................................................................08
6.1. Introduction to Artificial Intelligence and Machine Learning....................................08
6.1.1. Definition of Artificial Intelligence.................................................................08
6.1.2. Definition of Machine Learning.......................................................................09
6.1.3. Machine Learning Algorithms.........................................................................10
6.1.4. Applications of Machine Learning...................................................................11
6.2. Techniques of Machine Learning................................................................................12
6.2.1. Supervised Learning.........................................................................................12
6.2.2. Unsupervised Learning.....................................................................................16
6.2.3. Semi- supervised Learning..............................................................................18
6.2.4. Reinforcement Learning...................................................................................19
6.2.5. Some Important Considerations in Machine Learning....................................19
6.3. Data Preprocessing.....................................................................................................20
6.3.1. Data Preparation..............................................................................................20
6.3.2. Feature Engineering.........................................................................................21
6.3.3. Feature Scaling.................................................................................................22
6.3.4. Datasets............................................................................................................24
6.3.5. Dimensionality Reduction with Principal Component Analysis......................24
6.4. Math Refresher............................................................................................................25
6.4.1. Concept of Linear Algebra...............................................................................25
6.4.2. Eigenvalues, Eigenvectors, and Eigen decomposition....................................30
6.4.3. Introduction to Calculus...................................................................................30
6.4.4. Probability and Statistics..................................................................................31
6.5. Supervised learning.....................................................................................................34
6.5.1. Regression........................................................................................................34
6.5.1.1. Linear Regression.................................................................................35
6.5.1.2. Multiple Linear Regression..................................................................35
6.5.1.3. Polynomial Regression.........................................................................36
6.5.1.4. Decision Tree Regression.....................................................................37
6.5.1.5. Random Forest Regression...................................................................37
6.5.2. Classification....................................................................................................38
6.5.2.1. Linear Models......................................................................................39
6.5.2.1.1. Logistic Regression..................................................................39
6.5.2.1.2. Support Vector machines.........................................................39
6.5.2.2. Nonlinear Models.................................................................................40
6.5.2.2.1. K-Nearest Neighbors (KNN)....................................................40
6.5.2.2.2. Kernel Support Vector Machines (SVM)................................40
6.5.2.2.3. Naïve Bayes..............................................................................41
6.5.2.2.4. Decision Tree Classification....................................................41
6.5.2.2.5. Random Forest Classification..................................................42
6.6. Unsupervised learning................................................................................................43
6.6.1. Clustering.........................................................................................................43
6.6.1.1. Clustering Algorithms..........................................................................43
6.6.1.2. K-means Clustering..............................................................................44
6.7. Introduction to Deep Learning....................................................................................45
6.7.1. Meaning and Importance of Deep Learning....................................................45
6.7.2. Artificial Neural Networks...............................................................................46
6.7.3. TensorFlow.......................................................................................................47
7. Reason for choosing Machine Learning.............................................................................47
8. Learning Outcome..............................................................................................................48
9. Bibliography.......................................................................................................................49
9.1. Content source.............................................................................................................49
9.2. Picture from.................................................................................................................49
9.3. Book referred..............................................................................................................49
1. Introduction
1.1.A Taste of Machine Learning
 Arthur Samuel, an American pioneer in the field of computer gaming and
artificial intelligence, coined the term "Machine Learning" in 1959.
 Over the past two decades Machine Learning has become one of the mainstays
of information technology.
 With the ever-increasing amounts of data becoming available there is good
reason to believe that smart data analysis will become even more pervasive as a
necessary ingredient for technological progress.
1.2. Relation to Data Mining

 Data mining uses many machine learning methods, but with different goals; on
the other hand, machine learning also employs data mining methods as
"unsupervised learning" or as a preprocessing step to improve learner accuracy.
1.3. Relation to Optimization
Machine learning also has intimate ties to optimization: many learning problems
are formulated as minimization of some loss function on a training set of
examples. Loss functions express the discrepancy between the predictions of the
model being trained and the actual problem instances.
1.4. Relation to Statistics

Michael I. Jordan suggested the term data science as a placeholder to call the
overall field.
Leo Breiman distinguished two statistical modelling paradigms: data model and
algorithmic model, wherein "algorithmic model" means more or less the machine
learning algorithms like Random forest.

2. Technology Learnt
2.1. Introduction to AI & Machine Learning
2.1.1. Definition of Artificial Intelligence
 Data Economy
 World is witnessing real time flow of all types structured and unstructured data
from social media, communication, transportation, sensors, and devices.
 International Data Corporation (IDC) forecasts that 180 zettabytes of data will
be generated by 2025.
 This explosion of data has given rise to a new economy known as the
Data Economy.
 Data is the new oil that is precious but useful only when cleaned and processed.
 There is a constant battle for ownership of data between enterprises to
derive benefits from it.
 Define Artificial Intelligence
Artificial intelligence refers to the simulation of human intelligence in machines that are
programmed to think like humans and mimic their actions. The term may also be applied to
any machine that exhibits traits associated with a human mind such as learning and problem-
solving.

2.1.2. Definition of Machine Learning


 Relationship between AI and ML

Machine Learning is an approach or subset of Artificial Intelligence that is based on the idea
that machines can be given access to data along with the ability to learn from it.
 Define Machine Learning
Machine learning is an application of artificial intelligence (AI) that provides systems the
ability to automatically learn and improve from experience without being explicitly
programmed. Machine learning focuses on the development of computer programs that can
access data and use it learn for themselves.
 Features of Machine Learning
 Machine Learning is computing-intensive and generally requires a large amount
of training data.
 It involves repetitive training to improve the learning and decision making of
algorithms.
 As more data gets added, Machine Learning training can be automated for
learning new data patterns and adapting its algorithm.
2.1.3. Machine Learning Algorithms
 Traditional Programming vs. Machine Learning Approach

 Traditional Approach
Traditional programming relies on hard-coded rules.
 Machine Learning Approach
Machine Learning relies on learning patterns based on sample data.

 Machine Learning Techniques


 Machine Learning uses a number of theories and techniques from
Data Science.

 Machine Learning can learn from labelled data (known as supervised


learning) or unlabeled data (known as unsupervised learning).
2.1.4. Applications of Machine Learning
 Image Processing
 Optical Character Recognition (OCR)
 Self-driving cars
 Image tagging and recognition
 Robotics
 Industrial robotics
 Human simulation
 Data Mining
 Association rules
 Anomaly detection
 Grouping and Predictions
 Video games
 Pokémon
 PUBG
 Text Analysis
 Spam Filtering
 Information Extraction
 Sentiment Analysis
 Healthcare
 Emergency Room & Surgery
 Research
 Medical Imaging & Diagnostics
2.2. Techniques of Machine Learning
2.2.1. Supervised Learning
 Define Supervised Learning
Supervised learning is the machine learning task of learning a function that maps an input to
an output based on example input-output pairs. It infers a function from labeled training data
consisting of a set of training examples.
In supervised learning, each example is a pair consisting of an input object (typically a
vector) and a desired output value (also called the supervisory signal).

 Supervised Learning Flow


 Data Preparation
Clean data
Label data (x, y)
Feature Engineering
Reserve 80% of data for Training (Train_X) and 20% for Evaluation
(Train_E)
 Training Step
Design algorithmic logic
Train the model with Train X
Derive the relationship between x and y, that is, y = f(x)
 Evaluation or Test Step
Evaluate or test with Train E
If accuracy score is high, you have the final learned algorithm y = f(x)
If accuracy score is low, go back to training step
 Production Deployment
Use the learned algorithm y = f(x) to predict production data.
The algorithm can be improved by more training data, capacity, or algo redesign.
 Testing the Algorithms
 Once the algorithm is trained, test it with test data (a set of data instances
that do not appear in the training set).
 A well-trained algorithm can predict well for new test data.
 If the learning is poor, we have an underfitted situation. The algorithm will
not work well on test data. Retraining may be needed to find a better fit.

 If learning on training data is too intensive, it may lead to overfitting–a


situation where the algorithm is not able to handle new testing data that it
has not seen before. The technique to keep data generic is called
regularization.
 Examples of Supervised Learning
 Voice Assistants
 Gmail Filters
 Weather Apps
 Types of Supervised Learning

 Classification
 Answers “What class?”
 Applied when the output has finite and discreet values Example: Social
media sentiment analysis has three potential outcomes, positive,
negative, or neutral
 Regression
 Answers “How much?”

 Applied when the output is a continuous number


 A simple regression algorithm: y = wx + b. Example: relationship
between environmental temperature (y) and humidity levels (x)
2.2.2. Unsupervised Learning
 Define Unsupervised Learning
Unsupervised learning is the training of machine using information that is neither classified
nor labeled and allowing the algorithm to act on that information without guidance.
Here the task of machine is to group unsorted information according to similarities, patterns
and differences without any prior training of data.
 Types of Unsupervised Learning

 Clustering
The most common unsupervised learning method is cluster analysis. It is used to find data
clusters so that each cluster has the most closely matched data.
 Visualization Algorithms
Visualization algorithms are unsupervised learning algorithms that accept unlabeled data and
display this data in an intuitive 2D or 3D format. The data is separated into somewhat clear
clusters to aid understanding.
 Anomaly Detection
This algorithm detects anomalies in data without any prior training.

2.2.3. Semi- supervised Learning


 Define Semi-supervised Learning
Semi-supervised learning is a class of machine learning tasks and techniques that also make
use of unlabeled data for training – typically a small amount of labeled data with a large
amount of unlabeled data.

Semi-supervised learning falls between unsupervised learning (without any labeled training
data) and supervised learning (with completely labeled training data).
 Example of Semi-supervised Learning
 Google Photos automatically detects the same person in multiple photos
from a vacation trip (clustering –unsupervised).
 One has to just name the person once (supervised), and the name tag
gets attached to that person in all the photos.
2.2.4. Reinforcement Learning
 Define Reinforcement Learning
Reinforcement Learning is a type of Machine Learning that allows the learning system to
observe the environment and learn the ideal behavior based on trying to maximize some
notion of cumulative reward.

It differs from supervised learning in that labelled input/output pairs need not be presented,
and sub-optimal actions need not be explicitly corrected. Instead the focus is finding a
balance between exploration (of uncharted territory) and exploitation (of current knowledge)
 Features of Reinforcement Learning
 The learning system (agent) observes the environment, selects and takes
certain actions, and gets rewards in return (or penalties in certain cases).
 The agent learns the strategy or policy (choice of actions) that maximizes
its rewards over time.
 Example of Reinforcement Learning
 In a manufacturing unit, a robot uses deep reinforcement learning to identify
a device from one box and put it in a container.
 The robot learns this by means of a rewards-based learning system,
which incentivizes it for the right action.
2.2.5. Some Important Considerations in Machine Learning
 Bias & Variance Tradeoff
 Bias refers to error in the machine learning model due to wrong assumptions.
A high-bias model will underfit the training data.
 Variance refers to problems caused due to overfitting. This is a result of
over- sensitivity of the model to small variations in the training data. A
model with
many degrees of freedom (such as a high-degree polynomial model) is likely
to have high variance and thus overfit the training data.
 Bias & Variance Dependencies
 Increasing a model’s complexity will reduce its bias and increase its variance.

 Conversely, reducing a model’s complexity will increase its bias and reduce its
variance. This is why it is called a tradeoff.
 What is Representational Learning
In Machine Learning, Representation refers to the way the data is presented. This often make
a huge difference in understanding.

2.3. Data Preprocessing


2.3.1. Data Preparation
 Data Preparation Process
 Machine Learning depends largely on test data.
 Data preparation involves data selection, filtering, transformation, etc.

 Data preparation is a crucial step to make it suitable for ML.


 A large amount of data is generally required for the most common forms
of ML.
 Types of Data
 Labelled Data or Training Data

 Unlabeled Data
 Test Data
 Validation Data
2.3.2. Feature Engineering
 Define Feature Engineering
The transformation stage in the data preparation process includes an important step known as
Feature Engineering.

Feature Engineering refers to selecting and extracting right features from the data that are
relevant to the task and model in consideration.
 Aspects of Feature Engineering
 Feature Selection
Most useful and relevant features are selected from the available data
 Feature Addition
New features are created by gathering new data
 Feature Extraction
Existing features are combined to develop more useful ones
 Feature Filtering
Filter out irrelevant features to make the modelling step easy

2.3.3. Feature Scaling


 Define Feature Scaling
 Feature scaling is an important step in the data transformation stage of
data preparation process.
 Feature Scaling is a method used in Machine Learning for standardization
of independent variables of data features.
 Techniques of Feature Scaling

 Standardization
 Standardization is a popular feature scaling method, which gives data
the property of a standard normal distribution (also known as
Gaussian distribution).
 All features are standardized on the normal distribution (a
mathematical model).
 The mean of each feature is centered at zero, and the feature column
has a standard deviation of one.
 Normalization
 In most cases, normalization refers to rescaling of data features
between 0 and 1, which is a special case of Min-Max scaling.

 In the given equation, subtract the min value for each feature from
each feature instance and divide by the spread between max and min.
 In effect, it measures the relative percentage of distance of each
instance from the min value for that feature.
2.3.4. Datasets
 Machine Learning problems often need training or testing datasets.
 A dataset is a large repository of structured data.
 In many cases, it has input and output labels that assist in Supervised Learning.
2.3.5. Dimensionality Reduction with Principal Component Analysis
 Define Dimensionality Reduction
 Dimensionality reduction involves transformation of data to new dimensions
in a way that facilitates discarding of some dimensions without losing any key
information.

 Define Principal Component Analysis (PCA)


 Principal component analysis (PCA) is a technique for dimensionality
reduction that helps in arriving at better visualization models.
 Applications of PCA
 Noise reduction
 Compression
 Preprocess

Supervised learning
2.3.6. Regression
 In statistical modeling, regression analysis is a set of statistical processes for
estimating the relationships among variables.
 It includes many techniques for modeling and analyzing several variables, when the
focus is on the relationship between a dependent variable and one or more
independent variables (or 'predictors').

More specifically, regression analysis helps one understand how the typical value of the
dependent variable (or 'criterion variable') changes when any one of the independent
variables is varied, while the other independent variables are held fixed.
2.5.1.1. Linear Regression

 Linear regression is a linear approach for modeling the relationship


between a scalar dependent variable y and an independent variable x.

 where x, y, w are vectors of real numbers and w is a vector of


weight parameters.
 The equation is also written as:

y = wx + b
 where b is the bias or the value of output for zero input
2.5.1.2. Multiple Linear Regression
It is a statistical technique used to predict the outcome of a response
variable through several explanatory variables and model the relationships
between them.
It represents line fitment between multiple inputs and one output, typically:

2.5.1.3. Polynomial Regression

 Polynomial regression is applied when data is not formed in a straight line.


 It is used to fit a linear model to non-linear data by creating new
features from powers of non-linear features.
2.5.1.4. Decision Tree Regression
o A decision tree is a graphical representation of all the possible solutions
to a decision based on a few conditions.
o Decision Trees are non-parametric models, which means that the number
of parameters is not determined prior to training. Such models will
normally overfit data.
o In contrast, a parametric model (such as a linear model) has a
predetermined number of parameters, thereby reducing its degrees of
freedom. This in turn prevents overfitting.

o max_depth –limit the maximum depth of the tree


o min_samples_split –the minimum number of samples a node must
have before it can be split
o min_samples_leaf –the minimum number of samples a leaf node must have
o min_weight_fraction_leaf –same as min_samples_leaf but expressed as
a fraction of total instances
o max_leaf_nodes –maximum number of leaf nodes
o max_features –maximum number of features that are evaluated for
splitting at each node
2.5.1.5. Random Forest Regression
 Ensemble Learning uses the same algorithm multiple times or a group
of different algorithms together to improve the prediction of a model.
 Random Forests use an ensemble of decision trees to perform
regression tasks.
2.3.7. Classification
It specifies the class to which data elements belong to.
It predicts a class for an input variable.
It is best used when the output has finite and discreet values.

There are 2 types of classification, binomial and multi-class.


2.5.2.1. Linear Models
2.5.2.1.1. Logistic Regression
 This method is widely used for binary classification problems. It can
also be extended to multi-class classification problems.
 A binary dependent variable can have only two values, like 0 or 1, win
or lose, pass or fail, healthy or sick, etc.

 The probability in the logistic regression is often represented by the


Sigmoid function (also called the logistic function or the S-curve)

2.5.2.1.2. Support Vector machines


 SVMs are very versatile and are also capable of performing linear or
nonlinear classification, regression, and outlier detection.
 They involve detecting hyperplanes which segregate data into classes.
 The optimization objective is to find “maximum margin hyperplane” that
is farthest from the closest points in the two classes (these points are
called support vectors).
2.5.2.2. Nonlinear Models
2.5.2.2.1. K-Nearest Neighbors (KNN)
K-nearest Neighbors algorithm is used to assign a data point to clusters
based on similarity measurement.

A new input point is classified in the category such that it has the
greatest number of neighbors from that category.

2.5.2.2.2. Kernel Support Vector Machines (SVM)


Kernel SVMs are used for classification of nonlinear data.
In the chart, nonlinear data is projected into a higher dimensional space via
a mapping function where it becomes linearly separable.
A reverse projection of the higher dimension back to original feature
space takes it back to nonlinear shape.
2.5.2.2.3. Naïve Bayes
 According to Bayes model, the conditional probability P(Y|X) can be
calculated as:

 This means you have to estimate a very large number of P(X|Y)


probabilities for a relatively small vector space X.

2.5.2.2.4. Decision Tree Classification


 The advantage of decision trees is that they require very little
data preparation.
 They do not require feature scaling or centering at all.
 They are also the fundamental components of Random Forests, one
of the most powerful ML algorithms.

 Start at the tree root and split the data on the feature using the
decision algorithm, resulting in the largest information gain (IG).
2.5.2.2.5. Random Forest Classification
 Random decision forests correct for decision trees' habit of overfitting to
their training set.

 Random forests or random decision forests are an ensemble learning


method for classification, regression and other tasks that operates by
constructing a multitude of decision trees at training time and outputting
the class that is the mode of the classes (classification) or mean prediction
(regression) of the individual trees.
2.4. Unsupervised learning
2.4.1. Clustering
2.6.1.1. Clustering Algorithms

 Clustering means
 Clustering is a Machine Learning technique that involves the
grouping of data points.

 Prototype Based Clustering


 Prototype-based clustering assumes that most data is located near
prototypes; example: centroids (average) or medoid (most frequently
occurring point)
 K-means, a Prototype-based method, is the most popular method for
clustering that involves:
 Training data that gets assigned to matching cluster based on
similarity
 Iterative process to get data points in the best clusters possible
2.6.1.2. K-means Clustering

 K-means Clustering Algorithm


Step 1: randomly pick k centroids
Step 2: assign each point to the nearest centroid
Step 3: move each centroid to the center of the respective cluster
Step 4: calculate the distance of the centroids from each point again
Step 5: move points across clusters and re-calculate the distance from
the centroid
Step 6: keep moving the points across clusters until the Euclidean
distance is minimized
 Elbow Method
 One could plot the Distortion against the number of clusters K. Intuitively, if K
increases, distortion should decrease. This is because the samples will be
close to their assigned centroids. This plot is called the Elbow method.
 It indicates the optimum number of clusters at the position of the elbow, the
point where distortion begins to increase most rapidly.
 Euclidian Distance
 K-means is based on finding points close to cluster centroids. The
distance between two points x and y can be measured by the squared
Euclidean distance between them in an m-dimensional space.

 Examples of K-means Clustering


 Grouping articles (example: Google news)
 Grouping customers who share similar interests
 Classifying high risk and low risk patients from a patient pool
2.5. Introduction to Deep Learning
2.5.1. Meaning and Importance of Deep Learning
 Define Deep Learning
Deep Learning is a specialized form of Machine Learning that uses supervised, unsupervised,
or semi-supervised learning to learn data representations.

It is similar to the structure and function of the human nervous system.


 Why Deep Learning
The vast availability of Big Data enables machines to be trained.
Experts have discovered multi-layered learning networks that can be leveraged for
deep learning as they learn in layers.
Scientists have figured out that high-performing graphics processing units (GPU)
can be used for deep learning.
 ML Vs Deep Learning

2.5.2. Artificial Neural Networks


 Deep learning relies on multiple layers of training.
 Artificial Neural Network is a computing system made up of a number of
simple, highly interconnected processing elements which process information
by their dynamic state response to external inputs.

 It is an interconnected group of nodes akin to the vast network of layers of


neurons in a brain.
2.5.3. TensorFlow
 TensorFlow is the open source Deep Learning library provided by Google.

 It allows development of a variety of neural network applications such as


computer vision, speech processing, or text recognition.
 It uses data flow graphs for numerical computations.
PROBLEM STATEMENT -
 In the graph , we have fitted line using the data calculated B1 and
B ,each of these points are plotted there and some of them are on
line but most of them is missing.
 When we taken out the total residual of same, we can’’t define
them because some of positive and some of negative residual.
 Then we can find out the Residual Sum of Square ( it represents
the total error).
RESULT : B1= 9.0997 , Intercept =-34.6592
Reason for choosing Machine Learning
 Learning machine learning brings in better career opportunities
 Machine learning is the shining star of the moment.
 Every industry looking to apply AI in their domain, studying machine learning
opens world of opportunities to develop cutting edge machine learning
applications in various verticals – such as cyber security, image recognition,
medicine, or face recognition.
 Several machine learning companies on the verge of hiring skilled ML
engineers, it is becoming the brain behind business intelligence.

 Machine Learning Jobs on the rise


 The major hiring is happening in all top tech companies in search of those
special kind of people (machine learning engineers) who can build a hammer
(machine learning algorithms).
 The job market for machine learning engineers is not just hot but it’s sizzling.
 Machine Learning Jobs on Indeed.com - 2,500+(India) & 12,000+(US)
3. Learning Outcome
 Have a good understanding of the fundamental issues and challenges of
machine learning: data, model selection, model complexity, etc.
 Have an understanding of the strengths and weaknesses of many popular
machine learning approaches.
 Appreciate the underlying mathematical relationships within and across Machine
Learning algorithms and the paradigms of supervised and un-supervised learning.
 Be able to design and implement various machine learning algorithms in a range
of real-world applications.
 Ability to integrate machine learning libraries and mathematical and statistical
tools with modern technologies
 Ability to understand and apply scaling up machine learning techniques and
associated computing techniques and technologies.
4. Bibliography
4.1. All Content used in this report is from

 https://www.wikipedia.org/
 https://towardsdatascience.com/
 https://www.expertsystem.com/
 https://www.udemy.org/
 https://www.edureka.co/
 https://subhadipml.tech/
 https://www.forbes.com/
 https://medium.com/
 https://www.google.com/
4.2. All Pictures are from

 https://www.google.com/
 https://www.wikipedia.org/
 https://www.youtube.com/
 https://www.edureka.co/
4.3. Book I referred are
 Hands-on Machine Learning with Scikit-learn & Tensorflow By Aurelien Geron
 Python Machine Learning by Sebastian Raschka
MAHARAJA SURAJMAL INSTITUTE OF TECHNOLOGY

Summer/Industrial Training Evaluation Form F05 (MSIT-EXM-P-02)

(Year 2018 -- 2022)

Details of the Student Details of the Organisation

Name - SWATI Name and address of organisation – Udemy,5 th


Roll No.-03196302818 Floor ,Wework ,Golf Course Road,DLF Phase
Branch and Semester- ECE and 6th Sem Gurugram,Harayana,122003.

Mobile No- 9971964204 Broader Area – Machine learning and deep learning
with python &R

Name of Instructor- Abhishek Bansal and Pukhraj parikh

E-mail ID – swatishrm921@gmail.com Designation and Contact No- Programming Instructor

Student Performance Record

No. of days Number of Curriculum Scheduled for the Curriculum actually covered
Scheduled for days student by the student
the training actually
attended

Introduction, Data Introduction, Data Processing


Processing, Linear regression, , Linear regression,
Week 1 7 7 Classification Models Classification Models

K-Nearest Neighbors K-Nearest Neighbors


Classifier, Decision Tree Classifier, Decision Tree
Week 2 7 7 .Introduction to deep .Introduction to deep
Learning Learning

Week 3

Swati
(Signature of the student)

Any comments or suggestions for the student performance during the training program(to be filled by
instructor)
…………………………………………………………………………………………………………………………………………………………………
…………………………………………………………………………………………………………………………………
Abhishek Bansal
(Signature of the Instructor)
along with Seal

Note: Every student has to fill and submit this Performa duly signed by his/her instructor to the
faculty-in-charge by first week of September.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy