0% found this document useful (0 votes)
21 views7 pages

Untitled Document

Uploaded by

B2D Bored boi
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
21 views7 pages

Untitled Document

Uploaded by

B2D Bored boi
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 7

Chapter -1

“Machine Learning is a discipline that deals with programming the


systems so as to make them automatically learn and improve with
experience. Here, learning implies understanding the input data and
taking informed decisions based on the supplied data”. In simple words,
Machine Learning is a subset of AI which predicts results based on
incoming data. The utilities of ML are numerous. So as to detect spam
emails, forecast stock prices or to project class attendance one can
achieve results by means of earlier collected spam messages, previous
price history records or procure 5 years or more attendance data of a
class. ML will predict the results based upon previous data base
experience available with it.

AI is a technology to create intelligent machines that can recognize


human speech, can see (vision), assimilate knowledge, strategize and
solve problems as humans do. Broadly, AI entails all those technologies
or fields that aim to create intelligent machines.

Data can be defined as a representation of facts or instructions about


some entity (students, school, sports, business, animals etc.) that can
be processed or communicated by human or machines. Data is a
collection of facts, such as numbers, words, pictures, audio clips,
videos, maps, measurements, observations or even just descriptions of
things.

Structured data’ is most often categorized as quantitative data, and it's


the type of data most of us work with every day. Structured data has
predefined data types and format so that it fits well in the column/ fields
of database or spreadsheet. They are highly organised and easlyi
anaylsed

Unstructured data’ is most often categorized as qualitative data, and it


cannot be processed and analysed using conventional relational
database (RDBMS) methods.
Supervised Learning occurs in the presence of a supervisor or a teacher.
We train the machine with labeled data (i.e. some data is already tagged
with correct answer). It is then compared to the learning which takes
place in the presence of a supervisor or a teacher. A supervised learning
algorithm learns from labelled training data, and then becomes ready to
predict the outcomes for unforeseen data.

Many a times, perfectly labelled data sets are hard to find. In such
situations, data used to train the machine are neither labelled nor
classified. Unsupervised learning is a ML technique where we don’t need
to supply labelled data, instead we allow the machine learning model
(algorithm) to discover the patterns on its own. The task of the machine
is to assemble unsorted information according to resemblances,
patterns and variances without any former training of data.

“Reinforcement learning (RL)” as an area of machine learning concerned


with how software agents ought to take actions in an environment in
order to maximize some notion of cumulative reward. Reinforcement
learning is one of three basic machine learning paradigms, alongside
supervised learning and unsupervised learning

Deep Learning is inspired from human brain and the neurons in the
human brain

‘Artificial Neural Networks (ANN) can be described as layers of software


units called neurons (also called node), connected with different
neurons in a layered manner. These networks transform data from one
neuron to another neuron until they can classify it as an output. Neural
network is again a technique to build a computer program that learns
from data.

Input Node: This is the layer where information or initial data from the
real world gets introduced into the neural network. The information is
then passed onto the hidden node where computations can begin.

Hidden Node: There is no connection to the real world at this stage. This
is the point where the machine uses the information received from the
input node, it carries out computation and processing on it. There can
be more than one hidden layer.
Output Node: This is the final stage where the computations conclude,
and data is made available to the output layer from where it gets
transferred back into the real-world environment.

Deep learning is a branch of machine learning which is completely


based on artificial neural networks, as neural network mimics the human
brain so deep learning is also a kind of imitation of the human brain. In
deep learning, we don’t need to explicitly program everything”. It is
important to know that in deep learning, we do not need to explicitly
program everything

Chapter - 2’
. Rule – based Chatbot This is the simpler form of Chatbot which follows
a set of pre-defined rules in responding to user’s questions. For
example, a Chatbot installed at a school reception area, can retrieve data
from the school’s archive to answer queries on school fee structure,
course offered, pass percentage, etc

Chatbots are advanced forms of chatter-bots capable of holding


complex conversations in realtime. They process the questions (using
neural network layers) before responding to them. AI based Chatbots
also learn from previous experience and reinforced learning and it keeps
on evolving
the technology which enables the machines (software) to understand
and process the natural language (of humans), is called natural
language processing (NLP

in very simple terms CV is a field of study that enables the computers to


“see”. It is a subfield of AI and involves extraction of information from
digital images like the videos and photographs, analysing, and
understanding the content

The number in the matrix ranges between 0 to 255


CHAPTER -3
Ethics is defined as the discipline dealing with moral obligations and
duties of humans. It is a set of moral principles which govern the
behaviour and actions of individuals or groups

The ethics of AI is the part of the ethics technology specific to robots


and other artificially intelligent beings. It can be divided into roboethics,
a concern with the moral behaviour of humans as they design,
construct, use and treat artificially intelligent beings, and machine
ethics, which is concerned with the moral behaviour of artificial moral
agents (AMAs). With regard to artificial general intelligence (AGIs),
preliminary work has been conducted on approaches to integrating AGIs
which are full ethical agents with existing legal and social frameworks

AI bias, is a phenomenon that occurs when an algorithm produces


results that are systematically prejudiced towards certain gender,
language, race, wealth etc. and therefore produces skewed or leaned
output. Algorithms can have built-in biases because they are created by
individuals who have conscious or unconscious preferences that may
go undiscovered until the algorithms are used publically
the sources of AI bias
1. Data AI system are as good as the data we put into them. And putting
biased/skewed data into the system is the reason for AI bias. AI system
don’t have understanding of whether their training data is right or wrong
and have enough representation from a broader base.

Algorithm While algorithm, in itself doesn’t inject bias but can amplify
the biases caused by data. For example, an image classifier trained on
the images available in public domain - which happens to show more
women in the kitchen as opposed to men. AI algorithms are designed to
maximize accuracy. So an AI algorithm may decide that all people in the
kitchen are women, despite the fact the training data has some images
of men in the kitchen. It thereby incorporates a gender bias in the AI
system.

People The last issue is with the people who are developing the AI
system i.e. engineers, scientists, developers etc. They aim to get the
most accurate results with the available data. They are often lesser
focused on the broader context. It is rightly said that ethics and bias are
not the problem of the machine but that of the humans behind the
machine.

Chapter -5
Storytelling is defined as the art of narrating stories to engage an
audience. It originated in the ancient times with visual stories, such as
cave drawings, and then shifted to oral traditions, in which stories were
passed down from generation to generation by word of mouth. Later,
words formed into narratives, that included written, printed and typed
stories. Written language, as it is seen now, was arguably the first
technological innovation, that gave us as a species the power to convey
story in a physical format, and thus visualize, archive and share that
data with community members and future generations. It encourages
people to make use of their imagination and inventiveness (creativity) to
express themselves (verbal skills) which makes it a lot more than just a
recitation of facts and events.

The steps involved in telling an effective data story are given below:
1 Understanding the audience
2- Choosing the right data and visualisations
3- Drawing attention to key information
4- Developing a narrative
5-Engaging your audience

Data storytelling is a structured approach for communicating insights


drawn from data, and invariably involves a combination of three key
elements: data, visuals, and narrative. When the narrative is
accompanied with data, it helps to explain the audience what’s
happening in the data and why a particular insight has been generated.
When visuals are applied to data, they Purpose: To provide insight into
data storytelling and how it can bring a story to life. Say: “Now that you
have understood what storytelling is and why it is needed, let us learn
about a storytelling of a different kind - the art of data storytelling and in
the form of a narrative or story.” 146 can enlighten the audience to the
insights that they wouldn’t perceive without the charts or graphs.

Chapter - 8
Regression can be defined as a method or an algorithm in Machine
Learning that models a target value based on independent predictors. It
is essentially a statistical tool used in finding out the relationship
between a dependent variable and an independent variable. This method
comes to play in forecasting and finding out the cause and effect
relationship between variables.

Correlation is a measure of the strength of a linear relationship between


two quantitative variables

1 is a perfect positive correlation 0 is no correlation (the values don't


seem linked at all) -1 is a perfect negative correlation

Crosstabs Cross tabs help us establish a relationship between two


variables. This relationship is exhibited in a tabular form.

A scatter plot (aka scatter chart, scatter graph) uses dots to represent
values for two different numeric variables. The position of each dot on
the horizontal and vertical axis indicates values for an individual data
point. Scatter plots are used to observe relationships between variables

The Pearson correlation coefficient is used to measure the strength of a


linear association between two variables, where the value r = 1 means a
perfect positive correlation and the value r = -1 means a perfect negative
correlation. So, for example, you could use this test to find out whether
people's height and weight are correlated (the taller the people are, the
heavier they're likely to be)

Correlation and causation are terms which are mostly misunderstood


and often used interchangeably. Understanding both the statistical terms
is very important not only to make conclusions but more importantly,
making correct conclusions at the end. In this section we will
understand why correlation does not imply causation
Correlation is a statistical technique which tells us how strongly the pair
of variables are linearly related and change together. It does not tell us
why and how behind the relationship but it just says the relationship
exists.
Causation takes a step further than correlation. It says any change in the
value of one variable will cause a change in the value of another
variable, which means one variable makes the other happen. It is also
referred to as cause and effect.

A contingency table provides a way of portraying data that can facilitate


calculating probabilities. The table helps in determining conditional
probabilities quite easily. The table displays sample values in relation to
two different variables that may be dependent or contingent on one
another. Later on, we will use contingency tables again, but in another
manner

correlation is a measure of how closely two variables move together.


Pearson’s correlation coefficient is a common measure of correlation,
and it ranges from +1 for two variables that are perfectly in sync with
each other, to 0 when they have no correlation, to -1 when the two
variables are moving opposite to each other.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy