The Fundamentals of Machine Learning
The Fundamentals of Machine Learning
The Fundamentals of Machine Learning
OF MACHINE LEARNING
TABLE OF CONTENTS
3 WHAT IS MACHINE LEARNING?
8 HOW IT WORKS
16 CONTRIBUTORS
2
WHAT IS MACHINE LEARNING?
3
HOW DOES MACHINE LEARNING IMPACT OUR DAILY LIVES?
GOOD MACHINE LEARNING
If you asked someone on the street if they have ever heard of or utilized machine SHOULD NOT BE IN YOUR
learning, their answer would probably be no. What they don’t know is that FACE. IT SHOULD BE BEHIND
they’ve probably encountered it numerous times—just in one day. THE SCENES, TRACKING, AND
HELPING ACHIEVE GOALS
When you ask Siri what the weather forecast is, that’s machine learning. When MUCH MORE QUICKLY AND
you Google search something at work to help you do your job better or more EFFICIENTLY.
efficiently, you can thank machine learning. Another everyday example is our
spam folders—a machine learning algorithm is used to determine which emails Srinivas Bangalore
are inbox-worthy, and which are spam and don’t deserve attention. Similarly, —Director of Research and Technology
when Netflix suggests a show you should watch based on preference, it’s getting at Interactions
the suggestion from an algorithm.
4
BRIEF HISTORY OF
MACHINE LEARNING
From the 1950s to now, machine learning has significantly EXPLANATION BASED LEARNING 1981
developed. Below is a brief history of machine learning within the EBL, or Explanation Based Learning, was created in 1981 by
AI field. We show how the algorithms we described are motivated Gerald Dejong. This concept allowed a computer to analyze
by the need to solve very simple automation tasks, such as the training data and create a general rule it can follow by discarding
recognition of spoken words or written digits, and how AT&T unimportant data.
showed a strong leadership in this process.
MACHINE LEARNING RESEARCH GROUP 1985
THE TURING TEST 1950 Researchers from AT&T created the first research group for
In 1950, Alan Turing created the “Turing Test” to determine machine learning in 1985. They also began a series of machine
whether or not a computer was capable of real intelligence. In learning meetings that eventually turned into NIPS, the leading
order to pass the test, the computer had to be able to fool another conference on machine learning. This group was representative
human into believing it was also human. of the early machine learning community, breaking away from a
computer science field still mostly interested in expert systems.
THE FIRST COMPUTER PROGRAM 1952 These theoreticians were confronted with real world problems
Arthur Samuel created the first implementation of machine where machines had to replace humans in recognizing noisy
learning, the game of checkers, in 1952. The computer improved written digits: mainly check amounts and zip codes.
at the game the more it played by determining which moves
resulted in winning strategies, and incorporating those strategies AUTOMATED SPEECH RECOGNITION 1992
into the game. In 1992, Jay Wilpon (SVP of Natural Language Research at
Interactions) and a team of researchers at AT&T deployed the first
NEURAL NETWORKS FOR COMPUTERS 1957 nationwide automated speech recognition (ASR) using a machine
Frank Rosenblatt designed the first neural network for computers learning approach called Hidden Markov Models (HMMs). This
in 1957, which was meant to simulate the thought process of a saved billions of dollars in operating costs by spotting things like
human brain. collect calls.
5
1996 2014
1950
NATURAL LANGUAGE UNDERSTANDING 2001 MACHINE LEARNING AND SOCIAL MEDIA 2017
AT&T deployed natural language understanding in Interactive Acquired by Interactions in 2017, Digital Roots provides
Voice Response (IVR) systems in 2001, combining 3 of its machine companies with AI-based social media. Its technology allows
learning technologies: SVMs, HMMs, and Adaboost. brands to quickly filter, respond, and interact with followers
on social media.
DEEP LEARNING 2006
The concept of deep learning was successfully promoted,
increasing the power and accuracy of neural networks.
6
BRIEF 1950
Alan Turing created the “Turing Test”
HISTORY 1952
Arthur Samuel wrote the first
OF MACHINE computer learning program
LEARNING
1957
Frank Rosenblatt
1981 designed the first neural </>
Gerald Dejong introduced network for computers
the concept of Explanation
Based Learning (EBL)
1967
The “nearest neighbor”
algorithm was written
1985
AT&T created the first
research group dedicated 1992
to machine learning Jay Wilpon and team from AT&T
deployed the first nationwide
automated speech recognition
1996
Patrick Haffner and the team
at AT&T proposed the first
convolutional neural network
1997
Another group of AT&T
2011 researchers invented
A large community of the Adaboost algorithm
researchers started to work
on deep neural networks
2001
AT&T deployed natural language
understanding in IVR systems
2014
Google and Facebook made
machine learning, and in
particular deep learning, 2015 2017
their pivotal technology Interactions acquired Interactions acquired
AT&T’s Watson Digital Roots
7
HOW IT WORKS
The overall goal of machine learning is to build models that imitate and generalize
data. These models need to learn how to discriminate certain things to achieve a HOW DO YOU TELL THEM APART?
desired end result. Simply put, machine learning uses a variety of techniques, and
algorithms within these techniques, to reach a specific goal. Jay Wilpon explains how machine learning works
with an analogy of how algorithms decipher the
RECOGNIZING PATTERNS difference between types of fruits.
Machine learning learns from data, and uses that data to recognize patterns. Jay
Wilpon, Senior Vice President of Natural Language Research at Interactions, best
describes how machine learning works by using an analogy of fruits. For instance,
let’s assume someone handed you an orange and a grapefruit, and you’ve never
seen them before. How do you tell them apart? They’re both round, but the Size is one feature that can separate the two.
grapefruit is slightly bigger. You could then determine that size is one feature that
can separate the two. Now, let’s say someone hands you an apple. While the
shapes are similar, this fruit is red, triggering you to realize that color is another
potential differentiator. Finally, someone gives you a banana...now you can add
Color is another potential differentiator.
shape as another characterization.
This simple analogy is similar to how machine learning works. The job of machine
learning is not only to recognize that what it’s being handed is fruit, but also to
make sure that it is not calling a grapefruit a banana and vice versa.
You can add shape as another characterization.
8
MACHINE LEARNING TECHNIQUES
9
NEURAL NETWORKS
Deep neural networks (DNNs), also known as artificial neural networks (ANN),
represent a set of techniques used to build powerful learning systems. Unlike
algorithms such as SVMs and Adaboost, they add a number of “hidden” layers
that are used to extract intermediate representations. While invented in the
1980s, DNNs took off after 2010 thanks to powerful parallel hardware and
easy-to-use open source software. DNNs cover a huge range of different neural
architectures, the best known being:
10
THE IMPORTANCE OF
THE HUMAN ELEMENT
Regardless of how intelligent technology can be, at the end of the Interactions understands the crucial role the human element plays
day it will never be perfect. Humans can accelerate the process in artificial intelligence, which is why we’ve focused on integrating
of understanding by teaching the technology in real-time. For human intelligence into our technology. Our proprietary Adaptive
example, if machine learning comes across a piece of data it Understanding™ technology combines speech recognition, natural
cannot understand, a human can interfere and tell the technology language processing, and Human Assisted Understanding to
what it is, making it more accurate the next time it comes across provide our customers, and their customers, conversational and
that same piece of data. Technology does not have the same level engaging self-service. This enables continuous improvement and
of understanding as a human, and adding humans to the machine learning in live applications.
learning process can assist with decision making and allow the
technology to become more self-aware. This human touch can
personify machine learning, make it easier to relate to, and
in-turn, make it less intimidating.
11
WHO’S USING IT?
As previously mentioned, we encounter machine learning on a Trading floors - With its ability to efficiently assess data and
daily basis, whether we realize it or not. Aside from in our day- patterns, machine learning can assist with quick decision-making
to-day lives, industries from retail to government and more are in real-time.
depending on machine learning to get things done. Below is a
short list of how different industries are utilizing machine learning. Credit and risk management - Typically assessing credit risk is
This is not a complete list, as dozens of industries are using labor intensive and is prone to human-subjected errors. With
machine learning in a vast number of ways. machine learning, certain algorithms can help to provide mitigation
recommendations.
FINANCE
UTILITIES
With its quantitative nature, banking and finance are an ideal
application for machine learning. The technology is being used Utility companies can utilize machine learning in a number of ways,
in dozens of ways industry-wide, but here are a few of the most including uncovering hidden energy patterns, learning customer’s
commonly used: energy behaviors, and more.
12
HEALTHCARE OIL AND GAS
Diagnoses - Machine learning can analyze data and identify trends Energy sources - By analyzing different minerals in the ground,
or red flags within patients to potentially lead to earlier diagnoses machine learning provides the potential to discover new energy
and better treatments. sources.
Patient information - Data can be collected from a patient’s Streamlining oil distribution - Algorithms work to make oil
device to assess their health in real-time. distribution more efficient and cost-effective.
Drug discovery - Given its ability to detect patterns within data, Reservoir modeling - Certain machine learning techniques can
scientists are able to better predict drug side effects and results of focus on optimization of hydraulic fracturing, reservoir simulation,
drug experiments without actually performing them. and more.
Personalization - Machine learning allows online brands to Efficient transportation - Analysis of data can identify certain
suggest and advertise things you may like based on your browser patterns and trends to make routes more efficient for public
and search history. Brands use their collected data to give transportation, delivery companies, and more.
customers a unique and personalized experience.
13
CHALLENGES AND HESITATIONS
While machine learning has proved to have a profound impact across all
industries, there are still uncertainties and challenges regarding the technology.
First and foremost is the fear that technology will overcome humans. As we
discussed, technology is not perfect, and often needs the assistance of humans
to ensure accuracy. However, there is still a lot of fear and uncertainty regarding
the power of technology and its ability to become smarter than we are. At its
core, AI is a set of mathematical equations and algorithms that require human
training. This means that AI, and machine learning, are only as smart as we teach
them to be. When applied properly, AI is a perfect assistant to help humans
become more productive. Technology is not here to overcome us and overpower
us, but rather assist us and improve our quality of life.
A more technical issue with both machine learning and artificial intelligence is the
technology’s ability to handle unlabeled data. Because machine learning relies on
data to learn, it naturally requires a large amount of labeled data to work most
efficiently. However, there are many cases when data isn’t readily available or
is unlabeled. This makes creating algorithms more challenging. With on-going
research and new advancements, we’re training these systems to become
smarter and reach human-level accuracy, so that one day unlabeled data will be
just as sufficient as labeled data.
14
THE FUTURE
OF MACHINE
LEARNING
15
CONTRIBUTORS
JAY WILPON
SVP, NATURAL LANGUAGE RESEARCH, INTERACTIONS
With more than 150 published papers and patents in speech and natural language
research to his name, Jay Wilpon is one of the world’s pioneers and a chief evangelist
for speech and natural language technologies and services.
During his career, Jay has been a leading innovator for a number of industry-defining
voice enabled services, including AT&T’s How May I Help You service – the first
nationwide deployment of a true human-like spoken language understanding service.
Jay and his team are addressing the key challenges in speech, natural language
processing and multimodal dialog systems.
Jay has previously been awarded the distinguished honor of IEEE Fellow for his
leadership in the development of automatic speech recognition algorithms. For
pioneering leadership in the creation and deployment of speech recognition-based
services in the telephone network, Jay has also been awarded the honor of AT&T Fellow.
DAVID THOMSON
VP, SPEECH RESEARCH, INTERACTIONS
As Vice President of Speech Research, David manages Interactions’ R&D teams to
further Interactions’ goal to redefine the speech technology industry. David is at the
forefront of Interactions’ objective to create the most accurate, fastest, and highest
quality speech solutions. Prior to joining Interactions, David spent five years with
AT&T Labs, where he was responsible for the development of technology from speech
research. He has held senior executive-level positions at SpinVox, SpeechPhone, and
Fonix. He also spent 18 years at Lucent Technologies (now Alcatel-Lucent), where he
developed voice activated systems that have handled over 20 billion calls. David has
published 30 research papers and secured 11 patents in natural language research.
16
SRINIVAS BANGALORE
DIRECTOR OF RESEARCH AND TECHNOLOGY, INTERACTIONS
Dr. Srinivas Bangalore is currently the Director of Research and Technology at
Interactions. After receiving his PhD in Computer Science from The University of
Pennsylvania, he became a Principal Research Scientist at AT&T Labs—Research.
Dr. Bangalore has worked on many areas of Natural Language Processing including
Spoken Language Translation, Multimodal Understanding, Language Generation
and Question-Answering. He has co-edited three books on Supertagging, Natural
Language Generation, and Language Translation. He has authored over a 100 research
publications and holds over 100 patents in these areas. He has been awarded the Morris
and Dorothy Rubinoff award for outstanding dissertation, the AT&T Outstanding Mentor
Award, in recognition of his support and dedication to AT&T Labs Mentoring Program
and the AT&T Science & Technology Medal for technical leadership and innovative
contributions in Spoken Language Technology and Services. He has served on the
editorial board of Computational Linguistics Journal, Computer, Speech and Language
Journal and on program committees for a number of ACL and IEEE Speech Conferences.
PATRICK HAFFNER
LEAD INVENTIVE SCIENTIST, INTERACTIONS
Dr. Patrick Haffner has worked on machine learning algorithms since 1988. With
Yann LeCun, he was one of the pioneers in applying Neural Networks to speech and
image recognition, and led the deployment of the first NN used for an automation task
(check reading). With AT&T Labs Research, he was an expert in the learning algorithms
that enable data engineers to efficiently train machines using real world data, for
tasks ranging from language understanding to network monitoring. He was also an
expert advisor to the European Union for their funding programs on machine learning
and cognitive sciences. Dr. Haffner is a Lead Inventive Scientist at Interactions with
responsibility for managing the ever increasing variety of machine learning techniques
and software that an AI-driven company needs to use.
MICHAEL JOHNSTON
DIRECTOR OF RESEARCH AND INNOVATION, INTERACTIONS
Dr. Michael Johnston has over 25 years of experience in speech and language
technology. His research lies at the intersection of Natural Language Processing,
human-computer interaction, and spoken and multimodal dialog. More specifically,
his work focuses on the development of language and dialog processing techniques
that support spoken and multimodal interaction and the application of these to the
creation of novel systems and services. Dr. Johnston has over 50 technical papers and
32 patents in speech and language processing. Before joining Interactions, he held
positions at AT&T Labs Research, Oregon Graduate Institute and Brandeis University.
He is member of the board of AVIOS and editor and chair for the W3C EMMA
multimodal standard.
17
ABOUT INTERACTIONS
866.637.9049
18