CH-2 Data Science Emerging Technology

Download as pptx, pdf, or txt
Download as pptx, pdf, or txt
You are on page 1of 20

Chapter 2: Data Science

learning contents(points to be covered) are :


An Overview of Data Science

What are data and information

Data types and their representation

Data value Chain

Basic concepts of big data


Data Science 1
Overview of Data Science
• What is data science, data , information and big data?

 data science is:-

 defined as a multi-disciplinary field that uses scientific methods, processes,


algorithms, and systems in order to extract knowledge and insights from structured,
semi-structured and unstructured data.
 is much more than simply analyzing data.
 It offers a range of roles and requires a range of skills

 In case of academic discipline and profession, data science continues to evolve as one
of the most promising and in-demand career paths for skilled professionals.
Data Science 2
cont. ..
 Data professionals understand that they must advance past the traditional skills of
analyzing large amounts of data, data mining, and programming skills.
 Data scientists need to be curious and result-oriented, with exceptional industry-
specific knowledge and communication skills that allow them to explain highly
technical results to their non-technical counterparts.

Data Science 3
What are data and information

• what is Data and information ?


 Data can be defined as: - a representation of facts, figures, concepts, or
instructions in a formalized manner,

• which should be suitable for communication, interpretation, or


processing, by human or electronic machines.
 It can be described as unprocessed facts and figures.

 Is represented with the help of characters such as alphabets (A-Z, a-z), digits (0-
9) or special characters (+, -, /, *, <,>, =, etc.).

Data Science 4
cont. ..
 Information is defenied as :-
 the processed data on which decisions and actions are based.
 It is data that has been processed into a form that is meaningful to the recipient.
 or it is interpreted data; created from organized, structured, and processed data in
a particular context.

Data Science 5
Data Processing Cycle
• Data processing is the re-structuring or re-ordering of data by people or machines to
increase their usefulness and add values for a particular purpose.
• Basic steps that are consisted in the data processing :-

I. input
II. processing and
Data Processing Output
III. output.

Question
1. Define the above terms(input, processing and output) and discuss the main
differences between data and information with examples ?
Data Science 6
Data types and their representation
• Data types can be described from diverse perspectives. here some of the perspectives are:-

i. Data types from Computer programming perspective:- different languages may use
different terminology for the notion of data types. Common data types include:
 Integers(int)- is used to store whole numbers, mathematically known as integers
• Booleans(bool)- is used to represent restricted to one of two values: true or false
• Characters(char)- is used to store a single character
• Floating-point numbers(float)- is used to store real numbers
• Alphanumeric strings(string)- used to store a combination of characters and numbers

Data Science 7
cont. ..
ii. Data types from Data Analytics perspective :- interms of Analtic perspective there are
three common types of data types or structures. those are :
I. Structured Data:- is data that adheres to a pre-defined data model and is
therefore straightforward to analyze.
 It conforms to a tabular format with a relationship between the different rows
and columns.
 e.g Excel files or SQL databases (which has structured rows and columns that
can be sorted)
II. Semi-structured Data:- is a form of structured data that does not
conform with the formal structure of data models associated with
relational databases or other forms of data tables.
 It is also known as a self-describing structure. (why ?)
 e.g JSON , XML, sensor data
III. Unstructured data types :- is information that does not have either a
predefined data model or is not organized in a pre-defined manner.
Data Science 8
cont. ..
 Unstructured information is typically text-heavy but may contain data such as dates,
numbers, and facts as well.
 Do you think that understanding Unstructured data using traditional programs as
compared to data stored in structured databases easy?
 e.g audio, video files or No_x0002_SQL databases

Data Science 9
cont. ..
iii. Metadata :- is simply defined as data about data.
 From a technical point of view, this is not a separate data structure, but it is one
of the most important elements for Big Data analysis and big data solutions.
 provides additional information about a specific set of data.
for example, In a set of photographs, it describe when and where the photos were
taken. and also it provides fields for dates and locations which, by themselves, can
be considered structured data.
Because of this reason, metadata is frequently used by Big Data solutions for initial
analysis.
Compare metadata with structured, unstructured and semi-structured data ?
Data Science 10
Data value Chain
• It is introduced to describe the information flow within a big data system as a
series of steps needed to generate value and useful insights from data.
• The Big Data Value Chain identifies the following key high-level activities:

Data Science 11
cont. ..
 Data Acquisition :- is the process of gathering, filtering, and cleaning data before it is
put in a data warehouse or any other storage solution on which data analysis can be carried
out.
 is one of the major big data challenges in terms of infrastructure requirements. why
explain ?
 Data Analysis :- making the raw data acquired amenable to use in decision-making as
well as domain-specific usage.
 it involves exploring, transforming, and modeling data with the goal of highlighting
relevant data, synthesizing and extracting useful hidden information with high potential
from a business point of view. Data Science 12
cont. ..
 Data Curation :- It is the active management of data over its life cycle to ensure it meets the necessary
data quality requirements for its effective usage.

 it contains different activities such as content creation, selection, classification,


transformation, validation, and preservation.

 Data Storage :- is the persistence and management of data in a scalable way


that satisfies the needs of applications that require fast access to the data.
 Data Usage:- It covers the data-driven business activities that need access to
data, its analysis, and the tools needed to integrate the data analysis within the
business activity.
Data Science 13
Basic concepts of big data
• Big data is the term for a collection of data sets so large and complex .
• it becomes difficult to process using on-hand database management tools or
traditional data processing applications.
• it is characterized by 3V and more:
 Volume: large amounts of data (Zeta bytes)
 Velocity: Data is live streaming or in motion
 Variety: data comes in many different forms from
diverse sources
 Veracity: can we trust the data? How accurate is it?

Figure 2.4 Characteristics of big data


Data Science 14
Clustered Computing and Hadoop Ecosystem

 clusted computing :-
• Big data clustering software combines the resources of many smaller machines,
seeking to provide a number of benefits:
 Resource Pooling:
 High Availability:
 Easy Scalability:

Data Science 15
Hadoop and its Ecosystem

• Hadoop is an open-source framework intended to make interaction with big data easier.
 is a framework that allows for the distributed processing of large datasets
across clusters of computers using simple programming models.
 characteristics of Hadoop are:-

i. Economical:- Its systems are highly economical.

ii. Reliable:- stores copies of the data on different machines and is resistant to
hardware failure.
iii. Scalable:- is easily scalable both, horizontally and vertically.

iv. Flexible:- you can store as much structured


Data Science and unstructured data as you 16
cont. ..
• It comprises the following components and many others:
• Hadoop has an ecosystem that has
 HDFS: Hadoop Distributed File System
evolved from its four core  YARN: Yet Another Resource Negotiator
components:  MapReduce: Programming based Data Processing

 data management,  Spark: In-Memory data processing

 PIG, HIVE: Query-based processing of data services


 access,
 HBase: NoSQL Database
 processing, and  Mahout, Spark MLLib: Machine Learning algorithm

 storage. libraries
 Solar, Lucene: Searching and Indexing
 Zookeeper: Managing cluster

 Science
Data Oozie: Job Scheduling 17
cont. ..

Data Science 18
Big Data Life Cycle with Hadoop
• There are different stages of Big Data processing, some of them are:-

I. Ingesting data into the system :- data is ingested or transferred to Hadoop from
various sources such as relational databases, systems, or local files.
 Sqoop transfers data from RDBMS to HDFS, whereas Flume transfers event
data.
II. Processing the data in storage:- the data is stored and processed.
 The data is stored in the distributed file system, HDFS, and the NoSQL
distributed data, HBase. Spark and MapReduce perform data processing.

Data Science 19
cont. ..
III. Computing and analyzing data:- data is analyzed by processing
frameworks such as Pig, Hive, and Impala.

 Pig converts the data using a map and reduce and then analyzes it.
Hive is also based on the map and reduce programming and is
most suitable for structured data.
iv. Visualizing the results:- is performed by tools such as Hue and Cloudera Search.

 In this stage, the analyzed data can be accessed by users.

Data Science 20

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy