By Tatek, Mar.
2025
1
Introduction
• Data science is a multi-disciplinary field that uses scientific
methods, processes, algorithms, and systems to extract
knowledge and insights from structured, semi-structured
and unstructured data.
After completing this chapter, the students will be able to:
Describe what data science is and the role of data scientists.
Differentiate data and information.
Describe data processing life cycle
Understand different data types from diverse perspectives
Describe data value chain in emerging era of big data.
Understand the basics of Big Data.
Describe the purpose of the Hadoop ecosystem components.
2
1. What are data and information?
• Data is a representation of facts, concepts, or
instructions in a formalized manner, which
should be suitable for communication,
interpretation, or processing, by human or
electronic machines.
• It can be described as unprocessed facts and
figures.
• It is represented with the help of characters such
as:
alphabets (A-Z, a-z),
digits (0-9) or
special characters (+, -, /, *, <,>, =, etc.).
3
• Information is the processed data on which
decisions and actions are based.
• It is data that has been processed into a form that is
meaningful to the recipient.
• Furtherer more, information is interpreted data;
created from organized, structured, and processed
data in a particular context.
4
2. Data Processing Cycle
• Data processing is the re-structuring or re-ordering
of data by people or machines to increase their
usefulness and add values for a particular purpose.
• Data processing consists of basic steps:
Input
Processing
Output
• These three steps constitute the data processing
cycle.
5
Data Processing Cycle
Input Processing Output
a) Input − in this step, the input data is prepared in
some convenient form for processing.
• The form will depend on the processing machine.
• For example:
when computers are used, the input data can be
recorded on any one of the several types of storage
medium, such as hard disk, CD, flash disk and so
on. 6
b) Processing − in this step, the input data is changed
to produce data in a more useful form.
• For example:
Interest can be calculated on deposit to a bank,
A summary of sales for the month can be calculated
from the sales orders.
c) Output − at this stage, the result of the processed
data is collected.
• The particular form of the output data depends on
the use of the data.
• For example:
Output data may be payroll for employees.
7
Data types & their representation
• Data types can be described from diverse
perspectives.
• In computer science and computer programming,
for instance, a data type is simply an attribute of
data that tells the compiler or interpreter how the
programmer intends to use the data.
8
Data types from Computer programming
perspective
• Common data types in computer programming
language include:
Integers (int)- is used to store whole numbers,
mathematically known as integers
Booleans (bool)- is used to represent restricted to one of
two values: true or false
Characters (char)- is used to store a single character
Floating-point numbers (float)- is used to store real
numbers
Alphanumeric strings (string)- used to store a
combination of characters and numbers 9
Data types from Data Analytics
perspective
• From a data analytics point of view, it is important
to understand that there are three common types
of data types or structures:
• Structured
• Semi-structured
• Unstructured data types
10
1. Structured Data
• Structured data is data that adheres to a pre-
defined data model and is therefore
straightforward to analyze.
• Structured data conforms to a tabular format with a
relationship between the different rows and
columns.
• Common examples of structured data are Excel files
or SQL databases.
• Each of these has structured rows and columns that
can be sorted.
11
2. Semi-structured Data
• Semi-structured data is a form of structured data
that does not conform with the formal structure of
data models associated with relational databases or
other forms of data tables.
• But nonetheless, contains tags or other markers to
separate semantic elements and enforce hierarchies
of records and fields within the data.
• Therefore, it is also known as a self-describing
structure.
• Examples of semi-structured data include E-mails,
JSON, HTML and XML are forms of semi-structured
data.
12
3. Unstructured Data
• Unstructured data is information that either does not
have a predefined data model or is not organized in a
pre-defined manner.
• Unstructured information is typically text-heavy but
may contain data such as dates, numbers, and facts as
well.
• This results in irregularities and ambiguities that make
it difficult to understand using traditional programs as
compared to data stored in structured databases.
• Common examples of unstructured data include audio,
video, images, social media posts, E-mail body, PDFs,
NoSQL databases, …. 13
4. Metadata (Data about Data)
• Metadata is a data type, not a separate data
structure, but it is one of the most important
elements for Big Data analysis and big data
solutions.
• Metadata is data about data. It provides additional
information about a specific set of data.
• In a set of photographs, for example, metadata
could describe when and where the photos were
taken.
• The metadata then provides fields for dates and
locations which, by themselves, can be considered
structured data.
14
Data value Chain
• Data Value Chain is introduced to describe the
information flow within a big data system as a
series of steps needed to generate value and useful
insights from data.
• The Big Data Value Chain identifies the following
key high-level activities:
Data Acquisition
Data Analysis
Data Curation
Data Storage
Data Usage
15
Table: Data value Chain
16
1. Data Acquisition
• Data acquisition is the process of gathering,
filtering, and cleaning data before it is put in a data
warehouse or any other storage.
• Data acquisition is one of the major big data
challenges in terms of infrastructure requirements.
• The infrastructure required to support the
acquisition of big data:
Must deliver low, predictable latency in both capturing
data and in executing queries.
Be able to handle very high transaction volumes, often
in a distributed environment.
Support flexible and dynamic data structures. 17
2. Data Analysis
• Data Analysis is concerned with making the raw
data acquired to use in decision-making as well as
domain-specific usage.
• Data analysis involves exploring, transforming, and
modeling data with the goal of highlighting relevant
data, synthesizing and extracting useful hidden
information with high potential from a business
point of view.
• Related areas include data mining, business
intelligence, and machine learning.
18
3. Data Curation
• Data Curation is the active management of data over
its life cycle to ensure it meets the necessary data
quality requirements for its effective usage.
• Data curation processes can be categorized into
different activities such as:
Content creation
Selection
Classification
Transformation
Validation
Preservation
• Data curators (also known as scientific curators or
data annotators) hold the responsibility of ensuring
that data are trustworthy, discoverable, accessible,
reusable and fit their purpose.
19
4. Data Storage
• Data Storage is the persistence and management of
data in a scalable way that satisfies the needs of
applications that require fast access to the data.
• Relational Database Management Systems (RDBMS)
have been the main, and almost unique, a solution to
the storage paradigm for nearly 40 years.
• However, the ACID (Atomicity, Consistency, Isolation,
and Durability) properties that guarantee database
transactions lack flexibility with regard to schema
changes and the performance and fault tolerance
when data volumes and complexity grow, making them
unsuitable for big data scenarios.
• NoSQL technologies have been designed with the
scalability goal in mind and present a wide range of
solutions based on alternative data models.
20
5. Data Usage
• Data Usage covers the data-driven business
activities that need access to data, its analysis, and
the tools needed to integrate the data analysis
within the business activity.
• Data usage in business decision-making can
enhance competitiveness through the reduction of
costs, increased added value, or any other
parameter that can be measured against existing
performance criteria.
21
Basic concepts of big data
• Big data is a blanket term for the non-traditional
strategies and technologies needed to gather,
organize, process, and gather insights from large
datasets.
• While the problem of working with data that
exceeds the computing power or storage of a single
computer is not new, the pervasiveness, scale, and
value of this type of computing have greatly
expanded in recent years.
22
What is big data?
• Big data is the term for a collection of data sets so
large and complex that it becomes difficult to
process using on-hand database management tools
or traditional data processing applications.
• In this context, a “large dataset” means a dataset
too large to process or store with traditional tooling
or on a single computer.
• This means that the common scale of big datasets is
constantly shifting and may vary significantly from
organization to organization.
23
• Big data is characterized by 4V and more:
Volume: large amounts of data. Zeta bytes/Massive
datasets
Velocity: Data is live streaming or in motion
Variety: data comes in many different forms from
diverse sources
Veracity: can we trust the data? How accurate is it?
etc.
24
25
Clustered Computing and Hadoop Ecosystem
Clustered Computing
• Because of the quantities of big data,
individual computers are often
inadequate for handling the data at most
stages.
• To better address the high storage and
computational needs of big data,
computer clusters are a better fit.
26
• Big data clustering software combines the resources
of many smaller machines, seeking to provide a
number of benefits:
1. Resource Pooling
• Combining the available resources such as:
Storage space to hold data
CPU
Memory
• Processing large datasets requires large amounts of
all three of the resources.
2. High Availability
• Clusters can provide varying levels of fault tolerance
and availability guarantees to prevent hardware or
software failures from affecting access to data and
processing.
27
3. Easy Scalability
• Clusters make it easy to scale horizontally by
adding additional machines to the group.
• This means the system can react to changes
in resource requirements without expanding
the physical resources on a machine.
• Using clusters requires a solution for
managing cluster membership, coordinating
resource sharing, and scheduling actual
work on individual nodes.
• Cluster membership and resource allocation
is handled by software like Hadoop’s YARN
(Yet Another Resource Negotiator).
28
Hadoop and its Ecosystem
• Hadoop is an open-source framework
intended to make interaction with big
data easier.
• Hadoop is a framework that allows for
the distributed processing of large
datasets across clusters of computers
using simple programming models.
• It is inspired by a technical document
published by Google. 29
• The four key characteristics of Hadoop are:
• Economical: Its systems are highly
economical as ordinary computers can be
used for data processing.
• Reliable: It is reliable as it stores copies of
the data on different machines and is
resistant to hardware failure.
• Scalable: It is easily scalable both,
horizontally and vertically. A few extra nodes
help in scaling up the framework.
• Flexible: It is flexible and you can store as
much structured and unstructured data as
you need to and decide to use them later.
30
• Hadoop has an ecosystem that has
evolved from its four core components:
Data management
Access
Processing
Storage
31
• Hadoop is continuously growing to meet the needs
of Big Data. It comprises the following components
and many others:
HDFS: Hadoop Distributed File System
YARN: Yet Another Resource Negotiator
MapReduce: Programming based Data Processing
Spark: In-Memory data processing
PIG, HIVE: Query-based processing of data services
HBase: NoSQL Database
Mahout, Spark MLLib: Machine Learning algorithm
libraries
Solar, Lucene: Searching and Indexing
Zookeeper: Managing cluster
Oozie: Job Scheduling 32
Fig. Hadoop Ecosystem 33
• Big Data Life Cycle with Hadoop
1. Ingesting data into the system
• The first stage of Big Data processing is
Ingest.
• The data is ingested or transferred to
Hadoop from various sources such as
relational databases systems, or local
files.
• Sqoop transfers data from RDBMS to
HDFS, whereas Flume transfers event
data. 34
2. Processing the data in storage
• The second stage is Processing.
• In this stage, the data is stored and
processed.
• The data is stored in the distributed file
system, HDFS, and the NoSQL distributed
data, HBase.
• Spark and MapReduce perform data
processing.
35
3. Computing and analyzing data
• The third stage is to Analyze.
• Here, the data is analyzed by processing
frameworks such as Pig, Hive, and
Impala.
• Pig converts the data using a map and
reduce and then analyzes it.
• Hive is also based on the map and reduce
programming and is most suitable for
structured data. 36
4. Visualizing the results
• The fourth stage is Access, which is
performed by tools such as Hue and
Cloudera Search.
• In this stage, the analyzed data can be
accessed by users.
37
Thanks
Questions?
38