0% found this document useful (0 votes)
32 views32 pages

Ai Unit 4

The document provides an overview of various forms of machine learning, including supervised, unsupervised, semi-supervised, and reinforcement learning, detailing their definitions, categories, advantages, disadvantages, and applications. It explains supervised learning's reliance on labeled data for tasks like classification and regression, while unsupervised learning works with unlabeled data to find patterns. Additionally, it discusses semi-supervised learning as a blend of both approaches and highlights reinforcement learning's feedback-based learning process.

Uploaded by

got90992
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
32 views32 pages

Ai Unit 4

The document provides an overview of various forms of machine learning, including supervised, unsupervised, semi-supervised, and reinforcement learning, detailing their definitions, categories, advantages, disadvantages, and applications. It explains supervised learning's reliance on labeled data for tasks like classification and regression, while unsupervised learning works with unlabeled data to find patterns. Additionally, it discusses semi-supervised learning as a blend of both approaches and highlights reinforcement learning's feedback-based learning process.

Uploaded by

got90992
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 32

KLE Society’s

Bachelor of Computer Applications Dharwad

ARTIFICIAL INTELLIGENCE AND APPLICATIONS


BCA VI SEM (NEP)

UNIT IV
Learning
1. Forms of Learning:
1.1 Types of Machine Learning

Machine learning is a subset of AI, which enables the machine to automatically learn from data,
improve performance from past experiences, and make predictions. Machine learning contains a
set of algorithms that work on a huge amount of data. Data is fed to these algorithms to train them, and
on the basis of training, they build the model & perform a specific task.
These ML algorithms help to solve different business problems like Regression, Classification,
Forecasting, Clustering, and Associations, etc.
Based on the methods and way of learning, machine learning is divided into mainly four types, which
are:
1. Supervised Machine Learning
2. Unsupervised Machine Learning
3. Semi-Supervised Machine Learning
4. Reinforcement Learning
1. Supervised Machine Learning

As its name suggests, Supervised machine learning is based on supervision. It means in the supervised
learning technique, we train the machines using the "labelled" dataset, and based on the training, the
machine predicts the output. Here, the labelled data specifies that some of the inputs are already mapped
to the output. More preciously, we can say; first, we train the machine with the input and corresponding
output, and then we ask the machine to predict the output using the test dataset.
Let's understand supervised learning with an example. Suppose we have an input dataset of cats and
dog images. So, first, we will provide the training to the machine to understand the images, such as the
shape & size of the tail of cat and dog, Shape of eyes, colour, height (dogs are taller, cats are
smaller), etc. After completion of training, we input the picture of a cat and ask the machine to identify
the object and predict the output. Now, the machine is well trained, so it will check all the features of the
object, such as height, shape, colour, eyes, ears, tail, etc., and find that it's a cat. So, it will put it in the Cat
category. This is the process of how the machine identifies the objects in Supervised Learning.
The main goal of the supervised learning technique is to map the input variable(x) with the
output variable(y). Some real-world applications of supervised learning are Risk Assessment, Fraud
Detection, Spam filtering, etc.
Categories of Supervised Machine Learning
Supervised machine learning can be classified into two types of problems, which are given below:
• Classification
• Regression

a) Classification
Classification algorithms are used to solve the classification problems in which the output variable is
categorical, such as "Yes" or No, Male or Female, Red or Blue, etc. The classification algorithms predict
the categories present in the dataset. Some real-world examples of classification algorithms are Spam
Detection, Email filtering, etc.
Some popular classification algorithms are given below:
• Random Forest Algorithm
• Decision Tree Algorithm
• Logistic Regression Algorithm
• Support Vector Machine Algorithm

b) Regression
Regression algorithms are used to solve regression problems in which there is a linear relationship
between input and output variables. These are used to predict continuous output variables, such as
market trends, weather prediction, etc.
Some popular Regression algorithms are given below:
• Simple Linear Regression Algorithm
• Multivariate Regression Algorithm
• Decision Tree Algorithm
• Lasso Regression

Advantages and Disadvantages of Supervised Learning


Advantages:
• Since supervised learning work with the labelled dataset so we can have an exact idea about the
classes of objects.
• These algorithms are helpful in predicting the output on the basis of prior experience.
Disadvantages:
• These algorithms are not able to solve complex tasks.
• It may predict the wrong output if the test data is different from the training data.
• It requires lots of computational time to train the algorithm.

Applications of Supervised Learning


Some common applications of Supervised Learning are given below:
• Image Segmentation: Supervised Learning algorithms are used in image segmentation. In this
process, image classification is performed on different image data with pre-defined labels.
• Medical Diagnosis: Supervised algorithms are also used in the medical field for diagnosis
purposes. It is done by using medical images and past labelled data with labels for disease conditions.
With such a process, the machine can identify a disease for the new patients.
• Fraud Detection - Supervised Learning classification algorithms are used for identifying fraud
transactions, fraud customers, etc. It is done by using historic data to identify the patterns that can lead
to possible fraud.
• Spam detection - In spam detection & filtering, classification algorithms are used. These
algorithms classify an email as spam or not spam. The spam emails are sent to the spam folder.
• Speech Recognition - Supervised learning algorithms are also used in speech recognition. The
algorithm is trained with voice data, and various identifications can be done using the same, such as
voice-activated passwords, voice commands, etc.

2. Unsupervised Machine Learning


Unsupervised learning is different from the Supervised learning technique; as its name suggests, there
is no need for supervision. It means, in unsupervised machine learning, the machine is trained using the
unlabeled dataset, and the machine predicts the output without any supervision.
In unsupervised learning, the models are trained with the data that is neither classified nor labelled, and
the model acts on that data without any supervision.
The main aim of the unsupervised learning algorithm is to group or categories the unsorted
dataset according to the similarities, patterns, and differences. Machines are instructed to find the
hidden patterns from the input dataset.
Let's take an example to understand it more preciously; suppose there is a basket of fruit images, and
we input it into the machine learning model. The images are totally unknown to the model, and the task
of the machine is to find the patterns and categories of the objects.
So, now the machine will discover its patterns and differences, such as colour difference, shape
difference, and predict the output when it is tested with the test dataset.
Categories of Unsupervised Machine Learning
Unsupervised Learning can be further classified into two types, which are given below:
• Clustering
• Association

1) Clustering
The clustering technique is used when we want to find the inherent groups from the data. It is a way to
group the objects into a cluster such that the objects with the most similarities remain in one group and
have fewer or no similarities with the objects of other groups. An example of the clustering algorithm is
grouping the customers by their purchasing behaviour.
Some of the popular clustering algorithms are given below:
• K-Means Clustering algorithm
• Mean-shift algorithm
• DBSCAN Algorithm
• Principal Component Analysis
• Independent Component Analysis

2) Association
Association rule learning is an unsupervised learning technique, which finds interesting relations among
variables within a large dataset. The main aim of this learning algorithm is to find the dependency of one
data item on another data item and map those variables accordingly so that it can generate maximum
profit. This algorithm is mainly applied in Market Basket analysis, Web usage mining, continuous
production, etc.
Some popular algorithms of Association rule learning are Apriori Algorithm, Eclat, FP-growth
algorithm.
Advantages and Disadvantages of Unsupervised Learning Algorithm
Advantages:
• These algorithms can be used for complicated tasks compared to the supervised ones because
these algorithms work on the unlabeled dataset.
• Unsupervised algorithms are preferable for various tasks as getting the unlabeled dataset is
easier as compared to the labelled dataset.

Disadvantages:
• The output of an unsupervised algorithm can be less accurate as the dataset is not labelled, and
algorithms are not trained with the exact output in prior.
• Working with Unsupervised learning is more difficult as it works with the unlabelled dataset that
does not map with the output.

Applications of Unsupervised Learning


• Network Analysis: Unsupervised learning is used for identifying plagiarism and copyright in
document network analysis of text data for scholarly articles.
• Recommendation Systems: Recommendation systems widely use unsupervised learning
techniques for building recommendation applications for different web applications and e-commerce
websites.
• Anomaly Detection: Anomaly detection is a popular application of unsupervised learning, which
can identify unusual data points within the dataset. It is used to discover fraudulent transactions.
• Singular Value Decomposition: Singular Value Decomposition or SVD is used to extract
particular information from the database. For example, extracting information of each user located at a
particular location.

3. Semi-Supervised Learning

Semi-Supervised learning is a type of Machine Learning algorithm that lies between Supervised
and Unsupervised machine learning. It represents the intermediate ground between Supervised
(With Labelled training data) and Unsupervised learning (with no labelled training data) algorithms and
uses the combination of labelled and unlabeled datasets during the training period.
Although Semi-supervised learning is the middle ground between supervised and unsupervised
learning and operates on the data that consists of a few labels, it mostly consists of unlabeled data. As
labels are costly, but for corporate purposes, they may have few labels. It is completely different from
supervised and unsupervised learning as they are based on the presence & absence of labels.
To overcome the drawbacks of supervised learning and unsupervised learning algorithms, the
concept of Semi-supervised learning is introduced. The main aim of semi-supervised learning is to
effectively use all the available data, rather than only labelled data like in supervised learning. Initially,
similar data is clustered along with an unsupervised learning algorithm, and further, it helps to label the
unlabeled data into labelled data. It is because labelled data is a comparatively more expensive
acquisition than unlabeled data.
We can imagine these algorithms with an example. Supervised learning is where a student is under the
supervision of an instructor at home and college. Further, if that student is self-analysing the same
concept without any help from the instructor, it comes under unsupervised learning. Under semi-
supervised learning, the student has to revise himself after analyzing the same concept under the
guidance of an instructor at college.
Advantages and disadvantages of Semi-supervised Learning
Advantages:
• It is simple and easy to understand the algorithm.
• It is highly efficient.
• It is used to solve drawbacks of Supervised and Unsupervised Learning algorithms.
Disadvantages:
• Iterations results may not be stable.
• We cannot apply these algorithms to network-level data.
• Accuracy is low.

4. Reinforcement Learning

Reinforcement learning works on a feedback-based process, in which an AI agent (A software


component) automatically explore its surrounding by hitting & trail, taking action, learning from
experiences, and improving its performance. Agent gets rewarded for each good action and get
punished for each bad action; hence the goal of reinforcement learning agent is to maximize the rewards.
In reinforcement learning, there is no labelled data like supervised learning, and agents learn from their
experiences only.
The reinforcement learning process is similar to a human being; for example, a child learns various
things by experiences in his day-to-day life. An example of reinforcement learning is to play a game,
where the Game is the environment, moves of an agent at each step define states, and the goal of the
agent is to get a high score. Agent receives feedback in terms of punishment and rewards.
Due to its way of working, reinforcement learning is employed in different fields such as Game theory,
Operation Research, Information theory, multi-agent systems.
A reinforcement learning problem can be formalized using Markov Decision Process(MDP). In MDP,
the agent constantly interacts with the environment and performs actions; at each action, the
environment responds and generates a new state.
Categories of Reinforcement Learning
Reinforcement learning is categorized mainly into two types of methods/algorithms:
• Positive Reinforcement Learning: Positive reinforcement learning specifies increasing the
tendency that the required behaviour would occur again by adding something. It enhances the strength
of the behaviour of the agent and positively impacts it.
• Negative Reinforcement Learning: Negative reinforcement learning works exactly opposite to
the positive RL. It increases the tendency that the specific behaviour would occur again by avoiding the
negative condition.

Real-world Use cases of Reinforcement Learning


• Video Games: RL algorithms are much popular in gaming applications. It is used to gain super-
human performance. Some popular games that use RL algorithms are AlphaGO and AlphaGO Zero.
• Resource Management: The "Resource Management with Deep Reinforcement Learning" paper
showed that how to use RL in computer to automatically learn and schedule resources to wait for
different jobs in order to minimize average job slowdown.
• Robotics: RL is widely being used in Robotics applications. Robots are used in the industrial and
manufacturing area, and these robots are made more powerful with reinforcement learning. There are
different industries that have their vision of building intelligent robots using AI and Machine learning
technology.
• Text Mining Text-mining, one of the great applications of NLP, is now being implemented with
the help of Reinforcement Learning by Salesforce company.

Advantages and Disadvantages of Reinforcement Learning


Advantages
• It helps in solving complex real-world problems which are difficult to be solved by general
techniques.
• The learning model of RL is similar to the learning of human beings; hence most accurate results
can be found.
• Helps in achieving long term results.

Disadvantage
• RL algorithms are not preferred for simple problems.
• RL algorithms require huge data and computations.
• Too much reinforcement learning can lead to an overload of states which can weaken the results.

2. Decision Tree Classification Algorithm

• Decision Tree is a Supervised learning technique that can be used for both classification and
Regression problems, but mostly it is preferred for solving Classification problems. It is a tree-structured
classifier, where internal nodes represent the features of a dataset, branches represent the
decision rules and each leaf node represents the outcome.

• In a Decision tree, there are two nodes, which are the Decision Node and Leaf Node. Decision
nodes are used to make any decision and have multiple branches, whereas Leaf nodes are the output of
those decisions and do not contain any further branches.

• The decisions or the test are performed on the basis of features of the given dataset.

• It is a graphical representation for getting all the possible solutions to a problem/decision


based on given conditions.

• It is called a decision tree because, similar to a tree, it starts with the root node, which expands
on further branches and constructs a tree-like structure.

• In order to build a tree, we use the CART algorithm, which stands for Classification and
Regression Tree algorithm.

• A decision tree simply asks a question, and based on the answer (Yes/No), it further split the tree
into subtrees.

• Below diagram explains the general structure of a decision tree:


Note: A decision tree can contain categorical data (YES/NO) as well as numeric data.
2.1 Why use Decision Trees?

There are various algorithms in Machine learning, so choosing the best algorithm for the given dataset
and problem is the main point to remember while creating a machine learning model. Below are the two
reasons for using the Decision tree:
• Decision Trees usually mimic human thinking ability while making a decision, so it is easy to
understand.
• The logic behind the decision tree can be easily understood because it shows a tree-like structure.

2.3 Decision Tree Terminologies

Root Node: Root node is from where the decision tree starts. It represents the entire dataset, which
further gets divided into two or more homogeneous sets.
Leaf Node: Leaf nodes are the final output node, and the tree cannot be segregated further after getting
a leaf node.
Splitting: Splitting is the process of dividing the decision node/root node into sub-nodes according to
the given conditions.
Branch/Sub Tree: A tree formed by splitting the tree.
Pruning: Pruning is the process of removing the unwanted branches from the tree.
Parent/Child node: The root node of the tree is called the parent node, and other nodes are called the
child nodes.

2.4 How does the Decision Tree Algorithm Work?


In a decision tree, for predicting the class of the given dataset, the algorithm starts from the root node
of the tree. This algorithm compares the values of root attribute with the record (real dataset) attribute
and, based on the comparison, follows the branch and jumps to the next node.
For the next node, the algorithm again compares the attribute value with the other sub-nodes and move
further. It continues the process until it reaches the leaf node of the tree. The complete process can be
better understood using the below algorithm:

• Step-1: Begin the tree with the root node, says S, which contains the complete dataset.
• Step-2: Find the best attribute in the dataset using Attribute Selection Measure (ASM).
• Step-3: Divide the S into subsets that contains possible values for the best attributes.
• Step-4: Generate the decision tree node, which contains the best attribute.
• Step-5: Recursively make new decision trees using the subsets of the dataset created in step -3.
Continue this process until a stage is reached where you cannot further classify the nodes and called the
final node as a leaf node.

Example: Suppose there is a candidate who has a job offer and wants to decide whether he should accept
the offer or Not. So, to solve this problem, the decision tree starts with the root node (Salary attribute by
ASM). The root node splits further into the next decision node (distance from the office) and one leaf
node based on the corresponding labels. The next decision node further gets split into one decision node
(Cab facility) and one leaf node. Finally, the decision node splits into two leaf nodes (Accepted offers and
Declined offer). Consider the below diagram:

Attribute Selection Measures


While implementing a Decision tree, the main issue arises that how to select the best attribute for the
root node and for sub-nodes. So, to solve such problems there is a technique which is called as Attribute
selection measure or ASM. By this measurement, we can easily select the best attribute for the nodes
of the tree. There are two popular techniques for ASM, which are:
• Information Gain
• Gini Index

1. Information Gain:
• Information gain is the measurement of changes in entropy after the segmentation of a dataset
based on an attribute.
• It calculates how much information a feature provides us about a class.
• According to the value of information gain, we split the node and build the decision tree.
• A decision tree algorithm always tries to maximize the value of information gain, and a
node/attribute having the highest information gain is split first. It can be calculated using the below
formula:
• Information Gain= Entropy(S)- [(Weighted Avg) *Entropy(each feature)
Entropy: Entropy is a metric to measure the impurity in a given attribute. It specifies randomness in
data. Entropy can be calculated as:
Entropy(s)= -P(yes)log2 P(yes)- P(no) log2 P(no)
Where,
• S= Total number of samples
• P(yes)= probability of yes
• P(no)= probability of no

2. Gini Index:
• Gini index is a measure of impurity or purity used while creating a decision tree in the
CART(Classification and Regression Tree) algorithm.
• An attribute with the low Gini index should be preferred as compared to the high Gini index.
• It only creates binary splits, and the CART algorithm uses the Gini index to create binary splits.
• Gini index can be calculated using the below formula:

Gini Index= 1- ΣjPj2


Pruning: Getting an Optimal Decision tree
Pruning is a process of deleting the unnecessary nodes from a tree in order to get the optimal decision tree.
A too-large tree increases the risk of overfitting, and a small tree may not capture all the important
features of the dataset. Therefore, a technique that decreases the size of the learning tree without
reducing accuracy is known as Pruning. There are mainly two types of tree pruning technology used:
• Cost Complexity Pruning
• Reduced Error Pruning.

Advantages of the Decision Tree


• It is simple to understand as it follows the same process which a human follow while making any
decision in real-life.
• It can be very useful for solving decision-related problems.
• It helps to think about all the possible outcomes for a problem.
• There is less requirement of data cleaning compared to other algorithms.

Disadvantages of the Decision Tree


• The decision tree contains lots of layers, which makes it complex.
• It may have an overfitting issue, which can be resolved using the Random Forest algorithm.
• For more class labels, the computational complexity of the decision tree may increase.

3. Regression v/s. Classification in Machine Learning


Regression and Classification algorithms are Supervised Learning algorithms. Both the algorithms are
used for prediction in Machine learning and work with the labeled datasets. But the difference between
both is how they are used for different machine learning problems.
The main difference between Regression and Classification algorithms that Regression algorithms are
used to predict the continuous values such as price, salary, age, etc. and Classification algorithms are
used to predict/Classify the discrete values such as Male or Female, True or False, Spam or Not Spam,
etc.
Consider the below diagram:
3.1 Classification:

Classification is a process of finding a function which helps in dividing the dataset into classes based on
different parameters. In Classification, a computer program is trained on the training dataset and based
on that training, it categorizes the data into different classes.
The task of the classification algorithm is to find the mapping function to map the input(x) to the discrete
output(y).

Example: The best example to understand the Classification problem is Email Spam Detection. The
model is trained on the basis of millions of emails on different parameters, and whenever it receives a
new email, it identifies whether the email is spam or not. If the email is spam, then it is moved to the
Spam folder.

Types of ML Classification Algorithms:

Classification Algorithms can be further divided into the following types:


• Logistic Regression
• K-Nearest Neighbours
• Support Vector Machines
• Kernel SVM
• Naïve Bayes
• Decision Tree Classification
• Random Forest Classification

3.2 Regression:
Regression is a process of finding the correlations between dependent and independent variables. It
helps in predicting the continuous variables such as prediction of Market Trends, prediction of House
prices, etc.
The task of the Regression algorithm is to find the mapping function to map the input variable(x) to the
continuous output variable(y).

Example: Suppose we want to do weather forecasting, so for this, we will use the Regression algorithm.
In weather prediction, the model is trained on the past data, and once the training is completed, it can
easily predict the weather for future days.
3.3 Types of Regression Algorithm:
• Simple Linear Regression
• Multiple Linear Regression
• Polynomial Regression
• Support Vector Regression
• Decision Tree Regression
• Random Forest Regression

3.4 Difference between Regression and Classification


Regression Algorithm Classification Algorithm
In Regression, the output variable In Classification, the output variable
must be of continuous nature or real must be a discrete value.
value.
The task of the regression algorithm is The task of the classification
to map the input value (x) with the algorithm is to map the input
continuous output variable(y). value(x) with the discrete output
variable(y).
Regression Algorithms are used with Classification Algorithms are used
continuous data. with discrete data.
In Regression, we try to find the best In Classification, we try to find the
fit line, which can predict the output decision boundary, which can divide
more accurately. the dataset into different classes.
Regression algorithms can be used to Classification Algorithms can be
solve the regression problems such as used to solve classification problems
Weather Prediction, House price such as Identification of spam
prediction, etc. emails, Speech Recognition,
Identification of cancer cells, etc.
The regression Algorithm can be The Classification algorithms can be
further divided into Linear and Non- divided into Binary Classifier and
linear Regression. Multi-class Classifier.

4. Artificial Neural Network:

The term "Artificial neural network" refers to a biologically inspired sub-field of artificial intelligence
modeled after the brain. An Artificial neural network is usually a computational network based on
biological neural networks that construct the structure of the human brain. Similar to a human brain has
neurons interconnected to each other, artificial neural networks also have neurons that are linked to
each other in various layers of the networks. These neurons are known as nodes.
Artificial neural network tutorial covers all the aspects related to the artificial neural network. In this
tutorial, we will discuss ANNs, Adaptive resonance theory, Kohonen self-organizing map, Building
blocks, unsupervised learning, Genetic algorithm, etc.
What is Artificial Neural Network?

The term "Artificial Neural Network" is derived from Biological neural networks that develop the
structure of a human brain. Similar to the human brain that has neurons interconnected to one another,
artificial neural networks also have neurons that are interconnected to one another in various layers of
the networks. These neurons are known as nodes.
The given figure illustrates the typical diagram of Biological Neural Network.
The typical Artificial Neural Network looks something like the given figure.

Dendrites from Biological Neural Network represent inputs in Artificial Neural Networks, cell nucleus
represents Nodes, synapse represents Weights, and Axon represents Output.
Relationship between Biological neural network and artificial neural network:
BiologicalNeural ArtificialNeural
Network Network
Dendrites Inputs
Cell nucleus Nodes
Synapse Weights
Axon Output

An Artificial Neural Network in the field of Artificial intelligence where it attempts to mimic the
network of neurons makes up a human brain so that computers will have an option to understand things
and make decisions in a human-like manner. The artificial neural network is designed by programming
computers to behave simply like interconnected brain cells.
There are around 1000 billion neurons in the human brain. Each neuron has an association point
somewhere in the range of 1,000 and 100,000. In the human brain, data is stored in such a manner as to
be distributed, and we can extract more than one piece of this data when necessary from our memory
parallelly. We can say that the human brain is made up of incredibly amazing parallel processors.
We can understand the artificial neural network with an example, consider an example of a digital logic
gate that takes an input and gives an output. "OR" gate, which takes two inputs. If one or both the inputs
are "On," then we get "On" in output. If both the inputs are "Off," then we get "Off" in output. Here the
output depends upon input. Our brain does not perform the same task. The outputs to inputs
relationship keep changing because of the neurons in our brain, which are "learning."
The architecture of an artificial neural network:
To understand the concept of the architecture of an artificial neural network, we have to understand
what a neural network consists of. In order to define a neural network that consists of a large number
of artificial neurons, which are termed units arranged in a sequence of layers. Lets us look at various
types of layers available in an artificial neural network.
Artificial Neural Network primarily consists of three layers:

Input Layer:
As the name suggests, it accepts inputs in several different formats provided by the programmer.
Hidden Layer:
The hidden layer presents in-between input and output layers. It performs all the calculations to find
hidden features and patterns.
Output Layer:
The input goes through a series of transformations using the hidden layer, which finally results in output
that is conveyed using this layer.
The artificial neural network takes input and computes the weighted sum of the inputs and includes a
bias. This computation is represented in the form of a transfer function.

It determines weighted total is passed as an input to an activation function to produce the output.
Activation functions choose whether a node should fire or not. Only those who are fired make it to the
output layer. There are distinctive activation functions available that can be applied upon the sort of task
we are performing.
4.1 Advantages of Artificial Neural Network (ANN)

Parallel processing capability:


Artificial neural networks have a numerical value that can perform more than one task simultaneously.
Storing data on the entire network:
Data that is used in traditional programming is stored on the whole network, not on a database. The
disappearance of a couple of pieces of data in one place doesn't prevent the network from working.
Capability to work with incomplete knowledge:
After ANN training, the information may produce output even with inadequate data. The loss of
performance here relies upon the significance of missing data.
Having a memory distribution:
For ANN is to be able to adapt, it is important to determine the examples and to encourage the network
according to the desired output by demonstrating these examples to the network. The succession of the
network is directly proportional to the chosen instances, and if the event can't appear to the network in
all its aspects, it can produce false output.
Having fault tolerance:
Extortion of one or more cells of ANN does not prohibit it from generating output, and this feature makes
the network fault-tolerance.

4.2 Disadvantages of Artificial Neural Network:

Assurance of proper network structure:


There is no particular guideline for determining the structure of artificial neural networks. The
appropriate network structure is accomplished through experience, trial, and error.
Unrecognized behavior of the network:
It is the most significant issue of ANN. When ANN produces a testing solution, it does not provide insight
concerning why and how. It decreases trust in the network.
Hardware dependence:
Artificial neural networks need processors with parallel processing power, as per their structure.
Therefore, the realization of the equipment is dependent.
Difficulty of showing the issue to the network:
ANNs can work with numerical data. Problems must be converted into numerical values before being
introduced to ANN. The presentation mechanism to be resolved here will directly impact the
performance of the network. It relies on the user's abilities.

4.3 How do artificial neural networks work?


Artificial Neural Network can be best represented as a weighted directed graph, where the artificial
neurons form the nodes. The association between the neurons outputs and neuron inputs can be viewed
as the directed edges with weights. The Artificial Neural Network receives the input signal from the
external source in the form of a pattern and image in the form of a vector. These inputs are then
mathematically assigned by the notations x(n) for every n number of inputs.
Afterward, each of the input is multiplied by its corresponding weights ( these weights are the details
utilized by the artificial neural networks to solve a specific problem ). In general terms, these weights
normally represent the strength of the interconnection between neurons inside the artificial neural
network. All the weighted inputs are summarized inside the computing unit.
If the weighted sum is equal to zero, then bias is added to make the output non-zero or something else
to scale up to the system's response. Bias has the same input, and weight equals to 1. Here the total of
weighted inputs can be in the range of 0 to positive infinity. Here, to keep the response in the limits of
the desired value, a certain maximum value is benchmarked, and the total of weighted inputs is passed
through the activation function.
The activation function refers to the set of transfer functions used to achieve the desired output. There
is a different kind of the activation function, but primarily either linear or non-linear sets of functions.
Some of the commonly used sets of activation functions are the Binary, linear, and Tan hyperbolic
sigmoidal activation functions. Let us take a look at each of them in details:
Binary:
In binary activation function, the output is either a one or a 0. Here, to accomplish this, there is a
threshold value set up. If the net weighted input of neurons is more than 1, then the final output of the
activation function is returned as one or else the output is returned as 0.
Sigmoidal Hyperbolic:
The Sigmoidal Hyperbola function is generally seen as an "S" shaped curve. Here the tan hyperbolic
function is used to approximate output from the actual net input. The function is defined as:
F(x) =1/(1+e^-x)
Where e^-x is considered the Steepness parameter.
Types of Artificial Neural Network:
There are various types of Artificial Neural Networks (ANN) depending upon the human brain neuron
and network functions, an artificial neural network similarly performs tasks. The majority of the
artificial neural networks will have some similarities with a more complex biological partner and are
very effective at their expected tasks. For example, segmentation or classification.
Feedback ANN:
In this type of ANN, the output returns into the network to accomplish the best-evolved results
internally. As per the University of Massachusetts, Lowell Centre for Atmospheric Research. The
feedback networks feed information back into itself and are well suited to solve optimization issues. The
Internal system error corrections utilize feedback ANNs.
Feed-Forward ANN:
A feed-forward network is a basic neural network comprising of an input layer, an output layer, and at
least one layer of a neuron. Through assessment of its output by reviewing its input, the intensity of the
network can be noticed based on group behavior of the associated neurons, and the output is decided.
The primary advantage of this network is that it figures out how to evaluate and recognize input
patterns.

5. Support Vector Machine Algorithm

Support Vector Machine or SVM is one of the most popular Supervised Learning algorithms, which is
used for Classification as well as Regression problems. However, primarily, it is used for Classification
problems in Machine Learning.
The goal of the SVM algorithm is to create the best line or decision boundary that can segregate n-
dimensional space into classes so that we can easily put the new data point in the correct category in the
future. This best decision boundary is called a hyperplane.
SVM chooses the extreme points/vectors that help in creating the hyperplane. These extreme cases are
called as support vectors, and hence algorithm is termed as Support Vector Machine. Consider the below
diagram in which there are two different categories that are classified using a decision boundary or
hyperplane:

Example: SVM can be understood with the example that we have used in the KNN classifier. Suppose
we see a strange cat that also has some features of dogs, so if we want a model that can accurately
identify whether it is a cat or dog, so such a model can be created by using the SVM algorithm. We will
first train our model with lots of images of cats and dogs so that it can learn about different features of
cats and dogs, and then we test it with this strange creature. So as support vector creates a decision
boundary between these two data (cat and dog) and choose extreme cases (support vectors), it will see
the extreme case of cat and dog. On the basis of the support vectors, it will classify it as a cat. Consider
the below diagram:

SVM algorithm can be used for Face detection, image classification, text categorization, etc
5.1 Types of SVM
SVM can be of two types:
• Linear SVM: Linear SVM is used for linearly separable data, which means if a dataset can be
classified into two classes by using a single straight line, then such data is termed as linearly separable
data, and classifier is used called as Linear SVM classifier.
• Non-linear SVM: Non-Linear SVM is used for non-linearly separated data, which means if a
dataset cannot be classified by using a straight line, then such data is termed as non-linear data and
classifier used is called as Non-linear SVM classifier.

Hyperplane and Support Vectors in the SVM algorithm:


Hyperplane: There can be multiple lines/decision boundaries to segregate the classes in n-dimensional
space, but we need to find out the best decision boundary that helps to classify the data points. This best
boundary is known as the hyperplane of SVM.
The dimensions of the hyperplane depend on the features present in the dataset, which means if there
are 2 features (as shown in image), then hyperplane will be a straight line. And if there are 3 features,
then hyperplane will be a 2-dimension plane.
We always create a hyperplane that has a maximum margin, which means the maximum distance
between the data points.
Support Vectors:
The data points or vectors that are the closest to the hyperplane and which affect the position of the
hyperplane are termed as Support Vector. Since these vectors support the hyperplane, hence called a
Support vector.
How does SVM works?

Linear SVM:
The working of the SVM algorithm can be understood by using an example. Suppose we have a dataset
that has two tags (green and blue), and the dataset has two features x1 and x2. We want a classifier that
can classify the pair(x1, x2) of coordinates in either green or blue. Consider the below image:

So as it is 2-d space so by just using a straight line, we can easily separate these two classes. But there
can be multiple lines that can separate these classes. Consider the below image:
Hence, the SVM algorithm helps to find the best line or decision boundary; this best boundary or region
is called as a hyperplane. SVM algorithm finds the closest point of the lines from both the classes. These
points are called support vectors. The distance between the vectors and the hyperplane is called as
margin. And the goal of SVM is to maximize this margin. The hyperplane with maximum margin is
called the optimal hyperplane.

Non-Linear SVM:
If data is linearly arranged, then we can separate it by using a straight line, but for non-linear data, we
cannot draw a single straight line. Consider the below image:
So to separate these data points, we need to add one more dimension. For linear data, we have used two
dimensions x and y, so for non-linear data, we will add a third dimension z. It can be calculated as:
z=x2 +y2
By adding the third dimension, the sample space will become as below image:

So now, SVM will divide the datasets into classes in the following way. Consider the below image:
Since we are in 3-d Space, hence it is looking like a plane parallel to the x-axis. If we convert it in 2d space
with z=1, then it will become as:

Hence we get a circumference of radius 1 in case of non-linear data.


5.2 What are SVMs?
A support vector machine (SVM) is a supervised machine learning algorithm that classifies data by
finding an optimal line or hyperplane that maximizes the distance between each class in an N-
dimensional space.
5.3 Applications of SVMs
While SVMs can be applied for a number of tasks, these are some of the most popular applications of
SVMs across industries.

Text classification
SVMs are commonly used in natural language processing (NLP) for tasks such as sentiment analysis,
spam detection, and topic modeling. They lend themselves to these data as they perform well with high-
dimensional data.
Image classification
SVMs are applied in image classification tasks such as object detection and image retrieval. It can also
be useful in security domains, classifying an image as one that has been tampered with.
Bioinformatics
SVMs are also used for protein classification, gene expression analysis, and disease diagnosis. SVMs are
often applied in cancer research (link resides outside ibm.com) because they can detect subtle trends in
complex datasets.
Geographic information system (GIS)
SVMs can analyze layered geophysical structures underground, filtering out the 'noise' from
electromagnetic data. They have also helped to predict the seismic liquefaction potential of soil, which
is relevant to field of civil engineering.

6. Applications of AI:
Natural Language Processing:
large amounts of unstructured text, whether emails, social media conversations, online chats, survey
responses, and many other forms of data.
By bringing NLP into the workplace, companies can analyze data to find what’s relevant amidst the
chaos, and gain valuable insights that help automate tasks and drive business decisions.
So, how can natural language processing make your business smarter? Read on to find out.
Top 11 Natural Language Processing Applications
Natural language processing tools can help businesses analyze data and discover insights, automate
time-consuming processes, and help them gain a competitive advantage.
Let’s take a look at 11 of the most interesting applications of natural language processing in business:
1. Sentiment Analysis
2. Text Classification
3. Chatbots & Virtual Assistants
4. Text Extraction
5. Machine Translation
6. Text Summarization
7. Market Intelligence
8. Auto-Correct
9. Intent Classification
10. Urgency Detection
11. Speech Recognition

1. Sentiment Analysis
Natural language understanding is particularly difficult for machines when it comes to opinions, given
that humans often use sarcasm and irony. Sentiment analysis, however, is able to recognize subtle
nuances in emotions and opinions ‒ and determine how positive or negative they are.
When you analyze sentiment in real-time, you can monitor mentions on social media (and handle
negative comments before they escalate), gauge customer reactions to your latest marketing campaign
or product launch, and get an overall sense of how customers feel about your company.
You can also perform sentiment analysis periodically, and understand what customers like and dislike
about specific aspects of your business ‒ maybe they love your new feature, but are disappointed about
your customer service. Those insights can help you make smarter decisions, as they show you exactly
what things to improve.
2. Text Classification
Text classification, a text analysis task that also includes sentiment analysis, involves automatically
understanding, processing, and categorizing unstructured text.
Let’s say you want to analyze hundreds of open-ended responses to your recent NPS survey. Doing it
manually would take you a lot of time and end up being too expensive. But what if you could train a
natural language processing model to automatically tag your data in just seconds, using predefined
categories and applying your own criteria?
You might use a topic classifier for NPS survey responses, which automatically tags your data by topics
like Customer Support, Features, Ease of Use, and Pricing. Give it a try and see how it performs!
3. Chatbots & Virtual Assistants
Chatbots and virtual assistants are used for automatic question answering, designed to understand
natural language and deliver an appropriate response through natural language generation.
Standard question answering systems follow pre-defined rules, while AI-powered chatbots and virtual
assistants are able to learn from every interaction and understand how they should respond. The best
part: they learn from interactions and improve over time.
These intelligent machines are increasingly present at the frontline of customer support, as they can
help teams solve up to 80% of all routine queries and route more complex issues to human agents.
Available 24/7, chatbots and virtual assistants can speed up response times, and relieve agents from
repetitive and time-consuming queries.
4. Text Extraction
Text extraction, or information extraction, automatically detects specific information in a text, such as
names, companies, places, and more. This is also known as named entity recognition. You can also
extract keywords within a text, as well as pre-defined features such as product serial numbers and
models.
Applications of text extraction include sifting through incoming support tickets and identifying specific
data, like company names, order numbers, and email addresses without needing to open and read every
ticket.
You might also want to use text extraction for data entry. You could pull out the information you need
and set up a trigger to automatically enter this information in your database.
Keyword extraction, on the other hand, gives you an overview of the content of a text, as this free natural
language processing model shows. Combined with sentiment analysis, keyword extraction can add an
extra layer of insight, by telling you which words customers used most often to express negativity
toward your product or service.
5. Machine Translation
Machine translation (MT) is one of the first applications of natural language processing. Even though
Facebooks’s translations have been declared superhuman, machine translation still faces the challenge
of understanding context.
However, if you’ve been an avid user of Google Translate over the years, you’ll know that it has come a
long way since its inception, mainly thanks to huge advances in the field of neural networks and the
increased availability of large amounts of data.
Automated translation is particularly useful in business because it facilitates communication, allows
companies to reach broader audiences, and understand foreign documentation in a fast and cost-
effective way.
6. Text Summarization
Automatic summarization is pretty self-explanatory. It summarizes text, by extracting the most
important information. Its main goal is to simplify the process of going through vast amounts of data,
such as scientific papers, news content, or legal documentation.
There are two ways of using natural language processing to summarize data: extraction-based
summarization ‒ which extracts keyphrases and creates a summary, without adding any extra
information ‒ and abstraction-based summarization, which creates new phrases paraphrasing the
original source. This second approach is more common and performs better.
7. Market Intelligence
Marketers can benefit from natural language processing to learn more about their customers and use
those insights to create more effective strategies.
Analyzing topics, sentiment, keywords, and intent in unstructured data can really boost your market
research, shedding light on trends and business opportunities. You can also analyze data to identify
customer pain points and to keep an eye on your competitors (by seeing what things are working well
for them and which are not).
8. Auto-Correct
Natural Language Processing plays a vital role in grammar checking software and auto-correct functions.
Tools like Grammarly, for example, use NLP to help you improve your writing, by detecting grammar,
spelling, or sentence structure errors.
9. Intent Classification
Intent classification consists of identifying the goal or purpose that underlies a text. Apart from chatbots,
intent detection can drive benefits in sales and customer support areas.
By analyzing customer interactions like emails, chats, or social media posts, you can spot customers that
are ready to purchase. The faster you can detect and classify those leads, the more chances you have of
turning them into customers. Try this email classifier, and sort responses into categories like Interested,
Not Interested, and Unsubscribe.
Finally, looking for customer intent in customer support tickets or social media posts can warn you of
customers at risk of churn, allowing you to take action with a strategy to win them back.
10. Urgency Detection
NLP techniques can also help you detect urgency in text. You can train an urgency detection model using
your own criteria, so it can recognize certain words and expressions that denote gravity or discontent.
This can help you prioritize the most important requests and make sure they don’t get buried under a
pile of unresolved tickets.
Urgency detection helps you improve response times and efficiency, leading to a positive impact on
customer satisfaction.
11. Speech Recognition
Speech recognition technology uses natural language processing to transform spoken language into a
machine-readable format.
Speech recognition systems are an essential part of virtual assistants, like Siri, Alexa, and Google
Assistant, for example. However, there are more and more use cases of speech recognition in business.
For example, adding speech-to-text capabilities to business software, companies are able to
automatically transcribe calls, send emails, and even translate.
Discover Natural Language Processing Tools
Natural language processing has many exciting applications.
Natural language processing tools help businesses process huge amounts of unstructured data, like
customer support tickets, social media posts, survey responses, and more.
Not only are they used to gain insights to support decision-making, but also to automate time-consuming
tasks.
SaaS tools are the most accessible way to get started with natural language processing. With an AI-
platform like MonkeyLearn, you can start using pre-trained models right away, or build a customized
NLP solution in just a few steps (no coding needed).

7. Text Classification Process


The process of text classification starts with reading the document into the code. This is followed by
series of data pre- processing steps chosen based on business problem:
1. Tokenization: It breaks down longer string of text into smaller pieces for e.g. “This sentence needs
to be tokenized” will be broken to subsequent words:

{‘This’, ‘sentence’, ‘needs’, ‘to’, ‘be’, ‘tokenized’}


1 Text normalization: It aims to bring all the text content to same level. Some of the Text
normalization options are as follows:
2 Basic Normalization steps: Lowercasing text, Removal of Punctuations/Tags/whitespaces.
3 Stemming removes affixes (prefix, suffix, infixes, circumfixes) from the word, for example
studies will become studi and studying will become study.
4 Lemmatization: Obtains the canonical/dictionary form of the word. For example, studies and
studying will be converted to study. It is useful in context where words need to retain meaning after pre-
processing.
5 Stop Word Removal: Removal of common words such as the, and, is, a that provide no value to
the overall sentence.
6 Vectorization: The text sequences are transformed into numerical features that can be used in
the model. TF-IDF, Count vectorizer are some of the commonly used approaches for same.

7.1 Feature Selection techniques selects a subset of the features based on their importance. Document
Frequency is one of the common Feature Selection methods used, where words/features present below
certain frequency in the document are filtered out. Feature Extraction is an optional step in some
business scenarios, where additional features are created from pre-existing features. Clustering
methods is one technique used to add new features.
The final step in this process is the tagging of the data to predefined categories using one of the following
methods:
• Manual tagging
• Rule Based Filtering or string-matching algorithms such as fuzzy matching.
• Learning Algorithms such as Neural Networks that can utilize several hundred features to tag the
Text content. The Learning Algorithms can be classified into two approaches:
• Unsupervised Learning –Applied where there is lack of previously tagged data. Techniques like
clustering and associating rule-based algorithms can be applied to group together similar text. An
example is segmentation of customers into groups based on their details, purchase history and behavior.
These groups can then be further analyzed to identify patterns that allows for customizing customer
approach.
• Supervised Learning –Applied where there is enough volume of accurately categorized data
available. The ML algorithms learns the mapping function between the text and the tags based on already
categorized data. Algorithms such as SVM, Neural Networks, Random Forest are commonly used for text
classification.

Text Classification, both through supervised and unsupervised approach, finds application in various
fields such as social media, marketing, customer experience management, digital media etc. Some of
these will be elucidated below with use cases.
Unsupervised Learning:
The unsupervised approach to text classification looks for similar patterns and structures between text
to group them together. It finds application in real world in scenarios where the data volume is too large
to be completely classified, is in real time or the labels are not predefined.

8. Information Retrieval (IR)


In the context of Machine Learning (ML) is a subfield that focuses on the development of algorithms and
systems to efficiently and effectively search, retrieve, and present information from large collections of
unstructured or semi-structured data, such as text documents, images, or multimedia content. The goal
of information retrieval is to help users find relevant information based on their queries or information
needs.
Key concepts and components of Information Retrieval in Machine Learning include:
1. Document Collections: This refers to the corpus of documents or data that needs to be searched
and retrieved. It can include text documents, web pages, emails, multimedia content, and more.
2. Query Processing: Users submit queries (search terms or questions), and the IR system processes
these queries to identify relevant documents. Query processing may involve techniques like query
expansion, query reformulation, and relevance feedback.

3.Document Representation: Documents in the collection are typically represented in a way that
makes it easier to compare them to the user's query. Common representations include vector spaces,
term frequency-inverse document frequency (TF-IDF), and word embeddings.
4. Ranking and Scoring: IR systems assign a relevance score to each document in the collection with
respect to a given query. This score reflects the likelihood that the document is relevant to the user's
information needs. Various ranking algorithms, such as BM25, cosine similarity, and learning-to-rank
techniques, are used for this purpose.
5. Retrieval Models: These are mathematical models that formalize how documents are ranked and
scored. Common retrieval models include the Boolean model, vector space model, probabilistic model,
and language modeling.
6. Evaluation Metrics: To assess the performance of an IR system, various metrics like precision, recall,
F1 score, and mean average precision are used. These metrics measure how well the system retrieves
relevant documents and excludes irrelevant ones.
7. Relevance Feedback: In some cases, users can provide feedback on retrieved results, helping to
refine subsequent searches. This feedback can be explicit (e.g., user ratings) or implicit (e.g., click-
through data).
8. Machine Learning Techniques: Machine learning is often used to improve various aspects of
information retrieval. For example, ML algorithms can be used to learn better document
representations, query understanding, and ranking functions. Deep learning models like neural
networks and recurrent neural networks have also been applied to IR tasks.
9. Cross-Modal Retrieval: In addition to text-based retrieval, IR in ML also extends to multimedia
content, where systems aim to retrieve images, videos, or audio based on text queries and vice versa.
Information Retrieval in Machine Learning has practical applications in search engines,
recommendation systems, document classification, content tagging, and more. It plays a crucial role in
making vast amounts of information accessible and useful to users in a variety of domains.

9. Speech recognition

9.1 What is speech recognition?


Speech recognition, or speech-to-text, is the ability of a machine or program to identify words spoken
aloud and convert them into readable text. Rudimentary speech recognition software has a limited
vocabulary and may only identify words and phrases when spoken clearly. More sophisticated software
can handle natural speech, different accents and various languages.
Speech recognition uses a broad array of research in computer science, linguistics and computer
engineering. Many modern devices and text-focused programs have speech recognition functions in
them to allow for easier or hands-free use of a device.
Speech recognition and voice recognition are two different technologies and should not be confused:
• Speech recognition is used to identify words in spoken language.
• Voice recognition is a biometric technology for identifying an individual's voice.

9.2 How does speech recognition work?


Speech recognition systems use computer algorithms to process and interpret spoken words and
convert them into text. A software program turns the sound a microphone records into written language
that computers and humans can understand, following these four steps:
1. analyze the audio;
2. break it into parts;
3. digitize it into a computer-readable format; and
4. use an algorithm to match it to the most suitable text representation.

Speech recognition software must adapt to the highly variable and context-specific nature of human
speech. The software algorithms that process and organize audio into text are trained on different
speech patterns, speaking styles, languages, dialects, accents and phrasings. The software also separates
spoken audio from background noise that often accompanies the signal.
To meet these requirements, speech recognition systems use two types of models:
• Acoustic models. These represent the relationship between linguistic units of speech and audio
signals.
• Language models. Here, sounds are matched with word sequences to distinguish between
words that sound similar.

9.3 What applications is speech recognition used for?


Speech recognition systems have quite a few applications. Here is a sampling of them.
Mobile devices. Smartphones use voice commands for call routing, speech-to-text processing, voice
dialing and voice search. Users can respond to a text without looking at their devices. On Apple iPhones,
speech recognition powers the keyboard and Siri, the virtual assistant. Functionality is available in
secondary languages, too. Speech recognition can also be found in word processing applications like
Microsoft Word, where users can dictate words to be turned into text.
communicate with users and perform a variety of tasks triggered by voice commands.
Education. Speech recognition software is used in language instruction. The software hears the user's
speech and offers help with pronunciation.
Customer service. Automated voice assistants listen to customer queries and provides helpful
resources.
Healthcare applications. Doctors can use speech recognition software to transcribe notes in real time
into healthcare records.
Disability assistance. Speech recognition software can translate spoken words into text using closed
captions to enable a person with hearing loss to understand what others are saying. Speech recognition
can also enable those with limited use of their hands to work with computers, using voice commands
instead of typing.
Court reporting. Software can be used to transcribe courtroom proceedings, precluding the need for
human transcribers.
Emotion recognition. This technology can analyze certain vocal characteristics to determine what
emotion the speaker is feeling. Paired with sentiment analysis, this can reveal how someone feels about
a product or service.
Hands-free communication. Drivers use voice control for hands-free communication, controlling
phones, radios and global positioning systems, for instance.

10. Artificial Intelligence for Image Processing


Today’s image processing solutions powered by artificial intelligence (AI) can do things that were
unimaginable a few years ago. Advanced authentication systems rely on image processing for facial
recognition, while providers of online services and mobile applications can enhance their solutions with
such trending features as restoration of old photos, automated image and video editing, and synthetic
image generation.
10.1 Basics of digital image processing
Generally speaking, image processing is manipulating an image in order to enhance it or extract
information from it. There are two methods of image processing:
• Analog image processing is used for processing physical photographs, printouts, and other hard
copies of images
• Digital image processing is used for manipulating digital images with the help of computer
algorithms

In both cases, the input is an image. For analog image processing, the output is always an image. For
digital image processing, however, the output may be an image or information associated with that
image, such as data on features, characteristics, bounding boxes, or masks.
Today, image processing is widely used in medical visualization, biometrics, self-driving vehicles,
gaming, surveillance, law enforcement, and other spheres.
Here are some of the main purposes of image processing:
• Visualization — Represent processed data in an understandable way. For instance, giving visual
form to objects that aren’t visible
• Image sharpening and restoration — Improve the quality of processed images
• Image retrieval — Help with image search
• Object measurement — Measure objects in an image
• Pattern recognition — Distinguish and classify objects in an image, identify their positions, and
understand the scene

Digital image processing includes eight key phases

1. Image acquisition — Capture an image with a sensor (such as a camera) and convert it into a
manageable entity (for example, a digital image file). One popular image acquisition method is scraping.
2. Image enhancement — Improve the quality of an image in order to extract hidden information from
it for further processing.
3. Image restoration — Remove possible corruptions from an image in order to get a cleaner version.
This process is mostly based on probabilistic and mathematical models and can be used to get rid of blur
or noise, generate missing pixels, fix camera misfocus, remove watermarks, and eliminate other image
characteristics that may harm the training of a neural network.
4. Color image processing — Improve image quality and analyze image content based on color
information. Depending on the image type, we can talk about pseudocolor processing (when colors are
assigned grayscale values) or RGB processing (for images acquired with a full-color sensor).
5. Image compression and decompression — Reduce or restore the size and resolution of an image.
These techniques are often used for image augmentation when you slightly change an original image to
extend your dataset with quality relevant data. Image augmentation can help improve the way your
neural network model generalizes data and make sure it provides high-quality results.
6. Morphological processing — Describe the shapes and structures of objects in an image to create
datasets for training AI models. In particular, morphological analysis and processing can be applied at
the annotation stage, when you describe what you want your AI model to detect or recognize.
7. Object recognition — Identify specific features of particular objects in an image. AI-based image
recognition often uses such techniques as object detection, object recognition, and segmentation. This
technology is at the core of solutions like driverless automotive systems, medical diagnosis systems,
and AI-powered surveillance.
8. Representation and description — Visualize and describe processed data. Using special
visualization tools, you can turn arrays of numbers and values — the raw output of an AI system — into
readable images suitable for further analysis.
As each of these phases requires processing massive amounts of data, you can’t perform them manually.
The use of AI and machine learning (ML) boosts both the speed of data processing and the quality of
the final result. For instance, with the help of AI platforms, you can successfully accomplish such complex
tasks as object detection, facial recognition, and text recognition. But of course, in order to get high-
quality results, it’s important to pick the right methods and tools for image processing.

11. Computer Vision Applications

Computer vision is a subfield of AI (Artificial Intelligence), which enables machines to derive some
meaningful information from any image, video, or other visual input and perform the required
action on that information. Computer vision is like eyes for an AI system, which means if AI enables
the machine to think, computer vision enables the machines to see and observe the visual inputs.
Computer vision technology is based on the concept of teaching computers to process an image or a
visual input at pixels and derive meaningful information from it. Nowadays, Computer vision is in great
demand and used in different areas, including robotics, manufacturing, healthcare, etc. In this topic, we
will discuss some popular applications of Computer Vision, but before that, let's first understand some
common tasks that are performed by computer vision.
11.1 Below are some common tasks for which computer vision can be used:
• Image Classification: Image classification is a computer vision technique used to classify an
image, such as whether an image contains a dog, a person's face, or a banana. It means that with image
classification, we can accurately predict the class of an object present in an image.
• Object Detection: Object detection uses image classification to identify and locate the objects in
an image or video. With such detection and identification technique, the system can count objects in a
given image or scene and determine their accurate location, along with their labelling. For example, in a
given image, there is one person and one cat, which can be easily detected and classified using the object
detection technique.

• Object Tracking: Object tracking is a computer vision technique used to follow a particular
object or multiple items. Generally, object tracking has applications in videos and real-world
interactions, where objects are firstly detected and then tracked to get observation. Object tracking is
used in applications such as Autonomous vehicles, where apart from object classification and detection
such as pedestrians, other vehicles, etc., tracking of real-time motion is also required to avoid accidents
and follow the traffic rules.
• Semantic Segmentation: Image segmentation is not only about detecting the classes in an image
as image classification. Instead, it classifies each pixel of an image to specify what objects it has. It tries
to determine the role of each pixel in the image.

11.2 Computer Vision Applications


• As per the increasing demand for AI and Machine Learning technologies, computer vision also
has a great demand among different sectors. It has a massive impact on different industries, including
retail, security, healthcare, automotive, agriculture, etc. Below are some most popular applications of
computer vision:
• Defect detection using Computer Vision
• OCR using Computer vision
• Crop Monitoring
• Analysis of X-rays, MRI, and CT scans using Computer Vision
• Road Condition Monitoring
• 3D model Building using Computer vision
• Cancer Detection using Computer Vision
• Plant Disease Detection using Computer Vision
• Traffic Flow Analysis
12. Robotics
Robotics is a branch of engineering and computer sciences that includes the design, construction and
operation of machines that are capable of performing programmed tasks without additional human
involvement. At its core, robotics is about using technology to automate tasks while making them more
efficient and safe.
Historically, robots have been used for tasks that are too difficult or dangerous for humans to perform
— such as lifting heavy equipment — or for activities that are very repetitive, such as assembling
automobiles. By automating these tasks, robotics solutions can enhance productivity and improve
safety, freeing up human workers to focus on other more complex and creative endeavors.
It’s also worth noting that robots are not subject to the same limitations as humans. For example, a
human doing the same task over and over may become tired, bored or disengaged, but the robot will
continue to perform the same task with an unwavering level of efficiency and precision. Robotics
solutions are already making a major impact across numerous industries, from meticulously harvesting
crops to making deliveries and assembling cars.

12.1 Are AI and Robotics the Same Thing?

Although AI and robotics are sometimes used interchangeably, in reality, they are distinct — yet related
— fields. While both AI and robotics can potentially impact various industries and aspects of life in
significant ways, each serves a different purpose and operates in a unique way.
Simply put, AI neural network models are similar to biological neural networks, while robotics is
comparable to the human body. AI refers to the development of systems that can perform tasks that
typically require human intelligence, such as learning, problem-solving and decision-making. These
systems can work autonomously, without the need for constant instructions, since they’re programmed
to learn and adapt on their own.
Robotics, on the other hand, refers to the development of robots that can perform specific physical tasks.
These robots can be programmed to carry out simple, repetitive actions, such as sorting items or
assembling miniscule parts. While AI can be integrated into robotics to enhance the robot’s capabilities
and improve decision-making, it’s not always necessary. Some robotics applications simply require
robots to carry out predictable actions without the need for additional cognitive capabilities. While AI
and robotics are not the same things, they do complement each other and can work together to bring
about a wide range of benefits and advancements in various applications.

12.2 How AI Is Used in Robotics

AI has made substantial progress in recent years, and its integration with robotics has proven to be a
natural progression. While AI in robotics is not yet widespread, it’s rapidly gaining momentum as AI
systems become more advanced. The combination of AI and robotics holds tremendous potential,
leading to increased productivity and efficiency, improved safety and greater flexibility for workers in a
variety of professions.
One of the key ways in which AI is used in robotics is through machine learning. This technique enables
robots to learn and perform specific tasks through observing and mimicking human actions. AI gives
robots a computer vision that enables them to navigate, detect and determine their reactions
accordingly. This helps them go beyond simply performing repetitive tasks to become true “cognitive
collaborators.”
Another way that AI is used in robotics is through edge computing. AI applications in robotics require
the interpretation of massive amounts of data gathered by robot-based sensors in real time, which is
why this data is analyzed close to the machine, rather than being sent off to the cloud for processing.
This approach provides machines with real-time awareness, enabling robots to act on decisions at a rate
much quicker than human capabilities allow.
AI also helps robots learn to perform specific tasks through the use of various sensors, which may
include:
• Time-of-flight optical sensors
• Temperature and humidity sensors
• Ultrasonic sensors
• Vibration sensors
• Millimeter-wave sensors

These sensors help robots to learn and adapt, making them more intelligent and better equipped to act
and react in different scenarios.
These are just a few of the ways that artificial intelligence is used in conjunction with robotics.

12.3 Applications of AI in Robotics

In the world of robotics, AI has proven to be a valuable asset in a variety of applications. From customer
service to manufacturing, AI has made its mark and continues to revolutionize the way we think about
and interact with robots. Let’s take a closer look at some of the key areas where AI is being used
alongside robotics today.
Customer Service: AI-powered chatbots are becoming increasingly common in customer service
applications. These automated service agents can handle simple, repetitive requests without the need
for human involvement. The more these systems interact with humans, the more they learn. And as AI
systems become more sophisticated, we can expect to see more and more robots being used in customer
service in both online and brick-and-mortar environments.
Assembly: AI has proven to be an invaluable tool in robotic assembly applications, especially in complex
manufacturing industries such as aerospace. With the help of advanced vision systems, AI can enable
real-time course correction and can be used to help a robot automatically learn the best paths for certain
processes while in operation.
Packaging: AI is used in the packaging industry to improve efficiency, accuracy and cost-effectiveness.
By continuously refining and saving certain motions made by robotic systems, AI helps make installing
and moving robotic equipment easier for everyone.
Imaging: Across many industries — including assembly and logistics — accurate imaging is crucial.
With the assistance of AI, robots can achieve enhanced visual acuity and image recognition
competencies, enabling greater accuracy in even the smallest of details.
Machine Learning: Machine learning is a powerful tool for robots. By exploring their surroundings,
robots can learn more about their environment, find ways around obstacles and solve problems to
complete tasks more efficiently. From home robots like vacuum cleaners to manufacturing robots in
factories, machine learning is helping robots become more intelligent and adaptable in their work. These
are just a few of the many applications of AI in robotics today. As these technologies continue to expand
and grow in sophistication, it is likely that we will see even more innovative applications in the near
future.

12.4 What Is a Robotics Engineer?

As robotics continue to shape various industries, a robotics engineer plays a critical role in robotic
design, maintenance and functionality. A robotics engineer is a specialist responsible for building,
installing and maintaining the machines that perform tasks in sectors such as manufacturing, security,
aerospace and healthcare.
The day-to-day responsibilities of a robotics engineer include:
• Installing, repairing and testing equipment and components
• Performing predictive maintenance
• Incorporating relevant technical literature into their understanding of system operations
• Identifying new data sources
• Building working relationships
• Ensuring that software solutions meet customer needs
• Developing and deploying AI governance structure to manage ongoing implementation of AI
strategies
• Continuously evaluating and reimagining processes to incorporate conversational AI
• Maintaining knowledge of safety standards and regulations for the safe operation of a system

To become a robotics engineer, a bachelor’s or master’s degree in computer engineering, computer


science, electrical engineering or a related field is required. Fluency in multiple programming languages
and proficiency in algorithm design and debugging are also important qualifications. A successful
robotics engineer is also a continuous learner, a natural problem solver and is driven toward ongoing
improvement.
The average salary for a robotics engineer is $100,205* per year, making it a lucrative and in-demand
career path for those with the right qualifications and skills.
*Salary average according to Glassdoor as of February 2023.

12.5 Future of AI in Robotics

The future of AI in robotics is vast and exciting. The next stage of AI, known as AGI or Artificial General
Intelligence, holds the potential to reach levels of true human understanding. The key to this is
integrating the computational system of AI with a robot. The robot must possess mobility, senses (such
as touch, vision and hearing) and the ability to interact with physical objects, which will enable the
system to experience immediate sensory feedback from every action it takes. This feedback loop enables
the system to learn and comprehend, bringing it closer to achieving true AGI.
The current focus on AI in robotics is shifting from the question of what tasks robots can perform for
people, to what type of input a robot can provide the AI’s “mind.” By allowing AI to explore and
experiment with real objects, it will be possible for it to approach a deeper understanding, much like a
human child. With this integration of AI and robotics, we can expect to see significant advancements in
a wide range of industries, from manufacturing and healthcare to security and space exploration. The
future of AI in robotics is bright and holds the potential for tremendous progress in how we understand
and interact with the world. By combining the computational power of AI with the physical capabilities
of robots, we are opening up new doors for exploration and innovation, and the potential for true AGI is
within reach.

You might also like