Aws Aiml
Aws Aiml
Bachelor of Technology
in
Electronics And Communication Engineering
by
UNDURTY TEJA BABU (323103312L24)
Under the guidance of
1
Gayatri Vidya Parishad College of Engineering (Autonomous)
Visakhapatnam
CERTIFICATE
This is to certify that the Mini project-II/Intern-II titled AI-ML VIRTUAL INTERNSHIP a
bonafide record of the work done by UNDURTY TEJA BABU (323103312L24) in partial
fulfillment of the requirements for the award of the degree of Bachelor of Technology in
Electronics and Communication Engineering of the Gayatri Vidya Parishad College of
Engineering (Autonomous) affiliated to Andhra University, Visakhapatnam during the year
2025-2026.
2
INTERNSHIP CERTIFICATE
3
ACKNOWLEDGEMENT
I would like to express our deep sense of gratitude to our esteemed institute Gayatri Vidya
Parishad College of Engineering (Autonomous), which has provided us an opportunity to
fulfill our cherished desire.
I thank our Course coordinator Dr. CH. Sitha kumari, Associate Professor, Department of
Computer Science and Engineering , and our internship mentor Dr. D. Uma devi , Associate
Professor for the kind suggestions and guidance for the successful completion of our internship.
I am highly indebted to Dr. D. Deepika Rani , Associate Professor and Head of the
Department of Electronics and Communication Engineering, Gayatri Vidya Parishad
College of Engineering (Autonomous), for giving us an opportunity to do the internship in
college.
I express our sincere thanks to our Principal Dr. A.B. KOTESWARA RAO, Gayatri Vidya
Parishad College of Engineering (Autonomous) for his encouragement to us during this
project, giving us a chance to explore and learn new technologies in the form of mini projects.
I am grateful for EDUSKILLS and AICTE for providing us this learning opportunity
Finally, I am indebted to the teaching and non-teaching staff of the Electronics And
Communication Engineering Department for all their support in completion of our project.
4
INDEX
Sl. No Topic name Page number
CLOUD FOUNDATION 7
1 8
Introduction to cloud computing
2 Cloud economics and billing 9
6 Compute 15
7 Storage 16
8 Databases 17
9 Cloud Architecture 18
AI-ML 20
13 Forecasting 25-26
14 Computer vision 27
17 Conclusion 31
18 References 32
5
ABSTRACT
the basics of the cloud computing technologies which are termed as cloud
At the end of this internship I have understood the complete concepts of the cloud
foundations along with the basics topics of artificial intelligence which also as some
and operators, network and security engineers, system and database administrators. The
AI which stands for artificial intelligence refers to systems or machines that mimic
human intelligence to perform tasks and can iteratively improve themselves based on
6
TOPICS LEARNT IN CLOUD FOUNDATION:
6. Compute
7. Storage
8. Data bases
9. Cloud Architecture
7
I. INTRODUCTION TO CLOUD COMPUTING
Few services provided by AWS: Amazon Elastic Compute Cloud (Amazon EC2) ,
Amazon Simple Storage Service (Amazon S3), Amazon Virtual Private Cloud (Amazon
VPC), Amazon Relational Database Service (Amazon RDS) etc.
Three ways to interact with AWS: AWS Management Console(Easy to use graphical
interface), Command Line Interface (AWS CLI) (Access to services by discrete
commands or scripts), Software Development Kits (SDKs)(Access services directly from
your code (such as Java, Python, and others)).
The AWS Cloud Adoption Framework (AWS CAF) provides guidance and best practices
to help organizations identify gaps in skills and processes.
Six core perspectives: Business, People, and Governance perspectives (focus on business
capabilities.) Platform, Security, and Operations perspectives (focus on technical
capabilities.)
8
II. CLOUD ECONOMICS AND BILLING
AWS pricing model: There are three fundamental drivers of cost with AWS: compute,
storage, and outbound data transfer.
How we do pay for AWS:
1) Pay for what you use
2) Pay less when you reserve
3) Pay less when you use more
4) Pay even less as AWS grows
Services with no charge: Amazon VPC(Virtual Private Cloud), IAM(Identity and Access
Management), AWS Elastic Beanstalk, Cloud Formation, Automatic Scaling, AWS Ops
Works, Consolidated Billing.
Total Cost of Ownership (TCO): It is the financial estimate to help identify direct and
indirect costs of a system.
Uses of TCO:
• To compare the costs of running an entire infrastructure environment or specific
work load on-premises versus on AWS
• To budget and build the business case for moving to the cloud.
TCO considerations: Server cost00s, Storage costs, Network costs, IT labor costs.
Use the AWS Pricing Calculator to:
• Estimate monthly costs
• Identify opportunities to reduce monthly costs
• Model your solutions before building them
• Explore price points and calculations behind your estimate
• Find the available instance types and contract terms that meet your needs
• Name your estimate and create and name groups of services
AWS ORGANIZATIONS: It is a free account management service that enables you to
consolidate multiple AWS accounts into an organization that you create and centrally
manage. AWS Organizations include consolidated billing and account management
capabilities that help you tobetter meet the budgetary, security, and compliance needs of
your business.
BENEFITS:
Policy-based account management
Group based account management
Application programming interfaces (APIs) that automate account management
Consolidated billing.
9
III. AWS GLOBAL INFRASTRUCTURE
The AWS Global Infrastructure is designed and built to deliver a flexible, reliable,
scalable, andSecure cloud computing environment with high-quality global network
performance.
The AWS Cloud infrastructure is built around Regions. AWS has 22 Regions
Availability Zones. Availability Zones in turn consist of one or more data centers.
Each AWS Region has multiple, isolated locations that are known as availability
Zones. AWS Points of Presence are located in most of the major cities around the
world.
10
Governance service category.
11
IV. AWS CLOUD SECURITY
AWS SHARED RESPONSIBILITY MODEL: This shared model can help relieve the
customers operational burden as AWS operates, managesand controls the components
from the host operating system. The responsibility of this model is protecting
infrastructure that runs all the services offered inthe AWS Cloud.
Security of the Cloud: AWS is responsible for the physical infrastructure that hosts your
resources, including: Physical security of data centers, Hardware infrastructure, Software
infrastructure, Network infrastructure (routers, switches)
Security in the Cloud: Customer is responsible for Amazon Elastic Compute Cloud
(Amazon EC2) instance operating system (Including patching, maintenance),
Applications (Passwords, role-based access, etc.), Security group configuration, OS
or host-based firewalls (Including intrusion detection or prevention systems),
Network configurations, Account management (Login and permission settings for
each user).
AWS Identity and Access Management (IAM)allows you to control access to compute,
storage, database, application services and handles authentication in the AWS Cloud.
Essential components: IAM user, IAM group, IAM policy, IAM role
Authentication is a basic computer security concept. A user or system must first prove
their identity. You can assign two different types of access to users:
Programmatic access:
1) Authenticate using Access key ID, Secret access key
2) Provides AWS CLI and AWS SDK access
AWS Management Console access:
1) Authenticate using 12-digit Account ID and IAM user name, IAM password
12
2) If enabled, multi-factor authentication (MFA)prompts for an authentication code.
An IAM policy is a formal statement of permissions that will be granted to an entity. There
are two types of IAM policies.
1) Identity-based policies: Permissions policies that you can attach to a principal such as
an IAM user. Categorized as : Managed policies, Inline policies
2) Resource-based policies: These are JSON policy documents that you attach to a
resource, such as an S3 bucket.
13
V. NETWORKING AND CONTENT DELIVERY
A computer network is two or more client machines that are connected together to share
resources. A 32-bit IP address is called an IPv4 address and IPv6 addresses, which are 128
bits, are also available. IPv6 addresses can accommodate more user devices.
The Open Systems Interconnection (OSI) model is a conceptual model that is used to
explain how data travels over a network, consists of 7 layers, shows the common protocols
and addresses that are used to send data at each layer, also be used to understand how
communication takes place in a virtual private cloud (VPC).
A VPC is a logically isolated section of the AWS Cloud, to one Region and requires a
CIDR block, subdivided into subnets. A subnet belongs to one Availability Zone and
requires a CIDR block.
VPC networking options include: Internet gateway, NAT gateway, VPC ,VPC peering,
VPC sharing, AWS Site-to-Site VPN, AWS Direct Connect, AWS Transit Gateway, You
can use the VPC Wizard to implement your design.
VPC security:
1) Build security into your VPC architecture(Isolate subnets if possible, Choose the
appropriate gateway device or VPN connection for your needs, Use firewalls)
2) Security groups and network ACLs are firewall options that you can use to secure your
VPC.
14
Amazon Route 53 is a highly available and scalable cloud DNS web service that translates
domain names into numeric IP addresses, supports several types of routing policies.
15
VI. COMPUTE
Amazon Elastic Compute Cloud (Amazon EC2): It provides virtual machines where you
can host the same kinds of applications that you might run on a traditional on premises
server. It provides secure, resizable compute capacity in the cloud. EC2 instances can
support a variety of workloads.
Common uses for EC2 instances: Application servers, Web servers, Database
servers, Game servers, Mail servers, Media servers, Catalog servers, File servers,
Computing servers, Proxy servers
Right size: Choose the right balance of instance types. Notice when servers can be
either sized down or turned off, and still meet your performance requirements.
Increase elasticity: Design your deployments to reduce the amount of server
capacity that is idle by implementing deployments that are elastic, such as
deployments that use automatic scaling to handle peak loads.
Optimal pricing model: Recognize the available pricing options. Analyse your
usage patterns so that you can run EC2 instances with the right mix of pricing
options.
Optimize storage choices: Analyse the storage requirements of your deployments.
Reduce unused storage overhead when possible, and choose less expensive storage
options if they can still meet your requirements for storage performance.
Benefits of Lambda:
16
VII. STORAGE
Amazon EBS provides block-level storage volumes for use with Amazon EC2 instances.
Amazon EBS provides three volume types: General Purpose SSD, Provisioned IOPS SSD,
and magnetic.
Amazon Simple Storage Service (Amazon S3): Data is stored as objects in buckets,
virtually unlimited storage, Single object is limited to 5 TB, designed for 11 9s of
durability, Granular access to bucket and objects Amazon S3 offers a range of object-level
storage classes that are designed for different use: Amazon S3 Standard, Amazon S3
Intelligent- Tiering, Amazon S3 Standard-IA, Amazon S3 Standard, Amazon S3 One Zone-
IA, Amazon S3 Glacier, Amazon S3 Glacier Deep Archive
Key benefits:
You pay for only what you use, You can access Amazon S3 at anytime from anywhere
through a URL, Amazon S3 offers rich security controls.
Amazon Elastic File System (Amazon EFS): It provides file storage over a network;
Perfect for big data and analytics, media processing workflows, content management, web
serving, and home directories; Fully managed service that eliminates storage administration
tasks; Accessible from the console, an API, or the CLI; Scales up or down as files are
added or removed and you pay for what you use.
Amazon S3 Glacier pricing is based on Region. Its extremely low-cost design works well
for long-term archiving. The service is designed to provide 11 9s of durability for objects.
17
VIII. DATABASES
Use cases:
Web and mobile applications: High throughput, Massive storage scalability, High availability
Ecommerce applications: Low-cost database, Data security, Fully managed solution
Mobile and online games: Rapidly grow capacity, Automatic scaling, Database monitoring
Amazon DynamoDB: DynamoDB is a fast and flexible NoSQL database service for all
applications that need consistent, single-digit-millisecond latency at any scale. The core
DynamoDB components are tables, items, and attributes; runs exclusively on SSDs, and it
supports document and key-value store models& works well for mobile, web, gaming, ad
tech, and Internet of Things (IoT)applications. It’s accessible via the console, the AWS
CLI, and API calls.
Amazon Redshift features: Fast, fully managed data warehouse service; Easily scale with
no downtime; Columnar storage and parallel processing architectures;
Automatically and continuously monitors cluster; Encryption is built in.
18
IX. CLOUD ARCHITECTURE
Architecture is the art and science of designing and building large structures.
Large systems require architects to manage their size and complexity.
Any Company Corporation has three main departments:
• Fly and Snap –image acquisition, preprocessing, and storage
• Show and Sell –promoting, selling, and working with customers
• Make and Ship –manufacturing of products and delivery
Amazon CloudWatch
Amazon CloudWatch is a monitoring and observability service that is built for DevOps
engineers, developers, site reliability engineers (SRE), and IT managers. CloudWatch
monitors your AWS resources (and the applications that you run on AWS) in real time.
CloudWatch alarms
You can create a CloudWatch alarm that watches a single CloudWatch metric or the result
of a math expression based on CloudWatch metrics. You can create a CloudWatch alarm
based on a static threshold, anomaly detection, or a metric math expression. For an alarm
based on a static threshold, you must specify the: Namespace, Metric, Statistic, Period,
Conditions
20
TOPICS LEARNT IN AI – ML:
4. INTRODUCING FORECASTING
21
I, II. INTRODUCING MACHINE LEARNING
Machine learning is the scientific study of algorithms and statistical models to perform a
task by using inference instead of instructions. It is subset of AI, which is a broad branch of
computer science for building machines that can do human tasks.
The first type is supervised learning, where a model uses known inputs and outputs to
generalize future outputs. you can have different types of problems within supervised
learning, categorized into two categories:
1) Classification:
• Binary classification (classifying an observation into one of two categories.)
• Multiclass classification (classify an observation into one of three or more
categories.)
2) Regression: mapping inputs to a continuous value, like an integer.
Most business problems are supervised learning.
The second type is unsupervised learning, where the model doesn’t know inputs or outputs
—it finds patterns in the data without help.
The third type is reinforcement learning, an agent(what drives the learning ) continuously
learns, through trial and error, as it interacts in an environment(the place where the agent
learns).
The goal of the model is to try to correctly estimate the target value for new data.
The ML algorithm uses the features to predict the target.
22
Machine learning tools overview:
Jupyter Notebook is an open-source web application that enables you to create and share
documents that contain live code, equations, visualizations, and narrative text.
Jupyter Lab is also extensible and modular. You can write plugins that add new
components and integrate with existing ones.
Matplotlib is a library for creating scientific static, animated, and interactive visualizations
in Python.
The biggest problems that you directly influence are related to data, but you will also
deal with people, business and technology challenges.
23
III. MACHINE LEARNING PIPELINE WITH AMAZON SAGE MAKER
Imports and variables –This section imports the libraries that are used.
Download and extract –This section makes a web request and saves the bytes from the URL
as a stream.
Upload to Amazon S3 –With the extracted files in a folder, this section enumerates the
folder’s files.
To run statistics on your data, better understand it, you must ensure that it’s in the right
format for analysis.
Loading data can be done by using pandas, you can load data in many different formats such
as CSV, JSON, Excel, and Pickle.
Descriptive statistics - histogram, scatter plot, correlation matrix
Feature selection: selecting the features that are most relevant and discarding the Rest.
Cleaning data: Outliers as fall into two broad categories (a single variation for a single
variable (univariate) and a variation of two or more variables(multivariate)) Deleting the
outliers, Transforming the outlier ,Imputing a new value for the outlier.
24
Feature selection methods are available:
Split data into training, testing and validation sets to help you validate the models accuracy.
Can use K-fold cross validation can help with smaller datasets
Can use 2 key algorithms for supervised learning—XGBoost and linear learner
Use k-means for unsupervised learning
Use Amazon SageMaker training jobs to train models
Securing Data: Done by AWS Identity and Access Management (IAM) service,
encryption at rest & transit, AWS CloudTrail
25
IV. FORECASTING
You can think of time series data as falling into two broad categories.
The first type is univariate, which means that it has only one variable. The second type is
multivariate, which means that it has more than one variable.
In addition to these two categories, most time series datasets also follow one of the following
patterns:
Trend –A pattern that shows the values as they increase, decrease, or stay the
same over time
Seasonal – A repeating pattern that is based on the seasons in a year
Cyclical – Some other form of a repeating pattern
Irregular – Changes in the data over time that appear to be random or that have
no discern able pattern
Applications:
A common occurrence in real world forecasting problems is missing values in the raw data.
The missing data can be calculated in several ways:
Forward fill –Uses the last known value for the missing value.
Moving average –Uses the average of the last known values to calculate the missing
value.
Backward fill –Uses the next known value after the missing value. This practice is
known as lookahead, and it should be avoided.
Interpolation –Essentially uses an equation to calculate the missing value.
26
Some of the time challenges include –
Handling different time formats
Handling missing data through down sampling, up sampling and smoothing
Handling seasonality, such as weekdays and yearly cycles
Avoiding bad correlations
Supported domains:
Retail –Product demand
Inventory planning –Raw materials requirements
EC2 capacity –Capacity demand for Amazon Elastic Compute Cloud
Work force –Workload projections
Web traffic –Projected traffic to one or more websites
Metrics –Projecting metrics such as revenue, sales, or cash flow
Custom –Projections for a domain that you can’t map to one of the previous domains
The Root Mean Square Error (RMSE) is another method for evaluating the reliability of
your forecasts. Like we Quantile Loss, RMSE calculates how far off the forecasted values
were from the actual test data. The RMSE finds the difference between the actual target
value in the dataset and the forecasted value for that time period, and it then squares the
differences.
27
V. INTRODUCING COMPUTER VISION (CV)
Object detection provides the categories of the image and the location of the objects in the
image. The location is provided by a set of coordinates for a box that surrounds the image,
which is known as the bounding box.
Amazon Sage Maker Ground Truth enables you to build high-quality training datasets
for your machine learning models.Ground Truth can use active learning to automate the
labelling of your input data.Models must be trained for the specific domain that you want
to analyse.You can set custom labelling for the specific business case.Custom labelling
workflow.You must label images and create bounding boxes for objects.
28
VI. INTRODUCING NATURAL LANGUAGE PROCESSING
You can apply NLP to a wide range of problems. Some of the more common applications
include:
Search applications (such as Google and Bing)
Human machine interfaces (such as Alexa)
Sentiment analysis for marketing or political campaigns
Social research that is based on media analysis
Chatbots to mimic human speech in applications
Some of the more common use cases for Amazon Transcribe include:
Medical transcription
Video subtitles
Streaming content labeling
Customer call center monitoring
Amazon Polly
Amazon Polly can convert text into lifelike speech. You can input either plaintext files or a
file in Speech Synthesis Markup Language (SSML) format. SSML is a markup language
that you can use to provide special instructions for how speech should sound.
Amazon Translate
With Amazon Translate, you can create multilanguage experiences in your applications.
You can create systems for reading documents in one language, and then render or storing it
in another language. You can also use Amazon Translate as part of a document analysis
system.
29
CASE STUDY: PREDICTIVE MAINTENANCE FOR INDUSTRIAL
EQUIPEMNT
PROBLEM STATEMENT:
The manufacturing enterprise is facing challenges with high equipment downtime
and maintenance costs due to unforeseen breakdowns in a large number of industrial
machines and equipment. This unplanned downtime results in significant production losses
and increased maintenance expenses. To tackle this issue effectively, the company needs a
solution capable of analysing sensor data from these machines to predict potential
malfunctions or maintenance needs. By implementing a predictive maintenance strategy,
the company aims to proactively schedule maintenance tasks, prevent failures, and reduce
overall maintenance costs by minimizing downtime.
SOLUTION:
To address this problem, the enterprise can leverage AWS AI/ML services to build a
predictive maintenance solution. The solution involves the following steps:
1. Data Collection: Use AWS IoT Core to ingest sensor data from various industrial
equipment and machines in real-time. The sensor data may include measurements such as
temperature, vibration, pressure, and other relevant parameters.
2. Data Storage: Store the collected sensor data in a scalable and durable storage service
like Amazon S3.
3. Model Training: Build a predictive maintenance model using AWS SageMaker. The
company can choose from various built-in algorithms or implement custom models using
frameworks like TensorFlow or PyTorch. The model will be trained on historical sensor
data and labeled with instances of equipment failures or maintenance activities.
5. Real-time Inference: As new sensor data streams in from the industrial equipment, the
deployed model will analyze this data and predict the likelihood of equipment failure or the
need for maintenance.
31
CONCLUSION:
32
REFERENCES:
33
34