AWS-FAQS
AWS - Interview Questions And Answers
9 AWS Scenario-Based Interview Questions And Answers
Here are some AWS scenario-based
interview questions that interviewers may ask during the technical round:
1. If I have some private
servers on my premises and some of my workloads get distributed on the public
cloud, what do you call this architecture?
Hybrid cloud architecture
allows for sharing apps and data while maintaining the ability to be managed as
a unified information technology (IT) architecture. It combines public and
private clouds across a wide area network or broadband connection. For this
question, you can mention the type of architecture and the reason for its
usage.
Example: An example of
this kind of architecture is a hybrid cloud because we utilise both your
on-site servers, the private and public cloud. It is better if your private and
public clouds are all on the same network, as it makes this hybrid architecture
easier to use. Hybrid architecture places your public cloud servers in a
virtual private cloud and uses a virtual private network (VPN) to connect it to
your on-premises systems.Related: 33 Cloud Architect Interview Questions (With Sample Answers)
-------------------------------------------------------------------------------------------------------------------------------------
2. You have a video
transcoding application. A queue processes these videos. When video processing
gets interrupted in one instance, it resumes in another. Currently, there is a
massive backlog of videos to be processed. More instances are required for
this, but only until your backlog reduces. Which scenario can be the most
efficient?
Online video transcoding
involves changing a file's format to one more efficiently compressed so viewers
can stream content at the best quality without buffering.
With On-Demand Instances, there
are no long-term commitments and you can pay for the computing capability by
the hour or second. This eliminates high fixed costs and complicated tasks
related to planning, purchasing and maintaining hardware compatibility.Example:
You can use an on-demand
instance. The workload requires immediate completion, which makes it
an urgent requirement. In addition, once your backlog gets cleared, you rarely
require instances, so you may be unable to use reserved instances. Finally,
because the work is urgent, you cannot halt the work on your instance just
because the spot price increases, so you may not be able to use spot instances.
As a result, on-demand instances can be the best option here.
-------------------------------------------------------------------------------------------------------------------------------------
3. Suppose you have an AWS
application that requires 24x7 availability and can be down for a maximum of 15
minutes. How can you ensure backing up your Elastic Block Store (EBS) volume
database?
The text-based command line used
to access the operating system is a command-line interface (CLI). Application
programming interface (API) provides programmers with an effective technique to
build their software programs and to make them simpler and easier to
understand. You can answer how to use the two for cloud computing
administration.
Example: The important
procedures, in this case, can have automated backups because they operate
without requiring manual intervention. API and CLI are essential for automating
the backup process through scripts. The best option is to prepare for a timely
backup of the EC2 instances in EBS. In the event of a disaster or downtime, you
can recover the database instance using the EBS snapshot to keep on Amazon
S3.Related: 10 AWS Interview Questions And Answers
-------------------------------------------------------------------------------------------------------------------------------------
4. When you launch instances
in a cluster placement group, what network performance parameters can you
expect?
You can include peered virtual
private clouds (VPCs) in the same region in a cluster placement group.
Instances in the same cluster placement group share a high-bisection bandwidth
network segment and have a higher TCP/IP traffic throughout the limit per
flow.Example: If an instance gets launched in a placement group, you
can anticipate up to ten Gbps in single-flow and 20 Gbps in multi-flow or full
duplex network performance. The placement group's network traffic may not
exceed five Gbps (full duplex).Related: 12 AWS VPC Interview Questions (With Useful Sample Answers)
-------------------------------------------------------------------------------------------------------------------------------------
5. How can you prepare an
instance to serve traffic by configuring it with the application and its
dependencies?
You can use lifecycle hooks to
manage how your Amazon EC2 instances start and terminate as you scale in and
out. When an instance is launching, you might download and install the
software. When an instance is terminating, you might archive instance log files
in Amazon Simple Storage Service (S3).Example: Using lifecycle hooks
can help achieve this. They are effective because they enable you to pause the
formation or termination of an instance to look inside and conduct customised
operations like configuring the instance, obtaining the necessary files and
carrying out any other tasks. There may be several lifecycle hooks for each
auto-scaling group.
-------------------------------------------------------------------------------------------------------------------------------------
6. How can you safeguard EC2
instances running on a VPC?
You can establish a virtual
network in your own logically isolated area of the AWS cloud using Amazon
Virtual Private Cloud (Amazon VPC), also known as a virtual private cloud or
VPC. In the subnets of your VPC, you can establish AWS resources, such as Amazon
EC2 instances.Example: You may protect EC2 instances operating in a VPC
by using AWS Security groups linked with those instances to provide security at
the protocol and port access levels. You can set up secured access for the EC2
instance for inbound and outbound traffic. Like a firewall, AWS security groups
contain rules that filter the traffic flowing into and out of an EC2 instance
and prevent any unauthorised access to EC2 instances.
-------------------------------------------------------------------------------------------------------------------------------------
7. Where does an AMI fit
while designing an architecture for a solution?
You can launch an instance using
an Amazon Machine Image (AMI), a supported and maintained image offered by AWS.
When you launch an instance, you may specify an AMI. If you require numerous
instances with the same configuration, you can launch them all from a single
AMI.
Example: An instance
comes from an AMI, which acts as a blueprint for virtual machines. When
launching an instance, AWS provides pre-baked AMIs that you can select. Some of
these AMIs are not free; users can purchase them through the AWS Marketplace.
To save space on AWS, you can also choose to design your custom AMI. You can
modify your AMI if you do not require a particular collection of software on
your installation. It reduces costs as you are getting rid of unnecessary
items.
-------------------------------------------------------------------------------------------------------------------------------------
8. You are configuring an
Amazon S3 bucket to serve static assets for your public-facing web application.
Which method can ensure setting all objects you uploaded to the bucket to
public read?
You may securely manage access
to AWS resources with the help of the web service AWS Identity and Access
Management (IAM). IAM enables you to control who has access to resources and
gives authority to use them.Example: Setting the policy for the entire
bucket is preferable to altering each object. As this is a website, all
objects, by default, are accessible to the public. IAM is also useful to
provide more specific permissions.Related: 8 Popular AWS Certifications To Consider Pursuing
-------------------------------------------------------------------------------------------------------------------------------------
9. A company wants to use AWS
to deploy a two-tier web application. Complex query processing and table joins
are required by the application. But the company has limited resources and
requires high availability. Based on the requirements, what is the best
configuration that the company can choose?
The first tier of two-tier
architecture is for client placement. On the same server machine, which is a
part of the second tier, are the database server and the web application
server. The web application's second tier serves data and performs business logic
operations.Example: I believe DynamoDB addresses the
fundamental issues with database scalability, management, stability and
performance, though it lacks the RDBMS's features. DynamoDB does not support
query processing or complex transactions. For this kind of functionality, you
can run a relational engine on Amazon RDS or EC2.
AWS Interview Questions
1. Explain in depth what AWS is?
AWS stands for Amazon Web Service. It is a group of remote
computing services, which is also known asna cloud computing platform. This new
dimension of cloud computing is also known as Laas or infrastructure as a
service.
2. What are the three varieties of cloud services?
The three different varieties of cloud services include:
- Computing
- Storage
- Networking
-
3. Define Auto-scaling?
Auto-scaling is an activity that lets you dispatch advanced
instances on demand. Moreover, auto-scaling helps you to increase or decrease
resource capacity according to the application.
4. What do you mean by AMI?
AMI stands for Amazon Machine Image. It is a kind of
template that provides you related information (an operating system, an
application server, and applications) that is needed to launch the instance,
which is indeed a copy of the AMI working as a virtual server in the cloud.
With the help of different AMIs, you can easily launch instances.
5. Can you illustrate the relationship between an
instance and AMI?
With the help of just a single AMI, you can launch multiple
instances and that to even different types. At the same time, an instance type
is characterized by the host computer's hardware that is utilized for your
instance. Each instance provides different computer and memory capabilities.
Once the situation is launched, you will find it looking like a traditional
host, and you can communicate with its one would with any computer.
6. What does geo-targeting in CloudFront mean?
Suppose you want your business to produce and show
personalized content to the audience based on their geographic location without
making any changes to the URL, head straight to geo-targeting. Geo-targeting
enables you to create customized content for the group of spectators of a
specific geographical area, all by keeping their needs ahead.
AWS Interview Questions For S3
7. What is AWS S3?
S3 stands for Simple Storage Service. AWS S3 can be
utilized to store and get any amount of data at any time and the best part from
anywhere on the web. The payment model for S3 is to pay as you go.
8. How can one send a request to Amazon S3?
You can send the request by utilizing the AWS SDK or REST
API wrapper libraries.
9. What is a default storage class in S3?
The standard frequency accessed is the default storage class
in S3.
10. What different storage classes accessible in Amazon
S3?
Storage class that is accessible in Amazon S3 are:
- Amazon
S3 standard
- Amazon
S3 standard infrequent access
- Amazon
S3 reduced repetition storage
- Amazon
glacier
11. What are the ways to encipher the data in S3?
Three different methods will let you encipher the data in S3
- Server-side
encryption – C
- Server-side
encryption – S3
- Server-side
encryption – KMS
Also, Check Detail Blog on AWS Solution Architect Associate
12.On what grounds the pricing policy of the S3 is
decided?
Following factors arentaken under consideration while
deciding S3:
- Transfer
of data
- Storage
that is utilized
- Number
of requests made
- Transfer
acceleration
- Storage
management
13. What are the different types of routing policies that
are available in Amazon route S3?
The various types of routing policies available are as
follows:
- latency
based
- Weighted
- Failover
- Simple
- Geolocation
14. What is the standard size of an S3 bucket?
The maximum size of an S3 bucket is five terabytes.
15. Is Amazon S3 an international service?
Yes, Definitely. Amazon S3 is an international
service. Its main objective is to provide an object storage facility through
the web interface, and it utilizes the Amazon scalable storage infrastructure
to function its global network.
16. What are the important differences between EBS and
S3?
Here we have listedndown some of the essential differences
between EBS and S3:
- EBS
is highly scalable, whereas S3 is less scalable.
- EBS
has blocked storage; on the other hand, S3 is object storage.
- EBS
works faster than S3, whereas S3 works slower than EBS.
- The
user can approach EBS only through the given EC2 instance,but S3 can be
accessible by anyone. It is a public instance.
- EBS
supports the file system interface, whereas S3 supports the web interface.
17. What is the process to upgrade or downgrade a system
that involves near-zero downtime?
With the help of these following steps, one can upgrade or
downgrade a system with near-zero downtime:
- Start
EC2 console
- Select
the AMI operating system
- Open
an instance with a recent instance type
- Install
the updates
- Install
applications
- Analyze
the instance to check whether it is working
- If
working then expand the new instance and cover it up with the older one
- After
it is extended the system with near-zero downtime can be upgraded and
downgraded
18. What all is included in AMI?
AMI includes the following:
- A
template for the root volume for the instance
- Opening
permission
- A
block mapping which helps to decide on the capacity to be attached when it
gets launched.
19. Are there any tools or techniques available that will
help one understand if you are paying more than you should be and how accurate
it is?
With the help of these below-mentioned resources, you will
know whether the amount you are paying for the resource is accurate or not:
- Check
the top services table: You will find this on the dashboard in
the cost management console that will display the top five most used
services. This will also demonstrate how much you are paying on the
resources in question.
- Cost
explorer: With the help of cost explorer, you can see and check
the usage cost for 13 months. Moreover, know the amount of the next three
months too.
- AWS
budget: This lets you plan your budget efficiently.
- Cost
allocation tags: Get to view that resource that has cost you more
in a particular month. Moreover, organize and track your resource as well.
20. Apart from the console, is there any substitute tool
available that will help me log into the cloud environment?
- AWS
CLI for Linux
- Putty
- AWS
CLI for Windows
- AWS
CLI for Windows CMD
- AWS
SDK
- Eclipse
21. Can you name some AWS services that are not
region-specific?
- IAM
- Route
53
- Web
application firewall
- CloudFront
22. Can you define EIP?
EIP stands for Elastic IP address. It is a static Ipv4
address that is provided by AWS to administer dynamic cloud computing
services.
Checkout Other Popular Categories As Well
AWS Interview Questions For VPC
23. What is VPC?
VPC stands for Virtual Private cloud. VPC enables you to
open AWS resources into the world of virtual networks. With its help, network configuration,
as per the users' business requirements, can be build-up and personalized.
24. Illustrate some security products and features that
are available in VPC?
- Security
groups: This plays the role of the firewall for the EC2 instances
and helps to control inbound and outbound traffic at the instance grade.
- Network
access control lists: This represents the role of the firewall
for the subnets and helps control inbound and outbound traffic at the
subnet grade.
- Flow
logs: Flow logs help apprehend incoming and the outbound traffic
from the network interfaces in your VPC.
25. How can an Amazon VPC be monitored?
One can control VPC by using the following:
- CloudWatch
and CloudWatch logs
- VPC
flow logs
26. How many subnets can one have as per VPC?
One can have up to 200nsubnets per VPC
27. Provide the default table that we get when one sets up
AWS VPC?
The list of default tables are as follows:
- Network
ACL
- Security
group
- Route
table
28. How can security to your VPC be controlled?
One can utilize security groups, network access controls
(ACLs), and flow logs to administer your AWS VPC security.
29. Does the property of the broadcast or multicast be
backed up by Amazon VPC?
No. As of now, Amazon VPI does not provide any
support for broadcast or multicast process.
30. Explain the difference between a Domain and a Hosted
Zone?
This is the frequently asked question.
Domain
A domain is a collection of data describing a self-contained
administrative and technical unit. For example www.vinsys.com is a
domain and a general DNS concept.
Hosted zone
A hosted zone is a container that holds information about
how you want to route traffic on the internet for a specific domain. For
example fls.vinsys.com is
a hosted zone.
31. What are NAT gateways?
NAT stands for NetworknAddress Translation. NAT enables
instances to associate in a private subnet with the help of the internet and
other AWS services. Furthermore, NAT prohibits the internet from having a
connection with the instances.
32. How many buckets can be set-up in AWS by default?
You can build-up up to 100 buckets in each AWS account by
default.
33. How is SSH agent forwarding set-up so that you do not
have to copy the key every time you log in?
Here are the steps to achieve the set-up for this:
- Go
to PuTTY configuration
- Log
in to category SSH --- Auth
- Allow
SSH agent forwarding to your instance.
AWS Interview Questions for Amazon EC2
34. What are the different varieties of EC2 instances based
on their expenditure?
The three varieties ofnEC2 instances based on their cost
are:
- On-demand
instance: This comes in a lesser amount but is not recommended for long
term use.
- Spot
instance: This is not much expensive and can be purchased through bidding.
- Reserved
instance: This one is recommended for those who are planning to utilize an
instance for a year or more.
35. What is the best security practice for Amazon EC2?
Go through the following steps for secure Amazon EC2 best
practice:
- Utilize
AWS identity and access management to control access to your AWS resource.
- Forbid
access by enabling only trusted hosts or networks tonaccess ports on your
instance.
- Analyze
the rules in your security groups regularly.
- Open
only permission that you need
- Stop
passport login, for instance, opened from your AMI
36. What are the steps to configure CloudWatch to reclaim
EC2 instance?
Here are the steps that will help you restore EC2 instance:
- Set
up an alarm with the help of Amazon CloudWatch
- In
the alarm, go to Define alert and go to the action tab
- Select
recover this instance option
Other Important AWS Interview Questions
37. What are the various types of AMI designs?
The types are
- Completely
baked AMI
- Slightly
baked AMI (JeOS AMI)
- Hybrid
AMI
38. How can a user gain access to a specific bucket?
One needs to cover the below-mentioned steps to gain access:
- Classify
your instances
- Elaborate
on how licensed users can administer the specific server
- Lockdown
your tags
- Attach
your policies to IAM users
39. How can a current instance be added to a new Autoscaling
group?
Have a look at then steps how you can add an existing
instance to a new auto-scaling group:
- Launch
EC2 console
- Under
instances select your instance
- Select
the action, instance setting and attach to the auto-scaling group
- Choose
a new auto-scaling group
- Adhere
to this group to the instance
- If
needed edit the instance
- After
you are done, you can add the instance to a new auto-scaling group
successfully.
40. What is SQS?
SQS stands for Simple Queue Service. SQS administers the
message queue service. Either you can move the data or message from one
application to another even though it is not in an active state. With the help
of SQS, one can be sent between multiple services.
41. What are the types of queues in SQS?
There are two types of queues in SQS:
- Standard
Queues: This type of queue provides a limitless number of
transactions per second. Standard Queue is a default queue type.
- FIFO
Queues: FIFO queues ensure that the order of messages is received
and is strictly conserved in the precise order that they sent.
42. What are the different types of instances available?
Below we have mentioned the following types of instances
that are available:
- General-purpose
- Storage
optimize
- Accelerated
computing
- Computer-optimized
- Memory-optimized
43. What aspects need to be considered while migrating to
Amazon Web Services?
Have a look at the aspects that need to be taken into
consideration:
- Operational
amount
- Workforce
Capacity
- Cost
evasion
- Operational
facility
- Business
quickness
44. What are the components of an AWS CloudFormation
template?
YAML or JSON are the two AWS Cloud formation templates that
consist of five essential elements.
- Template
framework
- Output
values
- Data
tables
- Resources
- File
format version
45. What are the key pairs in AWS?
Secure logs in information for your virtual machine are key
pairs. To associate with the instances, you can utilize the key pairs, which
consist of a public key and private key.
46. How many Elastic IPs are granted you to set up by AWS?
VPC Elastic IP addresses are granted for each AWS account.
47. What are the advantages of auto-scaling?
Here are the various advantages of auto-scaling:
- Autoscaling
provides fault tolerance
- Provides
much-improved availability
- Better
cost management policy.
48. How can old snapshots be auto-deleted?
Have a look at the steps to auto-delete old snapshots:
- For
best practices snapshots needs to be taken of EBS volumes on Amazon S3
- AWS
Ops automaton is utilized to handle all the snaps naturally
- This
lets you set up, copy and delete Amazon EBS snapshots.
49. What are the varieties of load balancers in AWS?
You will come across three types of load balancers in AWS:
Application load balancer Network load balancer Classic load balancer
50. Are there any different AWS IAM divisions that one can
control?
With the help of AWS IAM, you can do the following: Build up
and administer IAM users Build up and conduct IAM groups Administer the
security accreditation of the users Build up and apply policies to provide
access to AWS services and resources
51. What do you mean by subnet?
A subnet is a larger section of IP addresses that is divided
into pieces.
52. What is the portrayal of AWS CloudTrail?
Cloud trail is a specifically framed tool that is used for
logging and tracking API calls. It is also used to analyze all S3 bucket
accesses.
53. Illustrate Amazon Elastic Cache?
A web service that is used to make deployments easily
accessible, scale, and store data in the cloud is known as Amazon ElasticCache.
54. What are the types of AMI that are provided by AWS?
The following types of AMI provided by AWS are as follows:
Instance store backed EBS backed
55. What is the duration taken by the boot time for the
instance stored backed AMI?
The duration for the boot time for an Amazon instance store
blackened AMI is less than 5 minutes.
56. Do you require an internet gateway to utilize peering
connections?
EBS volumes cannot be connected to multiple instances. But
you can compare various EBS volumes to a single case.
57. What are the different types of Cloud services?
The different types of cloud services are: Software as a
service (SaaS) Data as a service (DaaS) Platform as a service (PaaS)
Infrastructure as a service (IaaS)
58. Can you state some important features of the Amazon
cloud search?
The essential features of Amazon cloud search are: Boolean
searches Prefix searches Range searches Entire text search AutoComplete advice
59. Is vertical scaling allowed in Amazon instance?
Yes. Vertical estimation can be scaled in Amazon instance.
60. What are the different layers of cloud architecture, as
is explained in AWS training?
Here are the different types of layers of cloud
architecture: Cloud controller Cluster controller Storage controller Node
controller
61. What do you mean by Cloud Watch?
Cloud watching is an administration tool that is available
in Amazon web services. With the assistance of a cloud watch, one can
administer various resources of the organization. You can easily look at
multiple things like health, applications, and networks.
62. Can you provide few examples of the DB engine that is
utilized in AWS RDS?
Here that the examples of DB engine that is used in AWS RDS:
Maria DB Oracle DB MS-SQL DB MYSQL DB Postgre DB
63. Is region based services on all services get supported
by Amazon?
No. Region, specific service on all services, is not
provided and supported by Amazon. However, most of the services are
region-specific.
64. Which is the cheapest AWS region?
The US standard is the cheapest region, but it is also the
most established AWS region.
65. What are the advantages of AWS?
One of AWS's main advantages is that it provides services to
its users at extremely low cost. The service is easy to utilize, and the users
should not think and worry about the security factor, servers, and databases.
It comes with several benefits that make the users depend upon them easily.
66. Which service would you utilize if you need to perform
real-time administration of AWS services and get actionable insights?
Amazon Cloud Watch is the service that one can utilize.
67. Is AWS RDS available for free?
Yes. It is. It is a free tier. RDS is used by the AWS
customers, to begin with, the management database service in the cloud for
free.
68. State the difference between block storage and file
storage?
Block storage works at a lower grade and helps to administer
the data asset of blocks, whereas file storage functions at a higher level in
the form of file or folders.
Above aws scenario-based interview questions for experienced
and for freshers are just some of the examples that you can come across while
appearing for big IT companies like TCS, cognizant, Infosys aws interview
questions.
Vinsys is AWS Authorized Training Partner In India Check Out
the Various AWS Training & certifications or connect with us to know which
is most suitable to you.
https://www.projectpro.io/article/top-50-aws-interview-questions-and-answers-for-2018/399
1. Differentiate between on-demand instances and spot
instances.
Spot Instances are spare unused Elastic Compute Cloud (EC2)
instances that one can bid for. Once the bid exceeds the existing spot price
(which changes in real-time based on demand and supply), the spot instance will
be launched. If the spot price exceeds the bid price, the instance can go away
anytime and terminate within 2 minutes of notice. The best way to decide on the
optimal bid price for a spot instance is to check the price history of the last
90 days available on the AWS console. The advantage of spot instances is that
they are cost-effective, and the drawback is that they can be terminated
anytime. Spot instances are ideal to use when –
- There
are optional nice-to-have tasks.
- You
have flexible workloads that can run when there is enough computing
capacity.
- Tasks
that require extra computing capacity to improve performance.
On-demand instances are made available whenever you require
them, and you need to pay for the time you use them hourly. These instances can
be released when they are no longer required and do not require any upfront
commitment. The availability of these instances is guaranteed by AWS, unlike
spot instances.
The best practice is to launch a couple of on-demand
instances which can maintain a minimum level of guaranteed compute resources
for the application and add on a few spot instances whenever there is an
opportunity.
Here's what valued users are saying about ProjectPro
I come from Northwestern University, which is ranked 9th in
the US. Although the high-quality academics at school taught me all the basics
I needed, obtaining practical experience was a challenge. This is when I was
introduced to ProjectPro, and the fact that I am on my second subscription
year...
2. What is the boot time for an instance store-backed
instance?
The boot time for an Amazon Instance Store -Backed AMI is
usually less than 5 minutes.
3. Is it possible to vertically scale on an Amazon
Instance? If yes, how?
Following are the steps to scale an Amazon Instance
vertically –
- Spin
up a larger Amazon instance than the existing one.
- Pause
the existing instance to remove the root ebs volume from the server and
discard.
- Stop
the live running instance and detach its root volume.
- Make
a note of the unique device ID and attach that root volume to the new
server.
- Start
the instance again.
4. Differentiate between vertical and horizontal scaling in
AWS.
The main difference between vertical and horizontal scaling
is how you add compute resources to your infrastructure. In vertical scaling,
more power is added to the existing machine. In contrast, in horizontal
scaling, additional resources are added to the system with the addition of more
machines into the network so that the workload and processing are shared among
multiple devices. The best way to understand the difference is to imagine
retiring your Toyota and buying a Ferrari because you need more horsepower.
This is vertical scaling. Another way to get that added horsepower is not to
ditch the Toyota for the Ferrari but buy another car. This can be related to
horizontal scaling, where you drive several cars simultaneously.
When the users are up to 100, an Amazon EC2 instance alone
is enough to run the entire web application or the database until the traffic
ramps up. Under such circumstances, when the traffic ramps up, it is better to
scale vertically by increasing the capacity of the EC2 instance to meet the
increasing demands of the application. AWS supports instances up to 128 virtual
cores or 488GB RAM.
When the users for your application grow up to 1000 or more,
vertical cannot handle requests, and there is a need for horizontal scaling,
which is achieved through a distributed file system, clustering, and load
balancing.
5. What is the total number of buckets that can be created
in AWS by default?
100 buckets can be created in each of the AWS accounts. If
additional buckets are required, increase the bucket limit by submitting a
service limit increase.
6. Differentiate between Amazon RDS, Redshift, and Dynamo
DB.
|
Features |
Amazon RDS |
Redshift |
Dynamo DB |
|
Computing Resources |
Instances with 64 vCPU and 244 GB RAM |
Nodes with vCPU and 244 GB RAM |
Not specified, SaaS-Software as a Service. |
|
Maintenance Window |
30 minutes every week. |
30 minutes every week. |
No impact |
|
Database Engine |
MySQL, Oracle DB, SQL Server, Amazon Aurora, Postgre SQL |
Redshift |
NoSQL |
|
Primary Usage Feature |
Conventional Databases |
Data warehouse |
Database for dynamically modified data |
|
Multi A-Z Replication |
Additional Service |
Manual |
In-built |
7. An organization wants to deploy a two-tier web
application on AWS. The application requires complex query processing and
table joins. However, the company has limited resources and requires high
availability. Which is the best configuration for the company based on the
requirements?
DynamoDB deals with core problems of database storage,
scalability, management, reliability, and performance but does not have an
RDBMS’s functionalities. DynamoDB does not support complex joins or query
processing, or complex transactions. You can run a relational engine on Amazon
RDS or Amazon EC2 for this kind of functionality.
8. What should be the instance’s tenancy attribute for
running it on single-tenant hardware?
The instance tenancy attribute must be set to a dedicated
instance, and other values might not be appropriate for this operation.
9. What are the important features of a classic load
balancer in Amazon Elastic Compute Cloud (EC2)?

- The
high availability feature ensures that the traffic is distributed among
Amazon EC2 instances in single or multiple availability zones. This
ensures a high scale of availability for incoming traffic.
- Classic
load balancer can decide whether to route the traffic based on the health
check’s results.
- You
can implement secure load balancing within a network by creating security
groups in a VPC.
- Classic
load balancer supports sticky sessions, which ensures a user’s traffic is
always routed to the same instance for a seamless experience.
10. What parameters will you consider when choosing the
availability zone?
Performance, pricing, latency, and response time are factors
to consider when selecting the availability zone.
11. Which instance will you use for deploying a 4-node
Hadoop cluster in AWS?
We can use a c4.8x large instance or i2.large for this, but
using a c4.8x will require a better configuration on the PC.
12. How will you bind the user session with a specific
instance in ELB (Elastic Load Balancer)?
This can be achieved by enabling Sticky Session.
13. What are the possible connection issues you encounter
when connecting to an Amazon EC2 instance?
- Unprotected
private key file
- Server
refused key
- Connection
timed out
- No
supported authentication method available
- Host
key not found,permission denied.
- User
key not recognized by the server, permission denied.
14. Can you run multiple websites on an Amazon EC2 server
using a single IP address?
More than one elastic IP is required to run multiple
websites on Amazon EC2.
15. What happens when you reboot an Amazon EC2 instance?
Rebooting an instance is just similar to rebooting a PC. You
do not return to the image’s original state. However, the hard disk contents
are the same as before the reboot.
16. How is stopping an Amazon EC2 instance different from
terminating it?
Stopping an Amazon EC2 instance result in a normal shutdown
being performed on the instance, and the instance is moved to a stop state.
However, when an EC2 instance is terminated, it is transferred to a stopped
state, and any EBS volumes attached to it are deleted and cannot be recovered.
Advanced AWS Interview Questions and Answers
Here are a few AWS Interview Questions and Answers for
experienced professionals to further strengthen their knowledge of AWS services
useful in cloud computing.
17. Mention the native AWS security logging capabilities.
AWS CloudTrail:
This AWS service facilitates security analysis, compliance
auditing, and resource change tracking of an AWS environment. It can also
provide a history of AWS API calls for a particular account. CloudTrail is an
essential AWS service required to understand AWS use and should be enabled in
all AWS regions for all AWS accounts, irrespective of where the services are
deployed. CloudTrail delivers log files and an optional log file integrity
validation to a designated Amazon S3 (Amazon Simple Storage Service) bucket
once almost every five minutes. When new logs have been delivered, AWS
CloudTrail may be configured to send messages using Amazon Simple Notification
Service (Amazon SNS). It can also integrate with AWS CloudWatch Logs and AWS Lambda for
processing purposes.
AWS Config:
AWS Config is another significant AWS service that can
create an AWS resource inventory, send notifications for configuration changes
and maintain relationships among AWS resources. It provides a timeline of
changes in resource configuration for specific services. If multiple changes
occur over a short interval, then only the cumulative changes are recorded.
Snapshots of changes are stored in a configured Amazon S3 bucket and can be set
to send Amazon SNS notifications when resource changes are detected in AWS.
Apart from tracking resource changes, AWS Config should be enabled to
troubleshoot or perform any security analysis and demonstrate compliance over
time or at a specific time interval.
AWS Detailed Billing Reports:
Detailed billing reports show the cost breakdown by the
hour, day, or month, by a particular product or product resource, by each
account in a company, or by customer-defined tags. Billing reports indicate how
AWS resources are consumed and can be used to audit a company’s consumption of
AWS services. AWS publishes detailed billing reports to a specified S3 bucket
in CSV format several times daily.
Amazon S3 (Simple Storage Service) Access Logs:
Amazon S3 Access logs record information about individual
requests made to the Amazon S3 buckets and can be used to analyze traffic
patterns, troubleshoot, and perform security and access auditing. The access
logs are delivered to designated target S3 buckets on a best-effort basis. They
can help users learn about the customer base, define access policies, and set
lifecycle policies.
Elastic Load Balancing Access Logs:
Elastic Load Balancing Access logs record the individual
requests made to a particular load balancer. They can also analyze traffic
patterns, perform troubleshooting, and manage security and access auditing. The
logs give information about the request processing durations. This data can
improve user experiences by discovering user-facing errors generated by the
load balancer and debugging any errors in communication between the load
balancers and back-end web servers. Elastic Load Balancing access logs get delivered
to a configured target S3 bucket based on the user requirements at five or
sixty-minute intervals.
Amazon CloudFront Access Logs:
Amazon CloudFront Access logs record individual requests
made to CloudFront distributions. Like the previous two access logs, Amazon
CloudFront Access Logs can also be used to analyze traffic patterns, perform
any troubleshooting required, and for security and access auditing. Users can
use these access logs to gather insight about the customer base, define access
policies, and set lifecycle policies. CloudFront Access logs get delivered to a
configured S3 bucket on a best-effort basis.
Amazon Redshift Logs:
Amazon Redshift logs collect
and record information concerning database connections, any changes to user
definitions, and activity. The logs can be used for security monitoring and
troubleshooting any database-related issues. Redshift logs get delivered to a
designated S3 bucket.
Amazon Relational Database Service (RDS) Logs:
RDS logs record information on access, errors, performance,
and database operation. They make it possible to analyze the security,
performance, and operation of AWS-managed databases. RDS logs can be viewed or
downloaded using the Amazon RDS console, the Amazon RDS API, or the AWS
command-line interface. The log files may also be queried from a specific
database table.
Amazon Relational Database Service (RDS) logs capture
information about database access, performance, errors, and operation. These
logs allow security, performance, and operation analysis of the AWS-managed
databases. Customers can view, watch, or download these database logs using the
Amazon RDS console, the AWS Command Line Interface, or the Amazon RDS API. the
log files may also be queried by using DB engine-specific database
tables.
Amazon VPC Flow Logs:
Amazon VPC Flow logs collect information specific to the IP
traffic, incoming and outgoing from the Amazon Virtual Private Cloud (Amazon
VPC) network interfaces. They can be applied, as per requirements, at the VPC,
subnet, or individual Elastic Network Interface level. VPC Flow log data is
stored using Amazon CloudWatch Logs. To perform any additional processing or
analysis, the VPC Flow log data can be exported using Amazon CloudWatch. It is
recommended to enable Amazon VPC flow logs for debugging or monitoring policies
that require capturing and visualizing network flow data.
Centralized Log Management Options:
Various options are available in AWS for centrally managing
log data. Most of the AWS audit and access logs data are delivered to Amazon S3
buckets, which users can configure.
Consolidation of all the Amazon S3-based logs into a
centralized, secure bucket makes it easier to organize, manage and work with
the data for further analysis and processing. The Amazon CloudWatch logs
provide a centralized service where log data can be aggregated.
18. What is a DDoS attack, and how can you handle it?
A Denial of Service (DoS) attack occurs when a malicious
attempt affects the availability of a particular system, such as an application
or a website, to the end-users. A DDoS attack or a Distributed Denial of
Service attack occurs when the attacker uses multiple sources to generate the
attack.DDoS attacks are generally segregated based on the layer of the Open
Systems Interconnection (OSI) model that they attack. The most common DDoS
attacks tend to be at the Network, Transport, Presentation, and Application
layers, corresponding to layers 3, 4, 6, and 7, respectively.
19. What are RTO and RPO in AWS?

The Disaster Recovery (DR) Strategy involves having backups
for the data and redundant workload components. RTO and RPO are objectives used
to restore the workload and define recovery objectives on downtime and data
loss.
Recovery Time Objective or RTO is the maximum acceptable
delay between the interruption of a service and its restoration. It determines
an acceptable time window during which a service can remain unavailable.
Recovery Point Objective or RPO is the maximum amount of
time allowed since the last data recovery point. It is used to determine what
can be considered an acceptable loss of data from the last recovery point to
the service interruption.
RPO and RTO are set by the organization using AWS and have
to be set based on business needs. The cost of recovery and the probability of
disruption can help an organization determine the RPO and RTO.
Upskill yourself for your dream job with
industry-level big data projects
with source code.
20. How can you automate EC2 backup by using EBS?
AWS EC2 instances can be backed up by creating snapshots of
EBS volumes. The snapshots are stored with the help of Amazon S3. Snapshots can
capture all the data contained in EBS volumes and create exact copies of this
data. The snapshots can then be copied and transferred into another AWS region,
ensuring safe and reliable storage of sensitive data.
Before running AWS EC2 backup, it is recommended to stop the
instance or detach the EBS volume that will be backed up. This ensures that any
failures or errors that occur will not affect newly created snapshots.
The following steps must be followed to back up an Amazon
EC2 instance:
- Sign
in to the AWS account, and launch the AWS console.
- Launch
the EC2 Management Console from the Services option.
- From
the list of running instances, select the instance that has to be backed
up.
- Find
the Amazon EBS volumes attached locally to that particular instance.
- List
the snapshots of each of the volumes, and specify a retention period for
the snapshots. A snapshot has to be created of each volume too.
- Remember
to remove snapshots that are older than the retention period.
21. Explain how one can add an existing instance to a new
Auto Scaling group?
To add an existing instance to a new Auto Scaling group:
- Open
the EC2 console.
- From
the instances, select the instance that is to be added
- Go
to Actions -> Instance Setting -> Attach to Auto Scaling Group
- Select
a new Auto Scaling group and link this particular group to the instance.
Scenario-Based
AWS Interview Questions
22. You have a web server on an EC2 instance. Your
instance can get to the web but nobody can get to your web server. How will you
troubleshoot this issue?
23. What steps will you perform to enable a server in a
private subnet of a VPC to download updates from the web?
24. How will you build a self-healing AWS cloud
architecture?
25. How will you design an Amazon Web Services cloud
architecture for failure?
26. As an AWS solution architect, how will you implement
disaster recovery on AWS?
27. You run a news website in the eu-west-1 region, which
updates every 15 minutes. The website is accessed by audiences across the globe
and uses an auto-scaling group behind an Elastic load balancer and Amazon
relation database service. Static content for the application is on Amazon S3
and is distributed using CloudFront. The auto-scaling group is set to trigger a
scale-up event with 60% CPU utilization. You use an extra large DB instance
with 10.000 Provisioned IOPS that gives CPU Utilization of around 80% with
freeable memory in the 2GB range. The web analytics report shows that the load
time for the web pages is an average of 2 seconds, but the SEO consultant
suggests that you bring the average load time of your pages to less than 0.5
seconds. What will you do to improve the website's page load time for your
users?
28. How will you right-size a system for normal and peak
traffic situations?
29. Tell us about a situation where you were given
feedback that made you change your architectural design strategy.
30. What challenges are you looking forward to for the
position as an AWS solutions architect?
31. Describe a successful AWS project which reflects your
design and implementation experience with AWS Solutions Architecture.
32. How will you design an e-commerce application using
AWS services?
33. What characteristics will you consider when designing
an Amazon Cloud solution?
34. When would you prefer to use provisioned IOPS over
Standard RDS storage?
35. What do you think AWS is missing from a solutions
architect’s perspective?
36. What if Google decides to host YouTube.com on AWS?
How will you design the solution architecture?
Explore Categories
Apache Hadoop Projects Apache Hive Projects Apache Hbase Projects Apache Pig Projects Hadoop HDFS Projects Apache Impala Projects Apache Flume Projects Apache Sqoop Projects Spark SQL Projects Spark GraphX Projects Spark Streaming Projects Spark MLlib Projects Apache Spark Projects PySpark Projects Apache Zepellin Projects Apache Kafka Projects Neo4j Projects Microsoft Azure Projects Google Cloud Projects GCP AWS Projects
AWS EC2 Interview Questions and Answers
37. Is it possible to cast-off S3 with EC2 instances? If
yes, how?
It is possible to cast-off S3 with EC2 instances using root
approaches backed by native occurrence storage.
38. How can you safeguard EC2 instances running on a VPC?

AWS Security groups associated with EC2 instances can help
you safeguard EC2 instances running in a VPC by providing security at the
protocol and port access level. You can configure both INBOUND and OUTBOUND
traffic to enable secured access for the EC2 instance.AWS security groups are
much similar to a firewall-they contain a set of rules which filter the traffic
coming into and out of an EC2 instance and deny any unauthorized access to EC2
instances.
39. How many EC2 instances can be used in a VPC?
There is a limit of running up to a total of 20 on-demand
instances across the instance family, you can purchase 20 reserved instances
and request spot instances as per your dynamic spot limit region.
40. What are some of the key best practices for security in
Amazon EC2?
- Create
individual AWS IAM (Identity and Access Management) users to control
access to your AWS recourses. Creating separate IAM users provides
separate credentials for every user, making it possible to assign
different permissions to each user based on the access requirements.
- Secure
the AWS Root account and its access keys.
- Harden
EC2 instances by disabling unnecessary services and applications by
installing only necessary software and tools on EC2 instances.
- Grant
the least privileges by opening up permissions that are required to
perform a specific task and not more than that. Additional permissions can
be granted as required.
- Define
and review the security group rules regularly.
- Have
a well-defined, strong password policy for all users.
- Deploy
anti-virus software on the AWS network to protect it from Trojans,
Viruses, etc.
41. A distributed application that processes huge amounts of
data across various EC2 instances. The application is designed to recover
gracefully from EC2 instance failures. How will you accomplish this in a
cost-effective manner?
An on-demand or reserved instance will not be ideal in this
case, as the task here is not continuous. Moreover. Launching an on-demand
instance whenever work comes up makes no sense because on-demand instances are
expensive. In this case, the ideal choice would be to opt for a spot instance
owing to its cost-effectiveness and no long-term commitments.
AWS S3 Interview Questions and Answers
42. Will you use encryption for S3?
It is better to consider encryption for sensitive data on S3
as it is a proprietary technology.
43. How can you send a request to Amazon S3?
Using the REST API or the AWS SDK wrapper libraries, which
wrap the underlying Amazon S3 REST API.
44. What is the difference between Amazon S3 and EBS?
|
|
Amazon S3 |
EBS |
|
Paradigm |
Object Store |
Filesystem |
|
Security |
Private Key or Public Key |
Visible only to your EC2 |
|
Redundancy |
Across data centers |
Within the data center |
|
Performance |
Fast |
Superfast |
45. Summarize the S3 Lifecycle Policy.
AWS provides a Lifecycle Policy in S3 as a storage cost
optimizer. In fact, it enables the establishment of data retention policies for
S3 objects within buckets. It is possible to manage data securely and set up
rules so that it moves between different object classes on a dynamic basis and
is removed when it is no longer required.
46. What does the AWS S3 object lock feature do?
We can store objects using the WORM
(write-once-read-many) approach using the S3 object lock. An S3 user can use
this feature to prevent data overwriting or deletion for a set period of time
or indefinitely. Several organizations typically use S3 object locks to satisfy
legal requirements that require WORM storage.
AWS IAM Interview Questions and Answers
47. What do you understand by AWS policies?

In AWS, policies are objects that regulate the permissions
of an entity (users, groups, or roles) or an AWS resource. In AWS, policies are
saved as JSON objects. Identity-based policies, resource-based policies,
permissions borders, Organizations SCPs, ACLs, and session policies are the six
categories of policies that AWS offers.
48. What does AWS IAM MFA support mean?
MFA refers to Multi-Factor Authentication. AWS IAM MFA adds
an extra layer of security by requesting a user's username and password and a
code generated by the MFA device connected to the user account for accessing
the AWS management console.
49. How do IAM roles work?
IAM Role is an IAM Identity formed in an AWS account and
granted particular authorization policies. These policies outline what each IAM
(Identity and Access Management) role is allowed and prohibited to perform
within the AWS account. IAM roles do not store login credentials or access
keys; instead, a temporary security credential is created specifically for each
role session. These are typically used to grant access to users, services, or
applications that need explicit permission to use an AWS resource.
50. What happens if you have an AWS IAM statement that
enables a principal to conduct an activity on a resource and another statement
that restricts that same action on the same resource?
If more than one statement is applicable, the Deny effect
always succeeds.
51. Which identities are available in the Principal element?
IAM roles & roles from within your AWS accounts are the
most important type of identities. In addition, you can define federated users,
role sessions, and a complete AWS account. AWS services like ec2, cloudtrail,
or dynamodb rank as the second most significant type of principal.
Unlock
the ProjectPro Learning Experience for FREE
AWS Cloud Engineer Interview Questions and Answers
52. If you have half of the workload on the public cloud
while the other half is on local storage, what architecture will you use for
this?
Hybrid Cloud Architecture.
53. What does an AWS Availability Zone mean?
AWS availability zones must be traversed to access the
resources that AWS has to offer. Applications will be designed
effectively for fault tolerance. Availability Zones have low latency
communications with one another to efficiently support fault tolerance.
54. What does "data center" mean for Amazon Web
Services (AWS)?
According to the Amazon Web Services concept, the data
center consists of the physical servers that power the offered AWS resources.
Each availability zone will certainly include one or more AWS data centers to
offer Amazon Web Services customers the necessary assistance and support.
55. How does an API gateway (rest APIs) track user requests?
As user queries move via the Amazon API Gateway REST APIs to
the underlying services, we can track and examine them using AWS X-Ray.
56. What distinguishes an EMR task node from a core node?
A core node comprises software components that execute
operations and store data in a Hadoop Distributed File System or HDFS.
There is always one core node in multi-node clusters. Software elements that
exclusively execute tasks are found in task nodes. Additionally, it is optional
and doesn't properly store data in HDFS.
AWS Technical Interview Questions
57. A content management system running on an EC2 instance
is approaching 100% CPU utilization. How will you reduce the load on the EC2
instance?
This can be done by attaching a load balancer to an auto
scaling group to efficiently distribute load among multiple instances.
58. What happens when you launch instances in Amazon VPC?
Each instance has a default IP address when launched in
Amazon VPC. This approach is considered ideal when connecting cloud resources
with data centers.
59. Can you modify the private IP address of an EC2 instance
while it is running in a VPC?
It is not possible to change the primary private IP
addresses. However, secondary IP addresses can be assigned, unassigned, or
moved between instances at any given point.
60. You are launching an instance under the free usage tier
from AMI, having a snapshot size of 50GB. How will you launch the instance
under the free usage tier?
It is not possible to launch this instance under the free
usage tier.
61. Which load balancer will you use to make routing
decisions at the application or transport layer that supports VPC or EC2?
Classic Load Balancer.
Terraform AWS Interview Questions and Answers
62. What is the Terraform provider?
Terraform is a platform for managing and configuring
infrastructure resources, including computer systems, virtual machines (VMs),
network switches, containers, etc. An API provider is in charge of meaningful
API interactions that reveal resources. Terraform works with a wide range of
cloud service providers.
63. Can we develop on-premise infrastructure using
Terraform?
It is possible to build on-premise infrastructure using
Terraform. We can choose from a wide range of options to determine which vendor
best satisfies our needs.
64. Can you set up several providers using Terraform?
Terraform enables multi-provider deployments, including SDN
management and on-premise applications like OpenStack and VMware.
65. What causes a duplicate resource error to be ignored
during terraform application?
Here are some of the possible reasons:
- Rebuild
the resources using Terraform after deleting them from the cloud
provider's API.
- Eliminate
some resources from the code to prevent Terraform from managing them.
- Import
resources into Terraform, then remove any code that attempts to copy them.
66. What is a Resource Graph in Terraform?
The resources are represented using a resource graph. You
can create and modify different resources at the same time. To change the
configuration of the graph, Terraform develops a strategy. It immediately
creates a framework to help us identify drawbacks.
Get confident to build end-to-end projects
Access to a curated library of 250+ end-to-end industry
projects with solution code, videos and tech support.
AWS Glue Interview Questions and Answers
67. How does AWS Glue Data Catalog work?
AWS Glue Data Catalog is a managed
AWS service that enables you to store, annotate, and exchange metadata in the
AWS Cloud. Each AWS account and region has a different set of AWS Glue Data
Catalogs. It establishes a single location where several systems can store and
obtain metadata to keep data in data silos and query and modify the data using
that metadata. AWS Identity and Access Management (IAM) policies restrict
access to the data sources managed by the AWS Glue Data Catalog.
68. What exactly does the AWS Glue Schema Registry do?
You can validate and control the lifecycle of streaming data
using registered Apache Avro schemas by the AWS Glue Schema Registry. Schema
Registry is useful for Apache Kafka, AWS Lambda, Amazon Managed Streaming for
Apache Kafka (MSK), Amazon Kinesis Data Streams, Apache Flink, and Amazon Kinesis Data Analytics for Apache Flink.
Prepare for Your Next Big Data Job Interview with Kafka Interview Questions and
Answers
69. What relationship exists between AWS Glue and AWS Lake
Formation?

The shared infrastructure of AWS Glue, which provides
serverless architecture, console controls, ETL code development, and task
monitoring, is beneficial for AWS Lake Formation.
70. How can AWS Glue Schema Registry keep applications
highly available?
The Schema Registry storage and control layer supports the
AWS Glue SLA, and the serializers and deserializers employ best-practice
caching techniques to maximize client schema availability.
71. What do you understand by the AWS Glue database?

The AWS Glue Data Catalog database is a container for
storing tables. You create a database when you launch a crawler or manually add
a table. The database list in the AWS Glue console contains a list of all of
your databases.
AWS Lambda Interview Questions and Answers
72. What are the best security techniques in Lambda?
In Lambda, you can find some of the best alternatives for
security. When it comes to limiting access to resources, you can use Identity
Access and Management. Another option that extends permissions is a privilege.
Access might be restricted to unreliable or unauthorized hosts. The security
group's regulations can be reviewed over time to maintain the pace.
73. What does Amazon elastic block store mean?

It is a virtual storage area network that allows for the
execution of tasks. Users do not need to worry about data loss even if a disk
in the RAID is damaged because it can accept flaws easily. Elastic Block Storage
allows for the provisioning and allocation of storage. It can also be linked to
the API if necessary.
74. How much time can an AWS Lambda function run for?
After making the requests to AWS Lambda, the entire
execution must occur within 300 seconds. Although the timeout is set at 3
seconds by default, you can change it to any value between 1 and 300 seconds.
75. Is the infrastructure that supports AWS Lambda
accessible?
No, the foundation on which AWS Lambda runs is inaccessible
after it begins managing the compute infrastructure on the user's behalf. It
enables Lambda to carry out health checks, deploy security patches, and execute
other standard maintenance.
76. Do the AWS Lambda-based functions remain operational if
the code or configuration changes?
Yes. When a Lambda function is updated, there will be a
limited timeframe, less than a minute—during which both the old and new
versions of the function can handle requests.
AWS Interview Questions and Answers for DevOps
77. What does AWS DevOps' CodePipeline mean?

AWS offers a service called CodePipeline that offers
continuous integration and continuous delivery features. It also offers
provisions for infrastructure upgrades. The user-defined set release model
protocols make it very simple to perform tasks like building, testing, and
deploying after each build.
78. How can AWS DevOps manage continuous integration and
deployment?

The source code for an application must be stored and
versioned using AWS Developer tools. The application is then built, tested, and
deployed automatically using the services to an AWS instance or a local
environment. When implementing continuous integration and deployment services,
it is better to start with CodePipeline and use CodeBuild and CodeDeploy as
necessary.
79. What role does CodeBuild play in the release process
automation?

Setting up CodeBuild first, then connecting it directly with
the AWS CodePipeline, makes it simple to set up and configure the release
process. This makes it possible to add build steps continually, and as a
result, AWS handles the processes for continuous integration and continuous
deployment.
80. Is it possible to use Jenkins and AWS CodeBuild together
with AWS DevOps?
It is simple to combine AWS CodeBuild with Jenkins to
perform and execute jobs in Jenkins. Creating and manually controlling each
worker node in Jenkins is no longer necessary because build jobs are pushed to
CodeBuild and then executed there.
81. How does CloudFormation vary from AWS Elastic Beanstalk?
AWS Elastic BeanStalk and CloudFormation are two core
services by AWS. Their architecture makes it simple for them to work together.
EBS offers an environment in which cloud-deployed applications can be deployed.
To manage the lifecycle of the apps, this is incorporated with CloudFormation's
tools. This makes using several AWS resources quite simple. This ensures great
scalability in terms of using it for various applications, from older
applications to container-based solutions.
Theoretical knowledge is not enough to crack any Big Data
interview. Get your hands dirty on Hadoop projects for
practice and master your Big Data skills!
AWS Interview Questions and Answers for Java Developer
82. Is one Elastic IP enough for all the instances you have
been running?
There are both public and private addresses for the
instances. Until the Amazon EC2 or instance is terminated or disabled, the
private and public addresses are still associated with them. Elastic addresses
can be used in place of these addresses, and they remain with the instance as
long as the user doesn't explicitly disconnect them. There will be a need for
more than one Elastic IP if numerous websites are hosted on an EC2 server.
83. What networking performance metrics can you expect when
launching instances in a cluster placement group?
The following factors affect network performance:
- Type
of instance
- Network
performance criteria
When instances are launched in a cluster placement group,
one should expect the following:
- Single
flow of 10 Gbps.
- 20
Gbps full-duplex
- The
network traffic will be restricted to 5 Gbps irrespective of the placement
unit.
84. What can you do to increase data transfer rates in
Snowball?
The following techniques can speed up data transport
solution in Snowballs:
- Execute
multiple copy operations simultaneously.
- Copy
data to a single snowball from many workstations.
- To
reduce the encryption overhead, it is best to transfer large files into
small batches of smaller files.
- Removing
any additional hops.
85. Consider a scenario where you want to change your
current domain name registration to Amazon Route S3 without affecting your web
traffic. How can you do it?
Below are the steps to transfer your domain name
registration to Amazon Route S3:
- Obtain
a list of the DNS records related to your domain name.
- Create
a hosted zone using the Route 53 Management Console, which will store your
domain's DNS records, and launch the transfer operation from there.
- Get
in touch with the domain name registrar you used to register. Examine the
transfer processes.
- When
the registrar communicates the need for the new name server delegations,
your DNS requests will be processed.
86. How do you send a request to Amazon S3?
There are different options for submitting requests to
Amazon S3:
- Use
REST APIs.
- Use
AWS SDK Wrapper Libraries.
AWS Interview Questions for Testers
87. Mention the default tables you receive while
establishing an AWS VPC.

When building an AWS VPC, we get the three default tables-
Network ACL, Security Group, and Route table.
88. How do you ensure the security of your VPC?
To regulate the security of your VPC, use security groups,
network access control lists (ACLs), and flow logs.
89. What does security group mean?
In AWS, security groups, which are essentially virtual
firewalls, are used to regulate the inbound and outbound traffic to instances.
You can manage traffic depending on different criteria, such as protocol, port,
and source and destination.
90. What purpose does the ELB gateway load balancer endpoint
serve?

Application servers in the service consumer virtual private
cloud (VPC) and virtual appliances in that VPC are connected privately using
ELB gateway load balancer endpoints.
91. How is a VPC protected by the AWS Network Firewall?

The stateful firewall by AWS Network firewall protects
against unauthorized access to your Virtual Private Cloud (VPC) by monitoring
connections and identifying protocols. This service's intrusion prevention
program uses active flow inspection to detect and rectify loopholes in security
using single-based detection. This AWS service employs web filtering to block
known malicious URLs.
AWS Data Engineer Interview Questions and Answers
92. What type of performance can you expect from Elastic
Block Storage service? How do you back it up and enhance the performance?
The performance of elastic block storage varies, i.e., it
can go above the SLA performance level and drop below it. SLA provides an
average disk I/O rate which can, at times, frustrate performance experts who
yearn for reliable and consistent disk throughput on a server. Virtual AWS instances do not behave this
way. One can back up EBS volumes through a graphical user interface like
elasticfox or the snapshot facility through an API call. Also, the performance
can be improved by using Linux software raid and striping across four volumes.
93. Imagine that you have an AWS application that requires
24x7 availability and can be down only for a maximum of 15 minutes. How will
you ensure that the database hosted on your EBS volume is backed up?
Automated backups are the key processes as they work in the
background without requiring manual intervention. Whenever there is a need to
back up the data, AWS API and AWS CLI play a vital role in automating the
process through scripts. The best way is to prepare for a timely backup of the
EBS of the EC2 instance. The EBS snapshot should be stored on Amazon S3 (Amazon
Simple Storage Service) and can be used to recover the database instance in
case of any failure or downtime.
94. You create a Route 53 latency record set from your
domain to a system in Singapore and a similar record to a machine in Oregon.
When a user in India visits your domain, to which location will he be routed?
Assuming that the application is hosted on an Amazon EC2
instance and multiple instances of the applications are deployed on different
EC2 regions. The request is most likely to go to Singapore because Amazon Route
53 is based on latency, and it routes the requests based on the location that
is likely to give the fastest response possible.
95. How will you access the data on EBS in AWS?
Elastic block storage, as the name indicates, provides
persistent, highly available, and high-performance block-level storage that can
be attached to a running EC2 instance. The storage can be formatted and mounted
as a file system, or the raw storage can be accessed directly.
96. How will you configure an instance with the application
and its dependencies and make it ready to serve traffic?
You can achieve this with the use of lifecycle hooks. They
are powerful as they let you pause the creation or termination of an instance
so that you can sneak peek in and perform custom actions like configuring the
instance, downloading the required files, and any other steps that are required
to make the instance ready. Every auto-scaling group can have multiple
lifecycle hooks.
Infosys AWS Interview Questions
97. What do AWS export and import mean?

AWS Import/Export enables you to move data across AWS
(Amazon S3 buckets, Amazon EBS snapshots, or Amazon Glacier vaults) using
portable storage devices.
98. What do you understand by AMI?

AMI refers to Amazon Machine Image. It's a
template that includes the details (an operating system, an application server,
and apps) necessary to start an instance. It is a replica of the AMI executing
as a virtual server in the cloud.
99. Define the relationship between an instance and AMI.
You can launch instances from a single AMI. An instance type
specifies the hardware of the host computer that hosts your instance. Each type
of instance offers different cloud computing and memory resources. Once an
instance has been launched, it becomes a standard host and can be used in the
same way as any other computer.
100. Compare AWS with OpenStack.
|
Services |
AWS |
OpenStack |
|
User Interface |
GUI-Console API-EC2 API CLI -Available |
GUI-Console API-EC2 API CLI -Available |
|
Computation |
EC2 |
Nova |
|
File Storage |
S3 |
Swift |
|
Block Storage |
EBS |
Cinder |
|
Networking |
IP addressing Egress, Load Balancing Firewall (DNS), VPC |
IP addressing load balancing firewall (DNS) |
|
Big Data |
Elastic MapReduce |
- |
If you are willing to leverage v
Here is a list of the 50 best
AWS interview questions
Q1) What is AWS?
Answer: AWS is a platform that provides
on-demand resources for hosting web services, storage, networking, databases,
and other resources over the internet. AWS enables you to pay for only the
resources you use, making it a more cost-effective option for businesses of all
sizes.
Q2) What is S3?
Answer: S3 is
the perfect storage solution for businesses of all sizes. You can store
unlimited data and only pay for what you use. Plus, the payment model is highly
flexible, so you can choose the plan that best suits your needs.
Q3) What is S3 used for?
Answer: S3 is used for various
of purposes, the most common of which is storage. Companies and individuals use
S3 to store data that they don’t want to keep on their own computers, such as
photos, videos, and documents. S3 can also be used to host websites and
applications.
Boost your earning potential
with AWS expertise. Explore our certified AWS Courses for a high-paying career
Q4) What are EBS volumes?
Answer: EBS can be thought of as
your very own block of elasticity. Attach it to an instance, and your data will
be preserved, regardless of what happens to the instance. Plus, it’s always
there when you need it – no deleting necessary!
Q5) What is stored in EBS
volume?
Answer: EBS is a storage system
that can be used to store persistent data for EC2 instances.
Q6) What is auto-scaling?
Answer: Auto scaling allows you
to automatically scale your instances up and down, depending on your CPU or
memory utilization. There are two components to auto-scaling: auto-scaling
groups and launch configurations.
Q7) What are the types of
volumes in EBS?
Answer: EBS offers a variety of
volume types to suit your needs, such as:
- General-purpose
- Provisioned IOPS
- Magnetic
- Cold HDD
- Throughput optimized
Q8) What are the pricing models
for EC2instances?
Answer: The pricing models for
EC2 instances come in all shapes and sizes like:
- On-demand
- Reserved
- Spot
- Scheduled
- Dedicated
Become AWS AWS Solutions
Architect in 6 weeks Book a free demo now
Q9) What are key pairs?
Answer: The security of your
instances and virtual machines relies on key pairs. These pairs consist of a
public-key and private-key, which are used to connect to your instances.
Q10) Tell us the volume types
for EC2 Instances.
Answer: Instance store
volumes and EBS
Q11) What are the different
types of instances?
Answer: Following are the types
of instances,
- General-purpose
- Computer Optimized
- Storage Optimized
- Memory-Optimized
- Accelerated Computing
Q12) What are the components of
AWS?
Answer: AWS is a comprehensive
cloud platform that consists of EC2, S3, Route53, EBS,
Cloudwatch, and Key-Paris.
Q13) What are the four
foundational services in AWS?
Answer: Compute, storage,
networking, database, and identity & access management (IAM).
Q14) what are the ways to
encrypt data in S3?
Answer:
- Server-Side Encryption -S3
(AES 256 encryption)
- Server-Side Encryption -KMS
(Key Management Service)
- Server-Side Encryption -C
(Client-Side)
Q15) What is VPC?
Answer: A VPC (Virtual
Private Cloud) is a private network in the cloud. It’s a way to create isolated
networks in AWS, similar to on-premises data centers. You can create a VPC with
public and private subnets and control access to resources in each subnet. You
can also route traffic between subnets and create VPN connections to your
on-premises network.
Build Your Career as a
AWS Solution Architect
- Live Projects
- Resume / Interview
Preparation
Q16) What is the difference
between VPC and subnet?
Answer: A subnet is an important
part of a VPC. A VPC can have all public subnets or a combination of public and
private subnets. A private subnet is a subnet that doesn’t have a route to the
internet gateway and can be used for VPNs.
Q17) What are edge locations?
Answer: Edge location is the
place where the best content will be cached. When a user tries to access some
content, the content will be searched in the edge location. If it is not
available then, the content will be made available from the original location,
but a copy will be stored in the edge location to be accessed more quickly in
the future.
Q18) What are the minimum and
maximum size of individual objects that you can store in S3?
Answer: The maximum size of an
individual object that you can store in S3 is 5TB, and the minimum size is 0
bytes.
Q19) What is the default storage
class in S3?
Answer: Amazon S3’s default
storage class is the S3 Standard. When you don’t specify a storage class, your
objects will be stored in this reliable and convenient class.
Q20) What are the parameters for
S3 pricing?
Answer: The pricing model for S3
is as below,
- Storage used
- Number of requests you
make
- Storage management
- Data transfer
- Transfer acceleration
Q21) What is an AWS glacier?
Answer: Amazon Glacier is a
low-cost cloud storage service for data that you don’t need right away. Store
your data archiving and backup needs with Amazon Glacier and save!
Q22) What is the maximum
individual archive that you can store in glaciers?
Answer: You can store a maximum
individual archive of up to 40 TB.
Q23) What is the difference
between S3 and Glacier?
Answer: Amazon S3 (Simple
Storage Service) is one of the most popular AWS cloud storage options. It
allows you to store and retrieve any amount of data at any time and from
anywhere on the network.
Amazon Glacier was built for the
long-term storage and digital archiving of data, making it a great option for
creating large backups for data recovery.
Q24) What is the prerequisite to
work with Cross-region replication in S3?
Answer: To use cross-region
replication, you must enable versioning on both the source and destination
buckets and make sure they are in different regions.
Q25) What are the different
storage classes in S3?
Answer: Following are the types
of storage classes in S3,
- Standard frequently
accessed
- Standard infrequently
accessed
- One-zone infrequently
accessed.
- Glacier
Q26) What is a Cloudwatch?
Answer: CloudWatch is the
perfect way to keep an eye on your AWS resources and applications in real-time.
With CloudWatch, you can track important metrics that will help you keep your
applications running smoothly.
Q27) What are the cloudwatch
metrics that are available for EC2 instances?
Answer: Disk reads, Diskwrites,
CPU utilization, network packets in, network packets out, networkIn,
networkOut, CPUCreditUsage, CPUCreditBalance.
Q28) What is CloudWatch vs
CloudTrail?
Answer: CloudWatch and
CloudTrail provide complementary services that help you keep an eye on your AWS
environment. CloudWatch focuses on the activity of AWS services and resources,
reporting on their health and performance.
CloudTrail, on the other hand,
is a log of all actions that have taken place inside your AWS environment,
so that you can track changes and activity over time.
Q29) Why do I need CloudWatch?
Answer: CloudWatch data and
reports empower users to keep track of application performance, resource use,
operational issues, and constraints. This helps organizations resolve technical
issues and optimize operations.
Q30) What are the types of
cloudwatch?
Answer: There are two types of
cloud monitoring: basic and detailed. Basic monitoring is free, while detailed
monitoring incurs a charge.
Q31) What is an AMI?
Answer: AMI is an acronym for
Amazon Machine Image. AMI is a template that stores the software required to
launch an Amazon instance.
Q32) What is an EIP?
Answer: EIP is an elastic IP
address that is perfect for dynamic cloud computing. When you need a static IP
address for your instances, EIP is the perfect solution.
Q33) What are reserved
instances?
Answer: Reserved instances are a
great way to ensure you always have a fixed capacity of EC2 instances
available. In reserved instances, you’ll have to get into a contract of 1 year
or 3 years. This will give you configurations, launch permission, and a block
device mapping that specifies the volume to attach to the instance when it is
launched.
Q34) What is cloudfront?
Answer: Cloudfront is the
perfect way to quickly and easily distribute your content. With its low latency
and high data transfer speeds, Cloudfront ensures that your viewers will have a
smooth and enjoyable experience.
Q35) What are roles?
Answer: Roles provide
permissions to resources in your account that you trust. With roles, you don’t
need to create any usernames or passwords – you can simply work with the
resources you need.
Q36) What are the policies in
AWS?
Answer: Policies are permissions
that you can attach to the users that you create.
Q37) What are AWS service-linked
roles?
Answer: AWS service-linked roles
provide you with an easy way to give the service the permissions it needs to
call other services on your behalf.
Q38) What are the ways to secure
access S3 bucket?
Answer: There are two ways that
you can control access to your S3 buckets,
- ACL -Access Control
List
- Bucket policies
Q39) What is VPC peering
connection?
Answer: Create a VPC peering
connection to connect your business’s two AWS accounts and enjoy the benefits
of instances in each VPC behaving as if they are in the same network.
Q40) How can you control the
security of your VPC?
Answer: You can use security
groups and NACLs to control the security of your VPC finely.
Q41) What are the database types
in RDS?
Answer: Following are the types
of databases in RDS,
- Aurora
- Oracle
- MYSQL server
- Postgresql
- MariaDB
- SQL server
Q42) What are the types of Route
53 Routing Policy?
Answer:
- Simple routing policy
- Weighted routing
policy
- Latency routing
policy
- Failover routing
policy
- Geolocation routing
policy
- Geoproximity routing
policy
- Multivalue answer routing
policy
Q43) What is the maximum size of
messages in SQS?
Answer: The maximum size of
messages in SQS is 256 KB.
Q44) What is a redshift?
Answer: Amazon redshift is a
data warehouse product that is fast, powerful, and fully managed. It can handle
petabytes of data in the cloud.
Q45) What are the different
types of storage gateway?
Answer: Following are the types
of storage gateway.
- File gateway
- Volume gateway
- Tape gateway
Q46) What are NAT gateways?
Answer: NAT stands for Network
Address Translation. It provides a way for instances in a private subnet to
connect to the internet while preventing the internet from initiating a
connection with those instances.
Q47) What is a snowball?
Answer: Snowball is the perfect
way to transport data to and from AWS. With its source appliances, you can move
large amounts of data quickly and securely, reducing your network costs and
transfer times.
Q48) What is VPC peering
connection?
Answer: With VPC peering, you
can connect your business’s two AWS accounts and enjoy the benefits of
instances in each VPC behaving as if they are in the same network. This can
help reduce costs and increase security.
Q49) What is SNS?
Answer: SNS is a powerful web
service that simplifies the process of notifications from the cloud. You can
configure SNS to receive email or message notifications in a matter of
minutes.
Q50) What is the difference
between SES and SNS?
Answer: SES takes care of the
engineering required to ensure the delivery of their e-mails. SNS is a secure
service that allows publishers and subscribers to communicate privately.
view:
1. What are the AWS components?
Below are the AWS components:
- Data
Management and Data Transfer
- Compute
and Networking
- Storage
- Automation
and Orchestration
- Operations
and Management
- Visualization
- Security
and Compliance
This question often comes up early in an interview,
particularly for entry-level roles. It’s a basic knowledge test, allowing you
to show that you understand the fundamentals. In most cases, a simple
list-based answer is sufficient, as that keeps your answer concise.
2. Which load balancer supports path-based routing?
The application load balancer supports path-based routing.
Like the question above, this is designed to test your
general knowledge of AWS and cloud computing. Usually, you don’t have to
provide any information beyond what’s shown above, as you’re essentially
showing that you understand a critical fact.
3. What is the availability zone and region in AWS?
A region represents a separate geographic area in AWS, and
availability zones are highly available data centers within each AWS region.
Also, each region has multiple isolated locations known as availability zones.
The code for the availability zone is its region code followed by a letter
identifier. The best example is us-east-1a.
Providing the example at the end lets you add that little
something extra to your response without going too far. It’s a way to highlight
your expertise when you’re answering AWS cloud engineer interview questions a
little bit more, which can make a difference.
4. Why is VPC needed?
VPC — or Virtual Private Cloud — is used to create a virtual
network in the AWS cloud. It provides complete control over a virtual
networking environment, including resource placement, connectivity, and
security.
The answer above is pretty concise, but it covers most of
the important points. However, if you want to take your answer to this and
similar AWS engineer interview questions up a notch, consider following it up
with an example from your past experience. For instance, mention a time you
utilized VPC in a project and the benefits you gained from doing so.
5. What types of instances do you know?
Below are the different types of AWS instances:
- General
Purpose
- Compute
Optimized
- Memory
Optimized
- Accelerated
Computing
- Storage
Optimized
The trick with AWS data engineer interview questions such as
these is the phrasing. Since the hiring manager is asking about the ones you
“know,” you may want to start with the instances you’re most experienced with.
Then, you can also mention ones that you’re familiar with, at least in a
general sense.
6. What is auto-scaling?
Auto-scaling monitors your applications and automatically
adjusts capacity to maintain a steady, predictable performance at the lowest
possible cost. It makes scaling simple with recommendations that allow you to
optimize performance, cost, or balance between them.
Here’s another opportunity to mention an example from your
past experience. If you successfully used auto-scaling to balance cost and
performance, discuss that project after you provide the definition to highlight
not just your knowledge but your ability to apply it effectively.
7. What restrictions apply to AWS Lambda function code?
AWS Lambda has the following restrictions:
- Maximum
disk space — 512 MB
- Default
deployment package size — 50 MB
- Memory
range — 128 to 3008 MB
- A
function's maximum execution timeout is 15 minutes
In most cases, simply listing the restrictions is enough to
answer this question well. However, you can also mention why the restrictions
are in place if you want to add something extra to your response.
8. How do you trace user requests to Rest APIs (API
Gateway)?
We can use AWS X-Ray to trace and analyze user requests as
they travel through your Amazon API Gateway REST APIs to the underlying
services.
Again, this is a question where you can follow up your basic
answer with an example. That can make your response more impressive.
9. What’s the difference between Amazon S3 and EC2?
The main difference between Amazon S3 and EC2 is that S3
stores large amounts of data while EC2 runs the servers in the cloud.
Along with mentioning that point, consider relaying examples
of times when you used each technology.
10. Explain how elasticity differs from scalability.
Elasticity enables you to meet dynamic changes, while
scalability provides the static increase in the workload. The main purpose of
elasticity is to match the resources allocated with the actual amount of
resources needed at any given point in time. Scalability handles the changing
needs of an application within the boundary of the infrastructure via
statically adding or removing resources to meet the application's demands if
needed.
The answer above is usually enough. However, time
permitting, you can include examples from your work history to drive the points
home.
10 AWS interview questions and answers for experienced
engineers
As you advance in the field, the interview process gets
increasingly challenging. Hiring managers have more expectations regarding your
knowledge of big data, database management, system operations, and more. You’re
more likely to engage with complex AWS technical interview questions than basic
ones like those outlined in the section above.
Fortunately, you can still prepare by reviewing AWS cloud
engineer interview questions and answers. They will give you a working
foundation, ensuring you’re ready for the types of questions the hiring manager
will likely ask.
Here are the top 10 AWS interview questions and answers for
experienced engineers:
1. What are the methods to deploy the latest application
package to the servers in the autoscaling group in AWS?
There are three deployment methods which are as follows:
- Using
Codepipeline — Deploy the code when the server gets created.
- Replacing
image — Deploy the code to a temporary server, take the image, and replace
the image in the autoscaling group.
- Changing
the start-up script/user data in the autoscaling launch config — Pull the
code and restart the application server whenever a new server is created.
If you have time, consider providing examples of when you’ve
used the various deployment methods. Even if you just briefly mention them, it
lets the hiring manager know you’re experienced. Plus, they can always ask
follow-up questions for more details.
2. Do we need to open outbound port 80/443 on the load
balancer's security group to allow return traffic from the server?
No, the outbound port is not required to be kept open.
You can leave your answer that simple if you like. However,
you may want to expand on your answer and explain why that’s the case if you
have the opportunity.
3. How do you encrypt the connection in transit to RDS?
You use the certificate bundle (TLS/SSL) created by RDS to
encrypt the connection in transit.
In most cases, a straightforward answer is fine here, but
you do have the option to expand. First, you could reference an example of a
time when you tackled that task. Second, you can talk about how the certificate
bundle encrypts the connection. The former highlights experience, while the
second showcases knowledge, so use the approach that makes the most sense for
your situation.
4. How do you enhance AWS security with best practices?
AWS security best practices are as follows:
- Use
accurate account information
- Use
MFA (multi-factor authentication)
- Validate
IAM roles
- Rotate
your keys
- Limit
security groups
Mention the best practices as a starting point. Then,
discuss how they impact security, using examples of the types of activity they
are meant to promote or prevent.
5. What are the advantages of Redshift in AWS?
The advantages of Redshift in AWS include:
- Wide
adoption
- Ease
of administration
- Ideal
for data lakes
- Ease
of querying
- Columnar
storage
- Performance
- Scalability
- Security
- Strong
AWS ecosystem
- Pricing
With this answer, you may want to steer away from a simple
list. Instead, list some advantages and explain why they’re beneficial based on
your experience. Then, present the remaining ones as a list, expanding on the
last point you share to make the end of your response more compelling.
6. What are the differences between a core node and task
node in EMR?
A core node contains software components that run tasks and
store data in a Hadoop Distributed File System, or HDFS. Multi-node clusters
have at least one core node. A task node contains software components that only
run tasks. Also, it does not store data in HDFS and is technically optional.
If you have an example from your work history that you can
use to demonstrate the difference, then consider doing so. Otherwise, a
fact-based response is sufficient.
7. How can you speed up data transfer in Amazon Snowball?
There are several ways to speed up data transfer in Amazon
Snowball, including:
- Use
the latest Mac or Linux Snowball client
- Batch
small files together
- Perform
multiple copy operations at one time
- Copy
from multiple workstations
- Transfer
directories but not files
- Don't
perform other operations on files during transfer
- Reduce
local network use
- Eliminate
unnecessary hops
Here’s another one of the AWS interview questions for cloud
developers where integrating some examples into your list can make for a
stronger answer. Reference a past project or workplace task you handled to
outline the difference a few of the techniques above make.
8. How do you upload a file larger than 100 MB in Amazon S3?
There are two main options for uploading a file larger than
100 MB in Amazon S3: use the AWS Command Line Interface or use the AWS SDK.
If you want to make your answer more impactful, you can
outline situations where one option may be better than another. Alternatively,
you can discuss your past experience using the approaches.
9. Describe AWS routing policies.
Below are the AWS routing policies:
- Simple
routing policy
- Failover
routing policy
- Geolocation
routing policy
- Geo-Proximity
routing (traffic flow only) policy
- Latency-based
routing policy
- IP-based
routing policy
- Multi-value
answer routing policy
- Weighted
routing policy
With AWS cloud interview questions like this, you typically
need to expand a little on each point. Since there are so many policies, just
briefly explain each policy’s general purpose in one sentence. Otherwise, your
response could end up being too long.
10. What are the consistency models in DynamoDB?
The consistency models in DynamoDB are as follows:
- Eventual
consistency model — maximizes the read throughput with low latency
- Strong
consistency model — provides updated data with high latency
You may want to touch on the benefits and drawbacks of each
option or outline use cases for them based on your past experience to make your
response more well-rounded.
https://flexiple.com/aws/interview-questions
1. Define and explain the three basic types of cloud
services and the AWS products that are built based on them.
2. What is the relation between the Availability Zone and
Region?
3. What is geo-targeting in CloudFront?
4. What are the steps involved in a CloudFormation
Solution?
5. How do you upgrade or downgrade a system with near-zero
downtime?
6. What are the tools and techniques that you can use in
AWS to identify if you are paying more than you should be, and how to correct
it?
7. Is there any other alternative tool to log into the
cloud environment other than the console?
8. What are the native AWS Security logging capabilities?
9. You are trying to provide a service in a particular
region, but you do not see the service in that region. Why is this happening,
and how do you fix it?
Your engineers should not be hiring. They should be
coding.
Help your team focus on what they were hired for. Flexiple
will manage your entire hiring process and scale your tech team.
Hire dream developers now
10. What are the different types of virtualization in AWS,
and what are the differences between them?
11. What are the differences between NAT Gateways and NAT
Instances?
12. What is an Elastic Transcoder?
13. What are the core components of AWS?
14. What is an EC2 instance, and how is it different from
traditional servers?
15. What is the AWS Well-Architected Framework, and why is
it important?
16. What is an IAM role in AWS, and why is it used?
17. How do you secure data at rest in AWS?
18. What is AWS Lambda, and what are its key benefits
19. Can you explain the concept of an Availability Zone
(AZ)?
Your engineers should not be hiring. They should be
coding.
Help your team focus on what they were hired for. Flexiple
will manage your entire hiring process and scale your tech team.
Hire dream developers now
20. How do you ensure high availability in AWS?
21. What is an Auto Scaling group, and why is it used?
22. Explain the difference between RDS and DynamoDB.
24. What is the difference between an EC2 instance and an
AMI?
25. How do you back up data in AWS?
26. What is the AWS Shared Responsibility Model?
27. What is Amazon Route 53, and what is its primary use?
29. What is the purpose of an S3 bucket policy?
30. What is Amazon Redshift, and when would you use it?
31. What is AWS Identity and Access Management (IAM)?
Your engineers should not be hiring. They should be
coding.
Help your team focus on what they were hired for. Flexiple
will manage your entire hiring process and scale your tech team.
Hire dream developers now
33. What is AWS Elastic Beanstalk, and how does it
simplify application deployment?
34. What is Amazon SNS, and how is it used for
notifications?
35. How do you monitor AWS resources and applications?
36. What is CloudWatch?
37. Name some of the AWS services that are not
region-specific.
38. How do you set up a system to monitor website metrics
in real-time in AWS?
39. What is a DDoS attack, and what services can minimize
them?
40. What services can be used to create a centralized
logging solution?
What are the AWS Interview Questions for Intermediate and
Experienced?
The AWS interview questions for intermediate and
experienced candidates delve into the deeper intricacies of the AWS ecosystem.
AWS interview questions for intermediate and experienced candidates explore
areas such as troubleshooting a failed multi-AZ RDS failover, distinguishing
between S3 storage classes like Intelligent-Tiering and One Zone-IA, and
pinpointing best practices for optimizing AWS Lambda function performance.
Questions of advanced level are designed for those with
substantial hands-on experience. Topics range from creating a multi-region,
highly available application using AWS services, to cost-optimization
strategies for data-intensive applications on EMR. Advanced developers have
more than five years of dedicated AWS experience. The importance of these
questions is not overstated. They gauge the depth of an individual's AWS
expertise and also differentiate between beginners, intermediates, and experts.
Mastery of these questions signifies that a candidate is well-prepared to
handle complex AWS challenges in real-world settings.
41. What is AWS's global infrastructure and the concept of
regions and availability zones?
42. Explain AWS Organizations and how it simplifies
managing multiple AWS accounts.
Your engineers should not be hiring. They should be
coding.
Help your team focus on what they were hired for. Flexiple
will manage your entire hiring process and scale your tech team.
Hire dream developers now
43. How can cost optimization be achieved in AWS, and what
tools can help?
44. Describe the AWS Well-Architected Framework and its
pillars.
45. What is the purpose of AWS Trusted Advisor?
46. Compare EC2 and AWS Lambda in terms of use cases and
scaling.
47. How do you choose the right EC2 instance type?
48. Differentiate between EC2 On-Demand Instances,
Reserved Instances, and Spot Instances.
49. What is AWS Elastic Beanstalk, and when should it be
used?
50. Explain Amazon ECS (Elastic Container Service) and its
benefits.
51. Describe Amazon VPC and its role in network isolation.
52. How do you set up VPN connections to an Amazon VPC with
AWS Direct Connect?
Your engineers should not be hiring. They should be
coding.
Help your team focus on what they were hired for. Flexiple
will manage your entire hiring process and scale your tech team.
Hire dream developers now
53. Explain AWS Security Groups and Network ACLs.
54. What is AWS IAM, and how is least privilege access
enforced?
55. What is AWS WAF, and how does it protect web
applications?
56. Describe Amazon S3 storage classes and their use
cases.
57. Explain the differences between AWS EFS and AWS EBS.
58. How can you transfer large datasets securely to AWS
using Snowball or Snowmobile?
59. Describe AWS Storage Gateway's role in hybrid cloud
storage.
60. What are Amazon RDS Multi-AZ deployments, and why are
they important?
61. Compare Amazon RDS, Amazon Aurora, and Amazon DynamoDB
for different workloads.
62. What is Amazon Redshift, and what are its advantages
for data warehousing?
Your engineers should not be hiring. They should be
coding.
Help your team focus on what they were hired for. Flexiple
will manage your entire hiring process and scale your tech team.
Hire dream developers now
63. How does AWS Database Migration Service (DMS)
facilitate database migrations?
64. Explain Amazon Neptune's role in graph database
applications.
65. How do you optimize database performance in Amazon
DynamoDB?
66. What is AWS CodePipeline, and how does it enable
CI/CD?
67. Describe AWS CodeBuild and AWS CodeDeploy in the CI/CD
pipeline.
68. What is AWS Elastic Beanstalk's primary purpose?
69. How is AWS CloudFormation used for infrastructure as
code (IAC)?
70. What are Blue-Green deployments, and how are they
implemented in AWS?
What are the Advanced AWS Interview Questions?
Advanced AWS interview questions revolve around complex
scenarios and real-world use cases. These questions delve into areas such as
optimization strategies for AWS services, intricate AWS architecture designs,
security best practices, and cost-management techniques.
Advanced AWS Questions involve Scenario-based questions
that present hypothetical, real-world situations requiring the candidate to
solve or suggest best practices. For instance, an interviewer ask, "How
would you design a fault-tolerant system using AWS services?", or
"Recommend a strategy to migrate a large database to AWS without
downtime."
Scenario-based questions test a candidate's practical
knowledge and ability to apply AWS solutions in real-world settings. They
evaluate problem-solving skills, depth of AWS expertise, and adaptability to
unforeseen challenges. Employers prioritize these qualities, ensuring
candidates are able to handle on-the-job tasks effectively.
71. Explain AWS Key Management Service (KMS) and its role
in data security.
72. Distinguish between AWS CloudFormation and Elastic
Beanstalk for infrastructure as code (IAC).
Your engineers should not be hiring. They should be
coding.
Help your team focus on what they were hired for. Flexiple
will manage your entire hiring process and scale your tech team.
Hire dream developers now
73. Describe AWS Step Functions and their usage in
building serverless workflows.
74. What is AWS Lambda@Edge, and when would you use it?
75. Explain AWS Direct Connect and its importance in
hybrid cloud scenarios.
76. Define AWS PrivateLink and its significance for
securing access to AWS services.
77. How do you set up and manage AWS VPC peering, and what
challenges might arise?
78. What is Amazon Cognito, and how can it be used for
user authentication and authorization?
79. Describe the use cases and architecture of Amazon
Elastic File System (EFS).
80. How would you architect a highly available and
scalable web application using AWS services?
What are the Technical and Non-Technical AWS Interview
Questions?
Technical AWS interview questions focus primarily on
specific AWS services, their applications, and best practices. Examples include
queries on Amazon S3 bucket policies, Elastic Load Balancer configurations, and
AWS Lambda functions. Non-technical AWS questions center around general cloud
strategies, cost management, and AWS use-case scenarios.
These questions include various miscellaneous questions
that cover a wide range of topics, such as AWS case studies or emerging trends
in 2024. These questions are best for collaborative works, as they stimulate
discussions on diverse AWS topics. Seeking new AWS innovations or services
becomes crucial; hence, the importance of these miscellaneous questions is
undeniable, as they keep candidates updated and adaptable.
81. Explain EC2 instance types and selection criteria.
82. Compare S3 and EBS storage options.
Your engineers should not be hiring. They should be
coding.
Help your team focus on what they were hired for. Flexiple
will manage your entire hiring process and scale your tech team.
Hire dream developers now
83. How does AWS Lambda work, and what are common use
cases?
84. Outline the basics of an AWS Virtual Private Cloud
(VPC).
85. What is AWS IAM, and how do you manage access control?
86. How do you create an AWS CloudFormation template
87. What is AWS Elastic Beanstalk, and how does it
simplify deployment?
88. Explain Amazon RDS Multi-AZ deployment.
89. How do you monitor AWS resources using CloudWatch and
CloudTrail?
90. Describe Amazon VPC peering and Direct Connect.
91. Describe your experience with AWS and cloud
technologies.
92. How do you stay updated with AWS best practices and
trends?
Your engineers should not be hiring. They should be
coding.
Help your team focus on what they were hired for. Flexiple
will manage your entire hiring process and scale your tech team.
Hire dream developers now
93. Share an experience addressing cost optimization in
AWS.
94. Provide an example of a challenging AWS situation you
resolved.
95. Explain how you collaborate in diverse teams on AWS
projects.
96. How do you handle prioritizing and managing multiple
AWS tasks?
97. Discuss key considerations for AWS data security and
compliance.
98. Share your experience with AWS migrations or
large-scale deployments.
99. How do you ensure knowledge sharing within your team
on AWS solutions?
100. Explain how you communicate technical concepts to
non-technical stakeholders.
What is AWS?
https://www.simplilearn.com/tutorials/aws-tutorial/aws-interview-questions
Basic AWS Interview Questions
1. Define and explain the three
basic types of cloud services and the AWS products that are built based on
them?
The three basic types of cloud
services are:
- Computing
- Storage
- Networking
Here are some of the AWS
products that are built based on the three cloud service types:
Computing - These include EC2,
Elastic Beanstalk, Lambda, Auto-Scaling, and Lightsat.
Storage - These include S3,
Glacier, Elastic Block Storage, Elastic File System.
Networking - These include
VPC, Amazon CloudFront, Route53
2. What is the relation between
the Availability Zone and Region?
AWS regions are separate
geographical areas, like the US-West 1 (North California) and Asia South
(Mumbai). On the other hand, availability zones are the areas that are present
inside the regions. These are generally isolated zones that can replicate themselves
whenever required.

3. What is auto-scaling?
Auto-scaling is a function that
allows you to provision and launch new instances whenever there is a demand. It
allows you to automatically increase or decrease resource capacity in relation
to the demand.
4. What is geo-targeting in
CloudFront?
Geo-Targeting is a concept where
businesses can show personalized content to their audience based on their
geographic location without changing the URL. This helps you create customized
content for the audience of a specific geographical area, keeping their needs
in the forefront.
5. What are the steps involved
in a CloudFormation Solution?
Here are the steps involved in a
CloudFormation solution:

- Create or use an existing
CloudFormation template using JSON or YAML format.
- Save the code in an S3
bucket, which serves as a repository for the code.
- Use AWS CloudFormation to
call the bucket and create a stack on your template.
- CloudFormation reads the
file and understands the services that are called, their order, the
relationship between the services, and provisions the services one after
the other.
6. How do you upgrade or
downgrade a system with near-zero downtime?
You can upgrade or downgrade a
system with near-zero downtime using the following steps of migration:
- Open EC2 console
- Choose Operating System AMI
- Launch an instance with the
new instance type
- Install all the updates
- Install applications
- Test the instance to see if
it’s working
- If working, deploy the new
instance and replace the older instance
- Once it’s deployed, you can
upgrade or downgrade the system with near-zero downtime.
Take home these interview
Q&As and get much more. Download the complete AWS Interview Guide here:
7. What are the tools and
techniques that you can use in AWS to identify if you are paying more than you
should be, and how to correct it?
You can know that you are paying
the correct amount for the resources that you are using by employing the
following resources:
- Check the Top Services
Table
It is a dashboard in the cost
management console that shows you the top five most used services. This will
let you know how much money you are spending on the resources in question.
- Cost Explorer
There are cost explorer services
available that will help you to view and analyze your usage costs for the last
13 months. You can also get a cost forecast for the upcoming three months.
- AWS Budgets
This allows you to plan a budget
for the services. Also, it will enable you to check if the current plan meets
your budget and the details of how you use the services.
- Cost Allocation Tags
This helps in identifying the
resource that has cost more in a particular month. It lets you organize your
resources and cost allocation tags to keep track of your AWS costs.
Learn From The Best AWS Mentors!
AWS Solutions Architect
Certification TrainingExplore Program
![]()
8. Is there any other
alternative tool to log into the cloud environment other than console?
These can help you log into the
AWS resources:
- Putty
- AWS CLI for Linux
- AWS CLI for Windows
- AWS CLI for Windows CMD
- AWS SDK
- Eclipse
9. What services can be used to
create a centralized logging solution?
The essential services that you
can use are Amazon CloudWatch Logs, store them in Amazon S3, and then use Amazon Elastic Search to
visualize them. You can use Amazon Kinesis Firehose to move the data from
Amazon S3 to Amazon ElasticSearch.

10. What are the native AWS
Security logging capabilities?
Most of the AWS services have
their logging options. Also, some of them have account-level logging, like in
AWS CloudTrail, AWS Config, and others. Let’s take a look at two services in
specific:
AWS CloudTrail
This is a service that provides
a history of the AWS API calls for every account. It lets you perform security
analysis, resource change tracking, and compliance auditing of your AWS
environment as well. The best part about this service is that it enables you to
configure it to send notifications via AWS SNS when new logs are delivered.
AWS Config
This helps you understand the
configuration changes that happen in your environment. This service provides an
AWS inventory that includes configuration history, configuration change
notification, and relationships between AWS resources. It can also be configured
to send information via AWS SNS when new logs are delivered.
11. What is a DDoS attack, and
what services can minimize them?
DDoS is a cyber-attack in which
the perpetrator accesses a website and creates multiple sessions so that the
other legitimate users cannot access the service. The native tools that can
help you deny the DDoS attacks on your AWS services are:
- AWS Shield
- AWS WAF
- Amazon Route53
- Amazon CloudFront
- ELB
- VPC

12. You are trying to provide a
service in a particular region, but you do not see the service in that region.
Why is this happening, and how do you fix it?
Not all Amazon AWS services are
available in all regions. When Amazon initially launches a new service, it
doesn’t get immediately published in all the regions. They start small and then
slowly expand to other regions. So, if you don’t see a specific service in your
region, chances are the service hasn’t been published in your region yet.
However, if you want to get the service that is not available, you can switch
to the nearest region that provides the services.
13. How do you set up a system
to monitor website metrics in real-time in AWS?
Amazon CloudWatch helps you to
monitor the application status of various AWS services and custom events. It
helps you to monitor:
- State changes in Amazon EC2
- Auto-scaling lifecycle
events
- Scheduled events
- AWS API calls
- Console sign-in events

14. What are the different types
of virtualization in AWS, and what are the differences between them?
The three major types of
virtualization in AWS are:
- Hardware Virtual Machine
(HVM)
It is a fully virtualized
hardware, where all the virtual machines act separate from each other. These
virtual machines boot by executing a master boot record in the root block
device of your image.
- Paravirtualization (PV)
Paravirtualization-GRUB is the
bootloader that boots the PV AMIs. The PV-GRUB chain loads the kernel specified
in the menu.
- Paravirtualization on HVM
PV on HVM helps operating
systems take advantage of storage and network I/O available through the host.
15. Name some of the AWS
services that are not region-specific
AWS services that are not
region-specific are:
- IAM
- Route 53
- Web Application
Firewall
- CloudFront
Want a Job at AWS? Find Out What
It Takes
Cloud Architect Master's ProgramExplore Program
![]()
16. What are the differences
between NAT Gateways and NAT Instances?
While both NAT Gateways and NAT
Instances serve the same function, they still have some key differences.

17. What is CloudWatch?
The Amazon CloudWatch has the
following features:
- Depending on multiple
metrics, it participates in triggering alarms.
- Helps in monitoring the AWS
environments like CPU utilization, EC2, Amazon RDS instances, Amazon SQS,
S3, Load Balancer, SNS, etc.
18. What is an Elastic
Transcoder?
To support multiple devices with
various resolutions like laptops, tablets, and smartphones, we need to change
the resolution and format of the video. This can be done easily by an AWS
Service tool called the Elastic Transcoder, which is a media transcoding in the
cloud that exactly lets us do the needful. It is easy to use, cost-effective,
and highly scalable for businesses and developers.
AWS Interview Questions for
Intermediate and Experienced
19. With specified private IP
addresses, can an Amazon Elastic Compute Cloud (EC2) instance be launched? If
so, which Amazon service makes it possible?
Yes. Utilizing VPC makes it
possible (Virtual Private Cloud).
20. Define Amazon EC2 regions
and availability zones?
Availability zones are
geographically separate locations. As a result, failure in one zone has no
effect on EC2 instances in other zones. When it comes to regions, they may have
one or more availability zones. This configuration also helps to reduce latency
and costs.
21. Explain Amazon EC2 root
device volume?
The image that will be used to
boot an EC2 instance is stored on the root device drive. This occurs when an
Amazon AMI runs a new EC2 instance. And this root device volume is supported by
EBS or an instance store. In general, the root device data on Amazon EBS is not
affected by the lifespan of an EC2 instance.
22. Mention the different types
of instances in Amazon EC2 and explain its features.
- General Purpose Instances:
They are used to compute a range of workloads and aid in the allocation of
processing, memory, and networking resources.
- Compute Optimized
Instances: These are ideal for compute-intensive applications. They can
handle batch processing workloads, high-performance web servers,
machine learning inference, and various other tasks.
- Memory Optimized: They
process workloads that handle massive datasets in memory and deliver them
quickly.
- Accelerated Computing: It
aids in the execution of floating-point number calculations, data pattern
matching, and graphics processing. These functions are carried out using
hardware accelerators.
- Storage Optimised: They
handle tasks that require sequential read and write access to big data
sets on local storage.
23. Will your standby RDS be
launched in the same availability zone as your primary?
No, standby instances are
launched in different availability zones than the primary, resulting in
physically separate infrastructures. This is because the entire purpose of
standby instances is to prevent infrastructure failure. As a result, if the
primary instance fails, the backup instance will assist in recovering all of
the data.
Want a Job at AWS? Find Out What
It Takes
Cloud Architect Master's ProgramExplore Program
![]()
Advanced AWS Interview Questions
and Answers
24. What is the difference
between a Spot Instance, an On-demand Instance, and a Reserved Instance?
Spot instances are unused EC2
instances that users can use at a reduced cost.
When you use on-demand
instances, you must pay for computing resources without making long-term
obligations.
Reserved instances, however,
allow you to specify attributes such as instance type, platform, tenancy,
region, and availability zone. When instances in certain availability zones are
used, reserved instances offer significant reductions and capacity reservations.
25. How would you address a
situation in which the relational database engine frequently collapses when
traffic to your RDS instances increases, given that the RDS instance replica is
not promoted as the master instance?
A larger RDS instance type is
required for handling significant quantities of traffic, as well as producing
manual or automated snapshots to recover data if the RDS instance fails.
26. What do you understand by
'changing' in Amazon EC2?
To make limit administration
easier for customers, Amazon EC2 now offers the option to switch from the
current 'instance count-based limitations' to the new 'vCPU Based
restrictions.' As a result, when launching a combination of instance types
based on demand, utilization is measured in terms of the number of vCPUs.
27. Define Snapshots in Amazon
Lightsail?
The point-in-time backups of EC2
instances, block storage drives, and databases are known as snapshots. They can
be produced manually or automatically at any moment. Your resources can always
be restored using snapshots, even after they have been created. These resources
will also perform the same tasks as the original ones from which the snapshots
were made.
AWS Scenario-based Questions
28. On an EC2 instance, an
application of yours is active. Once the CPU usage on your instance hits 80%,
you must reduce the load on it. What strategy do you use to complete the task?
It can be accomplished by
setting up an autoscaling group to deploy additional instances, when an EC2
instance's CPU use surpasses 80% and by allocating traffic across instances via
the creation of an application load balancer and the designation of EC2 instances
as target instances.
29. Multiple Linux Amazon EC2
instances running a web application for a firm are being used, and data is
being stored on Amazon EBS volumes. The business is searching for a way to
provide storage that complies with atomicity, consistency, isolation, and durability
while also increasing the application's resilience in the event of a breakdown
(ACID). What steps should a solutions architect take to fulfill these demands?
AWS Auto Scaling groups can
create an application load balancer that spans many availability zones. Mount a
target on each instance and save data on Amazon EFS.
30. Your business prefers to use
its email address and domain to send and receive compliance emails. What
service do you recommend to implement it easily and budget-friendly?
This can be accomplished by
using Amazon Simple Email Service (Amazon SES), a cloud-based
email-sending service.
Technical and Non-Technical AWS
Interview Questions
31. Describe SES.
Amazon offers the Simple Email
Service (SES) service, which allows you to send bulk emails to customers
swiftly at a minimal cost.
32. Describe PaaS.
PaaS supports the operation of
multiple cloud platforms, primarily for the development, testing, and
oversight of the operation of the program.
33. How many S3 buckets can be
created?
Up to 100 buckets can be created
by default.
34. What is the maximum limit of
elastic IPs anyone can produce?
A maximum of five elastic IP
addresses can be generated per location and AWS account.
AWS Questions for Amazon EC2
35. What is Amazon EC2?
EC2 is short for Elastic Compute Cloud, and it
provides scalable computing capacity. Using Amazon EC2 eliminates the need to
invest in hardware, leading to faster development and deployment of
applications. You can use Amazon EC2 to launch as many or as few
virtual servers as needed, configure security and networking, and manage
storage. It can scale up or down to handle changes in requirements, reducing
the need to forecast traffic. EC2 provides virtual computing environments
called “instances.”
36. What Are Some of the
Security Best Practices for Amazon EC2?
Security best practices for
Amazon EC2 include using Identity and Access Management (IAM) to control access
to AWS resources; restricting access by only allowing trusted hosts or networks
to access ports on an instance; only opening up those permissions you require,
and disabling password-based logins for instances launched from your AMI.
37. Can S3 Be Used with EC2
Instances, and If Yes, How?
Amazon S3 can be used for
instances with root devices backed by local instance storage. That way,
developers have access to the same highly scalable, reliable, fast, inexpensive
data storage infrastructure that Amazon uses to run its own global network of
websites. To execute systems in the Amazon EC2 environment, developers load
Amazon Machine Images (AMIs) into Amazon S3 and then move them between Amazon
S3 and Amazon EC2.
Amazon EC2 and Amazon S3 are two
of the best-known web services that make up AWS.
38. What is the difference
between stopping and terminating an EC2 instance?
While you may think that both
stopping and terminating are the same, there is a difference. When you stop an
EC2 instance, it performs a normal shutdown on the instance and moves to a
stopped state. However, when you terminate the instance, it is transferred to a
stopped state, and the EBS volumes attached to it are deleted and can never be
recovered.
39. What are the different types
of EC2 instances based on their costs?
The three types of EC2 instances
are:
- On-demand Instance
It is cheap for a short time but
not when taken for the long term
- Spot Instance
It is less expensive than the on-demand
instance and can be bought through bidding.
- Reserved Instance
If you are planning to use an
instance for a year or more, then this is the right one for you.
Build and Deploy Azure
Applications Like a Pro!
Azure Cloud ArchitectExplore Program
![]()
40. How do you set up SSH agent
forwarding so that you do not have to copy the key every time you log in?
Here’s how you accomplish this:
- Go to your PuTTY
Configuration
- Go to the category SSH
-> Auth
- Enable SSH agent forwarding
to your instance

41. What are Solaris and AIX
operating systems? Are they available with AWS?
Solaris is an operating system
that uses SPARC processor architecture, which is not supported by the public
cloud currently.
AIX is an operating system that
runs only on Power CPU and not on Intel, which means that you cannot create AIX
instances in EC2.
Since both the operating systems
have their limitations, they are not currently available with AWS.
42. How do you configure
CloudWatch to recover an EC2 instance?
Here’s how you can configure
them:
- Create an Alarm using
Amazon CloudWatch
- In the Alarm, go to Define
Alarm -> Actions tab
- Choose Recover this
instance option

43. What are the common types of
AMI designs?
There are many types of AMIs,
but some of the common AMIs are:
- Fully Baked AMI
- Just Enough Baked AMI (JeOS
AMI)
- Hybrid AMI
44. What are Key-Pairs in AWS?
The Key-Pairs are
password-protected login credentials for the Virtual Machines that are used to
prove our identity while connecting the Amazon EC2 instances. The Key-Pairs are
made up of a Private Key and a Public Key which lets us connect to the
instances.
AWS Interview Questions for S3
45. What is Amazon S3?
S3 is short for Simple Storage
Service, and Amazon
S3 is the most supported storage platform available. S3 is object
storage that can store and retrieve any amount of data from anywhere. Despite
that versatility, it is practically unlimited as well as cost-effective because
it is storage available on demand. In addition to these benefits, it offers
unprecedented levels of durability and availability. Amazon S3 helps to manage
data for cost optimization, access control, and compliance.
Iginite Your Knowledge in AWS
Cloud Operations
Cloud Operations on AWSENROLL NOW
![]()
46. How can you recover/login to
an EC2 instance for which you have lost the key?
Follow the steps provided below
to recover an EC2 instance if you have lost the key:
- Verify that the EC2Config
service is running
- Detach the root volume for
the instance
- Attach the volume to a
temporary instance
- Modify the configuration
file
- Restart the original
instance
47. What are some critical
differences between AWS S3 and EBS?
Here are some differences
between AWS S3 and EBS:

48. How do you allow a user to
gain access to a specific bucket?
You need to follow the four
steps provided below to allow access. They are:
- Categorize your instances
- Define how authorized users
can manage specific servers.
- Lockdown your tags
- Attach your policies to IAM
users
49. How can you monitor S3
cross-region replication to ensure consistency without actually checking the
bucket?
Follow the flow diagram provided
below to monitor S3 cross-region replication:

50. What is SnowBall?
To transfer terabytes of data
outside and inside of the AWS environment, a small application called SnowBall
is used.
Data transferring using SnowBall
is done in the following ways:
- A job is created.
- The SnowBall application is
connected.
- The data is copied into the
SnowBall application.
- Data is then moved to the
AWS S3.
51. What are the Storage Classes
available in Amazon S3?
The Storage Classes that are
available in the Amazon S3 are the following:
- Amazon S3 Glacier Instant
Retrieval storage class
- Amazon S3 Glacier Flexible
Retrieval (Formerly S3 Glacier) storage class
- Amazon S3 Glacier Deep
Archive (S3 Glacier Deep Archive)
- S3 Outposts storage class
- Amazon S3 Standard-Infrequent
Access (S3 Standard-IA)
- Amazon S3 One
Zone-Infrequent Access (S3 One Zone-IA)
- Amazon S3 Standard (S3
Standard)
- Amazon S3 Reduced
Redundancy Storage
- Amazon S3
Intelligent-Tiering (S3 Intelligent-Tiering)
Boost Your Cloud Skills. Land an
Architect Role
Cloud Architect Master's ProgramExplore Program
![]()
AWS Interview Questions for VPC
52. What Is Amazon Virtual
Private Cloud (VPC) and Why Is It Used?
A VPC is the best way of connecting to your
cloud resources from your own data center. Once you connect your datacenter to
the VPC in which your instances are present, each instance is assigned a
private IP address that can be accessed
from your data center. That way, you can access your public cloud resources as
if they were on your own private network.
53. VPC is not resolving the
server through DNS. What might be the issue, and how can you fix it?
To fix this problem, you need to
enable the DNS hostname resolution, so that the problem resolves itself.
54. How do you connect multiple
sites to a VPC?
If you have multiple VPN
connections, you can provide secure communication between sites using the AWS
VPN CloudHub. Here’s a diagram that will show you how to connect various sites
to a VPC:

55. Name and explain some
security products and features available in VPC?
Here is a selection of security
products and features:
- Security groups - This acts
as a firewall for the EC2 instances, controlling inbound and outbound
traffic at the instance level.
- Network access control
lists - It acts as a firewall for the subnets, controlling inbound and
outbound traffic at the subnet level.
- Flow logs - These capture
the inbound and outbound traffic from the network interfaces in your VPC.
56. How do you monitor Amazon
VPC?
You can monitor VPC by using:
- CloudWatch and CloudWatch
logs
- VPC Flow Logs
57. How many Subnets can you
have per VPC?
We can have up to 200 Subnets
per Amazon Virtual Private Cloud (VPC).
General AWS Interview Questions
58. When Would You Prefer
Provisioned IOPS over Standard Rds Storage?
You would use Provisioned IOPS
when you have batch-oriented workloads. Provisioned IOPS delivers high IO
rates, but it is also expensive. However, batch processing workloads do not
require manual intervention.
59. How Do Amazon Rds, Dynamodb,
and Redshift Differ from Each Other?
Amazon RDS is a database
management service for relational databases. It manages patching, upgrading,
and data backups automatically. It’s a database management service for
structured data only. On the other hand, DynamoDB is a NoSQL database service
for dealing with unstructured data. Redshift is a data warehouse product used
in data analysis.
60. What Are the Benefits of
AWS’s Disaster Recovery?
Businesses use cloud computing
in part to enable faster disaster recovery of critical IT systems without the
cost of a second physical site. The AWS cloud supports many popular disaster
recovery architectures ranging from small customer workload data center
failures to environments that enable rapid failover at scale. With data centers
all over the world, AWS provides a set of cloud-based disaster recovery
services that enable rapid recovery of your IT infrastructure and data.
61. How can you add an existing
instance to a new Auto Scaling group?
Here’s how you can add an
existing instance to a new Auto Scaling group:
- Open EC2 console
- Select your instance under
Instances
- Choose Actions ->
Instance Settings -> Attach to Auto Scaling Group
- Select a new Auto Scaling
group
- Attach this group to the
Instance
- Edit the Instance if needed
- Once done, you can
successfully add the instance to a new Auto Scaling group
Build Your Career as a Cloud
Architect
Cloud Architect Master's ProgramExplore Program
![]()
62. What are the factors to
consider while migrating to Amazon Web Services?
Here are the factors to consider
during AWS migration:
- Operational Costs - These
include the cost of infrastructure, ability to match demand and supply, transparency,
and others.
- Workforce
Productivity
- Cost avoidance
- Operational resilience
- Business agility
63. What is RTO and RPO in AWS?
RTO or Recovery Time Objective
is the maximum time your business or organization is willing to wait for a
recovery to complete in the wake of an outage. On the other hand, RPO or
Recovery Point Objective is the maximum amount of data loss your company is
willing to accept as measured in time.
64. If you would like to
transfer vast amounts of data, which is the best option among Snowball,
Snowball Edge, and Snowmobile?
AWS Snowball is basically a data
transport solution for moving high volumes of data into and out of a specified
AWS region. On the other hand, AWS Snowball Edge adds additional computing
functions apart from providing a data transport solution. The snowmobile is an
exabyte-scale migration service that allows you to transfer data up to 100 PB.
65. Explain what T2 instances
are?
The T2 Instances are intended to
give the ability to burst to a higher performance whenever the workload demands
it and also provide a moderate baseline performance to the CPU.
The T2 instances are General
Purpose instance types and are low in cost as well. They are usually used
wherever workloads do not consistently or often use the CPU.
66. What are the advantages of
AWS IAM?
AWS IAM allows an administrator
to provide multiple users and groups with granular access. Various user groups
and users may require varying levels of access to the various resources that
have been developed. We may assign roles to users and create roles with defined
access levels using IAM.
It further gives us Federated
Access, which allows us to grant applications and users access to resources
without having to create IAM Roles.
67. Explain Connection Draining
Connection Draining is an AWS
service that allows us to serve current requests on the servers that are either
being decommissioned or updated.
By enabling this Connection
Draining, we let the Load Balancer make an outgoing
instance finish its existing requests for a set length of time before sending
it any new requests. A departing instance will immediately go off if Connection
Draining is not enabled, and all pending requests will fail.
68. What is Power User Access in
AWS?
The AWS Resources owner is
identical to an Administrator User. The Administrator User can build, change,
delete, and inspect resources, as well as grant permissions to other AWS users.
Administrator Access without the
ability to control users and permissions is provided to a Power User. A Power
User Access user cannot provide permissions to other users but has the ability
to modify, remove, view, and create resources.
Transition to Cloud Operations
in Just 3 Days!
Cloud Operations on AWSENROLL NOW
![]()
AWS Questions for CloudFormation
69. How is AWS CloudFormation
different from AWS Elastic Beanstalk?
Here are some differences
between AWS CloudFormation and AWS Elastic Beanstalk:
- AWS CloudFormation helps
you provision and describe all of the infrastructure resources that are
present in your cloud environment. On the other hand, AWS Elastic
Beanstalk provides an environment that makes it easy to deploy and run
applications in the cloud.
- AWS CloudFormation supports
the infrastructure needs of various types of applications, like legacy
applications and existing enterprise applications. On the other hand, AWS
Elastic Beanstalk is combined with the developer tools to help you manage
the lifecycle of your applications.
70. What are the elements of an
AWS CloudFormation template?
AWS CloudFormation templates are
YAML or JSON formatted text files that are comprised of five essential
elements, they are:
- Template parameters
- Output values
- Data tables
- Resources
- File format version
71. What happens when one of the
resources in a stack cannot be created successfully?
If the resource in the stack
cannot be created, then the CloudFormation automatically rolls back and
terminates all the resources that were created in the CloudFormation template.
This is a handy feature when you accidentally exceed your limit of Elastic IP
addresses or don’t have access to an EC2 AMI.

AWS Questions for Elastic Block
Storage
72. How can you automate EC2
backup using EBS?
Use the following steps in order
to automate EC2 backup using EBS:
- Get the list of instances
and connect to AWS through API to list the Amazon EBS volumes that are
attached locally to the instance.
- List the snapshots of each
volume, and assign a retention period of the snapshot. Later on, create a
snapshot of each volume.
- Make sure to remove the
snapshot if it is older than the retention period.
73. What is the difference
between EBS and Instance Store?
EBS is a kind of permanent
storage in which the data can be restored at a later point. When you save data
in the EBS, it stays even after the lifetime of the EC2 instance. On the other
hand, Instance Store is temporary storage that is physically attached to a host
machine. With an Instance Store, you cannot detach one instance and attach it
to another. Unlike in EBS, data in an Instance Store is lost if any instance is
stopped or terminated.
74. Can you take a backup of EFS
like EBS, and if yes, how?
Yes, you can use the EFS-to-EFS
backup solution to recover from unintended changes or deletion in Amazon EFS. Follow
these steps:
- Sign in to the AWS
Management Console
- Click the launch
EFS-to-EFS-restore button
- Use the region selector in
the console navigation bar to select region
- Verify if you have chosen
the right template on the Select Template page
- Assign a name to your
solution stack
- Review the parameters for
the template and modify them if necessary
75. How do you auto-delete old
snapshots?
Here’s the procedure for
auto-deleting old snapshots:
- As per procedure and best
practices, take snapshots of the EBS volumes on Amazon S3.
- Use AWS Ops Automator to
handle all the snapshots automatically.
- This allows you to create,
copy, and delete Amazon EBS snapshots.

Iginite Your Knowledge in AWS
Cloud Operations
Cloud Operations on AWSENROLL NOW
![]()
AWS Interview Questions for
Elastic Load Balancing
76. What are the different types
of load balancers in AWS?
Three types of load balancers
are supported by Elastic Load Balancing:
- Application Load Balancer
- Network Load Balancer
- Classic Load Balancer
77. What are the different uses
of the various load balancers in AWS Elastic Load Balancing?
Application Load Balancer
Used if you need flexible
application management and TLS termination.
Network Load Balancer
Used if you require extreme
performance and static IPs for your applications.
Classic Load Balancer
Used if your application is
built within the EC2 Classic network
AWS Interview Questions for
Security
78. What Is Identity and Access
Management (IAM) and How Is It Used?
Identity and
Access Management (IAM) is a web service for securely controlling
access to AWS services. IAM lets you manage users, security credentials such as
access keys, and permissions that control which AWS resources users and
applications can access.
79. How can you use AWS WAF in
monitoring your AWS applications?
AWS WAF or AWS Web Application
Firewall protects your web applications from web exploitations. It helps you
control the traffic flow to your applications. With WAF, you can also create
custom rules that block common attack patterns. It can be used for three cases:
allow all requests, prevent all requests, and count all requests for a new
policy.
80. What are the different AWS
IAM categories that you can control?
Using AWS IAM, you can do the
following:
- Create and manage IAM users
- Create and manage IAM
groups
- Manage the security
credentials of the users
- Create and manage policies
to grant access to AWS services and resources
81. What are the policies that
you can set for your users’ passwords?
Here are some of the policies
that you can set:
- You can set a minimum
length of the password, or you can ask the users to add at least one
number or special characters in it.
- You can assign requirements
of particular character types, including uppercase letters, lowercase
letters, numbers, and non-alphanumeric characters.
- You can enforce automatic
password expiration, prevent reuse of old passwords, and request for a
password reset upon their next AWS sign in.
- You can have the AWS users
contact an account administrator when the user has allowed the password to
expire.
82. What is the difference
between an IAM role and an IAM user?
The two key differences between
the IAM role and IAM user are:
- An IAM role is an IAM
entity that defines a set of permissions for making AWS service requests,
while an IAM user has permanent long-term credentials and is used to
interact with the AWS services directly.
- In the IAM role, trusted
entities, like IAM users, applications, or an AWS service, assume roles
whereas the IAM user has full access to all the AWS IAM functionalities.
83. What are the managed
policies in AWS IAM?
There are two types of managed
policies; one that is managed by you and one that is managed by AWS. They are
IAM resources that express permissions using IAM policy language. You can
create, edit, and manage them separately from the IAM users, groups, and roles
to which they are attached.
84. Can you give an example of
an IAM policy and a policy summary?
Here’s an example of an IAM
policy to grant access to add, update, and delete objects from a specific
folder.

Here’s an example of a policy
summary:

85. How does AWS IAM help your
business?
IAM enables to:
- Manage IAM users and their
access - AWS IAM provides secure resource access to multiple users
- Manage access for federated
users – AWS allows you to provide secure access to resources in your AWS
account to your employees and applications without creating IAM roles
Get Certifications in AWS, Azure
and Google Cloud
Cloud Architect Master's ProgramExplore Program
![]()
AWS Interview Questions for
Route 53
86. What Is Amazon Route 53?
Amazon Route 53 is a scalable
and highly available Domain Name System (DNS). The name refers to TCP or UDP
port 53, where DNS server requests are addressed.
87. What Is Cloudtrail and How
Do Cloudtrail and Route 53 Work Together?
CloudTrail is a service that
captures information about every request sent to the Amazon Route 53 API by an
AWS account, including requests that are sent by IAM users. CloudTrail saves
log files of these requests to an Amazon S3 bucket. CloudTrail captures
information about all requests. You can use information in the CloudTrail log
files to determine which requests were sent to Amazon Route 53, the IP address
that the request was sent from, who sent the request, when it was sent, and
more.
88. What is the difference
between Latency Based Routing and Geo DNS?
The Geo Based DNS routing takes
decisions based on the geographic location of the request. Whereas, the Latency
Based Routing utilizes latency measurements between networks and AWS data
centers. Latency Based Routing is used when you want to give your customers the
lowest latency possible. On the other hand, Geo Based routing is used when you
want to direct the customer to different websites based on the country or
region they are browsing from.
89. What is the difference
between a Domain and a Hosted Zone?
Domain
A domain is a collection of data
describing a self-contained administrative and technical unit. For
example, www.simplilearn.com is a domain and a
general DNS concept.
Hosted zone
A hosted zone is a container
that holds information about how you want to route traffic on the internet for
a specific domain. For example, lms.simplilearn.com is a hosted zone.
90. How does Amazon Route 53
provide high availability and low latency?
Here’s how Amazon Route 53
provides the resources in question:
Globally Distributed Servers
Amazon is a global service and
consequently has DNS services globally. Any customer creating a query from any
part of the world gets to reach a DNS server local to them that provides low
latency.
Dependency
Route 53 provides a high level
of dependability required by critical applications
Optimal Locations
Route 53 uses a global anycast
network to answer queries from the optimal position automatically.
AWS Interview Questions for
Config
91. How does AWS config work
with AWS CloudTrail?
AWS CloudTrail records user API
activity on your account and allows you to access information about the
activity. Using CloudTrail, you can get full details about API actions such as
the identity of the caller, time of the call, request parameters, and response
elements. On the other hand, AWS Config records point-in-time configuration
details for your AWS resources as Configuration Items (CIs).
You can use a CI to ascertain
what your AWS resource looks like at any given point in time. Whereas, by using
CloudTrail, you can quickly answer who made an API call to modify the resource.
You can also use Cloud Trail to detect if a security group was incorrectly
configured.
92. Can AWS Config aggregate
data across different AWS accounts?
Yes, you can set up AWS Config
to deliver configuration updates from different accounts to one S3 bucket, once
the appropriate IAM policies are applied to the S3 bucket.
AWS Interview Questions for
Database
93. How are reserved instances
different from on-demand DB instances?
Reserved instances and on-demand
instances are the same when it comes to function. They only differ in how they
are billed.
Reserved instances are purchased
as one-year or three-year reservations, and in return, you get very low hourly
based pricing when compared to the on-demand cases that are billed on an hourly
basis.
94. Which type of scaling would
you recommend for RDS and why?
There are two types of scaling -
vertical scaling and horizontal scaling. Vertical scaling lets you vertically
scale up your master database with the press of a button. A database can only
be scaled vertically, and there are 18 different instances in which you can
resize the RDS. On the other hand, horizontal scaling is good for replicas.
These are read-only replicas that can only be done through Amazon Aurora.
95. What is a maintenance window
in Amazon RDS? Will your DB instance be available during maintenance events?
RDS maintenance window lets you
decide when DB instance modifications, database engine version upgrades, and
software patching have to occur. The automatic scheduling is done only for
patches that are related to security and durability. By default, there is a
30-minute value assigned as the maintenance window and the DB instance will
still be available during these events though you might observe a minimal
effect on performance.
96. What are the consistency
models in DynamoDB?
There are two consistency models
In DynamoDB. First, there is the Eventual Consistency Model, which maximizes
your read throughput. However, it might not reflect the results of a recently
completed write. Fortunately, all the copies of data usually reach consistency
within a second. The second model is called the Strong Consistency Model. This
model has a delay in writing the data, but it guarantees that you will always
see the updated data every time you read it.
97. What type of query
functionality does DynamoDB support?
DynamoDB supports GET/PUT
operations by using a user-defined primary key. It provides flexible querying
by letting you query on non-primary vital attributes using global secondary
indexes and local secondary indexes.
Iginite Your Knowledge in AWS
Cloud Operations
Cloud Operations on AWSENROLL NOW
![]()
AWS Interview Questions - Short
Answer Questions
1. Suppose you are a game
designer and want to develop a game with single-digit millisecond latency,
which of the following database services would you use?
Amazon DynamoDB
2. If you need to perform
real-time monitoring of AWS services and get actionable insights, which
services would you use?
Amazon CloudWatch
3. As a web developer, you are
developing an app, targeted primarily for the mobile platform. Which of the
following lets you add user sign-up, sign-in, and access control to your web
and mobile apps quickly and easily?
Amazon Cognito
4. You are a Machine Learning
Engineer who is on the lookout for a solution that will discover sensitive
information that your enterprise stores in AWS and then use NLP to classify the
data and provide business-related insights. Which among the services would you
choose?
AWS Macie
5. You are the system
administrator in your company, which is running most of its infrastructure on
AWS. You are required to track your users and keep tabs on how they are being
authenticated. You wish to create and manage AWS users and use permissions to
allow and deny their access to AWS resources. Which of the following services
suits you best?
AWS IAM
6. Which service do you use if
you want to allocate various private and public IP addresses to make them
communicate with the internet and other instances?
Amazon VPC
7. This service provides you
with cost-efficient and resizable capacity while automating time-consuming
administration tasks
Amazon Relational Database
Service
8. Which of the following is a
means for accessing human researchers or consultants to help solve problems on
a contractual or temporary basis?
Amazon Mechanical Turk
9. This service is used to make
it easy to deploy, manage, and scale containerized applications using
Kubernetes on AWS. Which of the following is this AWS service?
Amazon Elastic Container Service
10. This service lets you run
code without provisioning or managing servers. Select the correct service from
the below options
11. As an AWS Developer, using
this pay-per-use service, you can send, store, and receive messages between
software components. Which of the following is it?
Amazon Simple Queue Service
12. Which service do you use if
you would like to host a real-time audio and video conferencing application on
AWS, this service provides you with a secure and easy-to-use application?
Amazon Chime
13. As your company's AWS
Solutions Architect, you are in charge of designing thousands of similar
individual jobs. Which of the following services best meets your requirements?
AWS Batch
Want a Job at AWS? Find Out What
It Takes
Cloud Architect Master's ProgramExplore Program
![]()
AWS Interview Questions -
Multiple-Choice
1. Suppose you are a game
designer and want to develop a game with single-digit millisecond latency,
which of the following database services would you use?
- Amazon RDS
- Amazon Neptune
- Amazon Snowball
- Amazon DynamoDB
2. If you need to perform
real-time monitoring of AWS services and get actionable insights, which
services would you use?
- Amazon Firewall Manager
- Amazon GuardDuty
- Amazon CloudWatch
- Amazon EBS
3. As a web developer, you are
developing an app, targeted especially for the mobile platform. Which of the
following lets you add user sign-up, sign-in, and access control to your web
and mobile apps quickly and easily?
- AWS Shield
- AWS Macie
- AWS Inspector
- Amazon Cognito
4. You are a Machine Learning
Engineer who is on the lookout for a solution that will discover sensitive
information that your enterprise stores in AWS and then use NLP to classify the
data and provide business-related insights. Which among the services would you
choose?
- AWS Firewall Manager
- AWS IAM
- AWS Macie
- AWS CloudHSM
5. You are the system
administrator in your company, which is running most of its infrastructure on
AWS. You are required to track your users and keep tabs on how they are being
authenticated. You wish to create and manage AWS users and use permissions to
allow and deny their access to AWS resources. Which of the following services
suits you best?
- AWS Firewall Manager
- AWS Shield
- Amazon API Gateway
- AWS IAM
6. Which service do you use if
you want to allocate various private and public IP addresses in order to make
them communicate with the internet and other instances?
- Amazon Route 53
- Amazon VPC
- Amazon API Gateway
- Amazon CloudFront
7. This service provides you
with cost-efficient and resizable capacity while automating time-consuming
administration tasks
- Amazon Relational Database
Service
- Amazon Elasticache
- Amazon VPC
- Amazon Glacier
8. Which of the following is a
means for accessing human researchers or consultants to help solve problems on
a contractual or temporary basis?
- Amazon Mechanical Turk
- Amazon Elastic Mapreduce
- Amazon DevPay
- Multi-Factor Authentication
9. This service is used to make
it easy to deploy, manage, and scale containerized applications using
Kubernetes on AWS. Which of the following is this AWS service?
- Amazon Elastic Container
Service
- AWS Batch
- AWS Elastic Beanstalk
- Amazon Lightsail
10. This service lets you run
code without provisioning or managing servers. Select the correct service from
the below options
- Amazon EC2 Auto Scaling
- AWS Lambda
- AWS Batch
- Amazon Inspector
11. As an AWS Developer, using
this pay-per-use service, you can send, store and receive messages between
software components. Which of the following is it?
- AWS Step Functions
- Amazon MQ
- Amazon Simple Queue Service
- Amazon Simple Notification
Service
12. Which service do you use if
you would like to host real-time audio and video conferencing application on
AWS, this service provides you with a secure and easy-to-use application?
- Amazon Chime
- Amazon WorkSpaces
- Amazon MQ
- Amazon AppStream
13. As your company's AWS
Solutions Architect, you are in charge of designing thousands of similar
individual jobs. Which of the following services best meets your requirements?
- AWS EC2 Auto Scaling
- AWS Snowball
- AWS Fargate
- AWS Batch
14. You are a Machine Learning
engineer and you are looking for a service that helps you build and train Machine
Learning models in AWS. Which among the following are we referring to?
- Amazon SageMaker
- AWS DeepLens
- Amazon Comprehend
- Device Farm
15. Imagine that you are working
for your company's IT team. You are assigned to adjusting the capacity of AWS
resources based on the incoming application and network traffic. How would you
do it?
- Amazon VPC
- AWS IAM
- Amazon Inspector
- Amazon Elastic Load
Balancing
Want a Job at AWS? Find Out What
It Takes
Cloud Architect Master's ProgramExplore Program
![]()
16. This cross-platform video
game development engine that supports PC, Xbox, Playstation, iOS, and Android
platforms allows developers to build and host their games on Amazon's servers.
- Amazon GameLift
- AWS Greengrass
- Amazon Lumberyard
- Amazon Sumerian
17. You are the Project Manager
of your company's Cloud Architects team. You are required to visualize,
understand and manage your AWS costs and usage over time. Which of the
following services works best?
- AWS Budgets
- AWS Cost Explorer
- Amazon WorkMail
- Amazon Connect
18. You are the chief Cloud
Architect at your company. How can you automatically monitor and adjust
computer resources to ensure maximum performance and efficiency of all scalable
resources?
- AWS CloudFormation
- AWS Aurora
- AWS Auto Scaling
- Amazon API Gateway
19. As a database administrator.
you will employ a service that is used to set up and manage databases such as
MySQL, MariaDB, and PostgreSQL. Which service are we referring to?
- Amazon Aurora
- AWS RDS
- Amazon Elasticache
- AWS Database Migration
Service
20. A part of your marketing
work requires you to push messages onto Google, Facebook, Windows, and Apple
through APIs or AWS Management Console. Which of the following services do you
use?
- AWS CloudTrail
- AWS Config
- Amazon Chime
- AWS Simple Notification
Service
Basic AWS Interview Questions for Freshers
1. What is AWS?
AWS (Amazon Web Services) is a platform to provide secure
cloud services, database storage, offerings for computing power, content delivery,
and other services to help the business scale and develop.
2. Give the comparison between AWS and OpenStack.
Criteria AWS OpenStack
License Amazon
proprietary Open-source
Operating system Provided
as per the cloud administrator AMIs
provided by AWS
Performing repeatable operations Through templates Through
text files
Want to learn the basics of AWS Cloud Solutions? Check out
our AWS Certification Training Course!
3. What is the importance of buffer in Amazon Web
Services?
An Elastic Load Balancer ensures that the incoming traffic
is distributed optimally across various AWS instances. A buffer will
synchronize different components and make the arrangement also elastic to a
load or traffic burst. The components are prone to working in an unstable way
of receiving and processing requests. The buffer creates an equilibrium linking
various apparatus and crafts them to work at an identical rate to supply more
rapid services.
4. How are Spot Instances, On-demand Instances, and
Reserved Instances different from one another?
Both Spot Instances and On-demand Instances are pricing
models.
Spot Instance On-demand
Instance
With Spot Instance, customers can purchase compute
capacity with no upfront commitment at all. With
On-demand Instances, users can launch instances at any time based on the
demand.
Spot Instances are spare Amazon instances that you can bid
for. On-demand Instances are suitable for
the high-availability needs of applications.
When the bidding price exceeds the spot price, the
instance is automatically launched, and the spot price fluctuates based on
supply and demand for instances. On-demand
Instances are launched by users only with the pay-as-you-go model.
When the bidding price is less than the spot price, the
instance is immediately taken away by Amazon. On-demand
Instances will remain persistent without any automatic termination from Amazon.
Spot Instances are charged on an hourly basis. On-demand Instances are charged on a
per-second basis.
Ready to tackle the most common interview question?
Take a quick Quiz to check it out
Take a Quiz
5. Why do we make subnets?
Creating subnets means dividing a large network into
smaller ones. These subnets can be created for several reasons. For example,
creating and using subnets can help reduce congestion by making sure that the
traffic destined for a subnet stays in that subnet. This helps in efficiently
routing the to the network, which reduces the network’s load.
Learn more about Amazon Web Services from our AWS
Tutorial!
Watch this video on AWS Interview Questions for Beginners:
6. Is there a way to upload a file that is greater than
100 megabytes on Amazon S3?
Yes, it is possible by using the multipart upload utility
from AWS. With the multipart upload utility, larger files can be uploaded in
multiple parts that are uploaded independently. You can also decrease upload
time by uploading these parts in parallel. After the upload is done, the parts
will be merged into a single object or file to create the original file from
which the parts were created.
To learn more about the Amazon S3 bucket, read the blog.
7. What is the maximum number of S3 buckets you can
create?
The maximum number of S3 buckets that can be created is
100.
Get 100% Hike!
Master Most in Demand Skills Now !
Email Address
+91 IN INDIA
Phone Number
By providing your contact details, you agree to our Terms
of Use & Privacy Policy
8. How can you save the data on root volume on an
EBS-backed machine?
We can save the data by overriding the terminate option.
9. When should you use the classic load balancer and the
application load balancer?
The classic load balancer is used for simple load
balancing of traffic across multiple EC2 instances.
Classic Load Balancer
While the application load balancing is used for more
intelligent load balancing, based on the multi-tier architecture or
container-based architecture of the application. Application load balancing is
mostly used when there is a need to route traffic to multiple services.
Classic Load Balancer
Want to learn about AWS DevOps? Check out our blog on What
is AWS DevOps.
10. How many total VPCs per account/region and subnets per
VPC can you have?
We can have a total of 5 VPCs for every account/region and
200 subnets for every VPC that you have.
11. Your organization has decided to have all its
workloads on the public cloud. But, due to certain security concerns, your
organization has decided to distribute some of the workload on private servers.
You are asked to suggest a cloud architecture for your organization. What will
your suggestion be?
A hybrid cloud. The hybrid cloud architecture is where an
organization can use the public cloud for shared resources and the private
cloud for confidential workloads.
12. Which one of the storage solutions offered by AWS
would you use if you need extremely low pricing and data archiving?
AWS Glacier is an extremely low-cost storage service
offered by Amazon that is used for data archiving and backup purposes. The
longer you store data in Glacier, the less it will cost you.
Are you interested in learning AWS from experts? Enroll in
our AWS Course in Bangalore and be a master of it!
13. You have connected four instances to ELB. To
automatically terminate your unhealthy instances and replace them with new
ones, which functionality would you use?
Auto-scaling groups
95% learner satisfaction score post completion of the
program*
500% salary hike received by a working professional post
completion of the course*
Fresher earned 30 LPA salary package on completion of the
course*
53% of learners received 50% and above salary hike post
completion of the program*
85% of the learners achieved their training objectives
within 9 months of course completion*
95% learner satisfaction score post completion of the
program*
500% salary hike received by a working professional post
completion of the course*
Fresher earned 30 LPA salary package on completion of the
course*
53% of learners received 50% and above salary hike post
completion of the program*
85% of the learners achieved their training objectives
within 9 months of course completion*
95% learner satisfaction score post completion of the
program*
Process Advisors
ey-logo
*Subject to Terms and Condition
14. The data on the root volumes of store-backed and
EBS-backed instances get deleted by default when they are terminated. If you
want to prevent that from happening, which instance would you use?
EBS-backed instances. EBS-backed instances use EBS volume
as their root volume. EBS volume consists of virtual drives that can be easily
backed up and duplicated by snapshots.
EBS Backed Instances
The biggest advantage of EBS-backed volumes is that the
data can be configured to be stored for later retrieval even if the virtual
machine or the instances are shut down.
15. How will you configure an Amazon S3 bucket to serve
static assets for your public web application?
By configuring the bucket policy to provide public read
access to all objects
That is all we have in our section on basic Amazon Web
Services interview questions section.
Intermediate AWS Interview Questions and Answers
16. Your organization wants to send and receive compliance
emails to its clients using its email address and domain. What service would
you suggest for achieving the same easily and cost-effectively?
Amazon Simple Email Service (Amazon SES), which is a
cloud-based email-sending service, can be used for this purpose.
17. Can you launch Amazon Elastic Compute Cloud (EC2)
instances with predetermined private IP addresses? If yes, then with which
Amazon service is it possible?
Yes. It is possible by using VPC (Virtual Private Cloud).
Looking for the Perfect Job Interview Attire? Worry Not.
Read our perfect guide on Interview Outfits to land your dream job.
18. If you launch a standby RDS, will it be launched in
the same availability zone as your primary?
No, standby instances are automatically launched in
different availability zones than the primary, making them physically
independent infrastructures. This is because the whole purpose of standby
instances is to prevent infrastructure failure. So, in case the primary goes
down, the standby instance will help recover all of the data.
19. What is the name of Amazon's Content Delivery Network?
Amazon CloudFront
20. Which Amazon solution will you use if you want to
accelerate moving petabytes of data in and out of AWS, using storage devices
that are designed to be secure for data transfer?
Amazon Snowball. AWS Snowball is the data transport
solution for large amounts of data that need to be moved into and out of AWS
using physical storage devices.
21. If you are running your DB instance as a Multi-AZ
deployment, can you use standby DB instances along with your primary DB
instance?
No, the standby DB instance cannot be used along with the
primary DB instances since the standby DB instances are supposed to be used
only if the primary instance goes down.
Interested in learning AWS? Enroll in our AWS Training in
Pune!
22. Your organization is developing a new multi-tier web
application in AWS. Being a fairly new and small organization, there’s limited
staff. However, the organization requires high availability. This new
application comprises complex queries and table joins. Which Amazon service
will be the best solution for your organization’s requirements?
DynamoDB will be the right choice here since it is
designed to be highly scalable, more than RDS or any other relational database
service.
23. You accidentally stopped an EC2 instance in a VPC with
an associated Elastic IP. If you start the instance again, what will be the
result?
Elastic IP will only be disassociated from the instance if
it’s terminated. If it’s stopped and started, there won’t be any change to the
instance, and no data will be lost.
24. Your organization has around 50 IAM users. Now, it
wants to introduce a new policy that will affect the access permissions of an
IAM user. How can it implement this without having to apply the policy at the
individual user level?
It is possible to use AWS IAM groups, by adding users in
the groups as per their roles and by simply applying the policy to the groups.
Get certified from the top AWS Training in Noida now!
Advanced AWS Interview Questions for Experienced
25. Your organization is using DynamoDB for its
application. This application collects data from its users every 10 minutes and
stores it in DynamoDB. Then every day, after a particular time interval, the
data (respective of each user) is extracted from DynamoDB and sent to S3. Then,
the application visualizes this data for the users. You are asked to propose a
solution to help optimize the backend of the application for latency at a lower
cost. What would you recommend?
ElastiCache. Amazon ElastiCache is a caching solution
offered by Amazon.
Elastic Cache
It can be used to store a cached version of the
application in a region closer to users so that when requests are made by the
users the cached version of the application can respond, and hence latency will
be reduced.
Become a master of AWS by going through this online AWS
Course in Coimbatore!
26. I created a web application with autoscaling. I
observed that the traffic on my application is the highest on Wednesdays and
Fridays between 9 AM and 7 PM. What would be the best solution for me to handle
the scaling?
Configure a policy in autoscaling to scale as per the
predictable traffic patterns.
27. How would you handle a situation where the relational
database engine crashes often whenever the traffic to your RDS instances
increases, given that the replica of the RDS instance is not promoted as the
master instance?
A bigger RDS instance type needs to be opted for handling
large amounts of traffic and creating manual or automated snapshots to recover
data in case the RDS instance goes down.
Learn more about AWS from this AWS Training in Chennai to
get ahead in your career!
28. You have an application running on your Amazon EC2
instance. You want to reduce the load on your instance as soon as the CPU
utilization reaches 100 percent. How will you do that?
It can be done by creating an autoscaling group to deploy
more instances when the CPU utilization exceeds 100 percent and distributing
traffic among instances by creating a load balancer and registering the Amazon
EC2 instances with it.
Watch this video on Free AWS Full Course:
29. What would I have to do if I wanted to access Amazon
Simple Storage buckets and use the information for access audits?
AWS CloudTrail can be used in this case as it is designed
for logging and tracking API calls, and it has also been made available for
storage solutions.
Learn the complete concepts of AWS through our AWS
Training in Hyderabad in 26 hours!
30. I created a key in the North Virginia region to
encrypt my data in the Oregon region. I also added three users to the key and
an external AWS account. Then, to encrypt an object in S3, when I tried to use
the same key, it was not listed. Where did I go wrong?
The data and the key should be in the same region. That
is, the data that has to be encrypted should be in the same region as the one
in which the key was created. In this case, the data is in the Oregon region,
whereas the key was created in the North Virginia region.
31. Suppose, you hosted an application on AWS that lets
the users render images and do some general computing. Which of the
below-listed services can you use to route the incoming user traffic?
Classic Load Balancer
Application Load Balancer
Network Load balancer
Application Load Balancer: It supports path-based routing
of the traffic and hence helps in enhancing the performance of the application
structured as smaller services.
Application Load Balancer
Using an application load balancer, the traffic can be
routed based on the requests made. In this case scenario, the traffic where
requests are made for rendering images can be directed to the servers only
deployed for rendering images, and the traffic where requests are made for
computing can be directed to the servers deployed only for general computing
purposes.
32. Suppose, I created a subnet and launched an EC2
instance in the subnet with default settings. Which of the following options
will be ready to use on the EC2 instance as soon as it is launched?
Elastic IP
Private IP
Public IP
Internet Gateway
Private IP. Private IP is automatically assigned to the instance
as soon as it is launched. While elastic IP has to be set manually, Public IP
needs an Internet Gateway which again has to be created since it’s a new VPC.
Cloud Computing EPGC IITR iHUB
33. Your organization has four instances for production
and another four for testing. You are asked to set up a group of IAM users that
can only access the four production instances and not the other four testing
instances. How will you achieve this?
We can achieve this by defining tags on the test and
production instances and then adding a condition to the IAM policy that allows
access to specific tags.
Go through the AWS Training in India to get a clear
understanding of AWS!
34. Your organization wants to monitor the read-and-write
IOPS for its AWS MySQL RDS instance and then send real-time alerts to its
internal operations team. Which service offered by Amazon can help your
organization achieve this scenario?
Amazon CloudWatch would help us achieve this. Since Amazon
CloudWatch is a monitoring tool offered by Amazon, it’s the right service to
use in the above-mentioned scenario.
35. Which of the following services can be used if you
want to capture client connection information from your load balancer at a
particular time interval?
Enabling access logs on your load balancer
Enabling CloudTrail for your load balancer
Enabling CloudWatch metrics for your load balancer
Enabling CloudTrail for your load balancer. AWS CloudTrail
is an inexpensive log monitoring solution provided by Amazon. It can provide
logging information for load balancers or any other AWS resources. The provided
information can be further used for analysis.
Learn more about AWS CloudWatch in the blog by
Intellipaat.
36. You have created a VPC with private and public
subnets. In what kind of subnet would you launch the database servers?
Database servers should be ideally launched on private
subnets. Private subnets are ideal for the backend services and databases of
all applications since they are not meant to be accessed by the users of the
applications, and private subnets are not routable from the Internet.
37. Is it possible to switch from an Instance-backed root
volume to an EBS-backed root volume at any time?
No, it is not possible.
38. Can you change the instance type of the instances that
are running in your application tier and are also using autoscaling? If yes,
then how? (Choose one of the following)
Yes, by modifying autoscaling launch configuration
Yes, by modifying autoscaling tags configuration
Yes, by modifying autoscaling policy configuration
No, it cannot be changed
Yes, by modifying the autoscaling launch configuration
Yes, by modifying the autoscaling tag configuration
Yes, by modifying the autoscaling policy configuration
No, it cannot be changed
Yes, the instance type of such instances can be changed by
modifying the autoscaling launch configuration. The tags configuration is used
to add metadata to the instances.
Do you know about the different types of AWS
Certifications? Read the Blog to find out.
39. Can you name the additional network interface that can
be created and attached to your Amazon EC2 instance launched in your VPC?
AWS Elastic Network Interface
40. Out of the following options, where does the user
specify the maximum number of instances with the autoscaling commands?
Autoscaling policy configuration
Autoscaling group
Autoscaling tags configuration
Autoscaling group configuration
Autoscaling Group configuration
41. Which service provided by AWS can you use to transfer
objects from your data center, when you are using Amazon CloudFront?
Amazon Direct Connect. It is an AWS networking service
that acts as an alternative to using the Internet to connect customers in
on-premise sites with AWS.
42. You have deployed multiple EC2 instances across multiple
availability zones to run your website. You have also deployed a Multi-AZ RDS
MySQL Extra Large DB Instance. The site performs a high number of small read
and write operations per second. After some time, you observed that there is
read contention on RDS MySQL. What would be your approach to resolve the
contention and optimize your website?
We can deploy ElastiCache in-memory cache running in every
availability zone. This will help in creating a cached version of the website
for faster access in each availability zone. We can also add an RDS MySQL read
replica in each availability zone that can help with efficient and better
performance for read operations. So, there will not be any increased workload
on the RDS MySQL instance, hence resolving the contention issue.
43. Your company wants you to propose a solution so that
the company’s data center can be connected to the Amazon cloud network. What
would your proposal be?
The data center can be connected to the Amazon cloud
network by establishing a virtual private network (VPN) between the VPC and the
data center. A virtual private network lets you establish a secure pathway or
tunnel from your premise or device to the AWS global network.
44. Which of the following Amazon Services would you
choose if you want complex querying capabilities but not a whole data
warehouse?
RDS
Redshift
ElastiCache
DynamoDB
Amazon RDS
45. You want to modify the security group rules while it
is being used by multiple EC2 instances. Will you be able to do that? If yes,
will the new rules be implemented on all previously running EC2 instances that
were using that security group?
Yes, the security group that is being used by multiple EC2
instances can be modified. The changes will be implemented immediately and
applied to all the previously running EC2 instances without restarting the
instances
Become a Cloud and DevOps Architect
46. Which one of the following is a structured data store
that supports indexing and data queries to both EC2 and S3?
DynamoDB
MySQL
Aurora
SimpleDB
SimpleDB
47. Which service offered by Amazon will you choose if you
want to collect and process e-commerce data for near real-time analysis?
(Choose any two)
DynamoDB
Redshift
Aurora
SimpleDB
DynamoDB
Redshift
Aurora
SimpleDB
DynamoDB. DynamoDB is a fully managed NoSQL database
service that can be fed any type of unstructured data. Hence, DynamoDB is the
best choice for collecting data from e-commerce websites. For near-real-time
analysis, we can use Amazon Redshift.
48. If in CloudFront the content is not present at an edge
location, what will happen when a request is made for that content?
CloudFront will deliver the content directly from the
origin server. It will also store the content in the cache of the edge location
where the content was missing.
49. Can you change the private IP address of an EC2
instance while it is running or in a stopped state?
No, it cannot be changed. When an EC2 instance is
launched, a private IP address is assigned to that instance at boot time. This
private IP address is attached to the instance for its entire lifetime and can
never be changed.
50. Which of the following options will you use if you
have to move data over long distances using the Internet, from instances that
are spread across countries to your Amazon S3 bucket?
Amazon CloudFront
Amazon Transfer Acceleration
Amazon Snowball
Amazon Glacier
Amazon Transfer Acceleration. It throttles the data
transfer up to 300 percent using optimized network paths and Amazon Content
Delivery Network. Snowball cannot be used here as this service does not support
cross-region data transfer.
51. Which of the following services is a data storage
system that also has a REST API interface and uses secure HMAC-SHA1
authentication keys?
Amazon Elastic Block Store
Amazon Snapshot
Amazon S3
Amazon S3. It gets various requests from applications, and
it has to identify which requests are to be allowed and which are to be denied.
Amazon S3 REST API uses a custom HTTP scheme based on a keyed HMAC for the
authentication of requests.
52. What is EC2?
Launched in 2006, EC2 is a virtual machine that you can
use to deploy your servers in the cloud, giving you OS-level control. It helps
you have control over the hardware and updates, similar to the case of
on-premise servers. EC2 can run on either of these operating systems- Microsoft
and Linux. It can also support applications like Python, PHP, Apache, and more.
Learn more about AWS Secrets Manager in the blog by
Intellipaat.
53. What is Snowball?
Snowball is an application designed for transferring
terabytes of data into and outside of the AWS cloud. It uses secured physical
storage to transfer the data. Snowball is considered a petabyte-scale data
transport solution that helps with cost and time savings.
54. What is CloudWatch?
The Amazon CloudWatch is used for monitoring and managing
data and getting actionable insights for AWS, on-premise applications, etc. It
helps you monitor your entire task stack which includes the applications,
infrastructure, and services. Apart from this, CloudWatch also assists you in
optimizing your resource utilization and cost by providing analytics-driven
insights.
55. What is Elastic Transcoder?
In the AWS cloud, the Elastic Transcoder is used for
converting media files into versions that can be run/played on devices such as
Tablets, PCs, Smartphones, etc. It consists of advanced transcoding features
with conversion rates starting from $ 0.0075 per minute.
Learn more about Amazon systems manager in the blog by
intellipaat.
56. What does an AMI include?
AMI stands for Amazon Machine Images. It includes the
following:
Single or multiple Amazon Elastic Block Store (Amazon EBS)
snapshots. Templates for the root volume of the instance.
Launch permissions that let AWS accounts use AMI to launch
instances.
A block device mapping specifies what volumes are to be
attached to the instance during its launch.
57. What are the Storage Classes available in Amazon S3?
The following storage classes are available in Amazon S3:
S3 Standard- It is by and large the default storage class.
In cases where no specification about the storage class is provided while
uploading the object, Amazon S3 assigns the S3 Standard storage class by
default.
Reduced Redundancy- It is assigned when non-critical,
reproducible data needs to be stored. The Reduced Redundancy Storage class is
designed in a way that the above data categories can be stored with less
redundancy.
However, it is always advisable to go ahead with the S3
Standard storage class.
58. What are the native AWS security logging capabilities?
The native AWS security logging capabilities include AWS
CloudTrail, AWS Config, AWS detailed billing reports, Amazon S3 access logs,
Elastic load balancing Access logs, Amazon CloudFront access logs, Amazon VPC
Flow logs, etc. To learn about native AWS security logging capabilities in
detail, click here.
Take up the AWS Masters Certification Course by
Intellipaat and upgrade your skill set.
59. What kind of IP address can you use for your customer
gateway (CGW) address?
We can use the Internet routable IP address, which is a
public IP address of your NAT device.
AWS Scenario-Based Interview Questions
60. A Company has a running Web Application Server in the
N. Virginia region and the server has a large size EBS volume of approximately
500 GB, and to see the demand of business, the company needs to migrate the
server from the current region to another AWS account’s Mumbai location. Which
is the best way to migrate the server from the current location to the Mumbai
region? And what information AWS administrator does require about AWS A/C?
Create an AMI of the server running in the North Virginia
region. Once the AMI is created, The administrator will need the 12-digit
account number of the #2 AWS account. This is required for copying the AMI
which we have created.
Once the AMI is successfully copied into the Mumbai
region, you can launch the instance using the copied AMI in the Mumbai region.
Once the instance is running and if it’s completely operational, the server in
the North Virginia region could be terminated. This is the best way to migrate
a server to a different account without any hassle.
61. A start-up company has a web application based in the
US-east-1 region with multiple Amazon EC2 instances running behind an
Application Load Balancer across multiple Availability Zones. As the company's
user base grows in the US-West-1 region, the company needs a solution with low
latency and improved availability. What should a solutions architect do to
achieve it.?
You need to notice here that, currently, the web
application is in the US-ease-1, and the user base grows in the US-east-1
region. The very first step, provision multiple EC2 instances (web application
servers) and configure an Application Load Balancer in us-west-1. Now, create
Global Accelerator in AWS Global Accelerator which uses an endpoint group that
includes the load balancer endpoints in both regions.
Read this blog about AWS services to know more.
62. A company currently operates a web application backed
by an Amazon RDS MySQL database. It has automated backups that are run daily
and are not encrypted. A security audit requires future backups to be encrypted
and unencrypted backups to be destroyed. The company will make at least one
encrypted backup before destroying the old backups. What should be done to
enable encryption for future backups?
Create a snapshot of the database.
Copy it to an encrypted snapshot.
Restore the database from the encrypted snapshot.
63. A company is going to launch one branch in the UK and
needs to continue with its existing main branch in the USA. The company has
almost 15 GB of data which is stored in an S3 Bucket in the Ohio region and
data is stored with the default storage class. The Company also wants to
provide its updated & stored data in the London S3 bucket using one zone
accessibility storage class to save storage costs. In addition, the company
also wants the data must be updated automatically in S3’s London bucket; if any
data is modified or written in the S3 bucket in Ohio.
Configure the Cross Region Replication Rule in the Ohio
region bucket and select the destination bucket in the London region to
replicate the data and store it in the destination using one zone IA storage
class to save cost.
64. You are an AWS Architect in your company, and you are
asked to create a new VPC in the N.Virginia Region with two Public and two
Private subnets using the following CIDR blocks:
VPC CIDR = 10.10.10.0/24
Public Subnet
Subnet01 : 10.10.10.0/26
Subnet02 : 10.10.10.64/26
Private Subnet
Subnet03: 10.10.10.128/26
Subnet04: 10.10.10.192/26
Using the above CIDRs you created a new VPC, and you
launched EC2 instances in all subnets as per the need.
Now, you are facing an issue in private instances that you
are unable to update operating systems from the internet. So, what
architectural changes and configurations will you suggest to resolve the issue?
NAT G/W is to be installed in one public subnet and will
configure the route table associated with private subnets to add NAT G/W entry
to provide internet access to private instances.
65. The data on the root volumes of store-backed and
EBS-backed instances get deleted by default when they are terminated. If you
want to prevent that from happening, which instance would you use? Ensure that
if the EC2 instance is restarted, the data or configuration in the EC2 instance
should not be lost.
EBS-backed instances or instances with EBS Volume.
EBS-backed instances use EBS volume as their root volume. These volumes contain
Operating Systems, Applications, and Data. We can create Snapshots from these
volumes or AMI from Snapshots.
The main advantage of EBS-backed volumes is that the data
can be configured to be stored for later retrieval even if the virtual machine
or instances are shut down.
66. You have an application running on an EC2 instance.
You need to reduce the load on your instance as soon as the CPU utilization
reaches 80 percent. How will you accomplish the job?
It can be done by creating an autoscaling group to deploy
more instances when the CPU utilization of the EC2 instance exceeds 80 percent
and distributing traffic among instances by creating an application load
balancer and registering EC2 instances as target instances.
67. In AWS, three different storage services are
available, such as EFS, S3, and EBS. When should I use Amazon EFS vs. Amazon S3
vs. Amazon Elastic Block Store (EBS)?
Amazon Web Services (AWS) offers cloud storage services to
support a wide range of storage workloads.
Amazon EFS is a file storage service for use with Amazon
compute (EC2, containers, and serverless) and on-premises servers. Amazon EFS
provides a file system interface, file system access semantics (such as strong
consistency and file locking), and concurrently accessible storage for up to
thousands of Amazon EC2 instances.
Amazon EBS is a block-level storage service for use with
Amazon EC2. Amazon EBS can deliver performance for workloads that require the
lowest latency for access to data from a single EC2 instance.
Amazon S3 is an object storage service. Amazon S3 makes
data available through an Internet API that can be accessed anywhere.
68. A company's web application is using multiple Linux
Amazon EC2 instances and stores data on Amazon EBS volumes. The company is
looking for a solution to increase the resiliency of the application in case of
a failure and to provide storage that complies with atomicity, consistency,
isolation, and durability (ACID). What should a solution architect do to meet
these requirements?
Create an Application Load Balancer with AWS Auto Scaling
groups across multiple Availability Zones. Store data on Amazon EFS and mount a
target on each instance.
69. An application running on AWS uses an Amazon Aurora
Multi-AZ deployment for its database. When evaluating performance metrics, a
solutions architect discovered that the database reads were causing high I/O
and adding latency to the write requests against the database. What should the
solution architect do to separate the read requests from the write requests?
Create a read replica and modify the application to use
the appropriate endpoint.
70. A client reports that they wanted to see an audit log
of any changes made to AWS resources in their account. What can the client do
to achieve this?
Enable AWS CloudTrail logs to be delivered to an Amazon S3
bucket.
71. Usually, you have noticed that one EBS volume can be
connected with one EC2 instance, our company wants to run a business-critical
application on multiple instances in a single region and needs to store all
instances output in single storage within the VPC. Instead of using EFS, our
company is recommending the use of multi-attach volume with instances. As an
architect, you need to suggest to them what instance type and EBS volumes they
should use.
The instance type should be EC2 Nitro-based instances and
Provisioned IOPs io1 multi-attach EBS volumes.
72. A company is using a VPC peering connection option to
connect its multiple VPCs in a single region to allow for cross-VPC
communication. A recent increase in account creation and VPCs has made it
difficult to maintain the VPC peering strategy, and the company expects to grow
to hundreds of VPCs. There are also new requests to create site-to-site VPNs
with some of the VPCs. A solutions architect has been tasked with creating a
central networking setup for multiple accounts and VPNs. Which networking solution
would you recommend to resolve it?
Configure a transit gateway with AWS Transit Gateway and
connect all VPCs and VPNs.
73. An organization has multiple facilities in various
continents such as North America, Europe, and the Asia Pacific. The
organization is designing a new distributed application to manage and optimize
its global supply chain and its manufacturing process. It needs to design the
process in such a way that the booked order in one continent should be able to
support data failover with a short Recovery Time Objective (RTO). The uptime of
the application should not impact manufacturing, what kind of solution would
you recommend as a solution architect?
Use Amazon DynamoDB global tables feature for the database
AWS Cloud Computing Interview Questions
74. What do you understand by VPC?
VPC is the abbreviated form of Virtual Private Cloud. It
allows you to launch AWS resources that can be defined by you and fully
customize the network configurations. Through VPC, you can define and take full
control of your virtual network environment. For example- you can have a
private address range, internet gateways, subnets, etc.
75. What are key pairs?
When connecting to an Amazon EC2 instance, you need to
prove your identity. Key pairs are used to execute this. A key pair is a set of
security credentials that are used during identity proofing. It consists of a
public key and a private key.
76. What are policies and what are the different types of
policies?
Policies define the permissions required to execute an
operation, irrespective of the method used to perform it. AWS supports six
types of policies:
Identity-based policies
Resource-based policies
Permissions boundaries
Organizations SCPs
ACLs
Session policies
1- Identity-based policies- These are JSON permissions
policy documents that control what actions an identity can perform, under what
conditions, and on which resources. These policies are further classified into
2 categories:
Managed Policies– These policies are standalone
identity-based policies that can be attached to different users, and groups in
your AWS environment.
Inline policies- These policies are directly attached to a
single user, group, or role. In situations where inline policies are used, a
strict one-to-one relationship between a policy and an identity is maintained.
2- Resource-based policies- These policies are the ones
attached to a resource such as an Amazon S3 bucket. They define which actions
can be performed on the particular resource and under what circumstances.
3- IAM permissions boundaries- They refer to the maximum level
of permissions that identity-based policies can grant to the specific entity.
4- Service Control Policies (SCPs)- SCPs are the maximum
level of permissions for an organization or organizational unit.
5- Access Control lists- They define and control which
principals in another AWS account can access the particular resource.
6- Session policies- They are advanced policies that are
passed as a parameter when a temporary session is programmatically created for
a role or federated user.
77. Which of the following is not an option in security
groups?
List of users
Ports
IP addresses
List of protocols
List of users
List of Users
Hope these top AWS cloud Interview questions and answers
for freshers and experienced, help you in preparing for top AWS jobs in the
Cloud market.
78. Unable to ping Instance We launched a Windows 2019 IIS
server in the Ohio region and deployed a dynamic website in this server, in
addition, the webserver also connected with a backend MS-SQL server to store
and access data related to the application. Our users were able to access the
website over the Internet. The next day our client informed us that they were
able to access the website, but weren’t able to ping the server from the
Internet. To ensure the ICMP rule in the Security Group, we checked, and the
Security Group had allowed the rule from 0.0.0.0/0. Would you try to help
troubleshoot the issue?
If the client can access the website from his/her end, it
means the connection is perfect and there is no issue with connectivity and the
Security Group configuration also seems correct.
We can check the internal firewall of the Windows 2019 IIS
server. If it is blocking ICMP traffic, we should enable it.
AWS Glue Interview Questions
79. Mention some of the significant features of AWS Glue.
Here are some of the significant features of AWS Glue
listed below.
Serverless data integration
Data quality and monitoring
Glue can automatically find structured and semi-structured
data kept in your data lake on Amazon S3, data warehouse on Amazon Redshift,
and other storage locations.
Automatically creates Scala or Python code for your ETL
tasks.
Allows you to visually clean and normalize data without
any code.
80. In AWS Glue, how do you enable and disable a trigger?
You can execute the below commands to start or stop the
trigger using the AWS CLI.
aws glue start-trigger –name MyTrigger
aws glue stop-trigger –name MyTrigger
81. What is a connection in AWS Glue?
Connection in AWS Glue is a service that stores
information required to connect to a data source such as Redshift, RDS, S3, or
DynamoDB.
82. How can you start an AWS Glue workflow run using AWS
CLI?
Using the start-workflow-run command of AWS CLI and
passing the workflow name, one can start the Glue workflow.
83. What is AWS Glue?
AWS Glue is a data integration service offered by Amazon
that makes it easy to discover, prepare, move, and transform your data for
analytics and application development.
84. What data sources does AWS Glue support?
AWS Glue can integrate with more than 80 data sources on
AWS, on-premises, and on other clouds. The service natively supports data
stored in the following databases in your Amazon Virtual Private Cloud (Amazon
VPC) running on Amazon Elastic Compute Cloud (Amazon EC2):
Amazon Aurora
Amazon RDS for MySQL
Amazon RDS for Oracle
Amazon RDS for PostgreSQL
Amazon RDS for SQL Server
Amazon Redshift
Amazon DynamoDB
Amazon S3
MySQL, Oracle, Microsoft SQL Server, and PostgreSQL
85. What programming language can we use to write my ETL
code for AWS Glue?
We can use either Scala or Python.
86. Can we write our code?
Yes. We can write your code using the AWS Glue ETL
library.
87. How does AWS Glue keep my data secure?
AWS Glue provides server-side encryption for data at rest
and SSL for data in motion.
88. How do we monitor the execution of my AWS Glue jobs?
AWS Glue provides the status of each job and pushes all
notifications to CloudWatch.
AWS S3 Interview Questions
89. Explain what S3 is.
S3 stands for Simple Storage Service. It is an
object-based storage service on AWS. It is a pay-as-you-go service with the
help of which you can store and extract any amount of data at any time from
anywhere on the web.
90. What is the Replication Rule feature supported by AWS
S3?
Amazon S3 offers a lot of useful features. One of the
features is the Replication Rule feature, which allows users to replicate the
data to a secondary region.
91. What are the different ways to encrypt a file in S3?
To encrypt a file in Amazon S3, users need to choose an
appropriate encryption option. AWS S3 offers multiple encryption options such
as:
Server-Side Encryption with Amazon S3 Managed Keys
(SSE-S3)
Server-Side Encryption with AWS Key Management Service
(SSE-KMS)
Server-Side Encryption with Customer-Provided Keys (SSE-C)
Client-Side Encryption.
92. What is static website hosting in S3?
Static website hosting in S3 is a feature that allows
users to host static web content directly from an S3 bucket.
93. Is there any possible way to restore the deleted S3
objects?
Yes, you can restore the deleted S3 objects easily if you
have a versioning-enabled bucket.
94. What is the maximum number of S3 buckets you can
create?
The maximum number of S3 buckets that can be created is
100.
95. How will you configure an Amazon S3 bucket to serve
static assets for your public web application?
By configuring the bucket policy to provide public read
access to all objects.
96. What is an S3 general-purpose bucket?
A bucket is a container for objects stored in Amazon S3,
and you can store any number of objects in a bucket. General purpose buckets
are the original S3 bucket type, and a single general purpose bucket can
contain objects stored across all storage classes except S3 Express One Zone.
They are recommended for most use cases and access patterns.
97. How is Amazon S3 data organized?
Amazon S3 is a simple key-based object store. When you
store data, you assign a unique object key that can later be used to retrieve
the data. Keys can be any string, and they can be constructed to mimic
hierarchical attributes. Alternatively, you can use S3 Object Tagging to
organize your data across all of your S3 buckets and/or prefixes.
AWS EC2 Interview Questions
98. What do you understand about Amazon EC2?
Elastic Cloud Compute (EC2) is a cloud computing service
provided by Amazon Web Services (AWS) that is used for hosting applications in
the cloud.
99. What is the Security Group in Amazon EC2?
Amazon Security Groups control both inbound and outbound
traffic, acting like a virtual firewall for your EC2 instances.
100. Explain Stop vs. Terminate an Amazon EC2 instance.
Both stopping and terminating the Amazon EC2 instance have
their purpose and consequences.
Feature Stop Terminate
Action Halts the
EC2 instance Deletes the instance
Restart Possible Not possible
Use Case Temporary
pause Long-term shutdowns
101. What is the use of regions and availability zones in
Amazon EC2 configuration?
To deploy your EC2 instance geographically, you need to
choose the region. To distribute these instances across multiple zones for
improved fault tolerance and high availability, you need to select the specific
availability zone within that region.
102. What is the placement group in EC2?
A placement group in Amazon EC2 is a logical grouping of
instances within a single availability zone.
103. You have an application running on your Amazon EC2
instance. You want to reduce the load on your instance as soon as the CPU
utilization reaches 100 percent. How will you do that?
It can be done by creating an autoscaling group to deploy
more instances when the CPU utilization exceeds 100 percent and distributing
traffic among instances by creating a load balancer and registering the Amazon
EC2 instances with it.
104. Suppose, I created a subnet and launched an EC2
instance in the subnet with default settings. Which of the following options
will be ready to use on the EC2 instance as soon as it is launched?
Elastic IP
Private IP
Public IP
Internet Gateway
Private IP. Private IP is automatically assigned to the
instance as soon as it is launched. While elastic IP has to be set manually,
Public IP needs an Internet Gateway which again has to be created since it’s a
new VPC.
105. Can you name the additional network interface that
can be created and attached to your Amazon EC2 instance launched in your VPC?
AWS Elastic Network Interface
106. You want to modify the security group rules while
it is being used by multiple EC2 instances. Will you be able to do that? If
yes, will the new rules be implemented on all previously running EC2 instances
that were using that security group?
Yes, the security group that is being used by multiple
EC2 instances can be modified. The changes will be implemented immediately and
applied to all the previously running EC2 instances without restarting the
instances.
107. Can you change the private IP address of an EC2
instance while it is running or in a stopped state?
No, it cannot be changed. When an EC2 instance is
launched, a private IP address is assigned to that instance at boot time. This
private IP address is attached to the instance for its entire lifetime and can
never be changed.
AWS Interview Questions for Solution Architect
108. What is Amazon Virtual Private Cloud (VPC), and
why is it used?
Amazon Virtual Private Cloud (VPC) is a service offered
by AWS that enables users to create their isolated virtual network inside the
AWS cloud.
109. What is an AWS Load Balancer?
AWS Load Balancer is a resource that is used to
distribute the traffic across the EC2 instance.
110. What is AWS SNS?
Amazon Simple Notification Service (SNS) is a fully
managed push notification service that sends messages to mobile and other
distributed systems.
111. What are the different types of load balancers in
EC2?
There are 3 types of Load Balancers in EC2.
Application Load Balancer (ALB)
Network Load Balancer (NLB)
Classic Load Balancer (CLB)
112. What is AWS CloudFormation?
AWS CloudFormation is an IaC (Infrastructure as Code)
service offered by AWS so that you can deploy your infrastructure redundantly
with the help of a code (json or Yaml).
113. What would I have to do if I wanted to access
Amazon Simple Storage buckets and use the information for access audits?
AWS CloudTrail can be used in this case as it is
designed for logging and tracking API calls, and it has also been made
available for storage solutions.
114. What are the native AWS security logging
capabilities?
The native AWS security logging capabilities include
AWS CloudTrail, AWS Config, AWS detailed billing reports, Amazon S3 access
logs, Elastic load balancing Access logs, Amazon CloudFront access logs, Amazon
VPC Flow logs, etc.
115. You have an application running on an EC2
instance. You need to reduce the load on your instance as soon as the CPU
utilization reaches 80 percent. How will you accomplish the job?
It can be done by creating an autoscaling group to
deploy more instances when the CPU utilization of the EC2 instance exceeds 80
percent and distributing traffic among instances by creating an application
load balancer and registering EC2 instances as target instances.
116. Is it possible to switch from an Instance-backed
root volume to an EBS-backed root volume at any time?
No, it is not possible.
117. Your organization has around 50 IAM users. Now, it
wants to introduce a new policy that will affect the access permissions of an
IAM user. How can it implement this without having to apply the policy at the
individual user level?
It is possible to use AWS IAM groups, by adding users
in the groups as per their roles and by simply applying the policy to the
groups.
AWS VPC Interview Questions
118. What is the difference between stateful and
stateless filtering?
Stateful filtering evaluates the origin IP address for
any request coming to your server, whereas stateless filtering evaluates both
the origin and the destination IP address.
119. What are the internet gateways in VPC?
Internet gateways are components that allow resources
within your VPC to communicate to and from the Internet.
120. What are the security groups in a VPC?
Security groups are essential elements that handle the
flow of data within the Virtual Private Cloud. They possess the capability to
function as formidable firewalls by employing distinct regulations that govern
the movement of information. These security groups exist both within the
Virtual Private Cloud and in the EC2 segments of the Amazon Web Services
console.
121. What is an Elastic IP address?
An Elastic IP address is more like a public IP address,
which is connected to the AWS account until a user terminates it.
122. What is a NAT device?
NAT stands for Network Address Translation and as the
name implies, it replaces the IP address for devices that are running in our
network, i.e., when we send any traffic from any device to the internet, it
will replace the IP address of the device with another IP address.
123. How many total VPCs per account/region and subnets
per VPC can you have?
We can have a total of 5 VPCs for every account/region
and 200 subnets for every VPC that you have.
124. You accidentally stopped an EC2 instance in a VPC
with an associated Elastic IP. If you start the instance again, what will be
the result?
Elastic IP will only be disassociated from the instance
if it’s terminated. If it’s stopped and started, there won’t be any change to
the instance, and no data will be lost.
125. You have created a VPC with private and public
subnets. In what kind of subnet would you launch the database servers?
Database servers should be ideally launched on private
subnets. Private subnets are ideal for the backend services and databases of
all applications since they are not meant to be accessed by the users of the
applications, and private subnets are not routable from the Internet.
126. How many subnets can we create per VPC?
Currently, you can create 200 subnets per VPC.
127. Can I monitor the network traffic in my VPC?
Yes. You can use Amazon VPC traffic mirroring and
Amazon VPC flow logs features to monitor the network traffic in your Amazon
VPC.
AWS Sal
Scenario based qsns
Q : Your website experiences varying levels of
traffic throughout the day. How can you ensure that your Amazon EC2 instances
automatically scale up and down based on demand?
You can use Amazon EC2 Auto Scaling to automatically adjust
the number of instances based on predefined scaling policies. Define scaling
policies based on metrics like CPU utilization or network traffic. When traffic
increases, EC2 Auto Scaling adds more instances, and when traffic decreases, it
removes instances, ensuring optimal performance and cost efficiency.
Mastering
Amazon Auto Scaling: Practical POCs and Use Cases
Q : You have an application that requires
extremely low-latency communication between instances. How can you achieve this
on Amazon EC2?
To achieve low-latency communication between instances, you
can use EC2 Placement Groups. Placement Groups enable instances to be placed in
close proximity within the same Availability Zone (AZ). There are two types of
Placement Groups: Cluster Placement Groups for low-latency and High-Performance
Computing (HPC) workloads and Spread Placement Groups for critical instances
that require maximum separation to minimize the risk of simultaneous failure.
Q : Your application needs to store sensitive data, and
you want to ensure that the data is encrypted at rest on EC2 instances. How can
you enable this encryption?
To encrypt data at rest on EC2 instances, you can use Amazon
Elastic Block Store (EBS) volumes with encryption enabled. When creating or
modifying an EBS volume, you can specify the use of AWS Key Management Service
(KMS) to manage the encryption keys. Data written to the EBS volume is
automatically encrypted, and it remains encrypted at rest.
Q : Your team is developing a containerized application
and wants to deploy it on EC2 instances. Which service can you use to manage
the containers on EC2 efficiently?
You can use Amazon Elastic Container Service (ECS) or Amazon
Elastic Kubernetes Service (EKS) to manage containers on EC2 instances. ECS is
a fully-managed service for running containers at scale, while EKS provides
Kubernetes management capabilities for container orchestration. Both services
simplify the process of deploying, managing, and scaling containerized
applications on EC2 instances.
Q : Your application requires GPU capabilities for
machine learning or graphics-intensive workloads. How can you launch EC2
instances with GPU support?
You can launch EC2 instances with GPU support by selecting
an instance type that offers GPU resources. Examples of such instances include
the “p3” and “g4” instance families. These instances are optimized for
different GPU workloads, and you can choose the one that best fits your
specific use case.
Q : You need to ensure that your EC2 instances are
running in a private network and are not directly accessible from the internet.
How can you achieve this?
To run EC2 instances in a private network without direct
internet access, you can place them in an Amazon Virtual Private Cloud (VPC)
with private subnets. To access the instances securely, you can set up a
bastion host (jump host) in a public subnet, which acts as a gateway for
connecting to the private instances through SSH or RDP.
Q : You want to enhance the security of your EC2
instances by restricting incoming traffic only to specific IP addresses. How
can you implement this security measure?
To restrict incoming traffic to specific IP addresses on EC2
instances, you can configure security group rules. Security groups act as
virtual firewalls and allow you to control inbound and outbound traffic. By
specifying the desired IP ranges in the inbound rules, you can ensure that only
traffic from those IP addresses is allowed to reach the instances.
Q : Your organization needs to store and share data
files across multiple EC2 instances. What service can you use to achieve
scalable and durable file storage?
You can use Amazon Elastic File System (EFS) to achieve
scalable and durable file storage for multiple EC2 instances. EFS provides a
managed file system that can be mounted concurrently by multiple instances
within a VPC. It supports the Network File System (NFS) protocol and
automatically scales capacity as data grows.
Q : Your team wants to minimize the cost of running EC2
instances for non-production environments (e.g., development and testing). How
can you achieve cost savings without compromising availability?
To minimize costs for non-production environments while
maintaining high availability, you can use EC2 Spot Instances. Spot Instances
allow you to bid on spare EC2 capacity, and they can significantly reduce costs
compared to On-Demand Instances. However, keep in mind that Spot Instances can
be terminated when the Spot price exceeds your bid, so they are best suited for
stateless and fault-tolerant workloads.
Q : Your application requires the ability to quickly
recover from instance failure and ensure data durability. What type of Amazon
EBS volume is recommended for such scenarios?
To ensure data durability and quick recovery from instance
failure, you can use Amazon EBS volumes with the “io1” (Provisioned IOPS) type.
“io1” volumes provide the highest performance and reliability and are ideal for
critical workloads that demand consistent and low-latency performance.
Q : Your organization needs to control the launch
permissions of Amazon Machine Images (AMIs) and prevent accidental termination
of EC2 instances. What AWS service can help you manage these permissions
effectively?
You can use AWS Identity and Access Management (IAM) to
manage launch permissions for AMIs and control who can launch instances from
specific AMIs. IAM allows you to define policies that restrict or grant
permissions for different users or groups. Additionally, IAM roles can be used
to control what actions EC2 instances are allowed to perform, reducing the risk
of accidental terminations.
Q : Your organization needs to host a web application
that requires consistent CPU performance and low latency. Which EC2 instance
type would you recommend, and why?
For applications requiring consistent CPU performance and
low latency, I would recommend using an EC2 instance from the “c5” or “m5”
instance families. Both families are designed for compute-intensive workloads,
with the “c5” instances offering higher CPU performance and the “m5” instances
providing a balance of compute and memory resources.
Q : Your application involves batch processing of large
datasets. How can you optimize the EC2 instances for such a workload?
For batch processing of large datasets, you can use EC2
instances from the “r5” instance family, which is optimized for
memory-intensive workloads. By choosing an instance with sufficient memory, you
can avoid performance bottlenecks caused by frequent disk swapping, enhancing
the efficiency of your batch processing.
Q : You need to create a cost-effective, scalable, and
fault-tolerant web application architecture. How can you achieve this with EC2?
To create a cost-effective, scalable, and fault-tolerant web
application architecture, you can use EC2 instances with Elastic Load Balancing
(ELB) and Auto Scaling. ELB distributes incoming traffic among multiple EC2
instances, while Auto Scaling automatically adjusts the number of instances
based on demand, ensuring optimal performance and cost-efficiency.
Q : Your team is developing a microservices-based
application and wants to deploy it on EC2 instances. What are some best
practices to ensure the scalability and maintainability of the microservices
architecture?
To ensure the scalability and maintainability of a
microservices-based application on EC2, consider the following best practices:
- Deploy
each microservice on separate EC2 instances to achieve isolation.
- Use
containerization technology like Docker to package and deploy
microservices consistently.
- Implement
an orchestration service like Amazon ECS or Amazon EKS to manage the
containerized microservices efficiently.
- Design
microservices with loosely coupled communication to enable independent
scaling and deployment.
Q : Your organization needs to run a Windows-based
application on EC2 instances. How can you ensure that the instances are
automatically updated with the latest Windows patches?
To automatically update Windows-based EC2 instances with the
latest patches, you can use AWS Systems Manager Patch Manager. Patch Manager
simplifies the process of managing Windows updates by automating patching and
providing insights into compliance and patching status.
Q : Your application requires low-latency access to a
relational database. How can you optimize EC2 instances to minimize database
response times?
To minimize database response times and achieve low-latency
access, you can deploy EC2 instances in the same AWS Region and Availability
Zone as the database. Additionally, consider using Amazon RDS Read Replicas to
offload read traffic from the primary database, reducing the load and improving
overall database performance.
Q : Your application must handle spikes in traffic
during seasonal promotions. How can you ensure that the EC2 instances scale up
automatically during peak times and scale down during off-peak times?
To automatically scale EC2 instances during peak and
off-peak times, you can use Amazon EC2 Auto Scaling with scheduled scaling
policies. Set up a schedule to increase the desired capacity before the
expected peak traffic and decrease it afterward. EC2 Auto Scaling will adjust
the number of instances based on the schedule, ensuring you have the right
capacity when you need it.
Q : Your organization is migrating a legacy application
to AWS EC2. The application requires direct access to the underlying hardware.
What EC2 feature can you use to fulfill this requirement?
To gain direct access to the underlying hardware, you can
use Amazon EC2 Dedicated Hosts. EC2 Dedicated Hosts provide dedicated physical
servers to run your instances, allowing you to launch instances on specific
hardware for compliance, licensing, or regulatory requirements.
Q : Your team is running multiple applications on EC2
instances, and you want to optimize costs by leveraging unused compute
capacity. What EC2 pricing option should you choose?
To optimize costs and leverage unused compute capacity, you
can use Amazon EC2 Spot Instances. Spot Instances allow you to bid on spare EC2
capacity, typically providing significant cost savings compared to On-Demand
Instances. However, be aware that Spot Instances can be terminated when the
Spot price exceeds your bid.
Q : You need to migrate an on-premises virtual machine
(VM) to AWS EC2. What service can you use to simplify the VM migration process?
To simplify the migration of on-premises VMs to AWS EC2, you
can use AWS Server Migration Service (SMS). SMS allows you to automate,
schedule, and track incremental replications of VMs from your data center to
AWS, reducing the complexity of the migration process.
Q : Your application requires frequent changes and
updates, and you want to test new features without affecting the production
environment. How can you achieve this with EC2?
To test new features and changes without affecting the
production environment, you can create an Amazon Machine Image (AMI) of the
existing production EC2 instance. Launch a new EC2 instance using the AMI in a
separate testing environment. This isolated environment allows you to
experiment and validate changes before applying them to the production
instance.
Q : You want to implement data encryption in transit
for communication between your EC2 instances and Amazon S3. How can you achieve
this security measure?
To implement data encryption in transit between EC2
instances and Amazon S3, use Amazon S3 Transfer Acceleration with SSL/TLS
encryption enabled. By enabling Transfer Acceleration, data is transferred over
an optimized network path with encryption, improving upload and download speeds
while ensuring data security.
Q : Your application relies on stateful connections
between clients and servers, and you need to preserve these connections even if
an EC2 instance fails. What service can you use to achieve this?
To preserve stateful connections even if an EC2 instance
fails, you can use Elastic IP addresses (EIPs) in combination with Auto
Scaling. Associate an EIP with an EC2 instance to create a static public IP
address that remains associated with the instance even if it is terminated.
Auto Scaling will automatically replace any failed instances and associate the
EIP with the new instance, preserving the client connections.
Q : Your development team needs to share sensitive data
securely between EC2 instances. How can you set up a secure communication
channel for this purpose?
To set up a secure communication channel between EC2
instances, you can use Virtual Private Cloud (VPC) peering or AWS PrivateLink.
VPC peering allows you to connect VPCs within the same AWS account privately.
AWS PrivateLink enables secure and private communication between VPCs and
supported AWS services without traversing the internet.
Q : Your organization requires on-premises resources to
communicate securely with EC2 instances within a VPC. How can you establish a
secure connection between your on-premises network and the VPC?
To establish a secure connection between your on-premises
network and an EC2 instance within a VPC, you can use AWS Site-to-Site VPN or
AWS Direct Connect. Site-to-Site VPN creates an encrypted tunnel over the
internet, whereas Direct Connect provides a dedicated connection through a
private network link.
Q : Your team wants to ensure that only authorized
personnel can access the EC2 instances via SSH. What security measure should be
implemented?
To ensure that only authorized personnel can access the EC2
instances via SSH, you should configure the security group rules to allow
inbound SSH access only from specific IP addresses or ranges associated with
authorized personnel. Additionally, you can manage SSH access using IAM roles
and AWS Systems Manager Session Manager for secure remote management.
Q : Your organization wants to ensure that EC2
instances are protected against common security threats and vulnerabilities.
What service can you use to monitor and assess the security posture of your
instances?
You can use Amazon Inspector to monitor and assess the
security posture of your EC2 instances. Amazon Inspector automatically assesses
instances for vulnerabilities and security deviations based on predefined
rulesets, providing you with detailed findings and recommendations to enhance
the security of your environment.
Q : Your application requires high network performance
and low latency communication between EC2 instances in different Availability
Zones. What service can you use to achieve this requirement?
To achieve high network performance and low latency
communication between EC2 instances in different Availability Zones, you can
use Enhanced Networking with Elastic Network Adapter (ENA). ENA optimizes
network performance for EC2 instances, allowing for faster and more reliable
inter-instance communication.
Q : Your team wants to automate the process of managing
EC2 instances and their configurations. Which AWS service can you use for this
purpose?
You can use AWS Systems Manager to automate the process of
managing EC2 instances and their configurations. Systems Manager provides a
unified interface for managing EC2 instances, including tasks like patch
management, configuration management, and instance inventory.
Q : You need to run Windows-based applications on EC2
instances, and your team requires remote desktop access for management
purposes. How can you enable remote desktop access to Windows EC2 instances?
To enable remote desktop access to Windows EC2 instances,
you need to configure the Windows Firewall and EC2 Security Groups to allow
Remote Desktop Protocol (RDP) traffic (port 3389). Additionally, ensure that
you have the necessary credentials to log in to the instances remotely.
Q : Your team wants to monitor the performance of EC2
instances and set up alerts for abnormal behavior. What AWS service can help
you achieve this?
To monitor the performance of EC2 instances and set up
alerts, you can use Amazon CloudWatch. CloudWatch provides a comprehensive set
of monitoring and alerting capabilities, allowing you to collect and track
metrics, set alarms, and automatically react to changes in your EC2 instances’
performance.
Q : You want to deploy your web application to multiple
regions to ensure high availability and low latency. What AWS service can you
use to automate the deployment process across regions?
You can use AWS Elastic Beanstalk to automate the deployment
process of your web application across multiple regions. Elastic Beanstalk
simplifies application deployment by automatically handling capacity
provisioning, load balancing, scaling, and application health monitoring.
Q : Your organization needs to ensure data privacy and
compliance by restricting access to EC2 instances based on user roles. How can
you achieve this?
To restrict access to EC2 instances based on user roles, you
can use AWS Identity and Access Management (IAM) to manage user permissions.
Define IAM roles with specific permissions and assign them to users or groups.
Users can access the EC2 instances based on the permissions associated with
their roles.
Q : Your application requires a mix of Linux and
Windows instances to handle different tasks. Can you use the same security
groups for both Linux and Windows instances?
Yes, you can use the same security groups for both Linux and
Windows instances. Security groups are a stateful firewall that controls
inbound and outbound traffic based on rules you define, regardless of the
operating system.
Q : Your team wants to ensure that your EC2 instances
are accessible over the internet while still being protected from unauthorized
access. What security measure can you implement?
To ensure that your EC2 instances are accessible over the
internet while being protected, you can use a combination of security groups
and Network Access Control Lists (NACLs). Security groups control inbound and
outbound traffic for EC2 instances, while NACLs control traffic to and from
subnets, providing an additional layer of security.
Q : Your application requires persistent data storage
that survives instance termination. What storage option can you use on EC2 for
this purpose?
For persistent data storage that survives instance
termination, you can use Amazon Elastic Block Store (EBS) volumes. EBS volumes
are durable, block-level storage devices that can be attached to EC2 instances
and persist independently of the instance lifecycle.
Q : Your organization wants to ensure that EC2
instances are launched only within specific AWS Regions. How can you enforce
this policy?
To enforce the launching of EC2 instances within specific
AWS Regions, you can use AWS Service Control Policies (SCPs) with AWS
Organizations. SCPs allow you to set permissions that apply to the entire
organization or specific organizational units, ensuring that instances are
launched only in approved regions.
Q : Your application processes a large number of data
records, and you want to distribute the workload efficiently across multiple
EC2 instances. What AWS service can you use for this purpose?
To distribute the workload efficiently across multiple EC2
instances, you can use Amazon Elastic MapReduce (EMR). EMR is a managed service
that simplifies the processing of large datasets using popular data processing
frameworks like Apache Hadoop and Apache Spark.
Q : Your team is designing a solution for disaster
recovery and business continuity. How can you replicate EC2 instances and data
across AWS Regions?
To replicate EC2 instances and data across AWS Regions for
disaster recovery, you can use AWS Disaster Recovery Solutions such as AWS
Backup, AWS Database Migration Service (DMS), and AWS Lambda functions to
automate the replication process.
Q : Your application requires instances with large
amounts of storage for database backups and archiving. What EC2 instance family
is best suited for this use case?
For applications that require instances with large amounts
of storage, you can use EC2 instances from the “i3” or “d2” instance families.
These instance families are optimized for storage-intensive workloads, with
“i3” instances offering high-performance local NVMe SSD storage, and “d2”
instances providing cost-effective HDD storage.
Q : Your application needs to support both IPv4 and
IPv6 traffic. How can you ensure that EC2 instances can handle both types of
traffic?
To ensure that EC2 instances can handle both IPv4 and IPv6
traffic, you need to enable dual-stack networking on your VPC. With dual-stack
enabled, EC2 instances can communicate with both IPv4 and IPv6 addresses.
Q : Your organization needs to run a highly regulated
workload that requires strict access control and monitoring. What AWS service
can you use to enforce fine-grained access permissions and logging?
To enforce fine-grained access permissions and logging for a
highly regulated workload, you can use AWS Identity and Access Management (IAM)
with AWS CloudTrail. IAM allows you to manage user access to AWS resources,
while CloudTrail provides detailed logs of API calls made by users and
services.
Q : Your organization wants to reduce costs for
development and testing environments, which are only required during specific
hours of the day. How can you achieve cost savings?
To reduce costs for development and testing environments,
you can use EC2 Instance Scheduler. EC2 Instance Scheduler allows you to
automatically start and stop EC2 instances based on a defined schedule,
ensuring that instances are only running when needed.
Enjoyed the scenario-based questions on Amazon EC2?
Follow me on Medium for more engaging content on AWS, Azure,
GCP, and beyond.
Let’s connect on LinkedIn for
the latest updates.
Basic AWS Interview Questions
Starting with the fundamentals, this section introduces
basic AWS interview questions essential for building a foundational
understanding. It's tailored for those new to AWS or needing a refresher,
setting the stage for more detailed exploration later.
1. What is cloud computing?
Cloud computing provides on-demand access to IT resources
like compute, storage, and databases over the internet. Users pay only for what
they use instead of owning physical infrastructure.
Cloud enables accessing technology services flexibly as needed
without big upfront investments. Leading providers like AWS offer a wide range
of cloud services via the pay-as-you-go consumption model. Our AWS
Cloud Concepts course covers many of these basics.
2. What is the problem with the traditional IT approach
compared to using the Cloud?
Multiple industries are moving away from traditional IT to
adopt cloud infrastructures for multiple reasons. This is because the Cloud
approach provides greater business agility, faster innovation, flexible scaling
and lower total cost of ownership compared to traditional IT. Below are some of
the characteristics that differentiate them:
|
Traditional
IT |
Cloud
computing |
|
|
3. How many types of deployment models exist in the
cloud?
There are three different types of deployment models in the
cloud, and they are illustrated below:
- Private
cloud: this type of service is used by a single organization and
is not exposed to the public. It is adapted to organizations using
sensitive applications.
- Public
cloud: these cloud resources are owned and operated by
third-party cloud services like Amazon Web Services, Microsoft Azure, and
all those mentioned in the AWS market share section.
- Hybrid
cloud: this is the combination of both private and public clouds.
It is designed to keep some servers on-premises while extending the
remaining capabilities to the cloud. Hybrid cloud provides flexibility and
cost-effectiveness of the public cloud.
4. What are the five characteristics of cloud computing?
Cloud computing is composed of five main characteristics,
and they are illustrated below:
- On-demand
self-service: Users can provision cloud services as needed
without human interaction with the service provider.
- Broad
network access: Services are available over the network and
accessed through standard mechanisms like mobile phones, laptops, and
tablets.
- Multi-tenacy
and resource pooling: Resources are pooled to serve multiple
customers, with different virtual and physical resources dynamically
assigned based on demand.
- Rapid
elasticity and scalability: Capabilities can be elastically
provisioned and scaled up or down quickly and automatically to match
capacity with demand.
- Measured
service: Resource usage is monitored, controlled, reported, and
billed transparently based on utilization. Usage can be managed,
controlled, and reported, providing transparency for the provider and
consumer.
5. What are the main types of Cloud Computing?
There are three main types of cloud computing: IaaS, PaaS,
and SaaS
- Infrastructure
as a Service (IaaS): Provides basic building blocks for cloud IT
like compute, storage, and networking that users can access on-demand
without needing to manage the underlying infrastructure. Examples: AWS
EC2, S3, VPC.
- Platform
as a Service (PaaS): Provides a managed platform or environment
for developing, deploying, and managing cloud-based apps without needing
to build the underlying infrastructure. Examples: AWS Elastic Beanstalk,
Heroku
- Software
as a Service (SaaS): Provides access to complete end-user
applications running in the cloud that users can use over the internet.
Users don't manage infrastructure or platforms. Examples: AWS Simple Email
Service, Google Docs, Salesforce CRM.
You can explore these in more detail in our Understanding Cloud Computing course.
6. What is Amazon EC2, and what are its main uses?
Amazon EC2 (Elastic Compute Cloud) provides scalable virtual
servers called instances in the AWS Cloud. It is used to run a variety of
workloads flexibly and cost-effectively. Some of its main uses are illustrated
below:
- Host
websites and web applications
- Run
backend processes and batch jobs
- Implement
hybrid cloud solutions
- Achieve
high availability and scalability
- Reduce
time to market for new use cases
7. What is Amazon S3, and why is it important?
Amazon Simple Storage Service (S3) is a versatile, scalable,
and secure object storage service. It serves as the foundation for many
cloud-based applications and workloads. Below are a few features highlighting
its importance:
- Durable
with 99.999999999% durability and 99.99% availability, making it suitable
for critical data.
- Supports
robust security features like access policies, encryption, VPC endpoints.
- Integrates
seamlessly with other AWS services like Lambda, EC2, EBS, just to name a
few.
- Low
latency and high throughput make it ideal for big data analytics, mobile
applications, media storage and delivery.
- Flexible
management features for monitoring, access logs, replication, versioning,
lifecycle policies.
- Backed
by the AWS global infrastructure for low latency access worldwide.
8. Explain the concept of ‘Regions’ and ‘Availability
Zones’ in AWS
- AWS
Regions correspond to separate geographic locations where AWS
resources are located. Businesses choose regions close to their customers
to reduce latency, and cross-region replication provides better disaster
recovery.
- Availability
zones consist of one or more discrete data centers with redundant power,
networking, and connectivity. They allow the deployment of resources in a
more fault-tolerant way.
Our course AWS
Cloud Concepts provides readers with a complete guide to learning
about AWS’s main core services, best practices for designing AWS applications,
and the benefits of using AWS for businesses.
Become a Data Engineer
Become a data engineer through advanced Python learning
AWS Interview Questions for Intermediate and Experienced
AWS DevOps interview questions
Moving to specialized roles, the emphasis here is on how AWS
supports DevOps practices. This part examines the automation and optimization
of AWS environments, challenging individuals to showcase their skills in
leveraging AWS for continuous integration and delivery. If you're going for an
advanced AWS role, check out our Data Architect Interview Questions blog post to
practice some data infrastructure and architecture questions.
9. How do you use AWS CodePipeline to automate a CI/CD
pipeline for a multi-tier application?
CodePipeline can be used to automate the flow from code
check-in to build, test, and deployment across multiple environments to
streamline the delivery of updates while maintaining high standards of quality.
The following steps can be followed to automate a CI/CD
pipeline:
- Create
a Pipeline: Start by creating a pipeline in AWS CodePipeline,
specifying your source code repository (e.g., GitHub, AWS CodeCommit).
- Define
Build Stage: Connect to a build service like AWS CodeBuild to
compile your code, run tests, and create deployable artifacts.
- Setup
Deployment Stages: Configure deployment stages for each tier of
your application. Use AWS CodeDeploy to automate deployments to Amazon EC2
instances, AWS Elastic Beanstalk for web applications, or AWS ECS for
containerized applications.
- Add
Approval Steps (Optional): For critical environments, insert
manual approval steps before deployment stages to ensure quality and
control.
- Monitor
and Iterate: Monitor the pipeline's performance and adjust as
necessary. Utilize feedback and iteration to continuously improve the
deployment process.
10. What key factors should be considered in designing a
deployment solution on AWS to effectively provision, configure, deploy, scale,
and monitor applications?
Creating a well-architected AWS deployment involves
tailoring AWS services to your app's needs, covering compute, storage, and
database requirements. This process, complicated by AWS's vast service catalog,
includes several crucial steps:
- Provisioning: Set
up essential AWS infrastructure such as EC2, VPC, subnets or managed
services like S3, RDS, CloudFront for underlying applications.
- Configuring: Adjust
your setup to meet specific requirements related to the environment,
security, availability, and performance.
- Deploying: Efficiently
roll out or update app components, ensuring smooth version transitions.
- Scaling: Dynamically
modify resource allocation based on predefined criteria to handle load
changes.
- Monitoring: Keep
track of resource usage, deployment outcomes, app health, and logs to
ensure everything runs as expected.
11. What is Infrastructure as a Code? Describe in your
own words
Infrastructure as Code (IaC) is a method of managing and
provisioning computer data centers through machine-readable definition files,
rather than physical hardware configuration or interactive configuration tools.
Essentially, it allows developers and IT operations teams to
automatically manage, monitor, and provision resources through code, rather
than manually setting up and configuring hardware.
Also, IaC enables consistent environments to be deployed
rapidly and scalably by codifying infrastructure, thereby reducing human error
and increasing efficiency.
12. What is your approach to handling continuous
integration and deployment in AWS DevOps?
In AWS DevOps, continuous integration and deployment can be
managed by utilizing AWS Developer Tools. Begin by storing and versioning your
application's source code with these tools.
Then, leverage services like AWS CodePipeline for
orchestrating the build, test, and deployment processes. CodePipeline serves as
the backbone, integrating with AWS CodeBuild for compiling and testing code,
and AWS CodeDeploy for automating the deployment to various environments. This
streamlined approach ensures efficient, automated workflows for continuous
integration and delivery.
13. How does Amazon ECS benefit AWS DevOps?
Amazon ECS is a scalable container management service that
simplifies running Docker containers on EC2 instances through a
managed cluster, enhancing application deployment and operation.
14. Why might ECS be preferred over Kubernetes?
ECS offers greater flexibility, scalability, and simplicity
in implementation compared to Kubernetes, making it a preferred choice for some
deployments.
AWS solution architect interview questions
For solution architects, the focus is on designing AWS
solutions that meet specific requirements. This segment tests the ability to
create scalable, efficient, and cost-effective systems using AWS, highlighting
architectural best practices.
15. What is the role of an AWS solution architect?
AWS Solutions Architects design and oversee applications on
AWS, ensuring scalability and optimal performance. They guide developers,
system administrators, and customers on utilizing AWS effectively for their
business needs and communicate complex concepts to both technical and
non-technical stakeholders.
16. What are the key security best practices for AWS EC2?
Essential EC2 security practices include using IAM for
access management, restricting access to trusted hosts, minimizing permissions,
disabling password-based logins for AMIs, and implementing multi-factor
authentication for enhanced security.
17. What is AWS VPC and its purpose?
Amazon VPC enables the deployment of AWS resources within a
virtual network that is architecturally similar to a traditional data center
network, offering the advantage of AWS's scalable infrastructure.
18. What are the strategies to create a highly available
and fault-tolerant AWS architecture for critical web applications?
Building a highly available and fault-tolerant architecture
on AWS involves several strategies to reduce the impact of failure and ensure
continuous operation. Key principles include:
- Implementing
redundancy across system components to eliminate single points of failure
- Using
load balancing to distribute traffic evenly and ensure optimal performance
- Setting
up automated monitoring for real-time failure detection and response.
Systems should be designed for scalability to handle varying loads, with a
distributed architecture to enhance fault tolerance.
- Employing
fault isolation, regular backups, and disaster recovery plans are
essential for data protection and quick recovery.
- Additionally,
designing for graceful degradation maintains functionality during outages,
while continuous testing and deployment practices improve system
reliability.
19. Explain how you would choose between Amazon RDS,
Amazon DynamoDB, and Amazon Redshift for a data-driven application.
Choosing between Amazon RDS, DynamoDB, and Redshift for a
data-driven application depends on your specific needs:
- Amazon
RDS is ideal for applications that require a traditional
relational database with standard SQL support, transactions, and complex
queries.
- Amazon DynamoDB suits applications needing a
highly scalable, NoSQL database with fast, predictable performance at any
scale. It's great for flexible data models and rapid development.
- Amazon
Redshift is best for analytical applications requiring complex
queries over large datasets, offering fast query performance by using
columnar storage and data warehousing technology.
20. What considerations would you take into account when
migrating an existing on-premises application to AWS? Use an example of choice.
When moving a company's customer relationship management
(CRM) software from an in-house server setup to Amazon Web Services (AWS), it's
essential to follow a strategic framework similar to the one AWS suggests,
tailored for this specific scenario:
- Initial
Preparation and Strategy Formation
- Evaluate
the existing CRM setup to identify limitations and areas for improvement.
- Set
clear migration goals, such as achieving better scalability, enhancing
data analysis features, or cutting down on maintenance costs.
- Identify
AWS solutions required, like leveraging Amazon EC2 for computing
resources and Amazon RDS for managing the database.
- Assessment
and Strategy Planning
- Catalog
CRM components to prioritize which parts to migrate first.
- Select
appropriate migration techniques, for example, moving the CRM database
with AWS Database Migration Service (DMS).
- Plan
for a steady network connection during the move, potentially using AWS
Direct Connect.
- Execution
and Validation
- Map
out a detailed migration strategy beginning with less critical CRM
modules as a trial run.
- Secure
approval from key stakeholders before migrating the main CRM functions,
employing AWS services.
- Test
the migrated CRM's performance and security on AWS, making adjustments as
needed.
- Transition
to Cloud Operation
- Switch
to fully managing the CRM application in the AWS environment, phasing out
old on-premises components.
- Utilize
AWS's suite of monitoring and management tools for continuous oversight
and refinement.
- Apply
insights gained from this migration to inform future transitions,
considering broader cloud adoption across other applications.
This approach ensures the CRM migration to AWS is aligned
with strategic business objectives, maximizing the benefits of cloud computing
in terms of scalability, efficiency, and cost savings.
21. Describe how you would use AWS services to implement
a microservices architecture.
Implementing a microservice architecture involves breaking
down a software application into small, independent services that communicate
through APIs. Here’s a concise guide to setting up microservices:
- Adopt
Agile Development: Use agile methodologies to facilitate rapid
development and deployment of individual microservices.
- Embrace
API-First Design: Develop APIs for microservices interaction
first to ensure clear, consistent communication between services.
- Leverage
CI/CD Practices: Implement continuous integration and continuous
delivery (CI/CD) to automate testing and deployment, enhancing development
speed and reliability.
- Incorporate
Twelve-Factor App Principles: Apply these principles to create
scalable, maintainable services that are easy to deploy on cloud platforms
like AWS.
- Choose
the Right Architecture Pattern: Consider API-driven,
event-driven, or data streaming patterns based on your application’s needs
to optimize communication and data flow between services.
- Leverage
AWS for Deployment: Use AWS services such as container
technologies for scalable microservices or serverless computing to reduce
operational complexity and focus on building application logic.
- Implement
Serverless Principles: When appropriate, use serverless
architectures to eliminate infrastructure management, scale automatically,
and pay only for what you use, enhancing system efficiency and
cost-effectiveness.
- Ensure
System Resilience: Design microservices for fault tolerance and
resilience, using AWS's built-in availability features to maintain service
continuity.
- Focus
on Cross-Service Aspects: Address distributed monitoring,
logging, tracing, and data consistency to maintain system health and
performance.
- Review
with AWS Well-Architected Framework: Use the AWS Well-Architected
Tool to evaluate your architecture against AWS’s best practices, ensuring
reliability, security, efficiency, and cost-effectiveness.
By carefully considering these points, teams can effectively
implement a microservice architecture that is scalable, flexible, and suitable
for their specific application needs, all while leveraging AWS’s extensive
cloud capabilities.
22. What is the relationship between AWS Glue and AWS
Lake Formation?
AWS Lake Formation builds on AWS Glue's infrastructure,
incorporating its ETL capabilities, control console, data catalog, and
serverless architecture. While AWS Glue focuses on ETL processes, Lake
Formation adds features for building, securing, and managing data lakes,
enhancing Glue's functions.
For AWS Glue interview questions, it's important to
understand how Glue supports Lake Formation. Candidates should be ready to
discuss Glue's role in data lake management within AWS, showing their grasp of
both services' integration and functionalities in the AWS ecosystem. This
demonstrates a deep understanding of how these services collaborate to process
and manage data efficiently.
Advanced AWS Interview Questions and Answers
AWS data engineer interview questions
Addressing data
engineers, this section dives into AWS services for data handling,
including warehousing and real-time processing. It looks at the expertise
required to build scalable data pipelines with AWS.
23. Describe the difference between Amazon Redshift, RDS,
and S3, and when should each one be used?
- Amazon
S3 is an object storage service that provides scalable and
durable storage for any amount of data. It can be used to store raw,
unstructured data like log files, CSVs, images, etc.
- Amazon
Redshift is a cloud data warehouse optimized for analytics and
business intelligence. It integrates with S3 and can load data stored
there to perform complex queries and generate reports.
- Amazon
RDS provides managed relational databases like PostgreSQL, MySQL,
etc. It can power transactional applications that need ACID-compliant
databases with features like indexing, constraints, etc.
24. Describe a scenario where you would use Amazon
Kinesis over AWS Lambda for data processing. What are the key considerations?
Kinesis can be used to handle large amounts of streaming
data and allows reading and processing the streams with consumer applications.
Some of the key considerations are illustrated below:
- Data
volume: Kinesis can handle up to megabytes per second of data vs
Lambda's limit of 6MB per invocation, which is useful for high throughput
streams.
- Streaming
processing: Kinesis consumers can continuously process data in
real-time as it arrives vs Lambda's batch invocations, and this helps with
low latency processing.
- Replay
capability: Kinesis streams retain data for a configured period,
allowing replaying and reprocessing if needed, whereas Lambda not suited
for replay.
- Ordering: Kinesis
shards allow ordered processing of related records. Lambda on the other
hand may process out of order.
- Scaling
and parallelism: Kinesis shards can scale to handle load. Lambda
may need orchestraation.
- Integration: Kinesis
integrates well with other AWS services like Firehose, Redshift, EMR for
analytics.
Furthermore, for high-volume, continuous, ordered, and
replayable stream processing cases like real-time analytics, Kinesis provides
native streaming support compared to Lambda's batch approach.
To learn more about data streaming, our course Streaming Data with AWS Kinesis and Lambda helps
users learn how to leverage these technologies to ingest data from millions of
sources and analyze them in real-time. This can help better prepare for AWS
lambda interview questions.
25. What are the key differences between batch and
real-time data processing? When would you choose one approach over the other
for a data engineering project?
Batch processing involves collecting data over a period of
time and processing it in large chunks or batches. This works well for
analyzing historical, less frequent data.
Real-time streaming processing analyzes data continuously as
it arrives in small increments. It allows for analyzing fresh, frequently
updated data.
For a data engineering project, real-time streaming could be
chosen when:
- You
need immediate insights and can't wait for a batch process to run. For
example, fraud detection.
- The
data is constantly changing and analysis needs to keep up, like social
media monitoring.
- Low
latency is required, like for automated trading systems.
Batch processing may be better when:
- Historical
data needs complex modeling or analysis, like demand forecasting.
- Data
comes from various sources that only provide periodic dumps.
- Lower
processing costs are critical over processing speed.
So real-time is best for rapidly evolving data needing
continuous analysis, while batch suits periodically available data requiring
historical modeling.
26. What is an operational data store, and how does it
complement a data warehouse?
An operational data store (ODS) is a database designed to
support real-time business operations and analytics. It acts as an interim
platform between transactional systems and the data warehouse.
While a data warehouse contains high-quality data optimized
for business intelligence and reporting, an ODS contains up-to-date,
subject-oriented, integrated data from multiple sources.
Below are the key features of an ODS:
- It
provides real-time data for operations monitoring and decision-making
- Integrates
live data from multiple sources
- It
is optimized for fast queries and analytics vs long-term storage
- ODS
contains granular, atomic data vs aggregated in warehouse
An ODS and data warehouse are complementary systems. ODS
supports real-time operations using current data, while the data warehouse
enables strategic reporting and analysis leveraging integrated historical data.
When combined, they provide a comprehensive platform for both operational and
analytical needs.
AWS Scenario-based Questions
Focusing on practical application, these questions assess
problem-solving abilities in realistic scenarios, demanding a comprehensive
understanding of how to employ AWS services to tackle complex challenges.
|
Case
Type |
Description |
Solution |
|
Application
migration |
A
company plans to migrate its legacy application to AWS. The application is
data-intensive and requires low-latency access for users across the globe.
What AWS services and architecture would you recommend to ensure high
availability and low latency? |
|
|
Disaster
recovery |
Your
organization wants to implement a disaster recovery plan for its critical AWS
workloads with an RPO (Recovery Point Objective) of 5 minutes and an RTO
(Recovery Time Objective) of 1 hour. Describe the AWS services you would use
to meet these objectives. |
|
|
DDos
attacks protection |
Consider
a scenario where you need to design a scalable and secure web application
infrastructure on AWS. The application should handle sudden spikes in traffic
and protect against DDoS attacks. What AWS services and features would you
use in your design? |
|
|
Real-time
data analytics |
An IoT
startup wants to process and analyze real-time data from thousands of sensors
across the globe. The solution needs to be highly scalable and
cost-effective. Which AWS services would you use to build this platform, and
how would you ensure it scales with demand? |
|
|
Large-volume
data analysis |
A
financial services company requires a data analytics solution on AWS to
process and analyze large volumes of transaction data in real time. The
solution must also comply with stringent security and compliance standards.
How would you architect this solution using AWS, and what measures would you
put in place to ensure security and compliance? |
|
Non-Technical AWS Interview Questions
Besides technical prowess, understanding the broader impact
of AWS solutions is vital to a successful interview, and below are a few
questions, along with their answers. These answers can be different from one
candidate to another, depending on their experience and background.
27. How do you stay updated with AWS and cloud technology
trends?
- Expected
from candidate: The interviewer wants to know about your
commitment to continuous learning and how they keep your skills relevant.
They are looking for specific resources or practices they use to stay
informed.
- Example
answer: "I stay updated by reading AWS official blogs and
participating in community forums like the AWS subreddit. I also attend
local AWS user group meetups and webinars. These activities help me stay
informed about the latest AWS features and best practices."
28. Describe a time when you had to explain a complex AWS
concept to someone without a technical background. How did you go about it?
- Expected
from candidate: This question assesses your communication skills
and ability to simplify complex information. The interviewer is looking
for evidence of your teaching ability and patience.
- Example
answer: "In my previous role, I had to explain cloud storage
benefits to our non-technical stakeholders. I used the analogy of storing
files in a cloud drive versus a physical hard drive, highlighting ease of
access and security. This helped them understand the concept without
getting into the technicalities."
29. What motivates you to work in the cloud computing
industry, specifically with AWS?
- Expected
from candidate: The interviewer wants to gauge your passion for
the field and understand what drives you. They're looking for genuine
motivations that align with the role and company values.
- Example
answer: "What excites me about cloud computing, especially
AWS, is its transformative power in scaling businesses and driving
innovation. The constant evolution of AWS services motivates me to solve
new challenges and contribute to impactful projects."
30. Can you describe a challenging project you managed
and how you ensured its success?
- Expected
from candidate: Here, the focus is on your project management and
problem-solving skills. The interviewer is interested in your approach to
overcoming obstacles and driving projects to completion.
- Example
answer: "In a previous project, we faced significant delays
due to resource constraints. I prioritized tasks based on impact,
negotiated for additional resources, and kept clear communication with the
team and stakeholders. This approach helped us meet our project milestones
and ultimately deliver on time."
31. How do you handle tight deadlines when multiple
projects are demanding your attention?
- Expected
from candidate: This question tests your time management and
prioritization skills. The interviewer wants to know how you manage stress
and workload effectively.
- Example
answer: "I use a combination of prioritization and
delegation. I assess each project's urgency and impact, prioritize
accordingly, and delegate tasks when appropriate. I also communicate
regularly with stakeholders about progress and any adjustments needed to
meet deadlines."
32. What do you think sets AWS apart from other cloud
service providers?
- Expected
from candidate: The interviewer is looking for your understanding
of AWS's unique value proposition. The goal is to see that you have a good
grasp of what makes AWS a leader in the cloud industry.
- Example answer: "AWS
sets itself apart through its extensive global infrastructure, which
offers unmatched scalability and reliability. Additionally, AWS's
commitment to innovation, with a broad and deep range of services, allows
for more flexible and tailored cloud solutions compared to its
competitors."
·
https://www.geeksforgeeks.org/amazon-interview-experience-for-sde-1-amazon-wow-2020/?ref=asr13
·
live interview questions
plied through HackerEarth
coding challenge for 6 months+ experienced
Round 1(Hackerearth Round –
Coding Challenge): Two questions were asked to solved.
- Based
on the priority queue
- Based
on the map
I solved both questions
within 40 minutes.
After few weeks got a call
for scheduling interview rounds.
Round 2 (Technical
Interview): Two questions were asked, along with the behavioural
questions.
- First
question was based on the binary search
- Second
was a DP problem with binary tree.
I was able to solved both
the problems, some questions were asked from the previous experience.
Round 3(Technical Interview
– Taken by a SDE-2): Two questions were asked, along with the behavioural
questions.
- First
question was similar to the problem
https://www.geeksforgeeks.org/next-greater-element/
- Second
question was a DP problem.
The questions were followed
by questions from previous experience and behavioural questions.
Round 4(Technical Interview
– Taken by a Senior SDM): Two questions were asked
- First
question was OOPs concept based question, I was given with some features
which needed to add in the amazon alexa, and asked to implement the
generic and maintainable code, using different oops concept.
- Second
question was similar to https://www.geeksforgeeks.org/the-celebrity-problem/,
They have provided a single array of people and a function knows(a,b)
which return true if a knows b otherwise false.
Some more questions from
resume, past experiences, college projects and behavioural based were
asked
Round 5(Bar Raiser Round –
Taken by SDE-2): Two questions along with some resume based and
behavioural questions.
- First
question was based on LRU based, the question was framed in a way that you
first need to understand the question itself, and ask different questions.
- Second
question was similar to the problem https://www.geeksforgeeks.org/find-top-k-or-most-frequent-numbers-in-a-stream/.
Tips:
- Thoroughly
go through amazon leadership principals as most of the behavioural
questions were asked from that only.
- Be
vocal throughout your interview process, incase if you are going in the
wrong direction interviewer might help you.
- Ask
questions to clarify the questions that were asked
- Always
be prepared to dry run the code you have written as interviewer might not
code in same language as you do.
After giving all the
rounds, a few days later got informed that i was selected.
Are you feeling lost in OS,
DBMS, CN, SQL, and DSA chaos? Our Complete Interview Preparation Course is
your solution! Trusted by over 100,000+ Geeks, it covers all essential topics,
ensuring you're well-prepared for technical challenges. Join the ranks of
successful candidates and unlock your placement potential. Enroll now and start
your journey to a successful career!
Amazon CloudTrail:
https://www.youtube.com/watch?v=g6st17puXxE
1.
Q: You notice suspicious activity in your AWS account, like unauthorized IAM
user creation. How would you use CloudTrail to investigate?
2.
Q: You need to comply with a regulation that requires logging all API calls to
your S3 buckets. How would you implement this with CloudTrail?
3.
Q: Your CloudTrail logs are growing rapidly, leading to increased storage
costs. How can you optimize log storage while maintaining necessary
information?
Amazon
CloudWatch:
1.
Q: Your web application experiencing performance issues during peak hours. How
would you use CloudWatch to diagnose the bottleneck?
2.
Q: You need to send automated notifications when specific CloudWatch alarms are
triggered. How would you configure this?
3.
Q: You want to analyze historical CloudWatch data for trends and anomalies. How
would you achieve this?
Amazon
Lambda:
1.
Q: You have a new API endpoint that needs to process image uploads
asynchronously. How would you implement this using Lambda?
2.
Q: Your Lambda function occasionally experiences cold starts, slowing down
initial processing. How can you optimize its performance?
3.
Q: You need to monitor the performance and stability of your Lambda functions.
How would you utilize CloudWatch?
Amazon
ECS:
1.
Q: You're managing a microservices architecture in ECS. How would you handle
scaling different services based on individual resource needs?
2.
Q: You encounter high deployment failure rates when updating your ECS services.
How would you improve reliability and rollback if necessary?
3.
Q: You need to monitor the health and performance of your ECS clusters and
tasks. How would you utilize CloudWatch and other tools?
Amazon
DynamoDB:
1.
Q: You plan to store user data for your mobile app in DynamoDB. How would you
design the primary key for efficient queries?
2.
Q: You experience slow DynamoDB read speeds for certain queries. How would you
diagnose and improve performance?
3.
Q: You need to backup and restore your DynamoDB data regularly. How would you
achieve this?
Circle
CI:
1.
Q: Your team develops a microservices architecture on AWS. How would you
integrate Circle CI with your development workflow to automate deployments and
tests?
2.
Q: You encounter intermittent build failures in Circle CI that make deployment
unreliable. How would you troubleshoot the issue?
3.
Q: You need to track and visualize your team's build history and performance in
Circle CI. How would you achieve this?
Elastic
Block Store (EBS):
1.
Q: You need to choose the right EBS volume type for your production database
instance. What factors would you consider and which type would you recommend?
2.
Q: You encounter frequent EBS volume snapshots exceeding your billing budget.
How can you optimize snapshot costs?
3.
Q: Your application experiences slow disk performance during peak hours. How
can you diagnose and improve EBS performance?
Elastic
Kubernetes Service (EKS):
1.
Q: You're planning to migrate containerized applications to EKS. What
considerations should you have for security and access control?
2.
Q: You encounter pod deployment failures in your EKS cluster. How would you
approach troubleshooting and identifying the root cause?
3.
Q: You want to scale your EKS cluster based on application load. How would you
implement autoscaling?
AWS VPC Interview Questions
With Sample Answers
Below are some AWS VPC interview
questions with example answers that may inspire you as you prepare your own:
1. What is a VPC?
At the start of your interview, a
hiring manager may focus on more general questions about your past experiences
and skills. The hiring manager may ask this question to assess your ability to
provide simple explanations for complex ideas. In answer to this question,
focus on providing the definition and on highlighting your knowledge of the
subject. You can prepare your answer by considering how you might explain the
term to someone with no prior experience.Example: The virtual
private cloud, or VPC, is an online space within Amazon's cloud. In this space,
you can use AWS resources.Related: 10 AWS Interview Questions And Answers
2. What are subnets when
working with a VPC?
This is another simple question
that a hiring manager may ask to assess your AWS VPC knowledge. This question
helps them learn more about your understanding of the service and your ability
to provide clear explanations. To answer this question, focus on your knowledge
of subnets. Try and remain clear and concise.Example: A subnet
is a segment of the unique internet protocol (IP) address range for the VPC.
This is where you can locate groupings of resources.
3. What is a NAT device?
A network address translation
(NAT) device is a component that allows you to conserve IP addresses and add a
layer of protection. A hiring manager may ask you about this to assess your
ability to protect and maintain the server. In your answer, define NAT devices
and explain why they are important.Example: NAT devices function
by blocking inbound traffic coming from the Internet and enabling outbound
traffic to other services. When traffic leaves the VPC, the NAT device's
address replaces the IP address. NAT devices are important because they provide
additional security and prevent the depletion of IP addresses.Related: What Is Network Address Translation? (Types And FAQs)
4. What is the difference
between stateful and stateless filtering?
Filtering is important for the
VPC, as it allows you to assess and restrict certain types of traffic. A hiring
manager may ask you about these two types of filtering to determine what you
know about AWS VPC security. To answer this question, define both types of
filtering and explain their differences.Example: Stateful
filtering is a process that assesses the origin of the request. It triggers a
replay to the originating device. Stateless filtering assesses the destination
IP and source and does not examine the request or the replay. The primary
difference between these filtering types is that security groups use stateful
filtering while network access control lists (ACLs) use stateless filtering.Related: 8 Popular AWS Certifications To Consider Pursuing
5. What are the advantages of
using a default VPC?
A default VPC is a network that
the system creates automatically when a user utilises resources. It is an
individual network in the cloud, and a hiring manager may ask you about it
because you may use it frequently as a cloud architect or data professional. In
answer to the question, you can provide a brief definition of a default VPC and
then detail its advantages.Example: When you use Amazon's
Elastic Compute Cloud (EC2) resources for the first time, the system
automatically creates a default VPC in the AWS cloud account. The default VPC
is useful as it is where you can access high-level features. In the default
VPC, you can use different IPs and access network interfaces without creating a
new VPC.Related: AWS Cloud Engineer Interview Questions And Sample Answers
6. What are the internet
gateways in VPC?
Internet gateways are important
for the communication between the VPC and the Internet. A hiring manager may
ask you about internet gateways to assess your knowledge and evaluate your
ability to regulate traffic within the VPC. To answer this question, briefly
define internet gateways and describe their functions.Example: Internet
gateways are components within the system that allow the VPC to connect to the
Internet. They are horizontally scaled, and only one gateway can connect to
each VPC. NAT devices can translate IP addresses when the system uses internet
gateways.
7. What is a network ACL in a
VPC?
The network ACL provides rules
regarding which users and systems can gain access to the VPC. Network ACLs are
essential to maintaining security and order. A hiring manager may ask you about
network ACLs to better understand your VPC knowledge and assess your ability to
maintain the system.Example: ACL stands for the Access Control
List and the network ACL functions as a list of authorised entities that can
gain access to the system. The network ACL regulates inbound and outbound
traffic, and the VPC generates the ACL automatically. You can modify the
automatically generated ACL to include specific rules.
8. What are the security
groups in a VPC?
Organisations often store
important information in VPCs, so it is important that the VPC's security is
functional. If you are a professional who works with VPCs, a hiring manager may
ask this question to ensure that you know how to protect private data. To
answer this question, define a security group and its location.Example: Security
groups are components that manage traffic within the VPC. They can act as
firewalls by using specific rules to control the traffic. There are security
groups in the VPC and the EC2 sections of the AWS console.Related: What Is An AWS Solutions Architect? A Complete Guide
9. What is an elastic IP
address?
Questions about IP addresses are
very common in AWS VPC interviews. IP addresses are specific strings of
characters that identify devices that communicate over the Internet. A hiring
manager may ask you this question to determine whether you know the differences
between the various types of IP addresses. To answer this question, define an
elastic IP address and explain how it works.Example: An elastic
IP address is much like a public IP address, except that the elastic address
connects to the AWS account until a user terminates it. You can detach and
reattach elastic addresses to different instances within the VPC.
10. How does ELB affect a VPC?
Elastic load balancing (ELB) is a
service that balances loads for AWS deployments. This process allows the
computer to accomplish tasks faster. A hiring manager may ask you this question
to determine your knowledge of security and scalability for VPCs. To answer
this question, you can list the different types of ELB and explain how they
work within the VPC.Example: Elastic load balancing, or ELB, is
a process that distributes the incoming work to help the computer complete it
faster. The three types of ELB are network, application and classic. You can
use application and network ELB forms to route the traffic to specific targets
within your VPC.
11. What are the limits for
VPCs, gateways and subnets?
When using AWS VPC, you can
create components, such as subnets and gateways. A hiring manager may ask about
the system's limits to assess your level of experience using it. This question
can also assess your knowledge about connectivity within the VPC. To answer,
you can provide the specific limit for each component in the question.Example: Limits
within the AWS VPC system exist to maintain the system's operating quality.
Limits vary depending on the component. You can create five VPCs, five elastic
IP addresses and 50 customer gateways per region and 200 subnets per VPC.Related: What Is An AWS Certification? (With Types And Benefits)
12. What is a public IP
address?
A hiring manager may ask this
question to see how much you know about IP addresses. To answer, you can define
a public IP address. It is also helpful to compare it to the other types of
addresses.Example: A public IP address is an easily accessible
internet identifier. When you first launch a VPC, the system automatically
creates and assigns one. Each time you stop and restart instances, you
terminate them. A public address differs from a private address because when
you stop an instance, you terminate the address.
No comments:
Post a Comment