Trusted Advisor inspects the AWS environment to make recommendations for system performance, saving money, availability, and closing security gaps
Trusted Advisor checks the following categories
Cost Optimization
Recommendations that can potentially save money by highlighting unused resources and opportunities to reduce your bill.
Security
Identification of security settings and gaps, inline with the best practices, that could make the AWS solution less secure
Fault Tolerance
Recommendations that help increase the resiliency and availability of the AWS solution by highlighting redundancy shortfalls, current service limits, and over-utilized resources.
Performance
Recommendations that can help to improve the speed and responsiveness of the applications
AWS Support API provides programmatic access to AWS Support Center features to create, manage, and close the Support cases, and operationally manage the Trusted Advisor check requests and status.
Trusted Advisor Priority helps you focus on the most important recommendations to optimize your cloud deployments, improve resilience, and address security gaps.
Trusted Advisor notification feature helps stay up-to-date with the AWS resource deployment by performing an automated weekly refresh.
AWS Support API
API provides two different groups of operations:
Support case management operations to manage the entire life cycle of your AWS support cases, from creating a case to resolving it, and includes
Open a support case
Get a list and detailed information about recent support cases
Filter your search for support cases by dates and case identifiers, including resolved cases
Add communications and file attachments to your cases, and add the email recipients for case correspondence
Resolve your cases
AWS Trusted Advisor operations to access checks
Get the names and identifiers for the checks
Request that a check be run against your AWS account and resources
Get summaries and detailed information for your check results
Refresh the checks
Get the status of each check
AWS Certification Exam Practice Questions
Questions are collected from Internet and the answers are marked as per my knowledge and understanding (which might differ with yours).
AWS services are updated everyday and both the answers and questions might be outdated soon, so research accordingly.
AWS exam questions are not updated to keep up the pace with AWS updates, so even if the underlying feature has changed the question might not be updated
Open to further feedback, discussion and correction.
The Trusted Advisor service provides insight regarding which categories of an AWS account?
Security, fault tolerance, high availability, and connectivity
Security, access control, high availability, and performance
Performance, cost optimization, security, and fault tolerance (Note – Service limits is the latest addition)
Performance, cost optimization, access control, and connectivity
Which of the following are categories of AWS Trusted Advisor? (Select TWO.)
Loose Coupling
Disaster recovery
Infrastructure as a Code
Security
Service limits
Which AWS tool will identify security groups that grant unrestricted Internet access to a limited list of ports?
AWS Direct Connect is a network service that provides an alternative to using the Internet to utilize AWS cloud services
DX links your internal network to an AWS Direct Connect location over a standard Ethernet fiber-optic cable with one end of the cable connected to your router, the other to an AWS Direct Connect router.
Connections can be established with
Dedicated connections – 1Gbps, 10Gbps, and 100Gbps capacity.
Hosted connection – Speeds of 50, 100, 200, 300, 400, and 500 Mbps can be ordered from any APN partners supporting AWS DX. Also, supports 1, 2, 5 & 10 Gbps with selected partners.
Virtual interfaces can be created directly to public AWS services ( e.g. S3) or to VPC, bypassing internet service providers in the network path.
DX locations in public Regions or AWS GovCloud (US) can access public services in any other public Region.
Each AWS DX location enables connectivity to all AZs within the geographically nearest AWS region.
DX supports both the IPv4 and IPv6 communication protocols.
Direct Connect Advantages
Reduced Bandwidth Costs
All data transferred over the dedicated connection is charged at the reduced data transfer rate rather than Internet data transfer rates.
Transferring data to and from AWS directly reduces the bandwidth commitment to the Internet service provider
Consistent Network Performance
provides a dedicated connection and a more consistent network performance experience than the Internet which can widely vary.
AWS Services Compatibility
is a network service and works with all of the AWS services like S3, EC2, and VPC
Private Connectivity to AWS VPC
Using DX Private Virtual Interface a private, dedicated, high bandwidth network connection can be established between the network and VPC
Elastic
can be easily scaled to meet the needs by either using a higher bandwidth connection or by establishing multiple connections.
Direct Connect Anatomy
Amazon maintains AWS Direct Connect PoP across different locations (referred to as Colocation Facilities) which are different from AWS regions.
As a consumer, you can either purchase a rack space or use any of the AWS APN Partners which already have the infrastructure within the Colocation Facility and configure a Customer Gateway
Connection from the AWS Direct Connect PoP to the AWS regions is maintained by AWS itself.
Connection from the Customer Gateway to the Customer Data Center can be established using any Service Provider Network.
Connection between the PoP and the Customer gateway within the Colocation Facility is called Cross Connect.
Once a DX connection is created with AWS, an LOA-CFA (Letter Of Authority – Connecting Facility Assignment) would be received.
LOA-CFA can be handover to the Colocation Facility or the APN Partner to establish the Cross Connect
Once the Cross Connect and the connectivity between the CGW and Customer DataCenter are established, Virtual Interfaces can be created
AWS Direct Connect requires a VGW to access the AWS VPC.
Virtual Interfaces – VIF
Each connection requires a Virtual Interface
Each connection can be configured with one or more virtual interfaces.
Supports, Public, Private, and Transit Virtual Interface
Each VIF needs a VLAN ID, interface IP address, ASN, and BGP key.
To use the connection with another AWS account, a hosted virtual interface (Hosted VIF) can be created for that account. These hosted virtual interfaces work the same as standard virtual interfaces and can connect to public resources or a VPC.
Direct Connect Network Requirements
Single-mode fiber with
a 1000BASE-LX (1310 nm) transceiver for 1 gigabit Ethernet,
a 10GBASE-LR (1310 nm) transceiver for 10 gigabits, or
a 100GBASE-LR4 for 100 gigabit Ethernet.
802.1Q VLAN encapsulation must be supported
Auto-negotiation for a port must be disabled so that the speed and mode (half or full duplex) cannot be modified and should be manually configured
Border Gateway Protocol (BGP) and BGP MD5 authentication must be supported
Bidirectional Forwarding Detection (BFD) is optional and helps in quick failure detection.
Direct Connect Connections
Dedicated Connection
provides a physical Ethernet connection associated with a single customer
Customers can request a dedicated connection through the AWS Direct Connect console, the CLI, or the API.
support port speeds of 1 Gbps, 10 Gbps, and 100 Gbps.
supports multiple virtual interfaces (current limit of 50)
Hosted Connection
A physical Ethernet connection that an AWS Direct Connect Partner provisions on behalf of a customer.
Customers request a hosted connection by contacting a partner in the AWS Direct Connect Partner Program, which provisions the connection
Support port speeds of 50 Mbps, 100 Mbps, 200 Mbps, 300 Mbps, 400 Mbps, 500 Mbps, 1 Gbps, 2 Gbps, 5 Gbps, and 10 Gbps
1 Gbps, 2 Gbps, 5 Gbps or 10 Gbps hosted connections are supported by limited partners.
supports a single virtual interface
AWS uses traffic policing on hosted connections and excess traffic is dropped.
Direct Connect Virtual Interfaces – VIF
Public Virtual Interface
enables connectivity to all the AWS Public IP addresses
helps connect to public resources e.g. SQS, S3, EC2, Glacier, etc which are reachable publicly only.
can be used to access all public resources across regions
allows a maximum of 1000 prefixes. You can summarize the prefixes into a larger range to reduce the number of prefixes.
does not support Jumbo frames.
Private Virtual Interface
helps connect to the VPC for e.g. instances with a private IP address
supports
Virtual Private Gateway
Allows connections only to a single specific VPC with the attached VGW in the same region
Private VIF and Virtual Private Gateway – VGW should be in the same region
helps access one or more VPC Transit Gateways associated with Direct Connect Gateways.
supports Jumbo frames with 8500 MTU
Direct Connect Redundancy
Direct Connect connections do not provide redundancy and have multiple single points of failures w.r.t to the hardware devices as each connection consists of a single dedicated connection between ports on your router and an Amazon router.
Redundancy can be provided by
Establishing a second DX connection, preferably in a different Colocation Facility using a different router and AWS DX PoP.
IPsec VPN connection between the Customer DC to the VGW.
For Multiple ports requested in the same AWS Direct Connect location, Amazon itself makes sure they are provisioned on redundant Amazon routers to prevent impact from a hardware failure
High Resiliency – 99.9%
High resiliency for critical workloads can be achieved by using two single connections to multiple locations.
It provides resiliency against connectivity failures caused by a fiber cut or a device failure. It also helps prevent a complete location failure.
Maximum Resiliency – 99.99%
Maximum resiliency for critical workloads can be achieved using separate connections that terminate on separate devices in more than one location.
It provides resiliency against device, connectivity, and complete location failures.
Direct Connect LAG – Link Aggregation Group
A LAG is a logical interface that uses the Link Aggregation Control Protocol (LACP) to aggregate multiple connections at a single AWS Direct Connect endpoint, treating them as a single, managed connection.
LAG can combine multiple connections to increase available bandwidth.
LAG can be created from existing or new connections.
Existing connections (whether standalone or part of another LAG) with the LAG can be associated after LAG creation.
LAG needs following rules
All connections must use the same bandwidth and port speed of 1, 10, 100 Gbps.
All connections must be dedicated connections.
Maximum of four connections in a LAG. Each connection in the LAG counts toward the overall connection limit for the Region.
All connections in the LAG must terminate at the same AWS Direct Connect endpoint.
Multi-chassis LAG (MLAG) is not supported by AWS.
LAG doesn’t make the connectivity to AWS more resilient.
LAG connections operate in Active/Active mode.
LAG supports attributes to define a minimum number of operational connections for the LAG function, with a default value of 0.
Direct Connect Failover
Bidirectional Forwarding Detection – BFD is a detection protocol that provides fast forwarding path failure detection times. These fast failure detection times facilitate faster routing reconvergence times.
When connecting to AWS services over DX connections it is recommended to enable BFD for fast failure detection and failover.
By default, BGP waits for three keep-alives to fail at a hold-down time of 90 seconds. Enabling BFD for the DX connection allows the BGP neighbor relationship to be quickly torn down.
Asynchronous BFD is automatically enabled for each DX virtual interface, but will not take effect until it’s configured on your router.
AWS has set the BFD liveness detection minimum interval to 300, and the BFD liveness detection multiplier to 3
It’s a best practice not to configure graceful restart and BFD at the same time to avoid failover or connection issues. For fast failover, configure BFD without graceful restart enabled.
BFD is supported for LAGs.
Direct Connect Security
Direct Connect does not encrypt the traffic that is in transit by default. To encrypt the data in transit that traverses DX, you must use the transit encryption options for that service.
DX connections can be secured
with IPSec VPN to provide secure, reliable connectivity.
with MACsec to encrypt the data from the corporate data center to the DX location.
MAC Security (MACsec)
is an IEEE standard that provides data confidentiality, data integrity, and data origin authenticity.
provides Layer2 security for 10Gbps and 100Gbps Dedicated Connections only.
delivers native, near line-rate, point-to-point encryption ensuring that data communications between AWS and the data center, office, or colocation facility remain protected.
removes VPN limitation that required the aggregation of multiple IPsec VPN tunnels to work around the throughput limits of using a single VPN connection.
Questions are collected from Internet and the answers are marked as per my knowledge and understanding (which might differ with yours).
AWS services are updated everyday and both the answers and questions might be outdated soon, so research accordingly.
AWS exam questions are not updated to keep up the pace with AWS updates, so even if the underlying feature has changed the question might not be updated
Open to further feedback, discussion and correction.
You are building a solution for a customer to extend their on-premises data center to AWS. The customer requires a 50-Mbps dedicated and private connection to their VPC. Which AWS product or feature satisfies this requirement?
Amazon VPC peering
Elastic IP Addresses
AWS Direct Connect
Amazon VPC virtual private gateway
Is there any way to own a direct connection to Amazon Web Services?
You can create an encrypted tunnel to VPC, but you don’t own the connection.
Yes, it’s called Amazon Dedicated Connection.
No, AWS only allows access from the public Internet.
Yes, it’s called Direct Connect
An organization has established an Internet-based VPN connection between their on-premises data center and AWS. They are considering migrating from VPN to AWS Direct Connect. Which operational concern should drive an organization to consider switching from an Internet-based VPN connection to AWS Direct Connect?
AWS Direct Connect provides greater redundancy than an Internet-based VPN connection.
AWS Direct Connect provides greater resiliency than an Internet-based VPN connection.
AWS Direct Connect provides greater bandwidth than an Internet-based VPN connection.
AWS Direct Connect provides greater control of network provider selection than an Internet-based VPN connection.
Does AWS Direct Connect allow you access to all Availabilities Zones within a Region?
Depends on the type of connection
No
Yes
Only when there’s just one availability zone in a region. If there are more than one, only one availability zone can be accessed directly.
A customer has established an AWS Direct Connect connection to AWS. The link is up and routes are being advertised from the customer’s end, however, the customer is unable to connect from EC2 instances inside its VPC to servers residing in its datacenter. Which of the following options provide a viable solution to remedy this situation? (Choose 2 answers)
Add a route to the route table with an IPSec VPN connection as the target (deals with VPN)
Enable route propagation to the Virtual Private Gateway (VGW)
Enable route propagation to the customer gateway (CGW) (route propagation is enabled on VGW)
Modify the route table of all Instances using the ‘route’ command. (no route command available)
Modify the Instances VPC subnet route table by adding a route back to the customer’s on-premises environment.
A company has configured and peered two VPCs: VPC-1 and VPC-2. VPC-1 contains only private subnets, and VPC-2 contains only public subnets. The company uses a single AWS Direct Connect connection and private virtual interface to connect their on-premises network with VPC-1. Which two methods increase the fault tolerance of the connection to VPC-1? Choose 2 answers
Establish a hardware VPN over the internet between VPC-2 and the on-premises network. (Peered VPC does not support Edge to Edge Routing)
Establish a hardware VPN over the internet between VPC-1 and the on-premises network
Establish a new AWS Direct Connect connection and private virtual interface in the same region as VPC-2 (Peered VPC does not support Edge to Edge Routing)
Establish a new AWS Direct Connect connection and private virtual interface in a different AWS region than VPC-1 (need to be in the same region as VPC-1)
Establish a new AWS Direct Connect connection and private virtual interface in the same AWS region as VPC-1
Your company previously configured a heavily used, dynamically routed VPN connection between your on-premises data center and AWS. You recently provisioned a Direct Connect connection and would like to start using the new connection. After configuring Direct Connect settings in the AWS Console, which of the following options will provide the most seamless transition for your users?
Delete your existing VPN connection to avoid routing loops configure your Direct Connect router with the appropriate settings and verify network traffic is leveraging Direct Connect.
Configure your Direct Connect router with a higher BGP priority than your VPN router, verify network traffic is leveraging Direct Connect, and then delete your existing VPN connection.
Update your VPC route tables to point to the Direct Connect connection configure your Direct Connect router with the appropriate settings verify network traffic is leveraging Direct Connect and then delete the VPN connection.
Configure your Direct Connect router, update your VPC route tables to point to the Direct Connect connection, configure your VPN connection with a higher BGP priority. And verify network traffic is leveraging the Direct Connect connection
You are designing the network infrastructure for an application server in Amazon VPC. Users will access all the application instances from the Internet as well as from an on-premises network The on-premises network is connected to your VPC over an AWS Direct Connect link. How would you design routing to meet the above requirements?
Configure a single routing table with a default route via the Internet gateway. Propagate a default route via BGP on the AWS Direct Connect customer router. Associate the routing table with all VPC subnets (propagating the default route would cause conflict)
Configure a single routing table with a default route via the internet gateway. Propagate specific routes for the on-premises networks via BGP on the AWS Direct Connect customer router. Associate the routing table with all VPC subnets.
Configure a single routing table with two default routes: one to the internet via an Internet gateway the other to the on-premises network via the VPN gateway use this routing table across all subnets in your VPC. (there cannot be 2 default routes)
Configure two routing tables one that has a default route via the Internet gateway and another that has a default route via the VPN gateway Associate both routing tables with each VPC subnet. (as the instances have to be in the public subnet and should have a single routing table associated with them)
You are implementing AWS Direct Connect. You intend to use AWS public service endpoints such as Amazon S3, across the AWS Direct Connect link. You want other Internet traffic to use your existing link to an Internet Service Provider. What is the correct way to configure AWS Direct Connect for access to services such as Amazon S3?
Configure a public Interface on your AWS Direct Connect link. Configure a static route via your AWS Direct Connect link that points to Amazon S3. Advertise a default route to AWS using BGP.
Create a private interface on your AWS Direct Connect link. Configure a static route via your AWS Direct Connect link that points to Amazon S3 Configure specific routes to your network in your VPC.
Create a public interface on your AWS Direct Connect link. Redistribute BGP routes into your existing routing infrastructure advertise specific routes for your network to AWS
Create a private interface on your AWS Direct connect link. Redistribute BGP routes into your existing routing infrastructure and advertise a default route to AWS.
You have been asked to design network connectivity between your existing data centers and AWS. Your application’s EC2 instances must be able to connect to existing backend resources located in your data center. Network traffic between AWS and your data centers will start small, but ramp up to 10s of GB per second over the course of several months. The success of your application is dependent upon getting to market quickly. Which of the following design options will allow you to meet your objectives?
Quickly create an internal ELB for your backend applications, submit a DirectConnect request to provision a 1 Gbps cross-connect between your data center and VPC, then increase the number or size of your DirectConnect connections as needed.
Allocate EIPs and an Internet Gateway for your VPC instances to use for quick, temporary access to your backend applications, then provision a VPN connection between a VPC and existing on-premises equipment.
Provision a VPN connection between a VPC and existing on-premises equipment, submit a DirectConnect partner request to provision cross connects between your data center and the DirectConnect location, then cut over from the VPN connection to one or more DirectConnect connections as needed.
Quickly submit a DirectConnect request to provision a 1 Gbps cross connect between your data center and VPC, then increase the number or size of your DirectConnect connections as needed.
You are tasked with moving a legacy application from a virtual machine running inside your datacenter to an Amazon VPC. Unfortunately, this app requires access to a number of on-premises services and no one who configured the app still works for your company. Even worse there’s no documentation for it. What will allow the application running inside the VPC to reach back and access its internal dependencies without being reconfigured? (Choose 3 answers)
An AWS Direct Connect link between the VPC and the network housing the internal services (VPN or a DX for communication)
An Internet Gateway to allow a VPN connection. (Virtual and Customer gateway is needed)
An Elastic IP address on the VPC instance (Don’t need a EIP as private subnets can also interact with on-premises network)
An IP address space that does not conflict with the one on-premises (IP address cannot conflict)
Entries in Amazon Route 53 that allow the Instance to resolve its dependencies’ IP addresses (Route 53 is not required)
A VM Import of the current virtual machine (VM Import to copy the VM to AWS as there is no documentation it can’t be configured from scratch)
S3 Glacier is a storage service optimized for archival, infrequently used data, or “cold data.”
S3 Glacier is an extremely secure, durable, and low-cost storage service for data archiving and long-term backup.
provides average annual durability of 99.999999999% (11 9’s) for an archive.
redundantly stores data in multiple facilities and on multiple devices within each facility.
synchronously stores the data across multiple facilities before returning SUCCESS on uploading archives, to enhance durability.
performs regular, systematic data integrity checks and is built to be automatically self-healing.
enables customers to offload the administrative burdens of operating and scaling storage to AWS, without having to worry about capacity planning, hardware provisioning, data replication, hardware failure detection, recovery, or time-consuming hardware migrations.
offers a range of storage classes and patterns
S3 Glacier Instant Retrieval
Use for archiving data that is rarely accessed and requires milliseconds retrieval.
S3 Glacier Flexible Retrieval (formerly the S3 Glacier storage class)
Use for archives where portions of the data might need to be retrieved in minutes.
offers a range of data retrievals options where the retrieval time varies from minutes to hours.
Expedited retrieval: 1-5 mins
Standard retrieval: 3-5 hours
Bulk retrieval: 5-12 hours
S3 Glacier Deep Archive
Use for archiving data that rarely need to be accessed.
Data stored has a default retrieval time of 12 hours.
S3 Glacier Flexible Retrieval and S3 Glacier Deep Archive objects are not available for real-time access.
is a great storage choice when low storage cost is paramount, with data rarely retrieved, and retrieval latency is acceptable. S3 should be used if applications require fast, frequent real-time access to the data.
can store virtually any kind of data in any format.
allows interaction through AWS Management Console, Command Line Interface CLI, and SDKs or REST-based APIs.
AWS Management console can only be used to create and delete vaults.
Rest of the operations to upload, download data, and create jobs for retrieval need CLI, SDK, or REST-based APIs.
Use cases include
Digital media archives
Data that must be retained for regulatory compliance
Financial and healthcare records
Raw genomic sequence data
Long-term database backups
S3 Glacier Storage Classes
S3 Glacier Instant Retrieval
Use for archiving data that is rarely accessed and requires milliseconds retrieval.
Use for archives where portions of the data might need to be retrieved in minutes.
Data has a minimum storage duration period of 90 days and can be accessed in as little as 1-5 minutes by using an expedited retrieval
You can also request free Bulk retrievals in up to 5-12 hours.
S3 supports restore requests at a rate of up to 1,000 transactions per second, per AWS account.
S3 Glacier Deep Archive
Use for archiving data that rarely needs to be accessed.
S3 Glacier Deep Archive is the lowest cost storage option in AWS.
Retrieval costs can be reduced further using bulk retrieval, which returns data within 48 hours.
Data stored has a minimum storage duration period of 180 days
Data stored has a default retrieval time of 12 hours.
S3 supports restore requests at a rate of up to 1,000 transactions per second, per AWS account.
S3 Glacier Flexible Data Retrievals Options
Glacier provides three options for retrieving data with varying access times and costs: Expedited, Standard, and Bulk retrievals.
Expedited Retrievals
Expedited retrievals allow quick access to the data when occasional urgent requests for a subset of archives are required.
Data has a minimum storage duration period of 90 days
Data accessed are typically made available within 1-5 minutes.
There are two types of Expedited retrievals: On-Demand and Provisioned.
On-Demand requests are like EC2 On-Demand instances and are available the vast majority of the time.
Provisioned requests are guaranteed to be available when needed.
Standard Retrievals
Standard retrievals allow access to any of the archives within several hours.
Standard retrievals typically complete within 3-5 hours.
Bulk Retrievals
Bulk retrievals are Glacier’s lowest-cost retrieval option, enabling retrieval of large amounts, even petabytes, of data inexpensively in a day.
Bulk retrievals typically complete within 5-12 hours.
S3 Glacier Data Model
Glacier data model core concepts include vaults and archives and also include job and notification configuration resources
Vault
A vault is a container for storing archives.
Each vault resource has a unique address, which comprises the region the vault was created and the unique vault name within the region and account for e.g. https://glacier.us-west-2.amazonaws.com/111122223333/vaults/examplevault
Vault allows the storage of an unlimited number of archives.
Glacier supports various vault operations which are region-specific.
An AWS account can create up to 1,000 vaults per region.
Archive
An archive can be any data such as a photo, video, or document and is a base unit of storage in Glacier.
Each archive has a unique ID and an optional description, which can only be specified during the upload of an archive.
Glacier assigns the archive an ID, which is unique in the AWS region in which it is stored.
An archive can be uploaded in a single request. While for large archives, Glacier provides a multipart upload API that enables uploading an archive in parts.
An Archive can be up to 40TB.
Jobs
A Job is required to retrieve an Archive and vault inventory list
Data retrieval requests are asynchronous operations, are queued and some jobs can take about four hours to complete.
A job is first initiated and then the output of the job is downloaded after the job is completed.
Vault inventory jobs need the vault name.
Data retrieval jobs need both the vault name and the archive id, with an optional description
A vault can have multiple jobs in progress at any point in time and can be identified by Job ID, assigned when is it created for tracking
Glacier maintains job information such as job type, description, creation date, completion date, and job status and can be queried
After the job completes, the job output can be downloaded in full or partially by specifying a byte range.
Notification Configuration
As the jobs are asynchronous, Glacier supports a notification mechanism to an SNS topic when the job completes
SNS topic for notification can either be specified with each individual job request or with the vault
Glacier stores the notification configuration as a JSON document
Glacier Supported Operations
Vault Operations
Glacier provides operations to create and delete vaults.
A vault can be deleted only if there are no archives in the vault as of the last computed inventory and there have been no writes to the vault since the last inventory (as the inventory is prepared periodically)
Vault Inventory
Vault inventory helps retrieve a list of archives in a vault with information such as archive ID, creation date, and size for each archive
Inventory for each vault is prepared periodically, every 24 hours
Vault inventory is updated approximately once a day, starting on the day the first archive is uploaded to the vault.
When a vault inventory job is, Glacier returns the last inventory it generated, which is a point-in-time snapshot and not real-time data.
Vault Metadata or Description can also be obtained for a specific vault or for all vaults in a region, which provides information such as
creation date,
number of archives in the vault,
total size in bytes used by all the archives in the vault,
and the date the vault inventory was generated
S3 Glacier also provides operations to set, retrieve, and delete a notification configuration on the vault. Notifications can be used to identify vault events.
Archive Operations
S3 Glacier provides operations to upload, download and delete archives.
All archive operations must either be done using AWS CLI or SDK. It cannot be done using AWS Management Console.
An existing archive cannot be updated, it has to be deleted and uploaded.
Archive Upload
An archive can be uploaded in a single operation (1 byte to up to 4 GB in size) or in parts referred to as Multipart upload (40 TB)
Multipart Upload helps to
improve the upload experience for larger archives.
upload archives in parts, independently, parallelly and in any order
faster recovery by needing to upload only the part that failed upload and not the entire archive.
upload archives without even knowing the size
upload archives from 1 byte to about 40,000 GB (10,000 parts * 4 GB) in size
To upload existing data to Glacier, consider using the AWS Import/Export Snowball service, which accelerates moving large amounts of data into and out of AWS using portable storage devices for transport. AWS transfers the data directly onto and off of storage devices using Amazon’s high-speed internal network, bypassing the Internet.
Glacier returns a response that includes an archive ID that is unique in the region in which the archive is stored.
Glacier does not support any additional metadata information apart from an optional description. Any additional metadata information required should be maintained on the client side.
Archive Download
Downloading an archive is an asynchronous operation and is the 2 step process
Initiate an archive retrieval job
When a Job is initiated, a job ID is returned as a part of the response.
Job is executed asynchronously and the output can be downloaded after the job completes.
A job can be initiated to download the entire archive or a portion of the archive.
After the job completes, download the bytes
An archive can be downloaded as all the bytes or a specific byte range to download only a portion of the output
Downloading the archive in chunks helps in the event of a download failure, as only that part needs to be downloaded
Job completion status can be checked by
Check status explicitly (Not Recommended)
periodically poll the describe job operation request to obtain job information
Completion notification
An SNS topic can be specified, when the job is initiated or with the vault, to be used to notify job completion
About Range Retrievals
S3 Glacier allows retrieving an archive either in whole (default) or a range, or a portion.
Range retrievals need a range to be provided that is megabyte aligned.
Glacier returns a checksum in the response which can be used to verify if any errors in the download by comparing it with the checksum computed on the client side.
Specifying a range of bytes can be helpful when:
Control bandwidth costs
Glacier allows retrieval of up to 5 percent of the average monthly storage (pro-rated daily) for free each month
Scheduling range retrievals can help in two ways.
meet the monthly free allowance of 5 percent by spreading out the data requested
if the amount of data retrieved doesn’t meet the free allowance percentage, scheduling range retrievals enable a reduction of the peak retrieval rate, which determines the retrieval fees.
Manage your data downloads
Glacier allows retrieved data to be downloaded for 24 hours after the retrieval request completes
Only portions of the archive can be retrieved so that the schedule of downloads can be managed within the given download window.
Retrieve a targeted part of a large archive
Retrieving an archive in a range can be useful if an archive is uploaded as an aggregate of multiple individual files, and only a few files need to be retrieved
Archive Deletion
An archive can be deleted from the vault only one at a time
This operation is idempotent. Deleting an already-deleted archive does not result in an error
AWS applies a pro-rated charge for items that are deleted prior to 90 days, as it is meant for long-term storage
Archive Update
An existing archive cannot be updated and must be deleted and re-uploaded, which would be assigned a new archive id
S3 Glacier Vault Lock
S3 Glacier Vault Lock helps deploy and enforce compliance controls for individual S3 Glacier vaults with a vault lock policy.
Specify controls such as “write once read many” (WORM) can be enforced using a vault lock policy and the policy can be locked for future edits.
Once locked, the policy can no longer be changed.
S3 Glacier Security
S3 Glacier supports data in transit encryption using Secure Sockets Layer (SSL) or client-side encryption.
All data is encrypted on the server side with Glacier handling key management and key protection. It uses AES-256, one of the strongest block ciphers available
Security and compliance of S3 Glacier are assessed by third-party auditors as part of multiple AWS compliance programs including SOC, HIPAA, PCI DSS, FedRAMP, etc.
AWS Certification Exam Practice Questions
Questions are collected from Internet and the answers are marked as per my knowledge and understanding (which might differ with yours).
AWS services are updated everyday and both the answers and questions might be outdated soon, so research accordingly.
AWS exam questions are not updated to keep up the pace with AWS updates, so even if the underlying feature has changed the question might not be updated
Open to further feedback, discussion and correction.
What is Amazon Glacier?
You mean Amazon “Iceberg”: it’s a low-cost storage service.
A security tool that allows to “freeze” an EBS volume and perform computer forensics on it.
A low-cost storage service that provides secure and durable storage for data archiving and backup
It’s a security tool that allows to “freeze” an EC2 instance and perform computer forensics on it.
Amazon Glacier is designed for: (Choose 2 answers)
Active database storage
Infrequently accessed data
Data archives
Frequently accessed data
Cached session data
An organization is generating digital policy files which are required by the admins for verification. Once the files are verified they may not be required in the future unless there is some compliance issue. If the organization wants to save them in a cost effective way, which is the best possible solution?
AWS RRS
AWS S3
AWS RDS
AWS Glacier
A user has moved an object to Glacier using the life cycle rules. The user requests to restore the archive after 6 months. When the restore request is completed the user accesses that archive. Which of the below mentioned statements is not true in this condition?
The archive will be available as an object for the duration specified by the user during the restoration request
The restored object’s storage class will be RRS (After the object is restored the storage class still remains GLACIER. Read more)
The user can modify the restoration period only by issuing a new restore request with the updated period
The user needs to pay storage for both RRS (restored) and Glacier (Archive) Rates
To meet regulatory requirements, a pharmaceuticals company needs to archive data after a drug trial test is concluded. Each drug trial test may generate up to several thousands of files, with compressed file sizes ranging from 1 byte to 100MB. Once archived, data rarely needs to be restored, and on the rare occasion when restoration is needed, the company has 24 hours to restore specific files that match certain metadata. Searches must be possible by numeric file ID, drug name, participant names, date ranges, and other metadata. Which is the most cost-effective architectural approach that can meet the requirements?
Store individual files in Amazon Glacier, using the file ID as the archive name. When restoring data, query the Amazon Glacier vault for files matching the search criteria. (Individual files are expensive and does not allow searching by participant names etc)
Store individual files in Amazon S3, and store search metadata in an Amazon Relational Database Service (RDS) multi-AZ database. Create a lifecycle rule to move the data to Amazon Glacier after a certain number of days. When restoring data, query the Amazon RDS database for files matching the search criteria, and move the files matching the search criteria back to S3 Standard class. (As the data is not needed can be stored to Glacier directly and the data need not be moved back to S3 standard)
Store individual files in Amazon Glacier, and store the search metadata in an Amazon RDS multi-AZ database. When restoring data, query the Amazon RDS database for files matching the search criteria, and retrieve the archive name that matches the file ID returned from the database query. (Individual files and Multi-AZ is expensive)
First, compress and then concatenate all files for a completed drug trial test into a single Amazon Glacier archive. Store the associated byte ranges for the compressed files along with other search metadata in an Amazon RDS database with regular snapshotting. When restoring data, query the database for files that match the search criteria, and create restored files from the retrieved byte ranges.
Store individual compressed files and search metadata in Amazon Simple Storage Service (S3). Create a lifecycle rule to move the data to Amazon Glacier, after a certain number of days. When restoring data, query the Amazon S3 bucket for files matching the search criteria, and retrieve the file to S3 reduced redundancy in order to move it back to S3 Standard class. (Once the data is moved from S3 to Glacier the metadata is lost, as Glacier does not have metadata and must be maintained externally)
A user is uploading archives to Glacier. The user is trying to understand key Glacier resources. Which of the below mentioned options is not a Glacier resource?
AWS Glue is a fully-managed, ETL i.e extract, transform, and load service that automates the time-consuming steps of data preparation for analytics
is serverless and supports pay-as-you-go model. There is no infrastructure to provision or manage.
handles provisioning, configuration, and scaling of the resources required to run the ETL jobs on a fully managed, scale-out Apache Spark environment.
makes it simple and cost-effective to categorize the data, clean it, enrich it, and move it reliably between various data stores and streams.
also helps setup, orchestrate, and monitor complex data flows.
help automate much of the undifferentiated heavy lifting involved with discovering, categorizing, cleaning, enriching, and moving data, so more time can be spent on analyzing the data.
also supports custom Scala or Python code and import custom libraries and Jar files into the AWS Glue ETL jobs to access data sources not natively supported by AWS Glue.
supports server side encryption for data at rest and SSL for data in motion.
provides development endpoints to edit, debug, and test the code it generates.
Glue ETL engine to Extract, Transform, and Load data that can automatically generate Scala or Python code.
Glue Data Catalog is a central repository and persistent metadata store to store structural and operational metadata for all the data assets.
Glue crawlers scan various data stores to automatically infer schemas and partition structures to populate the Data Catalog with corresponding table definitions and statistics.
AWS Glue Streaming ETL enables performing ETL operations on streaming data using continuously-running jobs.
Glue Flexible scheduler that handles dependency resolution, job monitoring, and retries.
Glue Studio offers a graphical interface for authoring AWS Glue jobs to process data allowing you to define the flow of the data sources, transformations, and targets in the visual interface and generating Apache Spark code on your behalf.
Glue Data Quality helps reduces manual data quality efforts by automatically measuring and monitoring the quality of data in data lakes and pipelines.
Glue DataBrew is a visual data preparation tool that makes it easy for data analysts and data scientists to prepare data with an interactive, point-and-click visual interface without writing code. It helps to visualize, clean, and normalize data directly from the data lake, data warehouses, and databases, including S3, Redshift, Aurora, and RDS.
AWS Glue Data Catalog
AWS Glue Data Catalog is a central repository and persistent metadata store to store structural and operational metadata for all the data assets.
AWS Glue Data Catalog provides a uniform repository where disparate systems can store and find metadata to keep track of data in data silos, and use that metadata to query and transform the data.
For a given data set, Data Catalog can store its table definition, physical location, add business-relevant attributes, as well as track how this data has changed over time.
Data Catalog is Apache Hive Metastore compatible and is a drop-in replacement for the Hive Metastore for Big Data applications running on EMR.
Data Catalog also provides out-of-box integration with Athena, EMR, and Redshift Spectrum.
Table definitions once added to the Glue Data Catalog, are available for ETL and also readily available for querying in Athena, EMR, and Redshift Spectrum to provide a common view of the data between these services.
Data Catalog supports bulk import of the metadata from existing persistent Apache Hive Metastore by using our import script.
Data Catalog provides comprehensive audit and governance capabilities, with schema change tracking and data access controls, which helps ensure that data is not inappropriately modified or inadvertently shared
Each AWS account has one AWS Glue Data Catalog per region.
AWS Glue Crawlers
AWS Glue crawler connects to a data store, progresses through a prioritized list of classifiers to extract the schema of the data and other statistics, and then populates the Data Catalog with this metadata.
Glue crawlers scan various data stores to automatically infer schemas and partition structures to populate the Data Catalog with corresponding table definitions and statistics.
Glue crawlers can be scheduled to run periodically so that the metadata is always up-to-date and in-sync with the underlying data.
Crawlers automatically add new tables, new partitions to existing tables, and new versions of table definitions.
Dynamic Frames
AWS Glue is designed to work with semi-structured data and introduces a dynamic frame component, which can be used in the ETL scripts.
Dynamic frame is a distributed table that supports nested data such as structures and arrays.
Each record is self-describing, designed for schema flexibility with semi-structured data. Each record contains both data and the schema that describes that data.
A Dynamic Frame is similar to an Apache Spark dataframe, which is a data abstraction used to organize data into rows and columns, except that each record is self-describing so no schema is required initially.
Dynamic frames provide schema flexibility and a set of advanced transformations specifically designed for dynamic frames.
Conversion can be done between Dynamic frames and Spark dataframes, to take advantage of both AWS Glue and Spark transformations to do the kinds of analysis needed.
AWS Glue Streaming ETL
AWS Glue enables performing ETL operations on streaming data using continuously-running jobs.
AWS Glue streaming ETL is built on the Apache Spark Structured Streaming engine, and can ingest streams from Kinesis Data Streams and Apache Kafka using Amazon Managed Streaming for Apache Kafka.
Streaming ETL can clean and transform streaming data and load it into S3 or JDBC data stores.
Use Streaming ETL in AWS Glue to process event data like IoT streams, clickstreams, and network logs.
Glue Job Bookmark
Glue Job Bookmark tracks data that has already been processed during a previous run of an ETL job by persisting state information from the job run.
Job bookmarks help Glue maintain state information and prevent the reprocessing of old data.
Job bookmarks help process new data when rerunning on a scheduled interval
Job bookmark is composed of the states for various elements of jobs, such as sources, transformations, and targets. for e.g, an ETL job might read new partitions in an S3 file. Glue tracks which partition the job has processed successfully to prevent duplicate processing and duplicate data in the job’s target data store.
Glue Databrew
Glue DataBrew is a visual data preparation tool that enables users to clean and normalize data without writing any code.
is serverless, and can help explore and transform terabytes of raw data without needing to create clusters or manage any infrastructure.
helps reduce the time it takes to prepare data for analytics and machine learning (ML).
provides 250 ready-made transformations to automate data preparation tasks, such as filtering anomalies, converting data to standard formats, and correcting invalid values.
business analysts, data scientists, and data engineers can more easily collaborate to get insights from raw data.
AWS Certification Exam Practice Questions
Questions are collected from Internet and the answers are marked as per my knowledge and understanding (which might differ with yours).
AWS services are updated everyday and both the answers and questions might be outdated soon, so research accordingly.
AWS exam questions are not updated to keep up the pace with AWS updates, so even if the underlying feature has changed the question might not be updated
Open to further feedback, discussion and correction.
An organization is setting up a data catalog and metadata management environment for their numerous data stores currently running on AWS. The data catalog will be used to determine the structure and other attributes of data in the data stores. The data stores are composed of Amazon RDS databases, Amazon Redshift, and CSV files residing on Amazon S3. The catalog should be populated on a scheduled basis, and minimal administration is required to manage the catalog. How can this be accomplished?
Set up Amazon DynamoDB as the data catalog and run a scheduled AWS Lambda function that connects to data sources to populate the database.
Use an Amazon database as the data catalog and run a scheduled AWS Lambda function that connects to data sources to populate the database.
Use AWS Glue Data Catalog as the data catalog and schedule crawlers that connect to data sources to populate the database.
Set up Apache Hive metastore on an Amazon EC2 instance and run a scheduled bash script that connects to data sources to populate the metastore.
I recently recertified for the AWS Certified SysOps Administrator – Associate (SOA-C02) exam.
SOA-C02 is the updated version of the SOA-C01 AWS exam with hands-on labs included, which is the first with AWS.
NOTE: As of March 28, 2023, the AWS Certified SysOps Administrator – Associate exam will not include exam labs until further notice. This removal of exam labs is temporary while we evaluate the exam labs and make improvements to provide an optimal candidate experience.
SOA-C02 is the first AWS exam that included 2 sections
Objective questions
Hands-on labs
With Labs
SOA-C02 Exam consists of around 50 objective-type questions and 3 Hands-on labs to be answered in 190 minutes.
Labs are performed in a separate instance. Copy-paste works, so make sure you copy the exact names on resource creation.
Labs are pretty easy if you have worked on AWS.
Plan to leave 20 minutes to complete each exam lab.
NOTE: Once you complete a section and click next you cannot go back to the section. The same is for the labs. Once a lab is completed, you cannot return back to the lab.
Practice the Sample Lab provided when you book the exam, which would give you a feel of how the hands-on exam would actually be.
Without Labs
SOA-C02 exam consists of 65 questions in 130 minutes, and the time is more than sufficient if you are well-prepared.
SOA-C02 exam includes two types of questions, multiple-choice and multiple-response.
SOA-C02 has a scaled score between 100 and 1,000. The scaled score needed to pass the exam is 720.
Associate exams currently cost $ 150 + tax.
You can get an additional 30 minutes if English is your second language by requesting Exam Accommodations. It might not be needed for Associate exams but is helpful for Professional and Specialty ones.
AWS exams can be taken either remotely or online, I prefer to take them online as it provides a lot of flexibility. Just make sure you have a proper place to take the exam with no disturbance and nothing around you.
Also, if you are taking the AWS Online exam for the first time try to join at least 30 minutes before the actual time as I have had issues with both PSI and Pearson with long wait times.
Signed up with AWS for the Free Tier account which provides a lot of the Services to be tried for free with certain limits which are more than enough to get things going. Be sure to decommission anything, if you using anything beyond the free limits, preventing any surprises 🙂
Elastic Beanstalk helps to quickly deploy and manage applications in the AWS Cloud without having to worry about the infrastructure that runs those applications.
Understand Elastic Beanstalk overall – Applications, Versions, and Environments
SCP provides the maximum permission that a user can have, however, the user still needs to be explicitly given IAM policy.
Consolidated billing enables consolidating payments from multiple AWS accounts and includes combined usage and volume discounts including sharing of Reserved Instances across accounts.
Systems Manager is the operations hub and provides various services like parameter store, patch manager
Parameter Store provides secure, scalable, centralized, hierarchical storage for configuration data and secret management. Does not support secrets rotation. Use Secrets Manager instead
Session Manager provides secure and auditable instance management without the need to open inbound ports, maintain bastion hosts, or manage SSH keys.
Patch Manager helps automate the process of patching managed instances with both security-related and other types of updates.
collects monitoring and operational data in the form of logs, metrics, and events, and visualizes it.
EC2 metrics can track (disk, network, CPU, status checks) but do not capture metrics like memory, disk swap, disk storage, etc.
CloudWatch unified agent can be used to gather custom metrics like memory, disk swap, disk storage, etc.
CloudWatch Alarm actions can be configured to perform actions based on various metrics for e.g. CPU below 5%
CloudWatch alarm can monitor StatusCheckFailed_System status on an EC2 instance and automatically recover the instance if it becomes impaired due to an underlying hardware failure or a problem that requires AWS involvement to repair.
Load Balancer metrics SurgeQueueLength and SpilloverCount
HealthyHostCount, UnHealthyHostCount determines the number of healthy and unhealthy instances registered with the load balancer.
Reasons for 4XX and 5XX errors
CloudWatch logs can be used to monitor, store, and access log files from EC2 instances, CloudTrail, Route 53, and other sources. You can create metric filters over the logs.
With Organizations, the trail can be configured to log CloudTrail from all accounts to a central account.
CloudTrail log file integrity validation can be used to check whether a log file was modified, deleted, or unchanged after being delivered.
AWS Config is a fully managed service that provides AWS resource inventory, configuration history, and configuration change notifications to enable security, compliance, and governance.
supports managed as well as custom rules that can be evaluated on periodic basis or as the event occurs for compliance and trigger automatic remediation
Conformance pack is a collection of AWS Config rules and remediation actions that can be easily deployed as a single entity in an account and a Region or across an organization in AWS Organizations.
Egress-Only Internet Gateway – relevant to IPv6 only to allow egress traffic from private subnet to internet, without allowing ingress traffic
VPC Flow Logs enables you to capture information about the IP traffic going to and from network interfaces in the VPC and can help in monitoring the traffic or troubleshooting any connectivity issues
VPC Peering provides a connection between two VPCs that enables routing of traffic between them using private IP addresses.
VPC Endpoints enables the creation of a private connection between VPC to supported AWS services and VPC endpoint services powered by PrivateLink using its private IP address
Ability to debug networking issues like EC2 not accessible, EC2 not reachable, or not able to communicate with others or Internet.
Failover routing policy helps to configure active-passive failover.
Geolocation routing policy helps route traffic based on the location of the users.
Geoproximity routing policy helps route traffic based on the location of the resources and, optionally, shift traffic from resources in one location to resources in another.
Latency routing policy use with resources in multiple AWS Regions and you want to route traffic to the Region that provides the best latency with less round-trip time.
Weighted routing policy helps route traffic to multiple resources in specified proportions.
Focus on Weighted, Latency routing policies
Understand ELB, ALB, and NLB and what features they provide like
InstanceLimitExceeded – Concurrent running instance limit, default is 20, has been reached in a region. Request increase in limits.
InsufficientInstanceCapacity – AWS does not currently have enough available capacity to service the request. Change AZ or Instance Type.
Monitoring EC2 instances
System status checks failure – Stop and Start
Instance status checks failure – Reboot
EC2 supports Instance Recovery where the recovered instance is identical to the original instance, including the instance ID, private IP addresses, Elastic IP addresses, and all instance metadata.
EC2 Image Builder can be used to pre-baked images with software to speed up booting and launching time.
Lambda functions can be hosted in VPC with internet access controlled by a NAT instance.
RDS Proxy acts as an intermediary between the application and an RDS database. RDS Proxy establishes and manages the necessary connection pools to the database so that the application creates fewer database connections.
S3 data protection provides encryption at rest and encryption in transit
S3 default encryption can be used to encrypt the data with S3 bucket policies to prevent or reject unencrypted object uploads.
Multi-part handling for fault-tolerant and performant large file uploads
static website hosting, CORS
S3 Versioning can help recover from accidental deletes and overwrites.
Pre-Signed URLs for both upload and download
S3 Transfer Acceleration enables fast, easy, and secure transfers of files over long distances between the client and an S3 bucket using globally distributed edge locations in CloudFront.
Understand Glacier as archival storage. Glacier does not provide immediate access to the data even with expediated retrievals.
Storage Gateway allows storage of data in the AWS cloud for scalable and cost-effective storage while maintaining data security.
Gateway-cached volumes stores data is stored in S3 and retains a copy of recently read data locally for low latency access to the frequently accessed data
Gateway-stored volumes maintain the entire data set locally to provide low latency access
EFS is a cost-optimized, serverless, scalable, and fully managed file storage for use with AWS Cloud and on-premises resources.
supports data at rest encryption only during the creation. After creation, the file system cannot be encrypted and must be copied over to a new encrypted disk.
supports General purpose and Max I/O performance mode.
If hitting PercentIOLimit issue move to Max I/O performance mode.
FSx makes it easy and cost-effective to launch, run, and scale feature-rich, high-performance file systems in the cloud
FSx for Windows supports SMB protocol and a Multi-AZ file system to provide high availability across multiple AZs.
AWS Backup can be used to automate backup for EC2 instances and EFS file systems
Data Lifecycle Manager to automate the creation, retention, and deletion of EBS snapshots and EBS-backed AMIs.
AWS DataSync automates moving data between on-premises storage and S3 or Elastic File System (EFS).
Databases
RDS provides cost-efficient, resizable capacity for an industry-standard relational database and manages common database administration tasks.
Multi-AZ deployment provides high availability, durability, and failover support
Read replicas enable increased scalability and database availability in the case of an AZ failure.
Automated backups and database change logs enable point-in-time recovery of the database during the backup retention period, up to the last five minutes of database usage.
Aurora is a fully managed, MySQL- and PostgreSQL-compatible, relational database engine
Backtracking “rewinds” the DB cluster to the specified time and performs in-place restore and does not create a new instance.
Automated Backups that help restore the DB as a new instance
Know ElastiCache use cases, mainly for caching performance
Understand ElastiCache Redis vs Memcached
Redis provides Multi-AZ support helps provide high availability across AZs and Online resharding to dynamically scale.
ElastiCache can be used as a caching layer for RDS.
S3 default encryption can help encrypt objects, however, it does not encrypt existing objects before the setting was enabled. You can use S3 Inventory to list the objects and S3 Batch to encrypt them.
Understand KMS for key management and envelope encryption
KMS with imported customer key material does not support rotation and has to be done manually.
AWS WAF – Web Application Firewall helps protect the applications against common web exploits like XSS or SQL Injection and bots that may affect availability, compromise security, or consume excessive resources
AWS GuardDuty is a threat detection service that continuously monitors the AWS accounts and workloads for malicious activity and delivers detailed security findings for visibility and remediation.
AWS Secrets Manager can help securely expose credentials as well as rotate them.
Secrets Manager integrates with Lambda and supports credentials rotation
AWS Shield is a managed Distributed Denial of Service (DDoS) protection service that safeguards applications running on AWS
is an automated security assessment service that helps improve the security and compliance of applications deployed on AWS.
automatically assesses applications for exposure, vulnerabilities, and deviations from best practices.
AWS Certificate Manager (ACM) handles the complexity of creating, storing, and renewing public and private SSL/TLS X.509 certificates and keys that protect the AWS websites and applications.
Know AWS Artifact as on-demand access to compliance reports
Analytics
Amazon Athena can be used to query S3 data without duplicating the data and using SQL queries
OpenSearch (Elasticsearch) service is a distributed search and analytics engine built on Apache Lucene.
Opensearch production setup would be 3 AZs, 3 dedicated master nodes, 6 nodes with two replicas in each AZ.
Integration Tools
Understand SQS as a message queuing service and SNS as pub/sub notification service
Focus on SQS as a decoupling service
Understand SQS FIFO, make sure you know the differences between standard and FIFO
Understand CloudWatch integration with SNS for notification
Practice Labs
Create IAM users, IAM roles with specific limited policies.
Create a private S3 bucket
enable versioning
enable default encryption
enable lifecycle policies to transition and expire the objects
enable same region replication
Create a public S3 bucket with static website hosting
Set up a VPC with public and private subnets with Routes, SGs, NACLs.
Set up a VPC with public and private subnets and enable communication from private subnets to the Internet using NAT gateway
Create EC2 instance, create a Snapshot and restore it as a new instance.
Set up Security Groups for ALB and Target Groups, and create ALB, Launch Template, Auto Scaling Group, and target groups with sample applications. Test the flow.
Create Multi-AZ RDS instance and instance force failover.
Set up SNS topic. Use Cloud Watch Metrics to create a CloudWatch alarm on specific thresholds and send notifications to the SNS topic
Set up SNS topic. Use Cloud Watch Logs to create a CloudWatch alarm on log patterns and send notifications to the SNS topic.
Update a CloudFormation template and re-run the stack and check the impact.
Use AWS Data Lifecycle Manager to define snapshot lifecycle.
Use AWS Backup to define EFS backup with hourly and daily backup rules.
AWS Certified SysOps Administrator – Associate (SOA-C02) Exam Day
Make sure you are relaxed and get some good night’s sleep. The exam is not tough if you are well-prepared.
If you are taking the AWS Online exam
Try to join at least 30 minutes before the actual time as I have had issues with both PSI and Pearson with long wait times.
The online verification process does take some time and usually, there are glitches.
Remember, you would not be allowed to take the take if you are late by more than 30 minutes.
Make sure you have your desk clear, no hand-watches, or external monitors, keep your phones away, and nobody can enter the room.
AWS Certified DevOps Engineer – Professional (DOP-C02) Exam Learning Path
AWS Certified DevOps Engineer – Professional (DOP-C02) exam is the upgraded pattern of the DevOps Engineer – Professional (DOP-C01) exam which was released in March 2023.
I recently attempted the latest pattern and DOP-C02 is quite similar to DOP-C01 with the inclusion of new services and features.
AWS Certified DevOps Engineer – Professional (DOP-C02) Exam Content
AWS Certified DevOps Engineer – Professional (DOP-C02) exam is intended for individuals who perform a DevOps engineer role and focuses on provisioning, operating, and managing distributed systems and services on AWS.
DOP-C02 basically validates
Implement and manage continuous delivery systems and methodologies on AWS
Implement and automate security controls, governance processes, and compliance validation
Define and deploy monitoring, metrics, and logging systems on AWS
Implement systems that are highly available, scalable, and self-healing on the AWS platform
Design, manage, and maintain tools to automate operational processes
AWS Certified DevOps Engineer – Professional (DOP-C02) Exam Summary
Professional exams are tough, lengthy, and tiresome. Most of the questions and answers options have a lot of prose and a lot of reading that needs to be done, so be sure you are prepared and manage your time well.
Each solution involves multiple AWS services.
DOP-C02 exam has 75 questions to be solved in 170 minutes. Only 65 affect your score, while 10 unscored questions are for evaluation for future use.
DOP-C02 exam includes two types of questions, multiple-choice and multiple-response.
DOP-C02 has a scaled score between 100 and 1,000. The scaled score needed to pass the exam is 750.
Each question mainly touches multiple AWS services.
Professional exams currently cost $ 300 + tax.
You can get an additional 30 minutes if English is your second language by requesting Exam Accommodations. It might not be needed for Associate exams but is helpful for Professional and Specialty ones.
As always, mark the questions for review and move on and come back to them after you are done with all.
As always, having a rough architecture or mental picture of the setup helps focus on the areas that you need to improve. Trust me, you will be able to eliminate 2 answers for sure and then need to focus on only the other two. Read the other 2 answers to check the difference area and that would help you reach the right answer or at least have a 50% chance of getting it right.
AWS exams can be taken either remotely or online, I prefer to take them online as it provides a lot of flexibility. Just make sure you have a proper place to take the exam with no disturbance and nothing around you.
Also, if you are taking the AWS Online exam for the first time try to join at least 30 minutes before the actual time as I have had issues with both PSI and Pearson with long wait times.
AWS Certified DevOps Engineer – Professional (DOP-C02) Exam Topics
AWS Certified DevOps Engineer – Professional exam covers a lot of concepts and services related to Automation, Deployments, Disaster Recovery, HA, Monitoring, Logging, and Troubleshooting. It also covers security and compliance related topics.
AWS Systems Manager and its various services like parameter store, patch manager
Parameter Store provides secure, scalable, centralized, hierarchical storage for configuration data and secret management. Does not support secrets rotation. Use Secrets Manager instead
Session Manager provides secure and auditable instance management without the need to open inbound ports, maintain bastion hosts, or manage SSH keys.
Patch Manager helps automate the process of patching managed instances with both security-related and other types of updates.
CloudWatch logs can be used to monitor, store, and access log files from EC2 instances, CloudTrail, Route 53, and other sources. You can create metric filters over the logs.
With Organizations, the trail can be configured to log CloudTrail from all accounts to a central account.
Config is a fully managed service that provides AWS resource inventory, configuration history, and configuration change notifications to enable security, compliance, and governance.
supports managed as well as custom rules that can be evaluated on periodic basis or as the event occurs for compliance and trigger automatic remediation
Conformance pack is a collection of AWS Config rules and remediation actions that can be easily deployed as a single entity in an account and a Region or across an organization in AWS Organizations.
CodePipeline is a fully managed continuous delivery service that helps automate the release pipelines for fast and reliable application and infrastructure updates.
CodePipeline pipeline structure (Hint : run builds parallelly using runorder)
Understand how to configure notifications on events and failures
CodePipeline supports Manual Approval
CodeArtifact is a fully managed artifact repository service that makes it easy for organizations of any size to securely store, publish, and share software packages used in their software development process.
CodeGuru provides intelligent recommendations to improve code quality and identify an application’s most expensive lines of code. Reviewer helps improve code quality and Profiler helps optimize performance for applications
EC2 Image Builder helps to automate the creation, management, and deployment of customized, secure, and up-to-date server images that are pre-installed and pre-configured with software and settings to meet specific IT standards.
Disaster Recovery
Disaster recovery is mainly covered as a part of Re-silent cloud solutions.
Disaster Recovery whitepaper, although outdated, make sure you understand the differences and implementation for each type esp. pilot light, warm standby w.r.t RTO, and RPO.
Compute
Make components available in an alternate region,
Backup and Restore using either snapshots or AMIs that can be restored.
Use minimal low-scale capacity running which can be scaled once the failover happens
Use fully running compute in active-active confirmation with health checks.
CloudFormation to create, and scale infra as needed
Storage
S3 and EFS support cross-region replication
DynamoDB supports Global tables for multi-master, active-active inter-region storage needs.
RDS supports cross-region read replicas which can be promoted to master in case of a disaster. This can be done using Route 53, CloudWatch, and lambda functions.
Network
Route 53 failover routing with health checks to failover across regions.
CloudFront Origin Groups support primary and secondary endpoints with failover.
is a threat detection service that continuously monitors the AWS accounts and workloads for malicious activity and delivers detailed security findings for visibility and remediation.
AWS Security Hub is a cloud security posture management service that performs security best practice checks, aggregates alerts and enables automated remediation.
Firewall Manager helps centrally configure and manage firewall rules across the accounts and applications in AWS Organizations which includes a variety of protections, including WAF, Shield Advanced, VPC security groups, Network Firewall, and Route 53 Resolver DNS Firewall.
S3 Access Logs enable tracking access requests to an S3 bucket.
S3 Event Notification enables notifications to be triggered when certain events happen in the bucket and supports SNS, SQS, and Lambda as the destination. S3 needs permission to be able to integrate with the services.
helps define reserved concurrency limits to reduce the impact
Lambda Alias now supports canary deployments
Reserved Concurrency guarantees the maximum number of concurrent instances for the function
Provisioned Concurrency
provides greater control over the performance of serverless applications and helps keep functions initialized and hyper-ready to respond in double-digit milliseconds.
supports Application Auto Scaling.
Step Functions helps developers use AWS services to build distributed applications, automate processes, orchestrate microservices, and create data and machine learning (ML) pipelines.
SQS supports dead letter queues and redrive policy which specifies the source queue, the dead-letter queue, and the conditions under which SQS moves messages from the former to the latter if the consumer of the source queue fails to process a message a specified number of times.
CloudWatch integration with SNS and Lambda for notifications.
AWS S3 offers a range of S3 Storage Classes to match the use case scenario and performance access requirements.
S3 storage classes are designed to sustain the concurrent loss of data in one or two facilities.
S3 storage classes allow lifecycle management for automatic transition of objects for cost savings.
All S3 storage classes provide the same durability, first-byte latency, and support SSL encryption of data in transit, and data encryption at rest.
S3 also regularly verifies the integrity of the data using checksums and provides the auto-healing capability.
S3 Storage Classes Comparison
S3 Standard
STANDARD is the default storage class, if none specified during upload
Low latency and high throughput performance
Designed for 99.999999999% i.e. 11 9’s Durability of objects across AZs
Designed for 99.99% availability over a given year
Resilient against events that impact an entire Availability Zone and is designed to sustain the loss of data in a two facilities
Ideal for performance-sensitive use cases and frequently accessed data
S3 Standard is appropriate for a wide variety of use cases, including cloud applications, dynamic websites, content distribution, mobile and gaming applications, and big data analytics.
S3 Intelligent Tiering (S3 Intelligent-Tiering)
S3 Intelligent Tiering storage class is designed to optimize storage costs by automatically moving data to the most cost-effective storage access tier, without performance impact or operational overhead.
Delivers automatic cost savings by moving data on a granular object-level between two access tiers
one tier that is optimized for frequent access and
another lower-cost tier that is optimized for infrequently accessed data.
a frequent access tier and a lower-cost infrequent access tier, when access patterns change.
Ideal to optimize storage costs automatically for long-lived data when access patterns are unknown or unpredictable.
For a small monthly monitoring and automation fee per object, S3 monitors access patterns of the objects and moves objects that have not been accessed for 30 consecutive days to the infrequent access tier.
There are no separate retrieval fees when using the Intelligent Tiering storage class. If an object in the infrequent access tier is accessed, it is automatically moved back to the frequent access tier.
No additional fees apply when objects are moved between access tiers
Suitable for objects greater than 128 KB (smaller objects are charged for 128 KB only) kept for at least 30 days (charged for a minimum of 30 days)
Same low latency and high throughput performance of S3 Standard
Designed for 99.999999999% i.e. 11 9’s Durability of objects across AZs
Designed for 99.9% availability over a given year
S3 Standard-Infrequent Access (S3 Standard-IA)
S3 Standard-Infrequent Access storage class is optimized for long-lived and less frequently accessed data. for e.g. for backups and older data where access is limited, but the use case still demands high performance
Ideal for use for the primary or only copy of data that can’t be recreated.
Data stored redundantly across multiple geographically separated AZs and are resilient to the loss of an Availability Zone.
offers greater availability and resiliency than the ONEZONE_IA class.
Objects are available for real-time access.
Suitable for larger objects greater than 128 KB (smaller objects are charged for 128 KB only) kept for at least 30 days (charged for minimum 30 days)
Same low latency and high throughput performance of Standard
Designed for 99.999999999% i.e. 11 9’s Durability of objects across AZs
Designed for 99.9% availability over a given year
S3 charges a retrieval fee for these objects, so they are most suitable for infrequently accessed data.
S3 One Zone-Infrequent Access (S3 One Zone-IA)
S3 One Zone-Infrequent Access storage classes are designed for long-lived and infrequently accessed data, but available for millisecond access (similar to the STANDARD and STANDARD_IA storage class).
Ideal when the data can be recreated if the AZ fails, and for object replicas when setting cross-region replication (CRR).
Objects are available for real-time access.
Suitable for objects greater than 128 KB (smaller objects are charged for 128 KB only) kept for at least 30 days (charged for a minimum of 30 days)
Stores the object data in only one AZ, which makes it less expensive than Standard-Infrequent Access
Data is not resilient to the physical loss of the AZ resulting from disasters, such as earthquakes and floods.
One Zone-Infrequent Access storage class is as durable as Standard-Infrequent Access, but it is less available and less resilient.
Designed for 99.999999999% i.e. 11 9’s Durability of objects in a single AZ
Designed for 99.5% availability over a given year
S3 charges a retrieval fee for these objects, so they are most suitable for infrequently accessed data.
Reduced Redundancy Storage – RRS
NOTE – AWS recommends not to use this storage class. The STANDARD storage class is more cost-effective now.
Reduced Redundancy Storage (RRS) storage class is designed for non-critical, reproducible data stored at lower levels of redundancy than the STANDARD storage class, which reduces storage costs
Designed for durability of 99.99% of objects
Designed for 99.99% availability over a given year
Lower level of redundancy results in less durability and availability
RRS stores object on multiple devices across multiple facilities, providing 400 times the durability of a typical disk drive,
RRS does not replicate objects as many times as S3 standard storage and is designed to sustain the loss of data in a single facility.
If an RRS object is lost, S3 returns a 405 error on requests made to that object
S3 can send an event notification, configured on the bucket, to alert a user or start a workflow when it detects that an RRS object is lost which can be used to replace the lost object
S3 Glacier Instant Retrieval
Use for archiving data that is rarely accessed and requires milliseconds retrieval.
Storage class has a minimum storage duration period of 90 days
Designed for 99.999999999% i.e. 11 9’s Durability of objects across AZs
Designed for 99.99% availability
S3 Glacier Flexible Retrieval – S3 Glacier
S3 GLACIER storage class is suitable for low-cost data archiving where data access is infrequent and retrieval time of minutes to hours is acceptable.
Storage class has a minimum storage duration period of 90 days
Provides configurable retrieval times, from minutes to hours
Expedited retrieval: 1-5 mins
Standard retrieval: 3-5 hours
Bulk retrieval: 5-12 hours
GLACIER storage class uses the very low-cost Glacier storage service, but the objects in this storage class are still managed through S3
For accessing GLACIER objects,
the object must be restored which can take anywhere between minutes to hours
objects are only available for the time period (the number of days) specified during the restoration request
object’s storage class remains GLACIER
charges are levied for both the archive (GLACIER rate) and the copy restored temporarily
Vault Lock feature enforces compliance via a lockable policy.
Offers the same durability and resiliency as the STANDARD storage class
Designed for 99.999999999% i.e. 11 9’s Durability of objects across AZs
Designed for 99.99% availability
S3 Glacier Deep Archive
Glacier Deep Archive storage class provides the lowest-cost data archiving where data access is infrequent and retrieval time of hours is acceptable.
Has a minimum storage duration period of 180 days and can be accessed at a default retrieval time of 12 hours.
Supports long-term retention and digital preservation for data that may be accessed once or twice a year
Designed for 99.999999999% i.e. 11 9’s Durability of objects across AZs
Designed for 99.9% availability over a given year
DEEP_ARCHIVE retrieval costs can be reduced by using bulk retrieval, which returns data within 48 hours.
Ideal alternative to magnetic tape libraries
S3 Analytics – S3 Storage Classes Analysis
S3 Analytics – Storage Class Analysis helps analyze storage access patterns to decide when to transition the right data to the right storage class.
S3 Analytics feature observes data access patterns to help determine when to transition less frequently accessed STANDARD storage to the STANDARD_IA (IA, for infrequent access) storage class.
Storage Class Analysis can be configured to analyze all the objects in a bucket or filters to group objects.
AWS Certification Exam Practice Questions
Questions are collected from Internet and the answers are marked as per my knowledge and understanding (which might differ with yours).
AWS services are updated everyday and both the answers and questions might be outdated soon, so research accordingly.
AWS exam questions are not updated to keep up the pace with AWS updates, so even if the underlying feature has changed the question might not be updated
Open to further feedback, discussion and correction.
What does RRS stand for when talking about S3?
Redundancy Removal System
Relational Rights Storage
Regional Rights Standard
Reduced Redundancy Storage
What is the durability of S3 RRS?
99.99%
99.95%
99.995%
99.999999999%
What is the Reduced Redundancy option in Amazon S3?
Less redundancy for a lower cost
It doesn’t exist in Amazon S3, but in Amazon EBS.
It allows you to destroy any copy of your files outside a specific jurisdiction.
It doesn’t exist at all
An application is generating a log file every 5 minutes. The log file is not critical but may be required only for verification in case of some major issue. The file should be accessible over the internet whenever required. Which of the below mentioned options is a best possible storage solution for it?
AWS S3
AWS Glacier
AWS RDS
AWS S3 RRS (Reduced Redundancy Storage (RRS) is an Amazon S3 storage option that enables customers to store noncritical, reproducible data at lower levels of redundancy than Amazon S3’s standard storage. RRS is designed to sustain the loss of data in a single facility.)
A user has moved an object to Glacier using the life cycle rules. The user requests to restore the archive after 6 months. When the restore request is completed the user accesses that archive. Which of the below mentioned statements is not true in this condition?
The archive will be available as an object for the duration specified by the user during the restoration request
The restored object’s storage class will be RRS (After the object is restored the storage class still remains GLACIER. Read more)
The user can modify the restoration period only by issuing a new restore request with the updated period
The user needs to pay storage for both RRS (restored) and Glacier (Archive) Rates
Your department creates regular analytics reports from your company’s log files. All log data is collected in Amazon S3 and processed by daily Amazon Elastic Map Reduce (EMR) jobs that generate daily PDF reports and aggregated tables in CSV format for an Amazon Redshift data warehouse. Your CFO requests that you optimize the cost structure for this system. Which of the following alternatives will lower costs without compromising average performance of the system or data integrity for the raw data? [PROFESSIONAL]
Use reduced redundancy storage (RRS) for PDF and CSV data in Amazon S3. Add Spot instances to Amazon EMR jobs. Use Reserved Instances for Amazon Redshift. (Spot instances impacts performance)
Use reduced redundancy storage (RRS) for all data in S3. Use a combination of Spot instances and Reserved Instances for Amazon EMR jobs. Use Reserved instances for Amazon Redshift (Combination of the Spot and reserved with guarantee performance and help reduce cost. Also, RRS would reduce cost and guarantee data integrity, which is different from data durability )
Use reduced redundancy storage (RRS) for all data in Amazon S3. Add Spot Instances to Amazon EMR jobs. Use Reserved Instances for Amazon Redshift (Spot instances impacts performance)
Use reduced redundancy storage (RRS) for PDF and CSV data in S3. Add Spot Instances to EMR jobs. Use Spot Instances for Amazon Redshift. (Spot instances impacts performance)
Which of the below mentioned options can be a good use case for storing content in AWS RRS?
Storing mission critical data Files
Storing infrequently used log files
Storing a video file which is not reproducible
Storing image thumbnails
A newspaper organization has an on-premises application which allows the public to search its back catalogue and retrieve individual newspaper pages via a website written in Java. They have scanned the old newspapers into JPEGs (approx. 17TB) and used Optical Character Recognition (OCR) to populate a commercial search product. The hosting platform and software is now end of life and the organization wants to migrate its archive to AWS and produce a cost efficient architecture and still be designed for availability and durability. Which is the most appropriate? [PROFESSIONAL]
Use S3 with reduced redundancy to store and serve the scanned files, install the commercial search application on EC2 Instances and configure with auto-scaling and an Elastic Load Balancer. (RRS impacts durability and commercial search would add to cost)
Model the environment using CloudFormation. Use an EC2 instance running Apache webserver and an open source search application, stripe multiple standard EBS volumes together to store the JPEGs and search index. (Using EBS is not cost effective for storing files)
Use S3 with standard redundancy to store and serve the scanned files, use CloudSearch for query processing, and use Elastic Beanstalk to host the website across multiple availability zones. (Standard S3 and Elastic Beanstalk provides availability and durability, Standard S3 and CloudSearch provides cost effective storage and search)
Use a single-AZ RDS MySQL instance to store the search index and the JPEG images use an EC2 instance to serve the website and translate user queries into SQL. (RDS is not ideal and cost effective to store files, Single AZ impacts availability)
Use a CloudFront download distribution to serve the JPEGs to the end users and Install the current commercial search product, along with a Java Container for the website on EC2 instances and use Route53 with DNS round-robin. (CloudFront needs a source and using commercial search product is not cost effective)
A research scientist is planning for the one-time launch of an Elastic MapReduce cluster and is encouraged by her manager to minimize the costs. The cluster is designed to ingest 200TB of genomics data with a total of 100 Amazon EC2 instances and is expected to run for around four hours. The resulting data set must be stored temporarily until archived into an Amazon RDS Oracle instance. Which option will help save the most money while meeting requirements? [PROFESSIONAL]
Store ingest and output files in Amazon S3. Deploy on-demand for the master and core nodes and spot for the task nodes.
Optimize by deploying a combination of on-demand, RI and spot-pricing models for the master, core and task nodes. Store ingest and output files in Amazon S3 with a lifecycle policy that archives them to Amazon Glacier. (Master and Core must be RI or On Demand. Cannot be Spot)
Store the ingest files in Amazon S3 RRS and store the output files in S3. Deploy Reserved Instances for the master and core nodes and on-demand for the task nodes. (Need better durability for ingest file. Spot instances can be used for task nodes for cost saving.)
Deploy on-demand master, core and task nodes and store ingest and output files in Amazon S3 RRS (Input must be in S3 standard)
AWS Compute Optimizer helps analyze the configuration and utilization metrics of the AWS resources.
reports whether the resources are optimal, and generates optimization recommendations to reduce the cost and improve the performance of the workloads.
delivers intuitive and easily actionable resource recommendations to help quickly identify optimal AWS resources for the workloads without requiring specialized expertise or investing substantial time and money.
provides a global, cross-account view of all resources
analyzes the specifications and the utilization metrics of the resources from CloudWatch for the last 14 days.
provides graphs showing recent utilization metric history data, as well as projected utilization for recommendations, which can be used to evaluate which recommendation provides the best price-performance trade-off.
Analysis and visualization of the usage patterns can help decide when to move or resize the running resources, and still meet your performance and capacity requirements.
generates recommendations for the following resources:
Questions are collected from Internet and the answers are marked as per my knowledge and understanding (which might differ with yours).
AWS services are updated everyday and both the answers and questions might be outdated soon, so research accordingly.
AWS exam questions are not updated to keep up the pace with AWS updates, so even if the underlying feature has changed the question might not be updated
Open to further feedback, discussion and correction.
A company must assess the business’s EC2 instances and Elastic Block Store (EBS) volumes to determine how effectively the business is using resources. The company has not detected a pattern in how these EC2 instances are used by the apps that access the databases. Which option best fits these criteria in terms of cost-effectiveness?
Use AWS Systems Manager OpsCenter.
Use Amazon CloudWatch for detailed monitoring.
Use AWS Compute Optimizer.
Sign up for the AWS Enterprise Support plan. Turn on AWS Trusted Advisor.
Auto Scaling Launch Template vs Launch Configuration
Launch Configuration
Launch configuration is an instance configuration template that an Auto Scaling Group uses to launch EC2 instances.
Launch configuration is similar to EC2 configuration and involves the selection of the Amazon Machine Image (AMI), block devices, key pair, instance type, security groups, user data, EC2 instance monitoring, instance profile, kernel, ramdisk, the instance tenancy, whether the instance has a public IP address, and is EBS-optimized.
Launch configuration can be associated with multiple ASGs
Launch configuration can’t be modified after creation and needs to be created new if any modification is required.
Basic or detailed monitoring for the instances in the ASG can be enabled when a launch configuration is created.
By default, basic monitoring is enabled when you create the launch configuration using the AWS Management Console, and detailed monitoring is enabled when you create the launch configuration using the AWS CLI or an API
AWS recommends using Launch Template instead.
Launch Template
A Launch Template is similar to a launch configuration, with additional features, and is recommended by AWS.
Launch Template allows multiple versions of a template to be defined.
With versioning, a subset of the full set of parameters can be created and then reused to create other templates or template versions for e.g, a default template that defines common configuration parameters can be created and allow the other parameters to be specified as part of another version of the same template.
Launch Template allows the selection of both Spot and On-Demand Instances or multiple instance types.
Launch templates support EC2 Dedicated Hosts. Dedicated Hosts are physical servers with EC2 instance capacity that are dedicated to your use.
Launch templates provide the following features
Support for multiple instance types and purchase options in a single ASG.
Launching Spot Instances with the capacity-optimized allocation strategy.
Support for launching instances into existing Capacity Reservations through an ASG.
Support for unlimited mode for burstable performance instances.
Support for Dedicated Hosts.
Combining CPU architectures such as Intel, AMD, and ARM (Graviton2)
Improved governance through IAM controls and versioning.
Automating instance deployment with Instance Refresh.
AWS Certification Exam Practice Questions
Questions are collected from Internet and the answers are marked as per my knowledge and understanding (which might differ with yours).
AWS services are updated everyday and both the answers and questions might be outdated soon, so research accordingly.
AWS exam questions are not updated to keep up the pace with AWS updates, so even if the underlying feature has changed the question might not be updated
Open to further feedback, discussion and correction.
A company is launching a new workload. The workload will run on Amazon EC2 instances in an Amazon EC2 Auto Scaling group. The company needs to maintain different versions of the EC2 configurations. The company also needs the Auto Scaling group to automatically scale to maintain CPU utilization of 60%. How can a SysOps administrator meet these requirements?
Configure the Auto Scaling group to use a launch configuration with a target tracking scaling policy.
Configure the Auto Scaling group to use a launch configuration with a simple scaling policy.
Configure the Auto Scaling group to use a launch template with a target tracking scaling policy.
Configure the Auto Scaling group to use a launch template with a simple scaling policy.
AWS Certified Developer – Associate DVA-C02 exam is the latest AWS exam released on 27th February 2023 and has replaced the previous AWS Developer – Associate DVA-C01 certification exam.
I passed the AWS Developer – Associate DVA-C02 exam with a score of 835/1000.
DVA-C02 exam consists of 65 questions in 130 minutes, and the time is more than sufficient if you are well-prepared.
DVA-C02 exam includes two types of questions, multiple-choice and multiple-response.
DVA-C02 has a scaled score between 100 and 1,000. The scaled score needed to pass the exam is 720.
Associate exams currently cost $ 150 + tax.
You can get an additional 30 minutes if English is your second language by requesting Exam Accommodations. It might not be needed for Associate exams but is helpful for Professional and Specialty ones.
AWS exams can be taken either remotely or online, I prefer to take them online as it provides a lot of flexibility. Just make sure you have a proper place to take the exam with no disturbance and nothing around you.
Also, if you are taking the AWS Online exam for the first time try to join at least 30 minutes before the actual time as I have had issues with both PSI and Pearson with long wait times.
Signed up with AWS for the Free Tier account which provides a lot of Services to be tried for free with certain limits which are more than enough to get things going. Be sure to decommission anything, if you using anything beyond the free limits, preventing any surprises 🙂
Read the FAQs at least for the important topics, as they cover important points and are good for quick review
AWS DVA-C02 exam concepts cover solutions that fall within AWS Well-Architected framework to cover scalable, highly available, cost-effective, performant, and resilient pillars.
AWS Certified Developer – Associate DVA-C02 exam covers a lot of the latest AWS services like Amplify, X-Ray while focusing majorly on other services like Lambda, DynamoDB, Elastic Beanstalk, S3, EC2
AWS Certified Developer – Associate DVA-C02 exam is similar to DVA-C01 with more focus on the hands-on development and deployment concepts rather than just the architectural concepts.
If you had been preparing for the DVA-C01, DVA-C02 is pretty much similar except for the addition of some new services covering Amplify, X-Ray, etc.
S3 Client-side encryption encrypts data before storing it in S3
S3 encryption in transit can be enforced with S3 bucket policies using secureTransport attributes.
S3 encryption at rest can be enforced with S3 bucket policies using x-amz-server-side-encryption attribute.
S3 features including
S3 provides cost-effective static website hosting. However, it does not support HTTPS endpoint. Can be integrated with CloudFront for HTTPS, caching, performance, and low-latency access.
S3 versioning provides protection against accidental overwrites and deletions. Used with MFA Delete feature.
S3 Pre-Signed URLs for both upload and download provide access without needing AWS credentials.
simple, fully managed, scalable, serverless, and cost-optimized file storage for use with AWS Cloud and on-premises resources.
provides shared volume across multiple EC2 instances, while EBS can be attached to a single instance within the same AZ or EBS Multi-Attach can be attached to multiple instances within the same AZ
can be mounted with Lambda functions
supports the NFS protocol, and is compatible with Linux-based AMIs
supports cross-region replication and storage classes for cost management.
is a fully managed service that makes it easy for developers to publish, maintain, monitor, and secure APIs at any scale.
Powerful, flexible authentication mechanisms, such as AWS IAM policies, Lambda authorizer functions, and Amazon Cognito user pools.
supports Canary release deployments for safely rolling out changes.
define usage plans to meter, restrict third-party developer access, configure throttling, and quota limits on a per API key basis
integrates with AWS X-Ray for understanding and triaging performance latencies.
API Gateway CORS allows cross-domain calls
Amplify
is a complete solution that lets frontend web and mobile developers easily build, ship, and host full-stack applications on AWS, with the flexibility to leverage the breadth of AWS services as use cases evolve.
easy way to create and manage a collection of related AWS resources, and provision and update them in an orderly and predictable fashion.
Supports Serverless Application Model – SAM for the deployment of serverless applications including Lambda.
CloudFormation StackSets extends the functionality of stacks by enabling you to create, update, or delete stacks across multiple accounts and Regions with a single operation.
CodeCommit is a secure, scalable, fully-managed source control service that helps to host secure and highly scalable private Git repositories.
CodeBuild is a fully managed build service that compiles source code, runs tests, and produces software packages that are ready to deploy.
CodeDeploy helps automate code deployments to any instance, including EC2 instances and instances running on-premises.
CodePipeline is a fully managed continuous delivery service that helps automate the release pipelines for fast and reliable application and infrastructure updates.
CodeArtifact is a fully managed artifact repository service that makes it easy for organizations of any size to securely store, publish, and share software packages used in their software development process.