AWS Certified Solutions Architect Professional Practice Test (SAP-C02)
Use the form below to configure your AWS Certified Solutions Architect Professional Practice Test (SAP-C02). The practice test can be configured to only include certain exam objectives and domains. You can choose between 5-100 questions and set a time limit.

AWS Certified Solutions Architect Professional SAP-C02 Information
The AWS Certified Solutions Architect – Professional (SAP-C02) exam is a test for people who want to show advanced skills in cloud design using Amazon Web Services. It proves that you can handle large, complex systems and design solutions that are secure, reliable, and meet business needs. Passing this exam shows a higher level of knowledge than the associate-level test and is often needed for senior cloud roles.
This exam includes multiple-choice and multiple-response questions. It covers areas like designing for high availability, choosing the right storage and compute services, planning for cost, and managing security at scale. You will also need to understand how to migrate big applications to the cloud, design hybrid systems, and use automation tools to keep environments efficient and safe.
AWS suggests having at least two years of real-world experience before taking this test. The SAP-C02 exam takes 180 minutes, includes about 75 questions, and requires a scaled score of 750 out of 1000 to pass. Preparing usually means lots of practice with AWS services, using study guides, and trying practice exams. For many professionals, this certification is an important milestone toward becoming a cloud architect or senior cloud engineer.

Free AWS Certified Solutions Architect Professional SAP-C02 Practice Test
- 20 Questions
- Unlimited
- Design Solutions for Organizational ComplexityDesign for New SolutionsContinuous Improvement for Existing SolutionsAccelerate Workload Migration and Modernization
An organization operates a stateless multi-tenant REST API on Amazon ECS (AWS Fargate). The service is fronted by Application Load Balancers that run in two AWS Regions (us-east-1 and eu-west-1). Security teams must allow customers to allowlist a small, unchanging set of public IP addresses. New reliability objectives specify that the application must keep working if an entire AWS Region fails, client traffic must shift to the healthy Region within 30 seconds, and failover must happen without any DNS cache flushes or other client-side changes. Operations also want a fully managed AWS solution with minimal maintenance. Which approach best meets these requirements?
Deploy AWS Global Accelerator with an endpoint group in each Region that targets the existing Application Load Balancers and rely on Global Accelerator health checks for automatic routing.
Establish dedicated AWS Direct Connect connections into each Region and advertise more-specific BGP prefixes to move traffic to the standby Region when a failure is detected.
Place the ALBs behind a single Amazon CloudFront distribution and configure an origin group for automatic origin failover between Regions.
Create active-passive Amazon Route 53 failover records that point to the ALBs, configure health checks, and reduce the record TTL to 30 seconds.
Answer Description
AWS Global Accelerator offers two static anycast IP addresses and continuously probes regional endpoints. When an endpoint or Region becomes unhealthy, Global Accelerator removes it from service in well under one minute and immediately routes new connections to healthy endpoints, so the 30-second objective is met. Because the same static IP addresses are used before and after failover, client devices and corporate firewalls require no DNS updates or configuration changes.
Route 53 DNS failover still depends on recursive resolvers honoring a low TTL; many resolvers cache records longer than the configured TTL, so convergence can exceed the 30-second goal and exposes changing IP addresses. CloudFront origin failover is limited to GET/HEAD/OPTIONS requests and, by default, can spend up to 30 seconds trying the primary origin before switching, which may breach the RTO and does not provide static IPs. Direct Connect is a private network link; shifting traffic between Regions requires BGP route manipulation and cannot guarantee automated failover within the required timeframe. Therefore, deploying AWS Global Accelerator in front of the existing Application Load Balancers is the most reliable and operationally efficient solution.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is AWS Global Accelerator and how does it enable static IP addresses?
How do Global Accelerator health checks improve reliability during failover?
Why is Global Accelerator preferred over Route 53 for this scenario?
A production account (1111) hosts a business-critical Amazon RDS for PostgreSQL Multi-AZ DB instance in the us-east-1 Region. A new compliance mandate requires that:
- Encrypted backups must be retained for at least 35 days in a dedicated disaster-recovery (DR) account (2222) that belongs to the same AWS Organization.
- The DR backups must reside in the us-west-2 Region so they are isolated from the primary Region.
- Administrators in the DR account must be able to restore the database without assistance from the production account.
- The solution must rely on managed AWS capabilities and minimize ongoing manual work.
- A recovery point objective (RPO) of 24 hours is acceptable.
Which approach meets these requirements with the LEAST operational overhead?
In the production account, create an AWS Backup plan that performs daily snapshot backups of the RDS instance and copies them to a backup vault in the DR account in us-east-1. In the DR account, configure a second AWS Backup plan that automatically copies those snapshots to a backup vault in us-west-2 with a 35-day retention policy. Use customer-managed CMKs shared between the accounts for encryption.
Create a cross-Region read replica of the RDS instance in the DR account in us-west-2 and rely on the replica's automated backups, configured for 35-day retention, to satisfy the compliance mandate.
Enable cross-Region automated backups replication on the RDS instance from us-east-1 to us-west-2 and manually share each replicated backup with the DR account. Set the backup retention period to 35 days in us-west-2.
Use AWS Backup in the production account to define a single copy rule that sends daily RDS backups directly to a backup vault in the DR account in us-west-2 with a 35-day retention period.
Answer Description
AWS Backup can automate snapshot creation and retention while handling encryption keys. For Amazon RDS snapshots, a single copy job can be either cross-account or cross-Region, but not both. The simplest compliant design therefore uses two managed copy rules:
- A daily backup plan in the production account copies each snapshot to a backup vault that resides in the DR account but stays in the same Region (us-east-1).
- A second backup plan that runs in the DR account automatically copies the incoming snapshots to a vault in us-west-2 and sets the 35-day retention.
Because the snapshots end up in the DR account, operators there can restore the database independently. All steps are fully managed once the plans and customer-managed CMKs are in place, so operational effort is minimal.
The alternative proposals fail at least one requirement:
- Cross-Region automated backups replicate only inside the same account and still need manual snapshot sharing, increasing effort.
- A cross-Region read replica provides replication, not immutable backups, and retention is lost if the replica is removed.
- A single AWS Backup rule that is simultaneously cross-account and cross-Region is not supported for RDS snapshots.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
How do cross-account and cross-Region snapshot backups work in AWS Backup?
What role do customer-managed CMKs play in cross-account backups?
Why is a cross-Region read replica insufficient for compliance mandates involving immutable backups?
A solutions architect is troubleshooting a connectivity issue in a hybrid environment. An application running on an EC2 instance in a spoke VPC (10.20.0.0/16) cannot connect to an on-premises database server (192.168.10.50) on port 1433. The spoke VPC is connected to a central inspection VPC via an AWS Transit Gateway. The inspection VPC is connected to the on-premises data center via an AWS Direct Connect connection. All traffic from the spoke VPC to on-premises is routed through firewall appliances in the inspection VPC. On-premises network engineers have confirmed that their firewalls are not blocking the traffic. The architect needs to identify the component in the AWS network path that is blocking the connection. What is the MOST efficient first step to diagnose this issue?
Configure Route 53 Resolver Query Logging for the spoke VPC. Analyze the logs to ensure the on-premises database's hostname is correctly resolving to the IP address 192.168.10.50.
Enable VPC Flow Logs on the network interfaces for the application instance, the Transit Gateway attachment, and the inspection VPC firewall instances. Query the logs using Amazon Athena to find REJECT entries for traffic destined for 192.168.10.50 on port 1433.
Use the Route Analyzer feature in Transit Gateway Network Manager to analyze the path from the spoke VPC attachment to the Direct Connect gateway attachment, verifying that routes are correctly propagated.
Use VPC Reachability Analyzer to create and run an analysis with the application's EC2 instance network interface as the source and the on-premises database IP address (192.168.10.50) as the destination, specifying port 1433.
Answer Description
The correct answer is to use VPC Reachability Analyzer. This tool is specifically designed to perform static analysis of network paths between a source and a destination. It checks the configurations of route tables, security groups, network ACLs, and Transit Gateways without sending any live packets. This allows it to quickly identify the specific component that is blocking connectivity, making it the most efficient first step for this scenario.
- Using VPC Flow Logs and Amazon Athena is a valid troubleshooting method, but it is less efficient. It requires enabling logs, waiting for traffic to be captured, and then performing complex queries on potentially large datasets to find the problem. This is more time-consuming than using the purpose-built Reachability Analyzer.
- The Route Analyzer feature in Transit Gateway Network Manager is not the best tool for this task because it only analyzes routes within the Transit Gateway route tables. It does not analyze VPC route tables, security group rules, or network ACLs, which are common sources of connectivity problems.
- Configuring Route 53 Resolver Query Logging would be appropriate if the problem were related to DNS name resolution. However, the scenario describes a failure to connect to a specific IP address, which points to a network path issue, not a DNS issue.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
How does the VPC Reachability Analyzer work?
What is the difference between VPC Reachability Analyzer and VPC Flow Logs?
Why doesn’t Route Analyzer in Transit Gateway Network Manager identify all connectivity issues?
A financial services company runs a latency-sensitive payment-processing workload in the us-east-1 Region. The workload uses an Amazon ECS cluster (EC2 launch type) with stateless microservices behind an Application Load Balancer, an Amazon Aurora MySQL DB cluster, and an Amazon ElastiCache for Redis cluster that stores session data.
Compliance rules require a recovery point objective (RPO) of 1 minute and a recovery time objective (RTO) of 5 minutes for a complete Regional failure. Management insists on a solution that is less costly than an active-active multi-Region deployment but still meets the objectives.
Which solution meets these requirements?
Implement a pilot-light strategy that replicates only the Aurora database to another Region, stores container images in Amazon ECR with cross-Region replication, and creates the ECS cluster, Redis nodes, and load balancer with AWS CloudFormation when a disaster is declared.
Deploy a warm-standby environment in a second Region: add an Aurora global database secondary cluster and a Redis Global Datastore replica, run a scaled-down copy of the ECS services (one task per service) behind an Application Load Balancer, enable cross-Region image and data replication, and use Amazon Route 53 failover routing to switch traffic when health checks fail.
Deploy a fully active-active architecture in two Regions with separate Aurora writer clusters, application-level replication for the Redis data, full-sized ECS services, and weighted Amazon Route 53 routing between the Regions.
Use AWS Elastic Disaster Recovery (AWS DRS) to continuously replicate the ECS instances and Redis nodes to a second Region, convert the Aurora cluster to an Aurora global database, and rely on AWS DRS orchestration to launch all recovered resources after a disaster.
Answer Description
A warm-standby pattern keeps a scaled-down but fully functional copy of the entire stack running in a second Region. Aurora Global Database and ElastiCache for Redis Global Datastore replicate data with typical latencies of less than 1 second, so the RPO is well under 1 minute. Because the ECS services, ALB, and other infrastructure are already deployed (albeit at minimal size), the environment only needs to scale out and Route 53 needs to redirect traffic, which can be completed within the 5-minute RTO.
With AWS Elastic Disaster Recovery, compute resources are provisioned only after failover; its typical RTO is 5-20 minutes, so it cannot guarantee a 5-minute target. A pilot-light strategy runs no application tier at all, so provisioning the full stack usually takes tens of minutes, exceeding the RTO. An active-active design would meet the objectives but costs significantly more because both Regions run at full production scale at all times. Therefore, the warm-standby solution is the most cost-effective option that satisfies both the 1-minute RPO and 5-minute RTO.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a warm-standby pattern in AWS architectures?
How does Aurora Global Database ensure low replication latency?
What is the difference between a warm-standby and an active-active architecture?
A company operates a Kubernetes-based product-catalog microservice that runs in Amazon EKS clusters deployed in us-east-1, eu-west-1, and ap-southeast-2. The service performs read-only SQL queries against a single Amazon Aurora MySQL DB cluster located in us-east-1. During global marketing events, 90 % of traffic originates outside the primary Region and p99 end-to-end latency in Europe spikes above 400 ms. New performance objectives state:
- End-user latency must remain below 100 ms at the 99th percentile worldwide.
- Read traffic is expected to grow 5× within 12 months.
- Catalog updates occur only in us-east-1 and may be eventually consistent everywhere else within 2 minutes.
- The solution must keep code changes and ongoing cost to a minimum.
Which combination of actions best meets these objectives?
Create an Amazon ElastiCache for Redis Global Datastore: deploy a primary cluster in us-east-1 and read-only replica clusters in eu-west-1 and ap-southeast-2, and implement a write-through pattern so the service first reads from the Region-local Redis endpoint and updates the primary cluster after each catalog change.
Migrate the catalog to Amazon DynamoDB and place a DynamoDB Accelerator (DAX) cluster in each Region, routing reads to the local DAX endpoint and writes to DynamoDB in us-east-1.
Implement Amazon Aurora Global Database by adding read-only secondary clusters in eu-west-1 and ap-southeast-2, and modify the service so that read queries are routed to the Region-local reader endpoint.
Put the API behind an Amazon CloudFront distribution and use Lambda@Edge to cache GET responses for 120 seconds at edge locations worldwide while leaving the Aurora database unchanged.
Answer Description
ElastiCache Global Datastore replicates Redis data to up to two secondary Regions with typical cross-Region lag of less than 1 s, enabling sub-millisecond local reads while writes occur only in the primary Region. Using a write-through pattern keeps the cache hot and ensures replicas converge well inside the 2-minute consistency window. Because 90 % of requests are now served from in-memory caches close to users, p99 latency comfortably drops below 100 ms and Aurora writer CPU load is relieved without expensive multi-Region database clusters. Aurora Global Database would meet latency but requires full secondary clusters, dramatically increasing cost; migrating to DynamoDB with DAX demands a data-model change and additional engineering effort; edge caching with CloudFront cannot guarantee a warm cache globally, adds invalidation complexity, and leaves the database hot. Therefore, deploying an ElastiCache for Redis Global Datastore and adopting a write-through caching strategy is the most cost-effective way to satisfy all stated requirements.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is ElastiCache for Redis Global Datastore?
How does a write-through caching pattern work?
Why is Aurora Global Database not a cost-effective solution here?
A financial services company utilizes a multi-account AWS environment with a hub-and-spoke network architecture centered around an AWS Transit Gateway. The security team is mandated to perform deep packet inspection (DPI) on all east-west traffic between spoke VPCs. The inspection must be conducted by a fleet of third-party intrusion detection system (IDS) appliances deployed on EC2 instances within a dedicated 'inspection' VPC. The solution must be highly scalable, have minimal performance impact on application workloads, and centralize the inspection tooling. Which approach should a solutions architect recommend to meet these requirements?
Configure VPC Flow Logs for all traffic in the spoke VPCs. Stream the logs to a central Amazon S3 bucket and use Amazon Athena for analysis.
Deploy AWS Network Firewall in the inspection VPC. Configure the Transit Gateway to route all inter-VPC traffic through the Network Firewall endpoints for inspection.
In the inspection VPC, configure a Gateway Load Balancer (GWLB) with the IDS appliance fleet as a target group. Create GWLB Endpoints in each spoke VPC and modify route tables to direct all traffic through the GWLB.
Configure VPC Traffic Mirroring on the source Elastic Network Interfaces (ENIs) in the spoke VPCs. Set the mirror target to a Network Load Balancer (NLB) in the inspection VPC that fronts the IDS appliance fleet.
Answer Description
The correct answer is to configure VPC Traffic Mirroring on the relevant ENIs and set the target to a Network Load Balancer (NLB) in the inspection VPC. VPC Traffic Mirroring is designed to copy network traffic from an Elastic Network Interface (ENI) and forward it to a target for out-of-band inspection. This is the ideal solution for an Intrusion Detection System (IDS), which passively analyzes traffic without being in the direct path. Using a Network Load Balancer as the target allows the mirrored traffic to be distributed across a fleet of IDS appliances, ensuring scalability and high availability. This approach has minimal performance impact because it duplicates the traffic rather than routing it through the appliances, avoiding any added latency or a potential single point of failure in the production traffic path.
- Incorrect: Enabling VPC Flow Logs is incorrect because Flow Logs capture metadata about IP traffic (e.g., source/destination IPs, ports, protocol) but do not capture the actual packet payloads. Deep packet inspection (DPI) requires analyzing the full packet content, which is not available in Flow Logs.
- Incorrect: Deploying AWS Network Firewall is incorrect because the requirement is to use a specific fleet of third-party IDS appliances. AWS Network Firewall is a managed AWS service and would not fulfill this explicit requirement.
- Incorrect: Using a Gateway Load Balancer (GWLB) in an in-line configuration is incorrect for this use case. A GWLB is designed to transparently insert appliances into the traffic path for in-line inspection, which is typical for an Intrusion Prevention System (IPS). Since the requirement is for a passive IDS, routing all traffic through the appliances would add unnecessary latency and complexity. The more appropriate and less impactful solution is to copy the traffic using Traffic Mirroring.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is VPC Traffic Mirroring and how does it work?
How does a Network Load Balancer (NLB) support scalability for IDS appliances?
Why is VPC Flow Logs insufficient for deep packet inspection (DPI)?
A financial services company is modernizing a monolithic on-premises application by refactoring it into containerized microservices to be deployed on Amazon ECS. A key security requirement is that all east-west traffic (service-to-service communication) between the microservices must be routed through a fleet of third-party network security appliances for deep packet inspection. The company wants to use AWS Fargate to minimize infrastructure management overhead. Which architectural challenge must a solutions architect address to meet these requirements when using the Fargate launch type?
AWS Fargate tasks cannot be assigned security groups, which prevents the implementation of the network traffic filtering rules required by the security appliances.
The use of an Application Load Balancer (ALB) for Fargate services encrypts all east-west traffic, which prevents network security appliances from performing deep packet inspection.
Fargate tasks use the
awsvpc
network mode, giving each task a dedicated ENI within a subnet, which complicates routing intra-VPC traffic to a centralized inspection appliance.Fargate does not support the
host
network mode, which is required to bind the security appliances directly to the same underlying instance as the application containers.
Answer Description
The correct answer identifies the fundamental networking challenge with inspecting east-west traffic for Fargate tasks. Fargate tasks are required to use the awsvpc
network mode, where each task is assigned its own Elastic Network Interface (ENI) and a private IP address from the VPC's subnet. This means that when one microservice communicates with another in the same VPC, the traffic flows directly between their ENIs within the subnet. Standard VPC routing does not intercept this intra-subnet traffic, making it difficult to force it through a centralized inspection point. To solve this, an architect must implement an advanced networking pattern. Common solutions include using an AWS Transit Gateway to route traffic between different VPCs (or different subnets) to a dedicated 'inspection VPC' where the security appliances are hosted. Another modern approach involves using a service mesh that can control and redirect traffic at the application layer.
Incorrect answers are:
- Fargate tasks absolutely can and do use security groups. A security group is associated with each task's ENI, providing stateful, instance-level firewall capabilities. Stating that they cannot be assigned is factually incorrect.
- While it's true that Fargate does not support
host
network mode, this mode is irrelevant to the problem of inspecting traffic between separate tasks.Host
mode ties a container's networking directly to the underlying host's network stack, which is a concept antithetical to the Fargate serverless model. - An Application Load Balancer (ALB) primarily manages north-south (ingress) traffic from clients to the services. It does not typically handle direct east-west (service-to-service) communication. Even if it did, the challenge is routing the traffic for inspection, not the encryption itself.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is the `awsvpc` network mode in AWS Fargate?
How can a Transit Gateway help with routing intra-VPC traffic?
What role does a service mesh play in traffic control for microservices?
A large enterprise uses AWS Organizations to manage dozens of member accounts. The finance team has reported a significant, unexpected increase in costs, but the high-level views in AWS Cost Explorer are insufficient for identifying the root cause. The company has configured AWS Cost and Usage Reports (CUR) to be delivered hourly in Apache Parquet format to an Amazon S3 bucket in the management account.
A solutions architect needs to implement a scalable and cost-effective solution to perform complex, ad-hoc SQL queries on this CUR data. The goal is to identify specific resources and API operations contributing to the cost increase across the entire organization.
Which approach will achieve this with the LEAST operational overhead?
Set up an AWS Glue crawler to run on the S3 bucket containing the CUR data. Configure the crawler to populate the AWS Glue Data Catalog. Use Amazon Athena to run standard SQL queries against the table created by the crawler.
Create an Amazon EMR cluster configured with Apache Spark. Develop Spark SQL jobs to load the Parquet files from Amazon S3 into data frames and run queries from a Zeppelin notebook attached to the cluster.
Develop an AWS Lambda function triggered by Amazon S3 events when new CUR files are delivered. The function will parse the Parquet files and load the data into a provisioned Amazon RDS for PostgreSQL database for querying.
Use Amazon S3 Select to query individual CUR Parquet files directly in the S3 bucket. Develop a script that iterates through all CUR files for the desired time range, executes S3 Select queries on each, and aggregates the results in the client application.
Answer Description
The correct answer is to use AWS Glue and Amazon Athena. This combination is the AWS-recommended, serverless, and most operationally efficient method for querying AWS Cost and Usage Reports (CUR). When CUR is configured for Athena integration, AWS delivers the data in the optimal Apache Parquet format and provides a CloudFormation template to automatically create the necessary AWS Glue crawler and Data Catalog table. The Glue crawler automatically discovers the schema and partitions, making the data available for querying via Athena with standard SQL. This approach requires minimal setup and no infrastructure to manage, directly addressing the requirement for the least operational overhead.
Using Amazon EMR is incorrect because it introduces significant operational overhead. While EMR is a powerful big data platform, it requires provisioning and managing a cluster of EC2 instances, which is more complex and costly than the serverless Athena model for this use case.
Using AWS Lambda to load data into an Amazon RDS database is incorrect because it involves substantial development and maintenance effort. The architect would need to write and maintain code for parsing, data transformation, and loading. Furthermore, RDS is a transactional database and less cost-effective for the large-scale analytical queries typically run against CUR data compared to Athena.
Using Amazon S3 Select is incorrect because it is designed to query data within a single S3 object. While useful for simple filtering, it is not suitable for running complex, ad-hoc analytical queries that span the thousands of files that make up a complete CUR dataset. This approach would require a complex client-side application to orchestrate queries and aggregate results, creating high operational overhead.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is AWS Glue, and how does it work with Athena?
Why is Apache Parquet format recommended for CUR data?
How does Amazon Athena enable serverless querying for CUR data?
An enterprise uses AWS Organizations to manage more than 500 AWS accounts. The security team has created a dedicated security-tooling account in the us-east-1 Region and must meet the following requirements:
- AWS Security Hub must be enabled in every current and future account in all Regions.
- All findings must be visible only in the security-tooling account.
- No other account may designate itself as the Security Hub delegated administrator. The solution must follow the principle of least privilege and require minimal ongoing maintenance. Which approach BEST meets these requirements?
Use a CloudFormation StackSet to deploy a template that enables Security Hub and its default standards in every current account and Region; configure the StackSet for automatic deployment to new accounts.
From the organization management account, run securityhub enable-organization-admin-account in each enabled Region to set the security-tooling account as delegated administrator. In the delegated administrator account, run securityhub update-organization-configuration with AutoEnable=true and enable the default standards for all Regions. Attach an SCP at the organization root that denies securityhub:EnableOrganizationAdminAccount to every account except the management account.
Enable Security Hub only in the security-tooling account and create a cross-Region finding aggregator. In each member account, add an EventBridge rule that forwards Security Hub findings to the aggregator.
Enable Security Hub through AWS Control Tower guardrails when the landing zone is set up. Rely on the guardrails to enable Security Hub in new accounts and prevent changes to the delegated administrator.
Answer Description
Running enable-organization-admin-account from the organization management account designates the security-tooling account as the Security Hub delegated administrator in the current Region; repeating the call (or using central-configuration) in every active Region ensures that the same account is admin everywhere. In the delegated administrator account, update-organization-configuration with AutoEnable=true (and the default standards setting) automatically adds every existing and future member account as a Security Hub member in every Region, so findings flow to the delegated administrator without additional setup. Finally, an SCP applied at the root that explicitly denies securityhub:EnableOrganizationAdminAccount (except for the management account) blocks any other account from changing the delegated administrator. This combination satisfies all three requirements with a single point of administration and no per-account maintenance. The other options either rely on manual EventBridge forwarding, require per-account StackSet deployments, or depend on Control Tower guardrails that neither auto-enable Security Hub in every Region nor prevent other accounts from changing the administrator.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is the purpose of the enable-organization-admin-account command in this solution?
What does AutoEnable=true do in the update-organization-configuration command?
How does the Service Control Policy (SCP) prevent unauthorized changes to the delegated administrator?
A financial services company is modernizing a component of its legacy market data processing application. This component, currently hosted on a fleet of over-provisioned Amazon EC2 instances, handles unpredictable, high-throughput transaction bursts. A critical requirement is to maintain processing latency under 100ms per invocation to meet SLAs. The primary goals are to reduce costs associated with idle capacity and improve scalability. The development team has refactored the component into an AWS Lambda function. Which configuration should a Solutions Architect recommend to meet these requirements MOST effectively?
Configure Provisioned Concurrency for the function, setting the number of concurrent executions based on anticipated peak load.
Increase the function's memory allocation to the maximum and rely on on-demand scaling.
Create a scheduled Amazon EventBridge rule that invokes the function every minute to keep it warm.
Place the function in a target group behind an Application Load Balancer and configure aggressive health checks to trigger invocations.
Answer Description
The correct answer is to configure Provisioned Concurrency for the Lambda function. This feature is specifically designed for applications that require predictable, low-latency performance by keeping a specified number of execution environments initialized and ready to respond in double-digit milliseconds. This directly addresses the two main requirements: eliminating cold start latency to meet the sub-100ms SLA during sudden bursts and optimizing costs compared to running a constantly active EC2 fleet.
- Increasing the function's memory allocation does provide more CPU power, which can reduce both cold start duration and execution time. However, it does not eliminate the cold start itself. The first request to an idle function will still experience initialization latency, which could violate the strict SLA. This makes it a helpful optimization but not the primary solution for guaranteed low latency.
- Creating a scheduled Amazon EventBridge rule to invoke the function every few minutes is a technique known as a "pinger" or "warmer". This is an outdated practice that is now considered an anti-pattern. It typically only keeps a single execution environment warm, which is insufficient for handling high-throughput bursts of traffic, and it is less reliable and efficient than Provisioned Concurrency.
- While an Application Load Balancer (ALB) can use a Lambda function as a target, using its health check mechanism to keep the function warm is not a recommended or effective pattern for this use case. ALB health checks are designed to monitor target health and route traffic accordingly, not to manage Lambda execution environments for performance. Provisioned Concurrency is the purpose-built AWS feature for this scenario.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Provisioned Concurrency in AWS Lambda?
Why is the 'pinger' or 'warmer' approach considered an anti-pattern?
How does memory allocation impact AWS Lambda performance?
A financial analytics company runs a platform on EC2 instances within private subnets, distributed across multiple Availability Zones in the us-east-1
region. The application frequently downloads terabytes of data from a critical third-party data provider's API, which is hosted outside of AWS. To facilitate this, a NAT Gateway is deployed in each Availability Zone. A cost analysis reveals that the "NAT Gateway - Data Processed" fees are a major operational expense. The company wants to drastically reduce these data transfer costs while preserving the high-availability, multi-AZ posture of the application. The third-party provider has recently announced that they offer an endpoint service powered by AWS PrivateLink in the us-east-1
region.
What is the MOST cost-effective solution to reduce these charges?
Create a VPC interface endpoint for the third-party's endpoint service within the company's VPC. Reconfigure the application to use the endpoint's DNS name to access the API.
Consolidate to a single NAT Gateway in one Availability Zone and update the VPC route tables to direct all outbound traffic through it.
Use AWS Direct Connect to establish a dedicated connection to the
us-east-1
region and route the API requests through this connection.Set up a fleet of caching proxy servers on EC2 instances in public subnets. Direct the application's data requests through this caching layer.
Answer Description
The correct answer is to create a VPC interface endpoint for the third-party's service. By using AWS PrivateLink, traffic from the EC2 instances to the third-party API will be routed through the interface endpoint over the AWS private network, completely bypassing the NAT Gateways. This directly eliminates the "NAT Gateway - Data Processed" charge for this traffic, which is the primary cost driver identified in the scenario. While the VPC endpoint has its own hourly and data processing fees, the data processing fee for an interface endpoint is significantly lower ($0.01 per GB) than that of a NAT Gateway ($0.045 per GB), resulting in substantial savings.
- Consolidating to a single NAT Gateway would actually increase costs for a multi-AZ application. Traffic originating from instances in other Availability Zones would first incur an inter-AZ data transfer fee ($0.01/GB) to reach the NAT Gateway, and then the standard NAT Gateway processing fee would still apply on top of that. This approach also introduces a single point of failure, which contradicts the high-availability requirement.
- A caching proxy layer only reduces costs for redundant data requests. Any request for new or unique data would still need to be fetched from the internet, incurring data transfer processing fees. Since AWS PrivateLink eliminates the NAT Gateway processing cost for all traffic to the provider, it is a more comprehensive and cost-effective solution.
- AWS Direct Connect is a service used to establish a dedicated private network connection from an on-premises data center to AWS. It cannot be used to connect resources within a VPC to a third-party service hosted on the public internet.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is an AWS VPC interface endpoint?
How do the costs of NAT Gateway and PrivateLink compare?
Why is using AWS PrivateLink better than consolidating to a single NAT Gateway for a multi-AZ setup?
A global media company operates a real-time video-streaming service with backend infrastructure deployed on EC2 instances behind Network Load Balancers (NLBs) in the us-east-1
, eu-central-1
, and ap-southeast-2
Regions. The service uses a custom TCP-based protocol for streaming. Users connect to the geographically closest regional endpoint by means of DNS-based routing.
The company is experiencing two major issues:
- Some corporate clients have strict firewall egress rules and struggle to whitelist the multiple static public IP addresses that each regional NLB exposes (one per Availability Zone and Region).
- During a recent service impairment in one Region, users were not automatically routed to a healthy Region, resulting in a significant outage for a large user segment.
The company wants to implement a solution that provides static entry points for the application and improves availability with fast, automatic cross-Region failover. Which solution best meets these requirements?
Establish AWS Direct Connect connections to each AWS Region and use a Direct Connect gateway for inter-Region failover.
Deploy an AWS Global Accelerator and configure each regional NLB as an endpoint in its respective endpoint group.
Configure an Amazon CloudFront distribution with the regional NLBs as custom origins and use a CloudFront Function to manage failover between origins.
Use Amazon Route 53 with a combination of latency-based routing and failover routing policies. Configure health checks for each regional NLB.
Answer Description
Deploying AWS Global Accelerator is the best solution. Global Accelerator allocates two static anycast IP addresses that serve as fixed entry points, simplifying firewall whitelisting for corporate clients. It supports both TCP and UDP, so it works with the company's custom TCP-based streaming protocol. Because Global Accelerator operates at the network layer and continuously health-checks all registered regional endpoints, traffic is automatically and almost immediately redirected to the nearest healthy Region-without waiting for DNS TTLs to expire.
Incorrect answers:
- Amazon CloudFront is optimized for HTTP/HTTPS and WebSocket traffic and cannot proxy an arbitrary TCP-based streaming protocol.
- Amazon Route 53 can perform DNS-based latency or failover routing, but failover depends on DNS TTLs and client caches, and it still exposes many NLB IP addresses that clients must whitelist.
- AWS Direct Connect is a private network service for on-premises connectivity and does not provide global static IPs or cross-Region failover for public internet users.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is AWS Global Accelerator and why is it suited to this scenario?
How does AWS Global Accelerator ensure faster failover compared to Route 53?
Why is Amazon CloudFront not a good fit for the custom TCP-based streaming protocol in this scenario?
Your company operates several AWS accounts managed with AWS Organizations. In the shared dev account, application teams need to create and maintain IAM roles for their Lambda functions and ECS tasks. The security team has produced a guardrail policy that grants only the following permissions:
- Read access to two designated S3 buckets
- Write access to one DynamoDB table
Developers must be allowed to self-service creation and updates of IAM roles only if the resulting roles never exceed the permissions in the guardrail policy, and the security team does not want to manually review each policy or role that is created.
Which solution BEST enforces the principle of least privilege while meeting these requirements?
Require developers to tag their roles with Environment=Dev and apply an ABAC policy that allows actions only when the principal and resource share the same tag.
Enable the AWS Config managed rule IAM_POLICY_NO_STATEMENTS_WITH_ADMIN_ACCESS and use an EventBridge rule to invoke a Lambda function that automatically deletes any policy marked NON_COMPLIANT.
Create a customer-managed policy that contains the approved S3 and DynamoDB permissions and designate it as a permissions boundary. Grant the developer group iam:CreateRole, iam:PutRolePolicy, and iam:AttachRolePolicy permissions only when the iam:PermissionsBoundary condition key equals the boundary policy's ARN.
Attach a service control policy to the dev account that denies iam:CreateRole for all principals except the security team, and have the security team create roles for developers on request.
Answer Description
Using an IAM permissions boundary provides a built-in, automatic mechanism to restrict the maximum permissions that any developer-created role can receive. By requiring the iam:PermissionsBoundary condition key to match the ARN of the security team's guardrail policy, developers can freely call iam:CreateRole, iam:PutRolePolicy, or iam:AttachRolePolicy, but the resulting roles can never exceed the boundary's scope. This satisfies least-privilege goals without any manual review.
A service control policy that blocks iam:CreateRole (choice B) forces the security team to provision every role themselves, preventing developer self-service. AWS Config remediation (choice C) detects problems after the fact and allows overly broad roles to exist until the rule is evaluated and remediated, violating the requirement for immediate enforcement. ABAC based only on Environment tags (choice D) does not stop a developer from attaching AdministratorAccess or other excessive permissions to a role; it simply constrains which resources the role may act on.
Therefore, configuring and enforcing an IAM permissions boundary is the most secure and automated approach.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is an IAM permissions boundary, and how does it work?
How does a permissions boundary differ from an SCP (Service Control Policy)?
What is the purpose of the iam:PermissionsBoundary condition key?
A fintech company runs its order-management microservices as stateless Amazon ECS services fronted by an Application Load Balancer in us-east-1. Transactional data is stored in an Amazon Aurora MySQL cluster in the same Region. The company's business-continuity policy requires that the application must continue to operate if an entire AWS Region becomes unavailable, with a recovery time objective (RTO) under 60 seconds and a recovery point objective (RPO) under 30 seconds. Additionally, failover must be fully automated with no human intervention, and the solution should avoid the cost of running a fully active/active stack in every Region.
Which architecture meets these requirements MOST cost-effectively?
Enable Aurora Multi-AZ with two readable standbys in us-east-1, replicate automated snapshots to us-west-2, and restore the snapshots into a new Aurora cluster before redeploying the ECS services after an outage.
Create a cross-Region read replica of the Aurora cluster in us-west-2 and configure an Amazon CloudWatch alarm that invokes an AWS Lambda function to promote the replica and update Route 53 DNS records when health checks fail.
Migrate the relational schema to Amazon DynamoDB global tables in both Regions and use AWS Database Migration Service continuous replication to keep the tables synchronized; direct the application to the secondary Region during a failure.
Convert the Aurora cluster to an Aurora Global Database and add a secondary Aurora cluster in us-west-2. Use Amazon Route 53 Application Recovery Controller routing controls to automatically promote the secondary cluster and shift traffic when a regional failure is detected.
Answer Description
Aurora Global Database replicates storage to secondary Regions with a typical replication lag of less than one second, meeting the 30-second RPO. It also supports managed cross-Region failover that typically completes in well under a minute, satisfying the 60-second RTO. By combining it with Route 53 Application Recovery Controller routing controls, failover and DNS redirection can be triggered automatically without operators, achieving the operational-excellence goal.
A cross-Region read replica (distractor) offers a similar RPO but requires manual promotion via control-plane APIs; documented failover times are usually several minutes, so it misses the RTO. Multi-AZ deployments with replicated snapshots (distractor) provide only in-Region HA; restoring from snapshots after a regional outage far exceeds both RTO and RPO. Migrating to DynamoDB global tables (distractor) could achieve the objectives, but re-architecting the relational workload and running fully active resources in multiple Regions would significantly increase cost and complexity, violating the cost-effectiveness constraint.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is an Aurora Global Database, and how does it assist with RTO and RPO?
How does the Route 53 Application Recovery Controller help automate failover?
Why isn't a cross-Region read replica sufficient for this scenario?
During a modernization effort, an on-premises e-commerce platform is being refactored into microservices on AWS. The new Order service must publish an event that triggers the Payment service whenever a customer completes checkout. Requirements for the messaging layer are as follows: the Payment service must never process the same order more than once, messages that relate to the same order must be delivered in the exact sequence in which they were generated, holiday sales can create burst traffic of tens of thousands of orders per second so the solution must scale automatically without manual sharding, and operations staff want to keep queue-management overhead to a minimum.
Which solution meets these requirements?
Publish events to an Amazon SQS FIFO queue with high-throughput mode enabled. Set the MessageGroupId to the order ID and turn on content-based deduplication.
Publish events to an Amazon SNS FIFO topic. Configure an Amazon SQS FIFO queue subscription for the Payment service.
Publish events to an Amazon SQS FIFO queue that uses a single MessageGroupId value such as "payment" for every message. Configure the consumer to process messages sequentially.
Publish events to an Amazon SQS standard queue. Include the order ID in a message attribute so the consumer can ignore duplicate messages that occasionally appear.
Answer Description
An Amazon SQS FIFO queue guarantees strict message ordering and offers exactly-once processing through message deduplication. Enabling high-throughput mode allows the queue to scale automatically to many thousands of messages per second without manual sharding. Using the order ID as the MessageGroupId confines ordering to each individual order while permitting parallel processing across different orders, and turning on content-based deduplication prevents accidental re-submission of the same order from creating duplicates.
A standard queue does not guarantee ordering and can deliver duplicates, so the consumer would still have to guard against both issues. A FIFO queue that uses a single MessageGroupId preserves order and eliminates duplicates but limits throughput to a single message group, which cannot handle tens of thousands of orders per second. An SNS FIFO topic with an SQS FIFO queue subscription is a valid pattern for ordered, exactly-once delivery, but it adds an unnecessary additional service to manage and introduces complexity when an SQS FIFO queue alone can solve the problem.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Amazon SQS FIFO queue's MessageGroupId, and how does it work?
What is high-throughput mode in Amazon SQS FIFO queues?
What is content-based deduplication in Amazon SQS FIFO queues?
A financial-services company runs Monte Carlo risk simulations overnight in a single AWS Region. The job launches up to 1,000 Linux instances for 6 hours. At least 100 instances must always be running to meet the service-level agreement, but the rest of the workload can tolerate Spot interruptions. The chief architect wants to minimize compute cost, obtain Spot capacity from pools that have the lowest interruption risk, and automatically replace any Spot Instances that receive a rebalance recommendation.
Which solution meets all of these requirements?
Use AWS Batch with a single Spot compute environment that specifies the SPOT_CAPACITY_OPTIMIZED allocation strategy and a desired vCPU count of 1,000. Batch will automatically provision the required capacity and handle any Spot interruptions.
Request an EC2 Spot Fleet for 1,000 instances across all Availability Zones using the lowest-price allocation strategy. Set OnDemandTargetCapacity to 100 and rely on a Lambda function to relaunch tasks if Spot Instances are interrupted.
Deploy the workload in an EC2 Auto Scaling group that launches only Spot Instances with the lowest-price allocation strategy. Configure Capacity Rebalancing so the group starts new Spot Instances when interruptions occur.
Create an EC2 Auto Scaling group that uses a launch template for the simulation instances. Configure a mixed instances policy with OnDemandBaseCapacity set to 100, OnDemandPercentageAboveBaseCapacity set to 0, and SpotAllocationStrategy set to capacity-optimized. Enable Capacity Rebalancing for the group.
Answer Description
A mixed-instances EC2 Auto Scaling group can combine On-Demand and Spot capacity in one fleet. Setting OnDemandBaseCapacity to 100 guarantees that the first 100 instances that launch-and any time the group scales-are On-Demand, so the SLA is always met. Setting OnDemandPercentageAboveBaseCapacity to 0 forces every additional instance to be Spot, maximizing cost savings. Using the capacity-optimized SpotAllocationStrategy directs the request to pools with the greatest spare capacity, which have the lowest probability of interruption. Enabling Capacity Rebalancing instructs the Auto Scaling group to launch replacement Spot Instances proactively when Amazon EC2 sends a rebalance recommendation, maintaining capacity even during interruptions. The other options either do not guarantee the 100-instance baseline, choose a lowest-price strategy with a higher interruption risk, or rely on external scripts rather than automated rebalance replacement.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is the purpose of the 'capacity-optimized' SpotAllocationStrategy in EC2 Auto Scaling?
How does enabling Capacity Rebalancing in an EC2 Auto Scaling group improve stability?
What is the difference between OnDemandBaseCapacity and OnDemandPercentageAboveBaseCapacity in a mixed instances policy?
A solutions architect is tasked with optimizing a large fleet of m5.4xlarge
EC2 instances running a legacy, monolithic, stateful .NET Framework application on Windows Server. The application serves critical business functions and experiences unpredictable, spiky traffic patterns. An analysis using Amazon CloudWatch shows that the average CPU utilization is consistently below 20%, but the memory utilization is consistently high, between 80-90%. The company wants to significantly reduce costs without compromising performance or availability during peak loads. Any proposed solution must provide a systematic way to apply recommendations across a multi-account organization.
Which strategy should the solutions architect recommend to meet these requirements?
Refactor the monolithic application into containerized microservices on .NET Core and deploy it to an Amazon EKS cluster with Cluster Autoscaler.
Use AWS Compute Optimizer to get data-driven recommendations and begin migrating the fleet to an appropriate memory-optimized (R-series) instance type.
Use AWS Cost Explorer Rightsizing recommendations to identify underutilized instances and manually downsize them to
m5.2xlarge
instances.Implement an Amazon EC2 Auto Scaling group with a step scaling policy based on CPU utilization, setting the minimum size to a smaller instance like
m5.large
.
Answer Description
The correct answer is to use AWS Compute Optimizer to generate recommendations and migrate the instances to a memory-optimized instance family. The workload is clearly memory-bound, not CPU-bound, as indicated by the high memory utilization and low CPU utilization. The m5
instance family is general-purpose. A memory-optimized family, such as the R-series, is better suited for this workload. AWS Compute Optimizer is the ideal tool for this scenario because it analyzes historical utilization data (including memory, if the CloudWatch agent is configured) and recommends optimal instance types, often suggesting changes across instance families (e.g., M to R) and to Graviton-based instances for further cost savings. It also supports cross-account recommendations, which fulfills the requirement for a systematic approach across the organization.
- Using Cost Explorer Rightsizing recommendations to downsize to
m5.2xlarge
is incorrect because it keeps the instance within the same general-purpose family, failing to address the core issue of the workload being memory-bound. While Cost Explorer provides rightsizing recommendations, Compute Optimizer offers more detailed performance-oriented analysis and is the superior tool for this specific optimization task. - Implementing an Auto Scaling group is not suitable for a stateful, monolithic application without significant re-architecture. Furthermore, a scaling policy based on CPU utilization would be ineffective since the application's bottleneck is memory, not CPU.
- Refactoring the application to a containerized microservices architecture on Amazon EKS is a major modernization project, not a rightsizing strategy. While it could be a valid long-term goal, it does not address the immediate requirement to optimize the existing fleet for cost and performance.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is AWS Compute Optimizer?
Why are R-series instances better for memory-bound workloads?
What data does AWS Compute Optimizer analyze to provide recommendations?
A financial services company operates a critical, multi-tier application in the us-east-1
Region. The application consists of a fleet of EC2 instances in an Auto Scaling group behind an Application Load Balancer, an Amazon RDS for PostgreSQL Multi-AZ database, and an Amazon ElastiCache for Redis cluster.
To meet business continuity requirements, the company must implement a disaster recovery (DR) strategy in the us-west-2
Region with a Recovery Time Objective (RTO) of 15 minutes and a Recovery Point Objective (RPO) of 1 minute. The company has chosen a warm standby approach to balance recovery time with cost.
Which of the following designs BEST implements a warm standby strategy that meets these requirements?
In
us-west-2
, deploy a fully scaled-out duplicate of the production environment, including the Auto Scaling group, RDS database, and ElastiCache cluster. Use an Amazon Route 53 latency-based routing policy to distribute traffic betweenus-east-1
andus-west-2
.In
us-west-2
, create a cross-region read replica for the Amazon RDS for PostgreSQL database. Configure an Auto Scaling group with a minimum and desired capacity of one small EC2 instance. Provision a small, single-node ElastiCache for Redis cluster. For failover, promote the RDS read replica, scale up the Auto Scaling group, and update an Amazon Route 53 failover record.In
us-west-2
, create a cross-region read replica for the RDS for PostgreSQL database. Configure an Auto Scaling group with a minimum and desired capacity of zero. Configure Amazon ElastiCache for Redis with Global Datastore to replicate the cache. Use Amazon Route 53 to fail over traffic.In
us-west-2
, replicate application AMIs and Amazon RDS snapshots. In a disaster, deploy a new AWS CloudFormation stack using the replicated AMI and restore the database from the latest snapshot. Use Amazon Route 53 to redirect traffic to the new stack.
Answer Description
The correct answer describes a classic warm standby architecture. In the DR region (us-west-2
), a scaled-down but functional version of the application is running. This includes using an Amazon RDS cross-region read replica for asynchronous data replication, which is suitable for meeting a low RPO like 1 minute. A minimal Auto Scaling group (with a desired capacity of one) keeps the application tier running and ready to be scaled up quickly, which is a key characteristic of warm standby and essential for a 15-minute RTO. A small, single-node ElastiCache cluster is also provisioned to complete the minimal running environment. Upon failover, the RDS replica is promoted, the Auto Scaling group is scaled out, and Amazon Route 53 redirects traffic, all of which can be accomplished within the 15-minute RTO.
- The option suggesting deploying a new stack from AMIs and RDS snapshots describes a backup and restore strategy. This approach would have a much higher RTO and RPO, failing to meet the specified 15-minute RTO and 1-minute RPO.
- The option suggesting a fully scaled environment with latency-based routing describes a hot standby (or multi-site active/active) strategy. While it would meet the RTO/RPO, it is far more expensive than the required warm standby approach.
- The option suggesting an Auto Scaling group with a minimum capacity of zero describes a pilot light strategy, not a warm standby. A pilot light approach does not have servers running and ready to take traffic, which would make meeting a 15-minute RTO challenging.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
Why is a cross-region read replica used for Amazon RDS in a warm standby strategy?
What is the role of the Auto Scaling group in a warm standby setup?
How does Amazon Route 53 help achieve failover in a warm standby architecture?
A company has 20 AWS member accounts that are linked to a management account by AWS Organizations. Each account hosts multiple workloads that are owned by different internal teams. Finance must:
- Generate a monthly chargeback report that shows the AWS cost for every team.
- Roll the cost of any untagged resources into a bucket named "NoOwner" so that Finance can spot tagging gaps.
- View a cost forecast for the next three months.
- Receive an automated email when a team's monthly spend is forecasted to exceed its budget by more than 10 percent.
The cloud architect wants to meet these requirements with the least operational overhead while using only native AWS services.
Which combination of actions meets the requirements?
Activate a "Team" cost allocation tag in the management account. Create an AWS Cost Category that inherits the tag and assigns a default value such as "NoOwner" for unmatched resources. In AWS Cost Explorer, build a monthly report grouped by the cost category and enable a three-month forecast. Create an AWS cost budget for each cost-category value with a forecasted threshold of 110 % and configure an email alert.
Enable AWS Cost Anomaly Detection and create a monitor for every account. Use AWS Compute Optimizer to project future costs and configure a single AWS Budget that sends alerts when the organization's unblended cost exceeds 10 % of the previous month.
Enable the AWS Cost and Usage Report (CUR) for each member account. Deliver CUR files to Amazon S3, run scheduled Amazon Athena queries that group costs by the "Team" tag, and visualize the results in Amazon QuickSight. Use Amazon CloudWatch alarms on Athena query results to send email alerts when spend exceeds 110 % of budget.
Create a billing group for each team in AWS Billing Conductor. Use AWS Pricing Calculator to estimate quarterly spend for each billing group and configure Amazon EventBridge rules that send SNS email notifications when actual account-level spend in AWS Cost Explorer exceeds the estimate by 10 %.
Answer Description
The simplest native-service approach is to rely on the AWS cost-management stack that is already integrated with AWS billing data.
- Activate a user-defined cost allocation tag such as "Team" in the management account so that the tag becomes visible in billing data. Activating the tag is required before it can be used for reporting.
- Create an AWS Cost Category that inherits the Team tag and sets a default value (for example, NoOwner) for all line items that do not contain the tag. This automatically groups both tagged and untagged spend without any external processing.
- Use AWS Cost Explorer to build a monthly view that is grouped by the new cost-category values. Cost Explorer can also display a forecast of up to 12 months, which covers the next quarter and requires no additional setup.
- For each cost-category value, create an AWS Budget that has a forecasted alert threshold of 110 % (10 % over the team's budgeted amount). Configure the budget to send email (or SNS) notifications when the threshold is crossed.
This solution meets every requirement, is entirely managed by AWS, and avoids the data-pipeline maintenance associated with exporting and querying the Cost and Usage Report or building custom dashboards. The other options either require significantly more operational work (for example, CUR + Athena + QuickSight), do not identify untagged resources, or use services that do not provide the required forecasting and alerting capabilities.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is an AWS Cost Category and how does it help in cost management?
How does AWS Budgets enable forecasting and alerts?
What is the role of cost allocation tags, and why are they important?
A startup is refactoring its stateless microservice API that currently runs on EC2 instances. Baseline traffic is about 100 requests per minute, but social-media campaigns can cause spikes of up to 10 000 requests per minute that last 15-30 minutes. The 95th-percentile backend response time must stay below 200 ms during spikes, compute cost must be as low as possible when demand is minimal, and the five-person operations team must not manage server patching or cluster capacity. Which architecture best satisfies these business objectives?
Run the containers in an Amazon ECS service using AWS Fargate. Define a capacity-provider strategy that assigns a higher weight to FARGATE_SPOT than FARGATE, set the desired task count between 1 and 300, and attach a target-tracking scaling policy on the ALBRequestCountPerTarget metric to keep the average load at 50 requests per task.
Run the containers on an Amazon EKS cluster that uses only Spot Instances managed by Cluster Autoscaler and Karpenter, keeping one node permanently in the cluster.
Repackage the workload as AWS Lambda functions behind Amazon API Gateway and configure 1 000 units of Provisioned Concurrency for each function.
Deploy the containers to an Amazon ECS cluster backed by an EC2 Auto Scaling group of c6i.large On-Demand instances and use a target-tracking scaling policy that keeps average CPU utilization at 60 percent with predictive scaling enabled.
Answer Description
Running the containers in an Amazon ECS service on AWS Fargate removes the need to manage servers or clusters. A target-tracking scaling policy on the ALBRequestCountPerTarget metric automatically adds or removes tasks so that each task handles roughly the specified number of requests, allowing the service to scale from one task during idle periods to hundreds of tasks within minutes and keep latency under 200 ms. A capacity-provider strategy that gives higher weight to FARGATE_SPOT than FARGATE launches tasks on discounted Fargate Spot whenever spare capacity is available, cutting compute cost by up to 70 percent, while still permitting on-demand Fargate tasks if Spot capacity is unavailable. The alternative solutions either retain idle EC2 or Provisioned Concurrency capacity, add operational overhead, or rely solely on Spot Instances without a cost-effective fallback, so they do not meet both the cost and performance goals as effectively.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is AWS Fargate and how does it eliminate the need to manage servers?
What is a capacity-provider strategy and how do FARGATE_SPOT and FARGATE work together?
How does the ALBRequestCountPerTarget metric support auto-scaling in this solution?
Nice!
Looks like that's it! You can go back and review your answers or click the button below to grade your test.