Episode 93: AI Services: Rekognition & Comprehend
Domain 4 of the AWS Cloud Practitioner exam shifts from technology into the financial side of cloud usage—billing, pricing, and support. While it may not feel as exciting as databases or security, cost awareness is a crucial part of the AWS shared responsibility model. Customers control how they consume services, and that means understanding pricing levers, tools for visibility, and strategies for optimization. This domain teaches you how AWS charges for compute, storage, and transfer, and how to manage those costs through calculators, budgets, and consolidated billing. In practice, cost management is often the difference between a smooth cloud adoption and a painful surprise bill.
The pay-as-you-go model is the foundation of AWS pricing. Instead of purchasing hardware upfront, you pay only for what you use. Services are metered by different units: compute is charged by instance hours or vCPU-seconds, storage by gigabytes stored per month, and data transfer by gigabytes moved. This model provides flexibility but requires awareness. Leaving a large EC2 instance running idle still generates hourly costs, and storing massive amounts of rarely accessed data in S3 Standard can quietly accumulate charges. The advantage is agility: you can scale up instantly without capital investment, but the responsibility is to monitor and optimize usage actively.
Compute pricing offers multiple models. On-Demand Instances charge by the hour or second, depending on the instance type, and require no commitment—ideal for unpredictable or spiky workloads. Reserved Instances (RIs) provide significant discounts, up to 72 percent, in exchange for one- or three-year commitments, making them ideal for steady-state applications. Savings Plans offer similar discounts but with more flexibility, applying across EC2, Fargate, and Lambda usage, as long as you commit to a consistent spend. Spot Instances provide the deepest discounts, sometimes 90 percent off, but they can be interrupted when AWS needs capacity back. They’re perfect for workloads that tolerate interruptions, such as rendering, simulations, or batch analytics. For the exam, cues like “steady baseline” point to Reserved or Savings Plans, while “fault-tolerant batch” points to Spot.
The Free Tier provides limited usage at no cost, giving new users and developers a way to explore AWS services. It includes 12 months of free resources, such as 750 hours per month of a t2.micro or t3.micro EC2 instance, 5 GB of S3 storage, and 750 hours of RDS usage. Some services, like Lambda and DynamoDB, include “always free” allocations, with a million requests or limited read/write capacity every month. The Free Tier is an important onboarding tool, but exam questions may test whether you understand its limits—it is not unlimited or suitable for production-scale use. Awareness of where Free Tier ends and standard charges begin is essential.
Data transfer pricing is another frequent exam topic. Inbound data transfer into AWS is generally free, but outbound transfers cost money, especially across Regions or to the internet. For example, transferring data between Availability Zones in the same Region may incur small charges, while moving data from one Region to another can be more expensive. Services like CloudFront and S3 Transfer Acceleration reduce costs and latency for global data delivery. On the exam, if the scenario highlights “reduce egress costs,” think CloudFront caching or leveraging private VPC endpoints instead of routing over the internet.
S3 pricing is tiered by storage class. Standard is most expensive but offers frequent access and high durability. Standard-IA and One Zone-IA are cheaper but charge for retrieval. Glacier and Glacier Deep Archive are lowest cost but require retrieval delays. Intelligent-Tiering adjusts automatically based on usage. The exam expects you to recognize which storage class fits an access pattern: “archival with hours-long retrieval” points to Glacier, “rarely accessed but still immediately available” points to Standard-IA, and “frequently accessed” defaults to Standard. S3’s model emphasizes aligning data patterns with the right storage class to balance cost and performance.
Compute costs are driven by multiple factors: instance type, size, Region, and usage hours. Larger and more powerful instances cost more, Regions have different price levels, and running instances continuously generates predictable monthly charges. For example, an m5.large instance costs less than an m5.4xlarge, and the same instance may cost more in Asia Pacific than in North Virginia. For the exam, the important takeaway is that compute costs are flexible and predictable if you right-size and choose Regions strategically.
Support plans round out the pricing model. Basic support is free, offering access to documentation and forums. Developer support provides business-hours technical support for non-production use. Business support, a common enterprise choice, includes 24/7 access to engineers, architecture guidance, and faster response times. Enterprise support goes further, with a designated Technical Account Manager and proactive support features. On the exam, if the scenario involves “mission-critical workloads with 24/7 support,” Business or Enterprise is the right choice. If it’s “experimentation or learning,” Basic or Developer fits best.
The AWS Pricing Calculator is an essential tool for estimating costs. It allows architects to model usage of services like EC2, RDS, or S3 with different configurations and see projected monthly charges. While not a detailed exam focus, you should know that it exists and is used for planning and estimating. Similarly, Cost Explorer provides historical views of charges and usage trends, helping identify cost drivers and allocation patterns. For example, you might use Cost Explorer to see that S3 storage doubled last month, driving higher bills.
AWS Budgets extend visibility into proactive management. They allow you to set thresholds on spending or usage and trigger alerts when those are exceeded. For example, you could set a budget to notify your team if EC2 costs rise above $5,000 in a month. Unlike Cost Explorer, which is retrospective, Budgets are proactive. For the exam, if the scenario mentions “alert when costs exceed limits,” the answer is Budgets.
The Cost and Usage Report (CUR) provides the most granular level of billing detail. It includes line-item data for every charge, down to specific resource usage. Enterprises often use CUR data for chargeback or detailed analysis. It can be exported to S3 and queried with Athena for insights. While CUR may not appear often on the exam, knowing it’s the most detailed report is helpful for tricky billing questions.
Finally, consolidated billing allows multiple AWS accounts in an organization to pool usage for volume discounts and simplify invoicing. Shared Reserved Instances and Savings Plans discounts apply across accounts, reducing overall costs. For example, one account may run Reserved Instances, but other accounts in the organization can still benefit from the discount. Consolidated billing also enables chargeback models, where each team is billed for its portion of the shared spend. For the exam, if the scenario highlights “shared discounts across accounts,” the correct answer is consolidated billing.
A tagging strategy is one of the simplest and most effective cost management practices. Cost allocation tags allow you to label resources by project, environment, owner, or business unit. When enabled, these tags feed into Cost Explorer and the Cost and Usage Report, making it possible to slice charges by category. For example, tagging all EC2 instances with “project=analytics” lets finance teams track spend for that initiative. The exam may frame this as “allocate costs to departments,” with the correct answer being cost allocation tags. Without consistent tagging hygiene, organizations struggle to attribute costs accurately, leading to blurred accountability.
Rightsizing is another cornerstone of cost control. Many workloads run on oversized instances or continue idling after demand subsides. AWS provides tools like Compute Optimizer and Trusted Advisor to identify underutilized resources, such as EC2 instances with consistently low CPU usage or RDS instances with minimal connections. Rightsizing means matching capacity to actual demand—downsizing instances, terminating unused ones, or moving to serverless when possible. On the exam, if you see “idle resources” or “optimize usage,” rightsizing is the right principle.
Lifecycle policies for storage further reduce costs. In S3, you can define lifecycle rules to transition objects into cheaper storage classes automatically. For example, logs might remain in S3 Standard for 30 days, then move to Standard-IA, and finally to Glacier Deep Archive after a year. This automation ensures data is retained for compliance or analytics while minimizing expense. On the exam, if the question asks about reducing cost for data that “must be retained but rarely accessed,” lifecycle policies and S3 tiering are the answer.
Architecture choices also influence cost. Caching at multiple levels reduces expensive operations and egress. CloudFront caches content at edge locations, minimizing data transfer charges from Regions. ElastiCache keeps frequently accessed queries in memory, reducing database load. For example, serving product images through CloudFront reduces both latency for users and egress costs from S3. Exam cues like “reduce latency and lower egress” nearly always point to caching or CDN integration.
Reserved capacity strategies are critical for predictable workloads. For EC2, RDS, Redshift, and ElastiCache, Reserved Instances provide cost savings for steady baselines. For compute flexibility, Savings Plans extend these savings across EC2, Fargate, and Lambda. The exam distinguishes these: RIs tie savings to specific instance types or families, while Savings Plans offer more flexibility. The cue is “steady-state workload,” which maps to reserved pricing models.
Spot Instances demand a strategy of diversification and fault tolerance. Because they can be interrupted with little notice, workloads must checkpoint progress and resume gracefully. For example, a video rendering farm can use multiple Spot fleets across instance types, falling back to On-Demand if interruptions occur. The exam often signals Spot with keywords like “cost savings for fault-tolerant batch processing.” Remember that Spot delivers deep discounts but requires careful workload design.
Governance comes through budgets, guardrails, and visibility. Budgets set spending thresholds, triggering alerts before costs spiral. Service Control Policies (SCPs) can block actions like launching large instance types in development accounts. For example, an organization might enforce SCPs to deny provisioning GPU instances unless approved. Governance ensures that cost and security best practices are enforced automatically, not just by policy documents.
Trusted Advisor provides a suite of cost checks, including recommendations for underutilized resources, low utilization on Reserved Instances, or unattached EBS volumes. For example, it might flag hundreds of unused elastic IPs, suggesting they be released. The exam sometimes frames this as “a tool that provides cost optimization checks,” and the correct answer is Trusted Advisor.
Forecasting and variance analysis are important for long-term planning. AWS Budgets and Cost Explorer both support forecasting based on historical usage. This helps organizations predict future spend and investigate when actuals deviate from forecasts. For example, if a project consistently exceeds budget in one Region, variance analysis can uncover workload growth or misconfigured resources. Forecasting turns cost management from reactive firefighting into proactive control.
AWS Marketplace and third-party solutions add another layer. Marketplace products may include licensing fees that appear on AWS bills, while cost management vendors provide advanced features like anomaly detection or multi-cloud reporting. The exam may touch lightly on Marketplace, especially around cost visibility and procurement. The key is knowing that Marketplace spend is part of consolidated AWS billing and can be tracked with the same tools.
Cross-account chargeback and showback models are common in enterprises using consolidated billing. Showback provides visibility of costs to teams without directly charging them, while chargeback allocates actual costs back to business units. For example, finance may allocate S3 charges proportionally across departments based on tagged usage. The exam often signals this with “allocating costs by team or department,” pointing back to consolidated billing plus tags.
From an exam perspective, the goal is to map workload profiles to the correct pricing tool. If the workload is steady and predictable, Reserved Instances or Savings Plans fit. If it’s spiky and unpredictable, On-Demand is best. If it’s tolerant of interruptions, Spot provides the answer. If the need is simply to estimate or plan, the Pricing Calculator is correct, while monitoring and enforcing limits points to Budgets. The exam doesn’t expect deep accounting knowledge but does expect you to match requirements with pricing models and tools.
In conclusion, Domain 4 is about combining pricing models with governance, tagging, and monitoring to keep cloud spend predictable and efficient. AWS’s flexibility allows organizations to optimize at every level, from compute reservations to storage lifecycle rules, but the responsibility lies with the customer to apply these tools. For the exam, the lesson is clear: pick the simplest pricing model or tool that matches the workload profile, and always assume governance and tagging are part of the answer. By mastering these fundamentals, you’ll approach billing questions with the same confidence as technical ones, closing out your Cloud Practitioner preparation with a complete perspective.
