Back to Blog
Interview Q&A

Top 200 AWS Cloud Interview Questions & Answers

Fortress Institute2026-04-0545 min read

Basic Questions (1-80)

Q1. What is Amazon Web Services (AWS)?

Amazon Web Services is a comprehensive cloud computing platform provided by Amazon, offering over 200 fully featured services from data centers globally. AWS provides on-demand access to compute power, database storage, content delivery, and other IT resources through the internet with pay-as-you-go pricing. It enables organizations to move faster, lower IT costs, and scale applications globally without upfront infrastructure investment.

Q2. What are the main categories of AWS services?

AWS services are organized into categories including Compute (EC2, Lambda, ECS), Storage (S3, EBS, EFS, Glacier), Database (RDS, DynamoDB, Redshift, ElastiCache), Networking (VPC, Route 53, CloudFront, Direct Connect), Security (IAM, KMS, Shield, WAF), and Management & Monitoring (CloudWatch, CloudTrail, Config). Additional categories cover ML/AI, IoT, Developer Tools, Migration, and Analytics. Each category contains multiple services tailored to specific use cases.

Q3. What is an AWS Region?

An AWS Region is a physical location in the world where AWS clusters data centers, currently comprising over 30 geographic regions globally. Each Region is completely independent and isolated from other Regions to achieve the greatest possible fault tolerance and stability. Customers choose Regions based on latency, compliance, data residency requirements, and service availability.

Q4. What is an Availability Zone (AZ)?

An Availability Zone is one or more discrete data centers within an AWS Region, each with redundant power, networking, and connectivity. AZs within a Region are connected through low-latency, high-throughput, and highly redundant networking. By distributing applications across multiple AZs, architects achieve high availability and fault tolerance—a failure in one AZ does not affect other AZs.

Q5. What is Amazon EC2?

Amazon Elastic Compute Cloud (EC2) provides resizable compute capacity in the cloud, allowing users to launch virtual servers called instances. EC2 offers a wide selection of instance types optimized for different use cases—compute-optimized (C-family), memory-optimized (R/X-family), storage-optimized (I/D-family), and general-purpose (M/T-family). Users pay only for compute capacity consumed, with options for On-Demand, Reserved, Spot, and Dedicated Host pricing.

Q6. What is Amazon S3?

Amazon Simple Storage Service (S3) is an object storage service offering industry-leading scalability, data availability, security, and performance. S3 stores data as objects within buckets, supporting objects up to 5 TB in size with virtually unlimited total storage. S3 provides multiple storage classes (Standard, Intelligent-Tiering, Standard-IA, One Zone-IA, Glacier, Glacier Deep Archive) for cost optimization based on access patterns.

Q7. What is IAM in AWS?

AWS Identity and Access Management (IAM) enables secure control of access to AWS services and resources. IAM allows creation of users, groups, and roles, and management of permissions through policies written in JSON. Key IAM concepts include the principle of least privilege, identity federation, multi-factor authentication (MFA), and service roles that grant AWS services permissions to act on your behalf.

Q8. What is a VPC?

Amazon Virtual Private Cloud (VPC) lets you provision a logically isolated section of the AWS Cloud where you can launch AWS resources in a virtual network that you define. You have complete control over your virtual networking environment, including selection of your own IP address range, creation of subnets, and configuration of route tables and network gateways. VPC supports both IPv4 and IPv6 addressing.

Q9. What is the difference between a public subnet and a private subnet?

A public subnet has a route in its route table pointing to an Internet Gateway, allowing resources within it to communicate directly with the internet. A private subnet has no direct route to the Internet Gateway, preventing direct internet access; resources in private subnets access the internet through a NAT Gateway or NAT Instance placed in a public subnet. Public subnets host internet-facing resources like load balancers, while private subnets host databases and application servers.

Q10. What is Amazon RDS?

Amazon Relational Database Service (RDS) makes it easy to set up, operate, and scale a relational database in the cloud. RDS manages time-consuming database administration tasks such as hardware provisioning, database setup, patching, and backups. RDS supports multiple database engines: MySQL, PostgreSQL, MariaDB, Oracle, SQL Server, and Amazon Aurora, with Multi-AZ deployment for high availability and Read Replicas for read scaling.

Q11. What is Amazon DynamoDB?

Amazon DynamoDB is a fully managed, serverless, key-value and document NoSQL database service that delivers single-digit millisecond performance at any scale. DynamoDB automatically scales throughput capacity with demand and replicates data across multiple AZs for durability. It supports DynamoDB Streams for change data capture, global tables for multi-region replication, and DAX (DynamoDB Accelerator) for in-memory caching.

Q12. What is AWS Lambda?

AWS Lambda is a serverless compute service that runs code in response to events and automatically manages the underlying compute resources. Lambda executes functions in response to triggers from over 200 AWS services and SaaS applications, scaling automatically from a few requests per day to thousands per second. Functions are billed in 1-millisecond increments based on execution duration and memory allocated, with no charge when code is not running.

Q13. What is Amazon CloudFront?

Amazon CloudFront is a fast content delivery network (CDN) service that securely delivers data, videos, applications, and APIs to customers globally with low latency and high transfer speeds. CloudFront integrates with AWS services like S3, EC2, Elastic Load Balancing, and Route 53, and provides DDoS protection via AWS Shield Standard at no additional cost. It supports custom SSL certificates, field-level encryption, and Lambda@Edge for running code at edge locations.

Q14. What is Amazon Route 53?

Amazon Route 53 is a highly available and scalable cloud Domain Name System (DNS) web service designed to route end users to internet applications. Route 53 supports multiple routing policies: Simple, Weighted, Latency-based, Failover, Geolocation, Geoproximity, and Multivalue Answer routing. It also provides domain registration, health checks, and traffic flow management with a visual editor.

Q15. What are EC2 instance purchasing options?

AWS offers several EC2 purchasing options: On-Demand Instances pay by the second with no long-term commitment; Reserved Instances provide up to 72% discount for 1- or 3-year commitments; Spot Instances use unused EC2 capacity at up to 90% discount but can be interrupted; Dedicated Hosts are physical servers dedicated to your use for compliance/licensing needs; Savings Plans offer flexible pricing with committed spend. Choosing the right mix optimizes cost significantly.

Q16. What is an Amazon Machine Image (AMI)?

An Amazon Machine Image (AMI) is a template that contains a software configuration (operating system, application server, and applications) from which you launch EC2 instances. AMIs can be AWS-provided, AWS Marketplace AMIs, community AMIs, or custom AMIs you create from existing instances. AMIs are region-specific and must be copied to other regions if needed. Storing AMIs enables rapid, consistent instance provisioning.

Q17. What is Amazon EBS?

Amazon Elastic Block Store (EBS) provides persistent block-level storage volumes for use with EC2 instances, persisting independently from the instance's life. EBS volume types include gp3/gp2 (general-purpose SSD), io2/io1 (provisioned IOPS SSD for databases), st1 (throughput-optimized HDD for big data), and sc1 (cold HDD for infrequent access). EBS supports snapshots to S3 for backup and volume replication across AZs.

Q18. What is Amazon EFS?

Amazon Elastic File System (EFS) provides a simple, scalable, fully managed elastic NFS file system for use with AWS Cloud services and on-premises resources. EFS automatically grows and shrinks as you add and remove files, requiring no capacity management. It supports two performance modes (General Purpose and Max I/O), two throughput modes (Bursting and Provisioned), and four storage classes including EFS Standard and EFS Infrequent Access.

Q19. What is AWS CloudFormation?

AWS CloudFormation provides a common language for describing and provisioning all infrastructure resources in your cloud environment using Infrastructure as Code (IaC). CloudFormation templates (JSON or YAML) define AWS resources and their dependencies, allowing creation, update, and deletion of entire resource stacks as a single unit. Key concepts include stacks, change sets (preview changes), stack sets (deploy across multiple accounts/regions), and drift detection.

Q20. What is Amazon CloudWatch?

Amazon CloudWatch is a monitoring and observability service that collects monitoring and operational data in the form of logs, metrics, and events. CloudWatch provides actionable insights to monitor applications, respond to system-wide performance changes, optimize resource utilization, and get a unified view of operational health. Key features include dashboards, alarms, Logs Insights for log analysis, Container Insights, and Synthetics for canary monitoring.

Q21. What is AWS Auto Scaling?

AWS Auto Scaling monitors applications and automatically adjusts capacity to maintain steady, predictable performance at the lowest possible cost. It supports scaling for EC2 instances, ECS tasks, DynamoDB tables, Aurora replicas, and more. Auto Scaling Groups (ASGs) define minimum, maximum, and desired capacity; scaling policies can be target tracking (maintain a metric at a target value), step scaling, or scheduled scaling based on predictable load patterns.

Q22. What is Elastic Load Balancing (ELB)?

Elastic Load Balancing automatically distributes incoming application traffic across multiple targets—EC2 instances, containers, IP addresses, and Lambda functions—in one or more Availability Zones. AWS offers four types: Application Load Balancer (layer 7, HTTP/HTTPS with path-based routing), Network Load Balancer (layer 4, ultra-low latency TCP/UDP), Gateway Load Balancer (for third-party virtual appliances), and Classic Load Balancer (legacy). ELB integrates with Auto Scaling, ACM, WAF, and CloudWatch.

Q23. What is Amazon SNS?

Amazon Simple Notification Service (SNS) is a fully managed pub/sub messaging service for both application-to-application (A2A) and application-to-person (A2P) communication. SNS topics allow publishers to send messages to multiple subscribers simultaneously—endpoints can include SQS queues, Lambda functions, HTTP/HTTPS endpoints, email, SMS, and mobile push notifications. SNS supports message filtering so subscribers receive only relevant messages.

Q24. What is Amazon SQS?

Amazon Simple Queue Service (SQS) is a fully managed message queuing service that enables decoupling and scaling of microservices, distributed systems, and serverless applications. SQS offers two queue types: Standard queues provide maximum throughput, best-effort ordering, and at-least-once delivery; FIFO queues guarantee exactly-once processing and strict ordering. Key features include visibility timeout, dead-letter queues (DLQ), long polling, and message retention up to 14 days.

Q25. What is AWS CloudTrail?

AWS CloudTrail is a service that enables governance, compliance, operational auditing, and risk auditing of your AWS account by recording API calls. CloudTrail records the caller's identity, time of the API call, source IP address, request parameters, and response elements returned by AWS. Trails can deliver log files to S3 and CloudWatch Logs; CloudTrail Insights detects unusual API activity patterns automatically.

Q26. What is Amazon Glacier?

Amazon S3 Glacier (now S3 Glacier Flexible Retrieval) is a secure, durable, and extremely low-cost cloud storage class for data archiving and long-term backup. Glacier offers three retrieval options: Expedited (1-5 minutes), Standard (3-5 hours), and Bulk (5-12 hours). S3 Glacier Instant Retrieval provides millisecond access; Glacier Deep Archive (12-48 hour retrieval) is the lowest-cost storage option for data retained for 7-10 years.

Q27. What is the AWS Shared Responsibility Model?

The AWS Shared Responsibility Model defines the division of security responsibilities between AWS and the customer. AWS is responsible for security "of" the cloud—protecting the infrastructure (hardware, software, networking, facilities) that runs AWS services. Customers are responsible for security "in" the cloud—including data encryption, access management, OS patching, network configuration, and application-level security. The boundary varies by service type (IaaS vs. managed services).

Q28. What is an Internet Gateway?

An Internet Gateway (IGW) is a horizontally scaled, redundant, and highly available VPC component that enables communication between instances in your VPC and the internet. It serves two purposes: providing a target in VPC route tables for internet-routable traffic, and performing network address translation (NAT) for instances with public IPv4 addresses. Each VPC can only have one IGW attached at a time.

Q29. What is a NAT Gateway?

A NAT (Network Address Translation) Gateway enables instances in a private subnet to connect to the internet or other AWS services while preventing the internet from initiating connections with those instances. NAT Gateways are managed by AWS, highly available within an AZ, and support bandwidth scaling up to 100 Gbps. They are placed in public subnets and private subnet route tables point to the NAT Gateway for internet-bound traffic.

Q30. What is Amazon ElastiCache?

Amazon ElastiCache is a fully managed in-memory caching service supporting Redis and Memcached engines. ElastiCache improves application performance by retrieving data from fast, managed, in-memory caches instead of slower disk-based databases. Redis supports advanced data structures, persistence, pub/sub, Lua scripting, and cluster mode for horizontal scaling; Memcached is simpler, multi-threaded, and suitable for simple caching use cases.

Q31. What is AWS Elastic Beanstalk?

AWS Elastic Beanstalk is an easy-to-use service for deploying and scaling web applications and services developed in Java, .NET, PHP, Node.js, Python, Ruby, Go, and Docker. Beanstalk automatically handles capacity provisioning, load balancing, auto-scaling, and application health monitoring while retaining full control over underlying resources. It supports multiple deployment policies: All at once, Rolling, Rolling with additional batch, and Immutable.

Q32. What is AWS ECS?

Amazon Elastic Container Service (ECS) is a fully managed container orchestration service that makes it easy to deploy, manage, and scale containerized applications. ECS supports two launch types: EC2 (you manage the underlying instances) and Fargate (serverless, AWS manages infrastructure). Key ECS concepts include Task Definitions (container blueprints), Tasks (running instances of task definitions), Services (ensure desired task count), and Clusters (logical grouping of resources).

Q33. What is AWS EKS?

Amazon Elastic Kubernetes Service (EKS) is a managed service that makes it easy to run Kubernetes on AWS without needing to install and operate your own Kubernetes control plane. EKS runs the Kubernetes control plane across multiple AWS Availability Zones and automatically detects and replaces unhealthy control plane nodes. It supports both EC2 and Fargate for worker nodes and integrates with AWS services like IAM, VPC, CloudWatch, and ECR.

Q34. What is Amazon ECR?

Amazon Elastic Container Registry (ECR) is a fully managed Docker container registry that makes it easy to store, manage, share, and deploy container images. ECR integrates with ECS, EKS, and Lambda, eliminating the need to manage your own container registries. It supports image vulnerability scanning, lifecycle policies to automate image cleanup, cross-account and cross-region replication, and encryption at rest using AWS KMS.

Q35. What is Amazon Redshift?

Amazon Redshift is a fast, scalable, fully managed cloud data warehouse that makes it simple and cost-effective to analyze all your data using standard SQL and existing BI tools. Redshift uses columnar storage, data compression, and massively parallel processing (MPP) to deliver fast query performance. Redshift Spectrum allows querying data directly in S3 without loading it; Redshift Serverless removes the need to manage clusters.

Q36. What is AWS Direct Connect?

AWS Direct Connect is a cloud service solution that establishes a dedicated private network connection between your on-premises network and AWS, bypassing the public internet. Direct Connect provides consistent network performance, reduced bandwidth costs, and increased bandwidth throughput for data-intensive workloads. Connection speeds range from 50 Mbps to 100 Gbps; hosted connections and hosted VIFs allow sharing through AWS partners.

Q37. What is Amazon Kinesis?

Amazon Kinesis makes it easy to collect, process, and analyze real-time streaming data at any scale. Kinesis Data Streams captures gigabytes of data per second from hundreds of thousands of sources; Kinesis Data Firehose is the easiest way to reliably load streaming data into data stores (S3, Redshift, OpenSearch); Kinesis Data Analytics enables real-time analytics using SQL or Apache Flink; Kinesis Video Streams captures and stores video streams.

Q38. What is AWS Glue?

AWS Glue is a fully managed serverless data integration service that makes it easy to discover, prepare, and combine data for analytics, ML, and application development. Glue crawlers automatically discover data stores and populate the Glue Data Catalog; ETL jobs transform and move data between stores using auto-generated or custom Apache Spark/Python scripts. Glue Studio provides a visual interface for building ETL pipelines without code.

Q39. What is Amazon Athena?

Amazon Athena is an interactive query service that makes it easy to analyze data in Amazon S3 using standard SQL. Athena is serverless—no infrastructure to manage—and you pay only for the queries run ($5 per TB of data scanned). It supports CSV, JSON, ORC, Avro, and Parquet formats; using columnar formats and partitioning can reduce cost and improve performance significantly. Athena integrates with AWS Glue Data Catalog for metadata management.

Q40. What is AWS KMS?

AWS Key Management Service (KMS) is a managed service that makes it easy to create and control the cryptographic keys used to protect your data. KMS uses hardware security modules (HSMs) to protect the security of your keys and integrates with over 100 AWS services for encryption. Key types include AWS managed keys (free), customer managed keys (CMKs), and custom key store (CloudHSM-backed). KMS provides centralized key management, key rotation, and detailed audit logs via CloudTrail.

Q41. What is Amazon SageMaker?

Amazon SageMaker is a fully managed service that provides every developer and data scientist with the ability to build, train, and deploy ML models quickly. SageMaker removes the heavy lifting of each step of the ML process—from data labeling (Ground Truth) to experimentation (Experiments) to model monitoring (Model Monitor). It provides built-in algorithms, automatic model tuning (hyperparameter optimization), and managed infrastructure for distributed training.

Q42. What is Amazon Cognito?

Amazon Cognito provides authentication, authorization, and user management for web and mobile apps. Cognito User Pools are user directories that provide sign-up, sign-in, and access control functionality; Cognito Identity Pools (Federated Identities) grant users temporary AWS credentials to access AWS services directly. Cognito supports social identity providers (Google, Facebook, Apple) and SAML/OIDC enterprise identity providers.

Q43. What is Amazon API Gateway?

Amazon API Gateway is a fully managed service that makes it easy to create, publish, maintain, monitor, and secure APIs at any scale. API Gateway supports REST APIs, HTTP APIs (lower latency and cost), and WebSocket APIs for real-time two-way communication. Features include throttling, caching, request/response transformation, API keys, usage plans, Lambda authorizers, and integration with Cognito for authentication.

Q44. What is AWS Step Functions?

AWS Step Functions is a serverless orchestration service that lets you coordinate multiple AWS services into serverless workflows using a visual workflow editor called Workflow Studio. Step Functions uses the Amazon States Language (ASL) to define state machines with states including Task, Choice, Parallel, Map, Wait, Succeed, Fail, and Pass. It supports Standard Workflows (exactly-once execution) and Express Workflows (high-volume, at-least-once).

Q45. What is Amazon EventBridge?

Amazon EventBridge is a serverless event bus service that makes it easy to build event-driven applications at scale. EventBridge receives events from AWS services, custom applications, and SaaS applications, then routes them to targets like Lambda, SQS, SNS, Step Functions, and more using rules with event patterns. EventBridge Scheduler enables scheduled invocations; EventBridge Pipes connects event sources to targets with optional filtering and enrichment.

Q46. What is AWS CodePipeline?

AWS CodePipeline is a fully managed continuous delivery service that helps automate release pipelines for fast and reliable application and infrastructure updates. Pipelines consist of stages (Source, Build, Test, Deploy) with transitions and actions; each action integrates with AWS services or third-party tools. CodePipeline integrates with CodeCommit, CodeBuild, CodeDeploy, CloudFormation, Elastic Beanstalk, ECS, and popular third-party tools like GitHub, Jenkins, and Jira.

Q47. What is AWS CodeBuild?

AWS CodeBuild is a fully managed continuous integration service that compiles source code, runs tests, and produces software packages ready for deployment. Build specifications are defined in a buildspec.yml file specifying install, pre-build, build, and post-build phases. CodeBuild scales automatically and charges per minute of build time; it supports Docker for custom build environments and caches dependencies to speed builds.

Q48. What is AWS CodeDeploy?

AWS CodeDeploy is a fully managed deployment service that automates software deployments to EC2 instances, on-premises servers, Lambda functions, and ECS services. Deployment configurations control rollout speed: AllAtOnce, HalfAtATime, OneAtATime, or custom. AppSpec files define deployment hooks (lifecycle events) for installation validation; CodeDeploy supports Blue/Green deployments to shift traffic with automatic rollback on failure.

Q49. What is Amazon Aurora?

Amazon Aurora is a MySQL and PostgreSQL-compatible relational database built for the cloud that combines the performance and availability of high-end commercial databases with the simplicity and cost-effectiveness of open-source databases. Aurora is up to 5x faster than standard MySQL and 3x faster than PostgreSQL, provides up to 15 low-latency read replicas, and stores six copies of data across three AZs. Aurora Serverless v2 scales capacity instantly in fine-grained increments.

Q50. What is Amazon VPC Peering?

VPC peering is a networking connection between two VPCs that enables routing of traffic between them using private IPv4 or IPv6 addresses as if they were in the same network. VPC peering can be between VPCs in the same account, different AWS accounts, or different Regions (inter-region peering). Peering is non-transitive—if VPC A peers with B and B peers with C, traffic cannot flow from A to C through B without a direct A-C peering connection.

Q51. What is AWS Transit Gateway?

AWS Transit Gateway is a network transit hub that enables customers to connect their VPCs and on-premises networks through a central gateway, simplifying network architecture. Instead of managing many-to-many VPC peering connections, Transit Gateway acts as a hub-and-spoke model supporting thousands of VPC attachments. It supports inter-region peering, multicast, route tables for traffic segmentation, and integration with Direct Connect and VPN.

Q52. What is AWS Shield?

AWS Shield is a managed DDoS protection service that safeguards applications running on AWS. Shield Standard is automatically enabled for all AWS customers at no additional charge and protects against common, frequently occurring network and transport layer DDoS attacks. Shield Advanced ($3,000/month) provides enhanced protections for EC2, ELB, CloudFront, Route 53, and Global Accelerator, including 24/7 access to the AWS DDoS Response Team (DRT) and cost protection against scaling charges from attacks.

Q53. What is AWS WAF?

AWS Web Application Firewall (WAF) protects web applications from common web exploits and bots that could affect availability, compromise security, or consume excessive resources. WAF lets you create rules that control bot traffic and block common attack patterns such as SQL injection, cross-site scripting (XSS), and OWASP Top 10 vulnerabilities. WAF Web ACLs attach to CloudFront, ALB, API Gateway, and AppSync; Managed Rule Groups from AWS and AWS Marketplace provide pre-built protections.

Q54. What is Amazon Inspector?

Amazon Inspector is an automated vulnerability management service that continually scans AWS workloads for software vulnerabilities and unintended network exposure. Inspector automatically discovers EC2 instances, Lambda functions, and container images in ECR and scans them for known CVEs and network reachability issues. It provides a risk score for each finding prioritized by exploitability and blast radius, and integrates with Security Hub for centralized findings.

Q55. What is Amazon GuardDuty?

Amazon GuardDuty is a threat detection service that continuously monitors for malicious activity and unauthorized behavior across your AWS accounts and workloads. GuardDuty analyzes CloudTrail events, VPC Flow Logs, DNS logs, EKS audit logs, and S3 data events using machine learning, anomaly detection, and integrated threat intelligence. Findings are categorized by severity and integrated with Security Hub, Detective, and EventBridge for automated response.

Q56. What is AWS Security Hub?

AWS Security Hub provides a comprehensive view of your security state in AWS and helps you check your environment against security industry standards and best practices. Security Hub aggregates, organizes, and prioritizes security alerts (findings) from multiple AWS services (GuardDuty, Inspector, Macie, IAM Analyzer, Firewall Manager) and third-party products. It evaluates your environment against the AWS Foundational Security Best Practices and CIS AWS Foundations standards.

Q57. What is Amazon Macie?

Amazon Macie is a data security service that uses ML to automatically discover, classify, and protect sensitive data stored in Amazon S3. Macie recognizes sensitive data types such as PII (names, addresses, credit card numbers, SSNs), financial data, and credentials. It provides an inventory of S3 buckets with their public accessibility and encryption status, and generates findings for policy violations and sensitive data discoveries that integrate with Security Hub and EventBridge.

Q58. What is Amazon Rekognition?

Amazon Rekognition is a cloud-based computer vision service that makes it easy to add image and video analysis to your applications. Rekognition can detect objects, scenes, activities, text, faces, and celebrities in images and videos; it can also compare faces for verification and search faces in collections. Rekognition Video analyzes video stored in S3 or streaming video from Kinesis Video Streams for activities and person tracking.

Q59. What is Amazon Lex?

Amazon Lex is a fully managed AI service for building conversational interfaces (chatbots and virtual agents) using voice and text. Lex provides advanced natural language understanding (NLU) and automatic speech recognition (ASR) technologies—the same technology that powers Amazon Alexa. It supports intents, slots, slot types, and fulfillment via Lambda functions; integrates with Connect (contact center), Kendra (knowledge base), and Polly (text-to-speech).

Q60. What is AWS Fargate?

AWS Fargate is a serverless compute engine for containers that works with both Amazon ECS and Amazon EKS, removing the need to provision and manage servers. With Fargate, you specify resource requirements (vCPU and memory) at the task or pod level, and AWS manages the underlying compute infrastructure. Fargate provides task-level isolation, integrates with VPC networking and IAM for security, and bills per vCPU and memory consumed per second.

Q61. What is Amazon SES?

Amazon Simple Email Service (SES) is a cost-effective, flexible, and scalable email service for sending transactional, marketing, and bulk emails. SES supports SMTP interface, API, and SDKs for email sending; features include email delivery metrics (delivery rate, bounce, complaints), suppression list management, virtual Deliverability Manager, and configuration sets for tracking. SES also supports email receiving with rules for processing incoming mail.

Q62. What is AWS Secrets Manager?

AWS Secrets Manager helps you protect secrets needed to access your applications, services, and IT resources by securely storing, rotating, managing, and retrieving credentials, API keys, and other secrets. Secrets Manager supports automatic rotation for supported services (RDS, Redshift, DocumentDB, other databases) using Lambda functions, eliminating hard-coded credentials. Secrets are encrypted using KMS and accessed via SDK/CLI/console with fine-grained IAM control.

Q63. What is AWS Systems Manager?

AWS Systems Manager (SSM) is an operations hub for managing applications and infrastructure running in the AWS Cloud and on-premises environments. SSM capabilities include Parameter Store (secure hierarchical storage for configuration data), Session Manager (secure shell access without SSH keys or bastion hosts), Patch Manager (automated patching), Run Command (remote command execution), and OpsCenter (aggregated operational issue view).

Q64. What is Amazon SWF?

Amazon Simple Workflow Service (SWF) is a web service that makes it easy to coordinate work across distributed application components. SWF manages state and flow of work across tasks and allows you to write coordination logic separately from activity logic. While SWF is still available, AWS now recommends Step Functions for new workflow orchestration use cases. SWF guarantees task assignment, tracks progress, and stores task results for up to one year.

Q65. What is Amazon WorkSpaces?

Amazon WorkSpaces is a fully managed, persistent desktop virtualization service that enables users to access data, applications, and resources they need from any device, anywhere. WorkSpaces runs on Amazon's global infrastructure and supports Windows and Linux desktops, with options for personal desktops (dedicated) or shared pools. WorkSpaces Secure Browser and Thin Client extend the portfolio for secure web browsing and lightweight virtual desktop scenarios.

Q66. What is AWS Backup?

AWS Backup is a fully managed, policy-based service that centralizes and automates data protection across AWS services. Backup policies (plans) define backup frequency, retention periods, lifecycle rules (transition to cold storage), and backup vault destinations. Supported services include EC2, EBS, RDS, Aurora, DynamoDB, EFS, FSx, Storage Gateway, VMware workloads, and Timestream. AWS Backup Audit Manager provides compliance reporting.

Q67. What is Amazon Elastic MapReduce (EMR)?

Amazon EMR is the industry-leading cloud big data solution for petabyte-scale data processing, interactive analytics, and ML using open-source frameworks like Apache Spark, Hive, Presto, HBase, Flink, and Hadoop. EMR automatically provisions and scales EC2 capacity; clusters can run on EC2 or EKS, or use EMR Serverless for automatic resource management. Use cases include ETL, log analysis, machine learning, real-time streaming, and genomics.

Q68. What is AWS DataSync?

AWS DataSync is an online data movement and discovery service that simplifies and accelerates migrations to AWS as well as moving data between on-premises and AWS storage. DataSync automates data movement between NFS, SMB, HDFS, self-managed object storage, and AWS storage (S3, EFS, FSx). It uses a purpose-built network protocol to transfer up to 10 Gbps, with built-in data integrity verification, encryption in transit, and bandwidth throttling.

Q69. What is Amazon OpenSearch Service?

Amazon OpenSearch Service (formerly Elasticsearch Service) makes it easy to deploy, operate, and scale OpenSearch clusters for log analytics, real-time application monitoring, and clickstream analysis. OpenSearch Service supports OpenSearch and legacy Elasticsearch APIs; UltraWarm provides cost-effective storage for hot/warm tiering; cold storage and Serverless options provide further cost optimization. It integrates with Kinesis Data Firehose, CloudWatch Logs, and DMS for data ingestion.

Q70. What is Amazon Timestream?

Amazon Timestream is a fast, scalable, and serverless time series database service for IoT and operational applications. Timestream automatically scales up or down to adjust capacity and performance, with recent data in memory and historical data in a cost-optimized magnetic tier. It offers built-in time series analytics functions (smoothing, approximation, interpolation) and integrates with IoT Core, Kinesis, Telegraf, and Grafana for visualization.

Q71. What is AWS Organizations?

AWS Organizations helps you centrally manage and govern your environment as you grow and scale your AWS resources. Organizations lets you create accounts, group them into organizational units (OUs), and apply Service Control Policies (SCPs) to restrict what services and actions can be used in member accounts. It enables consolidated billing across all accounts, centralized security controls, and delegated administrator for services like Security Hub, GuardDuty, and Config.

Q72. What is AWS Control Tower?

AWS Control Tower is the easiest way to set up and govern a secure, multi-account AWS environment based on AWS best practices. Control Tower automates the setup of a landing zone with pre-configured security baseline accounts (management, log archive, audit), organizational units, and preventive and detective guardrails. Guardrails are SCPs (preventive) or Config rules (detective/proactive) that enforce governance policies across the organization.

Q73. What is Amazon QuickSight?

Amazon QuickSight is a fast, cloud-native, serverless business intelligence (BI) service that makes it easy to deliver insights to everyone in your organization. QuickSight connects to AWS data services (S3, Redshift, Athena, RDS, Aurora), databases, SaaS applications, and third-party files. SPICE (Super-fast, Parallel, In-memory Calculation Engine) enables rapid query responses; Q enables natural language queries (NLQ); Pixel-perfect reports support operational reporting.

Q74. What is Amazon AppFlow?

Amazon AppFlow is a fully managed integration service that enables secure data transfer between AWS services and SaaS applications such as Salesforce, SAP, Zendesk, Slack, ServiceNow, and Marketo. AppFlow supports bidirectional data transfer with data transformation capabilities (masking, filtering, merging, validating) and runs on a fully managed, serverless infrastructure. It eliminates custom connector development and reduces integration time from weeks to minutes.

Q75. What is Amazon Lightsail?

Amazon Lightsail is the easiest way to get started with AWS for developers who need a simple cloud platform for low-complexity applications and websites. Lightsail provides virtual servers, storage, databases, and networking as easy-to-understand bundled plans at predictable monthly pricing. It includes pre-configured application stacks (WordPress, LAMP, Node.js, Magento) and management features (snapshots, monitoring, DNS) with a simplified console suitable for beginners.

Q76. What is the AWS Well-Architected Framework?

The AWS Well-Architected Framework describes key concepts, design principles, and architectural best practices for designing and running workloads in the cloud. It is organized around six pillars: Operational Excellence, Security, Reliability, Performance Efficiency, Cost Optimization, and Sustainability. The Well-Architected Tool in the console allows you to review workloads against the framework's questions and get improvement recommendations.

Q77. What is Amazon MSK?

Amazon Managed Streaming for Apache Kafka (MSK) is a fully managed service that makes it easy to build and run applications that use Apache Kafka to process streaming data. MSK manages the provisioning, configuration, and maintenance of Kafka clusters, including patching, replication, and automatic recovery from broker failures. MSK Serverless automatically provisions and scales capacity; MSK Connect enables managed Kafka Connect deployments for data integration.

Q78. What is AWS Lake Formation?

AWS Lake Formation is a service that makes it easy to set up, secure, and manage your data lake. Lake Formation simplifies data ingestion (from databases, S3) into the lake, applies ML-powered transformations with Blueprints, and enforces fine-grained access control (row-level, column-level security) using LF-Tags. It integrates with Glue Data Catalog, Athena, Redshift Spectrum, EMR, and QuickSight for a governed analytics ecosystem.

Q79. What is Amazon WorkMail?

Amazon WorkMail is a secure, managed business email and calendar service with support for existing desktop and mobile email client applications. WorkMail supports Microsoft Outlook on Windows and macOS, iOS and Android mail clients, and any IMAP-compatible email clients. It provides message journaling, email flow rules, mobile device management policies, and integration with WorkDocs and WorkSpaces for a complete managed workplace solution.

Q80. What is AWS Artifact?

AWS Artifact is your go-to central resource for compliance-related information about AWS services. Artifact provides on-demand access to AWS security and compliance reports (SOC 1, SOC 2, SOC 3, PCI DSS, ISO 27001, ISO 9001, FedRAMP, HIPAA eligibility, GDPR DPA) and online agreements (Business Associate Addendum for HIPAA, GDPR DPA). Reports are automatically kept up-to-date and can be shared with auditors and regulators.

Intermediate Questions (81-150)

Q81. Explain EC2 instance metadata and user data.

EC2 instance metadata is data about your instance that you can access from within the running instance at the URL http://169.254.169.254/latest/meta-data/. Metadata includes instance ID, AMI ID, instance type, security groups, public/private IP, and IAM role credentials. Instance user data is provided at launch (up to 16 KB) and is accessible at http://169.254.169.254/latest/user-data/; it runs as root at instance startup for bootstrapping (installing software, configuring services). IMDSv2 requires session-oriented requests for enhanced security against SSRF attacks.

Q82. How does S3 versioning work?

S3 versioning preserves, retrieves, and restores every version of every object stored in a bucket. When versioning is enabled, every overwrite or delete creates a new version rather than replacing or deleting the object; deletions result in a delete marker being added, not actual removal. Versioning cannot be disabled once enabled—only suspended. MFA Delete requires MFA authentication to permanently delete object versions or change versioning state, providing an additional security layer.

Q83. What are S3 lifecycle policies?

S3 lifecycle policies automate the transition of objects between storage classes and the expiration (deletion) of objects based on age or other criteria. Transition actions move objects from Standard to Standard-IA (minimum 30 days), then to Glacier Flexible Retrieval (minimum 90 days total), then to Glacier Deep Archive; expiration actions permanently delete objects or delete expired delete markers. Lifecycle rules can filter by prefix, object tags, or object size to apply policies to subsets of objects.

Q84. Explain IAM policies: identity-based vs. resource-based.

Identity-based policies are JSON documents attached to IAM identities (users, groups, roles) that grant or deny permissions to perform actions on resources. Resource-based policies are JSON documents attached to resources (S3 buckets, SQS queues, KMS keys, Lambda functions) that specify who (principals) can access the resource. When both exist, the effective permissions are the union, unless there is an explicit Deny. Resource-based policies also enable cross-account access without assuming a role.

Q85. What is an IAM role and when would you use it?

An IAM role is an IAM identity with specific permissions that can be assumed by trusted entities—AWS services, IAM users in the same or different account, or web identity/SAML federated identities. Roles do not have permanent credentials; assuming a role provides temporary security credentials (via STS). Common uses: EC2 instance profile for S3/DynamoDB access (no hardcoded credentials), Lambda execution role, cross-account access, and federated identity for employees accessing AWS via corporate SSO.

Q86. What are VPC Security Groups vs. Network ACLs?

Security Groups are stateful virtual firewalls at the instance/ENI level—return traffic for allowed inbound traffic is automatically allowed outbound regardless of outbound rules. Network ACLs are stateless firewalls at the subnet level—you must explicitly allow both inbound and outbound traffic for each direction. Security Groups support only Allow rules (implicit deny); NACLs support both Allow and Deny rules evaluated in numbered order. Security Groups are the first line of defense for instance-level control; NACLs provide subnet-level control.

Q87. Explain the difference between horizontal and vertical scaling on AWS.

Vertical scaling (scale up) increases the size of an instance (more CPU, RAM, storage) by stopping the instance, changing the instance type, and restarting—causing downtime and having physical limits. Horizontal scaling (scale out) adds more instances to distribute load, achieved automatically with Auto Scaling Groups and load balancers—no downtime and virtually unlimited scale. AWS architecture best practices favor horizontal scaling for stateless applications to achieve elastic, fault-tolerant designs.

Q88. What is RDS Multi-AZ vs. Read Replica?

RDS Multi-AZ creates a synchronous standby replica in a different AZ for high availability and automatic failover (typically 60-120 seconds)—the standby is not accessible for reads. Read Replicas use asynchronous replication and can be read by application read traffic to offload the primary instance; they support cross-region replication and can be promoted to standalone DB instances. Multi-AZ is for durability/availability; Read Replicas are for read scalability. Aurora uses a different shared storage model achieving both benefits natively.

Q89. How does DynamoDB achieve single-digit millisecond performance?

DynamoDB achieves single-digit millisecond performance through its distributed, SSD-backed storage architecture, consistent hashing for partition key distribution across storage nodes, and in-memory buffering of writes. All data is automatically replicated across three AZs synchronously before acknowledging writes. Provisioned throughput (RCUs/WCUs) reserves capacity on nodes for predictable performance; DynamoDB Accelerator (DAX) is an in-memory cache delivering microsecond response times for read-heavy workloads.

Q90. Explain Lambda cold start and mitigation strategies.

A Lambda cold start occurs when a new execution environment is initialized for a function invocation—AWS must allocate infrastructure, start the runtime, and execute initialization code outside the handler. Cold starts typically add 100ms-1s latency depending on runtime (Java/C# are worse) and package size. Mitigation strategies include: Provisioned Concurrency (pre-initialized environments ready to respond), keeping functions warm with scheduled pings, minimizing deployment package size, avoiding VPC if possible (or using VPC lattice with hyperplane ENIs), and using interpreted runtimes (Python, Node.js) for latency-sensitive functions.

Q91. What are CloudFormation nested stacks and why use them?

CloudFormation nested stacks are stacks created within another CloudFormation stack using the AWS::CloudFormation::Stack resource type, referencing a child template URL in S3. Nested stacks allow decomposition of complex architectures into reusable, manageable components—a root stack orchestrates VPC, security, application, and database child stacks. Cross-stack references via Outputs/ImportValue allow sharing resource ARNs between stacks. Nested stacks help overcome the 500-resource limit per stack and enable team-based template ownership.

Q92. How does CloudFront caching work and how do you invalidate cache?

CloudFront caches content at edge locations based on cache behaviors defined by path patterns. Cache behavior settings control TTL (minimum, default, maximum), query string forwarding, header/cookie forwarding, and compression. Origin Cache-Control and Expires headers control caching; CloudFront respects these or uses configured TTL overrides. Cache invalidation submits an invalidation request specifying paths (/* for all) to remove objects before TTL expiration—first 1,000 invalidation paths per month are free. Versioned URLs (appending version to filenames) are preferred over invalidations for cost efficiency.

Q93. Explain AWS Auto Scaling policies: Target Tracking vs. Step Scaling.

Target Tracking scaling maintains a specified metric at a target value (e.g., keep CPU at 50%)—AWS automatically creates and manages CloudWatch alarms and adjusts capacity to keep the metric at the target. Step Scaling uses CloudWatch alarms and defines scaling adjustments based on the breach size (alarm threshold + adjustment steps); as the metric deviates further from threshold, larger adjustments are applied. Target Tracking is simpler and recommended for most cases; Step Scaling provides more granular control for complex scaling needs.

Q94. What is an Application Load Balancer listener rule?

ALB listener rules define conditions and actions that determine how the ALB routes requests received by the listener. Rule conditions include host-based routing (route api.example.com to one target group, app.example.com to another), path-based routing (/api/* to backend, /* to frontend), HTTP method, query string, source IP, and HTTP headers. Rule actions include forward (to target group or weighted group), redirect (301/302 with configurable URL), and fixed-response (return static response without forwarding). Rules are evaluated in priority order.

Q95. What is VPC Flow Logs?

VPC Flow Logs capture information about IP traffic going to and from network interfaces in your VPC. Flow logs can be published to CloudWatch Logs or S3; they can be created at VPC, subnet, or ENI level. Each flow log record includes source/destination IP, source/destination port, protocol, packet count, byte count, start/end time, and ACCEPT/REJECT action based on security group and NACL decisions. Flow logs are essential for network troubleshooting, security analysis, and compliance auditing.

Q96. How does AWS STS (Security Token Service) work?

AWS Security Token Service (STS) issues temporary, limited-privilege credentials for IAM users or federated users. STS operations include AssumeRole (for role switching and cross-account access), AssumeRoleWithWebIdentity (for OpenID Connect identity providers like Google, Cognito), AssumeRoleWithSAML (for SAML 2.0 federation), GetFederationToken, and GetSessionToken (for MFA-protected API calls). Temporary credentials include access key ID, secret access key, and session token, expiring in 15 minutes to 12 hours.

Q97. Explain S3 Transfer Acceleration.

Amazon S3 Transfer Acceleration uses CloudFront's globally distributed edge locations to accelerate uploads to S3 over long distances. Clients upload to the nearest CloudFront edge location using an accelerated endpoint (bucket.s3-accelerate.amazonaws.com); data is then transferred over AWS's optimized network backbone to the S3 bucket. Transfer Acceleration is beneficial when uploading from geographically distant clients; the Speed Comparison tool lets you test whether acceleration provides a speed benefit before enabling.

Q98. What is AWS Global Accelerator?

AWS Global Accelerator is a networking service that improves the availability and performance of applications for global users by routing traffic through AWS's global network infrastructure rather than the public internet. Global Accelerator provides two static Anycast IP addresses as a fixed entry point; traffic is directed to the optimal regional AWS endpoint (ALB, NLB, EC2) based on health, geography, and routing policies. Unlike CloudFront (cache-based CDN), Global Accelerator is suited for non-cacheable content, gaming, IoT, and dynamic applications requiring consistent performance.

Q99. What are DynamoDB Global Tables?

DynamoDB Global Tables provide a fully managed, multi-region, multi-active replication solution enabling reading and writing to any replica table in any selected AWS region. Global Tables use last-writer-wins reconciliation for concurrent writes and typically achieve sub-second replication between regions. Use cases include globally distributed applications requiring low-latency local reads and writes, disaster recovery, and regional data sovereignty compliance. Version 2019.11.21 (current) supports adding replicas without table re-creation and uses on-demand/provisioned billing.

Q100. Explain Amazon ECS task definition components.

An ECS task definition is a JSON blueprint describing application containers. Key components include: container definitions (Docker image, CPU/memory, port mappings, environment variables, secrets from Secrets Manager/SSM), task-level CPU and memory (for Fargate), task IAM role (permissions for containers), execution role (ECS agent permissions to pull images and write logs), network mode (awsvpc for Fargate, bridge/host for EC2), volumes (EFS mounts, bind mounts), and logging configuration (awslogs driver for CloudWatch). Task definitions are versioned; services reference specific revisions.

Q101. What is AWS X-Ray?

AWS X-Ray helps developers analyze and debug distributed applications in production or in development, such as microservices architectures. X-Ray collects data about requests that your application serves and provides tools for viewing, filtering, and gaining insights into that data to identify issues and optimization opportunities. X-Ray traces show the path of a request through your application with segments, subsegments, and annotations; the service map visualizes dependencies and response time distributions.

Q102. What is Amazon EventBridge vs. SNS vs. SQS?

SNS is a pub/sub service for fan-out scenarios where one message goes to multiple subscribers (push model)—great for notifications and parallel processing. SQS is a message queue for decoupling producers and consumers with buffering (pull model)—ideal for load leveling and ensuring messages are processed. EventBridge is an event router with rich filtering and routing based on event patterns, supporting 200+ AWS services and SaaS sources—ideal for event-driven architectures. Often combined: EventBridge routes events to SNS (fan-out) and SQS (buffering) for complementary functions.

Q103. How does AWS cost allocation work with tags?

AWS cost allocation tags allow you to assign metadata (key-value pairs) to AWS resources, enabling cost tracking by project, team, environment, or cost center in Cost Explorer and AWS Cost and Usage Reports. Tags must be activated in the Billing console as cost allocation tags; AWS-generated tags (aws:createdBy) are also available. Combined with AWS Budgets (budget alerts and actions), Cost Anomaly Detection (ML-based spending anomalies), and Savings Plans coverage reports, tagging enables granular FinOps practices.

Q104. Explain the differences between S3 storage classes.

S3 Standard offers 99.99% availability and 11 nines durability for frequently accessed data with no retrieval fee. Standard-IA (Infrequent Access) has a lower storage price but per-GB retrieval fee—suitable for data accessed monthly. One Zone-IA stores in a single AZ (20% cheaper) for reproducible data. Intelligent-Tiering automatically moves objects between access tiers based on usage patterns without retrieval fees. Glacier Instant Retrieval offers archive pricing with millisecond access; Glacier Flexible Retrieval and Deep Archive provide lowest-cost storage for rarely accessed compliance archives with hour-scale retrieval.

Q105. What is CloudWatch Logs Insights?

CloudWatch Logs Insights is an interactive, pay-per-query log analytics service enabling you to search, analyze, and visualize log data using a purpose-built query language. Queries support filtering (filter @message like /ERROR/), aggregation (stats count(*) by bin(5m)), sorting, limiting, and parsing structured/unstructured log formats. Query results can be added to CloudWatch dashboards; Contributor Insights analyzes log patterns to identify top contributors to operational metrics. Insights automatically discovers log fields in JSON and common log formats.

Q106. What is Amazon EKS Fargate profile?

An EKS Fargate profile defines which pods run on AWS Fargate by specifying namespace and label selectors—matching pods are scheduled on Fargate without managing EC2 nodes. Fargate for EKS provides right-sized, isolated compute per pod with vCPU and memory configuration at the pod level. Limitations include no support for DaemonSets, stateful workloads requiring local storage, privileged containers, or custom AMIs; each Fargate pod gets an ENI and runs in a dedicated micro-VM for security isolation.

Q107. Explain AWS WAF rate-based rules.

AWS WAF rate-based rules automatically block IP addresses that send requests at a rate exceeding a defined threshold (minimum 100 requests per 5-minute window) to protect against DDoS attacks and brute-force login attempts. Rate limits can be applied per IP or per forwarded-IP (when behind a proxy); rate-based rules can include scope-down statements to count only requests matching specific conditions (path, header). Once the rate drops below the threshold, the block is automatically lifted after the evaluation period.

Q108. What is AWS CodeArtifact?

AWS CodeArtifact is a fully managed artifact repository service that makes it easy to securely store, publish, and share software packages used in development. CodeArtifact supports npm, PyPI, Maven, Gradle, NuGet, Swift, and generic package formats. Upstream repositories allow pulling packages from public repositories (npmjs.com, PyPI, Maven Central) through CodeArtifact, caching packages locally; domain-level sharing allows multiple repositories across accounts to share packages with consistent IAM controls.

Q109. How does Amazon Route 53 health checking work?

Route 53 health checkers are located around the world and send requests to your endpoints at the interval you specify (default 30 seconds, fast 10 seconds). Health checks monitor endpoints (HTTP, HTTPS, TCP), other health checks (calculated health checks combining multiple child checks), and CloudWatch alarms. DNS failover routing policies use health check status to route traffic—active-active (multiple healthy records), active-passive (primary route with failover to secondary when primary is unhealthy), and private hosted zone health checks using CloudWatch alarms.

Q110. What is AWS PrivateLink?

AWS PrivateLink provides private connectivity between VPCs, AWS services, and on-premises applications without exposing traffic to the public internet. PrivateLink uses Interface VPC Endpoints—ENIs with private IP addresses in your VPC that serve as entry points for traffic destined to supported AWS services. Endpoint services allow you to expose your own service to other VPCs/accounts via PrivateLink, enabling private SaaS delivery. PrivateLink avoids bandwidth constraints and security exposure of VPC peering and transit gateway for service-specific connectivity.

Q111. What is S3 Object Lock?

S3 Object Lock prevents objects from being deleted or overwritten for a fixed amount of time or indefinitely, meeting regulatory requirements for WORM (Write Once, Read Many) storage. Retention modes include Governance (authorized users can override) and Compliance (no one, including root, can delete/overwrite until retention expires). Legal Hold places an indefinite lock on individual object versions regardless of retention period. Object Lock requires versioning-enabled buckets and must be enabled at bucket creation.

Q112. Explain DynamoDB partition key design best practices.

Effective DynamoDB partition key design maximizes data distribution across partitions to avoid hot partitions (throttling). Best practices include using high-cardinality attributes (user IDs, UUIDs), adding a random suffix/prefix for write-heavy uniform distribution (write sharding), using composite keys (partition key + sort key) for hierarchical data models, and prefixing partition keys with the data type for single-table design. Avoid low-cardinality keys (status flags), sequential IDs (monotonically increasing creates hot partitions), and business-meaningful keys that might have uneven distribution.

Q113. What is Amazon Cognito user pool vs. identity pool?

Cognito User Pools are user directories that handle user registration, authentication (username/password, social/SAML/OIDC federation), MFA, account recovery, and email/phone verification—returning JWT tokens (ID, access, refresh) after authentication. Cognito Identity Pools (Federated Identities) exchange authenticated identities (from User Pools, social providers, SAML, developer-authenticated) for temporary AWS credentials via STS, enabling direct AWS service access from client apps. They are complementary: User Pools authenticate users; Identity Pools authorize AWS resource access.

Q114. How does VPC peering affect DNS resolution?

By default, DNS hostnames and DNS resolution for resources in a peered VPC are resolved using each VPC's local DNS. To enable resolution of private DNS hostnames across a peering connection, you must enable DNS resolution support for the peering connection in both VPCs—this allows instances in one VPC to resolve private DNS hostnames of instances in the peered VPC to private IP addresses. Route 53 Resolver endpoints (inbound/outbound) handle DNS resolution for hybrid environments with on-premises DNS servers.

Q115. What is Amazon MSK connect?

Amazon MSK Connect is a feature of Amazon MSK that makes it easy to run fully managed Apache Kafka Connect workloads on AWS. MSK Connect automatically scales resources for Kafka Connect workers and enables use of community-built or custom connectors to stream data between Apache Kafka and external systems (databases, S3, Elasticsearch/OpenSearch, Salesforce). Connector plugins are stored in S3; MSK Connect monitors connector health and restarts failed tasks automatically without operational overhead.

Q116. What is the difference between AWS Secrets Manager and SSM Parameter Store?

Both store sensitive configuration data, but differ in features and cost. Secrets Manager costs $0.40/secret/month and provides native automatic rotation for supported databases, cross-account sharing, and resource-based policies. SSM Parameter Store has a free tier for standard parameters and $0.05/10,000 API interactions for advanced parameters; supports hierarchical parameter paths and versioning but requires custom Lambda functions for rotation. Use Secrets Manager for database credentials requiring rotation; Parameter Store for configuration data, feature flags, and non-sensitive values.

Q117. What is Amazon Kinesis Data Streams vs. Kinesis Data Firehose?

Kinesis Data Streams (KDS) is a real-time, ordered streaming service with configurable retention (1-365 days) where you manage consumers (custom applications using KCL or SDK, Lambda, Kinesis Data Analytics). KDS requires managing shard count (each shard: 1 MB/s write, 2 MB/s read) or using on-demand mode. Kinesis Data Firehose is fully managed with no consumer management—it buffers, transforms, and delivers data directly to S3, Redshift, OpenSearch, Splunk, or HTTP endpoints without writing consumer code. KDS is for real-time custom processing; Firehose is for simple reliable delivery to storage/analytics.

Q118. Explain AWS IAM Permission Boundaries.

IAM Permission Boundaries are managed policies attached to IAM entities (users/roles) that set the maximum permissions the entity can have—they define the boundary beyond which no policy can grant permissions. Even if an identity-based policy grants broader permissions, the effective permissions are limited to the intersection of the identity policy and the permission boundary. Boundaries are used to delegate permission management safely—developers can create roles for their applications but cannot grant permissions outside the boundary set by administrators.

Q119. What is Amazon WorkDocs?

Amazon WorkDocs is a fully managed, secure content creation, storage, and collaboration service. WorkDocs supports real-time document collaboration, version control, commenting, feedback workflows, and cross-platform access (web, desktop, mobile). The WorkDocs SDK enables developers to build custom applications and integrations; WorkDocs Drive mounts WorkDocs as a local drive. WorkDocs integrates with WorkSpaces, Active Directory, and supports SharePoint migration and SSO via IAM Identity Center.

Q120. How does Amazon SQS visibility timeout work?

When a consumer receives an SQS message, the message becomes invisible to other consumers for the visibility timeout period (default 30 seconds, max 12 hours). If the consumer processes and deletes the message before timeout, processing completes successfully. If the consumer fails or crashes, the message reappears in the queue after the timeout for reprocessing. Consumers can extend visibility timeout for long processing tasks using ChangeMessageVisibility. Dead-letter queues (DLQ) capture messages that fail after the maxReceiveCount threshold, isolating poison pill messages.

Q121. What is Amazon Elastic Transcoder vs. AWS Elemental MediaConvert?

Amazon Elastic Transcoder is an older, simpler media transcoding service for converting video files stored in S3 to formats playable on various devices; it's still available but AWS recommends MediaConvert for new projects. AWS Elemental MediaConvert is a file-based video transcoding service supporting a broader range of input/output formats, advanced encoding features (HDR, Dolby Vision, Dolby Atmos), captions processing, DRM, and ad insertion. MediaConvert provides better performance, more features, and is the recommended choice for professional broadcast and OTT workflows.

Q122. What is Amazon Textract?

Amazon Textract is a machine learning service that automatically extracts text and data from scanned documents, going beyond simple OCR to identify the contents of fields in forms and information stored in tables. Textract returns extracted text, form key-value pairs, table structures, and confidence scores; Queries API allows targeted extraction of specific information. Textract integrates with Augmented AI (A2I) for human review of low-confidence extractions and with Comprehend for NLP analysis of extracted text.

Q123. Explain AWS Glue Data Catalog.

The AWS Glue Data Catalog is a central metadata repository for all your data assets across your organization, compatible with the Apache Hive Metastore API. The catalog stores table definitions, schemas (column names and types), partition information, and data location for structured and semi-structured data in S3, JDBC databases, and other stores. Glue crawlers automatically populate the catalog by scanning data sources and inferring schemas; the catalog is used by Athena, Redshift Spectrum, EMR, and Glue ETL jobs as their metastore.

Q124. What is CloudFormation drift detection?

CloudFormation drift detection identifies differences between the expected stack resource configuration (defined in the template) and the actual configuration of the live resources (which may have been modified outside CloudFormation). Drift detection compares each resource's expected properties against its actual properties and reports resources as IN_SYNC, MODIFIED, DELETED, or NOT_CHECKED. Drift detection helps enforce infrastructure-as-code discipline and identify unauthorized or accidental manual changes that could cause stack update failures.

Q125. What is Amazon Comprehend?

Amazon Comprehend is a natural language processing (NLP) service that uses ML to find insights and relationships in text. Comprehend can identify the language of the text, extract key phrases and named entities (persons, organizations, locations, dates), determine sentiment (positive, negative, neutral, mixed), and classify documents into custom categories. Comprehend Medical specializes in clinical text—extracting medical entities, medications, diagnoses, and PHI. Comprehend Custom enables training custom entity recognition and classification models.

Q126. How does Amazon EBS snapshot work and what is EBS fast snapshot restore?

EBS snapshots are incremental backups stored in S3—the first snapshot copies all data, subsequent snapshots store only changed blocks since the last snapshot. Snapshots can be copied across regions, shared with other accounts, and used to create new EBS volumes or AMIs. EBS Fast Snapshot Restore (FSR) eliminates the latency of I/O operations on a volume initialized from a snapshot (lazy loading phenomenon), enabling full performance immediately. FSR can be enabled per snapshot per AZ; it incurs additional charges per enabled AZ-hour.

Q127. What is AWS Config and how does it differ from CloudTrail?

AWS Config provides a detailed inventory of your AWS resources and records configuration changes over time, evaluating resources against desired configurations using Config Rules. Config answers the question "What did my AWS resource look like at time X?" and "Is this resource compliant with my policies?" CloudTrail records API call history answering "Who made what API call when?" Config Rules use managed rules (pre-built) or custom Lambda functions to detect non-compliant resources; remediation actions can automatically correct violations.

Q128. What is Amazon ECS Service Discovery?

ECS Service Discovery uses AWS Cloud Map to automatically register ECS tasks with DNS or Cloud Map service registry, enabling services to discover each other by name rather than hardcoded IPs. When tasks launch, ECS automatically creates DNS records (A records for awsvpc network mode) in a Route 53 private hosted zone; health checking removes unhealthy tasks from the registry. Service Connect (newer) provides simpler service-to-service communication with built-in load balancing and observability through CloudWatch metrics without managing Cloud Map directly.

Q129. Explain AWS Cost Explorer vs. AWS Budgets.

AWS Cost Explorer is a visualization tool for analyzing historical AWS spending and usage patterns—it provides pre-built reports, custom report creation with filters/grouping, Cost Explorer API for programmatic access, and Rightsizing Recommendations for underutilized EC2/RDS. AWS Budgets enables proactive cost control by setting custom thresholds on cost, usage, RI coverage, RI utilization, Savings Plans coverage, and triggering alerts (email, SNS) or automated actions (apply SCPs, apply IAM policies, stop EC2/RDS instances) when thresholds are breached. Use Cost Explorer for analysis; Budgets for governance.

Q130. What is a Launch Template vs. Launch Configuration?

Launch Templates are the modern replacement for Launch Configurations in Auto Scaling Groups and EC2 Fleet. Launch Templates support versioning (maintain multiple versions, designate a default), the full range of EC2 instance types including Spot and On-Demand mix configurations, T2/T3 Unlimited credit specifications, placement groups, capacity reservations, and Elastic GPUs. Launch Configurations are immutable (cannot be modified), don't support versioning, and are being deprecated. AWS recommends migrating ASGs from Launch Configurations to Launch Templates.

Q131. What is Amazon Translate?

Amazon Translate is a neural machine translation service that delivers fast, high-quality, affordable, and customizable language translation. It supports real-time translation (TranslateText API) and asynchronous batch translation of documents in S3 (StartTextTranslationJob). Custom terminology allows you to specify how domain-specific terms (brand names, technical terms) should be translated; parallel data enables customization of translation style and domain. Translate supports 75+ languages and integrates with Comprehend for sentiment-aware translation workflows.

Q132. What is AWS IoT Core?

AWS IoT Core is a managed cloud service that lets connected devices interact with cloud applications and other devices securely and reliably. IoT Core supports MQTT, HTTPS, and WebSocket protocols; devices authenticate using X.509 certificates or custom authorizers. Key components include the Message Broker (pub/sub), Rules Engine (routes messages to AWS services via SQL-like rules), Device Shadow (persistent virtual representation of device state), and Fleet Indexing for querying device metadata. IoT Core integrates with Greengrass for edge computing.

Q133. Explain CloudFront signed URLs vs. signed cookies.

CloudFront signed URLs restrict access to individual files or dynamic content—suitable when serving one specific file and tracking which users access it. Signed cookies restrict access to multiple files or an entire website under a path pattern—suitable when protecting a streaming video catalog or multiple files without changing URLs. Both use RSA key pairs; key groups (managed in CloudFront) replace deprecated CloudFront key pairs. Signed URLs override cookie-based access; use signed URLs for downloadable files and signed cookies for premium subscription content.

Q134. What is Amazon Neptune?

Amazon Neptune is a fast, reliable, fully managed graph database service for building and running applications that work with highly connected datasets. Neptune supports two popular graph query languages: Apache TinkerPop Gremlin (property graph) and W3C SPARQL (RDF graph). Use cases include knowledge graphs, fraud detection, social networks, recommendation engines, and network/IT operations. Neptune is highly available with read replicas, point-in-time recovery, continuous backup to S3, and encryption at rest and in transit.

Q135. What is Amazon Quantum Ledger Database (QLDB)?

Amazon QLDB is a fully managed ledger database that provides a transparent, immutable, and cryptographically verifiable transaction log owned by a central trusted authority. QLDB uses a journal structure where data changes are appended but never modified or deleted; the complete history of all changes is queryable using PartiQL (SQL-compatible). Hash chaining ensures cryptographic verification of the data history; document-oriented data model with ACID transactions supports complex queries. Use cases include supply chain, regulatory compliance, and financial systems requiring audit trails.

Q136. What is Amazon DocumentDB?

Amazon DocumentDB (with MongoDB compatibility) is a fully managed document database service that supports MongoDB workloads. DocumentDB implements the Apache 2.0 open-source MongoDB 3.6, 4.0, and 5.0 APIs; applications using MongoDB drivers can connect to DocumentDB without code changes for most operations. DocumentDB separates compute from storage, providing up to 15 low-latency read replicas, continuous backup to S3 with point-in-time recovery, and automatic storage scaling up to 64 TB. Elastic Clusters support millions of reads/writes per second with elastic storage.

Q137. How does Amazon ECR vulnerability scanning work?

ECR supports two scanning types: Basic scanning uses the open-source Clair database for common CVE detection in OS packages; Enhanced scanning uses Amazon Inspector with the Snyk database for deeper analysis including application programming language packages (npm, pip, gem) and OS packages. Basic scanning can be configured for on-push or manual scanning; Enhanced scanning provides continuous scanning that re-evaluates images when new CVEs are published. Findings are available via ECR console, API, EventBridge events, and Inspector console.

Q138. What is AWS Certificate Manager (ACM)?

AWS Certificate Manager handles the complexity of creating, storing, and renewing SSL/TLS certificates for use with AWS services and internal connected resources. ACM provides free public certificates for use with ALB, CloudFront, API Gateway, and other integrated services; private certificates are available via ACM Private CA (priced service). ACM automatically renews managed certificates before expiration; certificates cannot be exported (private key stays in AWS)—for export, use ACM Private CA or import third-party certificates.

Q139. Explain EC2 Placement Groups.

EC2 Placement Groups influence the placement of instances on underlying hardware for specific use cases. Cluster placement groups pack instances closely within a single AZ for low-latency, high-throughput HPC and big data workloads (up to 10 Gbps inter-instance bandwidth). Spread placement groups place instances on distinct hardware racks across AZs for critical applications requiring maximum fault isolation (max 7 instances per AZ per group). Partition placement groups divide instances into logical partitions on separate racks for large distributed systems (Hadoop, Cassandra, Kafka) with partition-level fault isolation.

Q140. What is AWS Outposts?

AWS Outposts is a family of fully managed solutions delivering AWS infrastructure and services to virtually any on-premises or edge location for a truly consistent hybrid experience. Outposts bring native AWS services (EC2, ECS, EKS, RDS, EMR), APIs, and tools to any data center, co-location space, or on-premises facility. Outposts integrate with the nearest AWS Region for management plane operations, cloud burst, backup, and services not available locally. Use cases include data residency requirements, local data processing, and low-latency on-premises applications.

Q141. What is Amazon Forecast?

Amazon Forecast is a fully managed time series forecasting service that uses statistical and ML algorithms (including DeepAR+, NPTS, CNN-QR, ETS, ARIMA, Prophet) to deliver high-accuracy forecasts. Forecast learns from your historical time series data and related data (prices, promotions, holidays, weather) to automatically train and tune models. Use cases include retail demand planning, supply chain forecasting, capacity planning, financial planning, and workforce planning. AutoPredictor automatically selects and combines the best algorithms for your dataset.

Q142. What is Amazon Personalize?

Amazon Personalize is a fully managed ML service that makes it easy to build applications delivering highly customized real-time product and content recommendations. Personalize uses the same technology as Amazon.com's recommendation system—you provide interaction data (clickstreams, purchases), item catalogs, and user metadata; Personalize trains recommendation models and delivers results via API. Use cases include e-commerce product recommendations, video/content next-watch, personalized re-ranking, and similar items. No ML expertise is required.

Q143. Explain Amazon EFS performance modes.

EFS offers two performance modes set at creation. General Purpose mode (default) has lower per-operation latency and is recommended for latency-sensitive use cases like web serving, content management, and home directories—supports Elastic Throughput, Bursting, and Provisioned Throughput modes. Max I/O mode supports higher aggregate throughput and IOPS with slightly higher latencies—suited for big data and media processing applications with highly parallelized access from thousands of clients. The Elastic Throughput mode (default for new file systems) scales throughput automatically with workload; Bursting Throughput uses a credit system based on storage size.

Q144. What is AWS Wavelength?

AWS Wavelength embeds AWS compute and storage services within telecommunications providers' 5G networks, enabling mobile edge computing applications with single-digit millisecond latencies to mobile devices and end users. Wavelength Zones are AWS infrastructure deployments at the edge of 5G networks (Verizon, KDDI, SK Telecom, Vodafone); developers can deploy applications using familiar AWS services and APIs. Use cases include real-time gaming, AR/VR, autonomous vehicles, smart factory IoT, and live video streaming requiring ultra-low latency to mobile clients.

Q145. What is Amazon Kendra?

Amazon Kendra is an intelligent enterprise search service powered by ML that enables your organization's employees to search across all your content repositories with natural language queries. Kendra indexes documents from over 40 data connectors (S3, SharePoint, Confluence, ServiceNow, Salesforce, Databases, web crawlers) and returns precise answers (not just links) by understanding the context and meaning of questions. Features include FAQs, document excerpts, relevance tuning, incremental learning, and integration with Lex for conversational search.

Q146. What is AWS Batch?

AWS Batch enables developers, scientists, and engineers to easily and efficiently run hundreds of thousands of batch computing jobs on AWS. Batch dynamically provisions compute resources (EC2 or Fargate) based on the volume and resource requirements of submitted jobs, defined in job definitions. Job queues with priority levels route jobs to compute environments; Array jobs process large datasets as multiple child jobs; multi-node parallel jobs support distributed computing (MPI workloads). Batch integrates with Step Functions for complex workflow orchestration.

Q147. Explain Amazon Fraud Detector.

Amazon Fraud Detector is a fully managed service using ML to identify potentially fraudulent online activities—account takeovers, online payment fraud, fake account creation, and more. Fraud Detector uses your historical fraud data and Amazon's 20+ years of fraud detection expertise to build customized models without ML expertise. Models (Online Fraud Insights, Transaction Fraud Insights, Account Takeover Insights) are deployed as detection API endpoints returning fraud scores and outcomes; rules and outcome logic allow threshold-based decisions integrated with application workflows.

Q148. What is Amazon Polly?

Amazon Polly is a text-to-speech (TTS) service that uses advanced deep learning technologies to synthesize natural-sounding human speech. Polly supports 60+ voices across 29 languages and provides two synthesis engines: Standard TTS (concatenative approach) and Neural TTS (NTTS, more natural-sounding with superior quality). Neural voices support Newscaster style and Conversational style for specific use cases. SSML (Speech Synthesis Markup Language) tags control speech characteristics—pauses, emphasis, pronunciation, prosody, voice mixing.

Q149. What is Amazon Connect?

Amazon Connect is an easy-to-use omnichannel cloud contact center service enabling companies to deliver superior customer service at lower cost. Connect provides a self-service graphical interface to design contact flows (drag-and-drop), manage agents, and track performance metrics. Features include natural language processing via Lex chatbots, real-time and post-call analytics via Contact Lens (transcription, sentiment analysis, keyword alerts), agent assist suggestions, and outbound campaigns. Connect integrates with CRM systems (Salesforce, Zendesk) and Lambda for custom logic.

Q150. What is Amazon Pinpoint?

Amazon Pinpoint is a flexible and scalable outbound and inbound marketing communications service for customer engagement. Pinpoint enables targeted campaigns across channels: email, SMS, push notifications, voice messages, and in-app messaging. Analytics capabilities include campaign metrics, journey analytics (multi-step engagement workflows), and customer segment analysis. Pinpoint integrates with SES for email, SNS for push/SMS, and Lambda/EventBridge for real-time event-driven communication triggers based on customer behavior.

Advanced Questions (151-200)

Q151. How do you design a highly available, multi-region active-active architecture on AWS?

A multi-region active-active architecture routes user traffic to the nearest healthy region using Route 53 latency-based routing with health checks, or Global Accelerator for consistent Anycast IPs. Each region has independent compute (EC2 ASG or ECS Fargate), load balancing (ALB), caching (ElastiCache), and data tiers. DynamoDB Global Tables provides multi-master replication for session/user data; Aurora Global Database provides sub-second cross-region replication for relational data with RPO < 1 second and RTO < 1 minute. S3 Cross-Region Replication synchronizes assets; CloudFront with origin group failover provides CDN-level redundancy. Event-driven architectures use EventBridge's cross-region event buses for decoupled coordination.

Q152. Explain AWS Landing Zone and Control Tower guardrails architecture.

A Landing Zone is a well-architected, multi-account AWS environment based on security and compliance best practices, typically with Management, Log Archive, Audit, and workload accounts organized in OUs. Control Tower provisions the landing zone using AWS Organizations, CloudFormation StackSets, Service Catalog, Config, CloudTrail, IAM Identity Center, and Security Hub. Preventive guardrails are SCPs applied to OUs blocking non-compliant actions (e.g., prevent disabling CloudTrail). Detective guardrails are Config Rules detecting policy violations (e.g., unencrypted EBS volumes). Proactive guardrails are CloudFormation hooks preventing non-compliant resource provisioning in stacks.

Q153. What is the AWS CDK and how does it differ from CloudFormation?

AWS Cloud Development Kit (CDK) is an open-source software development framework for defining cloud infrastructure using familiar programming languages (TypeScript, Python, Java, C#, Go). CDK constructs represent cloud components at three levels: L1 (CfnXxx, direct CloudFormation resource mapping), L2 (higher-level abstractions with sensible defaults, e.g., aws-ec2.Vpc), and L3 (patterns combining multiple resources, e.g., aws-ecs-patterns.ApplicationLoadBalancedFargateService). CDK synthesizes to CloudFormation templates (cdk synth) and deploys with cdk deploy; it enables IDE support (type checking, autocomplete), unit testing with assertions, and reusable construct libraries for standardizing infrastructure patterns across teams.

Q154. Explain the Saga pattern for distributed transactions on AWS.

The Saga pattern manages data consistency across microservices in distributed transactions by breaking a transaction into a sequence of local transactions, each publishing an event or message triggering the next step. If a step fails, compensating transactions undo preceding steps. On AWS, choreography-based sagas use EventBridge/SNS for event-driven coordination (each service subscribes to events and emits results); orchestration-based sagas use Step Functions as a central orchestrator managing the saga workflow with compensation logic. Step Functions Standard Workflows provide exactly-once execution semantics with built-in error handling and compensation support via catch/retry states.

Q155. How do you implement blue/green deployments on AWS?

Blue/green deployments maintain two identical production environments (blue = current, green = new version) with traffic shifted between them. On EC2/ECS: ALB weighted target groups shift traffic incrementally from blue to green ASG/service; CodeDeploy automates traffic shifting with pre/post-traffic hooks for validation. On Lambda: Lambda function aliases with weighted routing shift traffic percentage to new function version; canary and linear deployment configurations provide progressive traffic shifting. On ECS with CodeDeploy: task set-based blue/green with ALB/NLB test listener validates the green environment before production traffic shift. All approaches enable instant rollback by returning traffic to the previous environment.

Q156. What is AWS Service Mesh with App Mesh?

AWS App Mesh is a service mesh that provides application-level networking using the Envoy proxy sidecar to give applications consistent visibility and network traffic controls. App Mesh manages service discovery, traffic routing (weighted routing for canary deployments, retry policies, circuit breaking, timeouts), observability (distributed tracing with X-Ray, metrics, logs), and mutual TLS between services. App Mesh integrates with ECS, EKS, EC2, and on-premises via Envoy; virtual nodes represent services, virtual routers define routing rules. For new deployments, AWS recommends Amazon VPC Lattice, which provides service-to-service connectivity without sidecar proxies.

Q157. Explain DynamoDB Single-Table Design.

Single-table design stores all application entities in a single DynamoDB table, using generic partition key (PK) and sort key (SK) attribute names with entity-type-prefixed values (USER#123, ORDER#456) to model relationships and hierarchies. Access patterns drive the design—overloaded GSIs serve multiple access patterns; adjacency lists model many-to-many relationships. Benefits include reduced operational overhead (one table), efficient single-request fetching of related items using begins_with queries, and avoiding N+1 query problems. The approach requires upfront access pattern analysis; ad-hoc querying requires GSIs or Athena on DynamoDB export to S3.

Q158. What is chaos engineering on AWS and how do you implement it?

Chaos engineering is the discipline of experimenting on a distributed system to build confidence in its ability to withstand turbulent conditions. On AWS, AWS Fault Injection Service (FIS) is the managed chaos engineering service supporting experiments that inject CPU stress, memory pressure, I/O disruption, network latency/packet loss, terminate EC2 instances, stop ECS tasks, and failover RDS/Aurora. Experiments use SSM documents for EC2/ECS actions and native FIS actions for managed services; guardrails (stop conditions using CloudWatch alarms) automatically halt experiments if impact exceeds defined thresholds. The Simian Army (Netflix open-source) concepts like Chaos Monkey (instance termination) are automated through FIS on AWS.

Q159. How do you secure data in S3 against public exposure?

S3 security layers include: Block Public Access settings (account and bucket level) override any policies/ACLs granting public access—enable for all buckets unless explicitly serving public content. Bucket policies with aws:SourceVpc/aws:SourceVpce conditions restrict access to requests from specific VPCs or VPC endpoints. S3 Object Ownership (BucketOwnerEnforced) disables ACLs, simplifying ownership management. Server-side encryption (SSE-S3, SSE-KMS, SSE-C) encrypts all objects at rest. S3 Access Points provide dedicated access endpoints with resource policies for specific workloads/teams. Macie continuously discovers sensitive data. S3 Access Analyzer and IAM Access Analyzer identify unintended public/cross-account access.

Q160. Explain AWS Well-Architected Framework: Reliability Pillar patterns.

The Reliability pillar focuses on workloads performing intended functions correctly and consistently. Key patterns include: using multiple AZs/regions for fault isolation with circuit breakers (bulkhead pattern); managing quotas and constraints with AWS Service Quotas monitoring; testing recovery procedures with fault injection; deploying immutable infrastructure for consistent rollout; using exponential backoff with jitter for retry logic to avoid thundering herd; automating recovery with Auto Scaling, CloudWatch alarms, and EventBridge; implementing health checks at all layers; designing stateless applications for horizontal scaling; and using global services (Route 53, CloudFront, IAM) for global durability.

Q161. What is AWS Nitro System?

The AWS Nitro System is the underlying platform for the latest generation of EC2 instances, offloading virtualization functions to dedicated hardware and software (Nitro Cards for VPC networking, EBS, and instance storage; Nitro Security Chip; Nitro Hypervisor). By offloading hypervisor functions, Nitro provides near bare-metal performance with all host resources available for customer instances. Nitro Enclaves provide isolated compute environments within EC2 instances for processing sensitive data (cryptographic operations, PII, ML inference on sensitive data) with no persistent storage and no network access—accessible only through a local vsock.

Q162. How does Amazon Aurora DSQL work?

Amazon Aurora DSQL (Distributed SQL) is a serverless distributed SQL database designed for globally distributed, highly available OLTP workloads with active-active multi-region architecture. DSQL uses an optimistic concurrency control model—transactions read without locking and only check for conflicts at commit time, achieving high throughput. It provides PostgreSQL-compatible API, automatic scaling from zero, and eliminates the complexity of managing distributed SQL infrastructure. DSQL decouples storage and compute, using Journal-based storage for high durability and low-latency cross-region replication.

Q163. Explain the event-driven architecture patterns on AWS.

Event-driven architectures (EDA) on AWS use events as first-class citizens to decouple services, improve scalability, and enable loose coupling. Core patterns include: Event Notification (SNS/EventBridge fan-out to multiple consumers without coupling); Event-Carried State Transfer (events contain full state changes, eliminating source queries); Event Sourcing (store state as immutable event log in Kinesis or DynamoDB, reconstruct state by replaying events); CQRS (separate read models, fed by event streams, from write models); Choreography (each service reacts to events independently) vs. Orchestration (Step Functions coordinates service interactions). EventBridge Schema Registry and Pipes reduce integration boilerplate.

Q164. What is AWS Graviton and what are its advantages?

AWS Graviton processors are custom ARM-based processors designed by AWS using 64-bit Arm Neoverse technology. Graviton3 (current generation) delivers up to 25% better compute performance, 2x higher floating-point performance, and 2x faster cryptographic operations compared to comparable x86 instances, at up to 20% lower cost. Graviton instances are available across EC2 (M7g, C7g, R7g), RDS, ElastiCache, Lambda, ECS/Fargate, and EKS. Most Linux workloads (containers, microservices, web servers, data processing) run without modification; compiled languages may require ARM-compatible build pipelines.

Q165. How do you implement serverless microservices on AWS?

Serverless microservices on AWS use Lambda functions per service domain, API Gateway for HTTP interfaces, DynamoDB for persistence, SQS/EventBridge for async communication, and Cognito/JWT for authentication. Service isolation is achieved through separate Lambda functions with dedicated IAM execution roles following least privilege. Shared infrastructure uses CDK/CloudFormation stacks; Lambda Layers share common code (SDKs, utilities) across functions. Cold start optimization uses Provisioned Concurrency for latency-sensitive endpoints, ARM/Graviton-based Lambda for cost/performance, and SnapStart (Java) for faster cold starts. Observability uses X-Ray for distributed tracing, CloudWatch for metrics/logs, and structured JSON logging with correlation IDs.

Q166. Explain VPC networking for hybrid cloud connectivity.

Hybrid cloud connectivity options include: AWS Site-to-Site VPN provides encrypted IPsec tunnels over the internet (up to 1.25 Gbps per tunnel, 2 tunnels for redundancy) between customer gateway (on-premises VPN device) and virtual private gateway/transit gateway. AWS Direct Connect provides dedicated 1/10/100 Gbps private connections (+ hosted connections at sub-Gbps via partners) with consistent low latency. Direct Connect Gateway enables connecting one DX connection to multiple VPCs across regions. Transit Gateway connects VPCs and on-premises networks hub-and-spoke; VPN over Direct Connect provides encrypted private connectivity for compliance. AWS Cloud WAN provides managed global network across regions and on-premises.

Q167. What is AWS DataSync vs. AWS Storage Gateway?

AWS DataSync is an online data transfer service for migrating active datasets or synchronizing data between on-premises/edge storage and AWS storage services—it's designed for one-time or periodic data movement at high speed (up to 10 Gbps) with integrity verification. AWS Storage Gateway is a hybrid cloud storage service providing on-premises applications seamless access to cloud storage—File Gateway (NFS/SMB to S3), Volume Gateway (iSCSI block storage backed by S3/EBS snapshots), and Tape Gateway (virtual tape library backed by S3/Glacier). DataSync is for data transfer/migration; Storage Gateway is for ongoing hybrid access patterns where on-premises apps consume cloud storage transparently.

Q168. Explain AWS Zero Trust security model implementation.

Zero Trust on AWS eliminates implicit trust based on network location, requiring continuous verification of every identity and device. Implementation includes: IAM Identity Center (formerly SSO) for centralized identity with MFA; VPC Verified Access for application access without VPN using identity and device context; Secrets Manager for dynamic credentials eliminating static secrets; PrivateLink for private service access eliminating internet exposure; Security Hub for continuous compliance; GuardDuty for threat detection; CloudTrail for immutable audit logs; Certificate Manager for mutual TLS; AWS Verified Permissions for fine-grained authorization using Cedar policy language; and Network Firewall for micro-segmentation within VPCs.

Q169. What is AWS Proton?

AWS Proton is an automated infrastructure provisioning and deployment service for container and serverless applications that enables platform teams to define and manage standard, reusable infrastructure templates while allowing developers to self-service deploy applications. Platform teams define environment templates (shared infrastructure) and service templates (application-specific resources) using CloudFormation or Terraform; developers choose templates and deploy without deep infrastructure knowledge. Proton manages template versioning and automatically propagates updates to deployed environments and services with controlled rollouts.

Q170. How does Amazon EBS io2 Block Express work?

EBS io2 Block Express is the next generation of high-performance SAN-like EBS storage built on the AWS Nitro System, delivering up to 256,000 IOPS, 4,000 MB/s throughput, and 64 TiB volume size per volume—4x the IOPS, throughput, capacity, and 1,000x more durable (99.999%) than io1. Block Express uses a high-performance EC2-to-EBS NVMe interface, reducing latency to sub-millisecond. It's designed for the most I/O-intensive workloads: SAP HANA, Oracle, Microsoft SQL Server, IBM Db2. Block Express volumes require Nitro-based EC2 instances (R5b, X2idn, X2iedn, Im4gn, Is4gen).

Q171. Explain AWS Lake Formation row-level and column-level security.

Lake Formation provides fine-grained access control for data in the Glue Data Catalog, allowing row-level and column-level security without application code changes. Column-level security restricts which columns an IAM principal can query—unauthorized columns are excluded from query results transparently. Row-level security uses data filter expressions (WHERE clause conditions) applied to principals, ensuring users see only rows matching their filter (e.g., users see only rows where region = their assigned region). Cell-level security combines column and row filters. These controls are enforced at the metadata catalog level and respected by Athena, Redshift Spectrum, EMR, and Glue ETL.

Q172. What is Amazon VPC Lattice?

Amazon VPC Lattice is an application networking service that consistently connects, monitors, and secures communications between services across multiple VPCs and accounts without complex VPC peering, Transit Gateways, or service meshes. Lattice uses a service network abstraction—services register with the service network and clients discover services by name; Lattice handles routing, load balancing, health checks, and auth policies. Auth policies use IAM and Sigv4 for request authentication and authorization; TLS is handled automatically. Lattice replaces App Mesh for service-to-service networking with a managed, sidecar-free approach.

Q173. Explain AWS CodeGuru Reviewer and Profiler.

Amazon CodeGuru Reviewer uses ML trained on millions of code reviews to identify critical security vulnerabilities and hard-to-find bugs in Java and Python code. It integrates with CodeCommit, GitHub, Bitbucket, and GitLab for pull request reviews, identifying issues like resource leaks, concurrency race conditions, input validation gaps, and AWS API misuse (unencrypted data, over-permissive IAM). CodeGuru Profiler identifies the most expensive lines of code in production applications by sampling CPU utilization; it suggests code changes that can reduce compute costs and eliminate latency bottlenecks, with heap summary analysis for memory optimization.

Q174. What is Amazon MemoryDB for Redis?

Amazon MemoryDB for Redis is a Redis-compatible, durable, in-memory database service delivering ultra-fast performance with multi-AZ transactional durability. Unlike ElastiCache Redis (primarily a cache backed by a persistent database), MemoryDB uses a distributed transactional log that provides data durability across AZs—enabling MemoryDB to serve as the primary database for microservices requiring both microsecond reads and durable writes. MemoryDB supports Redis data structures, Redis Cluster for horizontal scaling, and the full Redis API, making migration from ElastiCache straightforward for applications needing primary database guarantees.

Q175. Explain AWS Fargate security architecture.

Fargate security architecture provides isolation at the task level rather than the host level. Each Fargate task runs in its own kernel using microVM technology (Firecracker), providing strong isolation boundaries without sharing underlying host kernel resources between tenants. Task IAM roles provide temporary credentials scoped to task-level AWS API permissions via the ECS metadata credential endpoint (replacing EC2 instance profile credentials). ECS task policies and Fargate ephemeral storage (20 GB encrypted) provide data isolation; VPC networking assigns each task a unique ENI with security group controls. Fargate eliminates OS patch management responsibilities for underlying hosts.

Q176. What is AWS Backup Vault Lock?

AWS Backup Vault Lock implements WORM (Write Once, Read Many) protection for backup vaults, preventing backup jobs from being deleted—even by the root user and AWS—during the lock period. Vault Lock is useful for meeting SEC Rule 17a-4(f), FINRA, CFTC, and HIPAA compliance requirements that mandate immutable backup retention. Cool-off period (1-3 days in governance mode) allows testing before irreversible lock; compliance mode makes locks permanent after cool-off, and AWS cannot override them. Vault Lock integrates with AWS Backup policies and works across all supported backup services.

Q177. How do you perform cost optimization for Lambda functions?

Lambda cost optimization focuses on memory allocation tuning, execution duration reduction, and right-sizing. AWS Lambda Power Tuning (open-source Step Functions state machine) runs the function at different memory settings to find the optimal cost/performance trade-off—higher memory often reduces execution time, reducing total cost. Reducing initialization code (cold start payload) and moving initialization outside the handler (container reuse) reduces billable duration. ARM/Graviton2-based Lambda (arm64 architecture) is 20% cheaper per GB-second. Provisioned Concurrency reduces cold starts for consistent workloads but incurs steady-state charges—evaluate against on-demand cold start impact.

Q178. Explain Amazon EKS cluster networking with VPC CNI.

Amazon VPC CNI (Container Network Interface) plugin is the default networking plugin for EKS, assigning each pod a primary IP address from the VPC CIDR—pods are first-class VPC citizens with VPC routing, security groups, and network policy support. VPC CNI pre-allocates ENIs and secondary IP addresses on EC2 nodes for fast pod startup; the number of IPs per node is bounded by EC2 instance type limits. Security Groups for Pods extends VPC security groups to individual pods (requires trunk ENI support on Nitro instances). Prefix delegation assigns /28 IPv4 prefixes to ENIs, dramatically increasing pod density. Alternative CNIs (Calico, Cilium) provide network policy and eBPF-based performance improvements.

Q179. What is Amazon GuardDuty for EKS?

GuardDuty EKS Protection continuously monitors Kubernetes API server audit logs and EKS runtime activity (container behaviors) for malicious activity and misconfigurations in EKS clusters. EKS Audit Log Monitoring detects suspicious API calls like privilege escalation (binding cluster-admin roles), anonymous API access, unusual pod launch patterns, and cryptomining. EKS Runtime Monitoring uses the GuardDuty security agent (DaemonSet) to detect runtime threats: shell execution in containers, privilege escalation via setuid, network traffic to known malicious IPs, and credential access attempts. EKS Threat Findings integrate with Security Hub and EventBridge for automated remediation workflows.

Q180. Explain AWS Elemental MediaLive and multi-bitrate encoding.

AWS Elemental MediaLive is a broadcast-grade live video processing service that encodes live video streams for broadcast and streaming delivery. MediaLive ingests live video via RTP, RTMP, RTMPS, or from S3/MediaConnect sources and transcodes it into multiple output formats (HLS, DASH, UDP/TS, MS Smooth). Multi-bitrate encoding creates an adaptive bitrate (ABR) ladder—multiple renditions at different resolutions and bitrates (e.g., 1080p 8Mbps, 720p 4Mbps, 480p 2Mbps)—enabling client-side adaptive streaming via HLS or DASH. HEVC/H.265 encoding reduces bandwidth by ~50% compared to H.264 at equivalent quality; HDR (HLG, HDR10, Dolby Vision) support serves premium display devices.

Q181. What is AWS Clean Rooms?

AWS Clean Rooms allows companies and their partners to analyze and collaborate on combined datasets without sharing or revealing underlying raw data. Companies create a clean room (analysis environment), invite collaborators, associate their encrypted data (stored in S3), and create collaboration rules (analysis rules) specifying which queries are allowed and what aggregations/minimum group sizes are required to protect individual data privacy. Clean Rooms uses privacy-enhancing technologies including differential privacy, data masking, and aggregation controls—no raw data leaves each party's account. Use cases include advertising measurement, financial services risk modeling, and healthcare research consortiums.

Q182. Explain Amazon Bedrock and generative AI on AWS.

Amazon Bedrock is a fully managed service providing access to high-performing foundation models (FMs) from Amazon (Titan, Nova) and third parties (Anthropic Claude, Meta Llama, Mistral, Cohere, AI21 Labs, Stability AI) via a single API, without managing infrastructure. Bedrock enables fine-tuning models with proprietary data (fine-tuning and continued pre-training), building agents that dynamically call APIs and data sources (Bedrock Agents with action groups), and creating enterprise knowledge bases for RAG (Retrieval Augmented Generation) using integrated vector stores (OpenSearch Serverless, Aurora pgvector, Pinecone). Bedrock Guardrails implement responsible AI controls (content filtering, PII redaction, grounding checks).

Q183. What is Amazon OpenSearch Serverless?

Amazon OpenSearch Serverless is an on-demand serverless configuration for Amazon OpenSearch Service that automatically scales resources based on your indexing and search workloads—eliminating the need to provision, manage, and scale OpenSearch clusters. Serverless uses OpenSearch Compute Units (OCUs) scaled automatically; search OCUs and indexing OCUs scale independently. Collections (groupings of indices) are the billing unit; two collection types: Search (optimized for low-latency search with S3 data tiering) and Time Series (optimized for log and metrics ingestion). PrivateLink-enabled VPC access, encryption with KMS, and data access policies (SAML and IAM) provide enterprise-grade security.

Q184. How do you implement GitOps on AWS?

GitOps on AWS uses Git as the single source of truth for infrastructure and application deployments, with automated sync between Git state and cluster/infrastructure state. For Kubernetes (EKS), tools like ArgoCD or Flux CD installed as EKS add-ons continuously reconcile the cluster state with manifests stored in CodeCommit/GitHub; AWS App2Container and Amazon EKS Blueprints provide opinionated starting configurations. For infrastructure, CodePipeline triggers CloudFormation/CDK deployments on git commits with automated testing (cdk-nag for compliance, cfn-guard for policy validation). SSP (Service Software Platform) patterns with CDK Pipelines implement multi-account, multi-region GitOps delivery with approval gates.

Q185. Explain Amazon S3 Intelligent-Tiering lifecycle.

S3 Intelligent-Tiering automatically moves objects between access tiers based on changing access patterns without performance impact or operational overhead, charging a small monthly monitoring and automation fee per object. Objects not accessed for 30 days move to Infrequent Access tier (40% lower cost); 90 days → Archive Instant Access (68% lower); optional deeper archive tiers activate asynchronously: Archive Access (90-day threshold, 3-5 hour retrieval) and Deep Archive Access (180-day threshold, 12-hour retrieval) at lowest storage costs. Objects accessed at any tier move back to Frequent Access immediately; no retrieval fees within Intelligent-Tiering regardless of tier.

Q186. What is AWS Transfer Family?

AWS Transfer Family provides fully managed support for file transfers directly into and out of Amazon S3 or EFS using SFTP, FTPS, FTP, and AS2 protocols—enabling legacy application integration without code changes. Transfer Family manages server endpoints, authentication (service-managed with Cognito, AWS Managed AD, or custom Lambda authorizer integrating with LDAP/AD), and transfers files directly to S3 prefixes mapped to user home directories. Managed workflows execute post-upload processing steps (Lambda, decryption, tag/copy/delete) automatically. Use cases include partner data exchange, trading partner AS2 EDI, and automated ETL ingestion pipelines.

Q187. Explain AWS Compute Optimizer.

AWS Compute Optimizer uses ML to analyze the configuration and resource utilization metrics of your AWS resources and recommends optimal configurations to reduce costs and improve performance. Compute Optimizer analyzes EC2 instances, EC2 Auto Scaling groups, EBS volumes, Lambda functions, ECS on Fargate services, and commercial software licenses—comparing current resource utilization against optimal specifications. Recommendations include right-sizing (instance type changes), and identify over-provisioned/under-provisioned resources; enhanced infrastructure metrics (paid) extends the look-back period from 14 days to 93 days for more accurate recommendations based on cyclic patterns.

Q188. What is Amazon Verified Permissions?

Amazon Verified Permissions is a scalable, fine-grained permissions management and authorization service that enables you to externalize and centralize authorization logic from your applications using the Cedar policy language. Cedar policies express role-based access control (RBAC), attribute-based access control (ABAC), and relationship-based access control (ReBAC) in a human-readable, strongly typed language with automated policy analysis (entitlement checking, policy validation). Applications call the IsAuthorized API for real-time authorization decisions; policy stores are versioned for audit. Verified Permissions integrates with Cognito for identity context and supports complex resource hierarchy authorization.

Q189. Explain Amazon CloudWatch Synthetics and canaries.

CloudWatch Synthetics allows you to create canaries—configurable scripts that run on a schedule to monitor endpoints and APIs, simulating user actions and verifying availability and performance even when there is no real user traffic. Canaries are written in Node.js or Python and can use the Synthetics library for browser testing (Playwright-based), API testing (HTTP/HTTPS checks), link checking (broken link detection), and visual monitoring (screenshot comparisons using CloudWatch Evidently). Canary results (success/failure, screenshots, HAR files) are stored in S3; failed canaries trigger CloudWatch alarms integrating with SNS, PagerDuty, and automated remediation via EventBridge.

Q190. What is AWS Resilience Hub?

AWS Resilience Hub provides a central place to define, validate, and track the resilience of AWS applications against disruptions. You add an application (CloudFormation stack, Terraform state, Service Catalog app, or EKS resources), define an RTO/RPO resiliency policy, and Resilience Hub assesses the app against the policy, identifying single points of failure and recommended improvements. The Assessment Report details component-level resilience gaps and specific recommendations (enable Multi-AZ, add read replicas, configure auto scaling); Standard Operating Procedures (SOPs) and runbooks are automatically generated. Integration with AWS FIS enables automated resilience testing.

Q191. How does Amazon SageMaker Pipelines work?

Amazon SageMaker Pipelines is a purpose-built CI/CD service for ML, enabling repeatable, automated ML workflows with tracking and lineage. Pipelines are defined using the SageMaker Python SDK as a series of steps: Processing (data preprocessing with Spark/scikit-learn), Training (distributed model training), Evaluation (model quality assessment), Condition (branching based on metrics), RegisterModel (register to Model Registry), Transform (batch inference), Clarify (bias/explainability analysis), and Lambda (custom logic). Each pipeline execution tracks parameters, metrics, and artifacts with full lineage; Model Registry enforces approval gates before production deployment, with A/B testing via endpoint traffic shifting.

Q192. Explain Amazon SQS FIFO queues for ordering guarantees.

SQS FIFO queues guarantee strictly ordered, exactly-once message delivery within a message group, with a throughput limit of 300 messages/second (3,000 with batching) per queue. Message Group ID partitions ordering—messages with the same group ID are delivered in strict order; multiple producers can use different group IDs for parallel, ordered processing. Message Deduplication ID (or content-based deduplication using SHA-256 hash) prevents duplicate delivery within a 5-minute deduplication interval. High Throughput FIFO mode supports up to 70,000 messages/second by partitioning across multiple shards while maintaining per-group ordering; use cases include financial transactions, order management, and event sequences.

Q193. What is Amazon EventBridge Pipes?

Amazon EventBridge Pipes provides point-to-point integrations between event sources and targets with optional filtering, enrichment, and transformation—reducing the boilerplate code for common integration patterns. Pipes connect sources (SQS, Kinesis, DynamoDB Streams, MSK, MQ, Kafka, SQS) directly to targets (Lambda, Step Functions, EventBridge bus, SNS, SQS, API Gateway, API destinations, Firehose) with filtering to reduce unnecessary invocations, optional enrichment via Lambda/Step Functions/API Gateway/API destinations, and input transformation using JSONata or static mappings. Pipes maintain ordering for streaming sources and handle batching, retries, and DLQ routing automatically.

Q194. Explain AWS cost management strategies for EC2.

EC2 cost management combines multiple strategies: Savings Plans (Compute Savings Plans offer up to 66% discount with flexibility across instance families/regions/OS; EC2 Instance Savings Plans up to 72% for specific family/region commitment) as the primary discount mechanism. Spot Instances for fault-tolerant workloads (batch, ML training, stateless web tiers) at up to 90% discount with Spot interruption handling (two-minute warning, checkpointing). EC2 Fleet and Auto Scaling with mixed instances policy diversifies across Spot pools to reduce interruption. Graviton3 instances reduce cost per compute unit. AWS Compute Optimizer identifies right-sizing opportunities; Scheduled Scaling shuts down non-production environments outside business hours. Reserved Instances for predictable baselines, Spot for elastic burst.

Q195. What is AWS Network Firewall?

AWS Network Firewall is a managed, stateful network firewall and intrusion prevention system (IPS) service for VPCs. Network Firewall supports stateless rules (5-tuple matching: source/destination IP, source/destination port, protocol), stateful rules (Suricata-compatible rule strings for deep packet inspection, domain list filtering for URL/FQDN-based filtering), and managed threat intelligence rule groups updated by AWS. Firewall endpoints are deployed in dedicated subnets; traffic is routed through firewall endpoints using Gateway Load Balancer internally; centralized deployment using Transit Gateway Appliance Mode serves multiple VPCs. Network Firewall integrates with Firewall Manager for central policy management across accounts.

Q196. Explain cross-account IAM access patterns on AWS.

Cross-account access is implemented through IAM role assumption. The target account creates a role with a trust policy allowing the source account (or specific principals) to assume it; the source account's principal calls STS AssumeRole to get temporary credentials for the target role. AWS Organizations with SCPs add an extra permission layer—even if a role allows assumption, SCPs can prevent it. Resource-based policies (S3 buckets, KMS keys, Lambda, SQS) can allow cross-account access without role assumption. AWS RAM (Resource Access Manager) shares resources (Transit Gateways, Subnets, License Manager configs, Route 53 Resolver rules, Network Firewall policies) across accounts without role assumption complexity.

Q197. What is Amazon Braket?

Amazon Braket is a fully managed quantum computing service that helps researchers and developers explore and design quantum algorithms, test them on different quantum hardware technologies, and accelerate research and discovery. Braket provides access to different types of quantum computers (gate-based superconducting qubits from IonQ, Rigetti; trapped-ion from IonQ; photonic from Xanadu), quantum circuit simulators (local, on-demand SV1/TN1/DM1), and hybrid quantum-classical algorithms using Amazon Braket Hybrid Jobs (managed ML-style quantum training). The Amazon Braket SDK is Python-based; PennyLane integration enables quantum ML.

Q198. How does Amazon Aurora Serverless v2 work?

Aurora Serverless v2 scales database capacity instantly in fine-grained increments (0.5 ACU steps) from a minimum of 0.5 Aurora Capacity Units (ACUs) to 128 ACUs based on database load, providing immediate scaling without query disruption—unlike v1's coarse-grained scaling with connection drops. ACU is a combination of ~2 GB RAM, corresponding CPU, and networking; scale-up happens within seconds; scale-down to minimum (for near-zero cost) happens after configurable idle period. Serverless v2 supports Multi-AZ deployments with read replicas, global databases, Aurora ML, Aurora Backtrack, and all Aurora MySQL/PostgreSQL features—it's a drop-in replacement for provisioned instances in cluster configurations.

Q199. What is AWS AppRunner?

AWS App Runner is a fully managed service that makes it easy to deploy containerized web applications and APIs at scale without managing infrastructure, load balancers, or auto scaling configuration. App Runner builds and deploys from source code (GitHub with buildspec) or a pre-built container image from ECR, automatically provisions compute, networking, and traffic routing. Auto scaling adjusts concurrency based on request volume; VPC connectors enable private VPC resource access (RDS, ElastiCache). App Runner handles TLS termination, health checks, and custom domains; it's suited for simple HTTP services where developers want rapid deployment without infrastructure expertise.

Q200. Explain observability best practices for distributed systems on AWS.

AWS observability best practices implement the three pillars—metrics, logs, traces—with correlation. Metrics: CloudWatch custom metrics with EMF (Embedded Metric Format) for Lambda/ECS structured metric publishing; Container Insights for ECS/EKS; Managed Grafana for visualization using CloudWatch, Prometheus (AMP), and X-Ray data sources. Logs: structured JSON logging with correlation IDs (trace ID, request ID); CloudWatch Logs with subscription filters forwarding to OpenSearch for advanced search; log-based metrics for error rate tracking. Traces: X-Ray SDK instrumentation with automatic sampling, service map visualization, groups for trace filtering; Distro for OpenTelemetry (ADOT) enables vendor-agnostic instrumentation exportable to X-Ray and third-party APM tools. Synthetic monitoring (CloudWatch Synthetics) validates user journeys proactively.

Chat with us
📞 Call
DemoWhatsApp