AWS Certified Cloud Practitioner PDF

Summary

This document contains key potential questions and answers from sessions of the AWS Certified Cloud Practitioner (CCP) certification course. It covers various topics related to cloud computing concepts, including cloud service models (IaaS, PaaS, SaaS), benefits, deployment models, and major cloud providers.

Full Transcript

Key Potential Questions and Answers from Each Session of the AWS Certified Cloud Practitioner (CCP) Certification Course Session (1)  What is cloud computing? Cloud computing involves using computing resources, such as servers, s...

Key Potential Questions and Answers from Each Session of the AWS Certified Cloud Practitioner (CCP) Certification Course Session (1)  What is cloud computing? Cloud computing involves using computing resources, such as servers, storage, databases, and software, over the internet, rather than owning and directly managing them. Companies can access these resources on demand and pay only for what they use.  What are the main benefits of cloud computing? o Cost-effectiveness: It shifts costs from large upfront investments to ongoing usage-based expenses, reducing the need for expensive hardware and software. o Flexibility and Scalability: Cloud resources can be scaled up or down quickly based on demand. o Global Infrastructure: Cloud providers have data centres worldwide, enabling businesses to deploy applications closer to their users, improving performance and reducing latency. o Focus on Core Business: Companies can focus on their core business rather than managing complex IT systems. o Disaster Recovery and Business Continuity: Cloud providers offer automated data backup and redundant systems, ensuring continuous service availability. o Enhanced Collaboration: Enables real-time collaboration among teams from different locations. o Automatic Updates: Cloud providers manage software updates, patches, and security enhancements. o Environmentally friendly: Cloud computing reduces the need for physical hardware, thus contributing to a smaller carbon footprint. o Innovation and Speed: Cloud provides access to advanced technologies, such as AI and machine learning, enabling businesses to innovate faster.  What are the three main cloud service models? o Infrastructure as a Service (IaaS): This provides access to basic computing infrastructure such as servers, storage, and networks. Users manage the operating system, databases, and applications. o Platform as a Service (PaaS): This offers a platform for developing and deploying applications, with the provider managing the underlying infrastructure, operating systems, and databases. Users focus on the application and data. o Software as a Service (SaaS): This provides access to ready-to-use applications over the internet. The provider manages all aspects of the application. |Page1  What are the main cloud deployment models? o Public Cloud: Cloud services are available to the general public over the internet, with resources being shared among multiple users and organisations. o Private Cloud: Cloud infrastructure is dedicated to a single organisation, often for security or data privacy. o Hybrid Cloud: A combination of public and private cloud models.  Who are the major cloud service providers? o Amazon Web Services (AWS) is the leading cloud platform, offering a wide range of services. o Microsoft Azure is a major cloud provider. o Google Cloud Platform (GCP) is also a significant cloud provider. o Other providers include Alibaba and IBM.  What is latency, and why is it important in cloud computing? Latency refers to the time it takes for data to travel between a user and a server. It's crucial for online services because low latency ensures faster response times and a better user experience. Cloud providers use their global infrastructure to reduce latency by deploying applications closer to users.  What is redundancy, and why is it important in cloud computing? Redundancy refers to having backup systems in place to ensure continuous availability and prevent data loss. Cloud providers use multiple data centres to ensure that data is backed up and available even in the event of a failure.  What is a data centre in the context of cloud computing? A data centre is a physical building filled with servers. Cloud providers manage these data centres.  What is scalability in the context of cloud computing? Scalability is the ability of a system to handle increasing amounts of workload or traffic. Cloud computing allows businesses to easily scale their IT resources up or down based on demand.  What does "pay-as-you-go" mean in the context of cloud computing? It means that you only pay for the resources you use, instead of buying or maintaining hardware.  What is variable expenditure and how does it relate to cloud computing? Variable expenditure means you pay for services as you use them, contrasting with capital expenditure where you pay upfront for resources. Cloud computing enables a shift to variable expenses, saving costs. Session (2)  What is the AWS Certified Cloud Practitioner (CCP) certification and who is it for? The AWS CCP certification is an entry-level certification that provides a foundational understanding of AWS cloud services and concepts. It's ideal for anyone starting with AWS or aiming for a broad understanding of cloud services.  What kind of examination is the AWS CCP? The exam is multiple choice and multiple answer. There are about 65 questions, and the duration is 90 minutes with a passing score of 70%. It is an understanding based exam and not something that you can just read and pass.  What are the main topics covered in the AWS CCP exam? The exam topics are categorized as follows: Cloud Computing Concepts (26%), Security and Compliance Concepts (25%), AWS Core Services and Technology (33%), and Billing, Pricing, and Support (16%). |Page2  What is AWS? AWS (Amazon Web Services) is a leading cloud service provider that offers a vast array of computing resources and services over the internet. It was launched in 2006 and has grown to be the largest cloud provider in the world. AWS provides services to host applications, store data, manage computing resources, and build solutions.  What are the core services AWS offered when it started? When AWS started in 2006, it offered three core services: EC2 for computing, S3 for storage, and SQS for messaging.  What is the AWS pricing model? AWS uses a pay-as-you-go pricing model, where users pay for the resources they consume. Instead of capital expenditure, AWS offers variable expenses.  Why was AWS created? AWS was originally created as an extension of Amazon's internal IT infrastructure to solve their own complex IT problems. It was later offered as a commercial cloud platform when Amazon realized other companies would benefit from their scalable, on-demand IT resources.  What is the significance of AWS in the cloud industry? AWS pioneered the modern cloud industry and remains a leading force due to its innovation, breadth of services, and global reach. AWS significantly changed how organizations approach IT infrastructure by offering a scalable, flexible, and cost-effective alternative to traditional on-premises hardware.  What is the AWS Global Infrastructure? The AWS Global Infrastructure is the backbone of AWS, providing a reliable, secure, and scalable network of data centers worldwide. It is composed of regions, availability zones (AZs), data centres and edge locations.  What are AWS Regions? AWS Regions are geographical areas around the world where AWS has built clusters of data centres. As of November 2024, AWS has 34 geographic regions, with plans for six more. Each region is a vast area with multiple clusters. The regions are separate from each other to ensure geographical redundancy. Each region operates independently of other regions.  What are AWS Availability Zones (AZs)? Availability Zones are physically separate data centres within an AWS Region. Each AZ has its own infrastructure to ensure fault tolerance and high availability. Within a region, there are multiple AZs. If one AZ faces issues, others in the same region can take over, ensuring application availability and reliability.  What are AWS Data Centers? Data Centers are physical facilities where the AWS infrastructure resides, including servers, networking equipment, and storage systems. Data centers are located within availability zones. AWS does not disclose the exact number or locations of its data centers.  What are AWS Edge Locations? Edge locations are physical sites that are intentionally designed by AWS to provide fast content delivery. They are not the same as data centers. Edge locations cache data closer to end-users to reduce latency. They are part of the AWS Content Delivery Network (CDN) and work through the CloudFront service.  How do Edge Locations provide fast content delivery? Edge locations cache data closer to end users to reduce latency and ensure faster content delivery, especially for static assets.  What is an Origin Server? The origin server is the actual server that contains the original content or data, such as a website. Edge locations copy data from the origin server to be closer to the end-user. |Page3 Session (3)  What is the AWS Certified Cloud Practitioner (CCP) certification? o It is an exam that tests your knowledge of cloud computing and AWS, and it is considered a credible certification. Passing the exam demonstrates your understanding of cloud services and AWS, leading to a certified Cloud Practitioner badge.  What are the core services offered by AWS? o When AWS was launched in 2006, it offered three core services: EC2 (for compute), S3 (for storage), and SQS (for messaging). Today, AWS has expanded to over 200 services, organised into categories. These include compute, storage, database, networking and content delivery, management and governance services.  What is Amazon EC2? o Amazon EC2 (Elastic Compute Cloud) is a cloud-based compute service that provides scalable virtual machines to run applications in the cloud. These virtual machines are also known as instances. EC2 provides machines in the cloud that can run applications. It is considered one of the core services of AWS.  What are EC2 Instance Types? o EC2 instances come in a variety of types designed for different use cases. These are organised into families based on the resources they provide, such as CPU, memory, storage, and networking.  General Purpose Instances: Provide a balance of CPU, memory, and networking resources, suitable for a variety of workloads such as web servers and development environments.  Compute Optimized Instances: Ideal for applications that require high-performance processing power, such as batch processing and scientific modelling.  Memory Optimized Instances: Designed for memory-intensive applications, such as high-performance databases and real-time big data analytics.  Storage Optimized Instances: Provide high sequential read and write access to large datasets, suitable for data warehousing and distributed file systems.  Accelerated Computing Instances: Use hardware accelerators such as GPUs, ideal for machine learning, AI, and graphics rendering.  What are the pricing models for Amazon EC2? o EC2 offers several pricing models to give users flexibility in terms of cost optimization.  On-Demand Instances: Pay for compute capacity by the second with no long-term commitments, suitable for unpredictable or short-term workloads.  Reserved Instances: Require an upfront payment for a one- to three- year commitment to receive significant discounts compared to On- Demand pricing, suitable for steady-state workloads. |Page4  Spot Instances: Allow users to bid for unused capacity at a discounted rate, which AWS can terminate with short notice, suitable for flexible, fault-tolerant applications.  Dedicated Hosts: Provide physical servers dedicated to your use, useful for licensing software that requires dedicated hardware.  Savings Plans: Offer flexible pricing allowing users to save on compute usage in exchange for a commitment to use a specific amount of resources for one to three years.  What does 'scalability' mean in the context of EC2?  Scalability is the ability to increase or decrease the resources of a virtual machine as needed. This includes making the virtual machine bigger or smaller by adding or removing resources such as RAM, CPU, or storage, or even adding more machines.  What is meant by 'virtualisation' when discussing EC2 instances? o Virtualisation is the process of creating virtual versions of something such as a computer or server. In EC2 it involves creating virtual instances or virtual machines using software called a hypervisor.  What is a hypervisor in the context of EC2 virtualisation? o A hypervisor is software that creates and manages virtual machines. It sits between the hardware and the operating system, allocating resources like CPU, memory, and storage to each virtual machine. AWS uses hypervisors such as Xen or Nitro.  What are some other key AWS services mentioned in the video? o S3 (Simple Storage Service) for storing and retrieving data. o Lambda for serverless computing. o Elastic Beanstalk for simplifying the deployment of web applications. o ECS (Elastic Container Service) and EKS (Elastic Kubernetes Service) for managing containers. o VPC (Virtual Private Cloud) for secure isolated networks. o CloudFront for content delivery networks. o CloudWatch for monitoring and logging AWS resources. o Auto Scaling for automatically adjusting capacity of AWS services. o EBS (Elastic Block Store) for block storage for EC2 instances.  How does Amazon CloudFront work? o It is a Content Delivery Network (CDN) that delivers data with low latency. It works with edge locations to bring data closer to users and reduce latency.  What is the significance of edge locations in AWS? o Edge locations are data centers around the world that bring data closer to users in order to reduce latency and enable fast content delivery.  Why might a business choose AWS for their applications? o AWS is considered secure and reliable. It provides a wide range of services, constant innovation and enables businesses to securely run a wide range of applications. It offers more security, compliance and governance capabilities than other cloud providers, and it allows businesses to scale, deploy and manage applications with greater flexibility and cost-effectiveness.  What is the AWS Nitro System? o It is the foundation of AWS's modern compute infrastructure that provides built-in security at the chip level. It offers high performance by offloading the overhead from virtualization to give more resources to applications.  How can I get AWS certified? |Page5 o You need to study, understand the core services, and take practice tests. A score of 70% or higher on the exam leads to certification.  What is the format of the AWS CCP certification exam? o The exam has 65 multiple-choice questions, and the time limit is 90 minutes. Session (4)  What is the AWS Certified Cloud Practitioner (CCP) certification? o The CCP certification is a globally recognised course that certifies individuals in cloud computing, preparing them for tech careers or specialisation in the cloud. This training aims to provide a strong foundation in cloud computing and AWS, aiding success in the CCP exam.  What are the core AWS services a cloud practitioner should know for the exam? o A cloud practitioner should have an idea about approximately 20-25 AWS services, out of over 200, focusing on the core services required for the examination. These are organised into categories such as compute, storage, database, networking, content delivery, management, and governance services.  Can you give examples of services within the compute category? o The compute services include:  EC2 (Elastic Compute Cloud): Provides virtual servers for running applications.  AWS Lambda: A serverless computing service for executing code in response to events.  Elastic Beanstalk: A managed service for deploying web applications.  ECS (Elastic Container Service): Manages Docker containers.  EKS (Elastic Kubernetes Service): A managed Kubernetes service for containerised applications.  Lightsail: A user-friendly service for launching simple web services.  What are some of the AWS storage services? o The AWS storage services include:  S3 (Simple Storage Service): For storing and retrieving various types of data.  EBS (Elastic Block Store): Provides block storage for EC2 instances.  Glacier: Low-cost storage for infrequently accessed data.  FSx: A managed file system solution.  Storage Gateway: Connects on-premises storage with AWS.  What are examples of AWS database services? o AWS database services include:  RDS (Relational Database Service): Manages relational databases like MySQL and PostgreSQL.  DynamoDB: A highly scalable NoSQL database.  Aurora: A high-performance relational database compatible with MySQL and PostgreSQL.  Redshift: A data warehouse service for analyzing large datasets.  DocumentDB: A managed database compatible with MongoDB for document based applications.  What are some networking and content delivery services in AWS? o These services include: |Page6  VPC (Virtual Private Cloud): Provides secure isolated networks.  Route 53: A domain registration and DNS service.  CloudFront: A content delivery network (CDN) for low-latency data delivery.  AWS Direct Connect: A dedicated network connection to AWS.  API Gateway: Manages and deploys secure APIs.  What are the management and governance services in AWS?  These services include: o CloudWatch: Monitors and logs AWS resources and applications. o CloudTrail: Records user activities for security and transparency. o Config: Tracks and assesses AWS resources and configurations for compliance. o Service Manager: Manages AWS resources across environments from a central location. o Organizations: Manages multiple AWS accounts with consolidated billing.  What is serverless computing? o Serverless computing is a cloud execution model where users focus on writing code, and the cloud provider handles the provisioning, scaling, and management of the servers needed to run that code. It abstracts away the complexities of server management from the user.  What is "abstraction" in the context of technology, especially in relation to serverless computing? o Abstraction refers to hiding complex details and providing simple interfaces for users to operate a technology. In serverless computing, the complexities of managing servers are hidden, allowing users to focus solely on their code.  Does serverless computing mean no servers are involved? o No, serverless computing does not mean that no servers are involved. It means the cloud provider manages the servers on behalf of the user, abstracting away the server operations.  What is AWS Lambda? o AWS Lambda is a serverless compute service that allows you to run code without provisioning or managing servers. It executes code in response to events or predefined triggers.  What is a "Lambda function"? o A Lambda function is the code you write and deploy to AWS Lambda. It is triggered by events from other AWS services or API requests from API Gateway.  How are Lambda functions triggered? o Lambda functions are triggered by events such as file uploads to S3, database updates in DynamoDB, messages sent to SNS or SQS, or API requests through API Gateway.  What are some common use cases for AWS Lambda? o Common use cases for Lambda include:  Handling backend processes for websites.  Sending notifications when files are uploaded.  Automatic image resizing.  Automating daily reminders or scheduled tasks.  Processing user input or comments on blogs.  Chat moderation.  File format conversion |Page7  Sending welcome emails after signing up for a service.  Tracking user activity.  Chatbot services.  Automation of repeated tasks.  IoT data processing.  When is it more appropriate to use Lambda over EC2? o Lambda is ideal for microservices, occasional events, and tasks that are triggered by events. It is not suitable for applications that need to run continuously for long periods. EC2 instances are better for continuous applications.  How does AWS Lambda billing work? o AWS Lambda billing is based on two main factors:  The number of requests (invocations). The first 1 million requests per month are free, then $0.20 per 1 million requests.  The compute time (duration). This is measured in gigabyte-seconds, based on the memory allocated to the function and the duration of its execution. The first 400,000 gigabyte-seconds of compute time per month are free.  What are 'runtimes' in the context of AWS Lambda? o Runtimes refer to the programming languages supported by AWS Lambda, including Python, JavaScript (Node.js), Java, C#, Ruby and Go. The runtime specifies the environment in which the code will run.  What are the steps involved in how Lambda functions work? o The steps are: writing the code, deploying the code to Lambda, setting up triggers, executing the function when an event occurs and the function returns a response.  Why should you delete unused AWS resources such as S3 buckets and Lambda functions? o You should delete unused AWS resources to avoid incurring unnecessary charges. Session (5)  What is AWS Elastic Beanstalk? o Elastic Beanstalk is a platform as a service (PaaS) that allows developers to deploy web applications and services without managing the underlying infrastructure. It simplifies infrastructure management by automating server setup, load balancing, auto-scaling, and application health monitoring.  What does Elastic Beanstalk manage for the developer? o Elastic Beanstalk manages and scales applications and services on behalf of the developer. It handles the server setup, load balancing, auto-scaling, and application health monitoring. It abstracts away the complexities of hardware, operating systems, and middleware, allowing developers to focus solely on application development.  What type of applications can be deployed with Elastic Beanstalk? o Elastic Beanstalk can deploy web applications, APIs, mobile backends, enterprise applications, microservices, batch processing tasks, and testing and development environments. It supports popular languages such as Python, Java, Node.js, Ruby,.NET, PHP, and Docker.  How does Elastic Beanstalk relate to serverless computing? |Page8  While similar to serverless computing in that it abstracts away infrastructure, Elastic Beanstalk is not serverless. With Elastic Beanstalk, you can run long-running applications, unlike Lambda, which is better suited for microservices. Also, with Elastic Beanstalk, you can see the AWS services (like EC2, S3) that are set up for you, whereas Lambda does not show you the underlying infrastructure.  What are the key features of Elastic Beanstalk? o Key features include:  Simplified deployment: Automates application deployment with minimal settings.  Platform support: Supports various platforms such as Apache Tomcat, Node.js, Python, PHP, Ruby, Go, and Docker.  Scalability: Automatically scales applications based on demand using auto-scaling.  Integrated monitoring: Provides built-in monitoring with AWS CloudWatch for performance metrics.  Managed infrastructure: Manages EC2 instances, Elastic Load Balancing, Auto Scaling, and RDS if configured.  Customizability: Allows full control and customisation of AWS resources.  How does Elastic Beanstalk work? o A developer uploads their application code (zipped or as a Docker image) to Elastic Beanstalk. Elastic Beanstalk then analyzes the code, and, based on your configurations or default settings, provisions necessary AWS resources such as EC2 instances, S3 buckets, auto-scaling groups, and load balancers. It also sets up CloudWatch for monitoring.  What is an "environment" in Elastic Beanstalk? o An "environment" refers to the collection of AWS resources created by Elastic Beanstalk to run an application. This includes resources such as EC2 instances, auto-scaling groups, Elastic Load Balancers, security groups, CloudWatch alarms, and S3 buckets for logs and storage.  What are the two types of environment tiers in Elastic Beanstalk? o Elastic Beanstalk offers two types of environment tiers:  Web server environment tier: Designed for web applications and APIs.  Worker environment tier: Used for background jobs or long-running tasks, such as processing tasks using an SQS queue.  What is the Elastic Beanstalk workflow? o The workflow involves:  Preparing the application with supported files (e.g., zip or Docker image).  Creating an environment.  Uploading the application via the Elastic Beanstalk console, command line interface, or SDKs.  Managing and scaling the application, monitoring its health.  Deploying new versions of the application.  How does Elastic Beanstalk handle application updates? o Elastic Beanstalk handles rolling updates to avoid downtime. It also maintains different versions of an application, allowing developers to revert to a previous version if an update causes an issue.  How can the health of an application be monitored in Elastic Beanstalk? |Page9 o Elastic Beanstalk uses health checks via Elastic Load Balancing and provides status indicators: green for healthy, yellow for issues such as high latency, and red for an application that is not running. Health can also be monitored through CloudWatch metrics on the Elastic Beanstalk dashboard.  How is Elastic Beanstalk priced? o Elastic Beanstalk itself is free. However, users pay for the AWS resources that are provisioned on their behalf, such as EC2 instances, S3 storage, and load balancers.  How can you customise Elastic Beanstalk? o Customisation can be done via the AWS Management Console and by using EB extensions, such as JSON or YAML files, to modify environment settings, such as EC2 instance types, IAM roles, or CloudWatch alarms. Session(6)  What is containerization? o Containerization is the process of packaging software code along with all its dependencies into a single, lightweight, and portable unit called a container. This allows the software to run consistently across different environments.  What are containers? o Containers are lightweight, standalone, and portable units of software that include everything needed to run an application, such as the code, runtime, system tools, libraries, and settings. They enable applications to run reliably and consistently across various computing environments.  How does containerization differ from virtualization? o Virtualisation allows multiple applications to run on a single server by simulating hardware and software. Each virtual machine (VM) includes its own operating system, which consumes a lot of disk space, RAM, and CPU power. Containers, on the other hand, share the host operating system's kernel, making them more lightweight and efficient. Containers are faster to start and consume fewer resources compared to VMs.  What are the advantages of using containers? o Containers offer several benefits:  Consistency: Applications run the same way in development and production.  Efficiency: Containers are lightweight compared to VMs.  Portability: Containerized applications can run anywhere.  Scalability: Containers support rapid scaling to handle increased workloads.  Isolation: Containers provide process and resource isolation.  What is Docker? o Docker is a platform for automating the deployment, scaling, and management of applications in lightweight, portable containers. It is an open-source tool that enables the creation and use of containers.  What is container orchestration? o Container orchestration is the automated process of managing and coordinating multiple containerized applications. It manages how containers function across a cluster of servers or virtual machines.  What are the key functions of a container orchestrator? o Container orchestrators handle various tasks: | P a g e 10  Automated scheduling of containers based on resource needs.  Load balancing traffic across containers to prevent overload.  Automatic scaling of the number of containers in response to demand.  Self-healing by replacing failed containers.  Networking and security, ensuring secure access controls.  Monitoring container health.  What is Kubernetes? o Kubernetes is the most widely used container orchestrator, initially developed by Google. It automates the deployment, scaling, and management of containerized applications across a cluster of servers.  What is the difference between Docker and Kubernetes? o Docker is a tool for creating containers, while Kubernetes is a tool for managing and orchestrating containers. Docker helps package applications into containers, and Kubernetes helps manage and scale these containers across multiple machines.  Why is learning about containerization important for tech professionals? o Containerization is a modern technology and the future of code deployment. Companies are adopting modern tools, and tech professionals need to stay updated to meet industry demands. Understanding containerization is crucial for job interviews and career advancement in modern technology. Session (7)  What is containerization? o Containerization is the process of packaging software code with all its dependencies into a container, which is a lightweight, standalone, and portable unit of software that can run anywhere. It is considered the future of software deployment.  What is Docker? o Docker is a platform for creating, managing, and running containers, and it's a leading solution for containerization. It's an important tool for software engineers, cloud developers, web developers and data professionals.  What is container orchestration? o Container orchestration is the automated process of managing and coordinating multiple containers or containerized applications. It involves tools called container orchestrators to handle multiple containers in a cluster.  What is the leading container orchestrator? o Kubernetes is the most widely used container orchestrator, developed by Google and now managed by the Cloud Native Computing Foundation (CNCF).  What is Amazon ECS (Elastic Container Service)? o Amazon ECS is a container orchestration service provided by AWS. It allows you to run and manage Docker containers on a cluster of virtual machines. It is a native AWS tool for container orchestration.  What are the launch types for ECS? o There are two main launch types for ECS: | P a g e 11  EC2 launch type: With this option, you manage the virtual machines (EC2 instances) where your containers run. You are responsible for setting up, managing, and maintaining these VMs.  Fargate launch type: With this option, AWS manages the underlying infrastructure. You only need to specify the CPU and memory requirements for your containers. AWS handles the provisioning and managing of the compute resources.  What is AWS Fargate? o AWS Fargate is a serverless compute engine for containers. It allows you to run containers without managing the underlying infrastructure. You define your application's requirements, and Fargate takes care of provisioning the resources. Fargate is a serverless compute model for containers. It's similar to Lambda in that it manages the infrastructure for you.  How does Fargate bill? o Fargate bills based on the compute resources used by your containers, such as virtual CPUs and memory, charged per second with a minimum of one minute.  When should you use Fargate? o Fargate is ideal for applications that require scalability and flexibility without the overhead of managing servers. It's suitable for microservices, batch processing, and services with variable workloads.  What are some benefits of Fargate? o Benefits include: no need to manage EC2 instances, automatic scaling, and integration with other AWS services for monitoring, logging, and security.  What is a limitation of Fargate? o Fargate may have a higher cost compared to managing ECS with EC2 launch types, especially for consistently high resource usage or specific instance configurations.  How does ECS work? o ECS uses clusters to organise resources for containerized applications. It uses tasks as the smallest unit of work which runs one or more containers. Services ensure that tasks remain running.  What are use cases for ECS? o ECS is commonly used for deploying microservice architectures and running batch processing workloads.  How is ECS billed? o ECS billing depends on the launch type:  With the EC2 launch type, you're billed for the EC2 instances in your cluster, including the instance type, storage (EBS), and data transfer.  With the Fargate launch type, you're billed based on the vCPU and memory resources your tasks are using, charged per second with a minimum of 1 minute. o Additional costs may apply for other AWS services used with ECS, such as data transfer, elastic load balancing, and CloudWatch.  What is Amazon EKS (Elastic Kubernetes Service)? o Amazon EKS is a managed Kubernetes service that allows you to run and scale Kubernetes applications on AWS. It simplifies using Kubernetes by managing the Kubernetes control plane. AWS has created their own container orchestration service with ECS, and EKS allows users to bring their pre- existing kubernetes deployments into AWS.  What is the Kubernetes control plane? | P a g e 12 o The control plane refers to the set of components that manage the Kubernetes cluster, including the API server, etcd, controller manager, and scheduler.  Who is EKS ideal for? o EKS is ideal for users already familiar with Kubernetes or those needing portability for containerized workloads across multiple environments.  What are key differences between ECS and EKS? o Container Orchestration Tool: ECS uses AWS's own container orchestration service, while EKS uses Kubernetes. o Ease of Use: ECS is simpler and AWS-native, great for beginners, while EKS is more complex, ideal for those familiar with Kubernetes. o Flexibility: ECS is tightly integrated with AWS, while EKS works across different environments (AWS, on-premise, other clouds). o Management: With ECS, AWS fully manages the orchestration layer, while with EKS, AWS manages the Kubernetes control plane, but you manage the Kubernetes nodes. o Use Cases: ECS is ideal for those seeking simplicity and a fully integrated AWS ecosystem, while EKS is better for users requiring Kubernetes features or multi-cloud portability.  Can EKS use Fargate? o Yes, EKS can use Fargate to run Kubernetes workloads without needing to manage servers. AWS provisions the compute resources and manages scaling, allowing you to focus on deployment.  How is EKS priced? o EKS charges a fixed fee for the Kubernetes control plane per cluster. You also pay for the underlying resources, such as EC2 instances or Fargate tasks, and other AWS resources like storage or load balancing. Session (8)  Question: What is the AWS CCP certification? o Answer: The AWS CCP certification is a program that validates an individual's skills and knowledge of cloud computing concepts and the Amazon Web Services (AWS) platform. It signifies a foundational understanding of AWS services and their applications.  Question: What are the core AWS services that are important for the CCP exam? o Answer: The core services include compute (like EC2), storage (like S3), and messaging (like SQS and SNS). Other important services include those related to databases (such as RDS and DynamoDB), networking (like VPC and Route 53), and management (like CloudWatch and CloudTrail).  Question: Why is it important to understand data storage concepts before learning about AWS storage services? o Answer: A foundational understanding of data storage concepts, such as file systems, data types, and storage types, allows for a more effective use of AWS storage tools. This knowledge also helps in selecting the appropriate AWS storage service for specific needs.  Question: What is a file system, and why is it important? o Answer: A file system is a method or structure used to organise, store, and manage files on a storage medium. It is essential because it allows a storage | P a g e 13 medium to understand how to group data, control how data is stored and retrieved, keep track of file locations and metadata, and separate data into structured pieces of information. Without a file system, data would be one large body with no way to differentiate or access individual pieces.  Question: What are the common types of file systems? o Answer: Common file systems include FAT32 (File Allocation Table 32), NTFS (New Technology File System), and exFAT (Extended File Allocation Table). Each is suited for specific devices, operating systems, and use cases. There is also ext4 for Linux systems.  Question: What are the key characteristics of FAT32? o Answer: FAT32 is an older file system, but it is widely compatible with many devices. Its main limitation is that individual files cannot exceed 4 GB in size, and partitions must be smaller than 16 TB. It has slow read and write speeds. It is commonly used on flash drives for compatibility across devices.  Question: What are the key characteristics of NTFS? o Answer: NTFS is the default file system for Windows. It supports large files (up to 16 exabytes) and has features like file permissions and encryption. However, it is primarily compatible with Windows and has limited support on macOS and Linux. It has fast read and write speeds and is best for Windows systems and internal drives.  Question: What are the key characteristics of exFAT? o Answer: exFAT is optimized for flash drives and external media. It supports large files (up to 16 exabytes) and has no realistic partition size limits. It is more compatible with macOS than NTFS but has limited support on older devices. It is the fastest option among these three and is best for USB flash drives and external hard drives with files larger than 4 GB.  Question: What are the different types of data? o Answer: Data can be categorized into structured data, unstructured data, and semi-structured data. Structured data is highly organised in rows and columns, like in databases and spreadsheets. Unstructured data lacks a predefined format and is difficult to analyse directly, examples include images, videos, audio files, and social media content. Semi-structured data combines elements of both and includes tags or markers for some organisation, such as XML and JSON files.  Question: What are the main types of storage systems?  Answer: The main types of storage systems are file storage, block storage, and object storage.  Question: How does file storage work? o Answer: File storage organises data in a hierarchical structure using files and folders, much like a computer's file explorer. It relies on file systems to store and retrieve data, and it is best suited for smaller-scale projects. Cloud services like Google Drive and Dropbox utilise this type of storage. AWS has the Elastic File System (EFS), which is a scalable, fully managed file storage service.  Question: How does block storage work? o Answer: Block storage organises data into fixed-sized blocks, each with a unique address. It is ideal for high-performance environments like databases and virtual machines, and it provides direct access to data blocks, ensuring low latency and fast data retrieval. AWS's Elastic Block Store (EBS) is a block storage service. | P a g e 14  Question: How does object storage work?  Answer: Object storage stores data as discrete units called objects, each containing the data, metadata, and a unique identifier. These objects are stored in buckets and are accessed using their unique IDs. It's designed for unstructured data and massive data sets, and is ideal for media content, data backup, and cloud applications. Amazon S3 (Simple Storage Service) is an object storage service.  Question: What is metadata? o Answer: Metadata is information about data, like file size, creation date, and last accessed date. It provides context and aids in categorising and retrieving data.  Question: What is the difference between file storage and block storage? o Answer: File storage uses a hierarchical tree structure with files and folders, managed by a file system, and is suitable for smaller-scale projects. Block storage breaks data into blocks with unique addresses, providing flexibility and high performance for databases and virtual machines.  Question: What is the difference between block storage and object storage? o Answer: Block storage divides data into fixed-size blocks accessed by unique addresses and is designed for high-performance and scalability, for use in VMs and databases. Object storage treats data as objects, with associated metadata and unique identifiers, suitable for unstructured data and large datasets. Session (9)  What is Amazon S3 and what type of storage is it? o Amazon S3 is a cloud object storage service that provides scalability, data availability, security, and performance. It is an object storage service.  What are the different types of storage and how are they different? o There are three main types of storage: file storage, block storage, and object storage. o File storage uses a hierarchical directory structure, similar to a computer's file system. o Block storage is another type of storage. o Object storage stores data as objects in containers called buckets and each object is tagged with metadata and a unique identifier. It does not use a hierarchical directory structure.  What are the key components of an object in S3? o An object in S3 consists of three components: data, which is the actual file or content; metadata, which is information about the data (e.g., file type, size, date); and a unique identifier (key) that is used to access the object.  What are buckets in S3? o Buckets are containers for objects stored in S3. Every object stored in S3 is stored in a bucket. Each bucket name must be globally unique. Buckets are created in specific AWS regions.  What are the naming rules for S3 buckets? | P a g e 15 o Bucket names must be globally unique. They must be between 3 and 63 characters. They must be lowercase and cannot include special characters.  How does S3 ensure data durability and availability? o S3 provides 99.999999999% (11 9's) durability by storing copies of data across multiple availability zones. o It provides 99.99% availability, ensuring users can access their data when needed. o Data is automatically replicated across a minimum of three geographically separated Availability Zones within a region. o Continuous data integrity checks are performed using checksums. If corruption is detected, it's automatically repaired using other intact copies.  What are S3 Storage Classes and why are they important? o S3 offers different storage classes to optimise costs based on data access patterns. Understanding storage classes is crucial for cost management in S3. o The main classes include:  S3 Standard: General-purpose storage for frequently accessed data. It is the most expensive storage class.  S3 Intelligent Tiering: Automatically moves objects between tiers (frequent, infrequent access) based on access patterns to optimize cost.  S3 Standard-Infrequent Access (IIA): Lower cost for data that is accessed less often.  S3 One Zone-Infrequent Access (IIA): Similar to standard IIA but stores data in a single Availability Zone (less durable).  S3 Glacier: Low-cost storage for archiving data with varying retrieval times. It is a storage class within S3.  Glacier Instant Retrieval: Low cost with millisecond retrieval for archive data.  Glacier Flexible Retrieval: For long-term archives, retrieval times range from minutes to hours.  Glacier Deep Archive: The lowest cost storage for rarely accessed data; retrieval in hours.  How is data accessed in S3? o Data is retrieved through HTTP-based APIs using URLs.  How is S3 priced? o S3 uses a pay-as-you-go model, where you pay only for what you use. o Pricing is based on:  Storage: Cost depends on the chosen storage class.  Data transfers: Uploads (data ingress) are free in most cases, while downloads (data egress) are charged based on the volume and destination.  Data transfer within AWS regions is free, but transfers between regions or to the internet are charged.  Requests: Charges vary based on the number and type of requests.  Additional Features: Including versioning, lifecycle policies, and data management tools. o Data retrieval costs are influenced by the retrieval type and vary across the different storage classes. o Transfer acceleration for faster uploads also incurs extra costs.  What security features does S3 offer? | P a g e 16 o S3 uses Identity and Access Management (IAM) policies, bucket policies and Access Control Lists (ACLs) to control access. o Data is encrypted both at the server side and client side. o Public access can be blocked to prevent data exposure.  What are pre-signed URLs in S3? o Pre-signed URLs are temporary, secure URLs that can be used to share objects without modifying bucket permissions.  What are common S3 request types and how do they work? o GET: Fetches the full content of an object. o HEAD: Retrieves only the metadata of an object. o Range Get: Downloads a specific portion of an object. o SELECT: Uses SQL expressions to query data within an object.  What is versioning in S3? o Versioning allows multiple versions of the same object to coexist in a bucket. This protects against accidental deletion or overwrites.  What is intelligent tiering? o Intelligent tiering is a storage class that automatically moves data between tiers based on access frequency within the intelligent tiering class. There is also a monitoring fee with the intelligent tiering class.  What are some typical use cases for S3? o S3 is used for data backup, archival, web application hosting, big data analytics, and more.  How do you create an S3 bucket? o You create a bucket through the AWS Management Console. o You must choose a globally unique name for the bucket. o You must choose the AWS region where the bucket will be created. o It is highly recommended that you block public access. Session (10)  What is Amazon EBS? o Amazon Elastic Block Store (EBS) is a scalable, high-performance block storage service designed for use with Amazon EC2 instances. It provides persistent block-level storage.  What is the primary purpose of EBS? o EBS is designed to provide storage for EC2 instances. It acts like a virtual hard drive attached to an EC2 instance, providing persistent storage.  What type of storage does EBS provide? o EBS provides block level storage.  What is block storage? o Block storage breaks data into blocks and stores them. EBS is the main block storage service on AWS.  What is the difference between persistent and ephemeral storage? o Persistent storage (like EBS) retains data even if the associated EC2 instance is stopped, rebooted, or terminated. Ephemeral storage (instance store) is temporary, and data is lost when the EC2 instance is stopped, terminated, fails, or is rebooted.  What is ephemeral storage called for EC2 instances? | P a g e 17 o Ephemeral storage for EC2 instances is called instance store. It is physically attached to the EC2 instance provided by the host machine.  What are EBS volumes? o EBS volumes are the storage volumes created for EC2 instances. They use block storage and persist independently from EC2 instances.  What are the main characteristics of EBS? o EBS is scalable (can be scaled up or down), provides high performance block storage, and is persistent (retains data).  How does EBS compare to the storage that comes with an EC2 instance? o An EC2 instance comes with its own ephemeral storage, called instance store, which is temporary. EBS provides persistent storage for EC2 instances.  How does an EC2 instance get persistent storage? o An EC2 instance gets persistent storage from EBS.  Can an EC2 instance have both instance store and EBS storage? o Yes, an EC2 instance can have both its own instance store (ephemeral storage) and persistent storage using EBS.  What happens to data in the instance store if the EC2 instance is stopped? o Data stored in the instance store is lost when the instance is stopped, terminated, fails, or is rebooted.  What is a key feature of EBS for data protection? o Snapshots are point-in-time backups of EBS volumes that are stored in S3. They are used for backup, recovery, and replication.  How do EBS snapshots work? o Snapshots are incremental, meaning that after the initial full backup, only the changes since the last snapshot are saved. This makes them cost-efficient.  Where are EBS snapshots stored? o EBS snapshots are stored in Amazon S3.  Can EBS volumes be resized? o Yes, EBS volumes are scalable and can be resized dynamically without downtime.  What are some key features of EBS? o Key features of EBS include scalability, high availability and durability, multi- attach (attaching volumes to multiple instances), performance optimization, and encryption.  What are the main types of EBS volumes? o The main types of EBS volumes include:  General Purpose SSD (gp2, gp3): Cost-effective and default type.  Provisioned IOPS SSD (io1, io2): High-speed SSDs for critical workloads.  Throughput Optimized HDD (st1): For big data and streaming.  Cold HDD (sc1): For infrequent access and archival data.  How is EBS pricing determined? o EBS pricing is based on the volume type, storage size, provisioned IOPS, and snapshots. It follows a pay-as-you-go model. You are billed for the total provisioned size even if you don't fully use it.  How do you get billed for EBS volumes o You pay per GB per month for your EBS volumes.  What are the use cases for EBS? o EBS is used for database storage, applications needing low-latency performance, and as boot volumes for EC2 instances. | P a g e 18  What is the difference between Amazon S3 and EBS? o S3 is object storage for scalable long-term storage of large datasets and unstructured data, accessed over the internet or APIs. EBS is block storage for high-performance, low-latency storage tied to EC2 instances and accessed as volumes. o S3 is optimized for high scalability whereas EBS is optimized for performance. o S3 has virtually unlimited storage whereas EBS is limited by the EC2 instance storage and size of volume.  What does the term 'iops' mean? o IOPS refers to input/output operations per second. You can provision specific IOPS to improve the speed of your EBS volumes. Session (11)  What is Amazon EFS? o Amazon EFS is a fully managed, serverless file storage service provided by AWS. It is designed for shared access, allowing multiple EC2 instances (and other AWS services like containers and Lambda) to access the same file system concurrently.  What type of storage is EFS? o EFS provides file storage, which uses a file system and follows a hierarchical structure with directories and files.  How does EFS compare to EBS and S3? o EBS is a block storage service primarily used for single EC2 instance storage, optimized for high-performance and low latency. S3 is an object storage service. EFS is designed for shared access among multiple instances needing a central data source.  What does "fully managed" and "serverless" mean in the context of EFS? o Fully managed means AWS handles the management of the service, including infrastructure. Serverless means you don't need to manage or configure the underlying servers. This is similar to compute services like Lambda, where the complexities of infrastructure are abstracted away.  What are the key use cases for EFS? o EFS is ideal for applications like web hosting, content management, and data sharing where shared file access is more important than ultra-low latency. It is also suitable for applications requiring a common data source across a distributed system.  How does EFS handle scalability and availability? o EFS is scalable and elastic, automatically growing or shrinking as you add or remove files. It is designed for high availability and durability, with data stored redundantly across multiple availability zones. It offers 11 9's of durability.  What is the file system used by EFS? o EFS uses the Network File System (NFS) protocol, specifically NFS version 4, which is designed for cloud network environments.  What are the storage classes available for EFS? o EFS offers standard and infrequent access (IA) storage classes.  EFS Standard: This is the default storage class, suitable for frequently accessed files, and stores data across multiple availability zones. | P a g e 19  EFS One Zone: This standard class stores data in a single availability zone and is a lower cost option when data durability across multiple zones isn't required.  EFS Standard IA: This storage class is a low-cost tier for infrequently accessed files, with data replicated across multiple availability zones.  EFS One Zone IA: This class is for infrequently accessed data stored in a single availability zone, offering lower costs.  What is the concept of Multi-Attach in EBS and how does it compare with EFS? o While EBS is primarily designed for single instance attachment, certain EBS volume types (io1, io2) support multi-attach, allowing multiple EC2 instances to access the same volume. However, applications must manage data consistency as EBS does not manage it in this mode. EFS is designed for multi-instance access without requiring application-level consistency management. EFS is built for shared access and file-system needs.  How is EFS priced? o EFS pricing is based on the amount of storage you use and any optional features like throughput or life cycle management. It has a free tier of up to 5 GB for the first 12 months. You only pay for the storage consumed.  What is a NAS and how does it relate to EFS? o NAS stands for Network Attached Storage, which is a central storage designed for a network environment. EFS is designed as a NAS for the cloud, where EC2 instances can connect to it for shared data.  Does EFS have life cycle management? o Yes. EFS has lifecycle management that can automatically move infrequently accessed files to cheaper storage classes. This reduces storage costs.  What are the performance characteristics of EFS? o EFS provides scalable performance which automatically scales. It is also elastic. Compared to EBS, EFS has a higher latency, which means it is slower, but is suitable for applications that prioritise shared access over speed.  Can EFS be accessed from on-premises servers? o Yes, on-premises servers can be connected to EFS using AWS Direct Connect or AWS VPN. Session (12)  What is a database? A database is an organised collection of data held in a system for easy access, management, and efficient retrieval. It is a structured source of information that is designed for organised access, modification and retrieval.  What are the two broad categories of databases? Databases can be categorised into relational databases and non-relational databases.  What are relational databases? Relational databases are table-based databases where data is organised into tables with rows and columns. These tables are related to each other to avoid data redundancy. They use structured query language (SQL) for managing and querying the data.  What are non-relational databases? Non-relational databases, also known as NoSQL databases, handle unstructured or semi-structured data and can use formats like key-value pairs, documents, or graphs. They are highly flexible and scalable. | P a g e 20  What is SQL? SQL, or Structured Query Language, is a language used to query relational databases. It is a fundamental skill in tech for working with structured databases.  What is data redundancy? Redundant data refers to excess or unnecessary data which is avoided in relational databases by establishing relationships between tables.  What are primary and foreign keys? Primary keys are used to uniquely identify each record within a table, while foreign keys are used to establish relationships between tables.  What is a Database Management System (DBMS)? A DBMS is a software tool that provides the necessary functionality to create, manage, and manipulate databases. It allows users to store, retrieve, update, and organize data.  What is a relational database management system (RDBMS)? An RDBMS is a type of DBMS that specifically manages relational databases. It stores data in structured tables and operates on relational data. Examples include MySQL, PostgreSQL, Microsoft SQL Server, and Oracle.  What is a database engine? A database engine is a core software component of a DBMS that handles data storage, manipulation, and retrieval. In the context of AWS, especially RDS, the term 'database engine' refers to the database management systems themselves.  What are the challenges of managing databases? Managing databases involves tasks such as hardware setup, scaling resources, taking backups, and ensuring security, all of which can be complex and time-consuming.  What is Amazon Relational Database Service (RDS)? RDS is a fully managed relational database service provided by AWS that automates common database administration tasks. It simplifies the setup, operation, and scaling of relational databases in the cloud.  What are the key benefits of RDS? o Automated Backups: RDS automatically takes backups of your database, allowing for easy restoration. o Scalability: It allows for easy scaling of resources up or down. o High Availability: RDS provides automatic failover to a standby instance to minimise downtime. o Security: RDS offers built-in encryption for data at rest and in transit, along with integration with AWS IAM for access control.  What database engines does RDS support? RDS supports various relational database engines including MySQL, PostgreSQL, MariaDB, Oracle, Microsoft SQL Server, and Amazon Aurora.  What is Amazon Aurora? Amazon Aurora is a high-performance, cloud-native database compatible with MySQL and PostgreSQL.  What is the pricing model of RDS? RDS pricing is based on instance type, storage used, backup storage on S3, and data transfer. There is a free tier of 750 hours of a T2 micro instance available.  What is a T2 micro instance? A T2 micro instance is a specific type of virtual server with limited resources. It is part of AWS's free tier, available for one year to users.  What are the different tiers of database architectures? There are three different database architectures: one-tier, two-tier and three-tier. The most common architecture is three-tier, which includes a database, a server, and an application.  What is the relationship between RDS, EC2 and EBS? RDS databases are run by EC2 instances, and their storage is on EBS. However, RDS is a managed service so the user does not directly interact with EC2 or EBS instances. | P a g e 21 Session (13)  What is a relational database, and how does it differ from a non-relational database? o Relational databases are structured and organised in rows and columns, forming tables. They use tables to establish relationships. Non-relational databases (NoSQL) are becoming more popular and do not use a table based structure.  What is Amazon RDS, and what database engines can be used with it? o Amazon RDS (Relational Database Service) simplifies database management by automating time-consuming administrative tasks. It allows users to choose from various relational database engines, including MySQL, PostgreSQL, and Amazon Aurora.  What is Amazon Aurora, and why is it considered a 'cloud-native' database? o Amazon Aurora is a fully managed, cloud-native relational database service developed by AWS. It's considered cloud-native because it's built specifically for the cloud, following cloud architecture and using features like autoscaling and cloud configurations.  What does 'fully managed' mean in the context of database services like RDS and Aurora? o 'Fully managed' means that AWS handles the underlying infrastructure, maintenance, and upgrades, so users don't have to manage the infrastructure.  How does Aurora compare to MySQL in terms of performance and features? o Aurora is designed to be faster and more scalable than MySQL. It is up to five times faster than MySQL and three times faster than PostgreSQL. Aurora also offers advanced features like automatic scaling, higher availability, and enhanced performance due to its cloud-based architecture.  What are some key features of Amazon Aurora that make it suitable for modern applications? o Key features include high performance, high availability through data replication across multiple Availability Zones (AZs), autoscaling of storage and compute resources and compatibility with MySQL and PostgreSQL.  What is the significance of Aurora's compatibility with MySQL and PostgreSQL? o This compatibility means that applications built for MySQL or PostgreSQL can be migrated to Aurora without requiring significant modifications. This makes it easier to leverage the advanced features of Aurora.  How does Aurora ensure high availability and data durability? o Aurora creates six copies of your data across three Availability Zones (AZs), which ensures high availability and fault tolerance in case of hardware failures.  What are commercial databases, and why might they be needed instead of open- source databases like MySQL? o Commercial databases are licensed databases with special features for autoscaling, advanced security, and high-traffic environments. They are often used by companies like Visa, Netflix and Facebook. They are needed when applications require very high performance, security, and reliability that open- source databases may not provide.  What are the benefits of using Aurora compared to using MySQL in RDS? | P a g e 22 o Although both are relational databases, Aurora offers better performance, scalability, and automatic data replication across multiple AZs, which RDS does not do by default. Aurora also provides a serverless option, which RDS does not.  What is Amazon Aurora Serverless (also called Aurora Serverless v2)? o Aurora Serverless is an on-demand, autoscaling configuration of Aurora that dynamically adjusts database capacity based on application needs. It simplifies database management and is cost-efficient for workloads with unpredictable demand. It automatically scales compute resources and only charges for the capacity used.  How does Aurora Serverless differ from provisioned Aurora in terms of managing database resources? o With provisioned Aurora, you select and manage specific instance types and compute resources. With Aurora Serverless, the service automatically manages compute resources based on workload, removing the need for manual selection and provisioning.  What are Aurora Capacity Units (ACUs)? o ACUs are used by Aurora Serverless to measure database capacity. An ACU represents a combination of processing power and memory.  What are the key features of Aurora Serverless? o Key features include automatic scaling of compute resources, cost efficiency through pay-per-use billing, high availability, fast setup, and easy switching between provisioned and serverless configurations.  What is the potential cost saving that Aurora Serverless offers?  Aurora Serverless can save up to 90% in database costs compared to provisioning capacity manually.  What are some use cases for Amazon Aurora? o Aurora is suitable for various applications, such as e-commerce platforms, financial applications, and mobile/gaming apps due to its ability to handle high volumes of transactions and offer high performance.  Why might someone choose to use MySQL in RDS instead of Aurora? o The main reason to choose MySQL in RDS instead of Aurora would be cost. MySQL is generally less expensive than Aurora, which offers supercharged features.  Does Aurora use the global infrastructure of AWS? o Yes, Aurora uses the global infrastructure of AWS and replicates data across multiple availability zones.  How does Aurora handle data storage and scaling? o Aurora offers auto scaling storage up to 64 terabytes. It allows data to be stored across multiple availability zones automatically. Session (14)  Question: What are the two main categories of databases? o Answer: The two main categories of databases are relational databases (also known as table-based databases) and non-relational databases (also known as NoSQL databases). | P a g e 23  Question: What is the primary difference between relational and non-relational databases? o Answer: Relational databases store data in structured tables with rows and columns, and establish relationships between tables. Non-relational databases handle unstructured or semi-structured data and do not use a traditional table structure.  Question: What does NoSQL stand for? o Answer: NoSQL stands for "not only SQL," referring to a class of database systems that handle unstructured or semi-structured data, unlike traditional relational databases.  Question: What are the four main types of NoSQL databases? o Answer: The four main types of NoSQL databases are: key-value stores, document databases, column family stores, and graph databases.  Question: How do key-value stores work? o Answer: Key-value stores store data as key-value pairs, allowing for fast and simple data access. The key is used to retrieve the associated value, similar to how a Python dictionary works.  Question: How do document databases work? o Answer: Document databases store data in flexible, JSON-like formats such as BSON, making them suitable for handling complex data structures. They use a dynamic schema, allowing for flexibility in data structure.  Question: How do column family stores work? o Answer: Column family stores data in columns rather than rows and can efficiently store and retrieve large amounts of data, with rows acting as the "mother" of several columns.  Question: How do graph databases work? o Answer: Graph databases store data in nodes and edges, which are particularly useful for representing complex relationships between data.  Question: Is one type of database (relational or NoSQL) universally better than the other? o Answer: No, neither type of database is universally better. Each excels in different scenarios and use cases. Relational databases are better for structured data and complex queries, while NoSQL databases are better for flexible, unstructured data and large-scale, real-time applications.  Question: What are some examples of use cases for NoSQL databases? o Answer: NoSQL databases are commonly used for chat applications, multiplayer games, IoT sensors, content management systems, big data analytics, e-commerce, and social networking.  Question: What is Amazon DynamoDB? o Answer: Amazon DynamoDB is a NoSQL database service provided by AWS. It is a fully managed, serverless, key-value and document database service.  Question: What are the key features of DynamoDB? o Answer: Key features of DynamoDB include:  Scalability: It automatically scales up or down based on throughput needs.  Low latency: Provides consistent single-digit millisecond latency for read and write operations.  Serverless: No need to provision or manage servers.  Security: Offers encryption at rest and in transit. | P a g e 24  Global tables: Replicates data across regions with low latency.  Built-in backup and restore: Automated backups and point-in-time recovery.  Question: What types of applications is DynamoDB ideal for? o Answer: DynamoDB is ideal for applications needing real-time updates like live chat and gaming leaderboards, IoT applications, e-commerce applications, and mobile/web applications.  Question: What are the pricing options for DynamoDB? o Answer: DynamoDB offers flexible billing options, including on-demand (pay-per-request) and provisioned capacity (specified resources) models.  Question: What is the primary key in DynamoDB? o Answer: The primary key consists of the partition key which is a unique identifier that lets you retrieve data from a table and a sort key which is optional and allows sorting and searching. Session (15)  What is Amazon DocumentDB? o Amazon DocumentDB is a fully managed, scalable, and highly available document database service that is compatible with MongoDB. It is designed for operational workloads that require the document model.  How does DocumentDB relate to NoSQL databases? o DocumentDB is a type of NoSQL database that stores data as documents in a JSON-like format. NoSQL databases are non-relational, designed for high- performance environments, and are schema-flexible. Document databases are one of four types of NoSQL databases, the others being key-value stores, column family stores, and graph databases.  What is a document in the context of DocumentDB? o In DocumentDB, a document is a single piece of data that is stored independently in a JSON-like format. It is not a Windows document or a file, but rather a data item that can operate autonomously and be stored with a flexible schema.  How does DocumentDB compare to DynamoDB? o While both are NoSQL databases on AWS, DynamoDB is primarily designed for key-value storage but also supports documents, and is fully serverless with extreme scalability for high throughput workloads. DocumentDB, on the other hand, is specifically designed for the document model and is compatible with MongoDB.  What are the key features of DocumentDB? o Key features include:  Compatibility with MongoDB: It is fully compatible with MongoDB APIs.  Scalability: It allows for easy scaling of storage and compute resources.  High Availability: It provides multi-AZ deployments with automatic failovers.  Automated Backups: It automatically takes daily backups with a retention period of 35 days.  Monitoring and Alerts: It offers built-in monitoring and customizable alerts. | P a g e 25  Security and Encryption DocumentDB provides security and encryption.  What are the common use cases for DocumentDB? o DocumentDB is suitable for:  Content Management Systems (CMS): Such as blogs or knowledge bases.  Mobile and Web Applications: Particularly those requiring fast and flexible access to JSON-like documents.  Catalog Management: Like e-commerce platforms with dynamic, semi-structured data.  How is DocumentDB priced? o Pricing is based on the compute used (instance type and size), the actual storage consumed, and snapshot storage beyond the free tier. DocumentDB itself is not included in the AWS free tier.  Can existing MongoDB applications be migrated to DocumentDB? o Yes, because DocumentDB is compatible with MongoDB, existing MongoDB applications can be migrated to AWS and hosted in DocumentDB without changes. The AWS Database Migration Service can assist with the migration.  What are the different types of NoSQL databases? o The four types of NoSQL databases are:  Document databases: Store data as documents (e.g., Amazon DocumentDB, MongoDB)  Key-value stores: Store data as key-value pairs (e.g., Redis, DynamoDB).  Column family stores: Organize data into columns and rows, with each row containing multiple columns (e.g., Cassandra, Google Bigtable).  Graph databases: Store data as nodes and edges, which represent entities and relationships between entities (e.g., Amazon Neptune, Neo4j). Session (16)  What is Amazon Redshift? o Amazon Redshift is a fast, fully managed, scalable data warehouse service designed for analytics and big data processing. o It is a petabyte-scale data warehouse. o Redshift allows users to query structured and semi-structured data using SQL. o It integrates with business intelligence (BI) tools like Tableau and PowerBI. o It supports complex queries, advanced analytics, and machine learning (ML) integrations. o Redshift integrates with other AWS services like S3, AWS Glue, and Amazon SageMaker.  What is a data warehouse? o A data warehouse is a centralized system or repository where large volumes of structured data from multiple sources are collected, organized, and stored. | P a g e 26 o It is designed for the purpose of data analysis, business intelligence, and reporting. o Data warehouses are typically large-scale, enterprise-level systems, often leveraging cloud-based solutions. o They are optimized for read-heavy workloads and complex querying. o A data warehouse is often called a single source of truth because it consolidates data from various sources into one unified and accurate data source.  What is the difference between a database and a data warehouse? o Databases are connected to live transactional systems for real-time operations (OLTP), where data is created, read, updated, and deleted (CRUD). o Data warehouses are not connected to live systems; they are data sources primarily for analysis (OLAP) where data is fetched from multiple sources such as databases. o A database is a structured collection of data for efficient access and management, while a data warehouse is a centralized repository for analytical and reporting purposes.  What is a data lake? o A data lake is a scalable, centralized repository that stores raw, unstructured, semi-structured, and structured data in its native format. o It is a storage location for unprocessed data. o Data is often extracted from the lake, then transformed for specific purposes. o Data Lakes can be used for big data analytics, machine learning, and data exploration.  What is the ETL process? o ETL stands for Extract, Transform, and Load. o It is the process of transforming raw data into high-quality data optimized for analytics. o Data is extracted from a data lake, transformed, and then loaded into a data warehouse.  What are the key components of Redshift architecture? o Redshift works with nodes and clusters, including a leader node and compute nodes. o The leader node manages client connections and distributes queries to compute nodes. o The compute nodes handle the actual data storage and query execution. o Redshift uses columnar storage which reduces disk I/O and improves query performance. o Redshift uses Massively Parallel Processing (MPP) to distribute data and query execution across multiple nodes.  What are the different types of Redshift and how are they priced? o Redshift comes in two types: Redshift Serverless and Redshift Provisioned.  With provisioned Redshift, you manually provision and manage your clusters or nodes.  With serverless Redshift, you do not have to manage the clusters. o Pricing is based on node type (Dense Compute or Dense Storage), usage (on- demand or reserved instances), and data transfer.  Inbound data transfers are free, while outbound transfers incur charges.  What are the use cases for Redshift? | P a g e 27 o Redshift is used for business intelligence, performing analytics on sales, marketing, and operational data. o It is used for big data analytics, analysing data from IoT devices or web applications. o It enables data lake integration, allowing the querying of structured and unstructured data using Redshift Spectrum.  What are the different roles in data management and their functions? o Data Engineers: Responsible for getting data from its original source into systems where the data will be ready to be used for analysis, which includes collecting, cleaning, and preparing data, and creating automated pipelines for data processing. o Data Analysts: Responsible for performing data analysis, extracting insights, and generating reports from processed data. They often use tools like SQL, Python, and data visualization software. o Data Scientists: They use machine learning tools and processes to make predictions and create intelligent models based on data. They use advanced techniques and programming languages like Python to analyze data and build predictive models. Session (17) 1. What is a Virtual Private Cloud (VPC) in AWS?  A VPC is a secure, private cloud network that you create inside AWS for your own applications and AWS services.  It's like reserving a portion of the AWS cloud and making it private and dedicated for your use, isolated from the outside world.  It allows you to have control over your cloud and configure your own policies for how your applications operate and how resources are accessed on the internet.  A VPC is a logically isolated section of the AWS cloud where you can launch AWS resources.  A VPC is tied to a single region and spans multiple availability zones. 2. What are the benefits of using a VPC?  Isolation and Security: Isolates resources for security and compliance.  Custom Networking: Allows you to customize network configurations.  Scalability: Enables easy scaling of resources within the VPC.  Hybrid Connectivity: Allows you to connect your on-premises network to AWS. 3. What are subnets within a VPC?  Subnets are ranges of IP addresses within your VPC.  They are used to segment the VPC into smaller, manageable pieces and to separate resources based on security, accessibility, and other factors.  Subnets are tied to availability zones.  Public subnets are configured to allow access to the internet and are typically used for web servers, load balancers, and NAT gateways. | P a g e 28  Private subnets are not directly accessible from the internet and are typically used for resources like databases and internal applications. 4. What is an IP address and why is it important for networking?  An IP address is a unique identifier assigned to a device on a network.  It allows computers and other devices to communicate, share resources, and exchange information.  An IP address has two purposes: it identifies a device on a network and provides a location for routing data.  Without an IP address, a device is not known or identifiable on a network.  IP addresses are the names that are used to address computers on a network since they are not people. 5. What are the main types of IP addresses?  There are two main types of IP addresses: IPv4 and IPv6.  IPv4 is a 32-bit address, divided into four octets, and supports approximately 4.3 billion unique addresses.  IPv6 is a 128-bit address designed to address the IPv4 exhaustion issue, and it supports a vastly larger number of addresses (approximately 340 undecillion).  Private IP addresses are used within private networks and are not accessible from the open internet, while public IP addresses are globally unique addresses assigned for communication over the internet. 6. What is CIDR notation and how is it used?  CIDR (Classless Inter-Domain Routing) notation is a method of writing IP addresses, where an IP address is followed by a slash and a number, e.g., a.b.c.d/n.  The 'a.b.c.d' is the base IP address of the network, and the 'n' represents the number of bits used for the network portion, leaving the remaining bits for the host portion.  CIDR notation provides a flexible way to allocate IP address ranges for computer networks.  It helps define the range of IP addresses available for a network and manage routing on computer networks.  It replaced the old class-based IP addressing system, making it more efficient. 7. What are the functions of the internet gateway and the NAT gateway?  An internet gateway (IGW) allows communication between your VPC and the internet.  It enables resources in a public subnet to access the internet and allows external users to reach those resources.  A NAT gateway enables instances in a private subnet to access the internet while keeping them secure from direct inbound traffic.  The NAT gateway is deployed in a public subnet, and it routes traffic from private subnets to the internet through the internet gateway.  The NAT gateway provides outbound internet access. 8. What is a virtual private gateway and how does it connect to on-premises networks? | P a g e 29  A virtual private gateway is a component that allows you to connect your VPC to your on-premises network via a VPN connection or via AWS Direct Connect.  It allows for secure connection between an on-premise data center to your VPC.  A VPN connection creates a secure communication channel, while AWS Direct Connect can be used for high bandwidth, low latency connections. 9. What are security groups and network access control lists (NACLs)?  Both are used to secure your AWS resources.  Security groups act as a virtual firewall to control inbound and outbound traffic to AWS resources like EC2 instances.  Security groups are stateful, meaning if inbound traffic is allowed, corresponding outbound traffic is automatically allowed.  Network ACLs (NACLs) are also a firewall but operate at the subnet level and are stateless.  NACLs require both inbound and outbound traffic rules to be explicitly set up. 10. What is subnetting and why is it important?  Subnetting is the process of dividing a larger network into smaller networks, each called a subnet.  It is important for network performance, security, and management.  Subnetting improves network performance by reducing traffic, enhances security by segmenting sensitive areas, simplifies troubleshooting by localizing issues, and facilitates easier network expansion.  Subnetting allows you to configure the network into more manageable pieces. | P a g e 30

Use Quizgecko on...
Browser
Browser