Horizontal scaling means that you scale by adding more ec2 machinesinto your pool of resources whereasVertical scaling means that you scale by adding more power (CPU, RAM) to an existing ec2 machine.
Vertical Scaling usually means upgrade of server hardware.
Horizontal Scaling is adding more ec2 machines.
With horizontal-scaling it is often easier to scale dynamically by adding more machines into the existing pool — Vertical-scaling is often limited to the capacity of a single machine, scaling beyond that capacity often involves downtime and comes with an upper limit.
Good examples of horizontal scaling are Redis, MongoDB.
Good example of vertical scaling is MySQL — Amazon RDS (The cloud version of MySQL).
RAID : Stands for Redundant Array of Independent Disks
RAID is a data storage virtualization technology that combines multiple physical disk drive components into one or more logical units for the purposes of data redundancy, performance improvement, or both.
It is among the most commonly asked AWS interview questions. There is little room to be creative – you either know or you don’t know the answer.
Amazon Web Services or AWS is a set of cloud computing services and tools from Amazon. It offers over 200 comprehensive data center services globally. AWS is a cross-functional platform that offers a wide variety of services ranging from data warehousing to content delivery.
When you launch a new EC2 instance, the EC2 service attempts to place the instance in such a way that all of your instances are spread out across underlying hardware to minimize correlated failures. You can use placement groups to influence the placement of a group of interdependent instances to meet the needs of your workload. Depending on the type of workload, you can create a placement group using one of the following placement strategies:
Cluster – packs instances close together inside an Availability Zone. This strategy enables workloads to achieve the low-latency network performance necessary for tightly-coupled node-to-node communication that is typical of HPC applications.
Partition – spreads your instances across logical partitions such that groups of instances in one partition do not share the underlying hardware with groups of instances in different partitions. This strategy is typically used by large distributed and replicated workloads, such as Hadoop, Cassandra, and Kafka.
Spread – strictly places a small group of instances across distinct underlying hardware to reduce correlated failures.
Placement group strategies
You can create a placement group using one of the following placement strategies:
Cluster placement groups
A cluster placement group is a logical grouping of instances within a single Availability Zone. A cluster placement group can span peered VPCs in the same Region. Instances in the same cluster placement group enjoy a higher per-flow throughput limit for TCP/IP traffic and are placed in the same high-bisection bandwidth segment of the network.
The following image shows instances that are placed into a cluster placement group.
Cluster placement groups are recommended for applications that benefit from low network latency, high network throughput, or both. They are also recommended when the majority of the network traffic is between the instances in the group. To provide the lowest latency and the highest packet-per-second network performance for your placement group, choose an instance type that supports enhanced networking. For more information, see Enhanced Networking.
We recommend that you launch your instances in the following way:
Use a single launch request to launch the number of instances that you need in the placement group.
Use the same instance type for all instances in the placement group.
If you try to add more instances to the placement group later, or if you try to launch more than one instance type in the placement group, you increase your chances of getting an insufficient capacity error.
If you stop an instance in a placement group and then start it again, it still runs in the placement group. However, the start fails if there isn't enough capacity for the instance.
If you receive a capacity error when launching an instance in a placement group that already has running instances, stop and start all of the instances in the placement group, and try the launch again. Starting the instances may migrate them to hardware that has capacity for all of the requested instances.
Partition placement groups
Partition placement groups help reduce the likelihood of correlated hardware failures for your application. When using partition placement groups, Amazon EC2 divides each group into logical segments called partitions. Amazon EC2 ensures that each partition within a placement group has its own set of racks. Each rack has its own network and power source. No two partitions within a placement group share the same racks, allowing you to isolate the impact of hardware failure within your application.
The following image is a simple visual representation of a partition placement group in a single Availability Zone. It shows instances that are placed into a partition placement group with three partitions—Partition 1, Partition 2, and Partition 3. Each partition comprises multiple instances. The instances in a partition do not share racks with the instances in the other partitions, allowing you to contain the impact of a single hardware failure to only the associated partition.
Partition placement groups can be used to deploy large distributed and replicated workloads, such as HDFS, HBase, and Cassandra, across distinct racks. When you launch instances into a partition placement group, Amazon EC2 tries to distribute the instances evenly across the number of partitions that you specify. You can also launch instances into a specific partition to have more control over where the instances are placed.
A partition placement group can have partitions in multiple Availability Zones in the same Region. A partition placement group can have a maximum of seven partitions per Availability Zone. The number of instances that can be launched into a partition placement group is limited only by the limits of your account.
In addition, partition placement groups offer visibility into the partitions — you can see which instances are in which partitions. You can share this information with topology-aware applications, such as HDFS, HBase, and Cassandra. These applications use this information to make intelligent data replication decisions for increasing data availability and durability.
If you start or launch an instance in a partition placement group and there is insufficient unique hardware to fulfill the request, the request fails. Amazon EC2 makes more distinct hardware available over time, so you can try your request again later.
Spread placement groups
A spread placement group is a group of instances that are each placed on distinct racks, with each rack having its own network and power source.
The following image shows seven instances in a single Availability Zone that are placed into a spread placement group. The seven instances are placed on seven different racks.
Spread placement groups are recommended for applications that have a small number of critical instances that should be kept separate from each other. Launching instances in a spread placement group reduces the risk of simultaneous failures that might occur when instances share the same racks. Spread placement groups provide access to distinct racks, and are therefore suitable for mixing instance types or launching instances over time.
A spread placement group can span multiple Availability Zones in the same Region. You can have a maximum of seven running instances per Availability Zone per group.
If you start or launch an instance in a spread placement group and there is insufficient unique hardware to fulfill the request, the request fails. Amazon EC2 makes more distinct hardware available over time, so you can try your request again later.
Placement group rules and limitations
General rules and limitations
Before you use placement groups, be aware of the following rules:
You can create a maximum of 500 placement groups per account in each Region.
The name that you specify for a placement group must be unique within your AWS account for the Region.
You can't merge placement groups.
An instance can be launched in one placement group at a time; it cannot span multiple placement groups.
On-Demand Capacity Reservation and zonal Reserved Instances provide a capacity reservation for EC2 instances in a specific Availability Zone. The capacity reservation can be used by instances in a placement group. However, it is not possible to explicitly reserve capacity for a placement group.
You cannot launch Dedicated Hosts in placement groups.
Cluster placement group rules and limitations
The following rules apply to cluster placement groups:
The following previous generation instances: A1, C3, cc2.8xlarge, cr1.8xlarge, G2, hs1.8xlarge, I2, and R3.
A cluster placement group can't span multiple Availability Zones.
The maximum network throughput speed of traffic between two instances in a cluster placement group is limited by the slower of the two instances. For applications with high-throughput requirements, choose an instance type with network connectivity that meets your requirements.
For instances that are enabled for enhanced networking, the following rules apply:
Instances within a cluster placement group can use up to 10 Gbps for single-flow traffic. Instances that are not within a cluster placement group can use up to 5 Gbps for single-flow traffic.
Traffic to and from Amazon S3 buckets within the same Region over the public IP address space or through a VPC endpoint can use all available instance aggregate bandwidth.
You can launch multiple instance types into a cluster placement group. However, this reduces the likelihood that the required capacity will be available for your launch to succeed. We recommend using the same instance type for all instances in a cluster placement group.
Network traffic to the internet and over an AWS Direct Connect connection to on-premises resources is limited to 5 Gbps.
Partition placement group rules and limitations
The following rules apply to partition placement groups:
A partition placement group supports a maximum of seven partitions per Availability Zone. The number of instances that you can launch in a partition placement group is limited only by your account limits.
When instances are launched into a partition placement group, Amazon EC2 tries to evenly distribute the instances across all partitions. Amazon EC2 doesn’t guarantee an even distribution of instances across all partitions.
A partition placement group with Dedicated Instances can have a maximum of two partitions.
Spread placement group rules and limitations
The following rules apply to spread placement groups:
A spread placement group supports a maximum of seven running instances per Availability Zone. For example, in a Region with three Availability Zones, you can run a total of 21 instances in the group (seven per zone). If you try to start an eighth instance in the same Availability Zone and in the same spread placement group, the instance will not launch. If you need to have more than seven instances in an Availability Zone, then the recommendation is to use multiple spread placement groups. Using multiple spread placement groups does not provide guarantees about the spread of instances between groups, but it does ensure the spread for each group, thus limiting impact from certain classes of failures.
Spread placement groups are not supported for Dedicated Instances.
In the navigation pane, choose Placement Groups, Create placement group.
Specify a name for the group.
Choose the placement strategy for the group. If you choose Partition, choose the number of partitions within the group.
Choose Create group.
Tag a placement group
To help categorize and manage your existing placement groups, you can tag them with custom metadata. For more information about how tags work, see Tag your Amazon EC2 resources.
When you tag a placement group, the instances that are launched into the placement group are not automatically tagged. You need to explicitly tag the instances that are launched into the placement group. For more information, see Add a tag when you launch an instance.
You can view, add, and delete tags using the new console and the command line tools.
To view, add, or delete a tag for an existing placement group
Select a placement group, and then choose Actions, Manage tags.
The Manage tags section displays any tags that are assigned to the placement group. Do the following to add or remove tags:
To add a tag, choose Add tag, and then enter the tag key and value. You can add up to 50 tags per placement group. For more information, see Tag restrictions.
To delete a tag, choose Remove next to the tag that you want to delete.
Choose Launch Instance. Complete the wizard as directed, taking care to do the following:
On the Choose an Instance Type page, select an instance type that can be launched into a placement group.
On the Configure Instance Details page, the following fields are applicable to placement groups:
For Number of instances, enter the total number of instances that you need in this placement group, because you might not be able to add instances to the placement group later.
For Placement group, select the Add instance to placement group check box. If you do not see Placement group on this page, verify that you have selected an instance type that can be launched into a placement group. Otherwise, this option is not available.
For Placement group name, you can choose to add the instances to an existing placement group or to a new placement group that you create.
For Placement group strategy, choose the appropriate strategy. If you choose partition, for Target partition, choose Auto distribution to have Amazon EC2 do a best effort to distribute the instances evenly across all the partitions in the group. Alternatively, specify the partition in which to launch the instances.
Describe instances in a placement group
You can view the placement information of your instances using one of the following methods. You can also filter partition placement groups by the partition number using the AWS CLI.
To view the placement group and partition number of an instance using the console
On the Details tab, under Host and placement group, find Placement group. If the instance is not in a placement group, the field is empty. Otherwise, it contains the name of the placement group name. If the placement group is a partition placement group, Partition number contains the partition number for the instance.
Change the placement group for an instance
You can change the placement group for an instance in any of the following ways:
Move an existing instance to a placement group
Move an instance from one placement group to another
Remove an instance from a placement group
Before you move or remove the instance, the instance must be in the stopped state. You can move or remove an instance using the AWS CLI or an AWS SDK.
To move an instance to a placement group using the AWS CLI
If you need to replace a placement group or no longer need one, you can delete it. You can delete a placement group using one of the following methods.
Requirement
Before you can delete a placement group, it must contain no instances. You can terminate all instances that you launched into the placement group, move them to another placement group, or remove them from the placement group.