Amazon Web Services
AttentionThis page documents an earlier version. Go to the latest (v2.1)version.
Prerequisites
Download and install terraform.
Verify by the
terraform
command, it should print a help message that looks similar to that shown below.
$ terraform
Usage: terraform [--version] [--help] <command> [args]
...
Common commands:
apply Builds or changes infrastructure
console Interactive console for Terraform interpolations
destroy Destroy Terraform-managed infrastructure
env Workspace management
fmt Rewrites config files to canonical format
1. Create a terraform config file
Create a terraform config file called yugabyte-db-config.tf
and add following details to it. The terraform module can be found in the terraform-aws-yugabyte github repository.
provider "aws" {
# Configure your AWS account credentials here.
access_key = "ACCESS_KEY_HERE"
secret_key = "SECRET_KEY_HERE"
region = "us-west-2"
}
module "yugabyte-db-cluster" {
# The source module used for creating AWS clusters.
source = "github.com/Yugabyte/terraform-aws-yugabyte"
# The name of the cluster to be created, change as per need.
cluster_name = "test-cluster"
# Existing custom security group to be passed so that we can connect to the instances.
# Make sure this security group allows your local machine to SSH into these instances.
custom_security_group_id="SECURITY_GROUP_HERE"
# AWS key pair that you want to use to ssh into the instances.
# Make sure this key pair is already present in the noted region of your account.
ssh_keypair = "SSH_KEYPAIR_HERE"
ssh_key_path = "SSH_KEY_PATH_HERE"
# Existing vpc and subnet ids where the instances should be spawned.
vpc_id = "VPC_ID_HERE"
subnet_ids = ["SUBNET_ID_HERE"]
# Replication factor of the YugabyteDB cluster.
replication_factor = "3"
# The number of nodes in the cluster, this cannot be lower than the replication factor.
num_instances = "3"
}
NOTE: If you do not have a custom security group, you would need to remove the ${var.custom_security_group_id}
variable in main.tf
, so that the aws_instance
looks as follows:
resource "aws_instance" "yugabyte_nodes" {
count = "${var.num_instances}"
...
vpc_security_group_ids = [
"${aws_security_group.yugabyte.id}",
"${aws_security_group.yugabyte_intra.id}",
"${var.custom_security_group_id}"
]
2. Create a cluster
Init terraform first if you have not already done so.
$ terraform init
Now run the following to create the instances and bring up the cluster.
$ terraform apply
Once the cluster is created, you can go to the URL http://<node ip or dns name>:7000
to view the UI. You can find the node’s ip or dns by running the following:
$ terraform state show aws_instance.yugabyte_nodes[0]
You can access the cluster UI by going to any of the following URLs.
You can check the state of the nodes at any point by running the following command.
$ terraform show
3. Verify resources created
The following resources are created by this module:
module.yugabyte-db-cluster.aws_instance.yugabyte_nodes
The AWS instances.
For cluster named test-cluster
, the instances will be named yb-ce-test-cluster-n1
, yb-ce-test-cluster-n2
, yb-ce-test-cluster-n3
.
module.yugabyte-db-cluster.aws_security_group.yugabyte
The security group that allows the various clients to access the YugabyteDB cluster.
For cluster named test-cluster
, this security group will be named yb-ce-test-cluster
with the ports 7000, 9000, 9042 and 6379 open to all other instances in the same security group.
module.yugabyte-db-cluster.aws_security_group.yugabyte_intra
The security group that allows communication internal to the cluster.
For cluster named test-cluster
, this security group will be named yb-ce-test-cluster-intra
with the ports 7100, 9100 open to all other instances in the same security group.
module.yugabyte-db-cluster.null_resource.create_yugabyte_universe
A local script that configures the newly created instances to form a new YugabyteDB universe.
4. Destroy the cluster (optional)
To destroy what we just created, you can run the following command.
$ terraform destroy