October 21, 2020 By Vidyasagar Machupalli 4 min read

Learn how to provision and auto scale VSI on VPC using Terraform scripts and explore auto scale scenarios like SSL termination (offloading) and end-to-end encryption.

With Auto Scale for VPC, you can create an instance group to scale according to your requirements. Based on the target utilization metrics that you define, the instance group can dynamically add or remove instances to achieve your specified instance availability.

Let’s start with a quick one, like an introduction to VPC.

What is a VPC?

A virtual private cloud (VPC) is a public cloud offering that lets an enterprise establish its own private cloud-like computing environment on shared public cloud infrastructure. A VPC gives an enterprise the ability to define and control a virtual network that is logically isolated from all other public cloud tenants, creating a private, secure place on the public cloud.

Auto scaling on VPC

With Auto Scale for VPC, you can improve performance and costs by dynamically creating virtual server instances (VSIs) to meet the demands of your environment. You set scaling policies that define your desired average utilization for metrics like CPU, memory, and network usage. You can create an instance group in your IBM Cloud VPC to auto scale according to your requirements by using the IBM Cloud console. Based on the target utilization metrics that you define, the instance group can dynamically add or remove instances to achieve your specified instance availability.

What’s the outcome?

Following the instructions and executing the Terraform scripts, you will provision the following:

  • A VPC
  • Two subnets — one in each zone (zone 1 and zone 2) — Configure in main.tf file
  • One VSI in subnet-zone-1 (to start with)
  • Load balancer (backend pool and frontend listener)
  • Instance group using an instance template
  • Instance group manager with a scaling policy (CPU utilization) attached to the instance group

Pre-requisites

  • IBM Cloud IAM API key (refer to the IAM documentation)
  • SSH key (refer to the VPC SSH documentation)
  • If you want to enable SSH offloading (termination) or end-to-end encryption, you need to create a Certificate manager service, order a certificate, and pass that to load balancer HTTPS listener. Follow the instructions here to learn the certificate ordering process.

Step-by-step instructions

  1. Clone the vpc-tutorials GitHub repository and move to the vpc-autoscale directory:
    git clone https://github.com/IBM-Cloud/vpc-tutorials.git
    cd vpc-autoscale
  2. Copy the Terraform variables template:
    cp terraform.tfvars.template terraform.tfvars
  3. Update the values in the terraform.tfvars file based on your requirement and save the file. To understand the variables, refer to the variables.tf file.
  4. Initialize the Terraform providers by running the following:
    terraform init
  5. Execute the Terraform plan:
    terraform plan
  6. Apply the Terraform plan:
    terraform apply

Note: The max_membership_count to scale is set to 5, the aggregation_window is set to 90 secs, and cooldown is set to 120 secs. You can change these values in the main.tf file. Refer to the Instance-group-manager Terraform provider documentation.

SSL offloading (termination)

For all incoming HTTPS connections, the load balancer service terminates the SSL connection and establishes a plain-text HTTP communication with the backend server. CPU-intensive SSL handshakes and encryption/decryption tasks are shifted away from the backend servers, allowing them to use all their CPU cycles for processing application traffic.

End-to-end encryption

In this section, you will enable end-to-end encryption by creating a self-signed certificate in the VSI. You will use the SSL installation script – install-software-ssl.sh.

Clean-up

Run the following command to delete the resources you created. The action is irreversible:

terraform destroy

Terraform-as-a-Service

With IBM Cloud Schematics, you can simplify the above Terraform experience. Schematics delivers Terraform-as-a-Service so that you can use a high-level scripting language to model the resources that you want in your IBM Cloud environment and enable Infrastructure as Code (IaC).

Refer to this post to understand how Schematics works with an example.

What’s next?

How about automatically assigning a floating IP to a newly created VSI by monitoring Activity Tracker events and using Cloud Functions to interact with the VPC API? Yes, you can achieve this by following the instructions in this post: “Extend VPC Resources with Cloud Functions, Activity Tracker with LogDNA, and Schematics.” 

Explore other VPC scenarios

There are additional scripts in the vpc-tutorials repo that are used to set-up scenarios for VPC tutorials and to clean up VPC resources:

Questions and feedback

If you have feedback, suggestions, or questions about this post, please reach out to me on Twitter @VidyasagarMSC or use the feedback button on the tutorial to report a problem on its content. You can also open issues.

The tutorials section has a feedback form on the side where you can comment on the content. If you have suggestions on the existing tutorials or ideas for future additions, please submit your feedback.

Was this article helpful?
YesNo

More from Cloud

Apache Kafka use cases: Driving innovation across diverse industries

6 min read - Apache Kafka is an open-source, distributed streaming platform that allows developers to build real-time, event-driven applications. With Apache Kafka, developers can build applications that continuously use streaming data records and deliver real-time experiences to users. Whether checking an account balance, streaming Netflix or browsing LinkedIn, today’s users expect near real-time experiences from apps. Apache Kafka’s event-driven architecture was designed to store data and broadcast events in real-time, making it both a message broker and a storage unit that enables real-time…

Primary storage vs. secondary storage: What’s the difference?

6 min read - What is primary storage? Computer memory is prioritized according to how often that memory is required for use in carrying out operating functions. Primary storage is the means of containing primary memory (or main memory), which is the computer’s working memory and major operational component. The main or primary memory is also called “main storage” or “internal memory.” It holds relatively concise amounts of data, which the computer can access as it functions. Because primary memory is so frequently accessed,…

Cloud investments soar as AI advances

3 min read - These days, cloud news often gets overshadowed by anything and everything related to AI. The truth is they go hand-in-hand since many enterprises use cloud computing to deliver AI and generative AI at scale. "Hybrid cloud and AI are two sides of the same coin because it's all about the data," said Ric Lewis, IBM’s SVP of Infrastructure, at Think 2024. To function well, generative AI systems need to access the data that feeds its models wherever it resides. Enter…

IBM Newsletters

Get our newsletters and topic updates that deliver the latest thought leadership and insights on emerging trends.
Subscribe now More newsletters