November 11, 2019 By Douglas Paris-White 3 min read

Here are six must-see sessions at KubeCon 2019 (November 18 – 21, 2019) that focus on the challenges for DevOps teams doing cloud native development with Kubernetes.

Cloud native application development, as the best way to compete as a digital business, is driving the innovation of cloud platforms themselves.

Kubernetes, by enabling application development and operations teams to do cloud-based work at velocity and scale, has become a foundational technology in the ecosystem of continuous delivery tools and practices.

Automating operations is the key. By automating operations, Kubernetes frees DevOps teams to keep their focus on what the customers want from the applications that are designed to grow the business. The declarative configuration model gives teams the power to chart a normal state for an application in production. Kubernetes takes care of running pods on the right nodes in a cluster and across regions, scaling up and down instances of relevant microservices to meet user demand.

Making your plan for KubeCon 2019

Here are six sessions at KubeCon 2019 (November 18 – 21, 2019) that focus on the  challenges for DevOps teams doing cloud native development with Kubernetes.

Running Kubernetes in production at high scale

Containerd Mini-Summit: Phil Estes (Distinguished Engineer & CTO, Container and Linux OS) leads this program on the container engine that enables Kubernetes to operate at very high scale.

Managing a DevOps pipeline

Mario’s Adventure in Tekton Land (Tuesday, November 19, 4:25pm – 5:00pm): Vincent Demeester (Principal Software Engineer, Red Hat) and Andrea Frittoli  (Open Source Developer Advocate, IBM Cloud) take you through replacing a pipeline defined with bash scripts to one based on re-usable Tekton modules.

Managing deployment patterns

Advanced Model Inferencing Leveraging Knative, Istio, and Kubeflow Serving (Wednesday, November 20, 10:55am – 11:30am): Animesh Singh (STSM and Program Director) covers how to handle autoscaling, scale-to-zero, canary, and other deployment patterns using Kubeflow Serving and the native Kubernetes stack (Knative, Istio).  

Managing microservices

Ready to Serve! Speeding-Up Startup Time of Istio-Powered Workloads (Thursday, November 21, 4:25pm – 5:00pm): Michal Malka (Manager, IBM Research) and Etai Lev Ran (System Architect, IBM)  analyze the latency contributed by Istio service mesh to pod startup time, right from pod creation and up to the pod becoming ready to service requests. They’ll also examine various techniques to reduce it, including using Istio CNI to bootstrap the pod’s network, launching the sidecar proxy with an initial routing configuration, and using manual sidecar injection.

Enabling serverless workloads on Kubernetes

CNCF’s Serverless Working Group: Tell Me Where it Hurts (Thursday, November 21, 4:25pm – 5:55pm): Doug Davis (STSM & Offering Manager, Knative) provides a community update on the state of serverless in Kubernetes. This session involves active discussion, so come with your pain points and opinions on the interoperability and portable of serverless workloads

Kubernetes does drones?

Flying Kubernetes: Using Drones to Understand how Kubernetes Works (IBM Cloud Theater, 11:00-11:30 am on November 19 and 20): Watch IBM Cloud leaders Briana Frank and Jason McGee put Kubernetes through its paces with a fleet of drones.

Running an application in Kubernetes involves creating and applying a deployment configuration file. Scaling involves manually defining a replica set in that configuration or defining an autoscaling rule (specifying a minimum and maximum number of pods to run) to tell Kubernetes how to respond with traffic fluctuations. That same deployment configuration enables an application to self-recover; should one of its instances go down, Kubernetes takes action to match the configuration.

In terms of managing load among the worker nodes assigned to host instances of an application, Kubernetes supports affinity rules; by default, Kubernetes assigns pods to a worker that is currently not running any and then round-robins additional pods onto the workers assigned to an application. You can take control of pod-worker assignments, as needed, by configuring affinity more strictly.

Maybe you already know those Kubernetes concepts and have seen them in action with your applications. But have you seen the concepts used to control the behavior of drones?

The drones are programmed to watch Kubernetes for changes in configuration to an application with whose pods/instances they are associated. As a bonus, you’ll see the drones demonstrate a canary testing configuration for an application using the open source Istio microservices mesh.

More about Kubernetes, containers, and the IBM Cloud Kubernetes Service

Explore the IBM Cloud Kubernetes Service.

Check out our lightboarding video series on various Kube and containers topics:

More from Automation

Announcing the general availability of IBM Concert

< 1 min read - At Think 2024, we announced IBM Concert®. It provides generative AI driven insights for your applications and puts site reliability engineers (SREs) and developers in control, enabling them to simplify and optimize their operations across any environment. IBM Concert is now generally available. You can now start using IBM Concert to get a detailed view of your applications and environments and apply generative AI to get insights on how to optimize your applications so your business works better.  Powered by…

Making HTTPS redirects easy with IBM NS1 Connect

3 min read - HTTPS is now the standard for application and website traffic on the internet. Over 85% of websites now use HTTPS by default—it’s to the point where a standard HTTP request now seems suspicious.  This is great for the security of the internet, but it’s a huge pain for the website and application teams that are managing HTTPS records. It was easy to move HTTP records around with a simple URL redirect. HTTPS redirects, on the other hand, require changing the URL…

5 SLA metrics you should be monitoring

7 min read - In business and beyond, communication is king. Successful service level agreements (SLAs) operate on this principle, laying the foundation for successful provider-customer relationships. A service level agreement (SLA) is a key component of technology vendor contracts that describes the terms of service between a service provider and a customer. SLAs describe the level of performance to be expected, how performance will be measured and repercussions if levels are not met. SLAs make sure that all stakeholders understand the service agreement…

IBM Newsletters

Get our newsletters and topic updates that deliver the latest thought leadership and insights on emerging trends.
Subscribe now More newsletters