November 11, 2019 By Douglas Paris-White 3 min read

Here are six must-see sessions at KubeCon 2019 (November 18 – 21, 2019) that focus on the challenges for DevOps teams doing cloud native development with Kubernetes.

Cloud native application development, as the best way to compete as a digital business, is driving the innovation of cloud platforms themselves.

Kubernetes, by enabling application development and operations teams to do cloud-based work at velocity and scale, has become a foundational technology in the ecosystem of continuous delivery tools and practices.

Automating operations is the key. By automating operations, Kubernetes frees DevOps teams to keep their focus on what the customers want from the applications that are designed to grow the business. The declarative configuration model gives teams the power to chart a normal state for an application in production. Kubernetes takes care of running pods on the right nodes in a cluster and across regions, scaling up and down instances of relevant microservices to meet user demand.

Making your plan for KubeCon 2019

Here are six sessions at KubeCon 2019 (November 18 – 21, 2019) that focus on the  challenges for DevOps teams doing cloud native development with Kubernetes.

Running Kubernetes in production at high scale

Containerd Mini-Summit: Phil Estes (Distinguished Engineer & CTO, Container and Linux OS) leads this program on the container engine that enables Kubernetes to operate at very high scale.

Managing a DevOps pipeline

Mario’s Adventure in Tekton Land (Tuesday, November 19, 4:25pm – 5:00pm): Vincent Demeester (Principal Software Engineer, Red Hat) and Andrea Frittoli  (Open Source Developer Advocate, IBM Cloud) take you through replacing a pipeline defined with bash scripts to one based on re-usable Tekton modules.

Managing deployment patterns

Advanced Model Inferencing Leveraging Knative, Istio, and Kubeflow Serving (Wednesday, November 20, 10:55am – 11:30am): Animesh Singh (STSM and Program Director) covers how to handle autoscaling, scale-to-zero, canary, and other deployment patterns using Kubeflow Serving and the native Kubernetes stack (Knative, Istio).  

Managing microservices

Ready to Serve! Speeding-Up Startup Time of Istio-Powered Workloads (Thursday, November 21, 4:25pm – 5:00pm): Michal Malka (Manager, IBM Research) and Etai Lev Ran (System Architect, IBM)  analyze the latency contributed by Istio service mesh to pod startup time, right from pod creation and up to the pod becoming ready to service requests. They’ll also examine various techniques to reduce it, including using Istio CNI to bootstrap the pod’s network, launching the sidecar proxy with an initial routing configuration, and using manual sidecar injection.

Enabling serverless workloads on Kubernetes

CNCF’s Serverless Working Group: Tell Me Where it Hurts (Thursday, November 21, 4:25pm – 5:55pm): Doug Davis (STSM & Offering Manager, Knative) provides a community update on the state of serverless in Kubernetes. This session involves active discussion, so come with your pain points and opinions on the interoperability and portable of serverless workloads

Kubernetes does drones?

Flying Kubernetes: Using Drones to Understand how Kubernetes Works (IBM Cloud Theater, 11:00-11:30 am on November 19 and 20): Watch IBM Cloud leaders Briana Frank and Jason McGee put Kubernetes through its paces with a fleet of drones.

Running an application in Kubernetes involves creating and applying a deployment configuration file. Scaling involves manually defining a replica set in that configuration or defining an autoscaling rule (specifying a minimum and maximum number of pods to run) to tell Kubernetes how to respond with traffic fluctuations. That same deployment configuration enables an application to self-recover; should one of its instances go down, Kubernetes takes action to match the configuration.

In terms of managing load among the worker nodes assigned to host instances of an application, Kubernetes supports affinity rules; by default, Kubernetes assigns pods to a worker that is currently not running any and then round-robins additional pods onto the workers assigned to an application. You can take control of pod-worker assignments, as needed, by configuring affinity more strictly.

Maybe you already know those Kubernetes concepts and have seen them in action with your applications. But have you seen the concepts used to control the behavior of drones?

The drones are programmed to watch Kubernetes for changes in configuration to an application with whose pods/instances they are associated. As a bonus, you’ll see the drones demonstrate a canary testing configuration for an application using the open source Istio microservices mesh.

More about Kubernetes, containers, and the IBM Cloud Kubernetes Service

Explore the IBM Cloud Kubernetes Service.

Check out our lightboarding video series on various Kube and containers topics:

More from Automation

Deployable architecture on IBM Cloud: Simplifying system deployment

3 min read - Deployable architecture (DA) refers to a specific design pattern or approach that allows an application or system to be easily deployed and managed across various environments. A deployable architecture involves components, modules and dependencies in a way that allows for seamless deployment and makes it easy for developers and operations teams to quickly deploy new features and updates to the system, without requiring extensive manual intervention. There are several key characteristics of a deployable architecture, which include: Automation: Deployable architecture…

Understanding glue records and Dedicated DNS

3 min read - Domain name system (DNS) resolution is an iterative process where a recursive resolver attempts to look up a domain name using a hierarchical resolution chain. First, the recursive resolver queries the root (.), which provides the nameservers for the top-level domain(TLD), e.g.com. Next, it queries the TLD nameservers, which provide the domain’s authoritative nameservers. Finally, the recursive resolver  queries those authoritative nameservers.   In many cases, we see domains delegated to nameservers inside their own domain, for instance, “example.com.” is delegated…

Using dig +trace to understand DNS resolution from start to finish

2 min read - The dig command is a powerful tool for troubleshooting queries and responses received from the Domain Name Service (DNS). It is installed by default on many operating systems, including Linux® and Mac OS X. It can be installed on Microsoft Windows as part of Cygwin.  One of the many things dig can do is to perform recursive DNS resolution and display all of the steps that it took in your terminal. This is extremely useful for understanding not only how the DNS…

IBM Newsletters

Get our newsletters and topic updates that deliver the latest thought leadership and insights on emerging trends.
Subscribe now More newsletters