🚀 Blackbird is here! Accelerate your API development with AI-powered Code Gen, Instant Mocking with Prod-Like Environments. 👉 Start your free trial

Blog

The latest posts and insights about Ambassador Labs - our products, our ecosystem, as well as voices from across our community.

Kubernetes, Local Kubernetes

How to Run a Virtual Machine (VM) on a Local Kubernetes Cluster

Kubernetes has emerged as the de facto standard for container orchestration because it revolutionized the way applications are deployed and managed at scale. But sometimes, it is a good choice to run a GPU-intensive workload, tokenization, or LDAP/Active Directory applications on virtual machines (VMs) instead of containers for security, compatibility, or performance reasons. Virtual machines have long been the go-to technology for running diverse workloads, offering greater isolation, compatibility with legacy applications, and the ability to run operating systems of different types. However, managing VMs in a traditional infrastructure along with containerized applications on Kubernetes can be challenging and time-consuming. This is where Kubevirt steps in, offering a solution that allows you to run versatile and isolated virtual machines (VMs) inside your local Kubernetes cluster.

November 13, 2023 | 8 min read

Kubernetes

Kubernetes Best Practices 2024: Essential Tips for Modern Cluster Management

According to a report generated by Statista, around 61% of organizations worldwide have started using Kubernetes. One of the reasons for this is the many benefits that Kubernetes provides for businesses across the world. For instance, Kubernetes can automate self-healing, workload discovery, and scaling containerized applications. It also helps to simplify the process of deploying and running distributed apps at scale, thereby making it the ideal tool for managing microservices. In this article, I’ll highlight Kubernetes best practices that will help developers improve their project’s performance, security, and costs. 1. Kubernetes best practices for configuration

November 1, 2023 | 13 min read
Buoyant & Cloud-Native podcast

Reviving the Edge: ‘Livin’ On the Edge’ is Back

Flynn from Buoyant & Cloud-Native Happenings In the ever-evolving landscape of cloud-native technology, APIs, Kubernetes, and service meshes, it's crucial to stay up-to-date with the latest developments and insights from experts in the field. That’s why, there’s no better time to revitalize our popular podcast, Livin’ on the Edge. We’re proud to say we’re back and better than ever!

November 1, 2023 | 7 min read

Telepresence

Propagating Headers When Using GRPC in Rust

Motivation Many useful Kubernetes tools, such as OpenTelemetry and Telepresence, require header propagation to function. Today, I wanted to share with you one of my favorite tools to tackle this challenge. Let's dive into gRPC and Tonic!

October 27, 2023 | 3 min read

Kubernetes, Local Kubernetes

Quick Tips to Speed Up Your Local Kubernetes Development Process

Kubernetes is pivotal in deploying and managing containerized applications at scale, but local Kubernetes development comes with its own set of challenges. These include the complexity of creating accurate local development environments that mirror production setups, which can be time-consuming and delay testing. This tension between development speed and accuracy is crucial in local Kubernetes scenarios. For developers and teams striving to balance rapid iteration with accurate, production-like testing, local Kubernetes development proves essential. Constructing disposable development environments that reflect production setups is not only challenging but also resource-intensive. Here, local development using Kubernetes offers significant advantages. Accelerate Your Local Kubernetes Development

October 24, 2023 | 4 min read

Plugins, Docker, JetBrains, Telepresence

How to Integrate Docker & JetBrains into Telepresence

You are a developer who enjoys experimenting while striving for optimal solutions. In the past, this was straightforward because your development work occurred on your own workstation. However, you now find yourself in a situation where your applications run within a container managed by a Kubernetes cluster. To implement any changes, you must first build a container and then deploy it to the cluster to have them tested. When the container malfunctions, debugging becomes challenging; you are forced to rely on log outputs or various metrics to make educated guesses about the underlying issues.

October 20, 2023 | 11 min read
1...2021
22
2324...42