Why Cast AI?
Cast AI is an automation platform that operates cloud-native and AI infrastructure at scale. By embedding autonomous decision-making directly into Kubernetes and cloud environments, Cast AI continuously optimizes performance, reliability, and efficiency in production.
The old way doesn't work. As Kubernetes and AI environments grow, manual decisions donβt. Cast AI replaces tickets, alerts, and manual tuning with continuous automation that adapts infrastructure as conditions change. Efficiency and cost savings follow naturally from that automation.
Over 2,100 companies already rely on Cast AI, including Akamai, BMW, Cisco, FICO, HuggingFace, NielsenIQ, Swisscom, and TGS.
Global team, diverse perspectives
We're headquartered in Miami, but our impact is international. We take a global and intentional approach to diversity. Today, Cast AI operates across 34 countries spanning Europe, North America, Latin America, and APAC, bringing a wide range of perspectives into how we build and lead.
Unicorn momentum
In January 2026, we achieved unicorn status with a strategic investment from Pacific Alliance Ventures, the corporate venture arm of Shinsegae Group (a $50+ billion Korean conglomerate). Our valuation now exceeds $1 billion, and we're just getting started.
Join us as we build the future of autonomous infrastructure.
This is a location-specific opportunity. We are currently accepting applications from candidates residing in the following European countries: Bulgaria, Croatia, Estonia, Greece, Hungary, Latvia, Lithuania, Poland, Romania, Slovakia, Slovenia, and Ukraine.
About the role
You'll join one of several teams building the low-level systems behind CAST AI's multi-cloud automation platform. The work sits at the intersection of Kubernetes, cloud infrastructure, and Linux.
Depending on your strengths, you'll work in one of these areas:
β’ Cluster and node lifecycle - custom node bootstrapping across AWS, GCP, and Azure. Controllers and operators managing full node lifecycle. Bridging the gap between autoscaler decisions and running VMs - including working around limitations of managed Kubernetes platforms like EKS, GKE, and AKS.
β’ Cross-cloud provisioning - orchestrating infrastructure across regions and clouds. Workflow orchestration, provisioning integrations, networking and identity across cloud boundaries.
β’ Live container migration - zero-downtime relocation of running containers using CRIU checkpoint/restore, custom CRI-O and runc forks, and forked cloud CNI plugins. Memory page transfer, TCP session migration, process-level checkpoint/restore. Kernel-level Linux engineering applied to Kubernetes.
β’ Storage optimization - dynamic volume rightsizing, BtrFS compression, custom CSI drivers. Full storage stack from cloud disk APIs through filesystem internals to Kubernetes PV/PVC lifecycle.
Common across all teams: Go, deep Kubernetes, multi-cloud, building beyond what the ecosystem offers.
How we build
We invest heavily in agentic development and AI-powered tooling. Engineers work with code agents and automated workflows daily. We expect you to shape how these tools evolve.
Requirements:
β’ Strong backend fundamentals and production-grade coding, preferably Go.
β’ Hands-on Kubernetes internals, cloud infrastructure, or large-scale distributed systems experience.
β’ Clear English communication, collaborative and self-directed.
β’ Bias toward action and ownership.
What would make you stand out
β’ Deep knowledge of EKS, GKE, or AKS internals.
β’ Low-level Linux experience: process internals, networking (NAT, iptables, conntrack, SDN), filesystems, storage.
β’ Kubernetes or cloud-native OSS contributions.
β’ Cloud storage depth - block storage, CSI, volume management, filesystem internals.
β’ Container runtime experience (CRI-O, runc, containerd) or CRIU.
β’ Experience with AI coding agents and agentic development workflows.
β’ Comfort building where there's no playbook.
Responsibilities:
β’ Design, build, and operate backend services, Kubernetes operators, and controllers in Go across AWS, GCP, and Azure.
β’ Solve infrastructure-layer problems β networking, scheduling, storage, compute, or container runtime internals.
β’ Build and maintain high-throughput services (gRPC/REST) and async orchestration workflows.
β’ Shape public and internal APIs.
β’ Leverage and improve AI-powered development workflows.
β’ Participate in design reviews, code reviews, on-call, and customer deep-dives.
Whatβs in it for you?
β’ Competitive salary (β¬6,500 - β¬9,000 gross, depending on the level of experience)
β’ Enjoy a flexible, remote-first global environment.
β’ Collaborate with a global team of cloud experts and innovators, passionate about pushing the boundaries of Kubernetes technology.
β’ Equity options.
β’ Get quick feedback with a fast-paced workflow. Most feature projects are completed in 1 to 4 weeks.
β’ Spend 10% of your work time on personal projects or self-improvement.
β’ Learning budget for professional and personal development - including access to international conferences and courses that elevate your skills.
β’ Annual hackathon to spark new ideas and strengthen team bonds.
β’ Team-building budget and company events to connect with your colleagues.
β’ Equipment budget to ensure you have everything you need.
β’ Extra days off to help maintain a healthy work-life balance.
Hiring process
β’ Screening call with Recruiter
β’ Hiring Manager interview
β’ Technical interview (system design)
β’ Live coding
β’ Culture Check interview with an executive
*As part of our standard hiring process, we would like to inform you that a background check may be conducted at the final stage of recruitment through our third-party provider, Checkr.
*Please note that Cast AI does not provide any form of visa sponsorship/work permit.
#LI-Remote