Kubernetes AI Jobs

Discover the latest remote and onsite Kubernetes AI roles across top active AI companies. Updated hourly.

Check out 301 new Kubernetes AI roles opportunities posted on AI Chopping Block

Electrical Design Engineer

New
Top rated
Armada
Full-time
Full-time
Posted

Translate business requirements into requirements for AI/ML models; prepare data to train and evaluate AI/ML/DL models; build AI/ML/DL models by applying state-of-the-art algorithms, especially transformers; leverage existing algorithms from academic or industrial research when applicable; test, evaluate, and benchmark the AI/ML/DL models, and publish the models, data sets, and evaluations; deploy models in production by containerizing the models; work with customers and internal employees to refine the quality of the models; establish continuous learning pipelines for models with online learning or transfer learning; build and deploy containerized applications on cloud or on-premise environments.

$154,560 – $193,200
Undisclosed
YEAR

(USD)

United States
Maybe global
Remote
Python
Java
C++
PyTorch
TensorFlow

GTM Engineer

New
Top rated
LangChain
Full-time
Full-time
Posted

Design and deploy production-grade agents using LangGraph and LangSmith that handle technical support queries, troubleshoot integrations, and guide users through complex onboarding flows. Analyze customer friction points to build self-service AI systems that reduce support volume and improve customer experience. Act as the product owner and technical lead to proactively identify opportunities for improvement, propose architectures, and own the full lifecycle of the systems built. Participate in the feedback loop for the product team by identifying gaps in frameworks and contributing to the LangChain and LangGraph open-source ecosystem. Develop AI-native onboarding workflows that automate documentation retrieval and code generation to help enterprise customers move from prototypes to production faster.

$160,000 – $180,000
Undisclosed
YEAR

(USD)

New York, United States
Maybe global
Remote
Python
TypeScript
LangChain
RAG
Prompt Engineering

DevSecOps Engineer (TypeScript & Agentic AI)

New
Top rated
Arize AI
Full-time
Full-time
Posted

Debug and fix issues in the platform and ship pull requests with those fixes. Build internal tools and copilots powered by generative AI to enhance the team’s capabilities. Rapidly prototype proof-of-concepts for customer use cases. Work collaboratively across Engineering, Product, and Solutions teams to unblock customers and advance AI adoption.

Undisclosed

()

Buenos Aires, Argentina
Maybe global
Remote
TypeScript
Python
Go
OpenAI API
LangChain

Machine Learning Engineer

New
Top rated
HappyRobot
Full-time
Full-time
Posted

Design, build, and maintain scalable machine learning systems including data ingestion, preprocessing, training, testing, and deployment. Develop and optimize end-to-end ML pipelines encompassing data collection, labeling, training, validation, and monitoring to ensure reliability and reproducibility. Implement robust MLOps practices such as model versioning, experiment tracking, CI/CD for machine learning, and continuous monitoring in production environments. Collaborate with product and engineering teams to integrate and deploy models into real-time products with a focus on efficiency and scalability. Ensure data quality, observability, and performance across all AI systems. Stay current with the latest AI infrastructure, tooling, and research to support ongoing innovation.

Undisclosed

()

Spain
Maybe global
Remote
Python
Go
MLOps
MLflow
Docker

Senior Product Manager, Enterprise AI Platform

New
Top rated
H Company
Full-time
Full-time
Posted

Define the vision and roadmap for the Enterprise AI platform. Understand key enterprise use cases and pain points through deep engagement with forward deployed teams, turning common pain points into high leverage features. Partner with research, engineering, and design teams to translate AI capabilities into useful product features. Own the product lifecycle from ideation through launch.

Undisclosed

()

Paris or London, United Kingdom
Maybe global
Hybrid
Python
Prompt Engineering
Model Evaluation
MLOps
MLflow

AI Deployment Engineer, Codex | Korea

New
Top rated
OpenAI
Full-time
Full-time
Posted

Serve as the primary technical subject matter expert on OpenAI Codex for a portfolio of customers, embedding deeply with them to enable their engineering teams and build coding workflows. Partner directly with customers to design and implement AI-enhanced development workflows, from rapid prototyping through scalable production rollout. Build high-quality demos, reference implementations, and workflow automations, using Codex itself as part of the development process. Lead large-format workshops, technical deep dives, and hands-on enablement sessions that help engineering organizations adopt AI coding tools effectively and safely. Contribute technical content including examples, guides, patterns, and best practices to the OpenAI Cookbook to help the broader developer community accelerate their work with Codex. Gather high-fidelity product insights from real customer deployments and translate them into clear product proposals and model feedback for internal teams. Influence customer strategy and decision-making by framing how AI coding tools fit into their software development lifecycle, technical roadmap, and organizational workflows. Serve as a trusted advisor on solution architecture, operational readiness, model configuration, security considerations, and best-practice adoption.

Undisclosed

()

Seoul, South Korea
Maybe global
Remote
Python
JavaScript
Prompt Engineering
Model Evaluation
OpenAI API

Software Engineer (SF)

New
Top rated
Fractional AI
Full-time
Full-time
Posted

Work on a small, high-caliber team building AI products for clients, from requirements gathering and prototyping through system design, development, testing, and deployment. Own features end-to-end and develop domain expertise across a range of AI use cases. Spend most of the time coding and frequently interact with clients to ensure the solutions meet their needs.

$160,000 – $220,000
Undisclosed
YEAR

(USD)

San Francisco, United States
Maybe global
Hybrid
Python
JavaScript
TypeScript
PyTorch
TensorFlow

Senior / Staff Software Engineer (SF/NY)

New
Top rated
Fractional AI
Full-time
Full-time
Posted

You will work on a small, high-caliber team building AI products for clients, setting technical direction, writing code, and serving as the go-to person when challenges arise. Spend approximately 75% of your time coding and 25% interacting with clients, including CTOs, to understand problems, evaluate tradeoffs, and ensure solutions meet their needs.

$230,000 – $350,000
Undisclosed
YEAR

(USD)

San Francisco or New York, United States
Maybe global
Hybrid
Python
JavaScript
TypeScript
PyTorch
TensorFlow

Staff Software Engineer, Bots

New
Top rated
Cantina Labs
Full-time
Full-time
Posted

As a member of the Bots team, design, build, and scale systems that enhance user engagement with the AI-powered platform, including bot chat orchestration, AI image generation, AI video generation, and tooling for managing these features. Collaborate with cross-functional teams like product managers, designers, and data specialists to deliver high-quality, performant, and maintainable features. Experiment with and integrate new AI image, video, and voice generation technologies. Build tooling and infrastructure around various AI technologies. Gain exposure to the architecture and operations of a fast-growing social AI product. Contribute expertise to evolve team processes and technical infrastructure, ensuring scalability and reliability.

$230,000 – $290,000
Undisclosed
YEAR

(USD)

San Francisco, United States
Maybe global
Onsite
Go
AWS
Docker
Kubernetes
CI/CD

Vision Foundation Model Research Intern

New
Top rated
Intrinsic
Full-time
Full-time
Posted

Contribute to the technical development and integration of advanced robotic automation solutions for manufacturing automation, utilizing the Intrinsic platform, ROS and state-of-the-art AI capabilities. Collaborate closely with research and industry partners to successfully integrate AI and automation capabilities into factory settings. Document designs, processes, and results, communicating effectively with internal technical teams and partners.

$132,000 – $187,000
Undisclosed
YEAR

(USD)

Mountain View, United States
Maybe global
Onsite
C++
Python
Kubernetes
AI
Robotics

Want to see more AI Egnineer jobs?

View all jobs

Access all 4,256 remote & onsite AI jobs.

Join our private AI community to unlock full job access, and connect with founders, hiring managers, and top AI professionals.
(Yes, it’s still free—your best contributions are the price of admission.)

Frequently Asked Questions

Need help with something? Here are our most frequently asked questions.

Question text goes here

Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare, eros dolor interdum nulla, ut commodo diam libero vitae erat. Aenean faucibus nibh et justo cursus id rutrum lorem imperdiet. Nunc ut sem vitae risus tristique posuere.

[{"question":"What are Kubernetes AI jobs?","answer":"Kubernetes AI jobs involve orchestrating containerized machine learning applications at scale. Professionals in these roles manage container deployment for AI workloads, distribute computational tasks across nodes for model training, allocate GPU resources efficiently, and automate ML pipelines. They typically work with frameworks like TensorFlow and PyTorch while ensuring high availability for production AI systems through automated scaling and self-healing capabilities."},{"question":"What roles commonly require Kubernetes skills?","answer":"Roles requiring Kubernetes skills include Machine Learning Engineers who deploy models to production, MLOps Engineers working with platforms like Kubeflow, Data Engineers managing processing pipelines, Platform Engineers supporting agentic AI applications, DevOps/SRE professionals handling containerized deployments, and Cloud Architects designing scalable environments. These positions typically involve maintaining infrastructure that supports the complete machine learning lifecycle."},{"question":"What skills are typically required alongside Kubernetes?","answer":"Alongside Kubernetes, employers typically look for container fundamentals (especially Docker), distributed systems knowledge, CI/CD pipeline experience, and cloud platform familiarity. Programming skills are essential for deployment scripts, while experience with ML frameworks like TensorFlow or PyTorch is valuable for AI-specific implementations. Understanding storage solutions, Kubernetes operators, and automated infrastructure management rounds out the typical skill requirements."},{"question":"What experience level do Kubernetes AI jobs usually require?","answer":"Kubernetes AI jobs typically require mid to senior-level experience. Employers look for professionals who understand containerization concepts, have worked with distributed systems, and can manage complex ML workflows. Prior exposure to cloud environments where Kubernetes runs is important. Candidates should demonstrate practical experience with CI/CD pipelines and familiarity with at least one major ML framework."},{"question":"What is the salary range for Kubernetes AI jobs?","answer":"Kubernetes AI jobs command competitive salaries due to the specialized intersection of container orchestration and machine learning skills. Compensation varies based on experience level, location, and specific industry. Roles requiring both strong AI expertise and Kubernetes infrastructure management typically offer premium compensation compared to general software engineering positions, reflecting the high market value of these combined skill sets."},{"question":"Are Kubernetes AI jobs in demand?","answer":"Kubernetes AI jobs are in high demand as organizations increasingly adopt containerized applications for machine learning workloads. The growth is driven by enterprises scaling their AI operations, edge computing applications, and the need for platform-agnostic infrastructure. Companies seek professionals who can manage the complexity of distributed ML systems, particularly for high-availability production environments and automated ML pipelines."},{"question":"What is the difference between Kubernetes and Docker in AI roles?","answer":"Docker creates containerized applications while Kubernetes orchestrates those containers at scale. In AI roles, Docker is used to package ML applications with their dependencies, while Kubernetes manages deployment across clusters, automates scaling during training, and handles resource allocation for GPUs. Docker provides consistency between environments, while Kubernetes adds critical production capabilities like load balancing, self-healing, and distributed computing for AI workloads."}]