AWS AI Jobs

Discover the latest remote and onsite AWS AI roles across top active AI companies. Updated hourly.

Check out 352 new AWS AI roles opportunities posted on AI Chopping Block

Staff Software Engineer, AI Voice Agent

New
Top rated
Aircall
Full-time
Full-time
Posted

As a Software Engineer on the AI Voice Agent team, you will work on real-time speech pipeline systems including live audio buffering, streaming, latency optimization, and integrating with speech providers. You will build and improve conversation intelligence systems that manage the LLM layer for natural conversation flow, including prompt construction, context management, function calling, and dialogue management. You will develop the action framework that allows the AI Voice Agent to execute tasks during calls such as querying account data, creating tickets, and checking order status, handling API configuration, success/failure branching, authentication management, and runtime execution. Additionally, you will work on knowledge ingestion, storage, and retrieval for the voice agent and manage memory for retaining information across conversations to improve responses. You will collaborate with designers to create easy-to-use interfaces for agent lifecycle management including creation, configuration, testing, and deployment. You will contribute to building evaluation frameworks and metrics for voice AI quality, post-call analytics, and instrumentation, as well as participate in the on-call rotation.

$215,000 – $250,000
Undisclosed
YEAR

(USD)

San Francisco, United States
Maybe global
Onsite
Python
TypeScript
Prompt Engineering
Model Evaluation
LLM

Software Engineer, AI Voice Agent

New
Top rated
Aircall
Full-time
Full-time
Posted

As a Software Engineer on the AI Voice Agent team, you will work on real-time systems involving live audio such as buffering, streaming, and latency optimization, along with integrating speech providers. You will build and improve conversation intelligence systems, including prompt construction, context management, function calling, and dialogue management to make conversations feel natural. You will develop the action framework to execute configurable API calls, manage success/failure branching, authentication, and runtime execution during calls. You will work on knowledge ingestion, storage, retrieval, memory, and context for the voice agent to improve its performance over time. Additionally, you will collaborate on agent lifecycle tasks such as creation, configuration, testing, and deployment of voice agents and help build evaluation frameworks for model performance, call quality metrics, and call analytics. Participation in on-call rotations is also expected.

$130,000 – $220,000
Undisclosed
YEAR

(USD)

San Francisco, United States
Maybe global
Onsite
Python
TypeScript
Prompt Engineering
Model Evaluation
MLOps

Senior Software Engineer, AI Voice Agent

New
Top rated
Aircall
Full-time
Posted

As a Senior Software Engineer on the AI Voice Agent team, you will work on real-time systems involving live audio streaming and latency optimization integrated with speech providers. You will build and improve conversation intelligence systems that manage LLM layers, including prompt construction, context management, function calling, and dialogue management to create natural, actionable phone conversations. You will develop the action framework allowing configurable API calls with branching logic and runtime execution, supporting tasks like data lookup and ticket creation during calls. You'll manage knowledge ingestion, storage, and retrieval to enhance agent memory and learning over time. You will collaborate with designers to enable customers to create, configure, test, and deploy voice agents through intuitive product experiences. Additionally, you will help develop evaluation frameworks, analytics, call quality metrics, and monitoring instrumentation, and participate in on-call rotation duties.

$150,000 – $220,000
Undisclosed
YEAR

(USD)

Maybe global
Python
TypeScript
Prompt Engineering
Model Evaluation
AWS

Staff Software Engineer, Foundations (Managed AI)

New
Top rated
Crusoe
Full-time
Full-time
Posted

As a Staff Software Engineer in the Foundations department, responsibilities include leading the design and implementation of highly scalable systems for the Managed AI offerings, driving the long-term technical roadmap for the Foundations team to support growth and evolving AI workloads, working cross-functionally with Cloud Engineering to align technical goals and solve integration challenges, leading by example through high-quality code contributions and mentoring Senior and Staff-level engineers, championing reliability, observability, and performance by identifying and resolving systemic bottlenecks, and staying current with AI infrastructure trends to ensure efficient and powerful tools are utilized.

$208,000 – $253,000
Undisclosed
YEAR

(USD)

San Francisco, United States
Maybe global
Onsite
Go
Python
Kubernetes
AWS
GCP

Senior Platform/DevOps Engineer (Kubernetes-Linux)

New
Top rated
Armada
Full-time
Full-time
Posted

Translate business requirements into requirements for AI/ML models; prepare data to train and evaluate AI/ML/DL models; build AI/ML/DL models by applying state-of-the-art algorithms, especially transformers; leverage existing algorithms from academic or industrial research when applicable; test, evaluate, and benchmark AI/ML/DL models and publish the models, data sets, and evaluations; deploy models in production by containerizing them; work with customers and internal employees to refine model quality; establish continuous learning pipelines for models using online or transfer learning; build and deploy containerized applications on cloud or on-premise environments.

$154,560 – $193,200
Undisclosed
YEAR

(USD)

Bellevue, United States
Maybe global
Onsite
Python
Java
C++
Docker
Kubernetes

AI Engineer

New
Top rated
AppZen
Full-time
Full-time
Posted

The AI Engineer will design and develop intelligent agents powered by large language models (LLMs) using tool calling, orchestration frameworks, and advanced context management to enable reasoning, planning, and autonomous decision-making across complex workflows. Responsibilities include working hands-on with modern agentic stacks such as LangGraph and Autogen, implementing asynchronous and streaming architectures, and ensuring production-grade observability to build scalable real-world AI systems.

$160,000 – $180,000
Undisclosed
YEAR

(USD)

San Jose, United States
Maybe global
Onsite
Python
Go
LangChain
MLOps
Kubernetes

Tech Lead Manager, Data Infrastructure

New
Top rated
Cartesia
Full-time
Full-time
Posted

The Tech Lead Manager, Data Infrastructure at Cartesia is responsible for defining the overall multi-modal data strategy across pre-training and post-training, including human, synthetic, and web-scale data sources. They lead, manage, and mentor a team of data engineers and specialists. They design and oversee the construction of robust, scalable data pipelines for text, audio, and video, establish and enforce rigorous standards for data quality across the organization, deeply understand how data affects model capability and proactively identify and source novel datasets, and manage relationships and budgets with external data vendors and partners.

$250,000 – $375,000
Undisclosed
YEAR

(USD)

San Francisco, United States
Maybe global
Onsite
Python
Data Pipelines
MLOps
AWS
GCP

Forward Deployed AI Engineer

New
Top rated
Talent Labs
Full-time
Full-time
Posted

Drive the end-to-end technical deployment of Latent Labs models into customer environments, ensuring seamless integration with existing scientific and IT infrastructure. Design and build production-grade API integrations, data pipelines and model-serving infrastructure tailored to each customer’s requirements. Work on-site or embedded with pharma and biotech partners to scope technical requirements, troubleshoot issues and deliver solutions. Ensure deployments meet enterprise standards for security, performance and reliability. Serve as the technical point of contact for assigned customers, building trusted relationships with their scientific and engineering teams, including spending time working on-site at international partner locations as needed. Gather and synthesise customer feedback, translating it into actionable insights for product, research and platform teams. Collaborate with internal teams to shape the product roadmap based on real-world deployment learnings. Create technical documentation, integration guides and best-practice resources for customers. Stay on top of the latest developments in ML infrastructure, model serving and cloud-native tooling. Gain a strong working understanding of protein and cell biology as it relates to the product. Participate in knowledge sharing, including organizing and presenting at internal reading groups.

Undisclosed

()

San Francisco, United States
Maybe global
Hybrid
Python
AWS
Docker
Kubernetes
CI/CD

Director, Data Center Operations

New
Top rated
Together AI
Full-time
Full-time
Posted

The responsibilities include advancing inference efficiency end-to-end by designing and prototyping algorithms, architectures, and scheduling strategies for low-latency, high-throughput inference. Implementing and maintaining changes in high-performance inference engines, including kernel backends and speculative decoding, profiling and optimizing performance across GPU, networking, and memory layers to improve latency, throughput, and cost. Unifying inference with RL/post-training by designing and operating RL and post-training pipelines, making RL and post-training workloads more efficient with inference-aware training loops, and using these pipelines to train, evaluate, and iterate on frontier models. Co-designing algorithms and infrastructure so that objectives, rollout collection, and evaluation are tightly coupled to efficient inference, identifying bottlenecks across the training engine, inference engine, data pipeline, and user-facing layers. Running ablations and scale-up experiments to understand trade-offs between model quality, latency, throughput, and cost, and feeding these insights back into model, RL, and system design. Owning critical systems at production scale by profiling, debugging, and optimizing inference and post-training services under real production workloads, driving roadmap items requiring engine modification, and establishing metrics, benchmarks, and experimentation frameworks to validate improvements rigorously. Providing technical leadership by setting technical direction for cross-team efforts at the intersection of inference, RL, and post-training, and mentoring other engineers and researchers on full-stack ML systems work and performance engineering.

$200,000 – $280,000
Undisclosed
YEAR

(USD)

San Francisco
Maybe global
Onsite
Python
PyTorch
TensorFlow
MLOps
Model Evaluation

Regional Sales Lead, Singapore

New
Top rated
Tenstorrent
Full-time
Full-time
Posted

Lead and contribute to cross-functional efforts solving complex physical design challenges across IPs, projects, and advanced technology nodes. Develop and enhance RTL-to-GDS methodologies, including floorplanning, synthesis, placement and routing (P&R), static timing analysis (STA), signoff, and assembly. Architect and deploy AI/ML-driven solutions in production flows to improve engineering efficiency, turnaround time, and quality of results (QoR). Optimize EDA tools and custom CAD flows using data-driven and machine learning-based techniques, working closely with internal teams such as verification, extraction, timing, Design for Test (DFT), and electronic design automation (EDA) vendors.

$100,000 – $500,000
Undisclosed
YEAR

(USD)

Santa Clara or Austin or Fort Collins, United States
Maybe global
Remote
Python
PyTorch
TensorFlow
MLOps
Docker

Want to see more AI Egnineer jobs?

View all jobs

Access all 4,256 remote & onsite AI jobs.

Join our private AI community to unlock full job access, and connect with founders, hiring managers, and top AI professionals.
(Yes, it’s still free—your best contributions are the price of admission.)

Frequently Asked Questions

Need help with something? Here are our most frequently asked questions.

Question text goes here

Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare, eros dolor interdum nulla, ut commodo diam libero vitae erat. Aenean faucibus nibh et justo cursus id rutrum lorem imperdiet. Nunc ut sem vitae risus tristique posuere.

[{"question":"What are AWS AI jobs?","answer":"AWS AI jobs involve building, training, and deploying generative AI applications using specialized cloud services. These roles work with tools like SageMaker for custom model development, Bedrock for foundation models, and Lake Formation for data governance. Professionals in these positions create AI-driven applications, implement RAG systems with Kendra, and orchestrate machine learning pipelines using Step Functions and Lambda."},{"question":"What roles commonly require AWS skills?","answer":"Common roles requiring AWS skills include machine learning engineers, data scientists, software engineers, architects, and platform engineers. These professionals work on generative AI applications and AI-assisted development lifecycles. They implement end-to-end ML pipelines in SageMaker, design LLM-powered applications with Bedrock, create agentic workflows, and build AI-enhanced developer tools using Amazon Q Developer."},{"question":"What skills are typically required alongside AWS?","answer":"Alongside AWS expertise, professionals typically need experience with JupyterLab, Git, and IDE integrations like VS Code. Knowledge of LangChain for LLM orchestration, machine learning concepts, and data engineering practices are valuable. Familiarity with generative AI patterns like retrieval-augmented generation, prompt engineering, and AI application development workflows helps create effective solutions within the AWS ecosystem."},{"question":"What experience level do AWS AI jobs usually require?","answer":"AWS AI jobs typically require mid to senior-level experience with cloud infrastructure and AI development patterns. Employers look for professionals familiar with JupyterLab environments, ML workflows in SageMaker, and foundation model deployment via Bedrock. Experience building end-to-end machine learning pipelines, implementing RAG systems, and orchestrating AI workflows using Step Functions and Lambda is highly valued."},{"question":"What is the salary range for AWS AI jobs?","answer":"AWS AI job salaries vary based on experience, location, and specific role. Machine learning engineers and data scientists implementing SageMaker solutions generally command premium compensation. Platform engineers orchestrating AI infrastructure and architects designing generative AI applications often receive higher salaries. Software engineers using Amazon Q for AI-assisted development are increasingly valued for their productivity enhancements."},{"question":"Are AWS AI jobs in demand?","answer":"AWS AI jobs are experiencing strong demand as organizations adopt generative AI technologies. Companies are actively hiring professionals who can implement AI-driven development lifecycles using tools like Amazon Q Developer. There's particular demand for engineers who can work with Bedrock for foundation models, build RAG systems with Kendra, and design agentic workflows for business process automation."},{"question":"What is the difference between AWS and Azure in AI roles?","answer":"The key difference in AI roles is that AWS emphasizes fully managed services like Bedrock for foundation models and SageMaker for end-to-end ML workflows, while Azure offers a different ecosystem through Azure AI services. AWS positions focus more on serverless orchestration and agentic capabilities unique to their toolchain. The platforms have distinct approaches to generative AI implementation, with different service integrations and developer experiences."}]