The AI job market moves fast. We keep up so you don't have to.
Fresh roles added daily, reviewed for quality — across every corner of the AI ecosystem.
I'm strong in:
Edit filters
New AI Opportunities
Showing 61 – 79 of 79 jobs
Tag
Chief Technology Officer
Bjak
201-500
China
Full-time
Remote
false
About the RoleA1 is building a proactive AI system that carries work forward across conversations, tools, and time — enabling users to delegate ongoing tasks to AI agents that coordinate across software, data, and workflows.We are looking for a leader who can think clearly about systems, make strong technical decisions, and help build the engineering organisation from the ground up.Be part of founding team to shape the technical direction of the company, while helping build a strong engineering team across the globe.What You'll DoTechnical DirectionDefine the long-term architecture for A1’s AI systems, infrastructure, and developer platformEvaluate trade-offs between speed of iteration and long-term system designEnsure systems are designed for scalability, reliability, and long-term evolutionGuide key decisions across model integration, data pipelines, distributed systems, and product architectureEngineering LeadershipWork with engineers to translate product direction into clear technical executionHelp structure engineering workstreams and keep teams aligned on prioritiesMaintain high engineering standards while keeping the team focused on shippingEstablish engineering culture, development practices, and technical standards across the companyBuilding the TeamBuild and scale a world-class engineering team across key talent hubs including China and USIdentify strong technical leaders and help build a high-quality engineering organizationDefine hiring standards and interview processes to maintain a high engineering barCoordination and ExecutionWork closely with product, research, and leadership teamsEnsure technical workstreams move forward smoothly across teams and locationsHelp resolve cross-team technical and execution challengesWhat You Will NeedStrong technical foundation in system architecture, large-scale systems, distributed architecture.Ability to reason clearly about complex systems and make pragmatic technical decisionsExperience building or leading high-performing engineering teamsStrong judgment on technical trade-offs and engineering prioritiesComfortable operating in early-stage environments with high ambiguityClear communication and ability to align teamsWe are particularly interested in candidates who enjoy building teams, superior products and shaping engineering organisations.How We WorkWe operate as a small, senior, hands-on team. Engineers own features end-to-end — from design discussion through production monitoring.Code reviews and design reviews are expected for all meaningful changes. We discuss architecture openly, make decisions quickly, and ship frequently.Interview processIf there appears to be a fit, we'll reach out to schedule 3, but no more than 4 interviews.Applications are evaluated by our technical team members. Interviews will be conducted via virtual meetings and/or onsite.We value transparency and efficiency, so expect a prompt decision. If you've demonstrated the exceptional skills and mindset we're looking for, the process to offer may be shorter.
No items found.
2026-03-18 11:01
Intern, Software Engineer - Perception
Haydenai
101-200
$45 – $45 / hour
United States
Intern
Remote
false
About UsAt Hayden AI, we are on a mission to harness the power of computer vision to transform the way transit systems and other government agencies address real-world challenges.From bus lane and bus stop enforcement to transportation optimization technologies and beyond, our innovative mobile perception system empowers our clients to accelerate transit, enhance street safety, and drive toward a sustainable future.About the Perception TeamHayden AI's Perception Team develops and refines AI and algorithms that power the company's mobile enforcement systems, processing video and image data to detect traffic violations, parking infractions, and other urban compliance scenarios. The team works cross-functionally with cloud and device engineering to build and deploy perception pipelines that make decisions about the real world and create insights for our customers.About the RoleAs a Perception Engineering Intern at Hayden AI, you will contribute directly to the AI systems that power our mobile enforcement platform. You will work on computer vision and machine learning models that process real-world video and image data to detect traffic violations and urban compliance events.This role sits at the intersection of machine learning, data, and real-world deployment. You may work on improving model accuracy, building or refining training datasets, evaluating edge cases, optimizing inference performance, or developing tooling to support perception pipelines across cloud and on-device environments.You will not be working on side projects—you will own meaningful technical work that contributes to production systems. With guidance from a senior engineer, you’ll design experiments, implement improvements, test performance, and help ship enhancements that directly impact how our systems interpret and act on the physical world.This internship is ideal for someone excited about applied AI—especially computer vision, model evaluation, and deploying ML systems beyond the lab into real-world environments.This position is based in San Francisco and follows a hybrid schedule with at least 3 days in-office per week.Key ResponsibilitiesBelow are your primary responsibilities. These represent the core areas where you’ll make an impact. As part of a rapidly evolving team, we look forward to your impact expanding over time.Take ownership of a real project and see it through to completionBuild and ship features with support from senior engineersWrite clean, scalable codeTest your work and iterate quicklyBe involved in everything from design discussions to deploymentCollaborate with engineers in code reviews and team discussionsParticipate in standups, sprint planning, and retrospectivesSupport the team on ad hoc engineering tasks as they come upHelp improve performance, reliability, or usability where neededAsk questions, seek feedback, and apply it quicklyDeliverables or project examples:GPS data analysisTrain Deep learning modelCreate AI datasetsLidar/Camera data toolingTest cases for end-to-end system performanceDevelop a cloud service in the event processing pipelineAdd page or a new user flow to the Portal web applicationRequired QualificationsThe qualifications below outline the experience and skills most relevant to success in this role. We recognize that skills and potential come in many forms, and we welcome diverse experiences that advance our mission.Education:Currently enrolled in a Master’s or PhD program in Computer Science, Mechanical Engineering, or a related technical field.
Technical Experience:Experience in one or more of the following programming languages:Python, C++
Personal Attributes:Detail-oriented with a high bar for quality and accuracy.Curious and self-driven, motivated to dig into problems and find root causes.Strong communicator who can clearly document findings and surface issues to the right stakeholders.Collaborative team player who thrives in cross-functional environments.Organized and reliable, with the ability to manage multiple tasks and follow through consistently.Comfortable with ambiguity and able to make progress with limited direction.
No items found.
2026-03-18 11:01
Lead Field Marketing & Events Manager
Snorkel AI
501-1000
$172,000 – $300,000
United States
Full-time
Remote
false
About Snorkel
At Snorkel, we believe meaningful AI doesn’t start with the model, it starts with the data.
We’re on a mission to help enterprises transform expert knowledge into specialized AI at scale. The AI landscape has gone through incredible changes between 2015, when Snorkel started as a research project in the Stanford AI Lab, to the generative AI breakthroughs of today. But one thing has remained constant: the data you use to build AI is the key to achieving differentiation, high performance, and production-ready systems. We work with some of the world’s largest organizations to empower scientists, engineers, financial experts, product creators, journalists, and more to build custom AI with their data faster than ever before. Excited to help us redefine how AI is built? Apply to be the newest Snorkeler!About the Role
Snorkel AI is hiring Frontier AI Solutions Engineers who will partner with leading AI labs on their most challenging data problems. This is a high-impact, customer-facing role that combines technical depth with strong presales instincts. You'll partner with customer research teams to design complex data and environments that improve frontier model performance, demonstrating Snorkel's capabilities through research-driven engagements.
You'll work at the critical intersection of research, technical strategy, and customer partnership. This includes scoping training data needs, designing RL environments and tasks, developing evaluation frameworks, probing model behavior and failure modes, and translating customer research objectives into actionable technical plans. You'll develop technical specifications, analyze frontier model failure modes, and serve as a thought partner to customer research teams throughout the sales cycle and into early delivery phases.
Main Responsibilities
Partner with frontier AI research labs to design datasets and environments that improve model performance
Lead technical conversations with customer researchers to understand model capabilities, failure modes, data requirements, and success criteria
Probe model behavior through systematic evaluation to uncover weaknesses and identify high-impact data interventions
Design evaluation frameworks, calibration processes, and quality rubrics that establish measurable project success metrics
Develop technical specifications for data projects that balance research rigor with operational feasibility
Serve as thought partner to customer research teams throughout the sales cycle, building trust and credibility
Stay current on frontier AI research, RL environment design, post-training techniques, and evaluation methodologies
Preferred Qualifications
Strong expertise in frontier AI concepts including LLMs, training data pipelines, evaluation methodologies, post-training techniques (RLHF, DPO, RLAIF), and domain areas such as coding agents, reasoning, multimodal models, or RL environments
Experience in applied ML research, data science, or research-intensive technical roles with customer-facing or collaborative research experience
Proficiency in Python and familiarity with ML frameworks and LLM APIs
Excellent communication skills — ability to deliver technical presentations and explain complex concepts to diverse audiences
Familiarity with data curation workflows, synthetic data generation, LLM-as-a-Judge, or evaluation framework design
Ability to work in a fast-moving environment, comfortable with ambiguity and rapid iteration
B.S. in Computer Science, Machine Learning, or related field with 4+ years of experience in AI/ML solutions engineering or technical customer-facing roles
Compensation range for Tier 1 locations of San Francisco Bay Area and New York City, $172K - $300K OTE. All offers also include equity in the form of employee stock options. Our compensation ranges are determined by role, level, and location. Within the range, individual pay is determined by work location and additional factors, including job-related skills, experience, and relevant education or training.
Why Join Snorkel AI?
At Snorkel AI, we're building the future of data-centric AI. Our Expert Data-as-a-Service organization partners with world-class customers to solve some of the hardest data challenges — creating training and evaluation data that power the next generation of LLMs and AI systems. You'll work directly on projects that impact real production systems, while shaping how internal teams deliver faster, better, and more intelligently. This is a rare opportunity to own technical data workflows and be a founding member of the technical DaaS team.
#LI-CG1
Salary Range
-
Salary Range $172,000—$300,000 USDBe Your Best at Snorkel
Joining Snorkel AI means becoming part of a company that has market proven solutions, robust funding, and is scaling rapidly—offering a unique combination of stability and the excitement of high growth. As a member of our team, you’ll have meaningful opportunities to shape priorities and initiatives, influence key strategic decisions, and directly impact our ongoing success. Whether you’re looking to deepen your technical expertise, explore leadership opportunities, or learn new skills across multiple functions, you’re fully supported in building your career in an environment designed for growth, learning, and shared success.
Snorkel AI is proud to be an Equal Employment Opportunity employer and is committed to building a team that represents a variety of backgrounds, perspectives, and skills. Snorkel AI embraces diversity and provides equal employment opportunities to all employees and applicants for employment. Snorkel AI prohibits discrimination and harassment of any type on the basis of race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state, or local law. All employment is decided on the basis of qualifications, performance, merit, and business need.
We will ensure that individuals with disabilities are provided reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. Please contact us to request accommodation.
No items found.
2026-03-18 11:01
Software Engineer, AI Platform
Harvey
501-1000
CA$154,000 – CA$264,000
Canada
Full-time
Remote
false
Why HarveyAt Harvey, we’re transforming how legal and professional services operate — not incrementally, but end-to-end. By combining frontier agentic AI, an enterprise-grade platform, and deep domain expertise, we’re reshaping how critical knowledge work gets done for decades to come.This is a rare chance to help build a generational company at a true inflection point. With 1000+ customers in 58+ countries, strong product-market fit, and world-class investor support, we’re scaling fast and defining a new category in real time. The work is ambitious, the bar is high, and the opportunity for growth — personal, professional, and financial — is unmatched.Our team is sharp, motivated, and deeply committed to the mission. We move fast, operate with intensity, and take real ownership of the problems we tackle — from early thinking to long-term outcomes. We stay close to our customers — from leadership to engineers — and work together to solve real problems with urgency and care. If you thrive in ambiguity, push for excellence, and want to help shape the future of work alongside others who raise the bar, we invite you to build with us.At Harvey, the future of professional services is being written today — and we’re just getting started.Role OverviewHarvey’s products all depend on a shared AI foundation: the model layer and agent infrastructure that determine the quality of work our agents deliver. Legal is one of the hardest domains for AI: documents run to hundreds of pages, matters can span millions of files, and there is zero margin for error on accuracy.The AI Platform team builds the foundation that every product and agent team at Harvey builds upon. This team is early and there’s a lot to build: model routing, agent architecture, context management, evals. Your work here sets the ceiling for what Harvey’s AI can do.Representative ProjectsContext Engineering & Agent Infrastructure. Build the platform-level systems for context management, session state, and memory that all of Harvey’s agents and products rely on.Model Integration & Routing. Own the infrastructure that lets Harvey onboard new foundation models fast and route to the right one for every task - a capability every product team depends on.Evaluation Infrastructure. Build the shared eval tooling and frameworks that let every team across Harvey measure and improve AI quality systematically.Shared Abstractions. Create the SDKs, platform primitives, and developer tooling that make it dramatically easier for product teams to ship AI-powered features.What You’ll DoDesign and build abstractions and platform-level systems that improve all of Harvey’s agentic products.Own infrastructure for model integration, routing, and evaluation that helps Harvey choose and deploy the right foundation model for any given context.Build evaluation frameworks and tooling that let every team across Harvey iterate on AI quality effectively.Partner closely with product engineering teams, PMs, and design to launch cutting-edge AI products.Evaluate, prototype, and integrate the latest advancements in AI and agentic systems as they emerge.What You Have5+ years of experience building backend systems, with at least 1+ year focused on AI/ML engineering. Staff candidates will typically have 8+ years and a track record of technical leadership across teams.Experience building and shipping multi-model or multi-provider AI systems in production.Familiarity with context management, session state, or memory systems in AI or distributed systems. You’ve thought about what the model sees and why it matters.A track record of building internal platforms, SDKs, or shared infrastructure that other engineering teams actually adopted - and an understanding of why developer experience matters as much as raw capability.Strong judgment about abstractions. Opinionated about good design but pragmatic about shipping incrementally.Excitement about agentic AI and the infrastructure challenges of making autonomous systems reliable when the stakes are real.A bias toward full ownership: you navigate ambiguity well and don’t wait for a roadmap to start solving problems.Bonus: experience building evaluation frameworks, working with agent/function-calling architectures, familiarity with legal or other high-stakes professional services domains, or time at early-stage or hyper-growth startups where the underlying technology changes regularly.Compensation$154,000 - $264,000 CADHarvey is an equal opportunity employer and does not discriminate on the basis of race, gender, sexual orientation, gender identity/expression, national origin, disability, age, genetic information, veteran status, marital status, pregnancy or related condition, or any other basis protected by law.We are committed to providing reasonable accommodations to applicants with disabilities, and requests can be made by emailing accommodations@harvey.ai
No items found.
2026-03-18 10:47
Applied AI Research Engineer – ML Systems & Structured Data
Granica
11-50
$160,000 – $250,000
United States
Full-time
Remote
false
Applied AI Research Engineer – ML Systems & Structured DataLocation: Bay Area (Mountain View)
Employment Type: Full-time
Work Model: On-site
Department: Research
Compensation: $160K – $250K + EquityOverviewGranica is building the next generation of efficient AI infrastructure.Today’s AI systems are limited not only by model design but by the inefficiency of the data that feeds them. At enterprise scale, redundant data, inefficient representations, and poorly optimized learning pipelines create enormous cost and latency.Granica’s mission is to eliminate that inefficiency.We combine advances in information theory, machine learning, and distributed systems to design data infrastructure that continuously improves how information is represented and used by AI.Granica’s research effort is led by Prof. Andrea Montanari (Stanford) and focuses on building learning systems that operate efficiently on large-scale structured and tabular data.While much of the industry focuses on text or media models, Granica is building the foundations of AI systems that learn directly from structured enterprise data.This role focuses on building machine learning systems for structured and tabular data rather than general LLM application development.The RoleThe Applied AI Research Team sits at the intersection of theory and production.Your work will take ideas emerging from fundamental research and turn them into practical algorithms, optimized pipelines, and production-ready ML systems that operate across petabytes of structured enterprise data.This is a high-ownership role for engineers who can think like researchers and build like systems engineers.You will translate theory into measurable performance improvements and help define the engineering foundations of structured AI.What You’ll DoTurn research into working systemsTransform foundational ideas from Granica Research and Prof. Andrea Montanari’s group into scalable algorithms and prototypesBuild evaluation harnesses, datasets, and benchmarks that measure real signal from research ideasDefine and improve metrics that quantify progress in structured AI systemsInvent and optimize algorithmsDevelop efficient learning methods for relational, tabular, graph, and enterprise datasetsPrototype representation learning architectures and compression-aware modelsExplore new approaches for learning from heterogeneous structured dataBuild high-performance ML pipelinesImplement fast training and inference pipelines using PyTorch, JAX, or custom kernelsOptimize memory usage, compute utilization, and data movementImprove cost, latency, and throughput for large-scale ML workloadsBuild hybrid AI systemsDesign systems integrating symbolic, relational, and neural componentsEnable AI models to reason over structured datasets without relying on text intermediariesCollaborate across research and engineeringWork with Research Scientists to validate hypotheses at scaleWork with Systems Engineers to integrate algorithms into Granica’s data platformWork with Product Engineering to ship features powering real enterprise workloadsIterate fast and measure everythingRun controlled experiments and analyze performance improvementsDeliver results with clear benchmarks and reproducible evaluationsDrive the cycle from prototype → production → optimizationWhat You’ll BringTechnical DepthStrong background in machine learning, probabilistic modeling, optimization, or large-scale ML systemsExperience building algorithms for structured, relational, tabular, or graph dataAbility to reason from first principles about scaling behavior, efficiency, and information flowSystems EngineeringHands-on experience with PyTorch, JAX, TensorFlow, or similar ML frameworksStrong programming skills in PythonExperience with systems languages such as Rust, C++, or CUDA is a plusExperience building large-scale ML pipelines, evaluation frameworks, or distributed systemsApplied MindsetProven ability to turn research ideas into performant, reliable codeComfort working in research-driven environments with ambiguous problem definitionsStrong experimentation discipline and focus on measurable performance improvementsBonus ExperienceStructured representation learning, tabular ML, relational learning, or graph MLExperience with large-scale training infrastructure or distributed MLFamiliarity with data systems, query engines, or large-scale data pipelinesExperience building evaluation infrastructure for ML systemsOpen-source contributions or collaborative work bridging research and production systemsWhy This Role MattersThe world’s most valuable data is structured.Most AI systems today are not built to learn from it efficiently.Granica is building the systems that close this gap.Your work will help define the engineering foundations of structured AI — designing the algorithms, pipelines, and infrastructure that enable efficient learning from enterprise data at global scale.This role offers:high ownershipreal research impactimmediate production relevanceand the opportunity to shape a new generation of AI systems.Compensation & BenefitsCompetitive salary, meaningful equity, and substantial bonus for top performersFlexible time off plus comprehensive health coverage for you and your familySupport for research, publication, and deep technical explorationAt Granica, you will shape the fundamental infrastructure that makes intelligence itself efficient, structured, and enduring. Join us to build the foundational data systems that power the future of enterprise AI!
No items found.
2026-03-18 10:47
Forward Deployed Engineer, Agentic Platform (Public Sector)
Cohere
501-1000
Canada
Full-time
Remote
false
Who are we?Our mission is to scale intelligence to serve humanity. We’re training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, and agents. We believe that our work is instrumental to the widespread adoption of AI.We obsess over what we build. Each one of us is responsible for contributing to increasing the capabilities of our models and the value they drive for our customers. We like to work hard and move fast to do what’s best for our customers.Cohere is a team of researchers, engineers, designers, and more, who are passionate about their craft. Each person is one of the best in the world at what they do. We believe that a diverse range of perspectives is a requirement for building great products.Join us on our mission and shape the future!About North:North is Cohere's cutting-edge AI workspace platform, designed to revolutionize the way enterprises utilize AI. It offers a secure and customizable environment, allowing companies to deploy AI while maintaining control over sensitive data. North integrates seamlessly with existing workflows, providing a trusted platform that connects AI agents with workplace tools and applications.Why This Role?Cohere’s team partners with Canadian public sector organisations to unlock transformative value through secure, ethical deployment of Generative AI (GenAI) solutions. We work collaboratively to address complex societal challenges while maintaining the highest standards of data security and compliance. You will work directly with public sector customers to quickly understand their greatest problems and design and implement solutions using Cohere's stack.This role offers a unique opportunity to shape how enterprises harness the power of AI in real-world applications. As a bridge between our core North product and our clients’ engineering teams, you’ll be at the forefront of solving complex problems and securely integrating AI into critical sectors.We are seeking engineers with diverse skill sets, including backend, infrastructure, agent development, and deployments, who deeply care about customers and want to work at the cutting edge of Agentic AI.Location: Ottawa or Toronto required (proximity to government customers), 20-40% travel anticipated
Security Clearance: Active Top Secret clearance strongly preferred; candidates eligible and willing to obtain clearance will also be consideredIn this role, you will:Build and ship features for North, our AI workspace platformDevelop autonomous agents that talk to sensitive enterprise dataExperiment at a high velocity and with a high level of quality to engage our customers and ultimately deliver solutions that exceed their expectationsWork across the entire product lifecycle from conceptualization through productionLead end-to-end deployment of North in private cloud and on-premises environments, including planning, configuration, testing, and rolloutYou may be a good fit if:You have experience with and enjoy working directly with customersYou have shipped (lots of) Python in productionYou have built and deployed highly performant client-side or server-side RAG/agentic applicationsYou have strong coding abilities and are comfortable working across the stack. You’re able to read and understand, and even fix issues outside of the main code baseYou excel in fast-paced environments and can execute while priorities and objectives are a moving targetIf some of the above doesn’t line up perfectly with your experience, we still encourage you to apply! We value and celebrate diversity and strive to create an inclusive work environment for all. We welcome applicants from all backgrounds and are committed to providing equal opportunities. Should you require any accommodations during the recruitment process, please submit an Accommodations Request Form, and we will work together to meet your needs.Full-Time Employees at Cohere enjoy these Perks:🤝 An open and inclusive culture and work environment 🧑💻 Work closely with a team on the cutting edge of AI research 🍽 Weekly lunch stipend, in-office lunches & snacks🦷 Full health and dental benefits, including a separate budget to take care of your mental health 🐣 100% Parental Leave top-up for up to 6 months🎨 Personal enrichment benefits towards arts and culture, fitness and well-being, quality time, and workspace improvement🏙 Remote-flexible, offices in Toronto, New York, San Francisco, London and Paris, as well as a co-working stipend✈️ 6 weeks of vacation (30 working days!)
No items found.
2026-03-18 10:46
Senior Data Scientist
Faculty
501-1000
United Kingdom
Full-time
Remote
false
Why Faculty?
We established Faculty in 2014 because we thought that AI would be the most important technology of our time. Since then, we’ve worked with over 350 global customers to transform their performance through human-centric AI. You can read about our real-world impact here.We don’t chase hype cycles. We innovate, build and deploy responsible AI which moves the needle - and we know a thing or two about doing it well. We bring an unparalleled depth of technical, product and delivery expertise to our clients who span government, finance, retail, energy, life sciences and defence.Our business, and reputation, is growing fast and we’re always on the lookout for individuals who share our intellectual curiosity and desire to build a positive legacy through technology.AI is an epoch-defining technology, join a company where you’ll be empowered to envision its most powerful applications, and to make them happen.About the team Our Public Services Business Unit is committed to leveraging AI for the benefit of individual citizens and the public good.
From our work informing strategic government decisions, to optimising our NHS, through to reducing bureaucratic backlogs - we know that AI offers opportunities to drive improvements at every level of Government and we are proud to lead on some of the most impactful work happening in the sector.
Because of the nature of the work we do with our Government clients, you may need to be eligible for UK Security Clearance (SC) and willing to work on site with these clients from time to time.About the roleAs a Senior Data Scientist, you will lead high-impact AI projects and shape the technical direction of bespoke solutions. This role requires hands-on technical excellence combined with crucial team leadership.
You will define data science approaches, design robust software architectures, mentor junior colleagues, and ensure delivery rigour across projects all while building deep client relationships and solidifying our reputation as a leader in practical, measurable AI.What you'll be doing:Leading project teams that deliver bespoke algorithms and high-stakes AI solutions to clients across the sector.Conceiving the core data science approach and designing the associated robust software architecture for new engagements.Mentoring a small number of data scientists and supporting the professional growth of technical team members on projects.Partnering with commercial teams to build client relationships and shape project scope for technical feasibility.Contributing to Faculty’s thought leadership and reputation through delivering courses, public speaking, or open-source projects.Ensuring best practices are followed throughout the project lifecycle to guarantee high-quality, impactful delivery.Who we're looking for:You possess senior experience in a professional data science position or a quantitative academic field.You demonstrate strong programming skills, with the ability to be a fluent Python programmer, using core libraries (NumPy, Pandas) and a deep-learning framework (e.g., PyTorch).You have a deep expertise in core data science paradigms (supervised/unsupervised, NLP, validation), demonstrating a proficiency across the standard data science toolkit, including the ability to develop new, innovative algorithms.You bring a leadership mindset, focused on growing the technical capabilities of the team and nurturing a collaborative culture.You exhibit commercial awareness, with experience in client-facing work and the ability to translate business problems into a rigorous mathematical framework.You are skilled in project planning, assessing technical feasibility, estimating delivery timelines, and leading a team to deliver high-quality work on a strict schedule.Our Interview ProcessTalent Team Screen (30 minutes)Take Home Technical AssessmentTechnical Interview (90 minutes)Commercial Interview (60 minutes)Our Recruitment EthosWe aim to grow the best team - not the most similar one. We know that diversity of individuals fosters diversity of thought, and that strengthens our principle of seeking truth. And we know from experience that diverse teams deliver better work, relevant to the world in which we live. We’re united by a deep intellectual curiosity and desire to use our abilities for measurable positive impact. We strongly encourage applications from people of all backgrounds, ethnicities, genders, religions and sexual orientations.Some of our standout benefits:Unlimited Annual Leave PolicyPrivate healthcare and dentalEnhanced parental leaveFamily-Friendly Flexibility & Flexible workingSanctus CoachingHybrid WorkingIf you don’t feel you meet all the requirements, but are excited by the role and know you bring some key strengths, please don't hesitate in applying as you might be right for this role, or other roles. We are open to conversations about part-time hours.
No items found.
2026-03-18 10:46
Backend Engineer- Inference Services
Deepgram
201-500
$150,000 – $220,000
United States
Full-time
Remote
false
Company OverviewDeepgram is the leading platform underpinning the emerging trillion-dollar Voice AI economy, providing real-time APIs for speech-to-text (STT), text-to-speech (TTS), and building production-grade voice agents at scale. More than 200,000 developers and 1,300+ organizations build voice offerings that are ‘Powered by Deepgram’, including Twilio, Cloudflare, Sierra, Decagon, Vapi, Daily, Cresta, Granola, and Jack in the Box. Deepgram’s voice-native foundation models are accessed through cloud APIs or as self-hosted and on-premises software, with unmatched accuracy, low latency, and cost efficiency. Backed by a recent Series C led by leading global investors and strategic partners, Deepgram has processed over 50,000 years of audio and transcribed more than 1 trillion words. There is no organization in the world that understands voice better than Deepgram.Company Operating RhythmAt Deepgram, we expect an AI-first mindset—AI use and comfort aren’t optional, they’re core to how we operate, innovate, and measure performance.Every team member who works at Deepgram is expected to actively use and experiment with advanced AI tools, and even build your own into your everyday work. We measure how effectively AI is applied to deliver results, and consistent, creative use of the latest AI capabilities is key to success here. Candidates should be comfortable adopting new models and modes quickly, integrating AI into their workflows, and continuously pushing the boundaries of what these technologies can do.Additionally, we move at the pace of AI. Change is rapid, and you can expect your day-to-day work to evolve just as quickly. This may not be the right role if you’re not excited to experiment, adapt, think on your feet, and learn constantly, or if you’re seeking something highly prescriptive with a traditional 9-to-5.Opportunity:Deepgram is looking for a Backend Software Engineer to join the Engine team to lead the design and implementation of Deepgram’s products. You will design and implement secure, robust, and scalable services for speech processing; efficient, distributed compute orchestration; optimized scheduling, and more. Your skill at building highly reusable code that overcomes technical challenges is paired with an intuition for delightful user experiences. You will be a critical voice in Deepgram’s Product and Engineering teams, driving high impact products from start to finish.What You’ll DoImprove Deepgram’s core inference services including areas in networking, speech processing, audio transcoding, and latency and memory optimizationDevelop processes for measuring, building, and optimizing services to maximize system performanceDebug complex system issues that include networking, scheduling, and high performance computing interactionsRapidly customize backend services to support our customer needsPartner with Product to design and implement new services, features, and/or products end to endYou’ll Love This Role If YouThrive in a fast-paced, impact-driven environment where learning new skills on-the-fly is not only encouraged but a regular necessityEnjoy balancing decisions about product and feature maturity to decide when to make minimally invasive changes versus when to incorporate detailed design workIt’s Important To Us That You Have3+ years of experience in an industry roleProgramming experience in Rust (or C, C++), with competence in PythonExcellent communication and organizational skills, both written and verbal.A high level of experience and understanding of version control; preferably git.Comprehensive experience with UNIX-style systems.It Would Be Great if You HadExperience with modern machine learning, such as experience with a framework like Torch or implementation knowledge of architectures like CNNs, RNNS, and transformersExperience with audio processingBenefits & Perks*Holistic healthMedical, dental, vision benefitsAnnual wellness stipendMental health supportLife, STD, LTD Income Insurance PlansWork/life blendUnlimited PTOGenerous paid parental leaveFlexible schedule12 Paid US company holidaysQuarterly personal productivity stipendOne-time stipend for home office upgrades401(k) plan with company matchTax Savings ProgramsContinuous learningLearning / Education stipendParticipation in talks and conferencesEmployee Resource GroupsAI enablement workshops / sessions*For candidates outside of the US, we use an Employer of Record model in many countries, which means benefits are administered locally and governed by country-specific regulations. Because of this, benefits will differ by region — in some cases international employees receive benefits US employees do not, and vice versa. As we scale, we will continue to evaluate where we can create more alignment, but a 1:1 global benefits structure is not always legally or operationally possible.Backed by prominent investors including Y Combinator, Madrona, Tiger Global, Wing VC and NVIDIA, Deepgram has raised over $215M in total funding. If you're looking to work on cutting-edge technology and make a significant impact in the AI industry, we'd love to hear from you!Deepgram is an equal opportunity employer. We want all voices and perspectives represented in our workforce. We are a curious bunch focused on collaboration and doing the right thing. We put our customers first, grow together and move quickly. We do not discriminate on the basis of race, religion, color, national origin, gender, sexual orientation, gender identity or expression, age, marital status, veteran status, disability status, pregnancy, parental status, genetic information, political affiliation, or any other status protected by the laws or regulations in the locations where we operate.We are happy to provide accommodations for applicants who need them.
No items found.
2026-03-18 10:46
Manager, AI Deployment Engineering - Health & Life Sciences
OpenAI
5000+
$251,000 – $335,000
United States
Full-time
Remote
false
About the TeamThe AI Deployment Engineering team works at the intersection of cutting-edge AI research and real-world application. We partner directly with OpenAI’s most strategic customers to design, implement, and scale production deployments of generative AI systems. Our work ensures that organizations can safely and effectively integrate AI into critical workflows.Within this team, the Health & Life Sciences (HLS) segment focuses on enabling pharmaceutical manufacturers, biotechnology companies, medical device organizations, contract research organizations (CROs), and health systems to adopt AI responsibly in regulated environments. We work across clinical, scientific, regulatory, and operational functions to unlock meaningful impact while maintaining the highest standards of safety, privacy, and compliance.About the RoleWe are seeking a Manager, AI Deployment Engineering to lead our Healthcare & Life Sciences deployment efforts. In this role, you will build and mentor a high-performing team of deployment engineers dedicated to helping HLS organizations move from experimentation to production with OpenAI technologies.You will operate at both strategic and technical levels — partnering with executive stakeholders, technical leaders, and cross-functional internal teams to deliver secure, compliant, and high-impact AI deployments. This role requires a strong technical foundation, people leadership experience, and a deep appreciation for the complexity of regulated healthcare and life sciences environments.This role is based in San Francisco or Seattle. We use a hybrid work model of 3 days in the office per week.In this role, you will:Own the strategy and operating model of the HLS AI Deployment Engineering team, ensuring alignment with company objectives and the evolving needs of our customers.Hire, mentor, and develop a high-impact team of AI Deployment Engineers focused on HLS production deploymentsEstablish operating mechanisms, delivery standards, and best practices tailored to regulated environmentsFoster a culture of technical excellence, customer empathy, and responsible AI deploymentDrive Successful Enterprise Deployments and oversee end-to-end implementation of generative AI applications in production across healthcare and life sciences organizationsGuide customers through complex integration efforts spanning R&D, clinical development, regulatory affairs, medical affairs, and ITDevelop scalable frameworks for secure, compliant AI adoption in environments governed by HIPAA, GxP, FDA, EMA, and related regulatory standardsEnsure measurable impact through activation, adoption, and workflow transformation (e.g., drug discovery acceleration, clinical documentation support, regulatory submission drafting)Collaborate closely with Sales, Account Directors, Solutions Architects, Product, Security, and Legal teamsServe as a trusted technical advisor to executive and senior technical stakeholders at enterprise HLS customersProvide structured product feedback informed by real-world deployment challenges and industry requirementsYou’ll thrive in this role if you:Have 8+ years of experience in technical delivery, solutions engineering, or deployment roles, including people management experienceHave led enterprise-scale implementations of AI, ML, or platform technologiesBring experience in healthcare or life sciences environments, including familiarity with clinical research, drug development, regulatory operations, or health system infrastructureUnderstand compliance frameworks such as HIPAA, GxP, and global regulatory considerationsAre comfortable engaging with executive stakeholders while maintaining technical depthEnjoy operating in ambiguous, fast-moving environments and building structure where it does not yet existCare deeply about responsible AI and its application in high-stakes domainsAbout OpenAIOpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity. We are an equal opportunity employer, and we do not discriminate on the basis of race, religion, color, national origin, sex, sexual orientation, age, veteran status, disability, genetic information, or other applicable legally protected characteristic. For additional information, please see OpenAI’s Affirmative Action and Equal Employment Opportunity Policy Statement.Background checks for applicants will be administered in accordance with applicable law, and qualified applicants with arrest or conviction records will be considered for employment consistent with those laws, including the San Francisco Fair Chance Ordinance, the Los Angeles County Fair Chance Ordinance for Employers, and the California Fair Chance Act, for US-based candidates. For unincorporated Los Angeles County workers: we reasonably believe that criminal history may have a direct, adverse and negative relationship with the following job duties, potentially resulting in the withdrawal of a conditional offer of employment: protect computer hardware entrusted to you from theft, loss or damage; return all computer hardware in your possession (including the data contained therein) upon termination of employment or end of assignment; and maintain the confidentiality of proprietary, confidential, and non-public information. In addition, job duties require access to secure and protected information technology systems and related data security obligations.To notify OpenAI that you believe this job posting is non-compliant, please submit a report through this form. No response will be provided to inquiries unrelated to job posting compliance.We are committed to providing reasonable accommodations to applicants with disabilities, and requests can be made via this link.OpenAI Global Applicant Privacy PolicyAt OpenAI, we believe artificial intelligence has the potential to help people solve immense global challenges, and we want the upside of AI to be widely shared. Join us in shaping the future of technology.
No items found.
2026-03-18 10:46
Head of Internal Tools Engineering
Bjak
201-500
United States
Full-time
Remote
false
About the RoleYou will architect, build, and scale the internal technology ecosystem that accelerates workforce productivity, eliminates operational friction, and gives the company a compounding infrastructure advantage. You will treat internal tools with the same product rigour and user-centricity as any external product — because the quality of internal systems directly determines organisational velocity.You will lead a cross-functional engineering team, make high-stakes build-vs-buy decisions, and drive AI adoption into internal workflows. This role requires someone who can see the company’s operational architecture as a system design problem and solve it with software.What You Will DoInternal Platform Strategy & RoadmapOwn the end-to-end strategy and roadmap for all internal tools, platforms, and automation — treating internal technology as a product, not a cost centre.Make strategic build-vs-buy decisions, knowing when a custom-built solution creates a lasting advantage and when a SaaS tool is the right answer.Map current and next-state process flows across the entire internal toolchain and lead systems transformation for internal teams.Systems Architecture & EngineeringArchitect and maintain the full engineering lifecycle for internal platforms — from ideation and design through deployment, iteration, and deprecation.Build seamless, API-first ecosystems that integrate internal tools across HR systems, finance platforms, knowledge management, CRM, and developer infrastructure.Own system reliability and operational resilience: establish success metrics for uptime, performance, and employee productivity.Design scalable, secure architectures using cloud-native principles, microservices, and modern integration patterns.AI & AutomationLead the strategy for integrating AI and LLMs into internal workflows — reimagining how knowledge is shared, decisions are supported, and work is executed across the organisation.Deploy intelligent automation tools — including AI where appropriate — to streamline internal processes and improve decision-making.Evaluate and integrate AI-assisted troubleshooting, proactive recommendations, and intelligent automation into the internal platform layer.Stay ahead of technology trends and drive continuous experimentation — you build prototypes, not slide decks.Developer Experience & ProductivityReduce cognitive load for internal users by providing golden paths, standardised workflows, and self-service capabilities.Ensure frictionless onboarding and seamless integration across the tool ecosystem.Measure platform success through adoption rates, user satisfaction, DORA metrics, and productivity impact — not feature count.Team LeadershipBuild, lead, and mentor a high-performing team of engineers and engineering managers.Cultivate a collaborative engineering culture rooted in ownership, speed, and craftsmanship.Provide technical mentorship and create growth paths for individual contributors and managers alike.Foster psychological safety and a feedback-driven environment that empowers people to do their best work.Cross-Functional CollaborationPartner with People, Finance, Engineering, Legal, and Operations leadership to translate complex business needs into a unified technical vision.Serve as the bridge between business stakeholders and the engineering team — you speak both languages fluently.Align internal platform investments with broader company strategy and demonstrate measurable ROI.What You Will NeedMust-Have12+ years of experience in software engineering, with at least 5 years in engineering leadership (managing teams or managing managers).Strong hands-on technical background: you’ve built production systems and can still credibly review architecture, code, and system design.Deep understanding of cloud-based systems (AWS, GCP, or Azure), APIs, microservices, data pipelines, and modern infrastructure.Proven track record of building and scaling internal tools or platforms that serve cross-functional business teams — not just engineering.Experience making build-vs-buy decisions and managing a portfolio of custom-built and third-party tools.Strong experience designing and building internal platforms and automation systemsStrong product thinking: you define success in terms of user outcomes and business impact, not technical output.Experience navigating the full SDLC from ideation through deprecation — you know when to build, iterate, and retire.Excellent communication skills: you can articulate complex technical concepts to non-technical executives and translate business problems into engineering roadmaps.Nice-to-HaveExperience at a high-growth technology or AI-native company that scaled rapidly.Background in platform engineering, developer experience, or internal developer platforms (IDPs).Familiarity with HRIS, ERP, and business systems integration (Workday, Salesforce, NetSuite, etc.).Experience with cybersecurity best practices and compliance frameworks for internal systems.Prior experience leading internal technology through M&A integrations or multi-entity consolidation.Exposure to knowledge management systems, internal search, and enterprise AI assistants.Experience integrating AI/LLMs into internal workflows or productivity toolsWhat Success Looks LikeWithin 30 days: Completed a full audit of the existing internal toolchain, identified the top friction points, and presented a prioritised roadmap.Within 60 days: Shipped at least two high-impact internal tools or automations that measurably improve productivity, and established team operating rhythm.Within 12 months: The internal tools ecosystem is a recognised competitive advantage — teams actively request new capabilities, AI is embedded in daily workflows, and internal NPS is consistently high.Who You AreAn engineer who thinks like a CEO. You’ve built production systems, led engineering teams, and understand that the best internal tools don’t just save time — they change how an organisation thinks and moves. You’re obsessed with removing friction, allergic to manual workarounds, and energised by the idea that great internal infrastructure compounds into organisational speed. You believe internal tools deserve the same craft as customer-facing products. You see AI not as a buzzword but as the most important lever for internal productivity in a generation, and you want to be the person who pulls it.How We WorkWe operate as a small, senior, hands-on team. Engineers own features end-to-end — from design discussion through production monitoring.Code reviews and design reviews are expected for all meaningful changes. We discuss architecture openly, make decisions quickly, and ship frequently.Interview processIf there appears to be a fit, we'll reach out to schedule 3, but no more than 4 interviews.Applications are evaluated by our technical team members. Interviews will be conducted via virtual meetings and/or onsite.We value transparency and efficiency, so expect a prompt decision. If you've demonstrated the exceptional skills and mindset we're looking for, the process to offer may be shorter.
No items found.
2026-03-18 10:46
Automotive Engineering & Python Expert - Freelance AI Trainer
Mindrift
1001-5000
$13 / hour
Argentina
Part-time
Remote
false
Please submit your CV in English and indicate your level of English proficiency. Mindrift connects specialists with project-based AI opportunities for leading tech companies, focused on testing, evaluating, and improving AI systems. Participation is project-based, not permanent employment.What this opportunity involves While each project involves unique tasks, contributors may: Design graduate- and industry-level automotive engineering problems grounded in real practice; Evaluate AI-generated solutions for correctness, assumptions, and engineering logic; Validate analytical or numerical results using Python (NumPy, SciPy, Pandas); Improve AI reasoning to align with first principles and accepted engineering standards; Apply structured scoring criteria to assess multi-step problem solving. What we look for This opportunity is a good fit for automotive engineers with an experience in python open to part-time, non-permanent projects. Ideally, contributors will have: Degree in Automotive Engineering or related fields, e.g. Mechatronics, Manufacturing Engineering, Mechanical Engineering, Aerospace Engineering, etc. 3+ years of professional automotive engineering experience Strong written English (C1/C2) Strong Python proficiency for numerical validation Stable internet connection Professional certifications (e.g., PE, CEng, PMP) and experience in international or applied projects are an advantage.How it works Apply → Pass qualification(s) → Join a project → Complete tasks → Get paidProject time expectations For this project, tasks are estimated to require around 10–20 hours per week during active phases, based on project requirements. This is an estimate, not a guaranteed workload, and applies only while the project is active. CompensationOn this project, contributors can earn up to $13 per hour equivalent, depending on their level and pace of contribution.Compensation varies across projects depending on scope, complexity, and required expertise. Please note that other projects on the platform may offer different earning levels based on their requirements.
No items found.
2026-03-18 10:46
Deployed Engineer (Boston)
LangChain
101-200
$150,000 – $250,000
United States
Full-time
Remote
false
About UsAt LangChain, our mission is to make intelligent agents ubiquitous. We build the foundation for agent engineering in the real world, helping developers move from prototypes to production-ready AI agents that teams can rely on. We began as widely adopted open-source tools and have grown to also offer a platform for building, evaluating, deploying, and operating agents at scale.Today, LangChain, LangGraph, LangSmith, and Agent Builder are used by teams shipping real AI products across startups and large enterprises. Millions of developers trust LangChain to power AI teams at companies like Replit, Clay, Coinbase, Workday, Lyft, Cloudflare, Harvey, Rippling, Vanta, and 35% of the Fortune 500.With $125M raised at Series B from IVP, Sequoia, Benchmark, CapitalG, and Sapphire Ventures, we’re at a stage where we’re continuing to develop new products, growth is accelerating, and all team members have meaningful impact on what we build and how we work together. LangChain is a place where your contributions can shape how this technology shows up in the real world.About the TeamThe Deployed Engineering team works directly with companies building and running AI agents in production, helping turn ideas and prototypes into systems teams can rely on.This is a hands-on, highly technical team that partners closely with customer engineers across the full lifecycle, from pre-sales evaluations to post-deployment advisory work. The focus is on achieving the technical win, co-designing agent architectures, and helping customers operate agents reliably at scale using the LangChain suite.Deployed Engineers sit at the intersection of engineering, product, and go-to-market, shaping how LangChain is adopted in the field and feeding real-world insights back into the platform.About the RoleThe Deployed Engineer…You’ll work on some of the hardest problems in applied AI — not demos, not research, but systems that real teams depend on in production. The feedback loop is fast, the impact is visible, and the work you do directly shapes how AI agents are built in the real world.Location(s): BostonWhat You’ll DoCo-architect and co-build production AI agents with customer engineering teamsOwn the technical win in pre-sales by designing POCs, answering deep technical questions, and guiding evaluationsHelp customers deploy and operate agent-based applications such as conversational agents, research agents, and multi-step workflowsAdvise customers post-sale on architecture, best practices, and roadmap-level decisionsRun technical demos, trainings, and workshops for developer audiencesSurface field feedback and contribute reusable patterns, cookbooks, and example code that scale across customersOccasionally contribute code upstream when it meaningfully improves customer outcomesWhat You’ll Bring3+ years in a relevant technical role (software engineering, customer engineering, solutions engineering, founding/product engineering), ideally in a startup or scale-upStrong Python, JavaScript and systems fundamentalsHave designed agent-based or LLM-powered applications beyond simple API calls, including multi-step workflows, orchestration, and failure handlingAre comfortable working directly with customers during POCs, architecture reviews, and technical evaluationsCan explain technical tradeoffs clearly and build trust with developer audiencesTake responsibility for outcomes, not just recommendationsHave a bias toward action and enjoy figuring things out as you goAre excited about operating AI agents in production, not just building demosNice to Have’s:You’ve deployed AI agents in production, especially using LangChain, LangGraph, or similar frameworksWorked with LLM evaluation, observability, or guardrailsHave experience with cloud environments (AWS, GCP, Azure), containers, and basic Kubernetes conceptsHave shipped and operated production software and are comfortable owning systems under real-world constraintsCompensation & BenefitsWe offer competitive compensation that includes base salary, variable compensation for relevant roles, meaningful equity, benefits, and perks. Benefits include things like medical, dental, and vision coverage, flexible vacation, a 401(k) plan, and life insurance. Actual compensation and offerings will vary based on role, level, and location. Team members in the EU, UK, and APAC receive locally competitive benefits aligned with regional norms and regulations.Annual OTE range: $150,000–$250,000 USD
No items found.
2026-03-18 10:46
Solutions Architect (Austin)
LangChain
101-200
$170,000 – $190,000
United States
Full-time
Remote
false
About the RoleWe're looking for a Solutions Architect to join our Professional Services team. You'll work directly with enterprise customers to design, deploy, and optimize production-grade AI infrastructure and agent systems. You'll be responsible for architecting scalable, secure infrastructure deployments and building reliable, well-evaluated agent applications that solve real business problems.This role combines software development, infrastructure/platform engineering, and customer-facing skills. You'll work on everything from Kubernetes cluster design to multi-agent system architecture, requiring deep technical expertise across both infrastructure and agent engineering domains.This role offers direct impact on customer success, the opportunity to shape best practices, and work with cutting-edge AI technology. You'll join a collaborative team environment with a strong engineering culture.About Us:At LangChain, our mission is to make intelligent agents ubiquitous. We build the foundation for agent engineering in the real world, helping developers move from prototypes to production-ready AI agents that teams can rely on. We began as widely adopted open-source tools and have grown to also offer a platform for building, evaluating, deploying, and operating agents at scale.Today, LangChain, LangGraph, LangSmith, and Agent Builder are used by teams shipping real AI products across startups and large enterprises. Millions of developers trust LangChain to power AI teams at companies like Replit, Clay, Coinbase, Workday, Lyft, Cloudflare, Harvey, Rippling, Vanta, and 35% of the Fortune 500.With $125M raised at Series B from IVP, Sequoia, Benchmark, CapitalG, and Sapphire Ventures, we’re at a stage where we’re continuing to develop new products, growth is accelerating, and all team members have meaningful impact on what we build and how we work together. LangChain is a place where your contributions can shape how this technology shows up in the real world.
Key ResponsibilitiesInfrastructure & Platform Engineering: Design scalable, highly-available infrastructure for AI platform deployments (compute, storage, networking, security), enterprise integration patterns, Infrastructure as Code (Terraform, Helm), multi-region HA/DR strategies, and CI/CD pipelinesAgent Engineering & Development: Design multi-agent systems using different patterns, implement agent logic using modern frameworks (langchain/langgraph), design comprehensive evaluation frameworks, optimize prompts with A/B testing, and guide deployment/operationsCustomer Engagement & Assessment: Lead technical maturity assessments, work directly with enterprise customers to understand requirements and present recommendations, and partner with Engagement Managers and Product/Engineering teamsWhat We're Looking For
Required Experience7+ years of experience in a technical, hands-on customer-facing roles such as Solutions Architect or Forward Deployed Engineer. We also like former founders, so if you have an unusual background, but all the right skillsets, you are welcome to applyInfrastructure & Platform:3+ years of experience designing and deploying production infrastructure on cloud platforms (GCP, AWS, or Azure)Strong Kubernetes experience (GKE, EKS, or AKS) including cluster design, autoscaling, and multi-zone deploymentsExperience with Infrastructure as Code (Terraform, Helm) and GitOps practicesKnowledge of database systems (relational databases, in-memory data stores) including HA, replication, backup strategies, and sizingExperience designing high-availability and disaster recovery solutionsStrong understanding of networking, security (SSO/RBAC, TLS, secrets management), and observability (Prometheus, Grafana, Datadog)Experience with CI/CD pipelines for infrastructure and applicationsAgent Engineering & Development:1+ years of experience building production AI/ML applications or agentsStrong experience with LLM frameworks (LangChain, LangGraph, or similar) for building agent-based applicationsExperience with state management patterns (short-term and long-term memory)Experience designing and implementing evaluation frameworks for AI applicationsStrong prompt engineering skills with experience in optimization and A/B testingExperience with vector stores, RAG patterns, and knowledge organizationExperience with tool integration, API design, and error handling patternsStrong Python and/or TypeScript development skillsCustomer-Facing:Customer-facing experience with enterprise customersExperience conducting technical assessments or infrastructure auditsStrong communication skills with ability to explain technical concepts to diverse audiencesKey AttributesStrong problem-solving skills with ability to analyze complex requirements and design elegant solutionsExcellent customer-facing communication skills, able to explain technical concepts to diverse audiencesExperience working cross-functionally with engineering teams, product teams, and customersConsultative approach with ability to understand customer needs, provide recommendations, and guide implementationAbility to balance infrastructure architecture with agent development workStrong engineering background with hands-on development experienceLocation: Austin, TexasCompensation: We offer competitive compensation that includes base salary, variable compensation for relevant roles, meaningful equity, benefits, and perks. Benefits include things like medical, dental, and vision coverage, flexible vacation, a 401(k) plan, and life insurance. Actual compensation and offerings will vary based on role, level, and location. Team members in the EU, UK, and APAC receive locally competitive benefits aligned with regional norms and regulations.$170K to $190K
No items found.
2026-03-18 10:46
Research Scientist – Tabular & Structured Machine Learning
Granica
11-50
$160,000 – $250,000
United States
Full-time
Remote
false
Research Scientist – Tabular & Structured Machine LearningThe MissionAI today is limited not only by model design but by the inefficiency of the data that feeds it. At scale, each redundant byte, poorly organized dataset, and inefficient data path slows progress and compounds into enormous cost, latency, and energy waste.Granica’s mission is to remove that inefficiency. We combine advances in information theory, probabilistic modeling, and distributed systems to design self-optimizing data infrastructure: systems that continuously improve how information is represented, compressed, and used by AI.Granica’s research group is led by Prof. Andrea Montanari (Stanford), bridging advances in learning theory and information efficiency with large-scale distributed systems. Together, we share a conviction that the next leap in AI will come not only from larger models, but from more efficient learning systems and better data representations.Most modern AI research focuses on text, images, or video. Granica’s work focuses on the far less explored but economically critical domain of large-scale structured and tabular data, which powers the majority of enterprise decision-making systems.Granica is pioneering a new class of structured AI models: foundational models built to learn and reason from relational, tabular, and structured data. While others focus on unstructured text or media, we are exploring the next frontier: systems that understand and reason over the structured information that runs the global economy.This role focuses specifically on machine learning for structured and tabular data rather than general LLM application development.What You’ll Build and ResearchInvent and prototype algorithms that advance the foundations of machine learning for structured and tabular dataDevelop new representation learning techniques and information models for large enterprise datasetsBuild adaptive learners combining statistical learning theory, probabilistic modeling, and large-scale systems optimizationContribute to the development of large tabular models and structured foundation modelsDesign architectures integrating relational, symbolic, and neural learning componentsResearch and implement methods for dataset compression, selection, and representation to improve learning efficiencyDevelop cost models and optimization frameworks for large-scale structured learning systemsCollaborate closely with the Granica research group led by Prof. Andrea Montanari (Stanford) and with systems engineersRapidly prototype new algorithms and evaluate them on real enterprise datasetsPublish and contribute to the broader research community shaping the future of structured AI and efficient ML systemsWhat You’ll BringPhD in Machine Learning, Statistics, Computer Science, Applied Mathematics, or a related fieldResearch experience related to structured, relational, or tabular dataExperience in one or more of the following areas:Tabular or relational machine learningRepresentation learning for structured dataStatistical learning theory or generalizationProbabilistic modeling or Bayesian inferenceOptimization for machine learningScalable or distributed ML systemsExperience working with structured datasets or relational data systemsStrong grounding in statistics, optimization, information theory, or probabilistic inferenceHands-on experience with PyTorch, JAX, or TensorFlowStrong programming skills in Python or RustDemonstrated ability to translate theoretical ideas into working systems or prototypesCuriosity about how structure and relational information enable new forms of learning and reasoningA pragmatic research mindset: you value elegant ideas but also ship systems that work at scaleBonusResearch in tabular machine learning, relational representation learning, or structured data modelingExperience building large-scale ML infrastructure or distributed training systemsFamiliarity with data systems, query engines, or dataset optimization pipelinesPublications at top venues such as NeurIPS, ICML, ICLR, COLT, KDD, AAAIContributions to open-source ML systems or research-to-production toolingCompensation & BenefitsCompetitive salary, meaningful equity, and substantial bonus for top performersFlexible time off plus comprehensive health coverage for you and your familySupport for research, publication, and deep technical explorationAt Granica, you will shape the fundamental infrastructure that makes intelligence itself efficient, structured, and enduring. Join us to build the foundational data systems that power the future of enterprise AI!
No items found.
2026-03-18 10:46
Manager/Sr. Manager, Biopharma Marketing
PathAI
201-500
$181,500 – $278,300
No items found.
Full-time
Remote
false
Who We Are
PathAI's mission is to improve patient outcomes with AI-powered pathology. Our platform promises substantial improvements to the accuracy of diagnosis and the efficacy of treatment of diseases like cancer, leveraging modern approaches in machine learning and artificial intelligence. We have a track record of success in deploying AI algorithms for histopathology in translational research, pathology labs and clinical trials. Rigorous science and careful analysis is critical to the success of everything we do. Our team, composed of diverse employees with a wide range of backgrounds and experiences, is passionate about solving challenging problems and making a huge impact on patient outcomes.
Where You Fit
As the Associate Director, MLOps Lead, you will lead the team responsible for the backbone of our AI/ML Stack: the infrastructure that bridges ML research and massive-scale production. Your primary directive is to evolve our stack to meet the next scale of needs in large scale ML training & inference workloads.
You’re someone who enjoys designing and building for reliability, relishes collaboration and technical challenges, and takes pride in making things better – without taking yourself too seriously. Our technical space is broad: high-scale AI training & inference workloads, cloud infrastructure, Kubernetes, observability, distributed systems, and a bit of everything in between.
What You’ll Do
This role is critical for driving the scalability and efficiency of our Machine Learning Operations platform with high-impact & high growth strategic initiatives.
Vision and Roadmap: Develop and execute the long term vision & roadmap for MLOPs team to support ML development and deployment needs across the business units. Successfully manage the tension between short-term tactical deliveries and long-term architectural transformation for future growth.
Team Management: Lead and mentor a team of 6-7+ high-performing engineers. Strategically allocate resources to manage support for existing services while executing key strategic initiatives.
Cross-Functional Collaboration: Partner with leaders across machine learning, data science, product engineering, and infrastructure to proactively identify pain points, address bottlenecks, and facilitate the deployment of new solutions.
Foundation Model Readiness: Architect the compute and storage pipelines required for ML Engineers to manage millions of slides and complex derived artifacts without data fragmentation or synchronization latency.
Inference Modernization: Modernize the AI Product inference stack to support 5-10x growth of AI runs across global deployments.
System Observability: Collaborate with Site Reliability Engineering (SRE) to establish comprehensive metrics covering compute under-utilization, network bottlenecks, and granular cost and turn-around-time attribution.
Technology Refresh: Conduct "Build vs. Buy" assessments, leading "Stack Refresh" audits to benchmark our proprietary tools against best-in-class commercial and open-source alternatives to meet our future needs.
What You Bring
To be successful in this role with us, you'll at least need:
Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field (or equivalent experience).
2-3+ years of experience managing engineering team(s), with a focus on building production-grade frameworks for MLOps or ML Infrastructure.
Deep technical expertise with ML workloads on kubernetes, cloud computing platforms (AWS/GCP/Azure), workflow orchestration (Airflow, Kubeflow, or proprietary equivalents) and DevOps principles and infrastructure-as-code (Helm, Terraform).
Proven experience managing petabyte-scale datasets and high-throughput production inference pipelines.
Strong software engineering skills in complex, multi-language systems and experience with scalable service architecture.
Use of AI assistants (e.g. CoPilot, Cursor, Claude) across platform development lifecycle.
It Would Be Great If You Also Have
Exposure to ML frameworks like PyTorch or Scikit-learn.
Experience with large-scale data processing frameworks (e.g. Spark, Hive, Databricks, Amazon EMR)
Expertise in MLOps principles, including model lifecycle management, feature stores, model monitoring, and CI/CD for ML.
Familiarity with security and compliance best practices in ML systems.
We Want To Hear From You
At PathAI, we are looking for individuals who are team players, are willing to do the work no matter how big or small it may be, and who are passionate about everything they do. If this sounds like you, even if you may not match the job description to a tee, we encourage you to apply. You could be exactly what we're looking for.
PathAI is an equal opportunity employer, dedicated to creating a workplace that is free of harassment and discrimination. We base our employment decisions on business needs, job requirements, and qualifications — that's all. We do not discriminate based on race, gender, religion, health, personal beliefs, age, family or parental status, or any other status. We don't tolerate any kind of discrimination or bias, and we are looking for teammates who feel the same way.
The cash compensation outlined below includes base salary or hourly wage and on-target commission for employees in eligible roles. The summary below indicates if an employee in this position is eligible for annual bonus, overtime pay and equity awards. Individual compensation packages are tailored based on skills, experience, qualifications, and other job-related factors.
Annual Pay Range:
AD, MLOps: $181,500 - $278,300
Not Overtime Eligible
Eligible for Equity
No items found.
2026-03-18 10:46
Senior Analytics Engineer
You.com
201-500
$165,000 – $200,000
United States
Full-time
Remote
false
you.com is an AI-powered search and productivity platform designed to empower users with personalized, efficient, and trustworthy search experiences. As a cutting-edge technology company, we combine advanced AI models with user-first principles to deliver tools that enhance discovery, creativity, and productivity. At you.com, we are on a mission to create the most helpful search engine in the world—one that prioritizes transparency, privacy, and user control.
We’re building a team of innovators, problem-solvers, and visionaries who are passionate about shaping the future of AI and technology. At you.com, you’ll have the opportunity to work on impactful projects, collaborate with some of the brightest minds in the industry, and grow your career in an environment that values creativity, diversity, and curiosity. If you’re ready to make a difference and help us revolutionize the way people search and work, we’d love to have you join us!
About the Role
AI is transforming how enterprises sell, support, and operate — surfacing hidden knowledge, automating workflows, and collapsing weeks of effort into minutes. As an AI Engineer, Enterprise Solutions, you’ll act like a startup CTO, turning this vision into robust, production-ready AI solutions that ship fast and deliver measurable wins for our customers. You will get a front seat to AI applications in the economy along with defining and shaping the future.
Responsibilities
Build and Ship: Design and develop AI applications primarily in Python. Run evaluations to validate models and package solutions for Kubernetes, AWS, or adapt them to customer on-premises clusters.
Work with Customers: Lead discovery sessions, guide pilot projects, and ensure successful deployments. Collaborate mostly remotely with occasional on-site workshops.
Run and Improve: Monitor system performance and reliability. Add to the logging, billing and auth services. Build internal tooling to automate repetitive tasks.
Share What You Learn: Provide feedback on patterns, pain points, and reusable modules to the core product team to influence the future direction of the AI platform.
Qualifications
Required:
2+ years of experience writing solid production-quality software.
Hands-on experience with LLMs and a solid understanding of machine learning fundamentals.
Strong customer empathy: ability to listen, ask insightful questions, and translate real-world pain points into intuitive technical designs.
Product-minded approach: focus on outcomes, comfortable shaping scope with product managers, defining success metrics, and making trade-offs to accelerate learning.
Excellent communication skills, able to engage effectively with executives and engineers both remotely and on-site.
Versatile problem-solver who thrives in ambiguous environments and enjoys rapid learning.
Nice to Have:
Experience in forward-deployed or product-oriented roles.
Exposure to regulated industries such as healthcare or finance.
Proficiency with Terraform or Pulumi and experience across multiple cloud platforms.
Experience integrating with ERP, CRM, or other large enterprise systems.
Our salary bands are structured based on a combination of geographic tiers and internal leveling. Compensation is determined by multiple factors assessed during the interview process, with the final offer reflecting these considerations.Salary Band$165,000—$200,000 USDCompany Perks:
Hubs in San Francisco and New York City offering regular in-person gatherings and co-working sessions
Flexible PTO with U.S. holidays observed and a week shutdown in December to rest and recharge*
A competitive health insurance plan covers 100% of the policyholder and 75% for dependents*
12 weeks of paid parental leave in the US*
401k program, 3% match - vested immediately!*
$500 work-from-home stipend to be used up to a year of your start date*
$1,200 per year Health & Wellness Allowance to support your personal goals*
The chance to collaborate with a team at the forefront of AI research
*Certain perks and benefits are limited to full-time employees only
You.com participates in E-Verify. We will provide the Social Security Administration (SSA) and, if necessary, the Department of Homeland Security (DHS) with information from each new employee’s Form I-9 to confirm work authorization. (English/Spanish: E-Verify Participation/Right to Work) We are also an inclusive, equitable, and accessible workplace. Please let us know if you require accommodation for any portion of the recruitment and hiring process.
Beware of recruiting scams: You.com will only contact you through official @You.com email addresses and will never ask for payment or sensitive personal information during the hiring process.
No items found.
2026-03-18 10:46
Engineering Manager, Active Learning
Deepgram
201-500
$180,000 – $220,000
United States
Full-time
Remote
false
Company OverviewDeepgram is the leading platform underpinning the emerging trillion-dollar Voice AI economy, providing real-time APIs for speech-to-text (STT), text-to-speech (TTS), and building production-grade voice agents at scale. More than 200,000 developers and 1,300+ organizations build voice offerings that are ‘Powered by Deepgram’, including Twilio, Cloudflare, Sierra, Decagon, Vapi, Daily, Cresta, Granola, and Jack in the Box. Deepgram’s voice-native foundation models are accessed through cloud APIs or as self-hosted and on-premises software, with unmatched accuracy, low latency, and cost efficiency. Backed by a recent Series C led by leading global investors and strategic partners, Deepgram has processed over 50,000 years of audio and transcribed more than 1 trillion words. There is no organization in the world that understands voice better than Deepgram.Company Operating RhythmAt Deepgram, we expect an AI-first mindset—AI use and comfort aren’t optional, they’re core to how we operate, innovate, and measure performance.Every team member who works at Deepgram is expected to actively use and experiment with advanced AI tools, and even build your own into your everyday work. We measure how effectively AI is applied to deliver results, and consistent, creative use of the latest AI capabilities is key to success here. Candidates should be comfortable adopting new models and modes quickly, integrating AI into their workflows, and continuously pushing the boundaries of what these technologies can do.Additionally, we move at the pace of AI. Change is rapid, and you can expect your day-to-day work to evolve just as quickly. This may not be the right role if you’re not excited to experiment, adapt, think on your feet, and learn constantly, or if you’re seeking something highly prescriptive with a traditional 9-to-5.OpportunityDeepgram is looking for an Engineering Manager to lead the design and implementation of Deepgram’s internal data and ML training systems. You will lead a team of engineers in building crucial components for driving a data flywheel that powers fundamental product quality as well as a system for large-scale, distributed ML training on a HPC cluster. Your ability to partner effectively with leaders and ICs in Research and DataOps while leading talented engineers to build, maintain, and extend mission-critical systems will contribute significantly to Deepgram’s competitive advantage. You will be a critical voice in Deepgram’s DataOps, Research, and Engineering teams, driving high impact products from start to finish.What You’ll DoRecruit, hire, train, and support top engineering talent to build a world-class teamTransform cross-functional visions into detailed project plans, ensuring clarity across teams on commitments, risks, and timelinesDefine and own the technical strategy enabling acceleration of Deepgram’s ML training pipelinesPromote a robust team engineering culture, including a focus on rigorous engineering standards as well as continuous improvement on team practices and processesPartner with DataOps and Research to design and implement new services, features, and/or products end to endCoach and mentor engineers to achieve high personal growth while delivering on ambitious team goalsYou’ll Love This Role If YouThrive in a fast-paced, impact-driven environment where learning new skills on-the-fly is not only encouraged but a regular necessityEnjoy balancing decisions about product and feature maturity to decide when to make minimally invasive changes versus when to incorporate detailed design workSee management as an opportunity to empower a team to solve big problems through your own grit, learning, and empathyIt’s Important To Us That You HaveProven experience managing and leading and growing a team with a consistent record of mentoring and coaching team membersA strong technical background building world class solutionsThe ability to drive technical decisions in a scalable and thoughtful mannerPassion about good engineering practices, emerging technologies, and improving processesA passion for navigating the team through the real-world constraints of a startup: how to build quickly, iterate frequently, and run experimentsStrong practical experience with software architecture and implementation, plus the ability to engage with an Engineering team to help build and scale the existing infrastructureIt Would Be Great if You HadExperience working in high-growth startupsFamiliarity with hybrid cloud models (bare-metal datacenters and cloud service providers) and scaled databasesDeep experience with machine learning training, inference, or bothBenefits & Perks*Holistic healthMedical, dental, vision benefitsAnnual wellness stipendMental health supportLife, STD, LTD Income Insurance PlansWork/life blendUnlimited PTOGenerous paid parental leaveFlexible schedule12 Paid US company holidaysQuarterly personal productivity stipendOne-time stipend for home office upgrades401(k) plan with company matchTax Savings ProgramsContinuous learningLearning / Education stipendParticipation in talks and conferencesEmployee Resource GroupsAI enablement workshops / sessions*For candidates outside of the US, we use an Employer of Record model in many countries, which means benefits are administered locally and governed by country-specific regulations. Because of this, benefits will differ by region — in some cases international employees receive benefits US employees do not, and vice versa. As we scale, we will continue to evaluate where we can create more alignment, but a 1:1 global benefits structure is not always legally or operationally possible.Backed by prominent investors including Y Combinator, Madrona, Tiger Global, Wing VC and NVIDIA, Deepgram has raised over $215M in total funding. If you're looking to work on cutting-edge technology and make a significant impact in the AI industry, we'd love to hear from you!Deepgram is an equal opportunity employer. We want all voices and perspectives represented in our workforce. We are a curious bunch focused on collaboration and doing the right thing. We put our customers first, grow together and move quickly. We do not discriminate on the basis of race, religion, color, national origin, gender, sexual orientation, gender identity or expression, age, marital status, veteran status, disability status, pregnancy, parental status, genetic information, political affiliation, or any other status protected by the laws or regulations in the locations where we operate.We are happy to provide accommodations for applicants who need them.
No items found.
2026-03-18 10:32
Forward Deployed Engineer
HappyRobot
51-100
United Kingdom
Full-time
Remote
false
About HappyRobotHappyRobot is the AI-native operating system for the real economy—a system that closes the circuit between intelligence and action. By combining real-time truth, specialized AI workers, and an orchestrating intelligence, we help enterprises run complex, mission-critical operations with true autonomyOur AI OS compounds knowledge, optimizes at every level, and evolves over time. We’re starting with supply chain and industrial-scale operations, where resilience, speed, and continuous improvement matter most—freeing humans to focus on strategy, creativity, and other high-value tasks.
You can learn more about our vision in our Manifesto. HappyRobot has raised $62M to date, including our most recent $44M Series B in September 2025. Our investors include Y Combinator (YC), Andreessen Horowitz (a16z), and Base10—partners who believe in our mission to redefine how enterprises operate. We’re channeling this investment into building a world-class team: people with relentless drive, sharp problem-solving skills, and the passion to push limits in a fast-paced, high-intensity environment. If this resonates, you belong at HappyRobot.Role OverviewWe are looking for a versatile and highly skilled Forward Deployed Engineer to join our team. In this role, you will combine strong technical abilities with excellent communication skills, working directly with customers to ensure they maximize the value of HappyRobot’s AI platform. You will be involved in onboarding, implementation, and ongoing support, while also contributing to product development and innovation.What You’ll DoCustomer-Facing Engineering – Work closely with customers from onboarding to ongoing usage, helping them integrate and optimize our AI solutions.Technical Development – Build new features, MVPs, and scalable solutions that directly impact customer outcomes.Full-Stack Development – Utilize React, TypeScript, Node.js, and Python to develop robust applications and tools.AI/ML Applications – Design, implement, and iterate on AI/ML solutions, including LLM prompting, tuning of voices, and transcribers to optimize use cases.Integration & APIs – Manage APIs and integrations with third-party systems to ensure seamless functionality for customers.Cross-Functional Collaboration – Partner with Product, Engineering, and Customer Success teams to deliver tailored solutions.Iterative Problem-Solving – Continuously iterate and improve AI solutions based on customer feedback and evolving requirements.Project Management – Prioritize and manage multiple projects under tight deadlines while maintaining high-quality results.Must HaveStrong full-stack experience: React, TypeScript, Node.js.Hands-on proficiency in Python.Experience building AI/ML applications, including LLM prompting and tuning.Ability to manage APIs and integrate with third-party systems.Excellent communication skills with the ability to explain technical concepts to non-technical stakeholders.Proven ability to prioritize and manage multiple projects under tight deadlines.Founder mindset: highly independent, takes ownership, and thrives in a fast-paced environment.Why join us?Opportunity to work at a high-growth AI startup, backed by top investors.Rapidly growing and backed by top investors including a16z, Y Combinator, and Base10.Ownership & Autonomy - Take full ownership of projects and ship fast.Top-Tier Compensation - Competitive salary + equity in a high-growth startup.Work With the Best - Join a world-class team of engineers and buildersOur Operating Principles
Extreme OwnershipWe take full responsibility for our work, outcomes, and team success. No excuses, no blame-shifting — if something needs fixing, we own it and make it better. This means stepping up, even when it’s not “your job.” If a ball is dropped, we pick it up. If a customer is unhappy, we fix it. If a process is broken, we redesign it. We don’t wait for someone else to solve it — we lead with accountability and expect the same from those around us.CraftsmanshipPutting care and intention into every task, striving for excellence, and taking deep ownership of the quality and outcome of your work. Craftsmanship means never settling for “just fine.” We sweat the details because details compound. Whether it’s a product feature, an internal doc, or a sales call — we treat it as a reflection of our standards. We aim to deliver jaw-dropping customer experiences by being curious, meticulous, and proud of what we build — even when nobody’s watching.We are “majos”
Be friendly & have fun with your coworkers. Always be genuine & honest, but kind. “Majo” is our way of saying: be a good human. Be approachable, helpful, and warm. We’re building something ambitious, and it’s easier (and more fun) when we enjoy the ride together. We give feedback with kindness, challenge each other with respect, and celebrate wins together without ego.Urgency with Focus
Create the highest impact in the shortest amount of time. Move fast, but in the right direction. We operate with speed because time is our most limited resource. But speed without focus is chaos. We prioritize ruthlessly, act decisively, and stay aligned. We aim for high leverage: the biggest results from the simplest, smartest actions. We’re running a high-speed marathon — not a sprint with no strategy.Talent Density and Meritocracy
Hire only people who can raise the average; ‘exceptional performance is the passing grade.’ Ability trumps seniority. We believe the best teams are built on talent density — every hire should raise the bar. We reward contribution, not titles or tenure. We give ownership to those who earn it, and we all hold each other to a high standard. A-players want to work with other A-players — that’s how we win.First-Principles Thinking
Strip a problem to physics-level facts, ignore industry dogma, rebuild the solution from scratch. We don’t copy-paste solutions. We go back to basics, ask why things are the way they are, and rebuild from the ground up if needed. This mindset pushes us to innovate, challenge stale assumptions, and move faster than incumbents. It’s how we build what others think is impossible.The personal data provided in your application and during the selection process will be processed by Happyrobot, Inc., acting as Data Controller.By sending us your CV, you consent to the processing of your personal data for the purpose of evaluating and selecting you as a candidate for the position. Your personal data will be treated confidentially and will only be used for the recruitment process of the selected job offer.In relation to the period of conservation of your personal data, these will be eliminated after three months of inactivity in compliance with the GDPR and legislation on the protection of personal data.If you wish to exercise your rights of access, rectification, deletion, portability, or opposition in relation to your personal data, you can do so through security@happyrobot.ai, subject to the GDPR.For more information, visit https://www.happyrobot.ai/privacy-policyBy submitting your request, you confirm that you have read and understood this clause and that you agree to the processing of your personal data as described.
No items found.
2026-03-18 10:32
Senior Software Engineer, Agent Infrastructure
Cohere
501-1000
Canada
Full-time
Remote
false
Who are we?Our mission is to scale intelligence to serve humanity. We’re training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, and agents. We believe that our work is instrumental to the widespread adoption of AI.We obsess over what we build. Each one of us is responsible for contributing to increasing the capabilities of our models and the value they drive for our customers. We like to work hard and move fast to do what’s best for our customers.Cohere is a team of researchers, engineers, designers, and more, who are passionate about their craft. Each person is one of the best in the world at what they do. We believe that a diverse range of perspectives is a requirement for building great products.Join us on our mission and shape the future!About the role.We’re building the next generation of agentic AI infrastructure at Cohere. This team sits at the intersection of ML systems, distributed infrastructure, and developer experience, creating the platform that powers autonomous AI agents at scale.You’ll work on hard, forward-looking problems with few established patterns, including secure code execution, agent state management, model routing, identity and authentication, and resource management for long-running agent workflows.This role is a strong fit for someone who combines systems depth with ML intuition. You should be comfortable building reliable infrastructure, thinking through distributed systems tradeoffs, and understanding how emerging agentic capabilities shape platform design.What you’ll work on.Secure execution environments for agent-generated codeIdentity, authentication, and trust boundaries for agentsModel routing and orchestration across different model types and environmentsRate limiting, quotas, and resource management for agent workflowsState management, memory, and filesystem abstractions for agents.
In this role you will:Turn emerging ML research ideas into production-ready infrastructureBuild core platform capabilities for execution, storage, and state managementPrototype and evaluate new technologies, then help decide what should move into productionPartner with research teams to shape infrastructure based on what future agent systems will needYou may be a good fit if you have:Experience building production ML infrastructure with strong systems fundamentalsHands-on work with agentic systems, multi-agent workflows, or agent development frameworksFamiliarity with model routing and LLM provider frameworks across different model types and environmentsExperience with scalable, fault-tolerant distributed systems and KubernetesA track record of moving quickly on prototypes and making good decisions about productionization.
BonusExperience across on-prem, private cloud, and public cloud environmentsFamiliarity with storage systems, embedded databases, or filesystem abstractionsExperience with code execution sandboxes such as gVisor, Firecracker, Kata, or WASM runtimesInterest in emerging ML infrastructure, edge inference, or browser-native modelsOpen-source contributions to LLM or agent infrastructure projectsExperience with identity, workload auth, or capability-based security systems.If some of the above doesn’t line up perfectly with your experience, we still encourage you to apply! We value and celebrate diversity and strive to create an inclusive work environment for all. We welcome applicants from all backgrounds and are committed to providing equal opportunities. Should you require any accommodations during the recruitment process, please submit an Accommodations Request Form, and we will work together to meet your needs.Full-Time Employees at Cohere enjoy these Perks:🤝 An open and inclusive culture and work environment 🧑💻 Work closely with a team on the cutting edge of AI research 🍽 Weekly lunch stipend, in-office lunches & snacks🦷 Full health and dental benefits, including a separate budget to take care of your mental health 🐣 100% Parental Leave top-up for up to 6 months🎨 Personal enrichment benefits towards arts and culture, fitness and well-being, quality time, and workspace improvement🏙 Remote-flexible, offices in Toronto, New York, San Francisco, London and Paris, as well as a co-working stipend✈️ 6 weeks of vacation (30 working days!)
No items found.
2026-03-18 10:32
Software Engineer, Architecture, Reliability, & Compute
Scale AI
5000+
United States
Full-time
Remote
false
Role Overview
Scale’s rapidly growing International Public Sector team is focused on using AI to address critical challenges facing the public sector around the world. Our core work consists of:
Creating custom AI applications that will impact millions of citizens
Generating high-quality training data for national LLMs
Upskilling and advisory services to spread the impact of AI
As a Production AI Ops Lead, you will design and develop the production lifecycle of full-stack AI applications, while supporting end-to-end system reliability, real-time inference observability, sovereign data orchestration, high-security software integration, and the resilient cloud infrastructure required for our international government partners.
At Scale, we’re not just building AI solutions—we’re enabling the public sector to transform their operations and better serve citizens through cutting-edge technology. If you’re ready to shape the future of AI in the public sector and be a founding member of our team, we’d love to hear from you.
You will:
Own the production outcome: Take full accountability for the long-term performance and reliability of AI use cases deployed across international government agencies.
Ensure Full-Stack integrity: Oversee the end-to-end health of the platform, ensuring seamless integration between the AI core and all full-stack components, from APIs to UI, to maintain a responsive and production-ready environment.
Scale the feedback loop: Build automated systems to monitor model performance and data drift across geographically dispersed environments, ensuring the right levels of reliability.
Navigate global compliance: Manage the technical lifecycle within diverse regulatory frameworks.
Incident command: Lead the response for production issues in mission-critical environments, ensuring rapid resolution and building the guardrails to prevent them from happening again.
Bridge the gap: Translate deep technical performance metrics into clear insights for senior international government officials.
Drive product evolution: Partner with our Engineering and ML teams to ensure the lessons learned in the field directly influence the technical architecture and decisions of future use cases.
Ideally, you have:
Experience: 6+ years in a high-impact technical role (SRE, FDE or MLOps) with experience in the public sector.
Global perspective: Familiarity with international government security standards and the complexities of deploying sovereign AI.
System architecture proficiency: Proven experience maintaining production-grade applications with a deep understanding of the full request lifecycle-connecting frontend/API layers to the backend and AI core.
Modern AI Stack expertise: Proficiency in coding and the modern AI infrastructure, including Kubernetes, vector databases, agentic development, and LLM observability tools.
Ownership: You treat every production deployment as your own. You race toward solving hard problems before the customer even sees them.
Reliability: You understand that in the public sector, a model failure may be a risk to public safety or privacy.
Customer communication: The ability to explain to a high-ranking official why the performance of the system has degraded and how we are fixing it.
PLEASE NOTE: Our policy requires a 90-day waiting period before reconsidering candidates for the same role. This allows us to ensure a fair and thorough evaluation of all applicants.
About Us:
At Scale, our mission is to develop reliable AI systems for the world's most important decisions. Our products provide the high-quality data and full-stack technologies that power the world's leading models, and help enterprises and governments build, deploy, and oversee AI applications that deliver real impact. We work closely with industry leaders like Meta, Cisco, DLA Piper, Mayo Clinic, Time Inc., the Government of Qatar, and U.S. government agencies including the Army and Air Force. We are expanding our team to accelerate the development of AI applications.
We believe that everyone should be able to bring their whole selves to work, which is why we are proud to be an inclusive and equal opportunity workplace. We are committed to equal employment opportunity regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability status, gender identity or Veteran status.
We are committed to working with and providing reasonable accommodations to applicants with physical and mental disabilities. If you need assistance and/or a reasonable accommodation in the application or recruiting process due to a disability, please contact us at accommodations@scale.com. Please see the United States Department of Labor's Know Your Rights poster for additional information.
We comply with the United States Department of Labor's Pay Transparency provision.
PLEASE NOTE: We collect, retain and use personal data for our professional business purposes, including notifying you of job opportunities that may be of interest and sharing with our affiliates. We limit the personal data we collect to that which we believe is appropriate and necessary to manage applicants’ needs, provide our services, and comply with applicable laws. Any information we collect in connection with your application will be treated in accordance with our internal policies and programs designed to protect personal data. Please see our privacy policy for additional information.
No items found.
2026-03-18 10:32
No job found
Your search did not match any job. Please try again
