AWS AI Jobs

Discover the latest remote and onsite AWS AI roles across top active AI companies. Updated hourly.

Check out 352 new AWS AI roles opportunities posted on The Homebase

Product Manager, Models

New
Top rated
Heidi Health
Full-time
Full-time
Posted

As the Product Manager for Heidi's models platform, you will own the product strategy and roadmap for the platform including evaluation pipelines, fine-tuning infrastructure, model routing, and safety systems. Your responsibilities include prioritising your team's work across enablement requests, model safety and quality, and new capability bets; fixing platform issues that cause blocks for product teams; building evaluation tooling and fine-tuning workflows usable in clinical settings; deciding improvements based on clinician feedback, model quality signals, and product team needs; allocating engineering capacity among competing requests and clearly communicating deferrals; working with engineers on evaluation design, fine-tuning trade-offs, and model architecture decisions; setting model quality and safety targets based on clinical outcomes; consolidating duplicate infrastructure across product teams; and monitoring foundation model developments to adjust the roadmap accordingly. You will collaborate closely with engineers, researchers, product PMs, and clinical safety teams and report to product leadership. This is a platform role whose outputs impact every user-facing product at Heidi.

Undisclosed

()

Sydney, Australia
Maybe global
Remote
Python
Model Evaluation
MLOps
MLflow
Docker

Forward Deployed Engineer, Agentic Platform

New
Top rated
Cohere
Full-time
Full-time
Posted

Build and ship features for North, an AI workspace platform; develop autonomous agents that interact with sensitive enterprise data; experiment rapidly and with high quality to engage customers and deliver solutions that exceed expectations; work across the entire product lifecycle from conceptualization through production; lead end-to-end deployment of North in private cloud and on-premises environments including planning, configuration, testing, and rollout.

Undisclosed

()

Middle East
Maybe global
Onsite
Python
RAG
Docker
Kubernetes
AWS

Forward Deployed Engineer - ML

New
Top rated
Modal
Full-time
Full-time
Posted

As a Forward Deployed ML Engineer at Modal, you will work hands-on with companies like Suno, Lovable, Cognition, and Meta to architect and optimize production AI workloads on Modal. You will contribute to open-source projects, publish technical content demonstrating Modal's capabilities across the AI stack, and collaborate with Modal's product and sales teams as both an engineer and a product stakeholder. Additionally, you will build trusted relationships with technical leaders at companies doing frontier AI work and conduct technical demos, experiments, and proof-of-concepts that highlight Modal's performance advantages.

Undisclosed

()

Stockholm, Sweden
Maybe global
Onsite
Python
PyTorch
TensorFlow
MLOps
Docker

Research Product Manager — Structured AI Systems

New
Top rated
Granica
Full-time
Full-time
Posted

The Research Product Manager is responsible for advancing foundational work in tabular data learning, structured and relational representation learning, compression-aware AI, hybrid symbolic, relational, and neural systems, and large-scale systems, linking these research efforts to real production systems managing petabytes of data. The role involves productionizing structured AI models by collaborating with Research and Systems teams to design training on Parquet/Iceberg/Delta data, define training infrastructure requirements, inference architectures, and maintenance loops, while understanding storage and compute trade-offs, data layout, compute scheduling, model lifecycle, infrastructure bottlenecks, and evaluation pipelines. The role also involves defining economic value extraction by identifying buyers, economic value sources, quantification methods, and converting research advances into revenue and platform advantages, requiring strong enterprise infrastructure economic intuition. Additionally, the Research Product Manager identifies viable modeling advances for production, terminates non-viable research directions, defines integration paths into enterprise workloads, and works with the Chief Research Scientist on research agenda prioritization. The position requires deep understanding of large AI model training, deployment, and maintenance in production systems, as well as translating foundational modeling advances into economically valuable infrastructure, shaping technical execution and economic strategy.

$160,000 – $250,000
Undisclosed
YEAR

(USD)

Mountain View, United States
Maybe global
Hybrid
Python
MLflow
MLOps
Data Pipelines
AWS

Solutions Engineer (Autonomous Vehicles & Robotics)

New
Top rated
Encord
Full-time
Full-time
Posted

As a Solutions Engineer at Encord, you will be the core technical expert for customers building autonomous vehicles, robotics, and physical AI solutions, specializing in LiDAR data, sensor fusion, and perception. Your responsibilities include leading technical discovery with perception teams to understand their sensor stacks, model development pipelines, and data challenges; architecting complete solutions for complex multimodal datasets including LiDAR, camera, and radar fusion, and sensor calibration; acting as the technical authority on handling 3D point clouds, sensor fusion, temporal sequences, and multimodal annotation; building bespoke proofs of concept for LiDAR data ingestion, point cloud processing, coordinate transformations, and sensor calibration; developing custom integrations with robotics/AV stacks such as MCAP, ROS, Apollo, and Autoware; creating technical demos for LiDAR annotation, 3D bounding boxes, semantic segmentation, and multi-sensor fusion; debugging complex issues involving point cloud rendering, sensor calibration matrices, and multimodal data synchronization; guiding prospects through technical evaluations of LiDAR formats, sensor configurations, and annotation requirements; providing expert consultation on 3D annotation best practices, coordinate conventions, and quality control workflows; partnering with Account Executives to co-own technical wins in enterprise sales cycles; translating technical capabilities into business value for CTOs and senior stakeholders; and channeling customer feedback to Product and Engineering teams to shape the product roadmap.

Undisclosed

()

San Francisco, United States
Maybe global
Hybrid
Python
Computer Vision
Data Pipelines
AWS
GCP

Senior Fullstack Software Engineer

New
Top rated
Heidi Health
Full-time
Full-time
Posted

Build systems that integrate with the EHRs used in American healthcare to make Heidi feel like a native capability rather than a plugin. Develop systems that simplify the complexity of US healthcare billing, compliance, and payer constraints so clinicians do not have to manage these complexities. Write clean, testable code with strong interfaces, error handling, and observability, ensuring the workflows are reliable for clinicians, operators, and downstream systems. Focus on outcomes by ensuring that the built systems help clinicians and improve practice revenue. Create agentic workflow functionalities where AI assists with extraction, reconciliation, and drafting within workflows, incorporating human review, auditability, and control. Collaborate closely in a team environment with frequent pairing and shared ownership of design and implementation. Learn about healthcare organizational operations, especially those serving US customers, to translate requirements and constraints into product improvements.

$150,000 – $210,000
Undisclosed
YEAR

(USD)

London, United Kingdom
Maybe global
Hybrid
Python
JavaScript
TypeScript
Docker
Kubernetes

Senior Program Manager, Infrastructure Strategy and Business Operations

New
Top rated
Together AI
Full-time
Full-time
Posted

Advance inference efficiency end-to-end by designing and prototyping algorithms, architectures, and scheduling strategies for low-latency, high-throughput inference. Implement and maintain changes in high-performance inference engines, including kernel backends, speculative decoding, and quantization. Profile and optimize performance across GPU, networking, and memory layers to improve latency, throughput, and cost. Design and operate RL and post-training pipelines optimizing algorithms and systems where most cost is inference. Make RL and post-training workloads more efficient with inference-aware training loops and techniques for large-scale rollout collection and evaluation. Use pipelines to train, evaluate, and iterate on frontier models based on the inference stack. Co-design algorithms and infrastructure tightly coupling objectives, rollout collection, and evaluation to efficient inference and quickly identify bottlenecks across training engine, inference engine, data pipeline, and user-facing layers. Run ablations and scale-up experiments to understand trade-offs between model quality, latency, throughput, and cost, and feed insights back into model, RL, and system design. Profile, debug, and optimize inference and post-training services under real production workloads. Drive roadmap items requiring engine modifications including kernels, memory layouts, scheduling logic, and APIs. Establish metrics, benchmarks, and experimentation frameworks to rigorously validate improvements. Provide technical leadership, set technical direction for cross-team efforts intersecting inference, RL, and post-training, and mentor engineers and researchers on full-stack ML systems work and performance engineering.

$200,000 – $280,000
Undisclosed
YEAR

(USD)

San Francisco, United States
Maybe global
Onsite
Python
Reinforcement Learning
MLOps
MLflow
Docker

Helix AI Engineer, Agentic Systems

New
Top rated
Figure AI
Full-time
Full-time
Posted

Design, deploy, and maintain Figure's training clusters. Architect and maintain scalable deep learning frameworks for training on massive robot datasets. Work together with AI researchers to implement training of new model architectures at a large scale. Implement distributed training and parallelization strategies to reduce model development cycles. Implement tooling for data processing, model experimentation, and continuous integration.

$150,000 – $350,000
Undisclosed
YEAR

(USD)

San Jose, United States
Maybe global
Onsite
Python
PyTorch
AWS
Azure
GCP

Software Engineer, Backend

New
Top rated
Mirage
Full-time
Full-time
Posted

Design, build, and own backend systems end-to-end, including services, APIs, data pipelines, and infrastructure that power the products. Solve complex technical challenges across distributed systems, scaling, concurrency, and performance. Integrate and operate large generative AI models in production by deploying, serving, and scaling systems that combine internal research and external capabilities to unlock new product experiences. Instrument, experiment, and iterate in production to continuously improve system and product quality. Design and operate core platform infrastructure, including integrations with third-party providers, storage systems, security, and internal APIs.

$185,000 – $285,000
Undisclosed
YEAR

(USD)

Union Square or New York, United States
Maybe global
Onsite
Python
JavaScript
Java
Go
Docker

Lazo - Head of Engineering

New
Top rated
Silver.dev
Full-time
Full-time
Posted

The Head of Engineering at Lazo is responsible for owning the technology strategy and roadmap aligned with business and product OKRs, defining the reference architecture for agentic systems including LLMs and tool orchestration, establishing security and compliance baselines such as IAM, data privacy, and SOC2-readiness, and managing cost governance (FinOps). They present trade-offs, risks, and progress in leadership reviews. The role involves hands-on engineering and delivery, including shipping backend services in Python/TypeScript, orchestrating agents and toolchains, integrating external APIs and databases, building robust pipelines, and handling end-to-end DevOps using AWS/GCP, containerization, IaC, CI/CD, and observability, as well as on-call design. They work to reduce technical debt, improve latency and throughput, and manage infrastructure cost per workflow/client. Responsibilities also include defining SLOs and error budgets to reduce MTTR and change-fail rates, implementing data access policies and secure data flows for AI features, driving post-mortems and preventive engineering, hiring and mentoring engineers, setting performance scorecards, fostering a culture of thoughtful trade-offs and fast feedback loops, partnering with Product and AI teams for scalable solutions, collaborating with Ops, Growth, and Customer teams for reliability and launch readiness, and managing key vendors and build-versus-buy decisions with ROI narratives.

$72,000 – $96,000
Undisclosed
YEAR

(USD)

Argentina
Maybe global
Remote
Python
TypeScript
Docker
AWS
GCP

Want to see more AI Egnineer jobs?

View all jobs

Access all 4,256 remote & onsite AI jobs.

Join our private AI community to unlock full job access, and connect with founders, hiring managers, and top AI professionals.
(Yes, it’s still free—your best contributions are the price of admission.)

Frequently Asked Questions

Need help with something? Here are our most frequently asked questions.

Question text goes here

Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare, eros dolor interdum nulla, ut commodo diam libero vitae erat. Aenean faucibus nibh et justo cursus id rutrum lorem imperdiet. Nunc ut sem vitae risus tristique posuere.

[{"question":"What are AWS AI jobs?","answer":"AWS AI jobs involve building, training, and deploying generative AI applications using specialized cloud services. These roles work with tools like SageMaker for custom model development, Bedrock for foundation models, and Lake Formation for data governance. Professionals in these positions create AI-driven applications, implement RAG systems with Kendra, and orchestrate machine learning pipelines using Step Functions and Lambda."},{"question":"What roles commonly require AWS skills?","answer":"Common roles requiring AWS skills include machine learning engineers, data scientists, software engineers, architects, and platform engineers. These professionals work on generative AI applications and AI-assisted development lifecycles. They implement end-to-end ML pipelines in SageMaker, design LLM-powered applications with Bedrock, create agentic workflows, and build AI-enhanced developer tools using Amazon Q Developer."},{"question":"What skills are typically required alongside AWS?","answer":"Alongside AWS expertise, professionals typically need experience with JupyterLab, Git, and IDE integrations like VS Code. Knowledge of LangChain for LLM orchestration, machine learning concepts, and data engineering practices are valuable. Familiarity with generative AI patterns like retrieval-augmented generation, prompt engineering, and AI application development workflows helps create effective solutions within the AWS ecosystem."},{"question":"What experience level do AWS AI jobs usually require?","answer":"AWS AI jobs typically require mid to senior-level experience with cloud infrastructure and AI development patterns. Employers look for professionals familiar with JupyterLab environments, ML workflows in SageMaker, and foundation model deployment via Bedrock. Experience building end-to-end machine learning pipelines, implementing RAG systems, and orchestrating AI workflows using Step Functions and Lambda is highly valued."},{"question":"What is the salary range for AWS AI jobs?","answer":"AWS AI job salaries vary based on experience, location, and specific role. Machine learning engineers and data scientists implementing SageMaker solutions generally command premium compensation. Platform engineers orchestrating AI infrastructure and architects designing generative AI applications often receive higher salaries. Software engineers using Amazon Q for AI-assisted development are increasingly valued for their productivity enhancements."},{"question":"Are AWS AI jobs in demand?","answer":"AWS AI jobs are experiencing strong demand as organizations adopt generative AI technologies. Companies are actively hiring professionals who can implement AI-driven development lifecycles using tools like Amazon Q Developer. There's particular demand for engineers who can work with Bedrock for foundation models, build RAG systems with Kendra, and design agentic workflows for business process automation."},{"question":"What is the difference between AWS and Azure in AI roles?","answer":"The key difference in AI roles is that AWS emphasizes fully managed services like Bedrock for foundation models and SageMaker for end-to-end ML workflows, while Azure offers a different ecosystem through Azure AI services. AWS positions focus more on serverless orchestration and agentic capabilities unique to their toolchain. The platforms have distinct approaches to generative AI implementation, with different service integrations and developer experiences."}]