Docker AI Jobs

Discover the latest remote and onsite Docker AI roles across top active AI companies. Updated hourly.

Check out 252 new Docker AI roles opportunities posted on The Homebase

Founding Engineering Lead

New
Top rated
AIFund
Full-time
Full-time
Posted

Own the technical foundation of Meeno end-to-end including web, mobile, backend, data, and experimentation. Co-design product vision in close partnership with Meeno's team. Build core AI product primitives such as voice capture/playback, low-latency interactions, scene framework (content, branching, scoring hooks), feedback loops and user progression, and personalization. Architect systems for speed and iteration with weekly experiments rather than quarterly releases. Set the engineering standards for quality, reliability, security/privacy, and shipping culture. Hire and mentor engineers as the team scales, focusing on quality over quantity and leveraging AI and talent to maintain lean operations.

$180,000 – $220,000
Undisclosed
YEAR

(USD)

New York, United States
Maybe global
Onsite
Python
JavaScript
Prompt Engineering
OpenAI API
MLOps

Founding Platform Engineer

New
Top rated
Netic
Full-time
Full-time
Posted

Design and own the semantic layer that powers the system-of-record flywheel, enabling compounding AI products across teams. Build primitives, abstractions, and APIs for product teams to use as building blocks, ensuring ease of use for shipping AI-driven features. Partner closely with internal product and engineering teams to understand needs, eliminate friction, and design intuitive, well-documented systems that are hard to misuse. Architect systems that span data warehouses, OLTP databases, streaming systems, and vector stores, making tradeoffs based on latency, throughput, consistency, and access patterns. Work with leadership to define the long-term platform architecture, including build-vs-buy decisions, evolving the semantic layer, and scaling the system as product surface area grows.

Undisclosed

()

San Francisco, United States
Maybe global
Onsite
Python
JavaScript
Java
Docker
Kubernetes

2026 New Grad | Software Engineer, Full-Stack

New
Top rated
Loop
Full-time
Full-time
Posted

Ship critical infrastructure managing real-world logistics and financial data for large enterprises. Own the why by building deep context through customer calls and understanding Loop's value to customers, pushing back on requirements if better solutions exist. Work full-stack across system boundaries including frontend UX, LLM agents, database schema, and event infrastructures. Leverage AI tools to handle routine tasks enabling focus on quality, architecture, and product taste. Constantly optimize development loops, refactor legacy patterns, automate workflows, and fix broken processes to raise velocity.

$150,000 – $150,000
Undisclosed
YEAR

(USD)

San Francisco or Chicago or NYC, United States
Maybe global
Hybrid
Python
JavaScript
TypeScript
PyTorch
TensorFlow

New Grad | Software Engineer, AI

New
Top rated
Loop
Full-time
Full-time
Posted

Ship critical infrastructure by managing real-world logistics and financial data for the largest enterprise in the world. Own the why by building deep context through customer calls and understanding Loop’s value to customers, pushing back on requirements if there is a better, faster way to solve problems. Work with full-stack proficiency across system boundaries, from frontend UX to LLM agents, database schema, and event infrastructures. Leverage AI tools to handle the boilerplate work so focus can be on quality, architecture, and product taste. Constantly optimize development loops, refactor legacy patterns, automate workflows, and fix broken processes to raise the velocity bar.

$150,000 – $150,000
Undisclosed
YEAR

(USD)

San Francisco or Chicago or NYC, United States
Maybe global
Hybrid
Python
JavaScript
TypeScript
Hugging Face
Transformers

Software Engineer, Platform Systems

New
Top rated
OpenAI
Full-time
Full-time
Posted

Design and build distributed failure detection, tracing, and profiling systems for large-scale AI training jobs. Develop tooling to identify slow, faulty, or misbehaving nodes and provide actionable visibility into system behavior. Improve observability, reliability, and performance across OpenAI's training platform. Debug and resolve issues in complex, high-throughput distributed systems. Collaborate with systems, infrastructure, and research teams to evolve platform capabilities. Extend and adapt failure detection systems or tracing systems to support new training paradigms and workloads.

Undisclosed

()

London, United Kingdom
Maybe global
Onsite
Python
C++
Docker
Kubernetes
CI/CD

Software Engineer, Platform Systems

New
Top rated
OpenAI
Full-time
Full-time
Posted

Design and build distributed failure detection, tracing, and profiling systems for large-scale AI training jobs. Develop tooling to identify slow, faulty, or misbehaving nodes and provide actionable visibility into system behavior. Improve observability, reliability, and performance across OpenAI's training platform. Debug and resolve issues in complex, high-throughput distributed systems. Collaborate with systems, infrastructure, and research teams to evolve platform capabilities. Extend and adapt failure detection systems or tracing systems to support new training paradigms and workloads.

$310,000 – $460,000
Undisclosed
YEAR

(USD)

San Francisco, United States
Maybe global
Onsite
Python
C++
Docker
Kubernetes
CI/CD

Software Engineer

New
Top rated
AIFund
Full-time
Full-time
Posted

Design, develop, and maintain web applications and backend services that integrate ML-powered features. Collaborate closely with Machine Learning Engineers and Product Managers to understand ML system requirements and translate them into robust software solutions. Build reliable, scalable, and low-latency services that support ML inference, data workflows, and AI-driven user experiences. Use LLMs to build scalable and reliable AI agents. Own the full software development lifecycle: design, implementation, testing, deployment, monitoring, and maintenance. Ensure high standards for code quality, testing, observability, and operational excellence. Troubleshoot production issues and participate in on-call or support rotations when needed. Mentor junior engineers and contribute to technical best practices across teams. Act as a strong cross-functional partner between product, engineering, and ML teams.

Undisclosed

()

San Francisco Bay Area, United States
Maybe global
Hybrid
Python
Docker
Kubernetes
AWS
GCP

Evaluations - Platform Engineer

New
Top rated
Antimetal
Full-time
Full-time
Posted

Own the evaluation stack by building online and offline evaluation pipelines that measure agent quality across ephemeral, voluminous MELT data, code, and unstructured documents, and set metrics defining the experience. Define quality at scale by designing evaluations that capture trajectory quality in production incidents spanning hundreds of services with ephemeral, high-volume, and approximative ground truth, ensuring metrics predict real outcomes. Build platform abstractions for agents by designing core agent architectures and extending internal frameworks such as sub-agents, MCPs, and middleware to enable confident iteration and faster shipping with product, platform, and research teams. Productionize these systems by owning latency, observability, and uptime.

$225,000 – $325,000
Undisclosed
YEAR

(USD)

New York, United States
Maybe global
Onsite
Python
TypeScript
MLOps
MLflow
Docker

Evaluation Engineer

New
Top rated
Elicit
Full-time
Full-time
Posted

The Evaluation Engineer will own the technical foundation of the auto-evaluation systems by building a comprehensive system that runs fast, is easy to use, and supports quickly building new evaluations. Responsibilities include improving the speed of the basic evals infrastructure with minimal latency, designing interfaces suitable for ML engineers, product managers, and customers, and ensuring the system architecture allows team members to easily add examples and run evaluations. The role also involves ensuring evaluations are accurate and reliable by encoding knowledge about how pharma customers make decisions, providing appropriate statistical tests, and confidence intervals for trustworthy results. Additionally, the engineer is expected to spend most time on the core eval platform, collaborate with the evals team on specific evals, mentor an evals engineering intern, and learn how users interact with the evaluation system to improve it.

$140,000 – $200,000
Undisclosed
YEAR

(USD)

Oakland, United States
Maybe global
Hybrid
Python
TypeScript
Docker
CI/CD
AWS

Product Marketing Manager, Public Sector

New
Top rated
Scale AI
Full-time
Full-time
Posted

The role involves translating AI research into product solutions by working with client-side researchers on post-training, evaluations, safety, and alignment, and building necessary primitives, data, and tooling. The candidate will partner deeply with core customers and frontier research labs to address complex technical problems related to model improvement, performance, and deployment. They are expected to shape and propose model improvement work by translating customer and research objectives into clear proposals and execution plans. Responsibilities include leading the end-to-end lifecycle from discovery through shipping initial solutions and scaling pilots, independently managing technical working sessions with senior stakeholders, defining success metrics, surfacing risks, and driving programs to measurable outcomes. The role requires cross-functional collaboration with research, platform, operations, security, and finance teams to deliver production-grade results. Additionally, the candidate will build robust evaluation frameworks, close the loop with data quality and feedback, and share learnings to enhance execution across accounts.

$201,600 – $241,920
Undisclosed
YEAR

(USD)

Washington, United States
Maybe global
Onsite
Python
Prompt Engineering
Model Evaluation
MLOps
MLflow

Want to see more AI Egnineer jobs?

View all jobs

Access all 4,256 remote & onsite AI jobs.

Join our private AI community to unlock full job access, and connect with founders, hiring managers, and top AI professionals.
(Yes, it’s still free—your best contributions are the price of admission.)

Frequently Asked Questions

Need help with something? Here are our most frequently asked questions.

Question text goes here

Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare, eros dolor interdum nulla, ut commodo diam libero vitae erat. Aenean faucibus nibh et justo cursus id rutrum lorem imperdiet. Nunc ut sem vitae risus tristique posuere.

[{"question":"What are Docker AI jobs?","answer":"Docker AI jobs involve developing, deploying, and maintaining AI applications using containerization technology. These positions focus on creating reproducible AI workflows, packaging machine learning models with dependencies, and ensuring consistent execution across environments. Professionals in these roles typically work on MLOps pipelines, containerized AI applications, and implement solutions that seamlessly transition from development to production."},{"question":"What roles commonly require Docker skills?","answer":"Machine Learning Engineers, Data Scientists, AI Developers, and DevOps Engineers working on AI systems commonly require containerization skills. These professionals use containers to package models, ensure reproducibility, and streamline deployment pipelines. Full-stack developers building AI-powered applications and MLOps specialists implementing continuous integration workflows also frequently need proficiency with containerized environments and deployment strategies."},{"question":"What skills are typically required alongside Docker?","answer":"Alongside containerization expertise, employers typically seek proficiency in AI frameworks like TensorFlow, PyTorch, and Hugging Face. Familiarity with Docker Compose for multi-container applications, version control systems, and CI/CD pipelines is essential. Additional valuable skills include YAML configuration, cloud deployment knowledge, GPU acceleration techniques, and experience with MLOps practices that facilitate model development, testing, and production deployment."},{"question":"What experience level do Docker AI jobs usually require?","answer":"AI positions requiring containerization skills typically seek mid-level professionals with 2-4 years of practical experience. Entry-level roles may accept candidates with demonstrated proficiency in basic container commands, Dockerfile creation, and image management. Senior positions often demand extensive experience integrating containers into production ML pipelines, optimizing container resources, and implementing advanced deployment strategies across cloud and edge environments."},{"question":"What is the salary range for Docker AI jobs?","answer":"Compensation for AI professionals with containerization expertise varies based on location, experience level, industry, and additional technical skills. Junior roles typically start at competitive market rates, while senior positions command premium salaries. The most lucrative opportunities combine deep learning expertise, container orchestration experience, and cloud platform knowledge. Specialized industries like finance or healthcare often offer higher compensation for these in-demand skill combinations."},{"question":"Are Docker AI jobs in demand?","answer":"Containerization skills remain highly sought after in AI development, with strong demand driven by organizations implementing MLOps practices and scalable AI deployment strategies. Recent partnerships like Anaconda-Docker and trends in serverless AI containers have intensified hiring needs. The emergence of specialized tools like Docker Model Runner, Docker Offload, and Docker AI Catalog reflects the growing importance of containerized workflows in modern AI development and deployment practices."},{"question":"What is the difference between Docker and Kubernetes in AI roles?","answer":"In AI roles, containerization focuses on packaging individual applications with dependencies for consistent execution, while Kubernetes orchestrates multiple containers at scale. ML engineers might use Docker to create reproducible model environments but implement Kubernetes to manage production deployments across clusters. While containerization handles the model packaging, Kubernetes addresses the scalability, load balancing, and automated recovery needed for production AI systems serving multiple users simultaneously."}]