Find AI Work That Works for You
Latest roles in AI and machine learning, reviewed by real humans for quality and clarity.
I'm strong in:
Edit filters
New AI Opportunities
Showing 61 – 79 of 79 jobs
Tag
Intern of Technical Staff - Sovereign AI
Cohere
501-1000
Canada
Full-time
Remote
false
Who are we?Our mission is to scale intelligence to serve humanity. We’re training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, and agents. We believe that our work is instrumental to the widespread adoption of AI.We obsess over what we build. Each one of us is responsible for contributing to increasing the capabilities of our models and the value they drive for our customers. We like to work hard and move fast to do what’s best for our customers.Cohere is a team of researchers, engineers, designers, and more, who are passionate about their craft. Each person is one of the best in the world at what they do. We believe that a diverse range of perspectives is a requirement for building great products.Join us on our mission and shape the future!As a Sovereign AI Intern, you will:Design, train and improve upon cutting-edge models to serve public interest.Help us develop new techniques to train and serve models safer, better, and faster.Train extremely large-scale models on massive datasets.Learn from experienced senior machine learning technical staff.Work closely with product teams to develop solutions.You may be a good fit if you have:Proficiency in Python and related ML frameworksExperience using large-scale distributed training strategies.Strong communication and problem-solving skills.Bonus: Canadian citizenshipBonus: papers at top-tier venues (such as NeurIPS, ICML, ICLR, AIStats, MLSys, JMLR, AAAI, Nature, COLING, ACL, EMNLP).If some of the above doesn’t line up perfectly with your experience, we still encourage you to apply! We value and celebrate diversity and strive to create an inclusive work environment for all. We welcome applicants from all backgrounds and are committed to providing equal opportunities. Should you require any accommodations during the recruitment process, please submit an Accommodations Request Form, and we will work together to meet your needs.Full-Time Employees at Cohere enjoy these Perks:🤝 An open and inclusive culture and work environment 🧑💻 Work closely with a team on the cutting edge of AI research 🍽 Weekly lunch stipend, in-office lunches & snacks🦷 Full health and dental benefits, including a separate budget to take care of your mental health 🐣 100% Parental Leave top-up for up to 6 months🎨 Personal enrichment benefits towards arts and culture, fitness and well-being, quality time, and workspace improvement🏙 Remote-flexible, offices in Toronto, New York, San Francisco, London and Paris, as well as a co-working stipend✈️ 6 weeks of vacation (30 working days!)
No items found.
2026-01-16 5:59
Applied AI Engineer - Agentic Workflows (Singapore)
Cohere
501-1000
Singapore
Full-time
Remote
false
Who are we?Our mission is to scale intelligence to serve humanity. We’re training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, and agents. We believe that our work is instrumental to the widespread adoption of AI.We obsess over what we build. Each one of us is responsible for contributing to increasing the capabilities of our models and the value they drive for our customers. We like to work hard and move fast to do what’s best for our customers.Cohere is a team of researchers, engineers, designers, and more, who are passionate about their craft. Each person is one of the best in the world at what they do. We believe that a diverse range of perspectives is a requirement for building great products.Join us on our mission and shape the future!Why this role?We’re a fast-growing startup building production-grade AI agents for enterprise customers at scale. We’re looking for Applied AI Engineers who can own the design, build, and deployment of agentic workflows powered by Large Language Models (LLMs)—from early prototypes to production-grade AI agents, to deliver concrete business value in enterprise workflows.In this role, you’ll work closely with customers on real-world business problems, often building first-of-their-kind agent workflows that integrate LLMs with tools, APIs, and data sources. While our pace is startup-fast, the bar is enterprise-high: agents must be reliable, observable, safe, and auditable from day one.You’ll collaborate closely with customers, product, and platform teams, and help shape how agentic systems are built, evaluated, and deployed at scale.What You’ll DoWork with enterprise customers and internal teams to turn business workflows into scalable, production-ready agentic AI systems.Design and build LLM-powered agents that reason, plan, and act across tools and data sources with enterprise-grade reliability.Balance rapid iteration with enterprise requirements, evolving prototypes into stable, reusable solutions.Define and apply evaluation and quality standards to measure success, failures, and regressions.Debug real-world agent behavior and systematically improve prompts, workflows, tools, and guardrails.Contribute to shared frameworks and patterns that enable consistent delivery across customers.Required Skills & ExperienceBachelor’s degree in Computer Science or a related technical field.Strong programming skills in Python and/or JavaScript/TypeScript.3+ years of experience building and shipping production software; 2+ years working with LLMs or AI APIs.Hands-on experience with modern LLMs (e.g., GPT, Claude, Gemini), vector databases, and agent/orchestration frameworks (e.g., LangChain, LangGraph, LlamaIndex, or custom solutions).Practical experience with RAG, agent workflows, evaluation, and performance optimization.Strong agent design skills, including prompt engineering, tool use, multi-step agent workflows (e.g. ReAct), and failure handling.Ability to reason about and balance trade-offs between customization and reuse, as well as autonomy, control, cost, latency, and risk.Strong communication skills and experience leading technical discussions with customers or partners.Nice-to-HaveExperience working in a fast-moving startup environment.Prior work delivering AI or automation solutions to enterprise customers.Familiarity with human-in-the-loop workflows, fine-tuning, or LLM evaluation techniques.Experience with cloud deployment and production operations for AI systems.Background in applied ML, NLP, or decision systems.Additional RequirementsStrong written and verbal communication skills.Ability and interest to travel up to 25%, flexible.Why Join UsBuild production-grade AI agents used in real enterprise workflows.Operate at scale while retaining end-to-end ownership.Work on hard problems in agent design, evaluation, and reliability.Shape shared platforms and standards, not just individual features.Move fast with a high bar for quality, safety, and reliability.If some of the above doesn’t line up perfectly with your experience, we still encourage you to apply! We value and celebrate diversity and strive to create an inclusive work environment for all. We welcome applicants from all backgrounds and are committed to providing equal opportunities. Should you require any accommodations during the recruitment process, please submit an Accommodations Request Form, and we will work together to meet your needs.Full-Time Employees at Cohere enjoy these Perks:🤝 An open and inclusive culture and work environment 🧑💻 Work closely with a team on the cutting edge of AI research 🍽 Weekly lunch stipend, in-office lunches & snacks🦷 Full health and dental benefits, including a separate budget to take care of your mental health 🐣 100% Parental Leave top-up for up to 6 months🎨 Personal enrichment benefits towards arts and culture, fitness and well-being, quality time, and workspace improvement🏙 Remote-flexible, offices in Toronto, New York, San Francisco, London and Paris, as well as a co-working stipend✈️ 6 weeks of vacation (30 working days!)
No items found.
2026-01-16 5:59
Early Career AI/ML Engineer
Brain Co
1-10
United States
Full-time
Remote
false
About Brain Co.Brain Co. is an Applied AI startup founded by Elad Gil and Jared Kushner, and backed by many of Silicon Valley’s leading builders — including Patrick Collison (CEO of Stripe), Andrej Karpathy (Cofounder of OpenAI), Mike Krieger (CPO of Anthropic), Kevin Weil (CPO of OpenAI), and Aravind Srinivas (CEO of Perplexity).We are building an AI platform and applications for the world’s most important institutions - delivering impact on real-world problems.Our progress so farAutomated construction permitting for a sovereign government → 80% faster, unlocking $375M+ in valueOptimized supply chains for a leading global energy company → 30% lower cost, 99% reliability, preventing $100M+ in lossesStreamlined hospital patient care across national health systems → 40% better outcomes, 80% less admin workRaised a $30M Series A from top investorsBuilt a team of 40+ AI experts from Tesla, Google DeepMind, NVIDIA, and DatabricksAt Brain Co., your work will be deployed in the real world, not stuck in research. We move fast, with more demand than we can serve, and are looking for exceptional people to take ownership from day one.About The RoleAs an AI/ML Engineer at Brain Co., you will play a crucial role in deploying state-of-the-art models to automate various real world problems in sectors such as healthcare, government and energy. Part of the role will involve turning research breakthroughs into practical solutions for various nation states. This role is your opportunity to make a significant impact by making AI technology both accessible and influential.In This Role, You Will:Innovate and Deploy: Design and deploy advanced LLM models to tackle real-world problems, particularly in automating complex, manual processes in a range of real-world verticals.Optimize and Scale: Build scalable data pipelines, optimize models for performance and accuracy, and prepare them for production. Monitor and maintain deployed models to ensure they continue delivering value across various governments worldwide.Make a Difference: Engage in projects including but not limited to optimizing the world's most advanced energy production systems, modernizing core government workflows, or improving patient outcomes in advanced public healthcare systems. Your work will directly impact how AI benefits individuals, businesses, and society at large.Engage with Leaders: interact directly with government officials in various countries and apply the first of its kind AI solutions while working alongside experienced ex. Founders, AI researchers, and software engineers to understand complex business challenges and deliver AI-powered solutions. Join a dynamic team where ideas are exchanged freely and creativity flourishes. You will be able to wear many hats: software building, product management, sales, interpersonal skills.Learn and Lead: Keep abreast of the latest developments in machine learning and AI. Participate in code reviews, share knowledge, and set an example with high-quality engineering practices.You Might Thrive In This Role If You:Have 0-2 years of industry experience in applied machine learning or related AI work.Hold a BSc/Master’s/PhD degree in Computer Science, Machine Learning, Data Science, or a related field.Have hands-on experience building GenAI-focused applications (e.g., agents, reasoning workflows, or RAG) and a solid understanding of how large language models are architected and operated.Have personally implemented models in common ML frameworks such as PyTorch, Jax or TensorFlow.Possess a strong foundation in data structures, algorithms, and software engineering principles.Exhibit excellent problem-solving and analytical skills, with a proactive approach to challenges.Can work collaboratively with cross-functional teams.Thrive in fast-paced environments where priorities or deadlines may compete.Are eager to own problems end-to-end and willing to acquire any necessary knowledge to get the job done.BenefitsCompetitive salaryMedical, Dental, and Vision (100% Coverage)Paid Maternity and Paternity Leave401(k)Daily LunchesCommuter BenefitsUnlimited PTOWhy Join UsShip quickly, iterate constantly and see your work deployed at global scaleCollaborate with industry veterans from Tesla, DeepMind, Databricks, and moreAccelerate your career with ownership based on impact, not tenureEarn competitive compensation + meaningful equity in a high-growth companyThrive in a culture built on speed, curiosity, and impactIf you want to see your work deployed at scale with real impact, Brain Co. is the place to build.
No items found.
2026-01-16 4:14
Forward Deployed Engineer (FDE), Life Sciences - London
OpenAI
5000+
United Kingdom
Full-time
Remote
false
About the teamOpenAI’s Forward Deployed Engineering team partners with global pharma and biotech, CROs, and research institutions to deploy existing expertise across the R&D value chain to help customers design and ship production-grade AI systems. We operate at the intersection of customer delivery and core platform development, converting early deployments into repeatable system standards and evaluation practices that scale across regulated environments.About the roleWe are hiring a Forward Deployed Engineer (FDE) to push the frontier on what is possible today across drug discovery (e.g., target identification, molecular design, pre-clinical) and development (e.g., trial design, trial ops, biostats) by leading end-to-end deployments of our models inside life sciences organizations and research institutions. You will work with customers who are deep experts in their scientific or operational domains, translating real-world data, infrastructure, and constraints into production systems.You will measure success through production adoption, measurable workflow impact, and eval-driven feedback loops, including evaluation benchmarks and acceptance criteria, that inform product and model roadmaps. You’ll work closely with our Product, Research, Partnerships, GRC, Security, and GTM to deliver in regulated contexts, including inspection readiness with audit trails and traceable evidence.This role is based in London. We use a hybrid work model of 3 days in the office per week. We offer relocation assistance. Travel up to 50% is required.In this role you willDesign and ship production systems around models, owning integrations, data provenance, reliability, and on-call readiness across research, clinical, and operational workflows.Lead discovery and scoping from pre-sales through post-sales, translating ambiguous workflow needs into hypothesis-driven problem framing, system requirements, and an execution plan with measurable endpoints.Define and enforce launch criteria for regulated contexts, including validation evidence, audit readiness, outcome metrics, and drive delivery until we demonstrate sustained production impact.Build in sensitive scientific data environments where auditability, validation, and access controls shape architecture, operating procedures, and failure handling.Run evaluation loops that measure model and system quality against workflow-specific scientific benchmarks and use results to drive model and product changes.Distill deployment learnings into hardened primitives, reference architectures, validation templates, and benchmark harnesses that scale across regulated life sciences environments.You might thrive in this role if youBring 5+ years of software/ML engineering or technical deployment experience with customer-facing ownership in biotech, pharma, clinical research, or scientific software; PhD, MS, or equivalent applied experience in a life sciences relevant field encouraged.Have owned customer GenAI deployments end-to-end from scoping through production adoption, and improved them through evaluation design, error analysis, and iterative evidence generation that tightens acceptance criteria over time.Have delivered AI systems in trial design, regulatory writing, or scientific operations where validation strategy, auditability, compliance constraints, and reviewer expectations shaped system design and rollout.Communicate clearly across scientific, clinical, model research, technical, and executive audiences, translating technical tradeoffs into decision quality, risk posture, and measurable outcomes with credibility.Apply systems thinking with high execution standards, consistently turning failures, escalations, and audit findings into improved operating standards, validation artifacts, and repeatable deployment playbooks.About OpenAIOpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity. We are an equal opportunity employer, and we do not discriminate on the basis of race, religion, color, national origin, sex, sexual orientation, age, veteran status, disability, genetic information, or other applicable legally protected characteristic. For additional information, please see OpenAI’s Affirmative Action and Equal Employment Opportunity Policy Statement.Background checks for applicants will be administered in accordance with applicable law, and qualified applicants with arrest or conviction records will be considered for employment consistent with those laws, including the San Francisco Fair Chance Ordinance, the Los Angeles County Fair Chance Ordinance for Employers, and the California Fair Chance Act, for US-based candidates. For unincorporated Los Angeles County workers: we reasonably believe that criminal history may have a direct, adverse and negative relationship with the following job duties, potentially resulting in the withdrawal of a conditional offer of employment: protect computer hardware entrusted to you from theft, loss or damage; return all computer hardware in your possession (including the data contained therein) upon termination of employment or end of assignment; and maintain the confidentiality of proprietary, confidential, and non-public information. In addition, job duties require access to secure and protected information technology systems and related data security obligations.To notify OpenAI that you believe this job posting is non-compliant, please submit a report through this form. No response will be provided to inquiries unrelated to job posting compliance.We are committed to providing reasonable accommodations to applicants with disabilities, and requests can be made via this link.OpenAI Global Applicant Privacy PolicyAt OpenAI, we believe artificial intelligence has the potential to help people solve immense global challenges, and we want the upside of AI to be widely shared. Join us in shaping the future of technology.
No items found.
2026-01-16 2:44
Technical Program Management Intern, AI Tools Development
Zoox
1001-5000
$6,500 – $7,500 / month
United States
Intern
Remote
false
Zoox is transforming mobility with fully autonomous, electric vehicles designed from the ground up for a driverless future. Our mission is to make transportation safer, more sustainable, and accessible to everyone. At Zoox, innovation, collaboration, and a bold vision for the future drive everything we do.
Zoox’s internship program offers hands-on experience with cutting-edge technology, mentorship from some of the industry’s brightest minds, and the opportunity to make meaningful contributions to real projects. We seek interns who demonstrate strong academic performance, engagement beyond the classroom, intellectual curiosity, and a genuine interest in Zoox’s mission.
This internship opportunity is within the Manufacturing Operations Program Management Team, which focuses on Technical Program Management across all teams within the department. Interns on this team will have the chance to work on Creating AI Tools for MOPS Teams, gaining valuable experience and insight into Automation and AI.
In this Role, You Will:• Implement AI-powered tools including LLMs and Copilots to enhance efficiency and automation for Manufacturing Operations teams.• Analyze operational data to inform tool development and measure impact.• Build and improve AI solutions based on user feedback and requirements.• Document technical implementations and create user guides for applied tools.• Present project progress and results to team leadership and stakeholders.• Contribute to the team's AI/automation strategy and best practices.Program RequirementsCurrently pursuing a B.S., M.S., or other advanced degree in Computer Science or relevant engineering field.Plan to return to school after completing the internship.Maintain good academic standing.Available to commit to a 12-week internship beginning in May or June 2026.At least one prior industry internship, co-op, or relevant project experience.Willing and able to relocate to the Bay Area, California, for the duration of the internship.Interns may not use proprietary Zoox information in university theses, publications, or share it outside of Zoox.QualificationsFamiliarity with at least one programming language (Python, JavaScript, Java, C++).Basic understanding of AI/ML concepts through coursework or self-study.Familiarity with data analysis, process optimization, or workflow automation tools.Self-motivated, eager to learn, and able to work effectively in a fast-paced environment.Strong problem-solving and analytical thinking skills.Good written and verbal communication skills.Bonus QualificationsExperience building AI-powered applications or tools.Familiarity with LLMs, prompt engineering, or AI agent frameworks.Previous internship experience in technical program management or software development.Understanding of agile development methodologies.Experience with version control systems (Git).Compensation:The monthly salary range for this position is $6,500 to $7,500. Compensation will vary based on geographic location and level of education. Additional benefits may include medical insurance, and a housing stipend (relocation assistance will be offered based on eligibility).
About ZooxZoox is developing the first ground-up, fully autonomous vehicle fleet and the supporting ecosystem required to bring this technology to market. Sitting at the intersection of robotics, machine learning, and design, Zoox aims to provide the next generation of mobility-as-a-service in urban environments. We’re looking for top talent that shares our passion and wants to be part of a fast-moving and highly execution-oriented team.
Follow us on LinkedIn
AccommodationsIf you need an accommodation to participate in the application or interview process please reach out to accommodations@zoox.com or your assigned recruiter.
A Final Note:You do not need to match every listed expectation to apply for this position. Here at Zoox, we know that diverse perspectives foster the innovation we need to be successful, and we are committed to building a team that encompasses a variety of backgrounds, experiences, and skills.
No items found.
2026-01-15 17:59
Member of Technical Staff, MLE [Singapore]
Cohere
501-1000
Singapore
Full-time
Remote
false
Who are we?Our mission is to scale intelligence to serve humanity. We’re training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, and agents. We believe that our work is instrumental to the widespread adoption of AI.We obsess over what we build. Each one of us is responsible for contributing to increasing the capabilities of our models and the value they drive for our customers. We like to work hard and move fast to do what’s best for our customers.Cohere is a team of researchers, engineers, designers, and more, who are passionate about their craft. Each person is one of the best in the world at what they do. We believe that a diverse range of perspectives is a requirement for building great products.Join us on our mission and shape the future!Why This Role Is DifferentThis is not a typical “Applied Scientist” or “ML Engineer” role. As a Member of Technical Staff, Applied ML, you will:Work directly with enterprise customers on problems that push LLMs to their limits.
You’ll rapidly understand customer domains, design custom LLM solutions, and deliver production-ready models that solve high-value, real-world problems.Train and customize frontier models — not just use APIs.
You’ll leverage Cohere’s full stack: CPT, post-training, retrieval + agent integrations, model evaluations, and SOTA modeling techniques.Influence the capabilities of Cohere’s foundation models.
Techniques, datasets, evaluations, and insights you develop for customers will directly shape the next generation of Cohere’s frontier models.Operate with an early-startup level of ownership inside a frontier-model company.
This role combines the breadth of an early-stage CTO with the infrastructure and scale of a deep-learning lab.Wear multiple hats, set a high technical bar, and define what Applied ML at Cohere becomes.
Few roles in the industry combine application, research, customer-facing engineering, and core-model influence as directly as this one.What You’ll DoTechnical Leadership & Solution DesignContribute to the design and delivery of custom LLM solutions for enterprise customers.Translate ambiguous business problems into well-framed ML problems with clear success criteria and evaluation methodologies.Modeling, Customization & Foundations ContributionBuild custom models using Cohere’s foundation model stack, CPT recipes, post-training pipelines (including RLVR), and data assets.Develop SOTA modeling techniques that directly enhance model performance for customer use-cases.Contribute improvements back to the foundation-model stack — including new capabilities, tuning strategies, and evaluation frameworks.Customer-Facing Technical ImpactWork as part of Cohere’s customer facing MLE team to identify high-value opportunities where LLMs can unlock transformative impact to our enterprise customers.You May Be a Good Fit If You Have:Technical FoundationsStrong ML fundamentals and the ability to frame complex, ambiguous problems as ML solutions.Fluency with Python and core ML/LLM frameworks.Experience working with (or the ability to learn) large-scale datasets and distributed training or inference pipelines.Understanding of LLM architectures, tuning techniques (CPT, post-training), and evaluation methodologies.Demonstrated ability to meaningfully shape LLM performance.Experience & LeadershipA broad view of the ML research landscape and a desire to push the state of the art.MindsetBias toward action, high ownership, and comfort with ambiguity.Humility and strong collaboration instincts.A deep conviction that AI should meaningfully empower people and organizations.Join UsThis is a pivotal moment in Cohere’s history. As an MTS in Applied ML, you will define not only what we build — but how the world experiences AI. If you're excited about building custom models, solving generational problems for global organizations, and shaping frontier-model capabilities, we’d love to meet you.If some of the above doesn’t line up perfectly with your experience, we still encourage you to apply! We value and celebrate diversity and strive to create an inclusive work environment for all. We welcome applicants from all backgrounds and are committed to providing equal opportunities. Should you require any accommodations during the recruitment process, please submit an Accommodations Request Form, and we will work together to meet your needs.Full-Time Employees at Cohere enjoy these Perks:🤝 An open and inclusive culture and work environment 🧑💻 Work closely with a team on the cutting edge of AI research 🍽 Weekly lunch stipend, in-office lunches & snacks🦷 Full health and dental benefits, including a separate budget to take care of your mental health 🐣 100% Parental Leave top-up for up to 6 months🎨 Personal enrichment benefits towards arts and culture, fitness and well-being, quality time, and workspace improvement🏙 Remote-flexible, offices in Toronto, New York, San Francisco, London and Paris, as well as a co-working stipend✈️ 6 weeks of vacation (30 working days!)
No items found.
2026-01-15 5:59
AI Engineer
Ironclad
501-1000
$150,000 – $165,000
United States
Full-time
Remote
false
Ironclad is the leading AI contracting platform that transforms agreements into assets. Contracts move faster, insights surface instantly, and agents push work forward, all with you in control. Whether you’re buying or selling, Ironclad unifies the entire process on one intelligent platform, providing leaders with the visibility they need to stay one step ahead. That’s why the world’s most transformative organizations, from OpenAI to the World Health Organization and the Associated Press, trust Ironclad to accelerate their business.
We’re consistently recognized as a leader in the industry: a Leader in the Forrester Wave and Gartner Magic Quadrant for Contract Lifecycle Management, a Fortune Great Place to Work, and one of Fast Company’s Most Innovative Workplaces. Ironclad has also been named to Forbes’ AI 50 and Business Insider’s list of Companies to Bet Your Career On. We’re backed by leading investors including Accel, Y Combinator, Sequoia, BOND, and Franklin Templeton. For more information, visit www.ironcladapp.com or follow us on LinkedIn.
This is a hybrid role. Office attendance is required at least twice a week on Tuesdays and Thursdays for collaboration and connection. There may be additional in-office days for team or company events.AI Engineering @ IroncladIronclad is accelerating its AI initiatives to transform how legal teams manage contracts. We're looking for a AI/ML Engineer to help shape the future of legal AI. In this role, you’ll work with cutting-edge tools such as HuggingFace, OpenAI APIs, and Rivet to develop state-of-the-art models and intelligent systems that extract structured insights from contracts and power product features used by leading legal teams.You’ll partner closely with product managers and designers to translate user needs into AI-powered product capabilities. You’ll also lead efforts in model evaluation, iteration, and deployment—ensuring our systems are robust, explainable, and continuously improving. As part of our end-to-end ownership model, you’ll contribute to the MLOps stack to ensure our models are scalable, reliable, and performant in production.What you'll be doing:Build and Tune models and AI capabilities that ingest legal documents, extract key information, and deliver actionable insights to customersImplementing backend services that power document understanding, classification, and natural language searchExplore and apply cutting-edge techniques in NLP, LLM integration, Prompt / Context Engineering, and semantic search to advance Ironclad’s AI/ML visionDesign and maintain robust systems for prompt management, model training / tuning, orchestration, evaluation & inference, with a focus on scalability, reliability, and observabilityDevelop and scale retrieval-augmented generation (RAG) systems and agent-based orchestration frameworks for complex legal workflowsHelp build foundational AI/ML infrastructure, including tools for experimentation, model versioning, and online/offline performance monitoringPartner closely with product managers and designers to shape user-facing AI experiences and deliver value through APIs and intelligent featuresRequired skills:BS degree in Computer Science, Engineering, or a related fieldExperience working with large language models (e.g., GPT, Claude) or building RAG systemsExperience with embedding models and vector databasesExperience in training, evaluating, and serving machine learning modelsHands-on experience with Python, backend service development, and RESTful APIsExperience in NLP techniques such as text classification, entity extraction, document understanding, or question answeringProficiency with AI frameworks and APIs (e.g., PyTorch, TensorFlow, HuggingFace , OpenAI , LangChain)Experience with cloud platforms (e.g., GCP, AWS, Azure)Solid understanding of Git or other version control systemsStrong engineering fundamentals and attention to code quality, testing, and maintainabilityExcellent communication and collaboration skillsPassion for building AI products that solve real-world customer problemsSNice to have:Familiarity with Vertex AI or other managed ML platformsExperience working with PDF and DOCX parsing for document ingestionComfort with Docker and containerized developmentFamiliarity with CI/CD systems (e.g., GitHub Actions, CircleCI)Monitoring and observability with tools like Datadog or equivalentExperience with data processing and analysis using tools like SQL, Pandas, NumPy, or PySparkBase Salary Range: $150,000 - $165,000The base salary range represents the minimum and maximum of the salary range for this position based at our San Francisco headquarters. The actual base salary offered for this position will depend on numerous factors, including individual proficiency, anticipated performance, and the location of the selected candidate. Our base salary is just one component of Ironclad’s competitive total rewards package, which also includes equity awards (a new hire grant, along with opportunities for additional awards throughout your tenure), competitive health and wellness benefits, and a commitment to career growth and development.
US Employee Benefits at Ironclad:100% health coverage for employees (medical, dental, and vision), and 75% coverage for dependents with buy-up plan options availableMarket-leading leave policies, including gender-neutral parental leave and compassionate leaveFamily forming support through Maven for you and your partnerPaid time off - take the time you need, when you need itMonthly stipends for wellbeing, hybrid work, and (if applicable) cell phone useMental health support through Modern Health, including therapy, coaching, and digital toolsPre-tax commuter benefits (US Employees)401(k) plan with Fidelity with employer match (US Employees)Regular team events to connect, recharge, and have funAnd most importantly: the opportunity to help build the company you want to work at**UK Employee-specific benefits are included on our UK job postingsPursuant to the San Francisco Fair Chance Ordinance, we will consider for employment qualified applicants with arrest and conviction records.
No items found.
2026-01-15 4:44
Forward Deployed Engineer (FDE), Life Sciences - NYC
OpenAI
5000+
$220,000 – $280,000
United States
Full-time
Remote
false
About the teamOpenAI’s Forward Deployed Engineering team partners with global pharma and biotech, CROs, and research institutions to deploy existing expertise across the R&D value chain to help customers design and ship production-grade AI systems. We operate at the intersection of customer delivery and core platform development, converting early deployments into repeatable system standards and evaluation practices that scale across regulated environments.About the roleWe are hiring a Forward Deployed Engineer (FDE) to push the frontier on what is possible today across drug discovery (e.g., target identification, molecular design, pre-clinical) and development (e.g., trial design, trial ops, biostats) by leading end-to-end deployments of our models inside life sciences organizations and research institutions. You will work with customers who are deep experts in their scientific or operational domains, translating real-world data, infrastructure, and constraints into production systems.You will measure success through production adoption, measurable workflow impact, and eval-driven feedback loops, including evaluation benchmarks and acceptance criteria, that inform product and model roadmaps. You’ll work closely with our Product, Research, Partnerships, GRC, Security, and GTM to deliver in regulated contexts, including inspection readiness with audit trails and traceable evidence.This role is based in NYC. We use a hybrid work model of 3 days in the office per week. We offer relocation assistance. Travel up to 50% is required.In this role you willDesign and ship production systems around models, owning integrations, data provenance, reliability, and on-call readiness across research, clinical, and operational workflows.Lead discovery and scoping from pre-sales through post-sales, translating ambiguous workflow needs into hypothesis-driven problem framing, system requirements, and an execution plan with measurable endpoints.Define and enforce launch criteria for regulated contexts, including validation evidence, audit readiness, outcome metrics, and drive delivery until we demonstrate sustained production impact.Build in sensitive scientific data environments where auditability, validation, and access controls shape architecture, operating procedures, and failure handling.Run evaluation loops that measure model and system quality against workflow-specific scientific benchmarks and use results to drive model and product changes.Distill deployment learnings into hardened primitives, reference architectures, validation templates, and benchmark harnesses that scale across regulated life sciences environments.You might thrive in this role if youBring 5+ years of software/ML engineering or technical deployment experience with customer-facing ownership in biotech, pharma, clinical research, or scientific software; PhD, MS, or equivalent applied experience in a life sciences relevant field encouraged.Have owned customer GenAI deployments end-to-end from scoping through production adoption, and improved them through evaluation design, error analysis, and iterative evidence generation that tightens acceptance criteria over time.Have delivered AI systems in trial design, regulatory writing, or scientific operations where validation strategy, auditability, compliance constraints, and reviewer expectations shaped system design and rollout.Communicate clearly across scientific, clinical, model research, technical, and executive audiences, translating technical tradeoffs into decision quality, risk posture, and measurable outcomes with credibility.Apply systems thinking with high execution standards, consistently turning failures, escalations, and audit findings into improved operating standards, validation artifacts, and repeatable deployment playbooks.About OpenAIOpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity. We are an equal opportunity employer, and we do not discriminate on the basis of race, religion, color, national origin, sex, sexual orientation, age, veteran status, disability, genetic information, or other applicable legally protected characteristic. For additional information, please see OpenAI’s Affirmative Action and Equal Employment Opportunity Policy Statement.Background checks for applicants will be administered in accordance with applicable law, and qualified applicants with arrest or conviction records will be considered for employment consistent with those laws, including the San Francisco Fair Chance Ordinance, the Los Angeles County Fair Chance Ordinance for Employers, and the California Fair Chance Act, for US-based candidates. For unincorporated Los Angeles County workers: we reasonably believe that criminal history may have a direct, adverse and negative relationship with the following job duties, potentially resulting in the withdrawal of a conditional offer of employment: protect computer hardware entrusted to you from theft, loss or damage; return all computer hardware in your possession (including the data contained therein) upon termination of employment or end of assignment; and maintain the confidentiality of proprietary, confidential, and non-public information. In addition, job duties require access to secure and protected information technology systems and related data security obligations.To notify OpenAI that you believe this job posting is non-compliant, please submit a report through this form. No response will be provided to inquiries unrelated to job posting compliance.We are committed to providing reasonable accommodations to applicants with disabilities, and requests can be made via this link.OpenAI Global Applicant Privacy PolicyAt OpenAI, we believe artificial intelligence has the potential to help people solve immense global challenges, and we want the upside of AI to be widely shared. Join us in shaping the future of technology.
No items found.
2026-01-15 2:44
Member of Technical Staff - Data Ingestion Engineer
Reflection
1-10
United States
Full-time
Remote
false
Our MissionReflection’s mission is to build open superintelligence and make it accessible to all.We’re developing open weight models for individuals, agents, enterprises, and even nation states. Our team of AI researchers and company builders come from DeepMind, OpenAI, Google Brain, Meta, Character.AI, Anthropic and beyond.About the RoleData is playing an increasingly crucial role at the frontier of AI innovation. Many of the most meaningful advances in recent years have come not from new architectures, but from better data.As a member of the Data Team, your mission is to build and operate the ingestion systems that turn the open web and other large-scale data sources into reliable, well-structured corpora for training frontier models. You will own the machinery that acquires, extracts, normalizes, versions, and delivers data to our pre-training pipelines. You’ll work directly with world-class researchers to close the loop between what we collect and how it impacts model performance.This role is ideal for engineers who love building robust distributed systems, but who also want to run experiments, reason about tradeoffs in data acquisition, and iterate quickly based on measurable impact.Working closely with our pre-training and data quality teams, you will:Build and operate large-scale data ingestion systems for pre-training, including web crawling, extraction, and dataset deliveryRun experiments to evaluate crawling strategies, extraction methods, and ingestion tradeoffsAnalyze ingested data to identify gaps, redundancy, and areas to improveBuild ingestion pipelines that scale reliably across large data campaignsDevelop specialized crawlers for high-priority data sourcesReview code, debug production issues, and continuously improve ingestion infrastructureAbout You:Curious about how training data influences model capabilities, and can iterate quickly based on measurable downstream impactAble to collaborate tightly across functions: researchers, infra, operations, and external partners.Enjoy working in a hybrid research–engineering roleSkills and Qualifications:Experience building web crawling, data ingestion, or large-scale data acquisition systems using Ray, Beam, Spark, or similar technologies.Familiarity with how LLMs are trained and evaluated, and an intuition for what makes data useful for trainingComfortable working with very large datasets (multi-TB to PB scale) and building systems that are observable, testable, and maintainableComfortable designing experiments and using data to guide system improvementsExcellent communication skills. You can explain system behavior. You consider and communicate tradeoffs clearlyWhat We Offer:We believe that to build superintelligence that is truly open, you need to start at the foundation. Joining Reflection means building from the ground up as part of a small talent-dense team. You will help define our future as a company, and help define the frontier of open foundational models.We want you to do the most impactful work of your career with the confidence that you and the people you care about most are supported.Top-tier compensation: Salary and equity structured to recognize and retain the best talent globally.Health & wellness: Comprehensive medical, dental, vision, life, and disability insurance.Life & family: Fully paid parental leave for all new parents, including adoptive and surrogate journeys. Financial support for family planning.Benefits & balance: paid time off when you need it, relocation support, and more perks that optimize your time. Opportunities to connect with teammates: lunch and dinner are provided daily. We have regular off-sites and team celebrations.
No items found.
2026-01-15 1:59
Staff Software Engineer, Cloud Infrastructure
Tenstorrent
1001-5000
$100,000 – $500,000
United States
Full-time
Remote
false
Tenstorrent is leading the industry on cutting-edge AI technology, revolutionizing performance expectations, ease of use, and cost efficiency. With AI redefining the computing paradigm, solutions must evolve to unify innovations in software models, compilers, platforms, networking, and semiconductors. Our diverse team of technologists have developed a high performance RISC-V CPU from scratch, and share a passion for AI and a deep desire to build the best AI platform possible. We value collaboration, curiosity, and a commitment to solving hard problems. We are growing our team and looking for contributors of all seniorities.Tenstorrent is accelerating the future of AI and high-performance compute by building industry-leading CPU and AI architectures. As an Automotive and Robotics SoC Architect, you will define scalable, top-down system architectures that unify our CPU and AI technologies for next-generation automotive applications. This senior technical role shapes the architectural direction of our automotive and robotics portfolio, ensuring our products meet the industry's highest expectations for performance, safety, reliability, and security. This position is central to how Tenstorrent delivers world-class automotive solutions and requires strong technical leadership, systems thinking, and cross-functional collaboration.
This role is remote, based out of North America.
We welcome candidates at various experience levels for this role. During the interview process, candidates will be assessed for the appropriate level, and offers will align with that level, which may differ from the one in this posting.
Who You Are
A systems thinker who can architect complex SoCs from concept to execution.
A strong communicator who can articulate technical direction across engineering teams and external partners.
Someone with deep knowledge of safety-critical systems and the unique needs of automotive environments.
An innovator who can identify future use cases and propose next-generation architectural solutions.
A leader who thrives in a highly technical, cross-functional, fast-moving environment.
What We Need
Bachelor’s, Master’s, or Ph.D. in Electrical Engineering, Computer Engineering, or related field.
Extensive experience designing complex SoCs, ideally in automotive applications.
Proficiency in hardware description languages such as Verilog or VHDL.
Experience with hardware/software co-design and co-verification.
Knowledge of automotive safety standards (e.g., ISO 26262) and security principles.
Someone comfortable with up to 25% international travel.
Experience with both cameras, sensors, and others is a plus.
What You Will Learn
How cutting-edge CPU and AI architectures are adapted for automotive-grade environments.
Best-in-class methodologies for safety-critical SoC design, verification, and system integration.
How to translate emerging automotive use cases into scalable, future-proof SoC architectures.
Approaches to hardware-level security, robustness, and cyber-resilience in automotive compute systems.
Cross-functional collaboration strategies that drive innovation across architecture, software, DV, and product teams.
Compensation for all engineers at Tenstorrent ranges from $100k - $500k including base and variable compensation targets. Experience, skills, education, background and location all impact the actual offer made.
Tenstorrent offers a highly competitive compensation package and benefits, and we are an equal opportunity employer.This offer of employment is contingent upon the applicant being eligible to access U.S. export-controlled technology. Due to U.S. export laws, including those codified in the U.S. Export Administration Regulations (EAR), the Company is required to ensure compliance with these laws when transferring technology to nationals of certain countries (such as EAR Country Groups D:1, E1, and E2). These requirements apply to persons located in the U.S. and all countries outside the U.S. As the position offered will have direct and/or indirect access to information, systems, or technologies subject to these laws, the offer may be contingent upon your citizenship/permanent residency status or ability to obtain prior license approval from the U.S. Commerce Department or applicable federal agency. If employment is not possible due to U.S. export laws, any offer of employment will be rescinded.
No items found.
2026-01-14 19:14
HR Operations Partner
Together AI
201-500
$160,000 – $230,000
No items found.
Full-time
Remote
false
About the Role
As an AI Researcher, you will be pushing the frontier of foundation model research and make them a reality in products. You will be working on developing novel architectures, system optimizations, optimization algorithms, and data-centric optimizations, that go beyond state-of-the-arts. As a team, we have been pushing on all these fronts (e.g., Hyena, FlashAttention, FlexGen, and RedPajama). You will also work closely together with the machine learning systems, NLP/CV, and engineering teams for inspiration of research problems and to jointly work on solutions to practical challenges. You will also interact with customers to help them in their journey of training, using, and improving their AI applications using open models. Your research skills will be vital in staying up-to-date with the latest advancements in machine learning, ensuring that we stay at the cutting edge of open model innovations.
Requirements
Strong background in Machine Learning
Experience in building state-of-the-art models at large scale
Experience in developing algorithms in areas such as optimization, model architecture, and data-centric optimizations
Passion in contributing to the open model ecosystem and pushing the frontier of open models
Excellent problem-solving and analytical skills
Bachelor's, Master's, or Ph.D. degree in Computer Science, Electrical Engineering, or a related field
Responsibilities
Develop novel architectures, system optimizations, optimization algorithms, and data-centric optimizations, that significantly improve over state-of-the-art
Take advantage of the computational infrastructure of Together to create the best open models in their class
Understand and improve the full lifecycle of building open models; release and publish your insights (blogs, academic papers etc.)
Collaborate with cross-functional teams to deploy your models and make them available to a wider community and customer base
Stay up-to-date with the latest advancements in machine learning
About Together AI
Together AI is a research-driven artificial intelligence company. We believe open and transparent AI systems will drive innovation and create the best outcomes for society, and together we are on a mission to significantly lower the cost of modern AI systems by co-designing software, hardware, algorithms, and models. We have contributed to leading open-source research, models, and datasets to advance the frontier of AI, and our team has been behind technological advancement such as FlashAttention, Hyena, FlexGen, and RedPajama. We invite you to join a passionate group of researchers in our journey in building the next generation AI infrastructure.
Compensation
We offer competitive compensation, startup equity, health insurance and other competitive benefits. The US base salary range for this full-time position is: $160,000 - $230,000 + equity + benefits. Our salary ranges are determined by location, level and role. Individual compensation will be determined by experience, skills, and job-related knowledge.
Equal Opportunity
Together AI is an Equal Opportunity Employer and is proud to offer equal employment opportunity to everyone regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability, gender identity, veteran status, and more.
Please see our privacy policy at https://www.together.ai/privacy
No items found.
2026-01-14 12:29
Lead Software Engineer (Machine Learning)
Faculty
501-1000
United Kingdom
Full-time
Remote
false
Why Faculty?
We established Faculty in 2014 because we thought that AI would be the most important technology of our time. Since then, we’ve worked with over 350 global customers to transform their performance through human-centric AI. You can read about our real-world impact here.We don’t chase hype cycles. We innovate, build and deploy responsible AI which moves the needle - and we know a thing or two about doing it well. We bring an unparalleled depth of technical, product and delivery expertise to our clients who span government, finance, retail, energy, life sciences and defence.Our business, and reputation, is growing fast and we’re always on the lookout for individuals who share our intellectual curiosity and desire to build a positive legacy through technology.AI is an epoch-defining technology, join a company where you’ll be empowered to envision its most powerful applications, and to make them happen.About the teamOur Energy, Transition and Environment business unit is pioneering meaningful change in the clean energy revolution. Our vision is to accelerate the transition to net-zero emissions and drive efficiencies for a new era of utility companies.We believe that the responsible, and intelligent, deployment of AI is critical to the success of this mission. We partner with a wide range of clients - from major energy operators, to GreenTech startups, and national infrastructure providers - to build solutions which return measurable impact and move us towards a smarter, cleaner, and more sustainable world.About the roleJoin us as a Lead Software Engineer, with a focus on Machine Learning, to spearhead the technical direction and delivery of complex, innovative AI projects. You will act as a technical expert, applying your skills across various projects from client-side deployments to advising on AI strategy, while ensuring architectural decisions are sound and reliable.
This role demands a balance of deep technical expertise and strong leadership, focusing on driving innovation, fostering team growth, and building reusable solutions across the organisation. If you're ready to manage high-risk projects and deliver practical, innovative outcomes, this is your chance to shape our future.What you'll be doing:Setting the technical direction and overseeing delivery of high-risk, ill-defined software and infrastructure projects while balancing strategic trade-offs and helping teams prioritise in shifting environments, taking full ownership of successful outcomes for our most challenging projects.Designing and developing reliable, production-grade ML systems and justifying critical architectural decisions to ensure robust delivery.Developing clear, comprehensively scoped roadmaps for novel solutions to help customers achieve their strategic goals and accurately estimating effort on large workstreams to ensure successful and timely deliveryEngaging with technical and non-technical customers at all stages of the customer lifecycle, giving reasoned and credible advice and opinions on a broad range engineering topicsCollaborating proactively both within multidisciplinary delivery teams and across the engineering community at Faculty to overcome technical challengesCoaching team members on specific technologies and driving the development of shared organisational resources and libraries to streamline delivery and improve engineering methods across the company.Leading the hiring and selection process while mentoring multiple individuals and managers to define the future shape of the engineering team.Who we're looking for:You are a recognised technical expert who sets the standard for code quality and solution design, possessing the breadth of knowledge to solve almost any problem.You have an entrepreneurial mindset and are proactive in recommending new technologies or ways of working to keep our offering ahead of the competition.You bring expert-level experience in at least one major Cloud Solution Provider (AWS, GCP, or Azure) and have led teams to build full-stack web applications.You are a proven leader, capable of managing other managers and setting team-wide development goals to elevate client delivery.You thrive in high-stakes environments, demonstrating the ability to turn innovative ideas into practical, measurable outcomes for global energy operators.You are a compelling communicator who can confidently defend technical rationales to senior stakeholders and guide both technical and non-technical teams.The Interview ProcessTalent Team Screen (30 minutes)
Introduction to the role (45 minutes)
Pair Programming Interview (90 minutes)
System Design Interview (90 minutes)
Commercial & Leadership Interview (60 minutes)
#LI-PRIO
Our Recruitment EthosWe aim to grow the best team - not the most similar one. We know that diversity of individuals fosters diversity of thought, and that strengthens our principle of seeking truth. And we know from experience that diverse teams deliver better work, relevant to the world in which we live. We’re united by a deep intellectual curiosity and desire to use our abilities for measurable positive impact. We strongly encourage applications from people of all backgrounds, ethnicities, genders, religions and sexual orientations.Some of our standout benefits:Unlimited Annual Leave PolicyPrivate healthcare and dentalEnhanced parental leaveFamily-Friendly Flexibility & Flexible workingSanctus CoachingHybrid Working (2 days in our Old Street office, London)If you don’t feel you meet all the requirements, but are excited by the role and know you bring some key strengths, please do apply or reach out to our Talent Acquisition team for a confidential chat - talent@faculty.ai Please know we are open to conversations about part-time roles or condensed hours.
No items found.
2026-01-14 9:14
Infrastructure Engineer
Dataiku
1001-5000
France
Germany
Netherlands
Full-time
Remote
false
Dataiku is The Universal AI Platform™, giving organizations control over their AI talent, processes, and technologies to unleash the creation of analytics, models, and agents. Providing no-, low-, and full-code capabilities, Dataiku meets teams where they are today, allowing them to begin building with AI using their existing skills and knowledge.Dataiku’s promise to our customers is to provide them with the software and support needed to accelerate their Data Science and Machine Learning maturity. Dataiku’s Data Science team is responsible for delivering on that promise.
As an AI Deployment Strategist / Data Scientist at Dataiku, you will have the opportunity to participate in our customers' journeys, from supporting their discovery of the platform to coaching users and co-developing data science applications from design to deployment. You will primarily work with our customers in the financial services and insurance industries.
You will gain hands-on experience coding in multiple languages (primarily Python, occasionally R, SQL, PySpark, JavaScript, etc.) and applying the latest big data technologies to real-world business use cases.
Our ideal candidate is comfortable learning new languages, technologies, and modelling techniques while being able to explain their work to other data scientists and clients.
Key Areas of Responsibility (What You’ll Do)
Help users discover and master the Dataiku platform through user training, office hours, demos, and ongoing consultative support.
Analyse and investigate various kinds of data and machine learning applications across industries and use cases.
Provide strategic input to the customer and account teams that help our customers achieve success.
Scope and co-develop production-level data science projects with our customers.
Mentor and help educate data scientists and other customer team members to aid in career development and growth.
Experience (What We’re Looking For)
French and English - fluent
Curiosity and a desire to learn new technical skills.
Empathy and an eagerness to share your knowledge with your colleagues, Dataiku’s customers, and the general public.
Ability to clearly explain complex topics to technical as well as non-technical audiences.
Over 5 years of experience with coding (Python, R, SQL).
Over 5 years of experience building ML models.
Understanding of underlying data systems and platform mechanics, such as Cloud architectures, K8S, Spark, and SQL.
Bonus points for any of these
Experience with Consulting and/or Customer-facing Data Science roles.
Experience in the manufacturing industry.
Experience with Spark, SAS, Data Engineering or MLOps.
Experience developing web apps in JavaScript, RShiny, or Dash.
Experience building APIs.
Experience using enterprise data science tools.
Passion for teaching or public speaking.
#LI-Hybrid
What are you waiting for!
At Dataiku, you'll be part of a journey to shape the ever-evolving world of AI. We're not just building a product; we're crafting the future of AI. If you're ready to make a significant impact in a company that values innovation, collaboration, and your personal growth, we can't wait to welcome you to Dataiku! And if you’d like to learn even more about working here, you can visit our Dataiku LinkedIn page.
Our practices are rooted in the idea that everyone should be treated with dignity, decency and fairness. Dataiku also believes that a diverse identity is a source of strength and allows us to optimize across the many dimensions that are needed for our success. Therefore, we are proud to be an equal opportunity employer. All employment practices are based on business needs, without regard to race, ethnicity, gender identity or expression, sexual orientation, religion, age, neurodiversity, disability status, citizenship, veteran status or any other aspect which makes an individual unique or protected by laws and regulations in the locations where we operate. This applies to all policies and procedures related to recruitment and hiring, compensation, benefits, performance, promotion and termination and all other conditions and terms of employment. If you need assistance or an accommodation, please contact us at: reasonable-accommodations@dataiku.com
Protect yourself from fraudulent recruitment activity
Dataiku will never ask you for payment of any type during the interview or hiring process. Other than our video-conference application, Zoom, we will also never ask you to make purchases or download third-party applications during the process. If you experience something out of the ordinary or suspect fraudulent activity, please review our page on identifying and reporting fraudulent activity here.
No items found.
2026-01-14 8:59
Senior Director and AGC, Product Legal (Privacy, IP, Employment)
Scale AI
5000+
$201,600 – $241,920
United States
Full-time
Remote
false
About the role
We’re hiring an AI Architect to sit at the intersection of frontier AI research, product, and go-to-market. You’ll partner closely with ML teams in high-stakes meetings, scope and pitch solutions to top AI labs, and translate research needs (post-training, evals, alignment) into clear product roadmaps and measurable outcomes. You’ll drive end-to-end delivery—partnering with AI research teams and core customers to scope, pilot, and iterate on frontier model improvements—while coordinating with engineering, ops, and finance to translate cutting-edge research into deployable, high-impact solutions.
What you’ll do
Translate research → product: work with client side researchers on post-training, evals, safety/alignment and build the primitives, data, and tooling they need.
Partner deeply with core customers and frontier labs: work hands-on with leading AI teams and frontier research labs to tackle hard, open-ended technical problems related to frontier model improvement, performance, and deployment.
Shape and propose model improvement work: translate customer and research objectives into clear, technically rigorous proposals—scoping post-training, evaluation, and safety work into well-defined statements of work and execution plans.
Translate research into production impact: collaborate with customer-side researchers on post-training, evaluations, and alignment, and help design the data, primitives, and tooling required to improve frontier models in practice.
Own the end-to-end lifecycle: lead discovery, write crisp PRDs and technical specs, prioritize trade-offs, run experiments, ship initial solutions, and scale successful pilots into durable, repeatable offerings.
Lead complex, high-stakes engagements: independently run technical working sessions with senior customer stakeholders; define success metrics; surface risks early; and drive programs to measurable outcomes.
Partner across Scale: collaborate closely with research (agents, browser/SWE agents), platform, operations, security, and finance to deliver reliable, production-grade results for demanding customers.
Build evaluation rigor at the frontier: design and stand up robust evaluation frameworks (e.g., RLVR, benchmarks), close the loop with data quality and feedback, and share learnings that elevate technical execution across accounts.
You have
Deep technical background in applied AI/ML: 5–10+ years in research, engineering, solutions engineering, or technical product roles working on LLMs or multimodal systems, ideally in high-stakes, customer-facing environments.
Hands-on experience with model improvement workflows: demonstrated experience with post-training techniques, evaluation design, benchmarking, and model quality iteration.
Ability to work on hard, ambiguous technical problems: proven track record of partnering directly with advanced customers or research teams to scope, reason through, and execute on deep technical challenges involving frontier models.
Strong technical fluency: you can read papers, interrogate metrics, write or review complex Python/SQL for analysis, and reason about model-data trade-offs.
Executive presence with world-class researchers and enterprise leaders; excellent writing and storytelling.
Bias to action: you ship, learn, and iterate.
How you’ll work
Customer-obsessed: start from real research needs; prototype quickly; validate with data.
Cross-functional by default: align research, engineering, ops, and GTM on a single plan; communicate clearly up and down.
Field-forward: expect regular customer time and research leads; light travel as needed.
What success looks like
Clear wins with top labs: pilots that convert to scaled programs with strong eval signals.
Reusable alignment & eval building blocks that shorten time-to-value across accounts.
Crisp internal docs (PRDs, experiment readouts, exec updates) that drive decisions quickly.
Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity based compensation, subject to Board of Director approval. Your recruiter can share more about the specific salary range for your preferred location during the hiring process, and confirm whether the hired role will be eligible for equity grant. You’ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.Please reference the job posting's subtitle for where this position will be located. For pay transparency purposes, the base salary range for this full-time position in the locations of San Francisco, New York, Seattle is:$201,600—$241,920 USDPLEASE NOTE: Our policy requires a 90-day waiting period before reconsidering candidates for the same role. This allows us to ensure a fair and thorough evaluation of all applicants.
About Us:
At Scale, our mission is to develop reliable AI systems for the world's most important decisions. Our products provide the high-quality data and full-stack technologies that power the world's leading models, and help enterprises and governments build, deploy, and oversee AI applications that deliver real impact. We work closely with industry leaders like Meta, Cisco, DLA Piper, Mayo Clinic, Time Inc., the Government of Qatar, and U.S. government agencies including the Army and Air Force. We are expanding our team to accelerate the development of AI applications.
We believe that everyone should be able to bring their whole selves to work, which is why we are proud to be an inclusive and equal opportunity workplace. We are committed to equal employment opportunity regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability status, gender identity or Veteran status.
We are committed to working with and providing reasonable accommodations to applicants with physical and mental disabilities. If you need assistance and/or a reasonable accommodation in the application or recruiting process due to a disability, please contact us at accommodations@scale.com. Please see the United States Department of Labor's Know Your Rights poster for additional information.
We comply with the United States Department of Labor's Pay Transparency provision.
PLEASE NOTE: We collect, retain and use personal data for our professional business purposes, including notifying you of job opportunities that may be of interest and sharing with our affiliates. We limit the personal data we collect to that which we believe is appropriate and necessary to manage applicants’ needs, provide our services, and comply with applicable laws. Any information we collect in connection with your application will be treated in accordance with our internal policies and programs designed to protect personal data. Please see our privacy policy for additional information.
No items found.
2026-01-14 7:29
Engagement Manager (Intelligence Community), Public Sector
Scale AI
5000+
$201,600 – $241,920
United States
Full-time
Remote
false
About the role
We’re hiring an AI Architect to sit at the intersection of frontier AI research, product, and go-to-market. You’ll partner closely with ML teams in high-stakes meetings, scope and pitch solutions to top AI labs, and translate research needs (post-training, evals, alignment) into clear product roadmaps and measurable outcomes. You’ll drive end-to-end delivery—partnering with AI research teams and core customers to scope, pilot, and iterate on frontier model improvements—while coordinating with engineering, ops, and finance to translate cutting-edge research into deployable, high-impact solutions.
What you’ll do
Translate research → product: work with client side researchers on post-training, evals, safety/alignment and build the primitives, data, and tooling they need.
Partner deeply with core customers and frontier labs: work hands-on with leading AI teams and frontier research labs to tackle hard, open-ended technical problems related to frontier model improvement, performance, and deployment.
Shape and propose model improvement work: translate customer and research objectives into clear, technically rigorous proposals—scoping post-training, evaluation, and safety work into well-defined statements of work and execution plans.
Translate research into production impact: collaborate with customer-side researchers on post-training, evaluations, and alignment, and help design the data, primitives, and tooling required to improve frontier models in practice.
Own the end-to-end lifecycle: lead discovery, write crisp PRDs and technical specs, prioritize trade-offs, run experiments, ship initial solutions, and scale successful pilots into durable, repeatable offerings.
Lead complex, high-stakes engagements: independently run technical working sessions with senior customer stakeholders; define success metrics; surface risks early; and drive programs to measurable outcomes.
Partner across Scale: collaborate closely with research (agents, browser/SWE agents), platform, operations, security, and finance to deliver reliable, production-grade results for demanding customers.
Build evaluation rigor at the frontier: design and stand up robust evaluation frameworks (e.g., RLVR, benchmarks), close the loop with data quality and feedback, and share learnings that elevate technical execution across accounts.
You have
Deep technical background in applied AI/ML: 5–10+ years in research, engineering, solutions engineering, or technical product roles working on LLMs or multimodal systems, ideally in high-stakes, customer-facing environments.
Hands-on experience with model improvement workflows: demonstrated experience with post-training techniques, evaluation design, benchmarking, and model quality iteration.
Ability to work on hard, ambiguous technical problems: proven track record of partnering directly with advanced customers or research teams to scope, reason through, and execute on deep technical challenges involving frontier models.
Strong technical fluency: you can read papers, interrogate metrics, write or review complex Python/SQL for analysis, and reason about model-data trade-offs.
Executive presence with world-class researchers and enterprise leaders; excellent writing and storytelling.
Bias to action: you ship, learn, and iterate.
How you’ll work
Customer-obsessed: start from real research needs; prototype quickly; validate with data.
Cross-functional by default: align research, engineering, ops, and GTM on a single plan; communicate clearly up and down.
Field-forward: expect regular customer time and research leads; light travel as needed.
What success looks like
Clear wins with top labs: pilots that convert to scaled programs with strong eval signals.
Reusable alignment & eval building blocks that shorten time-to-value across accounts.
Crisp internal docs (PRDs, experiment readouts, exec updates) that drive decisions quickly.
Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity based compensation, subject to Board of Director approval. Your recruiter can share more about the specific salary range for your preferred location during the hiring process, and confirm whether the hired role will be eligible for equity grant. You’ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.Please reference the job posting's subtitle for where this position will be located. For pay transparency purposes, the base salary range for this full-time position in the locations of San Francisco, New York, Seattle is:$201,600—$241,920 USDPLEASE NOTE: Our policy requires a 90-day waiting period before reconsidering candidates for the same role. This allows us to ensure a fair and thorough evaluation of all applicants.
About Us:
At Scale, our mission is to develop reliable AI systems for the world's most important decisions. Our products provide the high-quality data and full-stack technologies that power the world's leading models, and help enterprises and governments build, deploy, and oversee AI applications that deliver real impact. We work closely with industry leaders like Meta, Cisco, DLA Piper, Mayo Clinic, Time Inc., the Government of Qatar, and U.S. government agencies including the Army and Air Force. We are expanding our team to accelerate the development of AI applications.
We believe that everyone should be able to bring their whole selves to work, which is why we are proud to be an inclusive and equal opportunity workplace. We are committed to equal employment opportunity regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability status, gender identity or Veteran status.
We are committed to working with and providing reasonable accommodations to applicants with physical and mental disabilities. If you need assistance and/or a reasonable accommodation in the application or recruiting process due to a disability, please contact us at accommodations@scale.com. Please see the United States Department of Labor's Know Your Rights poster for additional information.
We comply with the United States Department of Labor's Pay Transparency provision.
PLEASE NOTE: We collect, retain and use personal data for our professional business purposes, including notifying you of job opportunities that may be of interest and sharing with our affiliates. We limit the personal data we collect to that which we believe is appropriate and necessary to manage applicants’ needs, provide our services, and comply with applicable laws. Any information we collect in connection with your application will be treated in accordance with our internal policies and programs designed to protect personal data. Please see our privacy policy for additional information.
No items found.
2026-01-14 7:29
Office Coordinator
X AI
5000+
$45 – $100 / hour
United States
Full-time
Remote
false
About xAI
xAI’s mission is to create AI systems that can accurately understand the universe and aid humanity in its pursuit of knowledge. Our team is small, highly motivated, and focused on engineering excellence. This organization is for individuals who appreciate challenging themselves and thrive on curiosity. We operate with a flat organizational structure. All employees are expected to be hands-on and to contribute directly to the company’s mission. Leadership is given to those who show initiative and consistently deliver excellence. Work ethic and strong prioritization skills are important. All engineers are expected to have strong communication skills. They should be able to concisely and accurately share knowledge with their teammates.About the Role
As an AI Tutor - Economics, you will be instrumental in enhancing the capabilities of our cutting-edge technologies by providing high-quality input and labels using specialized software. Your role involves collaborating closely with our technical team to support the training of new AI tasks, ensuring the implementation of innovative initiatives. You'll contribute to refining annotation tools and selecting complex problems from advanced economics domains, with a focus on macroeconomic forecasting, microeconomic incentives, and behavioral experiments. This position demands a dynamic approach to learning and adapting in a fast-paced environment, where your ability to interpret and execute tasks based on evolving instructions is crucial.
AI Tutor’s Role in Advancing xAI’s Mission
As an AI Tutor, you will play an essential role in advancing xAI's mission by supporting the training and refinement of xAI’s AI models. AI Tutors teach our AI models about how people interact and react, as well as how people approach issues and discussions in economics. To accomplish this, AI Tutors will actively participate in gathering or providing data, such as text, voice, and video data, sometimes providing annotations, recording audio, or participating in video sessions. We seek individuals who are comfortable and eager to engage in these activities as a fundamental part of the role, ensuring a strong alignment with xAI’s goals and objectives to innovate.
Scope
An AI Tutor will provide services that include labeling and annotating data in text, voice, and video formats to support AI model training. At times, this may involve recording audio or video sessions, and tutors are expected to be comfortable with these tasks as they are fundamental to the role. Such data is a job requirement to advance xAI’s mission, and AI Tutors acknowledge that all work is done for hire and owned by xAI.
Responsibilities
Use proprietary software applications to provide input/labels on defined projects.
Support and ensure the delivery of high-quality curated data.
Play a pivotal role in supporting and contributing to the training of new tasks, working closely with the technical staff to ensure the successful development and implementation of cutting-edge initiatives/technologies.
Interact with the technical staff to help improve the design of efficient annotation tools.
Choose problems from economics fields that align with your expertise, focusing on areas like macroeconomics, microeconomics, and behavioral economics.
Regularly interpret, analyze, and execute tasks based on given instructions.
Key Qualifications
Must possess a PhD in Economics or related field
Proficiency in reading and writing, both in informal and professional English.
Outstanding communication, interpersonal, analytical, and organizational capabilities.
Solid reading comprehension skills combined with the capacity to exercise autonomous judgment even when presented with limited data/material.
Strong passion for and commitment to technological advancements and innovation in economics.
Preferred Qualifications
Possesses experience with at least one publication in a reputable economics journal or outlet.
Teaching experience as a professor.
Location & Other Expectations
This position is based in Palo Alto, CA, or fully remote.
The Palo Alto option is an in-office role requiring 5 days per week; remote positions require strong self-motivation.
If you are based in the US, please note we are unable to hire in the states of Wyoming and Illinois at this time.
We are unable to provide visa sponsorship.
Team members are expected to work from 9:00am - 5:30pm PST for the first two weeks of training and 9:00am - 5:30pm in their own timezone thereafter.
For those who will be working from a personal device, please note your computer must be a Chromebook, Mac with MacOS 11.0 or later, or Windows 10 or later.
Compensation
$45/hour - $100/hour
The posted pay range is intended for U.S.-based candidates and depends on factors including relevant experience, skills, education, geographic location, and qualifications. For international candidates, our recruiting team can provide an estimated pay range for your location.
Benefits:
Hourly pay is just one part of our total rewards package at xAI. Specific benefits vary by country, depending on your country of residence you may have access to medical benefits. We do not offer benefits for part-time roles.xAI is an equal opportunity employer. For details on data processing, view our Recruitment Privacy Notice.
No items found.
2026-01-14 7:29
People Data & Operations Manager
Snorkel AI
501-1000
United States
Intern
Remote
false
About Snorkel
At Snorkel, we believe meaningful AI doesn’t start with the model, it starts with the data.
We’re on a mission to help enterprises transform expert knowledge into specialized AI at scale. The AI landscape has gone through incredible changes between 2015, when Snorkel started as a research project in the Stanford AI Lab, to the generative AI breakthroughs of today. But one thing has remained constant: the data you use to build AI is the key to achieving differentiation, high performance, and production-ready systems. We work with some of the world’s largest organizations to empower scientists, engineers, financial experts, product creators, journalists, and more to build custom AI with their data faster than ever before. Excited to help us redefine how AI is built? Apply to be the newest Snorkeler!About the Internship
This role offers a unique vantage point: you will conduct original research while witnessing how those ideas move through a high-growth startup’s Go-To-Market (GTM) motion to create measurable impact for leading enterprises and LLM developers.
You’ll work closely with Snorkel researchers on open-ended projects and produce clear research outputs (experiments, prototypes, internal writeups, and potentially publications depending on project fit and timing).
What You’ll Do
Innovate Human-AI Interaction: Design new paradigms for distilling human expertise into model behavior
Define Frontier Capabilities: Collaborate with leading labs to develop the data strategies unlocking the next generation of agentic, reasoning and multi-modal models.
See Real Impact in Weeks - something about being able to impact the business in fast cycles
Example Project Areas
Projects vary by mentor and collaboration needs, but may include:
Synthetic data generation + filtering for specialized tasks
Evaluation datasets and benchmarks for LLM / RAG / agent behavior
Data-centric methods for improving reliability, calibration, and failure-mode coverage
Evaluating HITL data annotation processes, gaps, and improvements
What We’re Looking For
PhD Researcher: Currently pursuing a PhD in CS, Math, or Stats with a focus on ML or NLP and a strong publication record (NeurIPS, ICML, ICLR, EMNLP).
Builder-Thinker: You write clean, scalable code to validate your theories and are comfortable navigating complex codebases.
Entrepreneurial Mindset: Curious about the business of AI—how startups scale, find product-market fit, and translate technical metrics into business ROI
Nice to have
Prior work on evaluation, data curation, synthetic data, weak supervision, NLP, or multimodal ML.
Experience collaborating with academic labs or participating in research programs.
Internship Details
Duration: Summer (flexible start/end)
Location: Hybrid (Redwood City/SF)
Compensation: Competitive, commensurate with experience
Why Snorkel AI Research
Snorkel offers the intellectual rigor of an elite academic lab with the velocity of a category-defining technology company. You will gain a macro-view of the AI ecosystem, working where the world’s largest models meet the world’s most complex data.
Be Your Best at Snorkel
Joining Snorkel AI means becoming part of a company that has market proven solutions, robust funding, and is scaling rapidly—offering a unique combination of stability and the excitement of high growth. As a member of our team, you’ll have meaningful opportunities to shape priorities and initiatives, influence key strategic decisions, and directly impact our ongoing success. Whether you’re looking to deepen your technical expertise, explore leadership opportunities, or learn new skills across multiple functions, you’re fully supported in building your career in an environment designed for growth, learning, and shared success.
Snorkel AI is proud to be an Equal Employment Opportunity employer and is committed to building a team that represents a variety of backgrounds, perspectives, and skills. Snorkel AI embraces diversity and provides equal employment opportunities to all employees and applicants for employment. Snorkel AI prohibits discrimination and harassment of any type on the basis of race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state, or local law. All employment is decided on the basis of qualifications, performance, merit, and business need.
We will ensure that individuals with disabilities are provided reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. Please contact us to request accommodation.
No items found.
2026-01-14 7:29
National Security & Technology Policy Fellow
Scale AI
5000+
$201,600 – $241,920
United States
Full-time
Remote
false
About the role
We’re hiring an AI Architect to sit at the intersection of frontier AI research, product, and go-to-market. You’ll partner closely with ML teams in high-stakes meetings, scope and pitch solutions to top AI labs, and translate research needs (post-training, evals, alignment) into clear product roadmaps and measurable outcomes. You’ll drive end-to-end delivery—partnering with AI research teams and core customers to scope, pilot, and iterate on frontier model improvements—while coordinating with engineering, ops, and finance to translate cutting-edge research into deployable, high-impact solutions.
What you’ll do
Translate research → product: work with client side researchers on post-training, evals, safety/alignment and build the primitives, data, and tooling they need.
Partner deeply with core customers and frontier labs: work hands-on with leading AI teams and frontier research labs to tackle hard, open-ended technical problems related to frontier model improvement, performance, and deployment.
Shape and propose model improvement work: translate customer and research objectives into clear, technically rigorous proposals—scoping post-training, evaluation, and safety work into well-defined statements of work and execution plans.
Translate research into production impact: collaborate with customer-side researchers on post-training, evaluations, and alignment, and help design the data, primitives, and tooling required to improve frontier models in practice.
Own the end-to-end lifecycle: lead discovery, write crisp PRDs and technical specs, prioritize trade-offs, run experiments, ship initial solutions, and scale successful pilots into durable, repeatable offerings.
Lead complex, high-stakes engagements: independently run technical working sessions with senior customer stakeholders; define success metrics; surface risks early; and drive programs to measurable outcomes.
Partner across Scale: collaborate closely with research (agents, browser/SWE agents), platform, operations, security, and finance to deliver reliable, production-grade results for demanding customers.
Build evaluation rigor at the frontier: design and stand up robust evaluation frameworks (e.g., RLVR, benchmarks), close the loop with data quality and feedback, and share learnings that elevate technical execution across accounts.
You have
Deep technical background in applied AI/ML: 5–10+ years in research, engineering, solutions engineering, or technical product roles working on LLMs or multimodal systems, ideally in high-stakes, customer-facing environments.
Hands-on experience with model improvement workflows: demonstrated experience with post-training techniques, evaluation design, benchmarking, and model quality iteration.
Ability to work on hard, ambiguous technical problems: proven track record of partnering directly with advanced customers or research teams to scope, reason through, and execute on deep technical challenges involving frontier models.
Strong technical fluency: you can read papers, interrogate metrics, write or review complex Python/SQL for analysis, and reason about model-data trade-offs.
Executive presence with world-class researchers and enterprise leaders; excellent writing and storytelling.
Bias to action: you ship, learn, and iterate.
How you’ll work
Customer-obsessed: start from real research needs; prototype quickly; validate with data.
Cross-functional by default: align research, engineering, ops, and GTM on a single plan; communicate clearly up and down.
Field-forward: expect regular customer time and research leads; light travel as needed.
What success looks like
Clear wins with top labs: pilots that convert to scaled programs with strong eval signals.
Reusable alignment & eval building blocks that shorten time-to-value across accounts.
Crisp internal docs (PRDs, experiment readouts, exec updates) that drive decisions quickly.
Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity based compensation, subject to Board of Director approval. Your recruiter can share more about the specific salary range for your preferred location during the hiring process, and confirm whether the hired role will be eligible for equity grant. You’ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.Please reference the job posting's subtitle for where this position will be located. For pay transparency purposes, the base salary range for this full-time position in the locations of San Francisco, New York, Seattle is:$201,600—$241,920 USDPLEASE NOTE: Our policy requires a 90-day waiting period before reconsidering candidates for the same role. This allows us to ensure a fair and thorough evaluation of all applicants.
About Us:
At Scale, our mission is to develop reliable AI systems for the world's most important decisions. Our products provide the high-quality data and full-stack technologies that power the world's leading models, and help enterprises and governments build, deploy, and oversee AI applications that deliver real impact. We work closely with industry leaders like Meta, Cisco, DLA Piper, Mayo Clinic, Time Inc., the Government of Qatar, and U.S. government agencies including the Army and Air Force. We are expanding our team to accelerate the development of AI applications.
We believe that everyone should be able to bring their whole selves to work, which is why we are proud to be an inclusive and equal opportunity workplace. We are committed to equal employment opportunity regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability status, gender identity or Veteran status.
We are committed to working with and providing reasonable accommodations to applicants with physical and mental disabilities. If you need assistance and/or a reasonable accommodation in the application or recruiting process due to a disability, please contact us at accommodations@scale.com. Please see the United States Department of Labor's Know Your Rights poster for additional information.
We comply with the United States Department of Labor's Pay Transparency provision.
PLEASE NOTE: We collect, retain and use personal data for our professional business purposes, including notifying you of job opportunities that may be of interest and sharing with our affiliates. We limit the personal data we collect to that which we believe is appropriate and necessary to manage applicants’ needs, provide our services, and comply with applicable laws. Any information we collect in connection with your application will be treated in accordance with our internal policies and programs designed to protect personal data. Please see our privacy policy for additional information.
No items found.
2026-01-14 7:29
Staff Software Engineer, Inference Infrastructure
Cohere
501-1000
United States
Full-time
Remote
false
Who are we?Our mission is to scale intelligence to serve humanity. We’re training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, and agents. We believe that our work is instrumental to the widespread adoption of AI.We obsess over what we build. Each one of us is responsible for contributing to increasing the capabilities of our models and the value they drive for our customers. We like to work hard and move fast to do what’s best for our customers.Cohere is a team of researchers, engineers, designers, and more, who are passionate about their craft. Each person is one of the best in the world at what they do. We believe that a diverse range of perspectives is a requirement for building great products.Join us on our mission and shape the future!Why this role?Are you energized by building high-performance, scalable and reliable machine learning systems? Do you want to help define and build the next generation of AI platforms powering advanced NLP applications? We are looking for Members of Technical Staff to join the Model Serving team at Cohere. The team is responsible for developing, deploying, and operating the AI platform delivering Cohere's large language models through easy to use API endpoints. In this role, you will work closely with many teams to deploy optimized NLP models to production in low latency, high throughput, and high availability environments. You will also get the opportunity to interface with customers and create customized deployments to meet their specific needs.You may be a good fit if you have:5+ years of engineering experience running production infrastructure at a large scaleExperience designing large, highly available distributed systems with Kubernetes, and GPU workloads on those clustersExperience with Kubernetes dev and production coding and supportExperience with GCP, Azure, AWS, OCI, multi-cloud on-prem / hybrid servingExperience in designing, deploying, supporting, and troubleshooting in complex Linux-based computing environmentsExperience in compute/storage/network resource and cost managementExcellent collaboration and troubleshooting skills to build mission-critical systems, and ensure smooth operations and efficient teamworkThe grit and adaptability to solve complex technical challenges that evolve day to dayFamiliarity with computational characteristics of accelerators (GPUs, TPUs, and/or custom accelerators), especially how they influence latency and throughput of inference.Strong understanding or working experience with distributed systems.Experience in Golang, C++ or other languages designed for high-performance scalable servers).If some of the above doesn’t line up perfectly with your experience, we still encourage you to apply! We value and celebrate diversity and strive to create an inclusive work environment for all. We welcome applicants from all backgrounds and are committed to providing equal opportunities. Should you require any accommodations during the recruitment process, please submit an Accommodations Request Form, and we will work together to meet your needs.Full-Time Employees at Cohere enjoy these Perks:🤝 An open and inclusive culture and work environment 🧑💻 Work closely with a team on the cutting edge of AI research 🍽 Weekly lunch stipend, in-office lunches & snacks🦷 Full health and dental benefits, including a separate budget to take care of your mental health 🐣 100% Parental Leave top-up for up to 6 months🎨 Personal enrichment benefits towards arts and culture, fitness and well-being, quality time, and workspace improvement🏙 Remote-flexible, offices in Toronto, New York, San Francisco, London and Paris, as well as a co-working stipend✈️ 6 weeks of vacation (30 working days!)
No items found.
2026-01-14 5:59
Site Reliability Engineer, Inference Infrastructure
Cohere
501-1000
Canada
Full-time
Remote
false
Who are we?Our mission is to scale intelligence to serve humanity. We’re training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, and agents. We believe that our work is instrumental to the widespread adoption of AI.We obsess over what we build. Each one of us is responsible for contributing to increasing the capabilities of our models and the value they drive for our customers. We like to work hard and move fast to do what’s best for our customers.Cohere is a team of researchers, engineers, designers, and more, who are passionate about their craft. Each person is one of the best in the world at what they do. We believe that a diverse range of perspectives is a requirement for building great products.Join us on our mission and shape the future!Why this role?Are you energized by building high-performance, scalable and reliable machine learning systems? Do you want to help define and build the next generation of AI platforms powering advanced NLP applications? We are looking for a Site Reliability Engineer to join the Model Serving team at Cohere. The team is responsible for developing, deploying, and operating the AI platform delivering Cohere's large language models through easy to use API endpoints. In this role, you will work closely with many teams to deploy optimized NLP models to production in low latency, high throughput, and high availability environments. You will also get the opportunity to interface with customers and create customized deployments to meet their specific needs.As a Site Reliability Engineer you will:Build self-service systems that automate managing, deploying and operating services.This includes our custom Kubernetes operators that support language model deployments.Automate environment observability and resilience. Enable all developers to troubleshoot and resolve problems.Take steps required to ensure we hit defined SLOs, including participation in an on-call rotation.Build strong relationships with internal developers and influence the Infrastructure team’s roadmap based on their feedback.Develop our team through knowledge sharing and an active review process.You may be a good fit if you have:5+ years of engineering experience running production infrastructure at a large scaleExperience designing large, highly available distributed systems with Kubernetes, and GPU workloads on those clustersExperience with Kubernetes dev and production coding and supportExperience with GCP, Azure, AWS, OCI, multi-cloud on-prem / hybrid servingExperience in designing, deploying, supporting, and troubleshooting in complex Linux-based computing environmentsExperience in compute/storage/network resource and cost managementExcellent collaboration and troubleshooting skills to build mission-critical systems, and ensure smooth operations and efficient teamworkThe grit and adaptability to solve complex technical challenges that evolve day to dayFamiliarity with computational characteristics of accelerators (GPUs, TPUs, and/or custom accelerators), especially how they influence latency and throughput of inference.Strong understanding or working experience with distributed systems.Experience in Golang, C++ or other languages designed for high-performance scalable servers).If some of the above doesn’t line up perfectly with your experience, we still encourage you to apply! We value and celebrate diversity and strive to create an inclusive work environment for all. We welcome applicants from all backgrounds and are committed to providing equal opportunities. Should you require any accommodations during the recruitment process, please submit an Accommodations Request Form, and we will work together to meet your needs.Full-Time Employees at Cohere enjoy these Perks:🤝 An open and inclusive culture and work environment 🧑💻 Work closely with a team on the cutting edge of AI research 🍽 Weekly lunch stipend, in-office lunches & snacks🦷 Full health and dental benefits, including a separate budget to take care of your mental health 🐣 100% Parental Leave top-up for up to 6 months🎨 Personal enrichment benefits towards arts and culture, fitness and well-being, quality time, and workspace improvement🏙 Remote-flexible, offices in Toronto, New York, San Francisco, London and Paris, as well as a co-working stipend✈️ 6 weeks of vacation (30 working days!)
No items found.
2026-01-14 5:59
No job found
Your search did not match any job. Please try again