Engineering Technical Lead Manager (TLM) - Enterprise
Use proprietary software applications to provide input/labels on defined projects. Support and ensure the delivery of high-quality curated data. Play a pivotal role in supporting and contributing to the training of new tasks, working closely with the technical staff to ensure the successful development and implementation of cutting-edge initiatives/technologies. Interact with the technical staff to help improve the design of efficient annotation tools. Choose problems from economics fields that align with your expertise, focusing on areas like macroeconomics, microeconomics, and behavioral economics. Regularly interpret, analyze, and execute tasks based on given instructions. Provide services including labeling and annotating data in text, voice, and video formats to support AI model training, and at times recording audio or video sessions as required by the role.
Member of Technical Staff - Data Quality Engineer (Pre-training)
As a member of the Data Team, you will own upstream data quality for LLM pre-training, working as a specialist or generalist across languages and modalities. You will partner closely with research and pre-training teams to translate requirements into measurable quality signals and provide actionable feedback to external data vendors. In addition to human-in-the-loop processes, you will design, validate, and scale automated QA methods to reliably measure data quality across large campaigns. You will build reusable QA pipelines that ensure delivery of high-quality data to pre-training teams for model training. Additionally, you will monitor and report on data quality over time, driving continuous iteration on quality standards, processes, and acceptance criteria.
Member of Technical Staff - Data Quality Engineer (Post-training)
As a member of the Data Team, your responsibilities include owning upstream data quality for LLM post-training and evaluation by analyzing expert-developed datasets and operationalizing quality standards for reasoning, alignment, and agentic use cases. You will partner closely with research and post-training teams to translate requirements into measurable quality signals and provide actionable feedback to external data vendors. You are expected to design, validate, and scale automated QA methods, including LLM-as-a-Judge frameworks, to reliably measure data quality across large campaigns. Additionally, you will build reusable QA pipelines that reliably deliver high-quality data to post-training teams for model training and evaluation. You will also monitor and report on data quality over time, driving continuous iteration on quality standards, processes, and acceptance criteria.
Evaluation Scenario Writer - AI Agent Testing Specialist
Design realistic and structured evaluation scenarios for LLM-based agents by creating test cases that simulate human-performed tasks and defining gold-standard behavior to compare agent actions against. Create structured test cases simulating complex human workflows, define gold-standard behavior and scoring logic to evaluate agent actions, analyze agent logs, failure modes, and decision paths. Work with code repositories and test frameworks to validate scenarios. Iterate on prompts, instructions, and test cases to improve clarity and difficulty. Ensure scenarios are production-ready, easy to run, and reusable.
Member of Technical Staff, Frontiers of Deep Learning Scaling
Use proprietary software applications to provide input/labels on defined projects. Support and ensure the delivery of high-quality curated data. Play a pivotal role in supporting and contributing to the training of new tasks, working closely with the technical staff to ensure the successful development and implementation of cutting-edge initiatives/technologies. Interact with the technical staff to help improve the design of efficient annotation tools. Choose problems from economics fields that align with your expertise, focusing on areas like macroeconomics, microeconomics, and behavioral economics. Regularly interpret, analyze, and execute tasks based on given instructions.
Freelance Software Developer (Ruby) / Quality Assurance (AI Trainer)
As an AI Tutor in Coding on the Mindrift platform, you will typically engage in code generation and code review, prompt evaluation, and complex data annotation. You will be involved in training and evaluation of large language models, benchmarking, and agent-based code execution in sandboxed environments. The role requires working across multiple programming languages including Python, JavaScript/TypeScript, Rust, and SQL. You will adapt guidelines for new domains and use cases and follow project-specific rubrics and requirements. Collaboration with project leads, solution engineers, and supply managers on complex or experimental projects is expected. Flexibility and quick adaptation to new requirements are essential.
Site Ops Manager
Use proprietary software applications to provide input and labels on defined projects, support and ensure the delivery of high-quality curated data, contribute to the training of new AI tasks by working closely with technical staff, interact with technical staff to help improve the design of efficient annotation tools, choose problems from economics fields that align with your expertise focusing on macroeconomics, microeconomics, and behavioral economics, and regularly interpret, analyze, and execute tasks based on given instructions. Additionally, provide services that include labeling and annotating data in text, voice, and video formats, sometimes recording audio or video sessions, which are fundamental parts of the role.
Site Ops Lead
Use proprietary software applications to provide input and labels on defined projects, support and ensure the delivery of high-quality curated data, contribute to the training of new tasks by working closely with technical staff to develop and implement advanced initiatives and technologies, interact with technical staff to help improve the design of efficient annotation tools, choose problems from economics fields such as macroeconomics, microeconomics, and behavioral economics that align with expertise, and regularly interpret, analyze, and execute tasks based on given instructions. The role also includes providing services that involve labeling and annotating data in text, voice, and video formats to support AI model training, recording audio or video sessions, and actively participating in gathering or providing data including text, voice, and video data with some annotations and recordings, all work done is for hire and owned by xAI.
Copy of Commercial Counsel
Use proprietary software applications to provide input/labels on defined projects. Support and ensure the delivery of high-quality curated data. Play a pivotal role in supporting and contributing to the training of new tasks, working closely with the technical staff to ensure the successful development and implementation of cutting-edge initiatives/technologies. Interact with the technical staff to help improve the design of efficient annotation tools. Choose problems from economics fields that align with your expertise, focusing on areas like macroeconomics, microeconomics, and behavioral economics. Regularly interpret, analyze, and execute tasks based on given instructions. Provide services that include labeling and annotating data in text, voice, and video formats to support AI model training. At times, recording audio or video sessions and being comfortable with these tasks, which are fundamental to the role.
Access all 4,256 remote & onsite AI jobs.
Frequently Asked Questions
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare, eros dolor interdum nulla, ut commodo diam libero vitae erat. Aenean faucibus nibh et justo cursus id rutrum lorem imperdiet. Nunc ut sem vitae risus tristique posuere.