Find AI Work That Works for You
Latest roles in AI and machine learning, reviewed by real humans for quality and clarity.
I'm strong in:
Edit filters
Latest AI Jobs
Showing 61 – 79 of 79 jobs
Tag
AI Agent Evaluation Analyst - AI Trainer
Mindrift
1001-5000
USD
0
0
-
55
No items found.
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $55/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
AI Trainer
Apply
September 22, 2025
AI Agent Evaluation Analyst - AI Trainer
Mindrift
1001-5000
USD
-
55
United States
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $55/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
Data Analyst
Data Science & Analytics
Apply
September 22, 2025
AI Agent Evaluation Analyst - AI Trainer
Mindrift
1001-5000
USD
0
0
-
15
India
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $15/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
AI Trainer
Apply
September 22, 2025
AI Agent Evaluation Analyst - AI Trainer
Mindrift
1001-5000
USD
0
0
-
55
United States
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $55/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
Data Analyst
Data Science & Analytics
Apply
September 22, 2025
AI Agent Evaluation Analyst - AI Trainer
Mindrift
1001-5000
USD
0
0
-
55
United States
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $55/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
Machine Learning Engineer
Data Science & Analytics
Apply
September 22, 2025
AI Agent Evaluation Analyst - AI Trainer
Mindrift
1001-5000
USD
0
0
-
38
Canada
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $38/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
Machine Learning Engineer
Data Science & Analytics
Apply
September 22, 2025
AI Agent Evaluation Analyst - AI Trainer
Mindrift
1001-5000
USD
0
0
-
55
United States
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $55/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
AI Trainer
Apply
September 22, 2025
AI Agent Evaluation Analyst - AI Trainer
Mindrift
1001-5000
USD
0
0
-
15
Brazil
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $15/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
Machine Learning Engineer
Data Science & Analytics
Apply
September 22, 2025
AI Agent Evaluation Analyst - AI Trainer
Mindrift
1001-5000
USD
0
0
-
15
India
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $15/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
No items found.
Apply
September 22, 2025
AI Agent Quality Specialist - AI Trainer
Mindrift
1001-5000
USD
0
0
-
38
Singapore
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $38/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
Machine Learning Engineer
Data Science & Analytics
Apply
September 22, 2025
AI Agent Quality Specialist - AI Trainer
Mindrift
1001-5000
USD
0
0
-
20
South Africa
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $20/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
Machine Learning Engineer
Data Science & Analytics
Apply
September 22, 2025
AI Agent Quality Specialist - AI Trainer
Mindrift
1001-5000
USD
-
38
Philippines
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $38/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
No items found.
Apply
September 22, 2025
AI Agent Quality Specialist - AI Trainer
Mindrift
1001-5000
USD
0
0
-
15
Argentina
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $15/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
Data Analyst
Data Science & Analytics
Apply
September 22, 2025
AI Agent Quality Specialist - AI Trainer
Mindrift
1001-5000
USD
0
0
-
38
Australia
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $38/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
Data Analyst
Data Science & Analytics
Apply
September 22, 2025
AI Agent Quality Specialist - AI Trainer
Mindrift
1001-5000
USD
0
0
-
44
No items found.
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $44/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
No items found.
Apply
September 22, 2025
AI Agent Quality Specialist - AI Trainer
Mindrift
1001-5000
USD
0
0
-
47
France
Part-time
Remote
true
At Mindrift, innovation meets opportunity. We believe in using the power of collective human intelligence to ethically shape the future of AI.
The Mindrift platform, launched and powered by Toloka, connects domain experts with cutting-edge AI projects from innovative tech clients. Our mission is to unlock the potential of GenAI by tapping into real-world expertise from across the globe. Who we're looking for:We’re looking for curious and intellectually proactive contributors, the kind of person who double-checks assumptions and plays devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?This is a flexible, project-based opportunity well-suited for: Analysts, researchers, or consultants with strong critical thinking skills. Students (senior undergrads / grad students) looking for an intellectually interesting gig. People open to a part-time and non-permanent opportunity.
About the project:We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem-solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.You do not need a coding background, but you must be curious, intellectually rigorous, and capable of evaluating the soundness and consistency of complex setups. If you’ve ever excelled in things like consulting, CHGK, Olympiads, case solving, or systems thinking — you might be a great fit.What you’ll be doing: Reviewing evaluation tasks and scenarios for logic, completeness, and realism. Identifying inconsistencies, missing assumptions, or unclear decision points. Helping define clear expected behaviors (gold standards) for AI agents. Annotating cause-effect relationships, reasoning paths, and plausible alternatives. Thinking through complex systems and policies as a human would to ensure agents are tested properly. Working closely with QA, writers, or developers to suggest refinements or edge case coverage. How to get started:Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.Requirements Excellent analytical thinking: Can reason about complex systems, scenarios, and logical implications. Strong attention to detail: Can spot contradictions, ambiguities, and vague requirements. Familiarity with structured data formats: Can read, not necessarily write JSON/YAML. Can assess scenarios holistically: What's missing, what’s unrealistic, what might break? Good communication and clear writing (in English) to document your findings.
We also value applicants who have: Experience with policy evaluation, logic puzzles, case studies, or structured scenario design. Background in consulting, academia, olympiads (e.g. logic/math/informatics), or research. Exposure to LLMs, prompt engineering, or AI-generated content. Familiarity with QA or test-case thinking (edge cases, failure modes, “what could go wrong”). Some understanding of how scoring or evaluation works in agent testing (precision, coverage, etc.). Benefits Get paid for your expertise, with rates that can go up to $47/hour depending on your skills, experience, and project needs. Take part in a flexible, remote, freelance project that fits around your primary professional or academic commitments. Participate in an advanced AI project and gain valuable experience to enhance your portfolio. Influence how future AI models understand and communicate in your field of expertise.
AI Trainer
Apply
September 22, 2025
Product Evangelist
Lovable
201-500
-
Sweden
Full-time
Remote
false
TL;DR - We’re looking for a magnetic, world-class product evangelist to represent Lovable at hackathons, customer HQs, and community events and to build educational content. You’ll inspire people to build, make them successful, and share their stories with the world. If you love teaching, building, and energizing communities, you’ll thrive here.Why Lovable?Lovable lets anyone and everyone build software with any language. From solopreneurs to Fortune 100 teams, millions of people use Lovable to transform raw ideas into real products - fast. We are at the forefront of a foundational shift in software creation, which means you have an unprecedented opportunity to change the way the digital world works. Over 2 million people in 200+ countries already use Lovable to launch businesses, automate work, and bring their ideas to life. And we’re just getting started.We’re a small, talent-dense team building a generation-defining company from Stockholm. We value extreme ownership, high velocity and low-ego collaboration. We seek out people who care deeply, ship fast, and are eager to make a dent in the world.What we’re looking forTrack record of public speaking, community engagement, or developer advocacy.Deep passion for building software and teaching others.Strong presentation skills: equally at ease in enterprise boardrooms and grassroots community events.Ability to produce educational content across text, video, and web formats.Comfortable on camera. Can record video tutorials and engaging live sessions.Obsessed with seeing others succeed and turning their dreams into reality.Passionate about AI and emerging technologies.Previous experience as a developer advocate, evangelist, or community builderYou have already built and shipped using Lovable!What you’ll doCreate and represent Lovable at hackathons, community events, and enterprise customer events - equipping people to build successfully.Be Lovable. Bring a magnetic presence that energizes people and inspires them to create with Lovable.Present polished, compelling talks and workshops for audiences big and small. Technical and non-technical.Be the vibe coding expert. Go deep on Lovable’s tools and show others how to unlock their potential.Produce text, video, and interactive content that educates and activates users.Host video sessions, livestreams, and lessons that spread Lovable know-how.Work together with the marketing team to champion community success; Celebrate user wins, share their stories, and help people pursue their dreams.How we hireFill in a short form then jump on an initial exploratory call.We'll send you a quick take-home asking to present your best work over a loom video presentation.Join us for a round of interviews to discuss your experience in more depthJoin us for trial work lasting 1-2 days remote or on-site. We'll see how you tick and you get to meet the team and explore whether joining Lovable feels right for you.About your applicationPlease submit your application in English. It’s our company language so you’ll be speaking lots of it if you join.We treat all candidates equally - if you’re interested please apply through our careers portal.
Technical Marketing Specialist
Marketing & Sales
Content Strategist
Marketing & Sales
Apply
September 20, 2025
Senior Project Manager, Research & Development
Xaira
101-200
USD
0
150000
-
185000
United States
Full-time
Remote
false
About Xaira Therapeutics Xaira is an innovative biotech startup focused on leveraging AI to transform drug discovery and development. The company is leading the development of generative AI models to design protein and antibody therapeutics, enabling the creation of medicines against historically hard-to-drug molecular targets. It is also developing foundation models for biology and disease to enable better target elucidation and patient stratification. Collectively, these technologies aim to continually enable the identification of novel therapies and to improve success in drug development. Xaira is headquartered in the San Francisco Bay Area, Seattle, and London.About the Role We are seeking a Senior Project Manager, Biologics R&D to support the planning and coordination of biologics programs from early-stage discovery through IND-enabling studies. This role is well-suited for a candidate with a strong scientific background who has transitioned into a project management role or who has taken on cross-functional project management responsibilities within research settings. The ideal candidate will have hands-on experience in biologics R&D, with the ability to manage timelines, track deliverables, and partner closely with scientific leads to keep projects moving forward.
This position emphasizes scientific understanding, team coordination, and milestone tracking, making it a strong fit for someone with a bench science background who has taken on increasing responsibility for project organization and cross-functional collaboration.
Key Responsibilities Project Planning and Execution Support day-to-day planning and coordination of biologics research programs, ensuring execution of key activities from discovery through IND-enabling work. Build and maintain project timelines that capture critical milestones, dependencies, and deliverables. Partner with scientific leads and functional stakeholders to align on scope, priorities, and progress. Milestone Management & Decision Support Collaborate with scientific teams to define key project milestones, decisional experiments and translate into actionable, resource-information timelines. Help compile project data, timelines, risks, and trade-offs to support informed decision- making. Coordinate and facilitate planning and documentation of decisive experiments, and update project plans based on results and next steps. Cross-Functional Project Coordination Facilitate cross-functional collaboration across research to ensure effective communication and documentation of decisions, alignment of priorities and timelines. Schedule and run cross-functional meetings, ensuring clear agendas, effective documentation, and follow-up on action items. Monitor interdependencies across functions and proactively identify, flag and resolve conflicts across workstreams. Operational Support & Project Monitoring Track project status, risks, and mitigation strategies; provide updates to internal stakeholders. Coordinate with functional leads to understand resource needs and potential impacts to timelines. Contribute to scenario planning and contingency thinking for critical path activities Identify operational challenges and recommend improvements to enhance efficiency and coordination. Resource Coordination & Timeline Scenarios Work closely with functional and team leads to understand resourcing needs, identify gaps and timeline impacts. Develop scenario plans for critical path development decisions to support contingency and investment planning to address potential delays, bottlenecks, or shifts in scope. Qualifications Advanced degree (MS, PhD, or equivalent) in life sciences or a related field strongly preferred. Minimum of 6+ years of industry experience, including hands-on biologics research with demonstrated project or timeline management responsibilities. Proven track record of managing programs from research through IND filing. Deep understanding of biologics R&D processes, including nonclinical, CMC, and regulatory components. Strong analytical, planning, and organizational skills; able to translate complex scientific ideas into clear plans. Excellent communication, time management collaboration, and stakeholder management skills. Experience with basic project tracking tools (Excel, Google Suite, etc); advanced PM certifications or software expertise preferred but not required Please note that this position will be required to be on-site in the SSF/Brisbane office. Remote candidates will not be considered. Preferred Attributes Comfortable managing projects in a fast-paced, cross-functional environment. Demonstrated ability to lead without authority and influence diverse teams. Passion for science and dedication to improving patient outcomes through innovative biologics development. Compensation We offer a competitive compensation and benefits package, seeking to provide an open, flexible, and friendly work environment to empower employees and provide them with a platform to develop their long-term careers. A Summary of Benefits is available for all applicants. We offer a competitive package that includes base salary, bonus, and equity. The base pay range for this position is expected to be $150,000 - $185,000 annually; however, the base pay offered may vary depending on the market, job-related knowledge, skills and capabilities, and experience.
Xaira Therapeutics an equal-opportunity employer. We believe that our strength is in our differences. Our goal to build a diverse and inclusive team began on day one, and it will never end.
TO ALL RECRUITMENT AGENCIES: Xaira Therapeutics does not accept agency resumes. Please do not forward resumes to our jobs alias or employees. Xaira Therapeutics is not responsible for any fees related to unsolicited resumes.
Project Manager
Product & Operations
Apply
September 19, 2025
Product Manager, Editor
Descript
101-200
USD
0
175000
-
265000
United States
Full-time
Remote
false
About the role Descript’s vision is to put video in every communicator’s toolkit. Back in the day you needed like six monitors and a bachelor’s degree to edit video. Descript lets you do it by editing docs & slides, and increasingly by just asking AI. In the future, maybe you won’t even need to ask! But building a new way to record and edit videos that look & sound good comes with a series of unique design, technology, and business challenges. In other words, we need really good product managers. As a Product Manager on the Editor team, you’ll be shaping the future of what creative collaboration looks like with AI, from recording to final polish. You’ll work alongside a small, flat, highly collaborative team of experienced PMs, engineers, designers, AI researchers, and marketers. This is an opportunity to get hands-on experience with cutting-edge AI technology in a product users love and grow fast in your PM craft. What you’ll do Develop a deep understanding of our customers, their creative goals, and the evolving video/AI landscape — and use those insights to guide the roadmap. Own and shape the recording experience so that anyone can look and sound great without special gear or expertise. Bring Descript’s taste for great video into the recording and editing experience, shaping interactions and defaults so creators naturally produce work that feels polished and professional. Partner with AI research, engineering, design, and other teams to translate cutting-edge technology into reliable, delightful experiences. Own and drive the feature roadmap for your team from idea generation through to execution and analysis. You’ll bring clarity and focus to your team, and help them deliver value to users reliably and effectively. Who you are You have seen (and shipped) some things: You have 3 or more years of product management experience, ideally with a product of technical complexity that incorporates generative AI models. You start with the customer: Plenty of “AI products” start with the technology and they all suck. You care deeply about understanding customers, and have experience in doing customer discovery and defining features that solve real problems. You are a fog sculptor: There are aspects to our product and its design that do not have real competitors yet. The product intuition and design conventions have not been decided. We need product managers who are attracted to ambiguous problems, know how to ask smart questions, and translate ideas into action. You care about craft: Our product is not a hype machine that runs on empty promises and vaporware. We think our product will be successful because people love to use it. That means sweating the details. But without the sweat, because you aren’t stressed out. You are funny, creative, maybe kind of weird: Ok, this isn’t a hard requirement but Descript is kind of a quirky place and the people who love it tend to like that about working here. So, just want to drop a hint early on. Also, the sense of humor is important because even though we take our work seriously it’s still just, like, video creation and editing. We aren’t flying rockets into space so let’s all have a good time. The base salary range for this role is $175,000 - $265,000/year. Final offer amounts will carefully consider multiple factors, including prior experience, expertise, and location, and may vary from the amount above.About Descript Descript is building a simple, intuitive, fully-powered editing tool for video and audio — an editing tool built for the age of AI. We are a team of 150 — with a proven CEO and the backing of some of the world's greatest investors (OpenAI, Andreessen Horowitz, Redpoint Ventures, Spark Capital). Descript is the special company that's in possession of both product market fit and the raw materials (passionate user community, great product, large market) for growth, but is still early enough that each new employee has a measurable influence on the direction of the company. Benefits include a generous healthcare package, catered lunches, and flexible vacation time. Our headquarters are located in the Mission District of San Francisco, CA. We're looking to hire people who are local and able to join us at the office when needed. We're flexible, and you're an adult—we don't expect or mandate that you're in the office every day. But we do believe there are valuable and serendipitous moments of discovery and collaboration that come from working together in person. Descript is an equal opportunity workplace—we are dedicated to equal employment opportunities regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability, gender identity, or Veteran status. We believe in actively building a team rich in diverse backgrounds, experiences, and opinions to better allow our employees, products, and community to thrive.
Product Manager
Product & Operations
Apply
September 19, 2025
Mid-Market Account Executive
Jasper
501-1000
USD
200000
200000
-
200000
United States
Full-time
Remote
true
Jasper is the leading AI marketing platform, enabling the world's most innovative companies to reimagine their end-to-end marketing workflows and drive higher ROI through increased brand consistency, efficiency, and personalization at scale. Jasper has been recognized as "one of the Top 15 Most Innovative AI Companies of 2024" by Fast Company and is trusted by nearly 20% of the Fortune 500 – including Prudential, Ulta Beauty, and Wayfair. Founded in 2021, Jasper is a remote-first organization with team members across the US, France, and Australia.About The RoleWe’re seeking a driven and forward-thinking Mid-Market Account Executive to join our growing team. This role offers the opportunity to partner with top-tier organizations, guiding them in discovering the full potential of our Marketing AI solutions.You’ll act as a strategic advisor, leveraging your technical acumen, exceptional sales skills, and ability to align our technology with the specific goals of VP and C-level decision-makers. This is a role tailor-made for self-starters who excel at forging relationships across diverse teams and delivering measurable value. If you’re ready to challenge conventional strategies, foster meaningful relationships, and grow your career in AI sales, we’d love to hear from you!This fully remote position reports to the Director of Mid-Market Sales and is open to candidates located within the continental US.What you will do at JasperDrive sales by taking ownership of end-to-end sales cycles, positioning yourself as a trusted product expert and advisor.Deliver compelling product demonstrations with a focus on the value proposition and ROI.Build and execute targeted strategies to source new pipeline opportunities from Ideal Customer Profile (ICP) accounts.Navigate IT and procurement channels with ease, simplifying technical discussions for diverse stakeholders.Develop ROI models that link product value to executive goals and business outcomes.Work collaboratively to align our solution with the unique challenges of each mid-market client.Partner and collaborate with SEs, VEs, Professional Services, and CS teams to ensure seamless integration and support of solutions.Actively share ideas and strategies within a team environment to achieve collective success.What you will bring to JasperProven sales experience with at least 4+ years of B2B SaaS sales experience, including a minimum of 2 years in Mid-Market sales.A proven track record of meeting or exceeding quotas, particularly in dynamic environments.Familiarity with team selling and consultative approaches.Experience in technical SaaS sales, with the ability to discuss systems, APIs, and integrations confidently.Ability to explain complex processes in clear, straightforward terms that drive stakeholder engagement.An interest in AI innovation, allowing you to stay on top of the latest advancements and discuss its benefits, trade-offs, and how it's used in real-world scenarios.Proficiency in developing ROI models customized to client objectives.A natural collaborator who thrives in cross-functional environments, leveraging diverse talents for unified outcomes.Passion for building relationships that extend beyond closing deals to foster long-term partnerships.Comfort working independently while contributing to a broader, growth-oriented team culture.A curiosity for understanding customer needs and addressing challenges in meaningful ways.Compensation RangeAt Jasper, we believe in pay transparency and are committed to providing our employees and candidates with access to information about our compensation practices. The expected OTE for this role is $200,000. Compensation may vary based on relevant experience, skills, competencies, and certifications.Benefits & PerksComprehensive Health, Dental, and Vision coverage beginning on the first day for employees and their families401(k) program with up to 2% company matchingEquity grant participationFlexible PTO with a FlexExperience budget ($900 annually) to help you make the most of your time away from workFlexWellness program ($1,800 annually) to help support your personal health goalsGenerous budget for home office set up $1,500 annual learning and development stipend 16 weeks of paid parental leaveOur goal is to be a diverse workforce that is representative at all job levels as we know the more inclusive we are, the better our product will be. We are committed to celebrating and supporting our differences and that diversity is essential to innovation and makes us better able to serve our customers. We hire people of all levels and backgrounds who are excited to learn and develop their skills. We are an equal opportunity employer. Applicants will not be discriminated against because of race, color, creed, sex, sexual orientation, gender identity or expression, age, religion, national origin, citizenship status, disability, ancestry, marital status, veteran status, medical condition, or any protected category prohibited by local, state or federal laws.By submitting this application, you acknowledge that you have reviewed and agree to Jasper's CCPA Notice to Candidates, available at legal.jasper.ai/#ccpa.
Enterprise Sales
Marketing & Sales
Apply
September 19, 2025
No job found
Your search did not match any job. Please try again