Top Research Scientist Jobs Openings in 2025

Looking for opportunities in Research Scientist? This curated list features the latest Research Scientist job openings from AI-native companies. Whether you're an experienced professional or just entering the field, find roles that match your expertise, from startups to global tech leaders. Updated everyday.

People also search for:
Jasper.jpg

Research Scientist Intern - Post-Training (Distillation)

Jasper
-
FR.svg
France
Intern
Remote
false
Jasper is the leading AI marketing platform, enabling the world's most innovative companies to reimagine their end-to-end marketing workflows and drive higher ROI through increased brand consistency, efficiency, and personalization at scale. Jasper has been recognized as "one of the Top 15 Most Innovative AI Companies of 2024" by Fast Company and is trusted by nearly 20% of the Fortune 500 – including Prudential, Ulta Beauty, and Wayfair. Founded in 2021, Jasper is a remote-first organization with team members across the US, France, and Australia.About The RoleJasper Research is seeking a highly motivated intern to advance the frontiers of open source image-generation and image-editing applications. In this role, you will be instrumental in developing new state-of-the-art open-source text-to-image and image editing models while collaborating closely with our talented team of researchers and engineers. The internship duration is 6 months.We have historically relied on third-party existing open-source image foundational models for its core applications (replace-background, image uncropping, image relighting …), mastering model fine-tuning methods. One of the team’s current core projects consists now of building our own open-weights foundational text-to-image model from scratch.This role is open to candidates located in Paris. It will be a hybrid setup, which requires you to come into the office when necessary. What you will do at JasperAs a Research Scientist Intern, you will work closely with our research team to design and implement new methods to improve the sampling speed of diffusion models, such as, amongst others, re-flow, shortcut models or adversarial diffusion distillation. This role offers a unique opportunity to contribute to the development of a foundational open-source model, addressing challenges in scalability, fidelity, and generalization. You will engage in both theoretical and applied research, collaborating with experts in machine learning, computer vision, and natural language processing. The main goals of this internship are:Research & Development: Conduct literature reviews, propose and implement innovative methods to accelerate the sampling from diffusion models in the context of text-to-image and image editing.Model Training & Evaluation: Participate in the fine-tuning stages of the training of large-scale text-to-image models, conduct rigorous ablations, design evaluation metrics, and analyze model performance.Documentation & Communication: Document research findings, prepare technical reports, and participate in the external communication of the results.Open Source & Community: Contribute to an ambitious open-source project, publish research findings, and engage with the broader AI community.What you will bring to JasperCurrently enrolled in a Ph.D. or M.Sc program in Machine Learning, applied mathematics or computer science (Ph.D, preferred).Experience with distillation and/or improved sampling techniques for diffusion modelsA genuine interest in the field and a strong motivation to contribute to open-source initiatives, with a potentially proven track record through personal projects or previous experience in deep learning, especially generative models (e.g., diffusion models, GANs, VAEs, transformers).Strong coding abilities in Python and deep learning frameworks (PyTorch, TensorFlow, JAX).The ideal candidate will possess a strong critical thinking and problem-solving mindset, coupled with excellent teamwork skills.Being available for period of 6 months.Nice to haveStrong programming Python skills, including software engineering best practices to produce high-quality code.Experience with distributed training and large-scale systems on GPU clustersExperience with large-scale data processingContributions to open-source projects.Proven track record of achieving significant results as demonstrated by first-authored publications in major conferences and journals such as CVPR, ECCV, ICCV, ICLR, NeurIPSBenefits & PerksFlexible, hybrid work environment. Our office is based at Station F in Paris, the vibrant hub of the French startup ecosystem. Our efficient and lean team at Station F thrives on innovation and collaboration.Competitive compensation package
Research Scientist
Product & Operations
Machine Learning Engineer
Data Science & Analytics
Computer Vision Engineer
Software Engineering
Apply
Hidden link
Jasper.jpg

Research Scientist Intern - Post-Training (RLHF)

Jasper
-
FR.svg
France
Full-time
Remote
false
Jasper is the leading AI marketing platform, enabling the world's most innovative companies to reimagine their end-to-end marketing workflows and drive higher ROI through increased brand consistency, efficiency, and personalization at scale. Jasper has been recognized as "one of the Top 15 Most Innovative AI Companies of 2024" by Fast Company and is trusted by nearly 20% of the Fortune 500 – including Prudential, Ulta Beauty, and Wayfair. Founded in 2021, Jasper is a remote-first organization with team members across the US, France, and Australia.About The RoleJasper Research is seeking a highly motivated intern to advance the frontiers of open source image-generation and image-editing applications. In this role, you will be instrumental in developing new state-of-the-art open-source text-to-image and image editing models while collaborating closely with our talented team of researchers and engineers. The internship duration is 6 months.We have historically relied on third-party existing open-source image foundational models for its core applications (replace-background, image uncropping, image relighting …), mastering model fine-tuning methods. One of the team’s current core projects consists now of building our own open-weights foundational text-to-image model from scratch.This role is open to candidates located in Paris. It will be a hybrid setup, which requires you to come into the office when necessary. What you will do at JasperAs a Research Scientist Intern, you will work closely with our research team to design the best-suited RLHF techniques for diffusion models. This role offers a unique opportunity to contribute to the development of a foundational open-source model, addressing challenges in scalability, fidelity, and generalization. You will engage in both theoretical and applied research, collaborating with experts in machine learning, computer vision, and natural language processing. The main goals of this internship areResearch & Development: Conduct literature reviews, propose and implement innovative methods to fine-tune text-to-image models with reinforcement learning methods (RLHF, DPO)Model Training & Evaluation: Participate in the fine-tuning stages of the training of large-scale text-to-image models, conduct rigorous ablations, design evaluation metrics, and analyze model performance.Documentation & Communication: Document research findings, prepare technical reports, and participate in the external communication of the results.Open Source & Community: Contribute to an ambitious open-source project, publish research findings, and engage with the broader AI community.What you will bring to JasperCurrently enrolled in a Ph.D. or M.Sc program in Machine Learning, applied mathematics or computer science (Ph.D, preferred).Experience with RLHF techniques with application to either Large Language Models or Diffusion Models.A genuine interest in the field and a strong motivation to contribute to open-source initiatives, with a potentially proven track record through personal projects or previous experience in deep learning, especially generative models (e.g., diffusion models, GANs, VAEs, transformers).Strong coding abilities in Python and deep learning frameworks (PyTorch, TensorFlow, JAX).The ideal candidate will possess a strong critical thinking and problem-solving mindset, coupled with excellent teamwork skills.Being available for period of 6 months.Nice to haveStrong programming Python skills, including software engineering best practices to produce high-quality code.Experience with distributed training and large-scale systems on GPU clustersExperience with large-scale data processingContributions to open-source projects.Proven track record of achieving significant results as demonstrated by first-authored publications in major conferences and journals such as CVPR, ECCV, ICCV, ICLR, NeurIPSBenefits & PerksFlexible, hybrid work environment. Our office is based at Station F in Paris, the vibrant hub of the French startup ecosystem. Our efficient and lean team at Station F thrives on innovation and collaboration.Competitive compensation package
Research Scientist
Product & Operations
Machine Learning Engineer
Data Science & Analytics
Computer Vision Engineer
Software Engineering
Apply
Hidden link
Mistral AI.jpg

AI Scientist - Audio

Mistral AI
-
FR.svg
France
GB.svg
United Kingdom
Full-time
Remote
false
About Mistral At Mistral we are on a mission to democratize AI, producing frontier intelligence for everyone, developed in the open, and built by engineers all over the world. We are a dynamic, collaborative team passionate about AI and its potential to transform society. Our diverse workforce thrives in competitive environments and is committed to driving innovation, with teams distributed between Europe, the USA and Asia. We are creative, low-ego and team-spirited. At Mistral, we develop models for the enterprise and for consumers, focusing on delivering systems which can really change the way in which businesses operate and which can integrate into our daily lives. All while releasing frontier models open-source, for everyone to try and benefit. Mistral is hiring experts in the training of large language models and distributed systems. Join us to be part of a pioneering company shaping the future of AI. What will you do *Research and develop novel methods to push the frontier of large language models*Work across use cases (e.g reasoning, code, agents) and modalities (e.g text, image and speech)*Build tooling and infrastructure to allow training, evaluation and analysis of AI models at scale*Work cross-functionally with other scientists, engineers and product teams to ship AI systems which have a real-world impact About you * An expert in speech input/output methodologies (specific to audio)*You are a highly proficient software engineer in at least one programming language (Python or other, e.g. Rust, Go, Java)*You have hands-on experience with AI frameworks (e.g. PyTorch, JAX) or distributed systems (e.g. Ray, Kubernetes)*You have high engineering competence. This means being able to design complex software and make it usable in production*You are a self-starter, autonomous and a team player Now, it would be ideal if* You have experience working with large-scale speech-language models*You have hands-on experience with training large transformer models in a distributed fashion*You can navigate the full MLOps stack, for instance, fine-tuning, evaluation and deployment *You have a strong publication record in a relevant scientific domain Note that this is not an exhaustive or necessary list of requirements. Please consider applying if you believe you have the skills to contribute to Mistral's mission. We value profile and experience diversity. Benefits France💰 Competitive cash salary and equity🥕 Food : Daily lunch vouchers🥎 Sport : Monthly contribution to a Gympass subscription 🚴 Transportation : Monthly contribution to a mobility pass🧑‍⚕️ Health : Full health insurance for you and your family🍼 Parental : Generous parental leave policy🌎 Visa sponsorship UK💰 Competitive cash salary and equity🚑 Insurance🚴 Transportation: Reimburse office parking charges, or 90GBP/month for public transport🥎 Sport: 90GBP/month reimbursement for gym membership🥕 Meal voucher: £200 monthly allowance for its meals💰 Pension plan: SmartPension (percentages are 5% Employee & 3% Employer)
Research Scientist
Product & Operations
Machine Learning Engineer
Data Science & Analytics
Apply
Hidden link
World Labs.jpg

Research Scientist – Generative Modeling

World Labs
-
US.svg
United States
Full-time
Remote
false
Role OverviewWe are seeking a talented Research Scientist specializing in generative modeling and diffusion models to join our modeling team. This role is ideal for someone who is an expert at pre-training or post-training of large-scale diffusion models for images, videos, or 3D assets or scenes.You will collaborate closely with researchers, engineers, and product teams to bring advanced 3D modeling and machine learning techniques into real-world applications, ensuring that our technology remains at the forefront of visual innovation. This role involves significant hands-on research and engineering work, driving projects from conceptualization through to production deployment.Key ResponsibilitiesDesign, implement, and train large-scale diffusion models for generating 3D worldsDevelop and experiment with post-training for large-scale diffusion models to add novel control signals, adapt to target aesthetic preferences, or distill for efficient inferenceCollaborate closely with research and product teams to understand and translate product requirements into effective technical roadmaps.Contribute hands-on to all stages of model development including data curation, experimentation, evaluation, and deployment.Continuously explore and integrate cutting-edge research in diffusion and generative AI more broadlyAct as a key technical resource within the team, mentoring colleagues, and driving best practices in generative modeling and ML engineeringIdeal Candidate Profile3+ years of experience in generative modeling or applied ML roles, ideally at a startup or other fast-paced research environmentExtensive experience with machine learning frameworks such as PyTorch or TensorFlow, especially in the context of diffusion models and other generative modelsDeep expertise in at least one area of generative modeling: pre-training, post-training, diffusion distillation, etc for diffusion modelsStrong history of publications or open-source contributions involving large-scale diffusion modelsStrong coding proficiency in Python and experience with GPU-accelerated computing.Ability to engage effectively with researchers and cross-functional teams, clearly translating complex technical ideas into actionable tasks and outcomes.Comfortable operating within a dynamic startup environment with high levels of ambiguity, ownership, and innovation.Nice to HaveContributions to open-source projects in the fields of computer vision, graphics, or ML.Familiarity with large-scale training infrastructure (e.g., multi-node GPU clusters, distributed training environments).Experience integrating machine learning models into production environments.Led or been involved with the development or training of large-scale, state-of-the-art generative models
Research Scientist
Product & Operations
Machine Learning Engineer
Data Science & Analytics
Apply
Hidden link
Articul8 AI.jpg

Principal AI Researcher (India)

Articul8
-
IN.svg
India
Full-time
Remote
true
About us:At Articul8 AI, we relentlessly pursue excellence and create exceptional AI products that exceed customer expectations. We are a team of dedicated individuals who take pride in our work and strive for greatness in every aspect of our business.Job Description:Articul8 AI is seeking an exceptional Principal AI Researcher to join us in shaping the future of Generative Artificial Intelligence (GenAI). As a member of our Applied Research team, you will be responsible for conducting cutting-edge research to advance the capabilities of our AI systems. Your role will involve designing, implementing, and evaluating novel approaches to improve our GenAI models, working at the intersection of GenAI research and product development.Responsibilities:Serve as the subject matter expert in various domains of GenAI research and development, including:Data pipelines: Design and optimize data processing workflows for large-scale model trainingTraining methodologies: Implement pre-training, mid-training, and post-training strategies and optimization techniquesReinforcement learning: Develop RL algorithms for GenAI and with applications in decision-making, personalization, and several other tasksMultimodal AI: Create systems that effectively process and generate across text, image, audio, and video modalitiesPersonalization: Design and implement tailored GenAI experiences by understanding user behavior, preferences, and contexts to deliver customized content and recommendationsKnowledge representation and retrieval: Develop techniques for effectively representation of information and knowledge elicitation, as well as search and retrieval.Play a technical leadership role in designing, developing, and scaling novel algorithms and models by taking them from research prototypes to production-ready systems that deliver real-world impactLead groundbreaking research initiatives in GenAI by identifying high-impact problems, designing innovative experiments, and developing solutions that advance both theoretical understanding and practical applications.Drive strategic decision-making processes by contributing to technology assessment, roadmap planning, and identifying areas for innovation that align with Articul8's business objectives.Partner with cross-functional teams to integrate cutting-edge research findings into products and maintain our technological leadership in the market.Monitor and analyze emerging trends in generative AI and related fields, sharing valuable research contributions through publications at prestigious conferences and journals.Mentor and guide junior team members and help building a strong culture of rapid innovation.Required Qualifications:Education: PhD/MSc degree in Computer Science, Machine Learning (ML), or a related field.Professional experience: 8+ years of experience as an AI researcher with a track record of applied research and/or product development (out of which, at least 3+ years should be on actively developing GenAI technologies).Core technical skills:Experience developing tools, libraries, and infrastructure for data preprocessing, model training/finetuning, and deployment of LLMs in research and production environments.A strong background in parallel/distributed computing on the cloud.Machine learning, deep learning, probability theory and statistics, natural language processing, computer vision, data wrangling and preparation, model evaluation and interpretation.Programming Skills: Proficiency in programming languages such as Python and experience working with version control systems (e.g., Git) and collaborating on code repositories is crucial.Preferred Qualifications:Experience with cloud computing platforms such as AWS, Azure, or GCP.Proven track record of publications in top-tier conferences and journals.Professional Attributes:Problem Solving: ability to break down complex problems into manageable components, devising creative solutions, and iteratively refining ideas based on feedback and experimental evidence.Collaboration and Communication: proficiency in working within cross-functional teams - communicating clearly, providing constructive criticism, delegating responsibilities, and respecting diverse perspectives.Critical Thinking: ability to carefully evaluate assumptions, questioning established methodologies, challenging own biases, and maintaining skepticism when interpreting results.Curiosity and Continuous Learning: ability to stay curious about advances in related fields and constantly seeking opportunities to expand knowledge base.Emotional Intelligence and Intellectual Humility: capable of displaying empathy, resilience, adaptability, and self-awareness. Ability to recognize own limitations, embracing uncertainty, acknowledging mistakes, and valuing others' contributions.If you're ready to join a team that's changing the game, apply now to become a part of the Articul8 team.
Research Scientist
Product & Operations
Machine Learning Engineer
Data Science & Analytics
Apply
Hidden link
Articul8 AI.jpg

Senior/Staff AI Researcher (Brazil)

Articul8
-
BR.svg
Brazil
Full-time
Remote
true
About us:At Articul8 AI, we relentlessly pursue excellence and create exceptional AI products that exceed customer expectations. We are a team of dedicated individuals who take pride in our work and strive for greatness in every aspect of our business.Job Description:Articul8 AI is seeking an exceptional Senior/Staff AI Researcher to join us in shaping the future of Generative Artificial Intelligence (GenAI). As a member of our Applied Research team, you will be responsible for conducting cutting-edge research to advance the capabilities of our AI systems. Your role will involve designing, implementing, and evaluating novel approaches to improve our GenAI models, working at the intersection of GenAI research and product development.Responsibilities:Serve as the subject matter expert in various domains of GenAI research and development, including:Data pipelines: Design and optimize data processing workflows for large-scale model trainingTraining methodologies: Implement pre-training, mid-training, and post-training strategies and optimization techniquesReinforcement learning: Develop RL algorithms for GenAI and with applications in decision-making, personalization, and several other tasksMultimodal AI: Create systems that effectively process and generate across text, image, audio, and video modalitiesPersonalization: Design and implement tailored GenAI experiences by understanding user behavior, preferences, and contexts to deliver customized content and recommendationsKnowledge representation and retrieval: Develop techniques for effectively representation of information and knowledge elicitation, as well as search and retrieval.Play a technical leadership role in designing, developing, and scaling novel algorithms and models by taking them from research prototypes to production-ready systems that deliver real-world impact.Partner with cross-functional teams to integrate cutting-edge research findings into products and maintain our technological leadership in the market.Monitor and analyze emerging trends in generative AI and related fields, sharing valuable research contributions through publications at prestigious conferences and journals.Required Qualifications:Education: PhD/MSc degree in Computer Science, Machine Learning (ML), or a related field.Professional experience: 5+ years of experience as an AI researcher with a track record of applied research and/or product development (out of which, at least 2+ years should be on actively developing GenAI technologies).Core technical skills:Experience developing tools, libraries, and infrastructure for data preprocessing, model training/finetuning, and deployment of LLMs in research and production environments.A strong background in parallel/distributed computing on the cloud.Machine learning, deep learning, probability theory and statistics, natural language processing, computer vision, data wrangling and preparation, model evaluation and interpretation.Programming Skills: Proficiency in programming languages such as Python and experience working with version control systems (e.g., Git) and collaborating on code repositories is crucial.Preferred Qualifications:Experience with cloud computing platforms such as AWS, Azure, or GCP.Proven track record of publications in top-tier conferences and journals.Professional Attributes:Problem Solving: ability to break down complex problems into manageable components, devising creative solutions, and iteratively refining ideas based on feedback and experimental evidence.Collaboration and Communication: proficiency in working within cross-functional teams - communicating clearly, providing constructive criticism, delegating responsibilities, and respecting diverse perspectives.Critical Thinking: ability to carefully evaluate assumptions, questioning established methodologies, challenging own biases, and maintaining skepticism when interpreting results.Curiosity and Continuous Learning: ability to stay curious about advances in related fields and constantly seeking opportunities to expand knowledge base.Emotional Intelligence and Intellectual Humility: capable of displaying empathy, resilience, adaptability, and self-awareness. Ability to recognize own limitations, embracing uncertainty, acknowledging mistakes, and valuing others' contributions.If you're ready to join a team that's changing the game, apply now to become a part of the Articul8 team. NOTE: This position is available via CLT contract only, Thank you!
Research Scientist
Product & Operations
Machine Learning Engineer
Data Science & Analytics
Apply
Hidden link
Observe.AI

Machine Learning Scientist - NLP

Observe
0
0
-
0
IN.svg
India
Full-time
Remote
false
About Us Observe.AI enables enterprises to transform how they connect with customers - through AI agents and copilots that engage, assist, and act across every channel. From automating conversations to guiding human agents in real time to uncovering insights that shape strategy, Observe.AI turns every interaction into a driver of loyalty and growth. Trusted by global leaders, we’re creating a future where every customer experience is smarter, faster, and more impactful. Why Join Us At our core, we are shaping how AI transforms real-world challenges in the contact center space. You will join a team that thrives on experimentation, rapid learning, and applied research, where every project is a chance to design, test, and bring new ideas into production, publish in top conferences, and share insights with the community via technical blogs. From developing agentic AI and LLM powered solutions to creating evaluation frameworks and deploying end to end pipelines, your work will have direct customer impact. We value curiosity, collaboration, and the courage to push boundaries, in areas of large language models, query understanding, intent classification, information retrieval, dialog systems, natural language generation, and resource constrained text systems. If you are excited to turn breakthroughs in Mult-Agent systems, LLMs, NLP, and ML into practical outcomes, and to work side by side with product and engineering teams to make ideas real, this is where your research mindset, technical expertise, and creativity will thrive.   What you’ll be doing: Experiment and Iterate on ML Algorithms: Design and implement use-case–specific machine learning algorithms - spanning Agentic AI, LLMs, and traditional ML models - while optimizing models and pipelines, and deploying end-to-end solutions to address real business challenges. Evaluation Frameworks: Develop robust evaluation frameworks, incorporating diverse datasets, appropriate metrics, and human-judge alignment - to rigorously assess language model performance and ensure standards of accuracy, efficiency, and reliability. Paper and Blog Submissions: Draft comprehensive scientific reports detailing approaches and results for submission to top conferences (e.g., EMNLP, NAACL, ACL, Interspeech), while also authoring blog posts on company blog space to broaden impact and share insights with the wider community. Cross-Functional Collaboration, Transform Trends into Practical Outcomes: Partner with product and engineering teams to align on model requirements and integrate feedback, while applying hands-on experimentation to turn emerging trends into practical, high-impact solutions through both independent work and collaborative innovation. What you’ll bring to the role: Experience: 3+ years of relevant experience in NLP/ML research or applied roles. Experimental Mindset and Applied Research: You thrive on exploration - designing, running, and iterating on experiments to generate insights and optimize outcomes. You embrace new approaches, fail fast, learn quickly from results, and push boundaries. LLM, SLM, NLP & ML Expertise: Strong foundation in Natural Language Processing and Machine Learning, with direct experience developing, fine-tuning, evaluating, optimizing and deploying Small/Large Language Models and other ML models like classifiers, embedding models. Bias to Action and Outcomes: You have deep expertise in a variety of research methods, but you don't hold those methods precious. You can flexibly apply necessary tools in an unconstrained way to drive impact for the business. Technical Proficiency: Advanced Python programming skills, with expertise in frameworks such as PyTorch, TensorFlow, and Hugging Face, along with frameworks of multi-agent systems and tool and function calling. Trend Awareness: Actively stay up to date with the latest breakthroughs in AgenticAI, Generative AI, models, methods, and tools, and apply them thoughtfully in your work. Collaborative Problem Solver: Comfortable working cross-functionally, sharing insights, and valuing diverse perspectives to drive collective problem-solving. Analytical & Critical Thinker: Strong ability to interpret model outputs, identify the categories of errors and LLM limitations, and implement solutions for improved performance. Perks & Benefits  Excellent medical insurance options and free online doctor consultations Yearly privilege and sick leaves as per Karnataka S&E Act Generous holidays (National and Festive) recognition and parental leave policies Learning & Development fund to support your continuous learning journey and professional development Fun events to build culture across the organization Flexible benefit plans for tax exemptions (i.e. Meal card, PF, etc.) Our Commitment to Inclusion and Belonging Observe.AI is an Equal Employment Opportunity employer that proudly pursues and hires a diverse workforce. Observe AI does not make hiring or employment decisions on the basis of race, color, religion or religious belief, ethnic or national origin, nationality, sex, gender, gender identity, sexual orientation, disability, age, military or veteran status, or any other basis protected by applicable local, state, or federal laws or prohibited by Company policy. Observe.AI also strives for a healthy and safe workplace and strictly prohibits harassment of any kind. We welcome all people. We celebrate diversity of all kinds and are committed to creating an inclusive culture built on a foundation of respect for all individuals. We seek to hire, develop, and retain talented people from all backgrounds. Individuals from non-traditional backgrounds, historically marginalized or underrepresented groups are strongly encouraged to apply. If you are ambitious, make an impact wherever you go, and you're ready to shape the future of Observe.AI, we encourage you to apply. For more information, visit www.observe.ai.
Machine Learning Engineer
Data Science & Analytics
NLP Engineer
Software Engineering
Apply
Hidden link
Xaira Therapeutics.jpg

Senior Scientist/Principal Scientist, Expression

Xaira
USD
140000
-
205000
US.svg
United States
Full-time
Remote
false
About Xaira Therapeutics Xaira is an innovative biotech startup focused on leveraging AI to transform drug discovery and development. The company is leading the development of generative AI models to design protein and antibody therapeutics, enabling the creation of medicines against historically hard-to-drug molecular targets. It is also developing foundation models for biology and disease to enable better target elucidation and patient stratification. Collectively, these technologies aim to continually enable the identification of novel therapies and to improve success in drug development. Xaira is headquartered in the San Francisco Bay Area, Seattle, and London.About the Role We are seeking a Senior or Principal Scientist to join our team and lead from the bench. The ideal candidate will have a strong foundation in protein expression in both eukaryotic, prokaryotic and cell free systems, with demonstrated expertise in cloning techniques essential for therapeutic protein development.  Candidates with pharmaceutical or industrial experience in protein production at scales ranging from microgram to gram scales are strongly preferred. About You The successful candidate will have a strong understanding of the biophysics of protein-protein interactions, protein chemistry and extensive experience in the molecular biology and biophysical techniques essential for protein engineering. You will be expected to work independently, drive scientific execution in the lab, and provide technical leadership on complex projects.  Supervisory or mentorship experience is desirable, as is the ability to collaborate effectively within cross-functional teams spanning discovery, process development, and analytical functions. Strong communication skills and the ability to present scientific data to diverse audiences are essential.  We are looking for a highly adaptable individual who can excel in a fast-paced, matrixed environment with shifting priorities and multiple responsibilities. This is a hands-on role suited for a scientist who is passionate about advancing drug discovery and development through innovation and scientific rigor.   Key Responsibilities Protein expression in multiple systems including mammalian, prokaryotic, insect and cell free systems is the primary responsibility. Downstream processing of samples including harvesting cells in small scale and up to 20 L using depth filters, affinity chromatography, ion exchange chromatography, and sizing. Use of various analytical techniques to evaluate process efficiency and quality. It is anticipated that this will require creative and unique approaches, including construct design and screening to increase yields, the development of novel strategies and methods for difficult to express proteins and to increase stability of problematic proteins.  The target proteins will be therapeutic proteins, antigens for the generation of therapeutics, assay reagents for screens, and proteins for structural studies. Experience in high throughput expression and purification is expected. The experience and flexibility to work on downstream purification is required. Leadership and mentoring of team members is expected. Excellent record keeping and organization is a must. Qualifications PhD with 8+ years of industry experience is required. Strong technical background in protein expression, tissue culture, aseptic techniques, and the operation and maintenance of bioreactors in an industry setting a must. Experienced molecular biologist with the ability to design and make expression constructs is required. Mammalian transient protein expression and cell line generation is required. Experience with insect cell expression using baculovirus and S2 cell line generation is strongly preferred. Experience with prokaryotic and cell free expression is a plus. Hands-on experience in generating and managing research cell banks is essential. Experience in protein purification and protein characterization is required. Experience with HPLC and FPLC techniques is desirable, specifically, knowledge of Agilent HPLC and Cytiva AKTA systems.  Downstream medium scale (10 - 20 L) sample preparation for purification. Protein analysis for expression including SDS-PAGE, CE-SDS, glycosylation analysis, cIEF and BLI/SPR. Demonstrated expertise in protein expression processes, ranging from 96-well plate formats to scales of up to 10-liter bioreactors, is required. Proven expertise in biological research, demonstrated by publications in peer-reviewed journals, issued patents, and successful IND filings. Track record of independently leading complex research projects while working collaboratively within cross-functional teams. Committed to fostering an inclusive, respectful environment that embraces diverse perspectives. Demonstrated inclusive leadership through active listening and thoughtful engagement with all viewpoints.   Compensation We offer a competitive compensation and benefits package, seeking to provide an open, flexible, and friendly work environment to empower employees and provide them with a platform to develop their long-term careers. A Summary of Benefits is available for all applicants. We offer a competitive package that includes base salary, bonus, and equity. The base pay range for this position is expected to be $140,000 - $205,000 annually; however, the base pay offered may vary depending on the market, job-related knowledge, skills and capabilities, and experience. Xaira Therapeutics an equal-opportunity employer. We believe that our strength is in our differences. Our goal to build a diverse and inclusive team began on day one, and it will never end. Xaira Therapeutics is an equal opportunity employer. We value diversity at our company. We do not discriminate on the basis of race, religion, color, national origin, sex, gender, gender expression, sexual orientation, age, marital status, veteran status, or disability status. We will ensure that individuals with disabilities are provided reasonable accommodation to participate in the job application or interview process, to perform critical job functions, and to receive other benefits and privileges of employment. Please contact Human Resources to request accommodation. TO ALL RECRUITMENT AGENCIES: Xaira Therapeutics does not accept agency resumes. Please do not forward resumes to our jobs alias or employees. Xaira Therapeutics is not responsible for any fees related to unsolicited resumes.  
Research Scientist
Product & Operations
Apply
Hidden link
Anduril Industries.jpg

Advanced Research Scientist

Anduril
USD
0
117300
-
175950
US.svg
United States
Full-time
Remote
false
Anduril Industries is a defense technology company with a mission to transform U.S. and allied military capabilities with advanced technology. By bringing the expertise, technology, and business model of the 21st century’s most innovative companies to the defense industry, Anduril is changing how military systems are designed, built and sold. Anduril’s family of systems is powered by Lattice OS, an AI-powered operating system that turns thousands of data streams into a realtime, 3D command and control center. As the world enters an era of strategic competition, Anduril is committed to bringing cutting-edge autonomy, AI, computer vision, sensor fusion, and networking technology to the military in months, not years.ABOUT THE TEAM Anduril’s Research Scientists excel at developing state-of-the-art algorithms and software that solve scientific problems with real-world applications. Working in small, innovative teams, our research scientists create impactful solutions that make a difference. Our research endeavors don’t end once we’ve published papers; our work is complete when our technology is deployed in mission-critical systems, ensuring success for our customers in government and industry.  Join us in our mission to expand the boundaries of what’s possible! WHAT YOU WILL DO Drive rapid prototyping initiatives for advanced R&D projects, focusing on specialized algorithm development in the context of radar systems, video sensors, space-based sensing, and Command and Control (C2) systems. Utilize high-fidelity modeling and simulation tools to assess and quantify the impact of innovative technologies on system performance. Collaborate with cross-disciplinary teams to ensure seamless integration of software and hardware, optimizing system functionalities for radar systems. Implement rigorous software quality assurance processes, using various testing methodologies to ensure reliability and efficiency of developed solutions. Engage with stakeholders to align R&D outcomes with mission-critical objectives, ensuring optimal performance and operational success. Mentor junior team members, fostering a culture of innovation and continuous improvement within the team. REQUIRED QUALIFICATIONS An M.S. or Ph.D. in Applied or Computational Mathematics, Electrical Engineering, Computer Science, Controls and Dynamical Systems, Aerospace Engineering, Physics, Statistics and Probability, or a related field. 2+ years of professional experience in embedded software/firmware engineering. Strong foundation in applied mathematics, including probability theory, optimization theory, linear algebra, and numerical analysis. Familiarity with functional programming languages (e.g., C/C++, Julia, Rust, Python, CUDA). Demonstrated experience in scientific computing, including algorithm implementation, optimization methods/theory, probabilistic/stochastic models, graphical models. Knowledge of digital signal processing (DSP) and image processing, as well as controls and estimation theory. Excellent written and verbal communication skills to convey complex technical concepts to diverse audiences. Adept at problem identification and principled approaches to problem formulation and solution. Effective data analysis, deep-diving, trouble-shooting. Open-minded, creative, imaginative. Agile learner. Enthusiastic collaboration, energized by driving team success. Ability to obtain and maintain a U.S. TS/SCI security clearance. PREFERRED QUALIFICATIONS Experience with either radar signal processing or image processing. Experience in GPU programming (CUDA programming) and rapid prototyping.     We request transcripts as part of the early application process to understand your academic background and how your coursework supports the skills deemed critical for the role. Transcripts help us assess your technical and analytical abilities, complementing our interview process in which we also evaluate practical experience and cultural fit. If you choose not to share your transcripts, you will need to provide detailed information regarding your academic performance in relevant courses, including projects and coursework specifics, to ensure we evaluate your academic accomplishments properly. If you do provide academic transcripts, feel free to redact non-technical information (e.g., student ID, dates, non-technical coursework, etc.). Unofficial transcripts obtained online acceptable for this assessment.US Salary Range$117,300—$175,950 USD  The salary range for this role is an estimate based on a wide range of compensation factors, inclusive of base salary only. Actual salary offer may vary based on (but not limited to) work experience, education and/or training, critical skills, and/or business considerations. Highly competitive equity grants are included in the majority of full time offers; and are considered part of Anduril's total compensation package. Additionally, Anduril offers top-tier benefits for full-time employees, including: Platinum Healthcare Benefits: For U.S. roles, we offer comprehensive medical, dental, and vision plans at little to no cost to you. For UK roles, Private Medical Insurance (PMI): Anduril will cover the full cost of the insurance premium for an employee and dependents. For AUS roles, Private health plan through Bupa: Coverage is fully subsidized by Anduril. Basic Life/AD&D and long-term disability insurance 100% covered by Anduril, plus the option to purchase additional life insurance for you and your dependents. Extremely generous company holiday calendar including a holiday hiatus in December, and highly competitive PTO plans. 16 weeks of paid Caregiver & Wellness Leave to care for a family member, bond with your baby, or tend to your own medical condition. Family Planning & Parenting Support: Fertility (eg, IVF, preservation), adoption, and gestational carrier coverage with additional benefits and resources to provide support from planning to parenting. Mental Health Resources: We provide free mental health resources 24/7 including therapy, life coaching, and more. Additional work-life services, such as free legal and financial support, available to you as well. A professional development stipend is available to all Andurilians. Daily Meals and Provisions: For many of our offices this means breakfast, lunch and fully stocked micro-kitchens. Company-funded commuter benefits available based on your region. Relocation assistance (depending on role eligibility). 401(k) retirement savings plan - both a traditional and Roth 401(k). (US roles only) The recruiter assigned to this role can share more information about the specific compensation and benefit details associated with this role during the hiring process. Anduril is an equal-opportunity employer committed to creating a diverse and inclusive workplace. The Anduril team is made up of incredibly talented and unique individuals, who together are disrupting industry norms by creating new paths towards the future of defense technology. All qualified applicants will be treated with respect and receive equal consideration for employment without regard to race, color, creed, religion, sex, gender identity, sexual orientation, national origin, disability, uniform service, Veteran status, age, or any other protected characteristic per federal, state, or local law, including those with a criminal history, in a manner consistent with the requirements of applicable state and local laws, including the CA Fair Chance Initiative for Hiring Ordinance. We actively encourage members of recognized minorities, women, Veterans, and those with disabilities to apply, and we work to create a welcoming and supportive environment for all applicants throughout the interview process. If you are someone passionate about working on problems that have a real-world impact, we'd love to hear from you! To view Anduril's candidate data privacy policy, please visit https://anduril.com/applicant-privacy-notice/.
Research Scientist
Product & Operations
Software Engineer
Software Engineering
Machine Learning Engineer
Data Science & Analytics
Computer Vision Engineer
Software Engineering
Apply
Hidden link
PathAI.jpg

Director of Translational Research, Oncology

PathAI
0
0
-
0
No items found.
Remote
true
Who We Are At PathAI, we’re building an AI-powered platform for pathology to advance the diagnosis and treatment of disease and to improve the lives of patients. We’re applying our work to drug development, clinical diagnosis, and global health. What You’ll Do The day-to-day responsibilities for the Director of Translational Research, Oncology at PathAI include: Oversee execution of multiple translational research projects/contracts with key clients to meet company goals and strengthen partner relationships Oversee execution of multiple programs to drive company-wide scientific and business strategy Develop client relationships and pursue new opportunities to drive business growth Responsible for professional development of team of program / project managers / support staff Responsible for hiring and training activities Collaborate and liaise across internal stakeholders including the business development, product, and machine learning teams, providing hypotheses for novel biological insights and strategy roadmaps for project completion. What You Bring Advanced degree in life sciences or biomedical engineering-related disciplines required, PhD preferred. At least 5 years work experience Experience building, leading and managing teams. Strong user of technology with a solid basis in quantitative analysis and data-driven decision making Intellectual curiosity and the ability to learn quickly in a complex space Excellent communication skills Publications of research in related fields  We Want To Hear From You At PathAI, we are looking for individuals who are team players, are willing to do the work no matter how big or small it may be, and who are passionate about everything they do. If this sounds like you, even if you may not match the job description to a tee, we encourage you to apply. You could be exactly what we're looking for.  PathAI is an equal opportunity employer, dedicated to creating a workplace that is free of harassment and discrimination. We base our employment decisions on business needs, job requirements, and qualifications — that's all. We do not discriminate based on race, gender, religion, health, personal beliefs, age, family or parental status, or any other status. We don't tolerate any kind of discrimination or bias, and we are looking for teammates who feel the same way.    #LI-Remote   
Research Scientist
Product & Operations
Apply
Hidden link
Scale AI.jpg

Head of Evaluation and Oversight Research

Scale AI
USD
260000
-
350000
US.svg
United States
Full-time
Remote
false
Scale is the leading data and evaluation partner for frontier AI companies, playing an integral role in advancing the science of evaluating and characterizing large language models (LLMs). Our research focuses on tackling the hardest problems in scalable oversight and the evaluation of advanced AI capabilities. We collaborate broadly across industry and academia and regularly publish our findings.  Our Research team is shaping the next generation of evaluation science for frontier AI models and works at the leading edge of model assessment and oversight. Some of our current research includes: Developing AI-assisted evaluation pipelines, where models help critique, grade, and explain outputs (e.g. RLAIF, model-judging-model). Advancing scalable oversight methods, such as rubric-guided evaluations, recursive oversight, and weak-to-strong generalization.  Designing benchmarks for frontier capabilities (e.g. reasoning, coding, multi-modal, and agentic tasks), inspired by efforts like MMMU, GPQA, SWE-Bench. Building evaluation frameworks for agentic systems, measuring multi-step workflows and real-world task success. You will: Lead a team of research scientists and engineers on foundational work in evaluation and oversight. Drive research initiatives on frameworks and benchmarks for frontier AI models, spanning reasoning, coding, multi-modal, and agentic behaviors. Design and advance scalable oversight methods, leveraging model-assisted evaluation, rubric-guided judgments, and recursive oversight. Collaborate with leading research labs across industry and academia. Publish research at top-tier venues and contribute to open-source benchmarking initiatives. Remain deeply engaged with the research community, both understanding trends and setting them. Ideally you'd have: Track record of impactful research in machine learning, especially in generative AI, evaluation, or oversight. Significant experience leading ML research in academia or industry. Strong written and verbal communication skills for cross-functional collaboration. Experience building and mentoring teams of research scientists and engineers. Publications at major ML/AI conferences (e.g. NeurIPS, ICML, ICLR, ACL, EMNLP, CVPR) and/or journals.   Our research interviews are crafted to assess candidates' skills in practical ML prototyping and debugging, their grasp of research concepts, and their alignment with our organizational culture. We do not ask LeetCode-style questions.Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity based compensation, subject to Board of Director approval. Your recruiter can share more about the specific salary range for your preferred location during the hiring process, and confirm whether the hired role will be eligible for equity grant. You’ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.Please reference the job posting's subtitle for where this position will be located. For pay transparency purposes, the base salary range for this full-time position in the locations of San Francisco, New York, Seattle is:$260,000—$350,000 USDPLEASE NOTE: Our policy requires a 90-day waiting period before reconsidering candidates for the same role. This allows us to ensure a fair and thorough evaluation of all applicants. About Us: At Scale, we believe that the transition from traditional software to AI is one of the most important shifts of our time. Our mission is to make that happen faster across every industry, and our team is transforming how organizations build and deploy AI.  Our products power the world's most advanced LLMs, generative models, and computer vision models. We are trusted by generative AI companies such as OpenAI, Meta, and Microsoft, government agencies like the U.S. Army and U.S. Air Force, and enterprises including GM and Accenture. We are expanding our team to accelerate the development of AI applications. We believe that everyone should be able to bring their whole selves to work, which is why we are proud to be an inclusive and equal opportunity workplace. We are committed to equal employment opportunity regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability status, gender identity or Veteran status.  We are committed to working with and providing reasonable accommodations to applicants with physical and mental disabilities. If you need assistance and/or a reasonable accommodation in the application or recruiting process due to a disability, please contact us at accommodations@scale.com. Please see the United States Department of Labor's Know Your Rights poster for additional information. We comply with the United States Department of Labor's Pay Transparency provision.  PLEASE NOTE: We collect, retain and use personal data for our professional business purposes, including notifying you of job opportunities that may be of interest and sharing with our affiliates. We limit the personal data we collect to that which we believe is appropriate and necessary to manage applicants’ needs, provide our services, and comply with applicable laws. Any information we collect in connection with your application will be treated in accordance with our internal policies and programs designed to protect personal data. Please see our privacy policy for additional information.
Research Scientist
Product & Operations
Machine Learning Engineer
Data Science & Analytics
Apply
Hidden link
Articul8 AI.jpg

Senior/Staff AI Researcher (India)

Articul8
-
IN.svg
India
Full-time
Remote
true
About us:At Articul8 AI, we relentlessly pursue excellence and create exceptional AI products that exceed customer expectations. We are a team of dedicated individuals who take pride in our work and strive for greatness in every aspect of our business.Job Description:Articul8 AI is seeking an exceptional Senior/Staff AI Researcher to join us in shaping the future of Generative Artificial Intelligence (GenAI). As a member of our Applied Research team, you will be responsible for conducting cutting-edge research to advance the capabilities of our AI systems. Your role will involve designing, implementing, and evaluating novel approaches to improve our GenAI models, working at the intersection of GenAI research and product development.Responsibilities:Serve as the subject matter expert in various domains of GenAI research and development, including:Data pipelines: Design and optimize data processing workflows for large-scale model trainingTraining methodologies: Implement pre-training, mid-training, and post-training strategies and optimization techniquesReinforcement learning: Develop RL algorithms for GenAI and with applications in decision-making, personalization, and several other tasksMultimodal AI: Create systems that effectively process and generate across text, image, audio, and video modalitiesPersonalization: Design and implement tailored GenAI experiences by understanding user behavior, preferences, and contexts to deliver customized content and recommendationsKnowledge representation and retrieval: Develop techniques for effectively representation of information and knowledge elicitation, as well as search and retrieval.Play a technical leadership role in designing, developing, and scaling novel algorithms and models by taking them from research prototypes to production-ready systems that deliver real-world impact.Partner with cross-functional teams to integrate cutting-edge research findings into products and maintain our technological leadership in the market.Monitor and analyze emerging trends in generative AI and related fields, sharing valuable research contributions through publications at prestigious conferences and journals.Required Qualifications:Education: PhD/MSc degree in Computer Science, Machine Learning (ML), or a related field.Professional experience: 5+ years of experience as an AI researcher with a track record of applied research and/or product development (out of which, at least 2+ years should be on actively developing GenAI technologies).Core technical skills:Experience developing tools, libraries, and infrastructure for data preprocessing, model training/finetuning, and deployment of LLMs in research and production environments.A strong background in parallel/distributed computing on the cloud.Machine learning, deep learning, probability theory and statistics, natural language processing, computer vision, data wrangling and preparation, model evaluation and interpretation.Programming Skills: Proficiency in programming languages such as Python and experience working with version control systems (e.g., Git) and collaborating on code repositories is crucial.Preferred Qualifications:Experience with cloud computing platforms such as AWS, Azure, or GCP.Proven track record of publications in top-tier conferences and journals.Professional Attributes:Problem Solving: ability to break down complex problems into manageable components, devising creative solutions, and iteratively refining ideas based on feedback and experimental evidence.Collaboration and Communication: proficiency in working within cross-functional teams - communicating clearly, providing constructive criticism, delegating responsibilities, and respecting diverse perspectives.Critical Thinking: ability to carefully evaluate assumptions, questioning established methodologies, challenging own biases, and maintaining skepticism when interpreting results.Curiosity and Continuous Learning: ability to stay curious about advances in related fields and constantly seeking opportunities to expand knowledge base.Emotional Intelligence and Intellectual Humility: capable of displaying empathy, resilience, adaptability, and self-awareness. Ability to recognize own limitations, embracing uncertainty, acknowledging mistakes, and valuing others' contributions.If you're ready to join a team that's changing the game, apply now to become a part of the Articul8 team.
Research Scientist
Product & Operations
Machine Learning Engineer
Data Science & Analytics
Apply
Hidden link
AGI.jpg

AI Researcher

AGI, inc
-
US.svg
United States
Full-time
Remote
false
AI Researcher Job DescriptionAt AGI Inc., we're not just redefining AI-human interaction—we're creating a world where humans and AI amplify each other's potential. Leveraging breakthrough agentic reasoning capabilities, we aim to bring genuinely useful AGI into everyday life. Backed by years of research, AGI Inc. is pioneering the large-scale deployment of intelligent Consumer AI Agents that simplify, streamline, and elevate everyday experiences.Our team comprises elite ex-entrepreneurs, AI researchers, and product experts from Stanford, Berkeley, Nvidia, and DeepMind, supported by top investors in Silicon Valley. Our founder, Div Garg, is a Stanford AI PhD dropout and previously founded MultiOn, a leading AI Agent startup that introduced browser agents to the world, supported by General Catalyst, Forerunner, Samsung, and Amazon.At AGI Inc., we believe in fusing human insight with cutting-edge AI to craft experiences that are as intuitive as they are groundbreaking, paving the way toward everyday AGI. Our agentic products feel like a natural extension of who you are—an ever-present partner for ideas, projects, and solutions.About the RoleAs an AI Researcher at AGI Inc., you'll be at the forefront of developing novel algorithms and techniques for our browser-based intelligent agents. You'll bridge the gap between theoretical AI research and practical product applications, working on challenges that push the boundaries of what browser agents can accomplish. Your research will directly influence our product roadmap and contribute to the scientific community through publications and open-source contributions. This role offers a unique opportunity to work in an environment that values both academic rigor and real-world impact.Key ResponsibilitiesPioneer Novel Research: Lead research initiatives in areas such as reinforcement learning, multi-agent systems, and natural language understanding to enhance our browser agents' capabilities.Translate Research to Products: Work closely with engineering and product teams to implement research findings into products that deliver tangible value to users.Contribute to the Scientific Community: Publish high-quality research papers at top AI conferences (NeurIPS, ICML, ICLR, ACL) and participate in the open-source AI ecosystem.Develop Technical Roadmaps: Help define our technical research strategy, identifying promising areas that align with our product vision.Mentor and Collaborate: Work with and mentor other researchers and engineers, fostering a culture of scientific excellence and innovation.QualificationsEducation: MS or PhD in Computer Science, Machine Learning, AI, or a related technical field from a top university. Exceptional candidates with bachelor's degrees and equivalent research experience will also be considered.Research Experience: Demonstrated research experience in machine learning, NLP, reinforcement learning, or multimodal systems, evidenced by publications, open-source contributions, or impactful projects.Technical Skills: Strong programming skills in Python and experience with deep learning frameworks (PyTorch, TensorFlow). Familiarity with large language models, web technologies, and browser automation is a plus.Problem-Solving: Exceptional analytical thinking and creativity in approaching complex AI problems with novel solutions.Communication: Ability to explain complex technical concepts clearly and collaborate effectively with cross-functional teams.Startup Mindset: Enthusiasm for fast-paced environments, comfort with ambiguity, and a desire to build products that millions of people will use.Why Join Us?Dual Impact: Contribute to cutting-edge AI research while seeing your work directly impact products used by real people.Resources: Access to substantial compute resources, large datasets, and the latest AI technologies to support your research.Autonomy: Freedom to pursue research directions aligned with our mission, with less bureaucracy than traditional research labs.Equity: Competitive compensation package with significant equity, aligning your success with the company's growth.Community: Work alongside world-class researchers, engineers, and designers who are passionate about advancing the field of AI.How to ApplyWe'd love to see what you've built. Please include your resume, links to your publications or GitHub repositories, and a brief research statement outlining your interests and how they align with our mission.We can't wait to see your contributions to AI—and to welcome you to the AGI Inc. team!
Research Scientist
Product & Operations
Machine Learning Engineer
Data Science & Analytics
Apply
Hidden link
Deepgram.jpg

Research Staff, Voice AI Foundations

Deepgram
USD
220000
150000
-
220000
US.svg
United States
Full-time
Remote
true
Company OverviewDeepgram is the leading voice AI platform for developers building speech-to-text (STT), text-to-speech (TTS) and full speech-to-speech (STS) offerings. 200,000+ developers build with Deepgram’s voice-native foundational models – accessed through APIs or as self-managed software – due to our unmatched accuracy, latency and pricing. Customers include software companies building voice products, co-sell partners working with large enterprises, and enterprises solving internal voice AI use cases. The company ended 2024 cash-flow positive with 400+ enterprise customers, 3.3x annual usage growth across the past 4 years, over 50,000 years of audio processed and over 1 trillion words transcribed. There is no organization in the world that understands voice better than DeepgramThe OpportunityVoice is the most natural modality for human interaction with machines. However, current sequence modeling paradigms based on jointly scaling model and data cannot deliver voice AI capable of universal human interaction. The challenges are rooted in fundamental data problems posed by audio: real-world audio data is scarce and enormously diverse, spanning a vast space of voices, speaking styles, and acoustic conditions. Even if billions of hours of audio were accessible, its inherent high dimensionality creates computational and storage costs that make training and deployment prohibitively expensive at world scale. We believe that entirely new paradigms for audio AI are needed to overcome these challenges and make voice interaction accessible to everyone. The RoleAs a Member of the Research Staff, you will pioneer the development of Latent Space Models (LSMs), a new approach that aims to solve the fundamental data, scale, and cost challenges associated with building robust, contextualized voice AI. Your research will focus on solving one or more of the following problems:Build next-generation neural audio codecs that achieve extreme, low bit-rate compression and high fidelity reconstruction across a world-scale corpus of general audio.Pioneer steerable generative models that can synthesize the full diversity of human speech from the codec latent representation, from casual conversation to highly emotional expression to complex multi-speaker scenarios with environmental noise and overlapping speech.Develop embedding systems that cleanly factorize the codec latent space into interpretable dimensions of speaker, content, style, environment, and channel effects -- enabling precise control over each aspect and the ability to massively amplify an existing seed dataset through “latent recombination”.Leverage latent recombination to generate synthetic audio data at previously impossible scales, unlocking joint model and data scaling paradigms for audio.  Endeavor to train multimodal speech-to-speech systems that can 1) understand any human irrespective of their demographics, state, or environment and 2) produce empathic, human-like responses that achieve conversational or task-oriented objectives.   Design model architectures, training schemes, and inference algorithms that are adapted for hardware at the bare metal enabling cost efficient training on billion-hour datasets and powering real-time inference for hundreds of millions of concurrent conversations.The ChallengeWe are seeking researchers who:See "unsolved" problems as opportunities to pioneer entirely new approachesCan identify the one critical experiment that will validate or kill an idea in days, not monthsHave the vision to scale successful proofs-of-concept 100xAre obsessed with using AI to automate and amplify your own impactIf you find yourself energized rather than daunted by these expectations—if you're already thinking about five ideas to try while reading this—you might be the researcher we need. This role demands obsession with the problems, creativity in approach, and relentless drive toward elegant, scalable solutions. The technical challenges are immense, but the potential impact is transformative. It's Important to Us That You HaveStrong mathematical foundation in statistical learning theory, particularly in areas relevant to self-supervised and multimodal learningDeep expertise in foundation model architectures, with an understanding of how to scale training across multiple modalitiesProven ability to bridge theory and practice—someone who can both derive novel mathematical formulations and implement them efficientlyDemonstrated ability to build data pipelines that can process and curate massive datasets while maintaining quality and diversityTrack record of designing controlled experiments that isolate the impact of architectural innovations and validate theoretical insightsExperience optimizing models for real-world deployment, including knowledge of hardware constraints and efficiency techniquesHistory of open-source contributions or research publications that have advanced the state of the art in speech/language AI How We Generated This Job DescriptionThis job description was generated in two parts.  The “Opportunity”, “Role”, and “Challenge” sections were generated by a human using Claude-3.5-sonnet as a writing partner.  The objective of these sections is to clearly state the problem that Deepgram is attempting to solve, how we intend to solve it, and some guidelines to help you decide if Deepgram is right for you. Therefore, it is important that this section was articulated by a human.  The “It’s Important to Us” section was automatically derived from a multi-stage LLM analysis (using o1) of key foundational deep learning papers related to our research goals.  This work was completed as an experiment to test the hypothesis that traits of highly productive and impactful researchers are reflected directly in their work. The analysis focused on understanding how successful researchers approach problems, from mathematical foundations through to practical deployment. The problems Deepgram aims to solve are immensely difficult and span multiple disciplines and specialties. As such, we chose seminal papers that we believe reflect the pioneering work and exemplary human characteristics needed for success. The LLM analysis culminates in an “Ideal Researcher Profile”, which is reproduced below along with the list of foundational papers.  Ideal Researcher ProfileAn ideal researcher, as evidenced by the recurring themes across these foundational papers, excels in five key areas: (1) Statistical & Mathematical Foundations, (2) Algorithmic Innovation & Implementation, (3) Data-Driven & Scalable Systems, (4) Hardware & Systems Understanding, and (5) Rigorous Experimental Design. Below is a synthesis of how each paper highlights these qualities, with references illustrating why they matter for building robust, impactful deep learning models. 1. Statistical & Mathematical FoundationsMastery of Core ConceptsMany papers, like Scaling Laws for Neural Language Models and Neural Discrete Representation Learning (VQ-VAE), reflect the importance of power-law analyses, derivation of novel losses, or adaptation of fundamental equations (e.g., in VQ-VAE's commitment loss or rectified flows in Scaling Rectified Flow Transformers). Such mathematical grounding clarifies why models converge or suffer collapse.Combining Existing Theories in Novel WaysPapers such as Moshi (combining text modeling, audio codecs, and hierarchical generative modeling) and Finite Scalar Quantization (FSQ's adaptation of classic scalar quantization to replace vector-quantized representations) show how reusing but reimagining known techniques can yield breakthroughs. Many references (e.g., the structured state-space duality in Transformers are SSMs) underscore how unifying previously separate research lines can reveal powerful algorithmic or theoretical insights.Logical Reasoning and Assumption TestingAcross all papers—particularly in the problem statements of Whisper or Rectified Flow Transformers—the authors present assumptions (e.g., "scaling data leads to zero-shot robustness" or "straight-line noise injection improves sample efficiency") and systematically verify them with thorough empirical results. An ideal researcher similarly grounds new ideas in well-formed, testable hypotheses. 2. Algorithmic Innovation & ImplementationCreative Solutions to Known BottlenecksEach paper puts forth a unique algorithmic contribution—Rectified Flow Transformers redefines standard diffusion paths, FSQ proposes simpler scalar quantizations contrasted with VQ, phi-3 mini relies on curated data and blocksparse attention, and Mamba-2 merges SSM speed with attention concepts.Turning Theory into PracticeWhether it's the direct preference optimization (DPO) for alignment in phi-3 or the residual vector quantization in SoundStream, these works show that bridging design insights with implementable prototypes is essential.Clear Impact Through Prototypes & Open-SourceMany references (Whisper, neural discrete representation learning, Mamba-2) highlight releasing code or pretrained models, enabling the broader community to replicate and build upon new methods. This premise of collaboration fosters faster progress. 3. Data-Driven & Scalable SystemsEmphasis on Large-Scale Data and Efficient PipelinesPapers such as Robust Speech Recognition via Large-Scale Weak Supervision (Whisper) and BASE TTS demonstrate that collecting and processing hundreds of thousands of hours of real-world audio can unlock new capabilities in zero-shot or low-resource domains. Meanwhile, phi-3 Technical Report shows that filtering and curating data at scale (e.g., "data optimal regime") can yield high performance even in smaller models.Strategic Use of Data for Staged TrainingA recurring strategy is to vary sources of data or the order of tasks. Whisper trains on multilingual tasks, BASE TTS uses subsets/stages for pretraining on speech tokens, and phi-3 deploys multiple training phases (web data, then synthetic data). This systematic approach to data underscores how an ideal researcher designs training curricula and data filtering protocols for maximum performance. 4. Hardware & Systems UnderstandingEfficient Implementations at ScaleMany works illustrate how researchers tune architectures for modern accelerators: the In-Datacenter TPU paper exemplifies domain-specific hardware design for dense matrix multiplications, while phi-3 leverages blocksparse attention and custom Triton kernels to run advanced LLMs on resource-limited devices.Real-Time & On-Device ConstraintsSoundStream shows how to compress audio in real time on a smartphone CPU, demonstrating that knowledge of hardware constraints (latency, limited memory) drives design choices. Similarly, Moshi's low-latency streaming TTS and phi-3-mini's phone-based inference highlight that an ideal researcher must adapt algorithms to resource limits while maintaining robustness.Architectural & Optimization DetailsPapers like Mamba-2 in Transformers are SSMs and the In-Datacenter TPU work show how exploiting specialized matrix decomposition, custom memory hierarchies, or quantization approaches can lead to breakthroughs in speed or energy efficiency. 5. Rigorous Experimental DesignControlled Comparisons & AblationsNearly all papers—Whisper, FSQ, Mamba-2, BASE TTS—use systematic ablations to isolate the impact of individual components (e.g., ablation on vector-quantization vs. scalar quantization in FSQ, or size of codebooks in VQ-VAEs). This approach reveals which design decisions truly matter.Multifold Evaluation MetricsFrom MUSHRA listening tests (SoundStream, BASE TTS) to FID in image synthesis (Scaling Rectified Flow Transformers, FSQ) to perplexity or zero-shot generalization in language (phi-3, Scaling Laws for Neural Language Models), the works demonstrate the value of comprehensive, carefully chosen metrics.Stress Tests & Edge CasesWhisper's out-of-distribution speech benchmarks, SoundStream's evaluation on speech + music, or Mamba-2's performance on multi-query associative recall demonstrate the importance of specialized challenge sets. Researchers who craft or adopt rigorous benchmarks and "red-team" their models (as in phi-3 safety alignment) are better prepared to address real-world complexities. SummaryOverall, an ideal researcher in deep learning consistently demonstrates:A solid grounding in theoretical and statistical principlesA talent for proposing and validating new algorithmic solutionsThe capacity to orchestrate data pipelines that scale and reflect real-world diversityAwareness of hardware constraints and system-level trade-offs for efficiencyThorough and transparent experimental practicesThese qualities surface across research on speech (Whisper, BASE TTS), language modeling (Scaling Laws, phi-3), specialized hardware (TPU, Transformers are SSMs), and new representation methods (VQ-VAE, FSQ, SoundStream). By balancing these attributes—rigorous math, innovative algorithms, large-scale data engineering, hardware-savvy optimizations, and reproducible experimentation—researchers can produce impactful, trustworthy advancements in foundational deep learning. Foundational PapersThis job description was generated through analysis of the following papers:Robust Speech Recognition via Large-Scale Weak Supervision (arXiv:2212.04356)Moshi: a speech-text foundation model for real-time dialogue (arXiv:2410.00037)Scaling Rectified Flow Transformers for High-Resolution Image Synthesis (arXiv:2403.03206)Scaling Laws for Neural Language Models (arXiv:2001.08361)BASE TTS: Lessons from building a billion-parameter Text-to-Speech model on 100K hours of data (arXiv:2402.08093)In-Datacenter Performance Analysis of a Tensor Processing Unit (arXiv:1704.04760)Neural Discrete Representation Learning (arXiv:1711.00937)SoundStream: An End-to-End Neural Audio Codec (arXiv:2107.03312)Finite Scalar Quantization: VQ-VAE Made Simple (arXiv:2309.15505)Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone (arXiv:2404.14219)Transformers are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality (arXiv:2405.21060)Backed by prominent investors including Y Combinator, Madrona, Tiger Global, Wing VC and NVIDIA, Deepgram has raised over $85 million in total funding. If you're looking to work on cutting-edge technology and make a significant impact in the AI industry, we'd love to hear from you!Deepgram is an equal opportunity employer. We want all voices and perspectives represented in our workforce. We are a curious bunch focused on collaboration and doing the right thing. We put our customers first, grow together and move quickly. We do not discriminate on the basis of race, religion, color, national origin, gender, sexual orientation, gender identity or expression, age, marital status, veteran status, disability status, pregnancy, parental status, genetic information, political affiliation, or any other status protected by the laws or regulations in the locations where we operate.We are happy to provide accommodations for applicants who need them.
Research Scientist
Product & Operations
Machine Learning Engineer
Data Science & Analytics
Apply
Hidden link
Deepgram.jpg

Research Staff, LLMs

Deepgram
USD
0
150000
-
220000
US.svg
United States
Full-time
Remote
true
Company OverviewDeepgram is the leading voice AI platform for developers building speech-to-text (STT), text-to-speech (TTS) and full speech-to-speech (STS) offerings. 200,000+ developers build with Deepgram’s voice-native foundational models – accessed through APIs or as self-managed software – due to our unmatched accuracy, latency and pricing. Customers include software companies building voice products, co-sell partners working with large enterprises, and enterprises solving internal voice AI use cases. The company ended 2024 cash-flow positive with 400+ enterprise customers, 3.3x annual usage growth across the past 4 years, over 50,000 years of audio processed and over 1 trillion words transcribed. There is no organization in the world that understands voice better than DeepgramThe OpportunityVoice is the most natural modality for human interaction with machines. However, current sequence modeling paradigms based on jointly scaling model and data cannot deliver voice AI capable of universal human interaction. The challenges are rooted in fundamental data problems posed by audio: real-world audio data is scarce and enormously diverse, spanning a vast space of voices, speaking styles, and acoustic conditions. Even if billions of hours of audio were accessible, its inherent high dimensionality creates computational and storage costs that make training and deployment prohibitively expensive at world scale. We believe that entirely new paradigms for audio AI are needed to overcome these challenges and make voice interaction accessible to everyone. The RoleDeepgram is currently looking for an experienced researcher to who has worked extensively with Large Language Models (LLMS) and has a deep understanding of transformer architecture to join our Research Staff. As a Member of the Research Staff, this individual should have extensive experience working on the hard technical aspects of LLMs, such as data curation, distributed large-scale training, optimization of transformer architecture, and Reinforcement Learning (RL) training.The ChallengeWe are seeking researchers who:See "unsolved" problems as opportunities to pioneer entirely new approachesCan identify the one critical experiment that will validate or kill an idea in days, not monthsHave the vision to scale successful proofs-of-concept 100xAre obsessed with using AI to automate and amplify your own impactIf you find yourself energized rather than daunted by these expectations—if you're already thinking about five ideas to try while reading this—you might be the researcher we need. This role demands obsession with the problems, creativity in approach, and relentless drive toward elegant, scalable solutions. The technical challenges are immense, but the potential impact is transformative.What You'll DoBrainstorming and collaborating with other members of the Research Staff to define new LLM research initiativesBroad surveying of literature, evaluating, classifying, and distilling current methodsDesigning and carrying out experimental programs for LLMsDriving transformer (LLM) training jobs successfully on distributed compute infrastructure and deploying new models into productionDocumenting and presenting results and complex technical concepts clearly for a target audienceStaying up to date with the latest advances in deep learning and LLMs, with a particular eye towards their implications and applications within our productsYou'll Love This Role if YouAre passionate about AI and excited about working on state of the art LLM researchHave an interest in producing and applying new science to help us develop and deploy large language modelsEnjoy building from the ground up and love to create new systems.Have strong communication skills and are able to translate complex concepts clearlyAre highly analytical and enjoy delving into detailed analyses when necessary It's Important to Us That You Have3+ years of experience in applied deep learning research, with a solid understanding toward the applications and implications of different neural network types, architectures, and loss mechanismProven experience working with large language models (LLMs) - including experience with data curation, distributed large-scale training, optimization of transformer architecture, and RL LearningStrong experience coding in Python and working with PytorchExperience with various transformer architectures (auto-regressive, sequence-to-sequence.etc)Experience with distributed computing and large-scale data processingPrior experience in conducting experimental programs and using results to optimize modelsIt Would Be Great if You HadDeep understanding of transformers, causal LMs, and their underlying architectureUnderstanding of distributed training and distributed inference schemes for LLMsFamiliarity with RLHF labeling and training pipelinesUp-to-date knowledge of recent LLM techniques and developmentsPublished papers in Deep Learning Research, particularly related to LLMs and deep neural networksBacked by prominent investors including Y Combinator, Madrona, Tiger Global, Wing VC and NVIDIA, Deepgram has raised over $85 million in total funding. If you're looking to work on cutting-edge technology and make a significant impact in the AI industry, we'd love to hear from you!Deepgram is an equal opportunity employer. We want all voices and perspectives represented in our workforce. We are a curious bunch focused on collaboration and doing the right thing. We put our customers first, grow together and move quickly. We do not discriminate on the basis of race, religion, color, national origin, gender, sexual orientation, gender identity or expression, age, marital status, veteran status, disability status, pregnancy, parental status, genetic information, political affiliation, or any other status protected by the laws or regulations in the locations where we operate.We are happy to provide accommodations for applicants who need them.
Research Scientist
Product & Operations
Machine Learning Engineer
Data Science & Analytics
Apply
Hidden link
Parallel.jpg

Research Scientist

Parallel
-
US.svg
United States
Full-time
Remote
false
At Parallel Web Systems, we are bringing a new web to life: it’s built with, by, and for AIs. Our work spans innovations across crawling, indexing, ranking, retrieval, and reasoning systems. Our first product is a set of APIs for AIs to do more with web data. We are a fully in-person team based in Palo Alto, CA. Our organization is flat; our team is small and talent dense.We want to talk to you if you are someone who can bring us closer to living our aspirational values:Own customer impact - It’s on us to ensure real-world outcomes for our customers.Obsess over craft - Perfect every detail because quality compounds.Accelerate change - Ship fast, adapt faster, and move frontier ideas into production.Create win-wins - Creatively turn trade-offs into upside.Make high-conviction bets - Try and fail. But succeed an unfair amount.Job: Our first dedicated research hire - you will answer the question: how to train and scale a model that can serve a web index?You: Have deep intuition on modern models and training. Like to argue how search, recommendations, and transformer models can converge. You care about your research being applied to product and systems that millions use.Our founder is Parag Agrawal. Previously, he was the CEO and CTO at Twitter. Our investors include First Round Capital, Index Ventures, Khosla Ventures, and many others.
Research Scientist
Product & Operations
Apply
Hidden link
Maincode.jpg

AI Research Group Leader

Maincode
-
AU.svg
Australia
Full-time
Remote
false
About the jobMaincode is building sovereign AI models in Australia. We train foundation models from scratch, design new reasoning architectures, and deploy them on state-of-the-art GPU clusters. This is not fine-tuning someone else’s work. This is creating new systems from first principles. We’re hiring a research leader for our AI Residency program. Our AI Residency Program is a focused research collaboration with top academic talent: late-stage PhD students, postdocs, and professors, working on foundational AI research at the architectural level. Each cohort of residents joins us for 6-months to pursue ambitious projects that align with our mission, combining academic rigor with industry-scale infrastructure and execution. This is a full-time, permanent role with ongoing responsibility for sourcing and selecting residents, guiding them through their projects, supporting publication efforts, and ensuring impactful, aligned research outcomes. The program is already up and running, now we’re looking for a research leader to own its day-to-day execution. You’ll make sure the right people are selected, the right projects are pursued, and the research produced is impactful, publishable, and aligned with Maincode’s long-term goals. You’ll also lead sourcing, interviewing, and hiring of residents, and ensure each one’s experience is productive and inspiring. This is a hybrid role based in Australia.What you’ll doRun the AI Residency Program end-to-end, ensuring smooth operations, clear timelines, and high-impact outcomes.Source, recruit, and hire outstanding residents, managing the full candidate pipeline from outreach to signed offers.Select and greenlight projects in consultation with Maincode’s research leadership to ensure strategic alignment.Guide and mentor residents throughout their 6-month program, drawing on your academic and industry experience.Maintain research quality, ensuring outputs meet top-tier publication standards and practical applicability.Coordinate with internal research teams so residency work complements and extends Maincode’s own research agenda.Represent the program externally, building relationships with universities, research groups, and other AI communities. Who you areA senior academic researcher (postdoc or above) with deep expertise in AI and a broad understanding of the field’s directions.Experienced in supervising PhD-level research and managing multiple projects simultaneously.Skilled in candidate sourcing, interviewing, and selection for high-level research roles.Have a strong publication record in top AI/ML venues (NeurIPS, ICLR, ICML, etc.).Strong strategic judgment in choosing projects and people that balance novelty with long-term impact.A confident communicator who can work seamlessly with both academic researchers and industry engineers.Motivated to strengthen sovereign AI capability in Australia through academic–industry collaboration.Why Maincode We are a small, mission-driven team building some of the most advanced AI systems in Australia. We operate our own GPU clusters, run large-scale training, and work closely across research and engineering to push the frontier of what’s possible. You’ll be surrounded by people who:Care deeply about model internals, not just outputs.Build things that work at scale.Take pride in learning, experimenting, and shipping.Believe Australia must have independent, world-class AI systems.If you want to run a program that brings world-class researchers into an environment where they can test ideas at scale and make a lasting impact on the science of AI, we’d love to hear from you.
Research Scientist
Product & Operations
Apply
Hidden link
Oumi.jpg

Research Scientist (Greece)

Oumi
USD
100000
-
220000
No items found.
Full-time
Remote
true
About OumiWhy we exist: Oumi is on a mission to make frontier AI truly open for all. We are founded on the belief that AI will have a transformative impact on humanity, and that developing it collectively, in the open, is the best path forward to ensure that it is done efficiently and safely.What we do: Oumi provides an all-in-one platform to build state-of-the-art AI models, end to end, from data preparation to production deployment, empowering innovators to build cutting-edge models at any scale. Oumi also develops open foundation models in collaboration with academic collaborators and the open community.Our Approach: Oumi is fundamentally an open-source first company, with open-collaboration across the community as a core principle. Our work is:Open Source First: All our platform and core technology is open sourceResearch-driven: We conduct and publish original research in AI, collaborating with our community of academic research labs and collaboratorsCommunity-powered: We believe in the power of open-collaboration and welcome contributions from researchers and developers worldwideRole OverviewThe Research Scientist will be an integral part of Oumi's research team, focusing on advancing the state-of-the-art in large language models (LLMs), vision language models (VLMs), and related technologies. This role involves conducting cutting-edge research, contributing to open-source projects, and collaborating with other researchers and engineers. Researchers at Oumi will work on various aspects of LLM/VLM development including training, evaluation, data curation, and benchmark development.What you’ll do:Model Development: Conduct research on training and evaluating new Large language models (LLMs), Vision Language Models (VLMs), and other AI models. This includes exploring new architectures, training techniques, and optimization methods.Data Curation: Develop methodologies for curating high-quality datasets for training and evaluating LLMs. This may involve data synthesis and other novel techniques.Benchmark Development: Develop evaluation benchmarks to measure the performance of LLMs across various tasks and domains.Research and Experimentation: Design and conduct experiments to validate research hypotheses and improve model performance.Open Source Contribution: Contribute to the Oumi open-source platform, models and projects, and other relevant tools and libraries.Collaboration: Collaborate with other researchers, engineers, and the broader community to advance the field of open-source AI.Publication: Publish research findings in leading conferences and journals.Platform Evaluation: Evaluate existing models and identify areas of improvement.Flexibility: Work with various models, including text and multimodal models, and both open and closed models.Problem Solving: Focus on the research that matters by skipping the plumbing and moving straight to research, building on the work of others and contributing back.What you’ll bring:Education: A Ph.D. or MSc. in computer science, machine learning, artificial intelligence, or a related field is preferred. Candidates with a strong publication record, or equivalent industry experience will be considered.Research Experience: Demonstrated experience in conducting original research in machine learning, with a strong publication record in top-tier conferences or journals.ML Expertise: Deep understanding of machine learning and deep learning concepts, with specific knowledge of large language models (LLMs) and/or vision language models (VLMs).Programming Skills: Strong programming skills in Python and experience using deep learning frameworks (e.g. PyTorch).Open Source: Familiarity with open-source projects and a passion for contributing to the open-source community.Initiative: A self-starter who can work independently and take ownership of initiatives.Values: Share Oumi's values: Beneficial for all, Customer-obsessed, Radical Ownership, Exceptional Teammates, Science-grounded.BenefitsCompetitive salary: $100,000 - $220,000Equity in a high-growth startupComprehensive health, dental and vision insurance21 days PTORegular team offsites and events
Research Scientist
Product & Operations
Machine Learning Engineer
Data Science & Analytics
Apply
Hidden link
Maincode.jpg

AI Researcher

Maincode
AUD
180000
150000
-
180000
AU.svg
Australia
Full-time
Remote
false
Maincode is building sovereign AI models in Australia. We are training foundation models from scratch, designing new reasoning architectures, and deploying them on state-of-the-art GPU clusters. This is not fine-tuning someone else’s work. This is building new systems from first principles.As an AI Researcher, you’ll help design the next generation of foundation models. You'll explore novel model architectures, training objectives, and data representations that push the limits of what models can do. You’ll ask the hard questions: where do today’s systems fall short in reasoning, learning efficiency, or alignment, and how could we do better? Then, you'll prototype solutions to find out.This is a deep applied research role. You'll write code to test your ideas, run experiments, and refine architectures until they show real-world potential. When your models are ready to scale, you’ll work side by side with our AI Engineers to take them from prototypes to large-scale training runs and deployed systems.If you love reading the latest papers, imagining what could be better, and proving it with working models, you'll feel right at home here.What you’ll doDesign and prototype new foundation model architectures, reasoning systems, and training algorithmsExplore better ways to represent data, design training objectives, and build models that reason, learn, and align more effectivelyBuild experimental models and pipelines to test your ideas on real-world dataStay deeply engaged with state-of-the-art research in venues like NeurIPS, ICLR, and ICML, and think critically about where the field is heading and how we can push it furtherWork closely with AI Engineers to transition prototypes into large-scale training and inference systemsHelp shape the intellectual direction of our sovereign AI efforts in Australia, advancing the science of how models learn and reasonContribute to a collaborative, open-minded team culture that values rigor, creativity, and building things that workWho you areAn experienced researcher with deep academic or applied horsepower, whether from machine learning, physics, neuroscience, applied mathematics, or another rigorous fieldHave built, trained, or experimented with deep learning models and are deeply curious about how to make them betterA regular reader and thinker in AI and machine learning research, with the ability to spot emerging trends, gaps in today’s models, and new directions to exploreSkilled in Python and machine learning frameworks like PyTorch, with a strong focus on hands-on experimentationComfortable turning abstract research questions into concrete model designs, experiments, and insightsExcited to work closely with AI Engineers to translate research into robust, scalable systemsMotivated to help build sovereign AI capability here in AustraliaWhy MaincodeWe are a small team building some of the most advanced AI systems in Australia. We are creating new foundation models from scratch, not just using what’s already out there.We operate our own GPU clusters, run large-scale training, and work closely across research and engineering to push the frontier of what’s possible.You'll be surrounded by people who:Care about model internals, not just outputsBuild things that work, at scaleTake pride in learning, experimenting, and shippingWant to help Australia build independent, world-class AI systems
Research Scientist
Product & Operations
Machine Learning Engineer
Data Science & Analytics
Apply
Hidden link
Together AI.jpg

Research Scientist, Post-Training

Together AI
USD
0
225000
-
300000
US.svg
United States
NL.svg
Netherlands
GB.svg
United Kingdom
Full-time
Remote
false
About Model Shaping The Model Shaping team at Together AI works on products and research for tailoring open foundation models to downstream applications. We build services that allow machine learning developers to choose the best models for their tasks and further improve these models using domain-specific data. In addition to that, we develop new methods for more efficient model training and evaluation, drawing inspiration from a broad spectrum of ideas across machine learning, natural language processing, and ML systems. About the Role As a Research Scientist in Post-Training, you will advance the methods for making foundation models more useful and reliable. You will analyze the limitations of current approaches to post-training across domains such as data curation, reinforcement learning, and model evaluation. Based on this analysis, you will design new techniques for model adaptation, as well as benchmarks for measuring the progress in the field. After evaluating your ideas through experimentation, you will present your findings to the global scientific community at leading ML/NLP conferences and collaborate with your teammates to integrate those improvements into Together’s platform. Responsibilities Define and drive the research agenda around efficiency and performance of foundation model training Study recent results from the broader AI research community, analyzing their relevance to the team’s research directions and ongoing projects Conduct experiments to empirically validate your hypotheses and compare the outcomes with relevant related work Share your findings both internally and externally (e.g., at top-tier conferences on ML and NLP) Partner with Machine Learning Engineers to integrate advanced methods into Together’s Model Shaping platform Requirements Can autonomously design, implement, and validate your research ideas Skilled at writing high-quality and efficient code in Python and PyTorch Have first-author publications at leading conferences on ML or NLP (ICLR, ICML, NeurIPS, ACL, EMNLP) Are a strong communicator, ready to both discuss your research plans with other scientists and explain them to broader audience Follow the latest advances in relevant subfields of AI Are passionate about seeing your research create real-world impact through Together AI's services and willing to work hands-on with production systems to achieve it Stand-out experience: Reinforcement learning of language models Curation of pre-training or post-training datasets and benchmarks Robust evaluation of foundation models Running large-scale experiments on GPU clusters About Together AI Together AI is a research-driven artificial intelligence company. We believe open and transparent AI systems will drive innovation and create the best outcomes for society, and together we are on a mission to significantly lower the cost of modern AI systems by co-designing software, hardware, algorithms, and models. We have contributed to leading open-source research, models, and datasets to advance the frontier of AI, and our team has been behind technological advancements such as FlashAttention, RedPajama, SWARM Parallelism, and SpecExec. We invite you to join a passionate group of researchers in our journey in building the next generation AI infrastructure. Compensation We offer competitive compensation, startup equity, health insurance, and other benefits, as well as flexibility in terms of remote work. The US base salary range for this full-time position is $225,000 - $300,000. Our salary ranges are determined by location, level and role. Individual compensation will be determined by experience, skills, and job-related knowledge. Equal Opportunity Together AI is an Equal Opportunity Employer and is proud to offer equal employment opportunity to everyone regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability, gender identity, veteran status, and more. Please see our privacy policy at https://www.together.ai/privacy
Research Scientist
Product & Operations
Apply
Hidden link
No job found
There is no job in this category at the moment. Please try again later