⚠️ Sorry, this job is no longer available.

AI Software Engineer Jobs

Latest roles in AI Software Engineer, reviewed by real humans for quality and clarity.

All Jobs

Showing 6179  of 79 jobs
Retell AI.jpg

Founding Senior Machine Learning Engineer

Retell AI
USD
280000
200000
-
280000
US.svg
United States
Full-time
Remote
false
ABOUT RETELL AIRetell AI is using the first principles to reimagine the call center with cutting edge voice AI.We believe voice is still the most natural way humans communicate, yet it has been trapped in outdated call centers for decades. Our mission is to bring intelligence, empathy, and speed to every phone conversation between businesses and their customers.Since launching 18 months ago, thousands of companies now utilize Retell’s AI voice agents to handle sales, support, and logistics calls that once required large teams of human agents. Backed by Y Combinator, Altman Capital, and other leading investors, we have scaled to $30M ARR with a team of 20 people, up from $5M at the start of 2025.Now, we’re scaling fast, and we’re looking for bold, ambitious people to help us build the gold standard for voice automation. If you want to work on deeply technical challenges, move fast, and make an outsized impact at one of the fastest-growing Voice AI startups in the world, you’ll love it here.Let’s build the future of voice automation together.We’re a top 50 AI app in a16z list: https://tinyurl.com/5853dt2xWe're also one of the top ranking startups on: https://leanaileaderboard.com/ABOUT THE ROLEThis is a hands-on, high-ownership role for ML engineers who want to build production models that actually ship, and perform under real-world constraints. As a Founding Senior Machine Learning Engineer at Retell, you’ll work across the ML stack to power human-like voice agents that handle millions of real-time phone conversations.You’ll fine-tune large language models and audio models, evaluate them with rigorous benchmarks (and human feedback), and deploy them into latency-sensitive, high-traffic systems. You’ll own model performance end-to-end—from training pipelines to post-deployment monitoring—and shape our ML strategy alongside the founding team.If you’re excited by hard technical challenges, fast iteration, and the opportunity to define how voice AI works at scale, this role is a rare chance to do it from the ground up.KEY RESPONSIBILITIESTrain & Tune Models – Fine-tune LLMs and audio models to maximize speed, accuracy, and production-readiness—pushing the frontier of real-time AI voice experiences.Benchmark & Evaluate – Build datasets, define rigorous metrics, and measure model performance across high-impact voice AI tasks to guide development.Deploy to Production – Work closely with engineering to ship models, monitor them in the wild, and ensure they stay fast, reliable, and accurate at scale.Run Human Evaluations – Build scalable pipelines to collect structured human feedback, benchmark subjective quality, and inform model iterations.Level Up Infrastructure – Design and maintain the ML infrastructure needed for fast experimentation, robust training, and continuous deployment. YOU MIGHT THRIVE IF YOUML Engineer with Real-World Experience – You’ve trained and shipped models in production. Bonus if you’ve worked with LLMs or audio models.Fluent in Modern ML Stack – You know your way around Python, PyTorch, and today’s ML tools—from training pipelines to evaluation benchmarks.Execution-Oriented – You move fast, take ownership, and focus on solving real problems over perfect ones.Startup-Ready – You’re adaptable, resilient, and energized by ambiguity and fast-changing priorities.Clear Communicator & Team Player – You collaborate well across functions and push decisions forward. JOB DETAILSJob Type: Full-time, 70 hr/week (50 hr/week onsite with flexible hours + 20 hr/week work from home)Cash: $200,000 - $280,000 base salary Equity: .0.15% – 0.35%Location: Redwood City, CA, USUS Visas: Retell AI is open to sponsoring work authorization for qualified candidates, including H1B/H-1B, TN, L-1, E-3, F-1 (OPT/CPT), and O-1 visas. OTHER BENEFITS100% coverage for medical, dental, and vision insurance$70/day DoorDash credit for unlimited breakfast, lunch, dinner, and snacks$200/month wellness reimbursement (gym, fitness classes, etc.)$300/month commuter reimbursement (gas, Caltrain, etc.)$75/month phone bill reimbursement$50/month internet reimbursement COMPENSATION PHILOSOPHYBest Offer Upfront: Choose from three cash-equity balance options, no negotiation needed.Top 1% Talent: Above-market pay (top 5 percentile) to attract high performers.High Ownership: Small teams, >$1M revenue/employee, and significant equity.Performance-Based: Offers tied to interview performance, not experience or past salaries. INTERVIEW PROCESSOnline Assessment (30 min): Coding questions on practical problem-solving (7 days to complete).Talent Screen (15min): chat with our recruiter to get a better sense of the role, the team, and what it’s like to work here.Technical Interview (45 min): Machine Learning specific coding Interview.Technical Interview (45 min): Live Practical Systems Design and Coding Interview.Onsite/Virtual Interviews (3 hrs): Hosted in our office if located in the Bay Area or virtual, with three rounds:ML System Design: A non-coding interview focused on whiteboarding and high-level system architecture.ML Question Deep Dive: In-depth discussion exploring your approach to a machine learning problem.Backend + AI Practical: A hands-on coding interview combining backend development with AI integration.Offer: Final stage, pending decision and offer discussion.
Machine Learning Engineer
Data Science & Analytics
Hidden link
Tavus.jpg

AI Researcher (Multimodal Audio/Video Generation)

Tavus
-
US.svg
United States
GB.svg
United Kingdom
Full-time
Remote
false
About UsTavus is a research lab pioneering human computing. We’re building AI Humans: a new interface that closes the gap between people and machines, free from the friction of today’s systems. Our real-time human simulation models let machines see, hear, respond, and even look real—enabling meaningful, face-to-face conversations. AI Humans combine the emotional intelligence of humans with the reach and reliability of machines, making them capable, trusted agents available 24/7, in every language, on our terms.Imagine a therapist anyone can afford. A personal trainer that adapts to your schedule. A fleet of medical assistants that can give every patient the attention they need. With Tavus, individuals, enterprises, and developers can all build AI Humans to connect, understand, and act with empathy at scale.We’re a Series A company backed by world-class investors including Sequoia Capital, Y Combinator, and Scale Venture Partners.Be part of shaping a future where humans and machines truly understand each other.The Role We’re looking for an AI Researcher to join our core AI team and push forward the science of audio-visual avatar generation. If you thrive in high-speed startup environments, enjoy experimenting with generative models, and love seeing your research ship into production then you’ll feel right at home.Your Mission 🚀Research and develop audio-visual generation models for conversational agents (e.g. Neural Avatars, Talking-Heads).Focus on models that are tightly coupled with conversation flow, ensuring verbal and non-verbal signals work seamlessly together.Experiment with diffusion models (DDPMs, LDMs, etc.), long-video generation, and audio generation.Collaborate with the Applied ML team to bring your research into real-world production.Stay ahead of the latest advancements in multimodal generation — and help shape the next wave.You’ll Be Great At This If You Have:A PhD (or near completion) in a relevant field, or equivalent hands-on research experience.Experience applying image/video generation models in practice.Strong foundations in generative modeling and rapid prototyping.Deep familiarity with diffusion models, including recent advances in efficiency.Good understanding of video-language models and multimodal generation.Proficiency in PyTorch and GPU-based inference.Nice-to-HavesExperience with long-video or audio generation.Skills in 3D graphics, Gaussian splatting, or large-scale training setups.Broader exposure to generative models and rendering.Familiarity with software engineering best practices.Publications in top-tier or respected venues (CVPR, NeurIPS, BMVC, ICASSP, etc.).Location Preferred: San Francisco (hybrid) or London (office opening soon). Remote within U.S. or Europe available for exceptional candidates.
Machine Learning Engineer
Data Science & Analytics
Research Scientist
Product & Operations
Hidden link
Tavus.jpg

Senior+ AI Researcher (Large Language Models)

Tavus
0
0
-
0
US.svg
United States
Full-time
Remote
false
About UsTavus is a research lab pioneering human computing. We’re building AI Humans: a new interface that closes the gap between people and machines, free from the friction of today’s systems. Our real-time human simulation models let machines see, hear, respond, and even look real; enabling meaningful, face-to-face conversations. AI Humans combine the emotional intelligence of humans with the reach and reliability of machines, making them capable, trusted agents available 24/7, in every language, on our terms.Imagine a therapist anyone can afford. A personal trainer that adapts to your schedule. A fleet of medical assistants that can give every patient the attention they need. With Tavus, individuals, enterprises, and developers can all build AI Humans to connect, understand, and act with empathy at scale.We’re a Series A company backed by world-class investors including Sequoia Capital, Y Combinator, and Scale Venture Partners.Be part of shaping a future where humans and machines truly understand each other.The Role We’re hiring a Senior AI Researcher to lead research on large language models for next-generation conversational avatars. This is a role for someone who doesn’t just want to follow the roadmap but you’ll help write it. You’ll drive research initiatives that define how avatars can think, adapt, and communicate naturally, across both verbal and non-verbal channels.Your Mission 🚀Lead research on LLMs and VLMs for Conversational Avatars, with a focus on modeling verbal + non-verbal interactions.Design and implement fine-tuning, adaptation, and conditioning techniques to control LLM behavior and align it with real-world conversational needs.Prototype, train, and optimize models that can operate in real-time generation settings with time budget constraints.Collaborate with the Applied ML and Product teams to move research into impactful production systems.Mentor other researchers, set research directions, and foster scientific excellence.You’ll Bring:A PhD or equivalent experience, plus 2–3+ years of postdoctoral or industry research on LLMs/VLMs.Proven expertise in generative language modeling and demonstrated innovation through prototyping.Strong track record in LLM adaptation, conditioning, or control methods.Experience with large model training at scale.Excellent programming skills (especially in PyTorch) and ability to move seamlessly between research and engineering tasks.Publications in top-tier conferences (EMNLP, COLING, NeurIPS, ICLR, CVPR, ICCV).Experience leading research efforts or directing technical agendas within teams.Nice-to-HavesBroader understanding of generative AI beyond text (multimodal, speech, vision).Strong grounding in deep learning methodologies that generalize across domains.Familiarity with best practices in software engineering and reproducible ML.Location Preferred: San Francisco (hybrid) or London (office opening soon). Remote within the U.S. or Europe considered for exceptional talent.
No items found.
Hidden link
Tavus.jpg

AI Researcher (Large Language Models)

Tavus
-
US.svg
United States
GB.svg
United Kingdom
Full-time
Remote
false
About UsTavus is a research lab pioneering human computing. We’re building AI Humans: a new interface that closes the gap between people and machines, free from the friction of today’s systems. Our real-time human simulation models let machines see, hear, respond, and even look real—enabling meaningful, face-to-face conversations. AI Humans combine the emotional intelligence of humans with the reach and reliability of machines, making them capable, trusted agents available 24/7, in every language, on our terms.Imagine a therapist anyone can afford. A personal trainer that adapts to your schedule. A fleet of medical assistants that can give every patient the attention they need. With Tavus, individuals, enterprises, and developers can all build AI Humans to connect, understand, and act with empathy at scale.We’re a Series A company backed by world-class investors including Sequoia Capital, Y Combinator, and Scale Venture Partners.Be part of shaping a future where humans and machines truly understand each other.The Role We’re looking for an AI Researcher to join our core AI team and push the boundaries of large language modeling in the context of conversational AI. If you thrive in fast-moving startup environments, enjoy experimenting with new ideas, and love seeing your work come to life in production then you’ll feel right at home.Your Mission 🚀Conduct research on large language modeling and adaptation for Conversational Avatars (e.g. Neural Avatars, Talking-Heads).Develop methods to model both verbal and non-verbal aspects of conversation, adapting and controlling avatar behavior in real time.Experiment with fine-tuning, adaptation, and conditioning techniques to make LLMs more expressive, controllable, and task-specific.Partner with the Applied ML team to take research from prototype to production.Stay up to date with cutting-edge advancements — and help define what comes next.You’ll Be Great At This If You Have:A PhD (or near completion) in a relevant field, or equivalent research experience.Hands-on experience with LLMs or VLMs and a strong foundation in generative language models.Experience in fine-tuning/adapting LLMs for control, conditioning, or downstream tasks.Solid background in deep learning and familiarity with foundation model methods.Strong PyTorch skills and comfort building deep learning pipelines.Nice-to-HavesKnowledge of large-scale model training and optimization.Broader understanding of generative AI across modalities.Exposure to software development best practices.A flexible, experimental mindset i.e. comfortable working across research and engineering.(Bonus) Publications at EMNLP, COLING, NeurIPS, ICLR, CVPR, ICCV.Location Preferred: San Francisco (hybrid) or London (office opening soon). Remote within the U.S. or Europe available for exceptional candidates.
Research Scientist
Product & Operations
Machine Learning Engineer
Data Science & Analytics
Hidden link
The Browser Company.jpg

Engineering Manager, Machine Learning

The Browser Company
USD
360000
300000
-
360000
US.svg
United States
Full-time
Remote
true
Hi, we're The Browser Company 👋 and we're building a better way to use the internet. Browsers are unique in that they are one of the only pieces of software that you share with your parents as well as your kids. Which makes sense, they're our doorway to the most important things — through them we socialize with loved ones, work on our passion projects, and explore our curiosities. But on their own, they don’t actually do a whole lot, they’re kind of just there. They don’t help us organize our messy lives or make it easier to compose our ideas. We believe that the browser could do so much more — it can empower and support the amazing things we do on the internet. That’s why we’re building one: a browser that can help us grow, create, and stay curious. To accomplish this lofty task, we’re building a diverse team of people from different backgrounds and experiences. This isn’t optional, it’s crucial to our mission, as we need a wide range of perspectives to challenge our assumptions and shape our browser through a bold, creative lens. With that in mind, we especially encourage women, people of color, and others from historically marginalized groups to apply.About The RoleAs our founding ML Engineering Manager at the Browser Company, you’ll build our strategy and define how we incorporate ML to advance Dia’s product vision. The browser holds most of the context about your workday, from memory of the tasks you’re trying to accomplish to access to all of your web apps. We use ML and AI to turn that context into high‑utility experiences that feel personal and improve over time, grounded in capabilities like computer‑use, memory, and web‑app integrations.Dia’s Assistant doesn’t just execute actions, it also suggests and recommend tasks, and learns which choices drive real outcomes. We leverage modern training techniques such as SFT, RL, and GEPA to optimize these experiences while protecting privacy. To move fast and learn continuously, we’re building a first‑class ML developer environment: curated data pipelines, tight experiment loops, offline/online evals with telemetry, and one‑click deploys with strong observability.Our Management PhilosophyEngineering Managers at Browser Company are hands-on technical leaders who build high-performing and psychologically-safe teams with a diverse group of individuals. You’ll work closely with your team to make product decisions, prioritize work, ship features, and promote an engineering culture of knowledge sharing and mentorship. Engineering Managers follow the same onboarding pathway as an IC engineer to learn our product data, and build processes, then ramp into leadership. Overall you will...Define and build our ML strategy, sequencing bets that improve Dia’s assistant and measurable user outcomes. You’ll stay ahead of the latest AI advancements and translate them into strategic opportunities for Dia’s roadmap.Prototype, architect, and ship LLM‑powered features; establish techniques to train models that improve over time and personalize experiences, partnering with Design and Product Engineering to balance quality, speed, and scale for real‑world use.Audit and evolve the ML stack and infrastructure - spanning Swift for on-device inference and Python for models and tooling - to support both encoder and decoder model families across client and server. Over time, scale the system using fine-tuned open-source LLMs.Partner on privacy and security by working with Security and Infra on data stewardship, deployment strategies, and responsible scaling. Establish ML Developer Experience by building tooling and workflows for high quality data curation, experimentation (fine-tuning, RL, prompting), evals, and continuous training to improve the models that power Dia.Support and build a talented team of machine learning engineers, helping them grow both technically and with a product mindset through fast iteration cycles. You’ll play a key role in contributing across the engineering org as we scale—owning processes, recruiting, and onboarding—and proactively improve architecture and practices to enhance performance, stability, and maintainability.Technical Projects You’ll Shape With Us…Browser context engine: learn a unified embedding space that fuses app integrations, enterprise connectors, and on-device signals to enable low-latency retrieval, routing, and personalizationPersonalize the command bar: develop ranking, intent understanding, and context-aware suggestions; fine-tune and evaluate models to measurably boost relevance and engagementEvolve our data flywheel by defining how we instrument product signals, design collection/storage pipelines, create labeling/evaluation loops, and continuously retrain to improve feature quality and personalization Qualifications6+ years of experience training, optimizing, and productionizing modern ML models, especially ones that run in a real-world product environment (bonus if you’ve worked closely with transformer models)3+ years mentoring and leading senior engineers with a track record of tech-leading critical work and setting a sustainable execution pace. You're able to execute on critical projects, own large, complex codebases and drive initiatives within your team.You love to keep up with the latest advancements in AI and are excited to apply the latest models and techniques to push the boundaries of what's possible in a browserYou have production experience with Python and have experience fine-tuning open-source LLMs and going beyond simple LoRA fine-tuningYou’re pragmatic, motivated by nebulous problems, and excited to work in a startup environment with quick product validation cycles.We’re primarily focused on hiring in North American time zones and require that folks have 4+ hours of overlap time with team members in Eastern Time Zone.Compensation and BenefitsOur total compensation package for full-time employees includes base salary, equity, and benefits. The annual salary range for this role is $300,000- $360,000 USD. The actual salary offered will vary based on experience level and interview performance.Benefits: We also offer a wide range of perks and benefits designed to support you, your family and to help you engage with your local community. To learn more, visit go.atlassian.com/perksandbenefits.Location: We’re a remote-friendly company and can hire in any country where Atlassian has a legal entity. If you live in New York (or want to visit), you’re welcome to work from our beautiful office in Williamsburg.The Browser Company is a well-funded, ambitious startup of close to 100 people (and growing!) who are passionate about building great products. We are a remote-first, distributed team, with the option to work from office in Brooklyn, New York. We strongly support diversity and encourage people from all backgrounds to apply.  🚙 To read more about what we value as a company, check out Notes on Roadtrips on our blog.
Machine Learning Engineer
Data Science & Analytics
Software Engineer
Software Engineering
Hidden link
Osmo.jpg

Senior Backend Engineer

Osmo
USD
0
160000
-
190000
US.svg
United States
Full-time
Remote
false
Who we are at Osmo:Osmo is a digital olfaction company, on a mission to give computers a sense of smell to improve the health and wellbeing of human life. Why? Our sense of smell both enriches and saves lives, and has a deep and direct connection to our emotions and memory. Olfactory Intelligence has applications across industries including fragrance, manufacturing, security, medicine, and more. We believe in the power of automation and thoughtfully applied AI/ML to solve problems beyond the reach of human intuition alone. Osmo is headquartered in New York, NY, with a new facility in New Jersey, and offices in Somerville, MA.Osmo is seeking a Senior Backend Engineer with deep expertise in distributed systems architecture to join our Software team. The ideal candidate has experience building robust, scalable systems that bring our Olfactory Intelligence to life.Key ResponsibilitiesDesign and scale high-performance backend systems and APIs that power real-world products.Evolve service architectures for scalability, reliability, and maintainability, balancing speed with technical excellence.Lead design reviews and shape architectural decisions that define the long-term health of our systems.Mentor engineers, raise the technical bar, and foster a culture of quality and collaboration.Own services end-to-end, from design and implementation to deployment, observability, and optimization.Drive meaningful impact by connecting strong engineering fundamentals with clear business and product outcomes.Required Qualifications5+ years of experience designing, building, and operating scalable backend systems and distributed architectures.Deep understanding of system design, performance optimization, and fault tolerance.Fluency in Python, with expertise in RESTful API design, frontend integration patterns, and data storage systems (e.g., BigQuery, Snowflake, PostgreSQL).Experience deploying and maintaining systems on cloud platforms such as Google Cloud Platform (GCP) or Amazon Web Services (AWS).Track record of leading complex initiatives and shaping technical direction across teams.Product-minded and pragmatic, balancing long-term architecture with fast, high-quality delivery.Excellent communicator who can bridge technical and non-technical audiences.Ways to stand outExperience building and integrating with third-party Application Programming Interfaces (APIs) and external data sources.Ability to collaborate across backend, data, and frontend domains to deliver cohesive systems.Interest in olfaction and the technical challenges of representing complex real-world phenomena in code.Salary Range: 160K - 190K annuallyBenefits: Medical, Dental, Vision, 401K, and more.If this role inspires you we’d encourage you to apply. We are committed to recruiting, developing, and retaining an incredible team optimized for a diversity of thought, background, and approaches. Base salary is just one component of Osmo’s total rewards package, which is designed to support the well-being, growth, and long-term success of our team members. Our comprehensive package includes health, dental, and vision coverage; a 401(k) retirement savings plan with company match; flexible paid time off and company holidays; and equity or incentive compensation for eligible roles. Actual compensation will vary based on factors such as experience, skills, location, internal equity, and other relevant business considerations. Osmo regularly reviews pay ranges to ensure they remain competitive, equitable, and aligned with current market data.All employment decisions and responsibilities are determined based on current ability and your ability to grow, without regard to race, color, gender identity, sex, sexual orientation, religion, age, marital status, physical, mental, or sensory disability, or any other characteristic protected by applicable law. Recruitment & Staffing Agencies: Osmo does not accept unsolicited resumes from any source other than candidates. The submission of unsolicited resumes by recruitment or staffing agencies to Osmo or its employees is strictly prohibited unless contacted directly by the Osmo Talent Acquisition team. Any resume submitted by an agency in the absence of a signed agreement will automatically become the property of Osmo and Osmo will not owe any referral or other fees with respect thereto.
Software Engineer
Software Engineering
Hidden link
WRITER.jpg

Customer architect

Writer
USD
0
144700
-
183000
US.svg
United States
Full-time
Remote
true
📐 About this role As a Customer architect, your objective will be to partner with your customers to support them in uncovering, and then creating solutions to serve their business needs, including creating custom templates and working through our APIs. Success will mean a meaningful increase in the number of customers actively leveraging these solutions in high-impact ways. You will work alongside your account partners, including a CSM and Implementation manager, as the technical resource on account to support the adoption and realization of value for our largest customers. Your positivity, sense of curiosity, and ability to create champions from early adopters in the AI writing space will help shape our entire culture. If you have a builder mentality, this is the role for you.🦸🏻‍♀️ Your responsibilitiesSupport the Customer success and Account management team, and work with an assigned group of customers to scope use-cases that need to be supported by a customized solution, validate the technical feasibility, and then own the implementation. Scope the customer’s business and technical needs, including quantifying the existing problems, understanding their tools, workflows, data structure, and integration points to deliver incredible contentOwn the creation of custom templates, deployment of our Knowledge Graph capabilities and act as the technical resource for customers to leverage our APIs Communicate limitations and propose solutions, including their scope, resource needs, and projected timelinesDevelop and maintain a deep understanding of the WRITER solutions, the broader AI and LLM landscape, as well as our customer’s ecosystem so you can offer prescriptive solution recommendationsWork closely with Engineering to QA the outputs from the implementation, calibrate, and iterate until we have an optimal output Develop new processes, systems and workflows to reduce time to value for our customersEnsure we have collateral, documentation and processes to support our customers in their technical integration needs Communicate the voice of the customer back to our Product and Engineering teams so we can make, relevant enhancements to our product⭐️ Is this you?5+ years of technical roles in SaaS including 3+ years of experience in solutions architect or technical account management role Experience managing a customer book of businessHave worked with Fortune 500 customers for a high growth, B2B SaaS companyExperience with generative AI and prompt engineering is a huge plusExperience working with APIsNice-to-have: Python chopsAn excellent communicator and collaborator, able to take complex technical concepts and translate to practical business language for non-technical audiencesHighly versatile, can lead a discovery session with multiple stakeholders up to a CIO level, uncovering use cases and business pain, while also being able to have the technical acumen to address complex technical hurdlesBusiness outcome oriented: a laser focus on delivering solutions that create meaningful business resultsTenured problem solver, able to identify the root cause of issues and work collaboratively to resolve An excellent collaborator, with experience working alongside account teams to drive towards custom results in a coordinated and unified manner. 🍩 Benefits & perks (US Full-time employees)Generous PTO, plus company holidaysMedical, dental, and vision coverage for you and your familyPaid parental leave for all parents (12 weeks)Fertility and family planning supportEarly-detection cancer testing through GalleriFlexible spending account and dependent FSA optionsHealth savings account for eligible plans with company contributionAnnual work-life stipends for:Home office setup, cell phone, internetWellness stipend for gym, massage/chiropractor, personal training, etc.Learning and development stipendCompany-wide off-sites and team off-sitesCompetitive compensation, company stock options and 401kWRITER is an equal-opportunity employer and is committed to diversity. We don't make hiring or employment decisions based on race, color, religion, creed, gender, national origin, age, disability, veteran status, marital status, pregnancy, sex, gender expression or identity, sexual orientation, citizenship, or any other basis protected by applicable local, state or federal law. Under the San Francisco Fair Chance Ordinance, we will consider for employment qualified applicants with arrest and conviction records.By submitting your application on the application page, you acknowledge and agree to WRITER's Global Candidate Privacy Notice.
Solutions Architect
Software Engineering
Hidden link
Bland.jpg

AI Deployment Engineer

Bland
USD
175000
120000
-
175000
US.svg
United States
Full-time
Remote
false
AI Deployment Engineer About Bland AIWe’re a series B startup, and have raised $65 million from Emergence Capital, Y Combinator, and the founders of PayPal and Twilio. We have grown to a 80+ person team, and we serve customers like Better.com, by delivering the most friendly, helpful, and human-like AI phone agents in the world.Why This Role ExistsEvery customer is different — the problems they’re solving, the data they have, and the impact they want to create with AI. We’re looking for someone who thrives in that ambiguity, who can take fuzzy goals and transform them into working, production-ready agents that deliver real value. You’ll work directly with customers to understand their workflows, build tailored agents, and help shape how Bland is adopted across their organization.What You’ll DoDesign and deploy AI solutions. Work closely with customers to translate their challenges into functional agents, integrating APIs and data sources to automate real business processes.Prototype fast, iterate faster. Build the first version quickly, get it into production, and refine based on real-world feedback.Collaborate deeply. Partner with customer teams across engineering, product, and operations to ensure the agent performs, scales, and delivers measurable outcomes.Own end-to-end delivery. From discovery call to deployment, you’ll lead the technical build, testing, and iteration — ensuring the experience feels natural, human, and on-brand.Drive adoption and expansion. Share results, train teams, and embed yourself within the customer organization to uncover new opportunities for automation and scale.Be the face of Bland. You are the customer’s champion, their best employee, and you treat them with unreasonable hospitality. You travel on-site, get to know our customers on a human level, and develop real relationships with our champions and other stakeholders, going above and beyond to host training sessions and dinners. Must-Have Qualifications1–5 years of experience in full-stack, AI, or solutions engineering roles where you owned builds from concept to production.Proven experience building and integrating AI or automation features into real-world applications.Hands-on experience integrating LLMs or AI SDKs into web applications.Strong comfort working with REST/JSON, scripting languages (Python or JavaScript), and modern dev tools (Git, NPM/PNPM).Track record of ownership and grit — moments where you built something from scratch, solved hard problems, or exceeded expectations.Excellent communication skills — you can explain complex AI concepts clearly to both technical and non-technical audiences.Thrives in a fast-moving, high-intensity environment — motivated by challenge, curiosity, and the pursuit of great work.Willingness to work in-person frequently (SF HQ and customer sites) to collaborate and accelerate learning. Nice-to-HavesHands-on experience experimenting with or deploying LLM-powered tools or agents.Prior startup or founder experience — you know what it takes to build without a playbook.Curiosity about AI agent design, orchestration, and automation systems.Experience working directly with customers to identify pain points and turn them into shipped solutions.A portfolio of personal or side projects that showcase creativity, technical depth, and persistence.Exceptional new grads with strong ownership and ambition are welcome to apply. Why you'll love working hereYou’ll be joining one of the fastest-moving AI teams in the world — where speed, creativity, and ownership are the default. You’ll ship real agents used by real customers, working directly with teams that are transforming how humans and AI collaborate.This role is perfect for someone who loves building, learning, and pushing boundaries a hands-on candidate who wants to see their work in the wild, solving real problems for real people.Relentlessness is the most important qualityIf you think you’re missing relevant experience but you’re a fast learner who’s excited for a new challenge – and you have the intangibles our team is looking for – please reach out. As long as you’re resourceful and a fast learner (and you can prove it to our team) we would love to meet you.Compensation & PerksSalary: $120k – $175k base + meaningful equity + benefits. Gorgeous office in Jackson Square, San Francisco (rooftop views & great coffee shops nearby).
Machine Learning Engineer
Data Science & Analytics
Software Engineer
Software Engineering
Hidden link
E2B.jpg

Product Engineer - AI

E2B
CZK
0
150000
-
200000
CZ.svg
Czech Republic
Full-time
Remote
false
💻 Languages: Next.js, TypeScript, Python✅ Skills: Building full-stack apps, being up to date with the latest AI frameworks and tools👉 Who we areE2B is a fast growing Series A startup with 7-figure revenue. We've raised over $32M in total since our funding in 2023 and are supported by great investors like Insight Partners. Our customers are companies like Perplexity, Hugging Face, Manus, or Groq. We're building the next hyperscaler for AI agents. 👉 About the roleYour job will be to inspire developers what they can build E2B. Part of that job is creating examples based on what we often see our users are doing and another part of that is leading by example by building experimental projects using E2B.You’ll be tasked with developing both smaller examples in our Cookbook and full-scale projects like Fragments and AI Analyst.This role requires a high amount of creativity and ability to ship projects on your own from 0 to 1.👉 What we’re looking for3+ years of experience in software engineering.Familiar with tools like AI SDK, LangChain, LlamaIndex, Haystack or Mastra.Detail oriented with great taste.Team player with high ownership.Excellent written and spoken communication skills.Thriving in fast-pace field and environment.Excited to work in person from Prague on a DevTool product.Bonus points if you have authored or contributed to open-source projects in the past.
Software Engineer
Software Engineering
Hidden link
Crusoe.jpg

Software Engineer II, SDN Networking

Crusoe
USD
0
131000
-
154000
US.svg
United States
Full-time
Remote
false
Crusoe's mission is to accelerate the abundance of energy and intelligence. We’re crafting the engine that powers a world where people can create ambitiously with AI — without sacrificing scale, speed, or sustainability.Be a part of the AI revolution with sustainable technology at Crusoe. Here, you'll drive meaningful innovation, make a tangible impact, and join a team that’s setting the pace for responsible, transformative cloud infrastructure.About This Role:As a Software Engineer II - Software Defined Networking, you will lead the development and execution of our Software Defined Networking strategy. You will work extensively with the Linux Kernel, utilizing technologies such as XDP/EBPF, DPDK, SmartNIC, and DPU/IPU to drive innovation and performance improvements. What You’ll Be Working On:Develop and execute the roadmap for the Software Defined Networking strategy at Crusoe Cloud.Be highly hands-on, guiding the engineering team through architecture decisions, design processes, design reviews, code reviews, and implementation tasks.Collaborate with the network infrastructure organization to develop industry-leading networking infrastructure.Lead Linux Kernel and driver development, system architecture design, production support, and cross-functional collaboration.What You’ll Bring to the Team:2+ years of related experience building and operating at scale in a production environment.Proven experience in system programming with C, C++ and/or Rust.Extensive knowledge of Linux Systems Internals and computer architecture.Expertise in Network Programming and Packet Processing pipelines.Hands-on experience with kernel bypass technologies such as XDP/EBPF, AF_XDP, and DPDK.In-depth understanding of TCP/IP and network accelerators like Mellanox/Nvidia SmartNIC (ConnectX6/7), DPU Bluefield3, and Intel IPU.Familiarity with SR-IOV, vDPA, and scalable functions.Strong background in kernel or embedded development, particularly with the Linux kernel.Experience with Open vSwitch, Openflow, and Open Virtual Networking.Knowledge of professional software engineering practices and best practices for the full software development life cycle, including coding standards, code reviews, source control management, build processes, testing, and operations.Demonstrated track record of contributions to the open source community (e.g., Open vSwitch/OVS, Open Virtual Networking/OVN, Multus, Cilium). Bonus Points:Advanced degree in Computer Science, Engineering, or a related field.Proven leadership experience in a technical role.Strong analytical and problem-solving skills.Excellent communication and collaboration skills.Experience with cloud networking platforms (AWS, Azure, GCP) and virtualization technologies (VMware, KVM).Benefits:Industry competitive payRestricted Stock Units in a fast growing, well-funded technology companyHealth insurance package options that include HDHP and PPO, vision, and dental for you and your dependentsEmployer contributions to HSA accountsPaid Parental LeavePaid life insurance, short-term and long-term disabilityTeladoc401(k) with a 100% match up to 4% of salaryGenerous paid time off and holiday scheduleCell phone reimbursementTuition reimbursementSubscription to the Calm appMetLife LegalCompany paid commuter benefit; $300 per monthCompensation: Compensation will be paid in the range of $131,000 - $154,000 per year + Bonus. Restricted Stock Units are included in all offers. Compensation will be determined based on the applicant's knowledge, education, experience, and abilities, as well as internal equity and alignment with market data.Crusoe is an Equal Opportunity Employer. Employment decisions are made without regard to race, color, religion, disability, genetic information, pregnancy, citizenship, marital status, sex/gender, sexual preference/ orientation, gender identity, age, veteran status, national origin, or any other status protected by law or regulation.
Software Engineer
Software Engineering
Hidden link
Deepgram.jpg

EMEA Account Executive

Deepgram
EUR
0
0
-
0
GB.svg
United Kingdom
Full-time
Remote
true
Company OverviewDeepgram is the leading voice AI platform for developers building speech-to-text (STT), text-to-speech (TTS) and full speech-to-speech (STS) offerings. 200,000+ developers build with Deepgram’s voice-native foundational models – accessed through APIs or as self-managed software – due to our unmatched accuracy, latency and pricing. Customers include software companies building voice products, co-sell partners working with large enterprises, and enterprises solving internal voice AI use cases. The company ended 2024 cash-flow positive with 400+ enterprise customers, 3.3x annual usage growth across the past 4 years, over 50,000 years of audio processed and over 1 trillion words transcribed. There is no organization in the world that understands voice better than Deepgram.Company Operating RhythmAt Deepgram, we expect an AI-first mindset—AI use and comfort aren’t optional, they’re core to how we operate, innovate, and measure performance.Every team member who works at Deepgram is expected to actively use and experiment with advanced AI tools, and even build your own into your everyday work. We measure how effectively AI is applied to deliver results, and consistent, creative use of the latest AI capabilities is key to success here. Candidates should be comfortable adopting new models and modes quickly, integrating AI into their workflows, and continuously pushing the boundaries of what these technologies can do.Additionally, we move at the pace of AI. Change is rapid, and you can expect your day-to-day work to evolve just as quickly. This may not be the right role if you’re not excited to experiment, adapt, think on your feet, and learn constantly, or if you’re seeking something highly prescriptive with a traditional 9-to-5.Note: this role is based out of the EMEA territory.Regional ExpectationsBased in an EMEA time zone with strong overlap to customer business hours.Excellent written and spoken English. Additional European languages are a strong plus.OpportunityDeepgram is looking for an Account Executive to drive new customers and revenue to our rapidly growing company. You will have the opportunity to sell a patented world-class voice AI platform that has been enabling our customers to achieve things previously not attainable. We are looking for hungry and talented sales professionals to help us take our top-line revenue to new heights!What You'll DoSelf-prospecting to build a strong sales pipeline of new logos, striving to exceed quarterly and annual sales targets, as well as partnering with the Marketing and SDR teams to generate meetingsWork closely with cross-functional teams (especially Sales Ops and Sales Engineers) to ensure sales effectivenessHave a depth of understanding regarding customer needs in the voice AI space, competition, and Deepgram products to be able to articulate Deepgram’s value proposition fully to highly technical buyersEstablish and maintain relationships with key stakeholders, work with CSMs to drive upsell opportunities and manage existing accounts ongoingYou’ll Love This Role If YouValue having leadership that has tenure and a track record of successThrive working with A Players as cross-functional partners across GTMUnderstand the importance of selling a product that is market validated and can deliver on its technical promiseBelieve there should be no cap to upside and want control of account up-sell potentialAre passionate about what Deepgram is building and how we are changing what is possible with speechIt’s Important to Us That You HaveProven experience in a technical full-cycle sales roleExperience building your own pipeline from scratch and closing accountsConsistent overachievement against sales targets in past experiencesStrong written and verbal communication skillsEmpathy and a unique ability to understand customer needsHighly organized with exceptional follow-up skills due to managing multiple opportunities at any given timeAbility to drive a consistent and repeatable sales process with a coachable/adaptable mindset based on other rep success.It Would Be Great If You HadFamiliarity with Challenger or Triangle sales methodologyExperience selling to a technical audience in a similar spaceExperience selling developer tools, infrastructure, or AI/ML technologies.Fluency in German, French, Dutch, or other major regional languages.Backed by prominent investors including Y Combinator, Madrona, Tiger Global, Wing VC and NVIDIA, Deepgram has raised over $85 million in total funding. If you're looking to work on cutting-edge technology and make a significant impact in the AI industry, we'd love to hear from you!Deepgram is an equal opportunity employer. We want all voices and perspectives represented in our workforce. We are a curious bunch focused on collaboration and doing the right thing. We put our customers first, grow together and move quickly. We do not discriminate on the basis of race, religion, color, national origin, gender, sexual orientation, gender identity or expression, age, marital status, veteran status, disability status, pregnancy, parental status, genetic information, political affiliation, or any other status protected by the laws or regulations in the locations where we operate.We are happy to provide accommodations for applicants who need them.
Enterprise Sales
Marketing & Sales
Hidden link
Black Forest Labs.jpg

Member of Technical Staff - 3D

Black Forest Labs
-
GE.svg
Germany
Remote
false
At Black Forest Labs, we’re on a mission to advance the state of the art in generative deep learning for media, building powerful, creative, and open models that push what’s possible.  Born from foundational research, we continuously create advanced infrastructure to transform ideas into images and videos.Our team pioneered Latent Diffusion, Stable Diffusion, and FLUX.1 – milestones in the evolution of generative AI. Today, these foundations power millions of creations worldwide, from individual artists to enterprise applications. We are looking for a 3D Researcher to bring precise camera control to our image and video generation models. Role and Responsibilities Training large-scale Diffusion (transformer) models for camera-controllable image and video generation Developing conditioning mechanisms for 3D camera parameters (poses, trajectories, intrinsics) in diffusion models Rigorously ablating design choices for 3D control and communicating results & decisions with the broader team Reasoning about the speed and quality trade-offs of 3D-aware architectures What we look for Experience training large-scale Diffusion models for image and video data Strong understanding of 3D projective geometry, camera models, and coordinate systems Experience with 3D-aware generative models or neural rendering techniques (NeRFs, 3DGS, etc.) Integrating geometric priors and 3D conditioning into neural networks Strong proficiency in PyTorch, transformer models and other NN architectures Deep understanding of training techniques such as FSDP, low precision training, and model parallelism Nice to have: Experience with multi-view consistency in generative models Understanding of camera calibration, structure-from-motion, or SLAM  
Machine Learning Engineer
Data Science & Analytics
Computer Vision Engineer
Software Engineering
Research Scientist
Product & Operations
Hidden link
Encord.jpg

Associate Software Engineer - (Graduate)

Encord
-
GB.svg
United Kingdom
Full-time
Remote
false
About Encord At Encord, we're building the AI infrastructure of the future. One of the biggest challenges AI companies face today is data quality. The success of any AI application relies heavily on the quality of its training data, yet for most teams, this crucial step is both the most costly and time-consuming. We’re here to change that. As former computer scientists, physicists, and quants, we’ve experienced firsthand how a lack of tools to prepare quality training data impedes progress in building AI. We believe AI is at a stage similar to the early days of computing or the internet—where the potential is clear, but the surrounding tools and processes are still catching up. That's why we started Encord. We are a talented and ambitious team of 100+, working at the cutting edge of computer vision and deep learning. Backed by $30M in Series B funding from top investors like CRV and Y Combinator, we’re one of the fastest-growing companies in our space. We have big plans ahead and are looking to expand our Engineering team to help us in building a better product for our customers. This is a full-time role starting in June or September 2026 depending on graduation dates. Who we are looking for We’re looking for curious, ambitious graduates who are excited to launch their software engineering careers and make an impact from day one. You’ll join us around June/September 2026 as part of a small, highly collaborative team that’s building the next generation of AI infrastructure. At Encord, you’ll work alongside serial founders, ex-computer scientists, physicists, and quants who care deeply about building great products and shipping fast. We’re looking for graduate full-stack engineers, whether you’re more front-end or back-end leaning, who are passionate about solving real customer problems and thrive in autonomous squads of 4–5 people. Encord is growing rapidly, with our product as the key driver of that growth. A year ago, we were a team of 50; now, we’re over 100 and continuing to accelerate. Joining us now means stepping in at a crucial stage of expansion for the company, product, and team. Your contributions here will have a real, visible impact. If you’re finishing university and looking for your first full-time engineering role where you can learn fast, build meaningful products, and grow alongside an exceptional team, Encord could be the perfect place for you. What to expect High performing team: You'll join the small team that has built a product that our customers love and consistently rate as the best in the market. Members of the team are veterans of both Big Tech and successful startups. While the team is more than a sum of its parts, every member on the team is impressive on their own.Impact: Your work will have direct and tangible impact for customers and in turn noticeably affect the trajectory of the company.End-to-end ownership: You will be entrusted with end-to-end ownership of your projects. From product, design and architectural decisions, all the way to deployment, monitoring and measuring impact on users. You will have the opportunity to work across the stack, covering everything from deployment to styling. We expect each other to take initiative, be proactive in problem-solving, and continuously seek improvements.Autonomy and collaboration: You will often work autonomously to take charge of your projects, make decisions, and drive your work forward. But as a lean and high trust team, we remain quick to collaborate and support each other with challenging problems together. You will work alongside amazing team-mates who will help you learn and upskill quickly.Scale and simplicity: You will tackle complex domains and immense scale with simple solutions. You'll leverage your expertise to solve challenging problems in simple ways. About You Eager to learn and grow: You’ve built small projects, coursework assignments, or personal apps, and are excited to apply what you’ve learned to real-world systems. You’re curious about how reliable, high-performance software is built and are keen to learn from experienced engineers.Motivated by impact:You want your work to make a meaningful difference. You’re excited by the idea of contributing to products that real people use and that deliver tangible results.A natural builder: You enjoy creating things from the ground up, whether that’s writing code, designing systems, or experimenting with new technologies. You’re comfortable exploring across the stack and love solving problemsComfortable with ambiguity: You don’t need every detail defined before you start. You enjoy figuring things out, learning as you go, and using feedback to guide your work.Collaborative and supportive: You value teamwork, communication, and helping others succeed. You take feedback well, share ideas openly, and contribute to a positive, high-performing team culture. QualificationsCurrently pursuing or holding a Bachelor’s or Master’s degree in Computer Science, Physics, Engineering, Mathematics, or a related technical field, or have gained equivalent expertise through coding bootcamps or relevant work experience.Proficient in one or more programming languages such as Python, Go, C#, C++, Java, Ruby on Rails, or JavaScript.
Software Engineer
Software Engineering
Hidden link
H Company.jpg

Implementation Strategist

H Company
-
FR.svg
France
Full-time
Remote
false
IMPLEMENTATION STRATEGISTAbout H: H Company is a next-generation AI research and product company pioneering the future of autonomous, agentic AI. Founded to build intelligence that acts, H Company is creating the foundational infrastructure for autonomous AI systems that drive real-world outcomes across industries.About the Team: The Implementation Strategist is the strategic link between customer and our engineering team. Your key stack is understand where the biggest customer problem stand and help find the right Agents to help our customer automate and optimize tasks, and allow them to focus on what really matters. The team is also responsible to shape the future vision of our products by distilling learnings from the field into the H platform.That platform empowers a wide range of professionals and businesses to safely, easily and autonomously deploy agents at scale for a broad range of tasks: from ensuring websites behave as expected to automating repeated tasks.Key Responsibilities:Dig into the data to understand and identify customer’s most important and impactful multi-step task to automateBe the Go To person for the customerPartner with Forward Deployed Engineer to Seamlessly integrate our solutions Coordinate trainings to ensure our solution best usageBuild and deliver demos and present results to strategic audience Requirements:3+ years of selling and implementing complex IT solutions to customersA double curriculum in Business and Engineering school would be ideal Technical Experience working with AI/ML systems would be appreciatedFluency in at least one backend language (e.g., Python, Go, Node.js, Java) in addition to English ;-)Strong problem-solving skills and ability to thrive in ambiguous, customer-facing environmentsExcellent communication skills, both written and verbalBias for action—able to ship quickly, iterate fast, and own the full lifecycle of solutionsWillingness to travel for on-site deployments or customer engagementsLocation:H's teams are distributed throughout France, the UK, and the USThis role is currently in ParisThe final decision for this will lie with the hiring manager for each individual roleWhat We Offer:Be part of the founding journey of one of the most exciting AI startups shaping the future of AI and agentic systems.Collaborate with a fun, dynamic and multicultural team, working alongside world-class AI engineers, researchers, and builders in a high-trust, high-impact environment.Enjoy a competitive compensation package, including salary and equityUnlock opportunities for professional growth, continuous learning, and career development
Implementation Lead
Software Engineering
Hidden link
Mistral AI.jpg

AI Deployment Strategist - Palo Alto

Mistral AI
USD
-
US.svg
United States
Full-time
Remote
false
About Mistral  At Mistral AI, we believe in the power of AI to simplify tasks, save time, and enhance learning and creativity. Our technology is designed to integrate seamlessly into daily working life. We democratize AI through high-performance, optimized, open-source and cutting-edge models, products and solutions. Our comprehensive AI platform is designed to meet enterprise needs, whether on-premises or in cloud environments. Our offerings include le Chat, the AI assistant for life and work. We are a dynamic, collaborative team passionate about AI and its potential to transform society.Our diverse workforce thrives in competitive environments and is committed to driving innovation. Our teams are distributed between France, USA, UK, Germany and Singapore. We are creative, low-ego and team-spirited. Join us to be part of a pioneering company shaping the future of AI. Together, we can make a meaningful impact. See more about our culture on https://mistral.ai/careers. Role Summary : AI Deployment Strategist As an AI Deployment Strategist, you will drive the adoption and deployment of Mistral’s AI solutions, working closely with customers from strategic vision to production implementation. This role sits at the intersection of business strategy, AI innovation, and hands-on deployment, ensuring our customers achieve transformative outcomes. You will partner with senior executives to design AI roadmaps, collaborate with the Applied AI team to deliver solutions in production, and ensure seamless transitions from presales to postsales. Your work will directly contribute to customer success, bridging the gap between strategy and execution. This role is ideal for those who thrive in a fast-paced environment, enjoy solving complex business challenges, and are passionate about turning AI potential into real-world impact. What you will do Strategic Discovery & Vision Setting• Lead executive-level workshops to identify business challenges and opportunities where Mistral’s AI can drive step-change improvements.• Co-create AI adoption roadmaps with customers, articulating the “art of the possible” and a clear path to value.• Collaborate with Account Executives to develop business cases, quantify ROI, and align solutions with customer objectives. AI Solution Design & Deployment• Architect end-to-end AI solutions, integrating Mistral’s models and platform into customer workflows and technical infrastructure.• Partner with the Applied AI team to design, prototype, and deploy AI solutions in production, ensuring scalability and impact.• Own the execution of pilot projects and proofs-of-value, demonstrating the potential of our technology and paving the way for full-scale deployment. Value Realization & Customer Success• Serve as a trusted advisor to customers, guiding their AI strategy and ensuring they maximize the value of their investment in Mistral.• Monitor key performance indicators (KPIs) tied to business outcomes, and communicate progress to executive sponsors.• Proactively identify expansion opportunities within accounts, building on initial successes to drive long-term partnerships. Cross-Functional Collaboration• Act as the bridge between customers and Mistral’s internal teams, synthesizing feedback to influence product and research roadmaps.• Develop reusable assets, best practices, and playbooks to scale go-to-market efforts and ensure consistent delivery excellence.• Travel (~30-60%) to foster deep client relationships and support on-site deployment. About you • 2+ years in a client-facing strategic role (e.g., management consulting, value engineering, or technical sales).• You hold a degree in a relevant scientific field (e.g., Computer Science, Data Science, Engineering, etc.)• Strong business acumen and problem-solving skills, with the ability to structure ambiguous challenges into actionable solutions.• Executive presence and communication skills to influence senior stakeholders (VP, C-level).• Foundational knowledge of AI/ML/Data Science, with the credibility to advise both technical and non-technical audiences.• Hands-on experience prototyping proofs-of-concept (Python, JavaScript, or similar) to demonstrate value.• Resilient, results-driven, and comfortable leading through influence in a collaborative environment.• Experience with sales qualification frameworks (e.g., MEDDPICC) and value-based selling is a plus. Why This Role Matters You will play a pivotal role in shaping how enterprises adopt and deploy AI, ensuring Mistral’s solutions deliver measurable impact. By partnering with the Applied AI team, you’ll help turn strategic visions into production-ready solutions, making Mistral an indispensable partner for our customers. What we offer 💰 Competitive cash salary and equity🚑 Healthcare: Medical/Dental/Vision covered👴🏻 401K: 6% matching🚴 Transportation: Reimburse office parking charges, or $120/month for public transport💡 Coaching: we offer BetterUp coaching on a voluntary basis🥎 Sport: $120/month reimbursement for gym membership🥕 Meal voucher: $400 monthly allowance for meals
Solutions Architect
Software Engineering
Hidden link
Descript.jpg

Head of Platform Engineering

Descript
USD
224000
-
296000
US.svg
United States
Full-time
Remote
false
Our vision at Descript is to build the next-generation platform for fast and easy creation of audio and video content. We are trusted by some of the world's top podcasters and influencers, as well as businesses like BBC, ESPN, HubSpot, Shopify, and The Washington Post for communicating via video. We've raised $100M from leading investors such as the OpenAI Startup fund, Andreessen Horowitz, Redpoint Ventures, and Spark Capital. About The Team We are seeking a Head of Platform Engineering to lead and scale our Platform organization. This organization at Descript is central to empowering our engineering teams to build and deliver products efficiently and effectively. Operating as an internal product team, the Platform group focuses on providing tools, infrastructure, and services that enable our engineers to deliver exceptional products to our users. You will be leading five teams across Platform, with expectations for continued growth in the next 6-12 months. The Platform organization encompasses: Infrastructure Team: Builds and maintains our core backend and development infrastructure, ensuring scalability, reliability, and performance. Builder Experience Team: Focuses on the foundations of our client applications, CI/CD, build, test, and release. AI Enablement Team: Develops and scales our AI/ML infrastructure to support cutting-edge research and integration of AI features into our products. Media Team: Scales and improves our proprietary media server which handles playback and media serving. Core Engineering: Responsible for monetization foundations (usage tracking, Stripe integration), identity services (auth, permissions, user/team management), and key enterprise integrations. Key Challenges Ahead: Infrastructure Scaling: Building a scalable, reliable, and performant infrastructure to support the growth of our engineering, data, and research teams. Product Transition: Leading the transition of our features and infrastructure from a desktop app-heavy product to a web-first and cloud-based user experience. Developer Experience: Establishing the foundations for a world-class developer experience, enhancing productivity and satisfaction across a fast-growing, full-stack engineering organization. AI/ML Integration: Scaling our AI/ML platform to support innovative research rapid productization of AI features, and a product feedback cycle for model improvements. This position reports directly to the VP of Engineering. What You'll Do Strategic Leadership: Develop and execute a strategic vision and roadmap for the Platform organization, aligning with company goals and ensuring cross-team collaboration. Team Development: Recruit, mentor, and grow engineering managers and engineers, fostering a culture of continuous learning and development. Operational Excellence: Ensure execution across the Platform teams is predictable, reliable, and sustainable, implementing best practices in project management and engineering processes. Cross-Functional Collaboration: Work closely with Product, Design, and other Engineering teams to ensure platform initiatives meet the needs of internal stakeholders. Innovation and Improvement: Drive innovation within the Platform organization, continually seeking ways to improve our infrastructure, tools, and processes. Culture Building: Help scale and evolve our company culture as we grow, promoting values of collaboration, inclusivity, and excellence. What You Bring Leadership Experience: Has 5+ years of engineering management experience, including leading multiple teams or an engineering organization, preferably in platform or developer experience domain. Technical Expertise: Demonstrates a strong technical background with experience in cloud platforms (GCP preferably), scalable infrastructure, and AI/ML technologies. Strategic Thinker: Can develop and communicate a clear vision, aligning teams and resources to achieve strategic objectives. Team Builder: Excels at creating collaborative, empowering, and high-performing team environments. Effective Communicator: Communicates clearly and effectively, both in writing and verbally, across technical and non-technical audiences. Adaptable and Resilient: Thrives in fast-paced, rapidly changing environments, with the ability to navigate ambiguity and drive results. Educational Background: Holds a Bachelor's degree in Computer Science, Engineering, or related field, or has equivalent professional experience. Nice to Haves Startup Experience: Experience managing engineering teams at a startup or high-growth company where the rate of change is extremely high. AI/ML Background: Familiarity with AI/ML platforms, tools, and practices. Developer Tools Expertise: Background in building developer tools and enhancing developer experience. The base salary range for this role is $224,000- $296,000/year. Final offer amounts will carefully consider multiple factors, including prior experience, expertise, location, and may vary from the amount above.  About Descript Descript is building a simple, intuitive, fully-powered editing tool for video and audio — an editing tool built for the age of AI. We are a team of 150 and the backing of some of the world's greatest investors (OpenAI, Andreessen Horowitz, Redpoint Ventures, Spark Capital).  Descript is the special company that's in possession of both product market fit and the raw materials (passionate user community, great product, large market) for growth, but is still early enough that each new employee has a measurable influence on the direction of the company. Benefits include a generous healthcare package, 401k matching program, catered lunches, and flexible vacation time. Our headquarters are located in the Mission District of San Francisco, CA. We're hiring for a mix of remote roles and hybrid roles.  For those who are remote, we have a handful of opportunities throughout the year for in person collaboration.  For our hybrid roles, we're flexible, and you're an adult—we don't expect or mandate that you're in the office every day. We do believe there are valuable and serendipitous moments of discovery and collaboration that come from working together in person.  Descript is an equal opportunity workplace—we are dedicated to equal employment opportunities regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability, gender identity, or Veteran status. We believe in actively building a team rich in diverse backgrounds, experiences, and opinions to better allow our employees, products, and community to thrive. 
DevOps Engineer
Data Science & Analytics
Software Engineer
Software Engineering
Hidden link
Cerebras Systems.jpg

Deployment Engineer, AI Inference

Cerebras Systems
0
0
-
0
US.svg
United States
CA.svg
Canada
Full-time
Remote
false
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.   Cerebras' current customers include global corporations across multiple industries, national labs, and top-tier healthcare systems. In January, we announced a multi-year, multi-million-dollar partnership with Mayo Clinic, underscoring our commitment to transforming AI applications across various fields. In August, we launched Cerebras Inference, the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services.About Us Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.    Cerebras' current customers include global corporations across multiple industries, national labs, and top-tier healthcare systems. In January, we announced a multi-year, multi-million-dollar partnership with Mayo Clinic, underscoring our commitment to transforming AI applications across various fields. In 2024, we launched Cerebras Inference, the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services.  About The Role  We are seeking a highly skilled Deployment Engineer to build and operate our cutting-edge inference clusters. These clusters would provide the candidate an opportunity to work with the world's largest computer chip, the Wafer-Scale Engine (WSE), and the systems that harness its unparalleled power.   You will play a critical role in ensuring reliable, efficient, and scalable deployment of AI inference workloads across our global infrastructure. On the operational side, you’ll own the rollout of the new software versions and AI replica updates, along the capacity reallocations across our custom-built, high-capacity datacenters.    Beyond operations, you’ll drive improvements to our telemetry, observability and the fully automated pipeline. This role involves working with advanced allocation strategies to maximize utilization of large-scale computer fleets.   The ideal candidate combines hands-on operation rigor with strong systems engineering skills and thrives on building resilient pipelines that keep pace with cutting-edge AI models.  This role does not require 24/7 hour on-call rotations.     Responsibilities  Deploy AI inference replicas and cluster software across multiple datacenters  Operate across heterogeneous datacenter environments undergoing rapid 10x growth  Maximize capacity allocation and optimize replica placement using constraint-solver algorithms  Operate bare-metal inference infrastructure while supporting transition to K8S-based platform  Develop and extend telemetry, observability and alerting solutions to ensure deployment reliability at scale  Develop and extend a fully automated deployment pipeline to support fast software updates and capacity reallocation at scale  Translate technical and customer needs into actionable requirements for the Dev Infra, Cluster, Platform and Core teams  Stay up to date with the latest advancements in AI compute infrastructure and related technologies.   Skills And Requirements  2-5 years of experience in operating on-prem compute infrastructure (ideally in Machine Learning or High-Performance Compute) or id developing and managing complex AWS plane infrastructure for hybrid deployments  Strong proficiency in Python for automation, orchestration, and deployment tooling  Solid understanding of Linux-based systems and command-line tools  Extensive knowledge of Docker containers and container orchestration platforms like K8S  Familiarity with spine-leaf (Clos) networking architecture  Proficiency with telemetry and observability stacks such as Prometheus, InfluxDB and Grafana  Strong ownership mindset and accountability for complex deployments  Ability to work effectively in a fast-paced environment.   Location   SF Bay Area.  Toronto  Why Join Cerebras People who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection  point in our business. Members of our team tell us there are five main reasons they joined Cerebras: Build a breakthrough AI platform beyond the constraints of the GPU. Publish and open source their cutting-edge AI research. Work on one of the fastest AI supercomputers in the world. Enjoy job stability with startup vitality. Our simple, non-corporate work culture that respects individual beliefs. Read our blog: Five Reasons to Join Cerebras in 2025. Apply today and become part of the forefront of groundbreaking advancements in AI! Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them. This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
DevOps Engineer
Data Science & Analytics
Hidden link
Giga.jpg

Senior Backend Engineer

Giga
USD
350000
200000
-
350000
US.svg
United States
Full-time
Remote
false
Location: San Francisco (100% In-Office) Experience: 5+ years Salary: $200K - $350k • Offers EquityAt Giga, we’re revolutionizing enterprise by deploying AI agents that resolve over 1 million customer tickets monthly via voice and chat. Industry leaders like DoorDash trust our AI to navigate complex, real-world conversations with clarity, accuracy, and scalability.Engineers at Giga work on problems like:building agents with almost no hallucination ratesa voice experience that's better than talking to humanscreating self-learning agents that optimize metrics Who You Are / Must-Haves:Exceptional EngineerBackend: Strong experience in Python (Django experience is a plus)Cloud: Proficient with AWS or Google CloudScalability: Familiarity with Kubernetes and Docker is a plusEntrepreneurial: Preference for candidates with experience at successful startups. Former founders are especially valuedSeasoned: 5+ years of hands-on engineering experience with roles in leadership Perks & BenefitsCompetitive total compensationEquinox membershipFull health, dental, and vision coverageOn-demand snacks, coffee, and DoorDash CreditsUbers covered home Interview ProcessScreening call w/ recruiting lead (20 min)Intro call w/ Founding Engineer (15 min)Live coding round w/ Founding Engineer (1 hr)In-office interview (full day)
Software Engineer
Software Engineering
Machine Learning Engineer
Data Science & Analytics
Hidden link
Cerebras Systems.jpg

AI Inference Support Engineer

Cerebras Systems
-
No items found.
Remote
false
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.   Cerebras' current customers include global corporations across multiple industries, national labs, and top-tier healthcare systems. In January, we announced a multi-year, multi-million-dollar partnership with Mayo Clinic, underscoring our commitment to transforming AI applications across various fields. In August, we launched Cerebras Inference, the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services.About The Role Join Cerebras’ new Global Support organization to help customers run production-grade AI inference. You’ll troubleshoot issues across model serving, deployment, and observability; resolve customer tickets; and partner with Engineering to improve the reliability, performance, and usability of our inference platform.  Responsibilities Own inbound tickets for inference issues (availability, latency/throughput, correctness, model loading, etc.).  Triage, reproduce, and debug across the stack: APIs/SDKs, model serving layers (e.g., vLLM), networking, etc.  Analyze logs/metrics/traces (e.g., Prometheus/Grafana/ELK) to drive fast resolution and clear RCAs.  Create and maintain high-quality runbooks, knowledge base articles, and “getting unstuck” guides.  Collaborate with Product/Eng to escalate defects, validate fixes, and influence roadmap via aggregated support insights.  Participate in follow-the-sun on-call rotations for P1/P2 incidents with defined SLAs.  Proactively identify pain points in both our solutions and those of our customers.  Advocate for customer needs internally, helping prioritize fixes, features, and reliability improvements.  Skills & Qualifications 4–6 years in technical support, SRE, or solutions engineering for distributed systems or ML/AI products.  Strong Linux fundamentals; confident with shell, systemd, containers (Docker), basic networking (TLS, DNS, HTTP/2, gRPC), and debugging with logs/metrics.  Proficiency in at least one scripting language (Python preferred) for repros, tooling, and log parsing.  Familiarity with modern LLM inference concepts: token streaming, batching, KV cache, etc.  Excellent customer communication: drive clarity from ambiguous reports, write crisp updates, and set accurate expectations.  Assets Exposure to one or more serving stacks (e.g. vLLM) and OpenAI-compatible APIs.  Observability practice (Prometheus, Grafana, Elk) and basic performance testing.  Ticketing/ITSM (e.g., Jira/ServiceNow/Zendesk), incident response, and SLA/SLO workflows.  Experience with GPUs/accelerators and performance tuning (throughput vs. latency trade-offs, batching/concurrency tuning).  Demonstrate humility, collaboration, and a commitment to continuous learning to support team and customer success.  Why Join Cerebras People who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection  point in our business. Members of our team tell us there are five main reasons they joined Cerebras: Build a breakthrough AI platform beyond the constraints of the GPU. Publish and open source their cutting-edge AI research. Work on one of the fastest AI supercomputers in the world. Enjoy job stability with startup vitality. Our simple, non-corporate work culture that respects individual beliefs. Read our blog: Five Reasons to Join Cerebras in 2025. Apply today and become part of the forefront of groundbreaking advancements in AI! Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them. This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
DevOps Engineer
Data Science & Analytics
Machine Learning Engineer
Data Science & Analytics
Software Engineer
Software Engineering
Hidden link
Cerebras Systems.jpg

AI Inference Support Engineer

Cerebras Systems
-
earth.svg
Europe
AE.svg
United Arab Emirates
Remote
false
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.   Cerebras' current customers include global corporations across multiple industries, national labs, and top-tier healthcare systems. In January, we announced a multi-year, multi-million-dollar partnership with Mayo Clinic, underscoring our commitment to transforming AI applications across various fields. In August, we launched Cerebras Inference, the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services.About The Role Join Cerebras’ new Global Support organization to help customers run production-grade AI inference. You’ll troubleshoot issues across model serving, deployment, and observability; resolve customer tickets; and partner with Engineering to improve the reliability, performance, and usability of our inference platform.  Responsibilities Own inbound tickets for inference issues (availability, latency/throughput, correctness, model loading, etc.).  Triage, reproduce, and debug across the stack: APIs/SDKs, model serving layers (e.g., vLLM), networking, etc.  Analyze logs/metrics/traces (e.g., Prometheus/Grafana/ELK) to drive fast resolution and clear RCAs.  Create and maintain high-quality runbooks, knowledge base articles, and “getting unstuck” guides.  Collaborate with Product/Eng to escalate defects, validate fixes, and influence roadmap via aggregated support insights.  Participate in follow-the-sun on-call rotations for P1/P2 incidents with defined SLAs.  Proactively identify pain points in both our solutions and those of our customers.  Advocate for customer needs internally, helping prioritize fixes, features, and reliability improvements.  Skills & Qualifications 4–6 years in technical support, SRE, or solutions engineering for distributed systems or ML/AI products.  Strong Linux fundamentals; confident with shell, systemd, containers (Docker), basic networking (TLS, DNS, HTTP/2, gRPC), and debugging with logs/metrics.  Proficiency in at least one scripting language (Python preferred) for repros, tooling, and log parsing.  Familiarity with modern LLM inference concepts: token streaming, batching, KV cache, etc.  Excellent customer communication: drive clarity from ambiguous reports, write crisp updates, and set accurate expectations.  Assets Exposure to one or more serving stacks (e.g. vLLM) and OpenAI-compatible APIs.  Observability practice (Prometheus, Grafana, Elk) and basic performance testing.  Ticketing/ITSM (e.g., Jira/ServiceNow/Zendesk), incident response, and SLA/SLO workflows.  Experience with GPUs/accelerators and performance tuning (throughput vs. latency trade-offs, batching/concurrency tuning).  Demonstrate humility, collaboration, and a commitment to continuous learning to support team and customer success.  Why Join Cerebras People who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection  point in our business. Members of our team tell us there are five main reasons they joined Cerebras: Build a breakthrough AI platform beyond the constraints of the GPU. Publish and open source their cutting-edge AI research. Work on one of the fastest AI supercomputers in the world. Enjoy job stability with startup vitality. Our simple, non-corporate work culture that respects individual beliefs. Read our blog: Five Reasons to Join Cerebras in 2025. Apply today and become part of the forefront of groundbreaking advancements in AI! Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them. This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
DevOps Engineer
Data Science & Analytics
Machine Learning Engineer
Data Science & Analytics
Solutions Architect
Software Engineering
Hidden link
No job found
Your search did not match any job. Please try again
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.