Location
Heidelberg, Germany
Heidelberg, Germany
Salary
(Yearly)
(Yearly)
(Yearly)
(Yearly)
(Hourly)
Undisclosed
Date posted
February 26, 2026
Job type
Full-time
Experience level
Senior
Summary this job with AI
Highlight
Highlight

Job Description

Our Mission

Aleph Alpha is one of the few companies in Europe doing serious foundation model pre-training. Our customers - in finance, manufacturing, public administration - need models that understand German, meet European regulatory requirements, and work reliably in high-stakes settings. We're building that in Heidelberg.

We are hiring a Performance Engineer to grow our pre-training efficiency team. If you are excited about making models fast, this is the role for you!

Team Culture

At Aleph Alpha, we foster a culture built on ownership, autonomy, and empowerment. Teams and individual contributors are trusted to take responsibility for their work and drive meaningful impact. We maintain a flat organizational structure with efficient, supportive management that enables quick decision‑making, open communication, and a strong sense of shared purpose.

About the role:

You will engineer the systems required to train foundation models at scale. Your objective is to maximize hardware utilization and training throughput on our large-scale GPU clusters (thousands of NVIDIA Blackwell GPUs). You will work at the intersection of deep learning frameworks, distributed systems, and GPU microarchitecture, eliminating bottlenecks from the Python layer down to the GPU kernel.

This role is for Aleph Alpha Research.

Your responsibilities:

  • End-to-End Optimization: Profile training loops using PyTorch Profiler, Nsight Systems and Nsight Compute to identify system- and kernel-level bottlenecks in order to maximize model throughput.

  • Distributed Strategy and Topology: Configure and tune composite parallelism strategies (e.g. TP, DP, HSDP/FSDP, EP), optimizing load balance, minimizing critical-path bottlenecks, and managing communication-to-computation trade-offs for large-scale LLM training.

  • Hardware-Aware Modeling: Partner with AI Researchers to define model architectures for hardware efficiency without compromising convergence.

You could be a great fit if you:

  • Are proficient in Python and the PyTorch library.

  • Have a strong engineering background in parallel and/or distributed systems with proven track record of excellence.

  • Have hands-on experience with modern machine learning techniques (especially large language models and their life cycle).

  • Deeply understand the CUDA programming model.

  • Have experience in distributed programming with APIs like NCCL or MPI.

  • Have experience analysing profiling traces with tools such as PyTorch Profiler and Nvidia Nsight.

  • Please note this role requires regular on-site collaboration in Heidelberg as a member of the Training Efficiency Team.

Strong candidates may also have:

  • Contributions to modern distributed training frameworks (e.g., TorchTitan, Megatron-LM, DeepSpeed).

  • Familiarity with low-precision training formats (MXFP4, MXFP8) and their impact on numerical stability and throughput.

  • A deep understanding of NCCL communication primitives, NVSHMEM or CUDA IPC and their performance.

  • A proven track record of implementing and optimising modern transformer-based model training.

  • A proven track record working on the NVIDIA Blackwell architecture.

Compensation and Benefits

  • Competitive salary and equity package

  • 30 days of paid vacation

  • Access to a variety of fitness & wellness offerings via Wellhub

  • Mental health support through nilo.health

  • JobRad® Bike Lease

  • Substantially subsidized company pension plan for your future security

  • Subsidized Germany-wide transportation ticket

  • Budget for additional technical equipment

  • Flexible working hours for better work-life balance and hybrid working model

Apply now
AlephAlpha is hiring a Senior Performance Engineer- Pretraining. Apply through The Homebase and and make the next move in your career!
Apply now
Companies size
201-500
employees
Founded in
2019
Headquaters
Heidelberg, Germany
Country
Germany
Industry
Computer Software
Social media
Visit website

Similar AI jobs

Here are other jobs you might want to apply for.

US.svg
United States

Staff Software Engineer, Model LifeCycle

Full-time
Machine Learning Engineer
GE.svg
Germany

Senior Performance Engineer- Pretraining

Full-time
Machine Learning Engineer
US.svg
United States

System Software Engineer

Full-time
Machine Learning Engineer
AR.svg
Argentina

AceUp - Lead ML Engineer (Generative AI & LLM Focus)

Full-time
Machine Learning Engineer
US.svg
United States

Member of Technical Staff - ML Training Systems

Full-time
Machine Learning Engineer
No items found.

Staff Engineer, API Core Platform

Full-time
Machine Learning Engineer
Open Modal