# Research Engineer, Machine Learning

**Company**: Mistral AI
**Location**: Palo Alto
**Work arrangement**: hybrid
**Experience**: senior
**Job type**: full-time
**Category**: Engineering
**Industry**: Technology
**Wikidata**: https://www.wikidata.org/wiki/Q119718658

**Apply**: https://jobs.lever.co/mistral/bada0014-0f32-4370-b55f-81c5595c7339
**Canonical**: https://yubhub.co/jobs/job_797494bd-994

## Description

About Mistral AI
==================
Mistral AI is a pioneering company that develops and provides high-performance, open-source AI models, products, and solutions.

Role Summary
=============
The Research Engineering team at Mistral AI spans Platform (shared infrastructure and clean code) and Embedded (inside research squads). Engineers can move along the research↔production spectrum as needs or interests evolve.

As a Research Engineer – ML track, you’ll build and optimize the large-scale learning systems that power our open-weight models. Working hand-in-hand with Research Scientists, you’ll either join:

* Platform RE Team: Enhance the shared training framework, data pipelines, and cluster tooling used by every team;
* Embedded RE Team: Sit inside a research squad (Alignment, Pre-training, Multimodal, …) and turn fresh ideas into repeatable, scalable code.

Responsibilities
==============
* Accelerate researchers by taking on the heavy parts of large-scale ML pipelines and building robust tools.
* Interface cutting-edge research with production: integrate checkpoints, streamline evaluation, and expose APIs.
* Conduct experiments on the latest deep-learning techniques (sparsified 70 B + runs, distributed training on thousands of GPUs).
* Design, implement, and benchmark ML algorithms; write clear, efficient code in Python.
* Deliver prototypes that become production-grade components for Le Chat and our enterprise API.

Requirements
==========
* Master’s or PhD in Computer Science (or equivalent proven track record).
* 4 + years working on large-scale ML codebases.
* Hands-on with PyTorch, JAX, or TensorFlow; comfortable with distributed training (DeepSpeed / FSDP / SLURM / K8s).
* Experience in deep learning, NLP, or LLMs; bonus for CUDA or data-pipeline chops.
* Strong software-design instincts: testing, code review, CI/CD.
* Self-starter, low-ego, collaborative.

What We Offer
============
* Competitive salary and equity.
* Healthcare: Medical/Dental/Vision covered for you and your family.
* Pension: 401K (6% matching).
* PTO: 18 days.
* Transportation: Reimburse office parking charges, or $120/month for public transport.
* Sport: $120/month reimbursement for gym membership.
* Meal stipend: $400 monthly allowance for meals (solution might evolve as we grow bigger).
* Visa sponsorship.
* Coaching: we offer BetterUp coaching on a voluntary basis.

## Skills

### Required
- PyTorch
- JAX
- TensorFlow
- Distributed Training
- Deep Learning
- NLP
- LLMs
- CUDA
- Data Pipelines
