Description
About the Role
The Anthropic Fellows Program is a 4-month full-time research opportunity that provides funding and mentorship to promising technical talent. As a Reinforcement Learning Fellow, you will work on an empirical project aligned with our research priorities, with the goal of producing a public output (e.g., a paper submission).
What to Expect
- 4 months of full-time research
- Direct mentorship from Anthropic researchers
- Access to a shared workspace (in either Berkeley, California or London, UK)
- Connection to the broader AI safety and security research community
- Weekly stipend of $3,850 USD / £2,310 GBP / $4,300 CAD + benefits (these vary by country)
- Funding for compute (~$15k/month) and other research expenses
Responsibilities
As a Reinforcement Learning Fellow, you will be responsible for:
- Building model-based tools to better understand AI training data and improve training data quality
- Conducting research and implementing solutions in areas such as RL algorithms
- Collaborating with other researchers and engineers to advance the state-of-the-art in reinforcement learning
Qualifications
- Fluency in Python programming
- Strong background in a discipline relevant to reinforcement learning (e.g., computer science, mathematics, or physics)
- Experience in areas of research or engineering related to reinforcement learning
Logistics
To participate in the Fellows program, you must have work authorization in the US, UK, or Canada and be located in that country during the program.
How to Apply
Applications and interviews are managed by Constellation, our official recruiting partner for this program. Clicking "Apply here" will redirect you to Constellation's application portal.
This listing is enriched and indexed by YubHub. To apply, use the employer's original posting:
https://job-boards.greenhouse.io/anthropic/jobs/5183052008