This position is no longer available
Hybrid

Anthropic was looking for an Anthropic AI Safety Fellow

Anthropic is looking for an Anthropic AI Safety Fellow to conduct hands-on, empirical research aimed at reducing catastrophic risks from advanced AI. This four-month fellowship provides direct mentorship from Anthropic researchers, the opportunity to work on a project aligned with our safety priorities, and the goal of producing a public output like a paper submission.

What You'll Do

  • Work on an empirical project aligned with Anthropic's AI safety research priorities.
  • Use external infrastructure, such as open-source models and public APIs, to conduct research.
  • Produce a public output, such as a paper submission.

What We're Looking For

  • Motivation to reduce catastrophic risks from advanced AI systems.
  • Excitement to transition into full-time empirical AI safety research and interest in a full-time role at Anthropic.
  • A strong technical background in computer science, mathematics, physics, cybersecurity, or related fields.
  • Ability to thrive in fast-paced, collaborative environments.
  • Capacity to implement ideas quickly and communicate clearly.
  • Fluency in Python programming.
  • Availability to work full-time on the Fellows program for 4 months.
  • Work authorization in the US, UK, or Canada and location in that country during the program.

Nice to Have

  • Experience with empirical ML research projects.
  • Experience working with Large Language Models.
  • Experience in one of our core research areas, such as Scalable Oversight or Adversarial Robustness.
  • Experience with deep learning frameworks and experiment management.
  • A track record of open-source contributions.

Technical Stack

  • Python
  • Open-source models
  • Public APIs
  • Deep learning frameworks

Team & Environment

You will receive direct mentorship from Anthropic researchers. Our culture is extremely collaborative, with frequent research discussions, and we work as a single cohesive team on large-scale research efforts.

Benefits & Compensation

  • Weekly stipend of 3,850 USD / 2,310 GBP / 4,300 CAN.
  • Access to country-specific benefits.
  • Funding for compute (~$15k/month) and other research expenses.
  • Access to a shared workspace in either Berkeley, California or London, UK.
  • Connection to the broader AI safety research community.

Work Mode

This is a hybrid role. Fellows can be based in London, UK, Ontario, CA, San Francisco, CA, or work remotely within the United States.

Required Skills
PythonDeep Learning FrameworksOpen-source ModelsPublic APIsAI SafetyMachine LearningResearchTechnical CommunicationCollaborationProblem Solving PythonDeep Learning FrameworksOpen-source ModelsPublic APIsAI SafetyMachine LearningResearchTechnical CommunicationCollaborationProblem Solving
Your first international client?

Don't lose them over invoicing

Clients ghost freelancers with unprofessional invoicing. Glopay gives you a real EU company partnership so they take you seriously from invoice #1.

Instant EU company partnership
Invoice builder with your branding
Automated payment reminders
Real-time payment tracking
Get EU company now
Ready in 24 hours
About company
Anthropic
Anthropic’s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole.
All jobs at Anthropic Visit website
Job Details
Category data
Posted 5 months ago