AI Safety Researcher

Anthropic·San Francisco, CA

AI SafetySeniorFull-timeRemote
$200K-$350KPosted 4 days ago

About the Role

Anthropic is looking for an AI Safety Researcher to advance our understanding of AI alignment and develop practical techniques for building safe and beneficial AI systems. You will conduct original research on constitutional AI, interpretability, and scalable oversight.

Requirements

  • PhD in Machine Learning, AI Safety, or related field
  • Track record of publications in AI safety or alignment
  • Strong mathematical foundations in statistics and optimization
  • Proficiency in Python and modern ML frameworks
  • Demonstrated ability to conduct independent research

Nice to Have

  • Experience with interpretability or mechanistic interpretability
  • Background in philosophy of mind or ethics
  • Experience with red-teaming or adversarial testing of LLMs
  • Familiarity with Constitutional AI or RLHF methods

Benefits

Significant equity in a high-growth AI company
Comprehensive health benefits
Flexible work arrangements
$10,000 annual conference and education budget
Relocation assistance
Sabbatical program

Skills

AI SafetyAlignmentInterpretabilityPythonPyTorchResearch

Related Jobs

Preparing for Your AI Career?

Vincony has all 400+ AI models in one place — compare responses, AI debate, Image/Video/Voice generator, and 20 more tools to help you learn and build with AI.

Visit Vincony.com