
Building safe and reliable AI systems for everyone
Anthropic, headquartered in SoMa, San Francisco, is an AI safety and research company focused on developing reliable, interpretable, and steerable AI systems. With over 1,000 employees and backed by Google, Anthropic has raised $29.3 billion in funding, including a monumental Series F round of $13 b...
Anthropic offers comprehensive health, dental, and vision insurance for employees and their dependents, along with inclusive fertility benefits via Ca...
Anthropic's culture is rooted in AI safety and reliability, with a focus on producing less harmful outputs compared to existing AI systems. The compan...

Anthropic • San Francisco, CA | New York City, NY
Anthropic is seeking a Biological Safety Research Scientist to design and develop safety systems for AI. You'll collaborate with experts to ensure responsible AI safety in the biological domain. This role requires a strong background in biological sciences.
You are a biological scientist with a strong understanding of safety mechanisms in AI systems. You have experience in designing and executing capability evaluations to assess the performance of AI models. Your background allows you to translate complex biosecurity concepts into practical technical safeguards. You thrive in collaborative environments, working closely with threat modeling experts and machine learning engineers to develop robust safety systems. You are committed to ensuring that AI technologies are safe and beneficial for society. You understand the balance between advancing legitimate life sciences research and preventing misuse by sophisticated threat actors.
Experience in AI safety or biosecurity is a plus. Familiarity with machine learning concepts and practices will enhance your ability to contribute effectively to the team. You are open to learning and adapting to new challenges in the rapidly evolving field of AI safety.
In this role, you will design and execute capability evaluations to assess the capabilities of new AI models. You will collaborate closely with internal and external experts to develop training data for safety systems, ensuring they are optimized for robustness against adversarial attacks while maintaining low false-positive rates for legitimate researchers. You will analyze the performance of safety systems and contribute to the development of oversight mechanisms that align with Anthropic's mission of creating reliable and interpretable AI systems. Your work will directly impact how frontier AI models handle dual-use biological knowledge, shaping the future of responsible AI safety.
At Anthropic, we provide competitive compensation and benefits, including optional equity donation matching, generous vacation and parental leave, and flexible working hours. You will work in a collaborative office space in San Francisco, surrounded by a team of dedicated researchers and engineers. We encourage you to apply even if your experience doesn't match every requirement, as we value diverse perspectives and backgrounds in our mission to build beneficial AI systems.
Apply now or save it for later. Get alerts for similar jobs at Anthropic.