
Building safe and reliable AI systems for everyone
Anthropic, headquartered in SoMa, San Francisco, is an AI safety and research company focused on developing reliable, interpretable, and steerable AI systems. With over 1,000 employees and backed by Google, Anthropic has raised $29.3 billion in funding, including a monumental Series F round of $13 b...
Anthropic offers comprehensive health, dental, and vision insurance for employees and their dependents, along with inclusive fertility benefits via Ca...
Anthropic's culture is rooted in AI safety and reliability, with a focus on producing less harmful outputs compared to existing AI systems. The compan...

Anthropic • Remote-Friendly (Travel-Required) | San Francisco, CA | Washington, DC
Anthropic is seeking a Policy Manager specializing in Chemical Weapons and High Yield Explosives to design evaluation methodologies and develop strategies for AI safety. This role requires a Ph.D. in Chemistry or related fields and 5-8 years of relevant experience.
You hold a Ph.D. in Chemistry, Chemical Engineering, or a related field with a focus on energetic materials, explosives, and/or chemical weapons. With 5-8+ years of experience in chemical weapons and/or explosives defense, you possess deep expertise in the field. You are adept at translating complex domain knowledge into actionable safety requirements and have a strong understanding of the implications of AI in sensitive areas such as chemical weapons and explosives. Your experience includes developing and executing strategies to identify and mitigate potential misuse of AI models in this domain. You are a collaborative team player, capable of working with software engineers to refine detection systems and automated enforcement tools for threats related to chemical weapons and explosives.
Experience in policy development related to chemical weapons and explosives is a plus. Familiarity with AI technologies and their applications in safety and security contexts will enhance your candidacy. You are proactive in monitoring emerging threats in the chemical and explosive landscape, ensuring that your strategies remain relevant and effective.
In this role, you will design and implement evaluation methodologies for assessing AI model capabilities relevant to chemical weapons and explosives synthesis. You will develop and execute strategies to identify and mitigate potential misuse in model outputs, creating threat models that include precursor identification, synthesis routes, and weaponization techniques. Your responsibilities will also involve reviewing and analyzing traffic to identify potential policy violations related to chemical and explosive content. You will collaborate with software engineers to develop and refine detection systems and automated enforcement tools for chemical and explosive threats. Rapid response to escalations involving dangerous queries will be a critical part of your role, as will collaborating across teams to establish safety benchmarks and develop appropriate model guardrails. You will translate your domain knowledge into actionable safety requirements and develop approaches to assess model knowledge boundaries for dual-use chemical information.
Anthropic offers a competitive compensation package and benefits, including optional equity donation matching, generous vacation, and parental leave. You will enjoy flexible working hours and a collaborative office environment in San Francisco. As a public benefit corporation, we are committed to creating reliable, interpretable, and steerable AI systems that are safe and beneficial for society. Join us in our mission to build beneficial AI systems while working with a team of dedicated researchers, engineers, policy experts, and business leaders.
Apply now or save it for later. Get alerts for similar jobs at Anthropic.