Anthropic
Overview
Research Engineer / Scientist, Tool Use Safety role at Anthropic. The Tool Use Team focuses on making Claude the world’s most capable, safe, reliable, and efficient model for tool use and agentic applications. This includes solving core problems such as tool use safety, tool call accuracy, long horizon tool use workflows, large-scale and dynamic tools, and tool use efficiency. The role involves advancing safe tool use and scaling research to enable Claude to assist users more autonomously while maintaining safety and reliability. Responsibilities
Design and implement novel and scalable reinforcement learning methodologies that push the state of the art of tool use safety Define and pursue research agendas that push the boundaries of what\'s possible Build rigorous, realistic evaluations that capture the complexity of real-world tool use safety challenges Ship research advances that directly impact and protect millions of users Collaborate with safety research (e.g. Safeguards, Alignment Science), capabilities research, and product teams to drive fundamental breakthroughs in safety, and work with teams to ship these into production Design, implement, and debug code across our research and production ML stacks Contribute to our collaborative research culture through pair programming, technical discussions, and team problem-solving You may be a good fit if you
Passionate about our safety mission Are driven by real-world impact and excited to see research ship in production Have strong machine learning research/applied-research experience, or a strong quantitative background such as physics, mathematics, or quantitative finance research Write clean, reliable code and have solid software engineering skills Communicate complex ideas clearly to diverse audiences Are hungry to learn and grow, regardless of years of experience Strong candidates may also have one or more of the following
Experience with tool use/agentic safety, trust & safety, or security Experience with reinforcement learning techniques and environments Experience with language model training, fine-tuning or evaluation Experience building AI agents or autonomous systems Published influential work in relevant ML areas, especially around LLM safety & alignment Deep expertise in a specialized area (e.g., RL, security, or mathematical foundations), even if still developing breadth in adjacent areas Experience shipping features or working closely with product teams Enthusiasm for pair programming and collaborative research Compensation
The annual salary range for this position is $315,000 - $425,000 USD. Logistics
Education requirements: We require at least a Bachelor's degree in a related field or equivalent experience. Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. Some roles may require more time in our offices. Visa sponsorship: We do sponsor visas. If we make you an offer, we will make every reasonable effort to obtain a visa, with support from an immigration lawyer. We encourage you to apply even if you do not meet every single qualification. We consider diverse perspectives important and welcome applicants from underrepresented groups. About Anthropic
Anthropic’s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for users and society. Our team includes researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems. The Tool Use Team focuses on tool use safety, tool call accuracy, long-horizon tool use workflows, large-scale tools, and tool use efficiency to support Claude and related applications.
#J-18808-Ljbffr
Research Engineer / Scientist, Tool Use Safety role at Anthropic. The Tool Use Team focuses on making Claude the world’s most capable, safe, reliable, and efficient model for tool use and agentic applications. This includes solving core problems such as tool use safety, tool call accuracy, long horizon tool use workflows, large-scale and dynamic tools, and tool use efficiency. The role involves advancing safe tool use and scaling research to enable Claude to assist users more autonomously while maintaining safety and reliability. Responsibilities
Design and implement novel and scalable reinforcement learning methodologies that push the state of the art of tool use safety Define and pursue research agendas that push the boundaries of what\'s possible Build rigorous, realistic evaluations that capture the complexity of real-world tool use safety challenges Ship research advances that directly impact and protect millions of users Collaborate with safety research (e.g. Safeguards, Alignment Science), capabilities research, and product teams to drive fundamental breakthroughs in safety, and work with teams to ship these into production Design, implement, and debug code across our research and production ML stacks Contribute to our collaborative research culture through pair programming, technical discussions, and team problem-solving You may be a good fit if you
Passionate about our safety mission Are driven by real-world impact and excited to see research ship in production Have strong machine learning research/applied-research experience, or a strong quantitative background such as physics, mathematics, or quantitative finance research Write clean, reliable code and have solid software engineering skills Communicate complex ideas clearly to diverse audiences Are hungry to learn and grow, regardless of years of experience Strong candidates may also have one or more of the following
Experience with tool use/agentic safety, trust & safety, or security Experience with reinforcement learning techniques and environments Experience with language model training, fine-tuning or evaluation Experience building AI agents or autonomous systems Published influential work in relevant ML areas, especially around LLM safety & alignment Deep expertise in a specialized area (e.g., RL, security, or mathematical foundations), even if still developing breadth in adjacent areas Experience shipping features or working closely with product teams Enthusiasm for pair programming and collaborative research Compensation
The annual salary range for this position is $315,000 - $425,000 USD. Logistics
Education requirements: We require at least a Bachelor's degree in a related field or equivalent experience. Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. Some roles may require more time in our offices. Visa sponsorship: We do sponsor visas. If we make you an offer, we will make every reasonable effort to obtain a visa, with support from an immigration lawyer. We encourage you to apply even if you do not meet every single qualification. We consider diverse perspectives important and welcome applicants from underrepresented groups. About Anthropic
Anthropic’s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for users and society. Our team includes researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems. The Tool Use Team focuses on tool use safety, tool call accuracy, long-horizon tool use workflows, large-scale tools, and tool use efficiency to support Claude and related applications.
#J-18808-Ljbffr