Research Engineer, Frontier Safety Loss of Control, DeepMind
- linkCopy link
- emailEmail a friend
Minimum qualifications:
- Bachelor's degree in Computer Science, Machine Learning, or a related technical field, or equivalent practical experience.
- 5 years of experience in engineering and agentic assistance, including software development in Python.
- Experience working in a frontier AI research and development environment.
- Experience working in a professional software engineering or research team environment.
- Experience working with technical stakeholders.
- Experience in frontier model risk.
Preferred qualifications:
- Experience of engineering or product design for AI tools or assistants, especially those focused on ML Research and Development (R&D).
- Experience with cybersecurity detection and response.
- Experience with collaborating or leading an applied ML project.
- Experience with Large Language Model (LLM) training and inference.
- Knowledge of AI control, chain-of-thought and other monitoring, faithfulness and monitorability and related research areas.
About the job
Our team develops monitoring and control for potentially misaligned AI to mitigate risks of extreme harms. Currently, this primarily involves: designing, building, and testing monitors for potentially dangerous behaviours; developing and implementing response policies to preserve AI usefulness while mitigating risks; and foreseeing ways in which our control tools might be bypassed or degraded. We are looking for an engineer who can rapidly iterate to solve never-before-seen problems with creativity and thoroughness.The Loss of Control team contributes to a defense in depth against the risk of misaligned AI systems being deployed. We take the possibility of very advanced AI seriously. We don’t think control is a suitable alternative to alignment in the limit of advancing intelligence. But while AI remains effectively monitorable, we think that control is an important part of an overall strategy for building safe AI.
We are looking for a research engineer for the Frontier Safety Loss of Control team within the AGI Safety and Alignment Team based in either San Francisco or London.
In this role, the core responsibility is to help Google prepare for the internal use of potentially misaligned AI systems. That means building defense-in-depth against AI that might persistently pursue goals that users and system developers did not intend.
Artificial intelligence will be one of humanity’s most transformative inventions. At Google DeepMind, we are a pioneering AI lab with exceptional interdisciplinary teams focused on advancing AI development to solve complex global challenges and accelerate high-quality product innovation for billions of users. We use our technologies for widespread public benefit and scientific discovery, ensuring safety and ethics are always our highest priority.
Responsibilities
- Identify ways that misaligned agents could cause harm and strategies for detecting and preventing harm.
- Implement technical controls to monitor agent thoughts, behaviour, and respond to mitigate potential harms.
- Integrate various agent behaviour signals from across the organisation to inform response policies.
- Conduct adversarial testing of controls.
- Work with internal product teams to ensure that control systems are adopted over all high-risk AI surfaces.
Information collected and processed as part of your Google Careers profile, and any job applications you choose to submit is subject to Google's Applicant and Candidate Privacy Policy.
Google is proud to be an equal opportunity and affirmative action employer. We are committed to building a workforce that is representative of the users we serve, creating a culture of belonging, and providing an equal employment opportunity regardless of race, creed, color, religion, gender, sexual orientation, gender identity/expression, national origin, disability, age, genetic information, veteran status, marital status, pregnancy or related condition (including breastfeeding), expecting or parents-to-be, criminal histories consistent with legal requirements, or any other basis protected by law. See also Google's EEO Policy, Know your rights: workplace discrimination is illegal, Belonging at Google, and How we hire.
If you have a need that requires accommodation, please let us know by completing our Accommodations for Applicants form.
Google is a global company and, in order to facilitate efficient collaboration and communication globally, English proficiency is a requirement for all roles unless stated otherwise in the job posting.
To all recruitment agencies: Google does not accept agency resumes. Please do not forward resumes to our jobs alias, Google employees, or any other organization location. Google is not responsible for any fees related to unsolicited resumes.