Google DeepMind · Mountain View

Research Scientist/Engineer, Model Threat Defense

11/13/2025

Description

  • Research Defense Strategies: Research techniques to detect distillation and techniques to actively defend against distillation.
  • Deploy Detection & Mitigation Systems: Design and build systems that detect abd mitigate unauthorized capability extraction.
  • Evaluate Impact: Rigorously measure the effectiveness of defense mechanisms, balancing the trade-offs between model robustness, defensive utility, and core model performance.
  • Collaborate and Publish: Work closely with world-class researchers across GDM, Google, and the industry to publish groundbreaking work, establish new benchmarks, and set the standard for responsible AI defense.

Qualifications

  • Ph.D. in Computer Science or a related quantitative field, or a B.S./M.S. in a similar field with 2+ years of relevant industry experience.
  • Demonstrated research or product expertise in a field related to model security, adversarial ML, post-training, or model evaluation.
  • Experience designing and implementing large-scale ML systems or counter-abuse infrastructure.
  • Deep expertise in one or more of the following areas: model distillation, model stealing, security, memorization, Reinforcement Learning, Supervised Fine-Tuning, or Embeddings.
  • Proven experience in Adversarial Machine Learning, with a focus on designing and implementing model defenses.
  • Strong software engineering skills and experience with ML frameworks like JAX, PyTorch, or TensorFlow.
  • A track record of landing research impact or shipping production systems in a multi-team environment.
  • Current or prior US security clearance.

Benefits

The US base salary range for this full-time position is between $166,000 - $244,000 + bonus + equity + benefits. Your recruiter can share more about the specific salary range for your targeted location during the hiring process.

Application

View listing at origin and apply!