Research Defense Strategies: Research techniques to detect distillation and techniques to actively defend against distillation.
Deploy Detection & Mitigation Systems: Design and build systems that detect abd mitigate unauthorized capability extraction.
Evaluate Impact: Rigorously measure the effectiveness of defense mechanisms, balancing the trade-offs between model robustness, defensive utility, and core model performance.
Collaborate and Publish: Work closely with world-class researchers across GDM, Google, and the industry to publish groundbreaking work, establish new benchmarks, and set the standard for responsible AI defense.
Qualifications
Ph.D. in Computer Science or a related quantitative field, or a B.S./M.S. in a similar field with 2+ years of relevant industry experience.
Demonstrated research or product expertise in a field related to model security, adversarial ML, post-training, or model evaluation.
Experience designing and implementing large-scale ML systems or counter-abuse infrastructure.
Deep expertise in one or more of the following areas: model distillation, model stealing, security, memorization, Reinforcement Learning, Supervised Fine-Tuning, or Embeddings.
Proven experience in Adversarial Machine Learning, with a focus on designing and implementing model defenses.
Strong software engineering skills and experience with ML frameworks like JAX, PyTorch, or TensorFlow.
A track record of landing research impact or shipping production systems in a multi-team environment.
Current or prior US security clearance.
Benefits
The US base salary range for this full-time position is between $166,000 - $244,000 + bonus + equity + benefits. Your recruiter can share more about the specific salary range for your targeted location during the hiring process.