Description
Apple Intelligence is powered by thoughtful data sampling, creation, and curation; high quality, detailed annotations; and application of these data to evaluate and mitigate safety concerns of new generative AI features. This role heavily draws on applied data science, scientific investigation and interpretation, cross-functional communication and collaboration, and metrics reporting and presentation to stakeholders & decision-makers. Responsibilities include:
- Develop metrics for evaluation of safety and fairness risks inherent to generative models and Gen-AI features
- Design datasets, identify data needs, and work on creative solutions, scaling and expanding data coverage through human and synthetic generation methods
- Develop sampling strategies and combine human annotation with auto-grading to deliver high-quality, high-confidence insights at a fast pace and large scale
- Use and implement data pipelines, and collaborate cross-functionally to execute end-to-end safety evaluations
- Distill project findings into recommendations for product engineering teams and safety policy development
- Develop ML-based enhancements to red teaming, model evaluation, and other processes to improve the quality of Apple Intelligence’s user-facing products
- Work with highly-sensitive content with exposure to offensive and controversial content