Description
Our organization supports a diverse array of programs passionate about evaluating ML algorithms and assessing model quality at scale, across domains like computer vision, audio, and multi-modal systems. You will collaborate with multi-functional teams, including domain experts and engineering leads, and adapt methodologies as new insights emerge.
In this role you will:
- Evaluate ML & MM-LLM Models: Analyze and validate computer vision, multi-modal, and large language models to ensure they meet accuracy, robustness, and usability standards.
- Develop Metrics: Design and implement metrics to measure the efficiency and accuracy of models.
- Failure Analysis: Conduct in-depth analysis on model failures across CV and MM-LLM pipelines to surface root causes and improvement areas.
- Data Processing: Clean, transform, and curate large-scale datasets for model evaluation and benchmarking.
- Model Optimization: Apply innovative techniques to optimize models for scalability and real-world deployment.
- Collaborate multi-functionally: Work closely with cross-functional teams, including software engineers, product managers, and other data scientists, to integrate models into production.
- Communicate Results: Present findings clearly and effectively to collaborators across levels of technical understanding.