Job Details
Job Information
Other Information
Job Description
Role Number: 200649572-0836
Summary
Would you like to play a part in building the next generation of generative AI applications at Apple? We’re looking for Machine Learning Engineers to work on ambitious projects that will impact the future of Apple, our products, and the broader world. This role is directed at assessing, quantifying, and improving the safety and inclusivity of Apple’s Generative-AI powered features and products. In this role you’ll have the opportunity to tackle innovative problems in machine learning, particularly focused on large language models for text generation, diffusion models for image generation, and mixed model systems for multimodal applications. As a member of Apple’s Responsible AI group you will be working on a wide array of new features and research in the generative AI space. Our team is currently interested in large generative models for vision and language, with particular interest on Responsible AI, safety, fairness, robustness, explainability, and uncertainty in models.
Description
This highly multifunctional role sits at the intersection of machine learning research and real-world product impact, focusing on the responsible development and safety evaluation of Apple Intelligence features. You will collaborate closely with top ML researchers, engineers, and product teams to define and deliver responsible AI technologies — from training and evaluating foundation models with safety in mind, to advancing alignment, robustness, and deployment safeguards for LLMs across Apple products.
A core part of this role involves developing, executing, and communicating pre- and post-ship safety evaluations, leveraging both human grading and model-based auto-grading methodologies. You will lead thoughtful data sampling, curation, and annotation to build high-quality evaluation datasets, and develop auto-grading infrastructure to support ongoing and future safety assessments. Your work will require strong applied data science skills, rigorous scientific interpretation, and clear cross-functional communication to ensure evaluation findings meaningfully inform the user experience and uphold Apple's Responsible AI values.
Minimum Qualifications
MS or PhD in Computer Science, Machine Learning, Statistics, or related field (or equivalent); publications in top ML venues (e.g., ACL, FAccT, ICML, NeurIPS, etc.) preferred
4+ years of research or product deployment experience in responsible AI, with strong fundamentals in LLMs, foundation models, and diffusion models
Proficient in Python and deep learning frameworks (e.g., JAX, PyTorch, TensorFlow), with experience writing production-quality code and leveraging foundation model-based AI programming tools (e.g., DSPy)
Experience working with generative models for evaluation and/or product development, including familiarity with common challenges, failure modes, and noisy crowd-based human evaluation data
Strong organizational skills with the ability to work effectively across large, multi-functional teams; comfort working with sensitive, offensive, or controversial content
Preferred Qualifications
Experience working in the Responsible AI space.
Prior scientific research and publication experience.
Strong organizational and operational skills working with large, multi-functional, and diverse teams.
Curiosity about fairness and bias in generative AI systems, and a strong desire to help make the technology more equitable.
Apple is an equal opportunity employer that is committed to inclusion and diversity. We seek to promote equal opportunity for all applicants without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, Veteran status, or other legally protected characteristics. Learn more about your EEO rights as an applicant (https://www.eeoc.gov/sites/default/files/2023-06/22-088_EEOC_KnowYourRights6.12ScreenRdr.pdf) .
Other Details

