Resume and JobRESUME AND JOB
Apple logo

AIML - ML Engineer, Safety Human Evaluation

Apple

Software and Technology Jobs

AIML - ML Engineer, Safety Human Evaluation

full-timePosted: Oct 1, 2025

Job Description

Join Us in Shaping the Future of Generative AI at Apple! Would you like to play a part in building the next generation of generative AI applications at Apple? We are looking for Machine Learning Engineers to work on ambitious projects that will impact the future of Apple, our products, and the broader world. This role is directed at assessing, quantifying, and improving the safety and inclusivity of Apple’s Generative-AI powered features and products. In this role you’ll have the opportunity to tackle innovative problems in machine learning, particularly focused on large language models for text generation, diffusion models for image generation, and mixed model systems for multimodal applications. As a member of the Apple HCMI/Responsible AI group, you will be working on Apple's generative models that will power a wide array of new features, as well as longer term research in the generative AI space. Our team is currently interested in large generative models for vision and language, with particular interest on Responsible AI, safety, fairness, robustness, explainability, and uncertainty in models. Apple Intelligence is powered by thoughtful data sampling, creation, and curation; high quality, detailed annotations; and application of these data to evaluate and mitigate safety concerns of new generative AI features. This role heavily draws on applied data science, scientific investigation and interpretation, cross-functional communication and collaboration, and metrics reporting and presentation to stakeholders & decision-makers. Responsibilities include: - Develop metrics for evaluation of safety and fairness risks inherent to generative models and Gen-AI features - Design datasets, identify data needs, and work on creative solutions, scaling and expanding data coverage through human and synthetic generation methods - Develop sampling strategies and combine human annotation with auto-grading to deliver high-quality, high-confidence insights at a fast pace and large scale - Use and implement data pipelines, and collaborate cross-functionally to execute end-to-end safety evaluations - Distill project findings into recommendations for product engineering teams and safety policy development - Develop ML-based enhancements to red teaming, model evaluation, and other processes to improve the quality of Apple Intelligence’s user-facing products - Work with highly-sensitive content with exposure to offensive and controversial content

Locations

  • Cambridge, Massachusetts, United States 02139
  • Cupertino, California, United States 95014
  • Seattle, Washington, United States 98117

Salary

Estimated Salary Rangemedium confidence

25,000,000 - 60,000,000 INR / yearly

Source: ai estimated

* This is an estimated range based on market data and may vary based on experience and qualifications.

Skills Required

  • Machine Learningintermediate
  • Generative AIintermediate
  • Large Language Modelsintermediate
  • Diffusion Modelsintermediate
  • Multimodal Applicationsintermediate
  • Responsible AIintermediate
  • Safety in AIintermediate
  • Fairness in AIintermediate
  • Robustness in AIintermediate
  • Explainability in AIintermediate
  • Uncertainty in Modelsintermediate
  • Data Samplingintermediate
  • Data Curationintermediate
  • Data Annotationintermediate
  • Applied Data Scienceintermediate
  • Scientific Investigationintermediate
  • Cross-functional Communicationintermediate
  • Collaborationintermediate
  • Metrics Reportingintermediate
  • Metrics Presentationintermediate
  • Dataset Designintermediate
  • Data Generationintermediate
  • Sampling Strategiesintermediate
  • Human Annotationintermediate
  • Auto-gradingintermediate
  • Data Pipelinesintermediate
  • Safety Evaluationsintermediate
  • Red Teamingintermediate
  • Model Evaluationintermediate

Required Qualifications

  • MS or PhD in Computer Science, Linguistics, Cognitive Science, HCI, Psychology, Mathematics, Physics, or a similar science or technology field with a strong basis in scientific data collection and analysis + at least 4 years of relevant work experience, or BA/BS with 8+ years of relevant work experience (experience, 4 years)
  • Experience gathering and analyzing language data, image data, and/or multi-modal data, including LLM-generated data (experience)
  • Strong experience designing human annotation projects, writing guidelines, and dealing with highly multi-labeled, nuanced, and often conflicting data (experience)
  • Proficiency in data science, machine learning, analytics, and programming with Python & Pandas; strong experience with one or more plotting & visualization libraries (experience)
  • Ability to collaborate with team members to prioritize competing projects, set and maintain a schedule for milestones and project completions, and communicate with all levels of team members as well as external stakeholders (experience)
  • Strong skills for rigorous model quality metrics development; interpretation of experiments and evaluations; and presentation to executives (experience)

Preferred Qualifications

  • Experience working in the Responsible AI space (experience)
  • Curiosity about fairness and bias in generative AI systems, and a strong desire to help make the technology more equitable (experience)
  • Prior scientific research and publication experience (experience)
  • Experience working with generative models for evaluation and/or product development, and up-to-date knowledge of common challenges and failures (experience)
  • Proven track record of contributing to diverse teams in a collaborative environment (experience)
  • A passion for building outstanding and innovative products. This position involves a wide variety of interdisciplinary skills (experience)

Responsibilities

  • Apple Intelligence is powered by thoughtful data sampling, creation, and curation; high quality, detailed annotations; and application of these data to evaluate and mitigate safety concerns of new generative AI features. This role heavily draws on applied data science, scientific investigation and interpretation, cross-functional communication and collaboration, and metrics reporting and presentation to stakeholders & decision-makers. Responsibilities include:
  • - Develop metrics for evaluation of safety and fairness risks inherent to generative models and Gen-AI features
  • - Design datasets, identify data needs, and work on creative solutions, scaling and expanding data coverage through human and synthetic generation methods
  • - Develop sampling strategies and combine human annotation with auto-grading to deliver high-quality, high-confidence insights at a fast pace and large scale
  • - Use and implement data pipelines, and collaborate cross-functionally to execute end-to-end safety evaluations
  • - Distill project findings into recommendations for product engineering teams and safety policy development
  • - Develop ML-based enhancements to red teaming, model evaluation, and other processes to improve the quality of Apple Intelligence’s user-facing products
  • - Work with highly-sensitive content with exposure to offensive and controversial content

Target Your Resume for "AIML - ML Engineer, Safety Human Evaluation" , Apple

Get personalized recommendations to optimize your resume specifically for AIML - ML Engineer, Safety Human Evaluation. Takes only 15 seconds!

AI-powered keyword optimization
Skills matching & gap analysis
Experience alignment suggestions

Check Your ATS Score for "AIML - ML Engineer, Safety Human Evaluation" , Apple

Find out how well your resume matches this job's requirements. Get comprehensive analysis including ATS compatibility, keyword matching, skill gaps, and personalized recommendations.

ATS compatibility check
Keyword optimization analysis
Skill matching & gap identification
Format & readability score

Tags & Categories

Hardware

Answer 10 quick questions to check your fit for AIML - ML Engineer, Safety Human Evaluation @ Apple.

Quiz Challenge
10 Questions
~2 Minutes
Instant Score

Related Books and Jobs

No related jobs found at the moment.

Apple logo

AIML - ML Engineer, Safety Human Evaluation

Apple

Software and Technology Jobs

AIML - ML Engineer, Safety Human Evaluation

full-timePosted: Oct 1, 2025

Job Description

Join Us in Shaping the Future of Generative AI at Apple! Would you like to play a part in building the next generation of generative AI applications at Apple? We are looking for Machine Learning Engineers to work on ambitious projects that will impact the future of Apple, our products, and the broader world. This role is directed at assessing, quantifying, and improving the safety and inclusivity of Apple’s Generative-AI powered features and products. In this role you’ll have the opportunity to tackle innovative problems in machine learning, particularly focused on large language models for text generation, diffusion models for image generation, and mixed model systems for multimodal applications. As a member of the Apple HCMI/Responsible AI group, you will be working on Apple's generative models that will power a wide array of new features, as well as longer term research in the generative AI space. Our team is currently interested in large generative models for vision and language, with particular interest on Responsible AI, safety, fairness, robustness, explainability, and uncertainty in models. Apple Intelligence is powered by thoughtful data sampling, creation, and curation; high quality, detailed annotations; and application of these data to evaluate and mitigate safety concerns of new generative AI features. This role heavily draws on applied data science, scientific investigation and interpretation, cross-functional communication and collaboration, and metrics reporting and presentation to stakeholders & decision-makers. Responsibilities include: - Develop metrics for evaluation of safety and fairness risks inherent to generative models and Gen-AI features - Design datasets, identify data needs, and work on creative solutions, scaling and expanding data coverage through human and synthetic generation methods - Develop sampling strategies and combine human annotation with auto-grading to deliver high-quality, high-confidence insights at a fast pace and large scale - Use and implement data pipelines, and collaborate cross-functionally to execute end-to-end safety evaluations - Distill project findings into recommendations for product engineering teams and safety policy development - Develop ML-based enhancements to red teaming, model evaluation, and other processes to improve the quality of Apple Intelligence’s user-facing products - Work with highly-sensitive content with exposure to offensive and controversial content

Locations

  • Cambridge, Massachusetts, United States 02139
  • Cupertino, California, United States 95014
  • Seattle, Washington, United States 98117

Salary

Estimated Salary Rangemedium confidence

25,000,000 - 60,000,000 INR / yearly

Source: ai estimated

* This is an estimated range based on market data and may vary based on experience and qualifications.

Skills Required

  • Machine Learningintermediate
  • Generative AIintermediate
  • Large Language Modelsintermediate
  • Diffusion Modelsintermediate
  • Multimodal Applicationsintermediate
  • Responsible AIintermediate
  • Safety in AIintermediate
  • Fairness in AIintermediate
  • Robustness in AIintermediate
  • Explainability in AIintermediate
  • Uncertainty in Modelsintermediate
  • Data Samplingintermediate
  • Data Curationintermediate
  • Data Annotationintermediate
  • Applied Data Scienceintermediate
  • Scientific Investigationintermediate
  • Cross-functional Communicationintermediate
  • Collaborationintermediate
  • Metrics Reportingintermediate
  • Metrics Presentationintermediate
  • Dataset Designintermediate
  • Data Generationintermediate
  • Sampling Strategiesintermediate
  • Human Annotationintermediate
  • Auto-gradingintermediate
  • Data Pipelinesintermediate
  • Safety Evaluationsintermediate
  • Red Teamingintermediate
  • Model Evaluationintermediate

Required Qualifications

  • MS or PhD in Computer Science, Linguistics, Cognitive Science, HCI, Psychology, Mathematics, Physics, or a similar science or technology field with a strong basis in scientific data collection and analysis + at least 4 years of relevant work experience, or BA/BS with 8+ years of relevant work experience (experience, 4 years)
  • Experience gathering and analyzing language data, image data, and/or multi-modal data, including LLM-generated data (experience)
  • Strong experience designing human annotation projects, writing guidelines, and dealing with highly multi-labeled, nuanced, and often conflicting data (experience)
  • Proficiency in data science, machine learning, analytics, and programming with Python & Pandas; strong experience with one or more plotting & visualization libraries (experience)
  • Ability to collaborate with team members to prioritize competing projects, set and maintain a schedule for milestones and project completions, and communicate with all levels of team members as well as external stakeholders (experience)
  • Strong skills for rigorous model quality metrics development; interpretation of experiments and evaluations; and presentation to executives (experience)

Preferred Qualifications

  • Experience working in the Responsible AI space (experience)
  • Curiosity about fairness and bias in generative AI systems, and a strong desire to help make the technology more equitable (experience)
  • Prior scientific research and publication experience (experience)
  • Experience working with generative models for evaluation and/or product development, and up-to-date knowledge of common challenges and failures (experience)
  • Proven track record of contributing to diverse teams in a collaborative environment (experience)
  • A passion for building outstanding and innovative products. This position involves a wide variety of interdisciplinary skills (experience)

Responsibilities

  • Apple Intelligence is powered by thoughtful data sampling, creation, and curation; high quality, detailed annotations; and application of these data to evaluate and mitigate safety concerns of new generative AI features. This role heavily draws on applied data science, scientific investigation and interpretation, cross-functional communication and collaboration, and metrics reporting and presentation to stakeholders & decision-makers. Responsibilities include:
  • - Develop metrics for evaluation of safety and fairness risks inherent to generative models and Gen-AI features
  • - Design datasets, identify data needs, and work on creative solutions, scaling and expanding data coverage through human and synthetic generation methods
  • - Develop sampling strategies and combine human annotation with auto-grading to deliver high-quality, high-confidence insights at a fast pace and large scale
  • - Use and implement data pipelines, and collaborate cross-functionally to execute end-to-end safety evaluations
  • - Distill project findings into recommendations for product engineering teams and safety policy development
  • - Develop ML-based enhancements to red teaming, model evaluation, and other processes to improve the quality of Apple Intelligence’s user-facing products
  • - Work with highly-sensitive content with exposure to offensive and controversial content

Target Your Resume for "AIML - ML Engineer, Safety Human Evaluation" , Apple

Get personalized recommendations to optimize your resume specifically for AIML - ML Engineer, Safety Human Evaluation. Takes only 15 seconds!

AI-powered keyword optimization
Skills matching & gap analysis
Experience alignment suggestions

Check Your ATS Score for "AIML - ML Engineer, Safety Human Evaluation" , Apple

Find out how well your resume matches this job's requirements. Get comprehensive analysis including ATS compatibility, keyword matching, skill gaps, and personalized recommendations.

ATS compatibility check
Keyword optimization analysis
Skill matching & gap identification
Format & readability score

Tags & Categories

Hardware

Answer 10 quick questions to check your fit for AIML - ML Engineer, Safety Human Evaluation @ Apple.

Quiz Challenge
10 Questions
~2 Minutes
Instant Score

Related Books and Jobs

No related jobs found at the moment.