Resume and JobRESUME AND JOB
OpenAI logo

Researcher, Robustness & Safety Training Careers at OpenAI - San Francisco, California | Apply Now!

OpenAI

Researcher, Robustness & Safety Training Careers at OpenAI - San Francisco, California | Apply Now!

full-timePosted: Feb 10, 2026

Job Description

Researcher, Robustness & Safety Training at OpenAI - San Francisco Careers

Join OpenAI's Safety Systems team as a Researcher, Robustness & Safety Training in San Francisco, California. This senior-level role is at the forefront of AI safety research, focusing on RLHF, adversarial training, and model robustness to ensure safe AGI deployment. Apply now for this high-impact position driving OpenAI's mission.

Role Overview

The Researcher, Robustness & Safety Training position at OpenAI in San Francisco is a pivotal role within the Model Safety Research team. This team is dedicated to advancing AI safety capabilities, ensuring that OpenAI's most powerful models can be deployed safely to benefit society. As AI systems grow more capable, new challenges emerge in enforcing safety policies, robustness against adversaries, privacy protection, and trustworthiness in critical domains.

In this role, you'll conduct cutting-edge research on topics like Reinforcement Learning from Human Feedback (RLHF), adversarial training, and robustness. You'll implement these innovations directly into OpenAI's core training pipelines and launch safety features in products used by millions. This isn't just research—it's about shaping the future of safe AGI, collaborating across teams to meet the highest safety standards, and making a tangible impact on humanity.

OpenAI's commitment to safety is unwavering. The Safety Systems team leads efforts to deploy models responsibly, learning from real-world use while mitigating risks. If you have 4+ years in AI safety, a PhD in ML or related fields, and a passion for aligned AGI, this San Francisco-based role offers unparalleled opportunities to influence global AI safety standards.

Key focus areas include balancing safety with helpfulness, defending against malicious actors, securing user privacy, and building trust in high-stakes applications like healthcare or autonomous systems. Your work will directly contribute to OpenAI's charter for universally beneficial AI.

Key Responsibilities

As a Researcher in Robustness & Safety Training at OpenAI, your responsibilities will span research, implementation, and strategy:

  • Lead state-of-the-art research in AI safety, specializing in RLHF, adversarial training, and model robustness techniques.
  • Develop and implement novel methods within OpenAI’s core model training infrastructure for enhanced safety.
  • Deploy safety improvements into production products, ensuring real-world reliability.
  • Define research directions to make AI systems safer, more aligned, and resilient to adversarial threats.
  • Collaborate with Trust & Safety, legal, policy, and engineering teams to uphold rigorous safety protocols.
  • Perform comprehensive evaluations of model safety, pinpointing risks and crafting mitigation strategies.
  • Engineer defenses against adversarial attacks, maintaining performance in hostile environments.
  • Tackle privacy risks through advanced techniques like differential privacy in large models.
  • Enhance fairness and reduce biases across diverse deployment scenarios.
  • Build trustworthy AI for safety-critical domains, validating behaviors under uncertainty.
  • Analyze deployment data to iteratively improve safety systems.
  • Publish and present findings, advancing the broader AI safety field.
  • Mentor junior researchers and contribute to team strategy.

These tasks position you at the intersection of research and deployment, ensuring OpenAI's AI benefits society safely.

Qualifications

To excel as a Researcher, Robustness & Safety Training in OpenAI's San Francisco office, you should possess:

  • 4+ years in AI safety research, with expertise in RLHF, adversarial training, robustness, fairness, and biases.
  • Ph.D. or equivalent in computer science, machine learning, statistics, or related disciplines.
  • Hands-on experience with safety protocols for deploying large-scale AI models.
  • Deep knowledge of deep learning architectures, optimization, and scaling laws.
  • Strong programming skills in Python, PyTorch/TensorFlow, and ML frameworks.
  • Demonstrated passion for OpenAI’s mission and alignment with its safety charter.
  • Proven ability to collaborate in cross-functional, fast-paced environments.
  • Track record of publishing in top ML conferences (NeurIPS, ICML, ICLR).
  • Experience evaluating model vulnerabilities and proposing scalable fixes.
  • Understanding of real-world AI risks like jailbreaks, data poisoning, and misuse.
  • Excitement for tackling open problems in AGI safety.

Ideal candidates thrive in collaborative settings and are driven by the challenge of safe AI deployment.

Salary & Benefits

OpenAI offers competitive compensation for this senior Researcher role in San Francisco, estimated at $320,000 - $500,000 USD yearly, including base salary, equity, and bonuses. Total compensation reflects experience and impact.

Benefits include:

  • Comprehensive medical, dental, vision coverage.
  • 401(k) with generous matching.
  • Unlimited vacation and flexible hours.
  • Parental leave up to 20 weeks.
  • Professional growth stipend ($10K+ annually).
  • Onsite fitness center, catered meals, wellness programs.
  • Relocation support for San Francisco move.
  • Equity in OpenAI, a leader in AGI.
  • Mental health resources and employee assistance.
  • Commuter benefits and subsidized housing options.
  • Access to cutting-edge hardware and compute clusters.
  • Opportunities for global conferences and sabbaticals.

This package supports top talent focused on AI safety innovation.

Why Join OpenAI?

OpenAI is the pioneer in safe AGI development, with products like ChatGPT transforming the world. Joining the Safety Systems team in San Francisco means working with world-class researchers on problems that matter: making powerful AI safe for everyone.

Our culture emphasizes transparency, collaboration, and impact. You'll shape safety standards for future AI, publish influential work, and deploy changes affecting billions. San Francisco's vibrant tech ecosystem complements OpenAI's innovative environment.

With a mission to benefit humanity, OpenAI invests heavily in safety. This role offers intellectual challenge, massive scale, and the chance to define AGI safety. Past team members have advanced RLHF globally—your contributions could do the same.

OpenAI provides resources unmatched elsewhere: unlimited compute, expert peers, and direct mission alignment. If you're passionate about robust AI, this is your opportunity to lead.

How to Apply

Ready to advance AI safety at OpenAI? Submit your resume, cover letter highlighting AI safety experience, and links to publications or GitHub. Include why you're excited about OpenAI's mission.

Applications are reviewed on a rolling basis. Top candidates advance to research interviews, safety deep-dives, and team fits. We prioritize diverse perspectives aligned with our charter.

Apply now via OpenAI's careers page for the Researcher, Robustness & Safety Training role in San Francisco. Shape the future of safe AI today!

This page optimized for searches like 'OpenAI AI safety researcher jobs San Francisco', 'RLHF researcher careers', 'AGI safety training positions'. Updated 2024.

Locations

  • San Francisco, California, United States

Salary

Estimated Salary Rangehigh confidence

336,000 - 550,000 USD / yearly

Source: ai estimated

* This is an estimated range based on market data and may vary based on experience and qualifications.

Skills Required

  • RLHF (Reinforcement Learning from Human Feedback)intermediate
  • Adversarial Trainingintermediate
  • AI Robustnessintermediate
  • Model Safety Researchintermediate
  • Deep Learningintermediate
  • Machine Learningintermediate
  • AI Alignmentintermediate
  • Fairness and Bias Mitigationintermediate
  • Privacy in AIintermediate
  • Security Risk Assessmentintermediate
  • PyTorchintermediate
  • TensorFlowintermediate
  • Large Language Modelsintermediate
  • Cross-Functional Collaborationintermediate
  • Research Strategyintermediate
  • Model Evaluationintermediate
  • Safety Policy Enforcementintermediate
  • Adversarial Robustnessintermediate

Required Qualifications

  • 4+ years of experience in AI safety research, particularly in RLHF, adversarial training, robustness, fairness, and biases (experience)
  • Ph.D. or equivalent degree in computer science, machine learning, or a related field (experience)
  • Experience in safety work for AI model deployment (experience)
  • In-depth understanding of deep learning research (experience)
  • Strong engineering skills in implementing ML models (experience)
  • Passion for AI safety and OpenAI's mission to build safe AGI (experience)
  • Alignment with OpenAI's charter for universally beneficial AI (experience)
  • Proven track record in state-of-the-art AI safety research (experience)
  • Experience collaborating with cross-functional teams (T&S, legal, policy) (experience)
  • Ability to identify model risks and propose mitigation strategies (experience)
  • Excitement about deploying safe AI models in real-world settings (experience)
  • Team player in collaborative research environments (experience)

Responsibilities

  • Conduct state-of-the-art research on AI safety topics including RLHF, adversarial training, and robustness
  • Implement new safety methods in OpenAI’s core model training pipelines
  • Launch safety improvements directly into OpenAI’s products
  • Set research directions and strategies for safer, more aligned AI systems
  • Collaborate with Trust & Safety, legal, policy, and other research teams
  • Evaluate and assess the safety of OpenAI models and systems
  • Identify areas of risk in AI deployments and propose mitigation strategies
  • Develop techniques to enforce nuanced safety policies without sacrificing helpfulness
  • Enhance model robustness against adversarial attacks and malicious use
  • Address privacy and security risks in powerful AI models
  • Build trustworthy AI for safety-critical domains
  • Contribute to OpenAI's mission of safe AGI deployment
  • Learn from real-world deployments to improve safety systems
  • Distribute AI benefits responsibly while prioritizing safety

Benefits

  • general: Competitive salary with equity in a high-growth AI company
  • general: Comprehensive health, dental, and vision insurance
  • general: 401(k) matching and retirement planning support
  • general: Unlimited PTO and flexible work policies
  • general: Generous parental leave and family benefits
  • general: Professional development stipend for conferences and courses
  • general: Onsite gym, wellness programs, and mental health support
  • general: Catered meals, snacks, and beverages daily
  • general: Commuter benefits and relocation assistance
  • general: Cutting-edge research environment with top talent
  • general: Impactful work shaping the future of safe AGI
  • general: Collaborative culture fostering innovation and trust
  • general: Access to state-of-the-art compute resources
  • general: Opportunities to publish groundbreaking safety research

Target Your Resume for "Researcher, Robustness & Safety Training Careers at OpenAI - San Francisco, California | Apply Now!" , OpenAI

Get personalized recommendations to optimize your resume specifically for Researcher, Robustness & Safety Training Careers at OpenAI - San Francisco, California | Apply Now!. Takes only 15 seconds!

AI-powered keyword optimization
Skills matching & gap analysis
Experience alignment suggestions

Check Your ATS Score for "Researcher, Robustness & Safety Training Careers at OpenAI - San Francisco, California | Apply Now!" , OpenAI

Find out how well your resume matches this job's requirements. Get comprehensive analysis including ATS compatibility, keyword matching, skill gaps, and personalized recommendations.

ATS compatibility check
Keyword optimization analysis
Skill matching & gap identification
Format & readability score

Tags & Categories

OpenAI researcher jobs San FranciscoAI safety researcher careersRLHF researcher OpenAIadversarial training jobsrobustness safety training OpenAIAGI safety research positionsmodel safety researcher San FranciscoAI alignment careers OpenAIPhD ML safety jobsOpenAI safety systems teamreinforcement learning human feedback jobsAI robustness researcherfairness bias mitigation AI jobsdeploy safe AI models careersOpenAI San Francisco researcheradvanced AI safety researchmalicious use AI defense jobsprivacy security AI researchertrustworthy AI safety critical domainssenior AI safety PhD OpenAIcross functional AI safety collaborationSafety Systems

Answer 10 quick questions to check your fit for Researcher, Robustness & Safety Training Careers at OpenAI - San Francisco, California | Apply Now! @ OpenAI.

Quiz Challenge
10 Questions
~2 Minutes
Instant Score

Related Books and Jobs

No related jobs found at the moment.

OpenAI logo

Researcher, Robustness & Safety Training Careers at OpenAI - San Francisco, California | Apply Now!

OpenAI

Researcher, Robustness & Safety Training Careers at OpenAI - San Francisco, California | Apply Now!

full-timePosted: Feb 10, 2026

Job Description

Researcher, Robustness & Safety Training at OpenAI - San Francisco Careers

Join OpenAI's Safety Systems team as a Researcher, Robustness & Safety Training in San Francisco, California. This senior-level role is at the forefront of AI safety research, focusing on RLHF, adversarial training, and model robustness to ensure safe AGI deployment. Apply now for this high-impact position driving OpenAI's mission.

Role Overview

The Researcher, Robustness & Safety Training position at OpenAI in San Francisco is a pivotal role within the Model Safety Research team. This team is dedicated to advancing AI safety capabilities, ensuring that OpenAI's most powerful models can be deployed safely to benefit society. As AI systems grow more capable, new challenges emerge in enforcing safety policies, robustness against adversaries, privacy protection, and trustworthiness in critical domains.

In this role, you'll conduct cutting-edge research on topics like Reinforcement Learning from Human Feedback (RLHF), adversarial training, and robustness. You'll implement these innovations directly into OpenAI's core training pipelines and launch safety features in products used by millions. This isn't just research—it's about shaping the future of safe AGI, collaborating across teams to meet the highest safety standards, and making a tangible impact on humanity.

OpenAI's commitment to safety is unwavering. The Safety Systems team leads efforts to deploy models responsibly, learning from real-world use while mitigating risks. If you have 4+ years in AI safety, a PhD in ML or related fields, and a passion for aligned AGI, this San Francisco-based role offers unparalleled opportunities to influence global AI safety standards.

Key focus areas include balancing safety with helpfulness, defending against malicious actors, securing user privacy, and building trust in high-stakes applications like healthcare or autonomous systems. Your work will directly contribute to OpenAI's charter for universally beneficial AI.

Key Responsibilities

As a Researcher in Robustness & Safety Training at OpenAI, your responsibilities will span research, implementation, and strategy:

  • Lead state-of-the-art research in AI safety, specializing in RLHF, adversarial training, and model robustness techniques.
  • Develop and implement novel methods within OpenAI’s core model training infrastructure for enhanced safety.
  • Deploy safety improvements into production products, ensuring real-world reliability.
  • Define research directions to make AI systems safer, more aligned, and resilient to adversarial threats.
  • Collaborate with Trust & Safety, legal, policy, and engineering teams to uphold rigorous safety protocols.
  • Perform comprehensive evaluations of model safety, pinpointing risks and crafting mitigation strategies.
  • Engineer defenses against adversarial attacks, maintaining performance in hostile environments.
  • Tackle privacy risks through advanced techniques like differential privacy in large models.
  • Enhance fairness and reduce biases across diverse deployment scenarios.
  • Build trustworthy AI for safety-critical domains, validating behaviors under uncertainty.
  • Analyze deployment data to iteratively improve safety systems.
  • Publish and present findings, advancing the broader AI safety field.
  • Mentor junior researchers and contribute to team strategy.

These tasks position you at the intersection of research and deployment, ensuring OpenAI's AI benefits society safely.

Qualifications

To excel as a Researcher, Robustness & Safety Training in OpenAI's San Francisco office, you should possess:

  • 4+ years in AI safety research, with expertise in RLHF, adversarial training, robustness, fairness, and biases.
  • Ph.D. or equivalent in computer science, machine learning, statistics, or related disciplines.
  • Hands-on experience with safety protocols for deploying large-scale AI models.
  • Deep knowledge of deep learning architectures, optimization, and scaling laws.
  • Strong programming skills in Python, PyTorch/TensorFlow, and ML frameworks.
  • Demonstrated passion for OpenAI’s mission and alignment with its safety charter.
  • Proven ability to collaborate in cross-functional, fast-paced environments.
  • Track record of publishing in top ML conferences (NeurIPS, ICML, ICLR).
  • Experience evaluating model vulnerabilities and proposing scalable fixes.
  • Understanding of real-world AI risks like jailbreaks, data poisoning, and misuse.
  • Excitement for tackling open problems in AGI safety.

Ideal candidates thrive in collaborative settings and are driven by the challenge of safe AI deployment.

Salary & Benefits

OpenAI offers competitive compensation for this senior Researcher role in San Francisco, estimated at $320,000 - $500,000 USD yearly, including base salary, equity, and bonuses. Total compensation reflects experience and impact.

Benefits include:

  • Comprehensive medical, dental, vision coverage.
  • 401(k) with generous matching.
  • Unlimited vacation and flexible hours.
  • Parental leave up to 20 weeks.
  • Professional growth stipend ($10K+ annually).
  • Onsite fitness center, catered meals, wellness programs.
  • Relocation support for San Francisco move.
  • Equity in OpenAI, a leader in AGI.
  • Mental health resources and employee assistance.
  • Commuter benefits and subsidized housing options.
  • Access to cutting-edge hardware and compute clusters.
  • Opportunities for global conferences and sabbaticals.

This package supports top talent focused on AI safety innovation.

Why Join OpenAI?

OpenAI is the pioneer in safe AGI development, with products like ChatGPT transforming the world. Joining the Safety Systems team in San Francisco means working with world-class researchers on problems that matter: making powerful AI safe for everyone.

Our culture emphasizes transparency, collaboration, and impact. You'll shape safety standards for future AI, publish influential work, and deploy changes affecting billions. San Francisco's vibrant tech ecosystem complements OpenAI's innovative environment.

With a mission to benefit humanity, OpenAI invests heavily in safety. This role offers intellectual challenge, massive scale, and the chance to define AGI safety. Past team members have advanced RLHF globally—your contributions could do the same.

OpenAI provides resources unmatched elsewhere: unlimited compute, expert peers, and direct mission alignment. If you're passionate about robust AI, this is your opportunity to lead.

How to Apply

Ready to advance AI safety at OpenAI? Submit your resume, cover letter highlighting AI safety experience, and links to publications or GitHub. Include why you're excited about OpenAI's mission.

Applications are reviewed on a rolling basis. Top candidates advance to research interviews, safety deep-dives, and team fits. We prioritize diverse perspectives aligned with our charter.

Apply now via OpenAI's careers page for the Researcher, Robustness & Safety Training role in San Francisco. Shape the future of safe AI today!

This page optimized for searches like 'OpenAI AI safety researcher jobs San Francisco', 'RLHF researcher careers', 'AGI safety training positions'. Updated 2024.

Locations

  • San Francisco, California, United States

Salary

Estimated Salary Rangehigh confidence

336,000 - 550,000 USD / yearly

Source: ai estimated

* This is an estimated range based on market data and may vary based on experience and qualifications.

Skills Required

  • RLHF (Reinforcement Learning from Human Feedback)intermediate
  • Adversarial Trainingintermediate
  • AI Robustnessintermediate
  • Model Safety Researchintermediate
  • Deep Learningintermediate
  • Machine Learningintermediate
  • AI Alignmentintermediate
  • Fairness and Bias Mitigationintermediate
  • Privacy in AIintermediate
  • Security Risk Assessmentintermediate
  • PyTorchintermediate
  • TensorFlowintermediate
  • Large Language Modelsintermediate
  • Cross-Functional Collaborationintermediate
  • Research Strategyintermediate
  • Model Evaluationintermediate
  • Safety Policy Enforcementintermediate
  • Adversarial Robustnessintermediate

Required Qualifications

  • 4+ years of experience in AI safety research, particularly in RLHF, adversarial training, robustness, fairness, and biases (experience)
  • Ph.D. or equivalent degree in computer science, machine learning, or a related field (experience)
  • Experience in safety work for AI model deployment (experience)
  • In-depth understanding of deep learning research (experience)
  • Strong engineering skills in implementing ML models (experience)
  • Passion for AI safety and OpenAI's mission to build safe AGI (experience)
  • Alignment with OpenAI's charter for universally beneficial AI (experience)
  • Proven track record in state-of-the-art AI safety research (experience)
  • Experience collaborating with cross-functional teams (T&S, legal, policy) (experience)
  • Ability to identify model risks and propose mitigation strategies (experience)
  • Excitement about deploying safe AI models in real-world settings (experience)
  • Team player in collaborative research environments (experience)

Responsibilities

  • Conduct state-of-the-art research on AI safety topics including RLHF, adversarial training, and robustness
  • Implement new safety methods in OpenAI’s core model training pipelines
  • Launch safety improvements directly into OpenAI’s products
  • Set research directions and strategies for safer, more aligned AI systems
  • Collaborate with Trust & Safety, legal, policy, and other research teams
  • Evaluate and assess the safety of OpenAI models and systems
  • Identify areas of risk in AI deployments and propose mitigation strategies
  • Develop techniques to enforce nuanced safety policies without sacrificing helpfulness
  • Enhance model robustness against adversarial attacks and malicious use
  • Address privacy and security risks in powerful AI models
  • Build trustworthy AI for safety-critical domains
  • Contribute to OpenAI's mission of safe AGI deployment
  • Learn from real-world deployments to improve safety systems
  • Distribute AI benefits responsibly while prioritizing safety

Benefits

  • general: Competitive salary with equity in a high-growth AI company
  • general: Comprehensive health, dental, and vision insurance
  • general: 401(k) matching and retirement planning support
  • general: Unlimited PTO and flexible work policies
  • general: Generous parental leave and family benefits
  • general: Professional development stipend for conferences and courses
  • general: Onsite gym, wellness programs, and mental health support
  • general: Catered meals, snacks, and beverages daily
  • general: Commuter benefits and relocation assistance
  • general: Cutting-edge research environment with top talent
  • general: Impactful work shaping the future of safe AGI
  • general: Collaborative culture fostering innovation and trust
  • general: Access to state-of-the-art compute resources
  • general: Opportunities to publish groundbreaking safety research

Target Your Resume for "Researcher, Robustness & Safety Training Careers at OpenAI - San Francisco, California | Apply Now!" , OpenAI

Get personalized recommendations to optimize your resume specifically for Researcher, Robustness & Safety Training Careers at OpenAI - San Francisco, California | Apply Now!. Takes only 15 seconds!

AI-powered keyword optimization
Skills matching & gap analysis
Experience alignment suggestions

Check Your ATS Score for "Researcher, Robustness & Safety Training Careers at OpenAI - San Francisco, California | Apply Now!" , OpenAI

Find out how well your resume matches this job's requirements. Get comprehensive analysis including ATS compatibility, keyword matching, skill gaps, and personalized recommendations.

ATS compatibility check
Keyword optimization analysis
Skill matching & gap identification
Format & readability score

Tags & Categories

OpenAI researcher jobs San FranciscoAI safety researcher careersRLHF researcher OpenAIadversarial training jobsrobustness safety training OpenAIAGI safety research positionsmodel safety researcher San FranciscoAI alignment careers OpenAIPhD ML safety jobsOpenAI safety systems teamreinforcement learning human feedback jobsAI robustness researcherfairness bias mitigation AI jobsdeploy safe AI models careersOpenAI San Francisco researcheradvanced AI safety researchmalicious use AI defense jobsprivacy security AI researchertrustworthy AI safety critical domainssenior AI safety PhD OpenAIcross functional AI safety collaborationSafety Systems

Answer 10 quick questions to check your fit for Researcher, Robustness & Safety Training Careers at OpenAI - San Francisco, California | Apply Now! @ OpenAI.

Quiz Challenge
10 Questions
~2 Minutes
Instant Score

Related Books and Jobs

No related jobs found at the moment.