companyOpenAI logo

Researcher - Loss of Control at OpenAI | San Francisco

OpenAISan Francisco
On-site Full-time

Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.


Unlock Your Potential

Generate Job-Optimized Resume

One Click And Our AI Optimizes Your Resume to Match The Job Description.

Is Your Resume Optimized For This Role?

Find Out If You're Highlighting The Right Skills And Fix What's Missing

Experience Level

Entry Level

Qualifications

QualificationsStrong technical background in AI and machine learning, demonstrated ability to collaborate effectively with multidisciplinary teams, and a commitment to ethical AI development.

About the job

About Our Team

The Safety Systems organization at OpenAI is dedicated to ensuring the responsible development and deployment of our most advanced AI models. We create evaluations, safeguards, and safety frameworks to ensure that our models operate as intended in real-world scenarios.

The Preparedness team plays a critical role within the Safety Systems organization, guided by OpenAI’s Preparedness Framework.

As frontier AI models hold the potential for significant benefits to humanity, they also present escalating risks. The Preparedness team is essential in preparing for the development of increasingly capable frontier AI models, focusing on identifying, tracking, and preparing for catastrophic risks associated with these technologies.

The mission of the Preparedness team includes:

  1. Monitoring and predicting the evolving capabilities of frontier AI systems, especially regarding risks that could have catastrophic consequences.

  2. Establishing concrete procedures, infrastructure, and partnerships to mitigate these risks and safely advance the development of powerful AI systems.

The Preparedness team integrates capability assessment, evaluations, internal red teaming, and mitigations for frontier models, coordinating overall AGI preparedness. This fast-paced and impactful work holds significant importance for both our organization and society.

About the Role

As frontier AI systems become more capable, they exhibit greater autonomy, the ability to pursue long-term goals, adapt to feedback, and utilize tools. While these advancements offer immense potential, they also raise the risk of models behaving in misaligned or deceptive ways, which can be difficult to supervise or contain. Addressing the risk of loss of control is a key challenge in the safe development and deployment of advanced AI systems.

In your role as a Researcher focused on loss of control mitigations, you will design and implement an end-to-end mitigation strategy aimed at reducing the risk of intentionally subversive or inadequately controllable behaviors in OpenAI’s products and internal operations. This position necessitates strong technical expertise and close collaboration across functions to ensure that safeguards are enforceable, scalable, and effective. You will directly contribute to establishing robust protections as model capabilities evolve.

About OpenAI

OpenAI is at the forefront of AI technology, dedicated to ensuring that the most powerful AI models are developed in a safe and responsible manner. Our mission is to advance digital intelligence that can benefit humanity as a whole, while minimizing risks associated with AI advancements.

Similar jobs

Tailoring 0 resumes

We'll move completed jobs to Ready to Apply automatically.