About the job
Your Contribution at Lila
At Lila, we are assembling a dynamic and empowered AI safety team dedicated to proactively addressing the potential risks associated with scientific superintelligence. This team will collaborate closely with all core departments, including science, model training, and lab integration, to craft a customized safety strategy that aligns with our unique objectives and deployment methods. Key responsibilities will encompass the development of technical safety strategies, engagement with the broader ecosystem, and the creation of essential technical documentation, including risk assessments and capability evaluations.
Your Key Responsibilities
- Design and execute evaluations to identify scientific risks—focusing on both established and emerging threats—from state-of-the-art scientific models integrated with automated physical laboratories.
- Develop initial proof-of-concept safety measures, such as machine learning models designed to detect and mitigate unsafe behaviors from scientific AI models and physical laboratory outputs.
- Gain a comprehensive understanding of various model capabilities, primarily within scientific contexts but also extending to non-scientific domains (e.g., persuasion, deception) to shape Lila's overarching safety strategy.
- Engage in high-quality research initiatives as needed to evaluate and restrict scientific capabilities effectively.
Qualifications for Success
- A Bachelor's degree in a relevant technical field (e.g., computer science, engineering, machine learning, mathematics, physics, statistics) or equivalent experience.
- Proficient programming skills in Python and hands-on experience with machine learning frameworks (such as Inspect) for large-scale evaluations and structured testing.
- Demonstrated experience in constructing evaluations or conducting red-teaming exercises pertaining to CBRN/cyber risks or frontier model capabilities, encompassing both unsafe and benign attributes.
- Background in designing and/or implementing AI safety frameworks in cutting-edge AI enterprises.
- Exceptional ability to communicate intricate technical concepts and issues to audiences without technical expertise.
Desirable Qualifications
- A Master’s or PhD in a field pertinent to safety evaluations of AI models within scientific areas or another technical discipline.
- Publications in AI safety, evaluations, or model behavior at leading ML/AI conferences (such as NeurIPS, ICML, ICLR, ACL) or model release documentation.
- Experience exploring risks arising from novel scientific advancements (e.g., biosecurity, computational biology) or utilizing specialized scientific tools (e.g., large-scale foundational models in science).

