Bio

Dr. Heidy Khlaaf is the Chief AI Scientist at the AI Now Institute focusing on the assessment and safety of AI within autonomous weapons systems. Overall, she specialises in the evaluation, specification, and verification of complex or autonomous software implementations in safety-critical systems. She has an extensive and broad range of expertise in leading system safety audits (e.g., IEC 61508, DO-178C), varying from UAVs to large nuclear power plants, that contribute to construction of safety cases for safety-critical software.

Dr. Khlaaf has helped establish and pioneer the field of AI Safety Engineering, and is known for her work leading the safety evaluation of Codex at OpenAI where she developed a framework that measures a model’s performance outcomes against a cross functional risk assessment, now a de facto methodology used across AI labs. She was previously the Engineering Director of the AI Assurance team at Trail of Bits, where she led the cyber evaluations as part of the launch of the UK AI Safety Institute, and unveiled the LeftoverLocals vulnerability.

Her unique expertise at the intersection of Systems Software Engineering and Machine Learning has allowed her to lead, contribute, and produce the development of various standards and auditing frameworks for safety related applications and their development. This includes policy and regulatory frameworks for US and UK Regulators that enable the assurance of AI and ML to be safely deployed within critical systems. She is currently part of the Network of Experts for UNSG’s AI Advisory Body, and an ISO SC 42 (Artificial Intelligence) Committee Member via the British Standards Institute. She has been featured in TIME, NPR, Politico, Vox, WIRED, and many other media outlets

She completed her Computer Science PhD at University College London in 2017, where she was advised by Nir Piterman. She was a recipient of the prestigious NSF GRFP award. Her work focused on the temporal verification, termination, and non-termination of infinite-state software systems. She has won a best paper award at CAV 2015, and a subsequent invitation to JACM, for her work on the first automated algorithm to verify CTL* verification for infinite-state systems.

Download CV

Featured In

Thesis

"The Past, Present, and Future(s): Verifying Temporal Software Properties", Heidy Khlaaf. PhD Dissertation. Department of Computer Science, University College London, 2018. PDF

Publications

* indicates first author(s)

LeftoverLocals: Listening to LLM Responses Through Leaked GPU Local Memory , T. Sorensen*, H. Khlaaf* arXiv:2401.16603 [cs], January 2024.

Toward Comprehensive Risk Assessments and Assurance of AI-Based Systems, H. Khlaaf* Trail of Bits, March 2023.

A Hazard Analysis Framework for Code Synthesis Large Language Models, H. Khlaaf*, P. Mishkin*, J. Achiam, G. Krueger, M. Brundage arXiv:2207.14157 [cs], July 2022.

Evaluating Large Language Models Trained on Code (Codex/Copilot Paper), Chen et al. arXiv:2107.03374 [cs], July 2021a.

"97 Things Every SRE Should Know", edited by Emil Stolarsky and Jaime Woo. O'Reilly Media Inc., November 2020.

Toward Trustworthy AI Development: Mechanisms for Supporting Verifiable Claims, April 2020. 59 co-authors from 29 organisations, including tech companies and academic groups such as: Open AI, Leverhulme Centre for the Future of Intelligence, University of Oxford, Partnership on AI, Adelard, Mila, Google Brain, and many others.

Disruptive Innovations and Disruptive Assurance: Assuring Machine Learning and Autonomy R. Bloomfield* and H. Khlaaf* with P. Ryan Conmy, G. Fletcher. IEEE Computer, 52(9): 82-89 (2019).

Verifying Increasingly Expressive Temporal Logics for Infinite-State Systems H. Khlaaf* with B. Cook and N. Piterman*. Journal of ACM, 64, 2, Article 15 (May 2017), 39 pages.

"T2: Temporal Property Verification" M. Brockschmidt* and H. Khlaaf* with B. Cook, S. Ishtiaq, and N. Piterman Tools and Algorithms for the Construction and Analysis of Systems, Eindhoven, Netherlands, 2016. PDF

"On Automation of CTL* Verification for Infinite-State Systems" H. Khlaaf* with B. Cook and N. Piterman*. Computer Aided Verification, San Francisco, USA, 2015. Best Paper Award at CAV 2015, Invited Submission to JACM. PDF

"Fairness for Infinite-State Systems" H. Khlaaf* with B. Cook and N. Piterman*. Tools and Algorithms for the Construction and Analysis of Systems, London, United Kingdom, 2015. PDF

"Faster Temporal Reasoning for Infinite-State Programs" H. Khlaaf* with B. Cook and N. Piterman. Formal Methods in Computer-Aided Design, Lausanne, Switzerland, 2014. PDF

"Abstract: Fairness for Infinite-State Systems" H. Khlaaf* with B. Cook and N. Piterman. 14th International Workshop on Termination, Vienna, Austria, 2014.

Tech

"Auditing safety-critical AI systems"
BSI-VdTÜV AI Forum On Auditing AI-Systems: From Basics to Applications (German Federal Office for Information Security ), Invited Speaker, Berlin, Germany, 2020. (~150 attendees)

"Applicable and Achievable Formal Verification"
SRECon 2019 Keynote Speaker, Dublin, Ireland, 2019. (~800 attendees)

"Standards We Love"
Papers We Love @ Strangeloop, Invited Speaker, St. Louis, Missouri, 2018. (~500 attendees)

"Lessons from F#: From Academic Prototypes to Safety-Critical Systems"
F# eXchange, Invited Speaker, London, UK, 2018.

"Determining Software Safety in Critical Systems"
Github Constellation, Invited Speaker, London, UK, 2018. (~350 attendees)

Academic

"T2: Temporal Property Verification". Tools and Algorithms for the Construction and Analysis of Systems, Eindhoven, Netherlands, 2016.

"On Automation of CTL* Verification for Infinite-State Systems". Computer Aided Verification, San Francisco, USA, 2015. Best Paper Award.

"Fairness for Infinite-State Systems". Tools and Algorithms for the Construction and Analysis of Systems, London, United Kingdom, 2015.

"Faster Temporal Reasoning for Infinite-State Programs". Formal Methods in Computer-Aided Design, Lausanne, Switzerland, 2014.

For a comprehensive list, please refer to my CV.

Features

"How AI tells Israel who to bomb"
Vox, May 2024.

"Israel under pressure to justify its use of AI in Gaza"
Politico, March 2024.

"A Flaw in Millions of Apple, AMD, and Qualcomm GPUs Could Expose AI Data"
WIRED, January 2024.

"Israel is using an AI system to find targets in Gaza. Experts say it's just the start"
NPR, December 2023.

"Autonomous Vehicles Are Driving Blind"
New York Times, October 2023.

"Sam Altman Is the Oppenheimer of Our Age"
New York Magazine, September 2023.

"To avoid AI doom, learn from nuclear safety"
MIT Technology Review, June 2023.

Writing

"How AI Can Be Regulated Like Nuclear Energy"
H. Khlaaf. TIME, October 2023.

"An Introspection of the State of Gender Equality in Climbing"
H. Khlaaf. HoldBreaker, January 2019.

"Cultural Ramifications of Technical Interviews"
H. Khlaaf. Model View Culture, Issue 23, June 2015.

Other

"Where The Wild Things Keep Playing"
Outdoor Research, a film by Krystle Wright.

"Glacéau Smartwater - Wellbeing"
Smartwater campaign, a film by Tubby Brother and Tiffany Soi.

Climbing


When not analyzing safety-critical systems, you will most likely find me climbing. I mostly enjoy bouldering and I am currently climbing around the V9 grade range outdoors. I climb both indoors and outdoors and my most recent trips have been to: Portland UK, Rocklands (South Africa), the Peak District UK, Dolomites Italy, Sintra Portugal, Magic Wood Switzerland, Albarracin Spain, Shawangunk Mountains, Brione Switzerland, Sardegna Italy, Fontainebleau France, Yosemite National Park, Grand Canyon National Park.