Harry Coppock
10 Downing Street Fellow · Research Scientist · Visiting Lecturer
HEA Fellow, PhD, PGCert, MSc, MEng
Hey, welcome to my website 🌱
I am an AI researcher passionate about ensuring AI benefits humanity. To achieve this I work on a broad range of challenges including AI evaluation, development of robust AI capabilities, and reform within Government.
Currently, I am a Research Scientist at the UK AI Security Institute (AISI), where I was the first technical hire. I am also a 10 Downing Street Fellow where I advise on and lead AI projects addressing Prime Minister priority objectives. Along with this, I am a Visiting Lecturer at Imperial College London where for one semester a year I teach the Deep Learning course.
Previously, I completed a PhD in AI at Imperial College London and spent time on secondment at The Alan Turing Institute. During my doctorate, I founded Maat, an AI due diligence consultancy that assessed the AI capabilities of companies involved in M&A and VC transactions, working with high-profile clients including Pfizer. Prior to this, I completed an MSc in AI at Imperial College London, following an undergraduate degree in Materials Science and Engineering.
In my spare time, I like to be in nature, working on my partner’s farm in Somerset, playing sport with friends and travelling to places that feel different to the UK.
Research Scientist
Since joining AISI as its first technical hire in November 2023, I’ve enjoyed building AISI’s AI evaluation capabilities. In this role, I’ve built out AISI’s autonomous and cyber evaluations, primarily involving capture-the-flag-style agentic evals. I’ve also been directly involved in 20+ frontier model pre-deployment evaluation cycles, working closely with labs such as Anthropic, OpenAI, and Google DeepMind to evaluate harmful capabilities of their models before public release (see our 2025 Frontier AI Trends Report for a redacted overview of these evaluations). To enable this line of work, I have also been involved with open-source evaluation software such as Inspect AI, Hibayes, and Inspect Scout. I’ve been delighted to see AISI become a world-leading lab, attracting great talent — now 170+ technical individuals — and demonstrating new ways of operating effectively within government.
A significant amount of my time has also been spent improving the science of AI evaluation, ensuring our measures are good indicators of latent capability. A subset of this work which is now public covers a rigorous framework for agentic benchmarks, also using generalised linear models to investigate LLM judge biases, and more recently on best practices for scalable oversight or transcription analysis. Internally, I have worked on inference time scaling laws, real world human uplift trials, automatic jailbreak methods and measuring the robustness properties.

10 Downing Street Innovation Fellow
In this Deputy Director role, I work closely with ministers, advising on and leading AI-related opportunities and projects. Alongside AISI, I contributed to the early work of the Incubator for Artificial Intelligence, which focused on building technical capability within government and reducing reliance on external procurement to deliver better outcomes for the UK through tech. I have had the privilege of serving across two governments, gaining significant political experience, including through the transition of power from a Conservative to a Labour government following the general election.

Visiting Lecturer
Teaching is something I enjoy, and I’m fortunate to have taught the Deep Learning course in the Department of Computing at Imperial College London for the past three years. I cover topics including efficient deep learning, the science of AI evaluation, scaling laws, generative models, and transformer architectures. You can find recordings of some of my lectures (Lecture 1, Lecture 2) and a set of lecture notes. To support my pedagogical journey, I have completed a PGCert in Higher Education, am a Fellow of the Higher Education Academy (HEA), and am actively involved in faculty working groups to ensure the department’s teaching methods adapt appropriately to AI advancements.
AI for Health
Ensuring the public has access to the best possible healthcare through the NHS is a key objective of mine. I actively work towards this goal by pursuing advancements in AI for healthcare, rigorously validating their clinical utility, and working to position the NHS to maximise its chances of adopting these technological advances for public benefit.
A example contribution of mine was a paper in Nature Machine Intelligence demonstrating that a widely accepted technology — with hundreds of millions of pounds invested — purporting to diagnose COVID status from respiratory audio was in fact driven by confounders and provided no additional clinical utility over simple symptom checkers.
I have also led a study in the NHS investigating barriers to using frontier LLMs for medication safety reviews (demo of use case). Medication errors account for half of all avoidable medication-related harm, making this an important area of work. However, we found poor translation from synthetic and medical-style QA benchmarks to real-world clinical workflows, as illustrated in the figure below. Next, I hope to work on progressing post-training methods, addressing the spiky capability landscape of current AI, ensuring better and more robust real-world outcomes for AI.
Currently, I am campaigning within the NHS for it to strengthen its technical capability — both to develop its own AI technologies and to enable it to procure external solutions in a way that avoids vendor lock-in and ensures only the highest quality products are deployed for patient care.