Hany Farid, Emily Cooper and Rebecca Wexler - who happen to be three of my favorite academics of all time EVER - authored an article about our work on AI voice clones and what this means in a court of law.
Featured in the Royal Academy of Engineering Ingenia Magazine
The Royal Academy of Engineering did a very kind profile on me for the Ingenia magazine.
The article can be found here: https://www.ingenia.org.uk/articles/qa-sarah-barrington-phd-student-studying-ai-harms-and-deepfakes/. Thank you to Jasmine Wragg for coordinating, and Florence Downs for being so lovely to work with!
How AI-Voice Clones Fool Humans: Our Latest Work in Nature Scientific Reports
Out in Nature Portfolio SciReports today, our work on detecting AI-powered voice clones. TL;DR: AI-generated voices are nearly indistinguishable from human voices. Listeners could only spot a 'fake' 60% of the time. By comparison, randomly guessing gets 50% accuracy.
Releasing The DeepSpeak Dataset
Tired of using the usual poor quality, non-consensual and limited deepfake datasets found online, we decided to make our own. Now in it’s second iteration, the dataset comprises over 50 hours of footage of 500 diverse individuals (recorded with their consent), and 50 hours of deepfake video footage.
In the Washington Post Today
Delighted to open today’s Washington Post newsletter to see a prominent headline about audio deepfakes- with quotes from Hany and myself in it!
Check out the article here: https://www.washingtonpost.com/technology/interactive/2024/ai-voice-detection-trump-harris-deepfake-election/
Writing about Bad Bunny in the Berkeley Science Review
Our piece about the AI-deepfake of Bad Bunny is out in the Berkeley Science Review now.
Keynote at the Alan Turing Institute Women in AI Security Workshop
On 17th July 2024, I was delighted to be able to speak as a keynote technical presenter at the Alan Turing Institute Women in AI Security Workshop.
Talking to Vox About FKA twigs, Scarlett Johansson and Audio Deepfakes
I was invited to share my thoughts with Vox about recent developments in the entertainment industry revolving around FKA twigs, Scarlett Johansson and the looming dawn of audio deepfakes.
Our Work on NPR
Was great to be a part of this recent NPR piece on voice cloning. I got to talk to Huo Jingnan about all things VC detection, and how there’s no ‘silver bullet’ from machine learning to fix this problem.
Invited to the White House to talk AI Voice Cloning
My PI, Hany Farid, and I were delighted to join the discussion about AI voice cloning technologies at The White House in January.
Conference: Presenting our Voice Cloning Work at the IEEE International Workshop on Information Forensics and Security
How do you detect a cloned voice? The simple answer… deep learning. Hugely enjoyed presenting our different detection algorithms and the relative benefits of each at the 2023 IEEE International Workshop on Information and Forensics (WIFS) in Nuremburg, Germany.
Talk & workshop: AI & Cybersecurity for the US Department of State TechWomen Initative
Had a fantastic day leading the AI & cybersecurity session for the U.S. Department of State #TechWomen23 initiative. I presented an overview of the AI landscape at present; ranging from deepfakes and misinformation to the potential threat of cyber-fuelled nuclear warfare.
Working for OpenAI at the Confidence-Building Measures for Artificial Intelligence Workshop, 2023
Selected student facilitator for the Berkeley Risk and Security Lab & OpenAI workshop on Confidence-Building Measures for Artificial Intelligence (Jan 2023, paper released August 2023).
Paper: Detecting Real vs. Deep-fake Cloned Voices
Recently, I was fortunate to work with Romit Barua and Gautham Koorma on a project exploring the relative performances of computation approaches to cloned-voice detection. Synthetic-voice cloning technologies have seen significant advances in recent years, giving rise to a range of potential harms.
Conference: Presenting at the 2023 Nobel Prize Summit
Excited, shocked (!) and honored to share that our work in deepfake detection was recognized at The Nobel Prize summit by the Digital Public Goods Alliance and United Nations UNDP as part of their campaign in combatting disinformation.
We were invited to attend the summit and participate in the most enriching and enlightening conversations about information integrity. We met world leaders, Laureates, policymakers, and technologists from the international community who are creating real change in their fields. We were also invited to the Royal Norwegian Embassy to present our work.
We were selected as one of an array of open-source innovations aiming to enhance information integrity worldwide. More info at bit.ly/Info-Pollution #nobelprizesummit #nobelprize #unitednations. Thank you to our lab Romit Barua Gautham Koorma Hany Farid UC Berkeley School of Information, and a special shout out to Nick Merrill for sharing this amazing opportunity.
Paper: Published in Nature Scientific Reports!
Hany & I have been working for over a year devising a study to examine how well we can estimate height and weight from a single image. We compare state-of-the-art AI, computer vision, and 3D modeling techniques with estimations from experts and non-experts. The results are surprising.
Code: Detecting Deep-Fakes Through Corneal Reflections
Classical computer science techniques for the re-projection of an image onto a non-planar surface have been extended to modeling the geometry of the human eye, as seen in Ko and Nayar (2004). Yet, there are few examples of the application of these methods to the analysis of corneal reflections in deep fake detection. The work of Hu et al. (2021) explored inconsistencies in inter-eye reflections, but focussed solely on specular highlights as opposed to the reflected scene itself. This work aims to build upon these two approaches in order to investigate whether there are discernible differences between the corneal reflections of real vs. deep-fake human portraits.
Findings
The findings from this approach provide several key contributions:
A framework for automatically extracting the corneal surface of a human in a single image and generating iris-based features.
For two publicly-available datasets, we ultimately find that there are no significant discernible differences between real and fake corneal reflections (based on the performance of three classification models).
We hypothesize that this lack of differentiation between real and fake portraits is primarily due to insufficient image quality once cropped (which typically reduces a 1024x1024px photo down to approximately 50x50px). This clearly impacts the pixel representation of the corneal surface and introduces noise that distorts the extracted features. The noisy nature of the dataset is evident from the outputs of Principal Component Analysis, which indicates that 10 components are required to capture just 75% of the variation in the data.
In spite of this, the feature extraction process highlights several features of interest that, qualitatively, can be seen to differentiate between real and fake portraits. This includes the shape of the pupil, which, for the majority of humans, should form a near-perfect circle shape when facing toward the camera. Yet, in several GAN-synthesised images from our dataset, we observe inconsistencies such as straight edges, corners, and even no clear pupil at all.
Additionally, edges observed in the reflections of real corneal surfaces (particularly in outdoor or natural light settings) exhibit discernable properties such as objects or information about the scene. By contrast, several deep-fake corneal reflections were observed to contain noise and abstract lighting, not resembling a coherent scene or object. However, we hypotheise that while evident to a human, the edge count feature was skewed by noise in the low resolution cropped images, thus not providing a strong predictor.
Analysis: https://github.com/sbarrington/corneal-reflections/tree/main/01%20Paper
Public Dataset: https://github.com/sbarrington/cornea-iris-dataset (DOI: 10.5281/zenodo.7396604)
Analysis snapshots:
Paper: The ‘Fungibility’ of Non-Fungible Tokens: A Quantitative Analysis of ERC-721 Metadata
Non-Fungible Tokens (NFTs), digital certificates of ownership for virtual art, have until recently been traded on a highly lucrative and speculative market. Yet, an emergence of misconceptions, along with a sustained market downtime, are calling the value of NFTs into question.
Winning the UC Berkeley & Binance LIFT Big Ideas Contest
Since late last year, I have been collaborating on two very exciting projects regarding the feasibility of applications for Web3 technologies as part of the University of California Big Ideas contest.
Stanford University CodeX Conference: DAOs & Systems for Resilient Societies
In April, I attended the Stanford University CodeX Blockchain Group's DAOs and Systems for Resilient Societies conference as a collaborator with my colleagues from the Open Earth Foundation (OEF). OEF is a non-profit research organisation whom I have worked with on several research projects, examining the feasibility and security implications of decentralised technologies for climate applications.