CMU-CS-03-119
Computer Science Department
School of Computer Science, Carnegie Mellon University



CMU-CS-03-119

Preserving Privacy by De-identifying Facial Images

Elaine Newton, Latanya Sweeney, Bradley Malin

March 2003

CMU-CS-03-119.ps
CMU-CS-03-119.ps.gz
CMU-CS-03-119.pdf


Keywords: Video surveillance, privacy, de-identification, privacy-preserving data mining, k-anonymity


In the context of sharing video surveillance data, a significant threat to privacy is face recognition software, which can automatically identify known people, such as from a database of drivers' license photos, and thereby track people regardless of suspicion. This paper introduces an algorithm to protect the privacy of individuals in video surveillance data by de-identifying faces such that many facial characteristics remain but the face cannot be reliably recognized. A trivial solution to de-identifying faces involves blacking out each face. This thwarts any possible face recognition, but because all facial details are obscured, the result is of limited use. Many ad hoc attempts, such as covering eyes or randomly perturbing image pixels, fail to thwart face recognition because of the robustness of face recognition methods. This paper presents a new privacy-enabling algorithm, named k-Same, that scientifically limits the ability of face recognition software to reliably recognize faces while maintaining facial details in the images. The algorithm determines similarity between faces based on a distance metric and creates new faces by averaging image components, which may be the original image pixels (k-Same Pixel) or eigenvectors (k-Same-Eigen). Results are presented on a standard collection of real face images with varying k.

25 pages


Return to: SCS Technical Report Collection
School of Computer Science homepage

This page maintained by reports@cs.cmu.edu