Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2010 May 1;28(5):807-813.
doi: 10.1016/j.imavis.2009.08.002.

Multi-PIE

Affiliations

Multi-PIE

Ralph Gross et al. Proc Int Conf Autom Face Gesture Recognit. .

Abstract

A close relationship exists between the advancement of face recognition algorithms and the availability of face databases varying factors that affect facial appearance in a controlled manner. The CMU PIE database has been very influential in advancing research in face recognition across pose and illumination. Despite its success the PIE database has several shortcomings: a limited number of subjects, single recording session and only few expressions captured. To address these issues we collected the CMU Multi-PIE database. It contains 337 subjects, imaged under 15 view points and 19 illumination conditions in up to four recording sessions. In this paper we introduce the database and describe the recording procedure. We furthermore present results from baseline experiments using PCA and LDA classifiers to highlight similarities and differences between PIE and Multi-PIE.

PubMed Disclaimer

Figures

Figure 1
Figure 1
Variation captured in the Multi-PIE face database.
Figure 2
Figure 2
Setup for the high resolution image capture. Subjects were seated in front of a blue background and recorded using a Canon EOS 10D camera along with a Macro Ring Lite MR-14EX ring flash.
Figure 3
Figure 3
Camera labels and approximate locations inside the collection room. There were 13 cameras located at head height, spaced in 15° intervals. Two additional cameras (08_1 and 19_1) were located above the subject, simulating a typical surveillance camera view. Each camera had one flash attached to it with three additional flashes being placed between cameras 08_1 and 19_ 1.
Figure 4
Figure 4
Panoramic image of the collection room. 14 of the 15 cameras used are highlighted with yellow circles, 17 of the 18 flashes are highlighted with white boxes with the occluded camera/flash pair being located right in front of the subject in the chair. The monitor visible to the left was used to ensure accurate positioning of the subject throughout the recording session.
Figure 5
Figure 5
Example high resolution images of one subject across all four recording session. For session 1 we recorded a smile image in addition to the neutral image.
Figure 6
Figure 6
Example images of the facial expressions recorded in the four different sessions. The images shown here were recorded by the camera directly opposite the subject with the flash attached to said camera illuminating the scene.
Figure 7
Figure 7
Montage of all 15 cameras views in the CMU Multi-PIE database, shown with frontal flash illumination. 13 of the 15 cameras were located at head height with two additional cameras mounted higher up to obtain views typically encountered in surveillance applications. The camera labels are shown in each image (see Figure 3).
Figure 8
Figure 8
Computation of flash-only images as difference between flash and non-flash images.
Figure 9
Figure 9
PCA performance for PIE and Multi-PIE across recording sessions. Since PIE only contains images from one session, gallery and probe images are identical, resulting in perfect recognition (PIE 68). For Multi-PIE, accuracies decrease with increasing time difference between the acquistition of gallery and probe images. We show results for a 68 subject subset of Multi-PIE (M-PIE 68) as well as for the full set of available subjects (M-PIE full).
Figure 10
Figure 10
Comparison of PCA and LDA recognition across illumination conditions in PIE and Multi-PIE. For matched experimental conditions (PCA PIE 68 in (a) and PCA M-PIE 68 in (b)), performance is similar, experimentally veryifying the similarity in the physical setup of the two collections. Whereas LDA performance over PIE nearly saturates at 95%, the average accuracy over Multi-PIE using the largest test set (71.3%) indicates further room for improvement.
Figure 11
Figure 11
PCA and LDA performance on Multi-PIE across illumination and sessions. Results shown are averages over all illumination conditions. Performance decreases with increasing time difference between the recording of gallery and probe images. Performance overall is lower than in Figure 9 due to the influence of the illumination differences.
Figure 12
Figure 12
PCA performance on Multi-PIE across expressions and illuminations. We use the neutral images (without flash illumination) recorded in the same session as gallery and the expression images under all illumination conditions as probe. The combined influence of illumination and expression reduces accuracies drastically, with PCA rates varying between 13.7% (for scream) and 21.1% (for squint). LDA accuracies are higher on average (41.4% vs. 18.5%), peaking at 50.1% (again for squint). As comparison we also show PCA recognition rates for identical gallery and probe illumination conditions (labeled “PCA M-PIE”).

References

    1. Belhumeur P, Hespanha J, Kriegman D. Eigenfaces vs. Fisherfaces: Recognition using class specific linear projection. IEEE Transactions on Pattern Analysis and Machine Intelligence. 1997;19(7):711–720.
    1. Beveridge R, Bolme D, Draper B, Teixeira M. The CSU face identification evaluation system. Machine Vision and Applications. 2005;16:128–138.
    1. Cootes T, Edwards G, Taylor C. Active appearance models. IEEE Transaction on Pattern Analysis and Machine Intelligence. 2001;23(6)
    1. Delac K, Grgic M, Grgic S. Independent comparative study of PCA, ICA, and LDA on the FERET data set. International Journal of Imaging Systems and Technology. 2005;15(5):252–260.
    1. Georghiades A, Kriegman D, Belhumeur P. From few to many: generative models for recognition under variable pose and illumination. IEEE Transaction on Pattern Analysis and Machine Intelligence. 2001;23(6):643–660.

LinkOut - more resources