US20050201595A1 - Pattern characteristic extraction method and device for the same - Google Patents

Pattern characteristic extraction method and device for the same Download PDF

Info

Publication number
US20050201595A1
US20050201595A1 US10/512,194 US51219404A US2005201595A1 US 20050201595 A1 US20050201595 A1 US 20050201595A1 US 51219404 A US51219404 A US 51219404A US 2005201595 A1 US2005201595 A1 US 2005201595A1
Authority
US
United States
Prior art keywords
vectors
input
extracting
feature
pattern
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10/512,194
Inventor
Toshio Kamei
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NEC Corp
Original Assignee
NEC Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NEC Corp filed Critical NEC Corp
Assigned to NEC CORPORATION reassignment NEC CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KAMEI, TOSHIO
Publication of US20050201595A1 publication Critical patent/US20050201595A1/en
Priority to US12/133,084 priority Critical patent/US8116571B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/169Holistic features and representations, i.e. based on the facial image taken as a whole
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2132Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on discrimination criteria, e.g. discriminant analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/7715Feature extraction, e.g. by transforming the feature space, e.g. multi-dimensional scaling [MDS]; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships

Definitions

  • the similarity between patterns such as characters or human faces has been determined by extracting feature vectors from input patterns, extracting feature vectors effective for identification from the feature vectors, and comparing the feature vectors obtained from the respective patterns.
  • non-patent reference 1 Moghaddam et al., “Probabilistic Visual Learning for Object Detection”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 17, No. 7, pp. 696-710, 1997) or linear discriminant analysis is performed on the principal components of the feature vector (non-patent reference 2: W.
  • covariance matrices, within-class covariance matrices, and between-class covariance matrices are calculated with respect to prepared learning samples, and basis vectors are obtained as solutions to the eigenvalue problems in the covariance matrices.
  • the features of input feature vectors are then transformed by using these basis vectors.
  • the transformation matrix W which maximizes equation (1) can be obtained as a set of generalized eigenvectors corresponding to M large eigenvalues of equation (6) as the eigenvalue problem of a column vector w i .
  • the transformation matrix W obtained in this manner will be referred to as a discriminant matrix.
  • ⁇ B w i ⁇ i ⁇ W w i (6)
  • the principal component analysis is performed on input feature vectors, and discriminant analysis is applied to principal components with large eigenvalues. More specifically, as shown in FIG. 2 , after principal components are extracted by projecting an input feature vector by using a basis matrix obtained by the principal component analysis, a feature vector effective for identification is extracted by projecting principal components by using the discriminant matrix obtained by discriminant analysis as a basis matrix.
  • the principal component analysis using the total covariance matrix ⁇ T is no more than sequentially selecting orthogonal axes within a feature space in the axial direction in which large covariances appear. For this reason, a feature axis effective for pattern identification is lost.
  • x 1 and x 2 are features which have large variances but are irrelevant to pattern identification
  • x 3 is effective for pattern identification but has a small variance (between-class variance/within-class variance, i.e., Fisher's ratio, is large, but the variance value itself is sufficiently smaller than those of x 1 and x 2 ).
  • x 1 and x 2 are features which have large variances but are irrelevant to pattern identification
  • x 3 is effective for pattern identification but has a small variance (between-class variance/within-class variance, i.e., Fisher's ratio, is large, but the variance value itself is sufficiently smaller than those of x 1 and x 2 ).
  • FIG. 3A is the distribution of data viewed from a direction almost perpendicular to the plane defined by x 1 and x 2 , with the black circles and white circles representing data points in different classes.
  • black and white circles cannot be identified.
  • black and white circles can be separated from each other. If, however, an axis with a large variance is selected, the plane defined by x 1 and x 2 is selected as a feature space, which is equivalent to performing discrimination by seeing FIG. 3A . This makes it difficult to perform discrimination.
  • the present invention has been made in consideration of the above problems in the prior art, and has as its object to provide a feature vector transformation technique for suppressing a reduction in feature amount effective for discrimination and performing efficient feature extraction when a feature vector effective for discrimination is to be extracted from an input pattern feature vector and feature dimensions are to be compressed.
  • a pattern feature extraction method is characterized by comprising the step of expressing one of a pattern feature and a feature from an image by using a plurality of feature vectors x 1 , the step of obtaining a discriminant matrix W i of each feature vector by linear discriminant analysis with respect to each of the plurality of feature vectors x 1 , the step of obtaining in advance a discriminant matrix W T by linear discriminant analysis with respect to a feature vector y obtained by arraying vectors y i obtained by linearly transforming the vectors x i using the discriminant matrix W i , and the step of performing linear transformation specified by the discriminant matrix W i and the discriminant matrix W T .
  • This pattern feature extraction method is characterized in that the step of performing linear transformation comprises the step of compressing the number of feature dimensions by transforming a feature vector of a pattern.
  • the above pattern feature extraction method is characterized in that the step of expressing comprises the step of extracting a feature vector x i formed from pixel values obtained from a plurality of sample points with respect to a plurality of sample point sets S i preset in an image, and the step of performing linear transformation comprises the step of extracting a feature amount from the image by transforming a feature vector for each image sample set.
  • the above pattern feature extraction method is characterized in that the step of expressing comprises the step of segmenting an image into a plurality of preset local regions, and expressing a feature amount as a feature vector x i extracted for each of the plurality of local regions, and the step of performing linear transformation comprises the step of extracting a feature amount from an image by transforming a feature vector of a local region.
  • the above pattern feature extraction method is characterized by further comprising the step of performing a two-dimensional Fourier transform for an image, wherein the step of expressing comprises the step of extracting a real component and an imaginary component of a two-dimensional Fourier transform as a feature vector x i , and the step of calculating a power spectrum of a two-dimensional Fourier transform, and extracting the power spectrum as a feature vector x 2 , and in the step of performing linear transformation, a feature amount is extracted from an image by transforming a feature vector.
  • This pattern feature extraction method is characterized in that in the step of performing linear transformation, a feature amount is extracted from an image by transforming a feature vector x 1 corresponding to a real component and an imaginary component of a Fourier component and a feature vector x 2 corresponding to a power spectrum of the Fourier component by linear transformation specified by a discriminant matrix W i and a discriminant matrix W T corresponding to principal components of a feature vector x i in such a manner that dimension reduction is realized.
  • This pattern feature extraction method is characterized in that the step of expressing further comprises the step of segmenting an image into a plurality of regions, and in the step of extracting the feature vector x 2 , a two-dimensional Fourier power spectrum is calculated in each of the segmented regions.
  • the method is characterized in that in the step of segmenting, a region is segmented into regions having different sizes in a multiple manner.
  • the method is characterized in by further comprising the step of reducing feature dimensions by performing feature extraction by kernel discriminant analysis on an obtained two-dimensional Fourier power spectrum and extracting an effective feature amount.
  • the method is characterized by further comprising the step of reducing feature dimensions by performing linear transformation using a discriminant matrix obtained in advance by linear discriminant analysis with respect to an obtained two-dimensional Fourier power spectrum.
  • This pattern feature extraction method is characterized in that the step of expressing further comprises the step of calculating a power spectrum of a two-dimensional Fourier transform, the step of segmenting an image into a plurality of regions and calculating a power spectrum of a two-dimensional Fourier transform for each of the regions, and the step of extracting a vector obtained combining the respective power spectra as a feature vector x 2 .
  • a pattern feature extraction device is a pattern feature extraction device for compressing feature dimensions of a pattern feature by using linear transformation, characterized by comprising basis matrix storage means for storing a basis matrix specified by a discriminant matrix W i of feature vectors obtained by linear discriminant analysis on a plurality of feature vectors x i representing a pattern feature and a discriminant matrix W T obtained in advance by linear discriminant analysis on a feature vector y obtained by combining vectors y i obtained by performing linear transformation of the vectors x i using a discriminant matrix, and linear transformation means for compressing feature dimension by transforming a feature vector of a pattern by using the basis matrix stored by the basis matrix storage means.
  • a computer-readable storage medium is a computer-readable storage medium which records a program for causing a computer to execute pattern feature extraction to compress feature dimensions of a pattern feature by using linear transformation, the program being characterized by including a program for executing a function of expressing a pattern feature by a plurality of feature vectors x i , obtaining in advance a discriminant matrix W i of feature vectors obtained by performing linear discriminant analysis on each of the feature vectors x i , and obtaining in advance a discriminant matrix W T by linear discriminant analysis on a feature vector y obtained by combining vectors y i obtained by linear transformation of the vectors x i , and a function of compressing feature dimensions by transforming a feature vector of a pattern by linear transformation specified by the discriminant matrix W i and the discriminant matrix W T .
  • An image feature extraction method is characterized by comprising the step of obtaining a Fourier spectrum vector by calculating a Fourier spectrum of an input normalized image by using a predetermined mathematic expression, the step of extracting a multiblock Fourier amplitude vector from a Fourier amplitude of a partial image of the normalized image, the step of obtaining normalized vectors of a Fourier spectrum vector and the multiblock intensity vector by performing projection of feature vectors with respect to the Fourier spectrum vector and the multiblock intensity vector by using a basis matrix, the step of combining the normalized vectors to form a combined Fourier vector and obtaining a projection vector of the coupled value by using a second basis matrix, and the step of extracting a Fourier feature by quantizing the projection vector.
  • FIG. 1 is a block diagram showing the arrangement of a pattern feature extraction device according to an embodiment of the present invention
  • FIG. 2 is a view for explaining the prior art
  • FIG. 3 is a view for explaining the distribution of pattern features
  • FIG. 4 is a block diagram showing the arrangement of a pattern feature extraction device according to the second embodiment of the present invention.
  • FIG. 5 is a view for explaining an embodiment of the present invention.
  • FIG. 6 is a view for explaining an embodiment of the present invention.
  • FIG. 7 is a block diagram showing the arrangement of a facial image matching system according to the third embodiment of the present invention.
  • FIG. 8 is a view for explaining an embodiment of the present invention.
  • FIG. 9 is a view for explaining an embodiment of the present invention.
  • FIG. 10 is a view for explaining an embodiment of the present invention.
  • FIG. 11 is a view for explaining an embodiment of the present invention.
  • FIG. 12 is a view for explaining an embodiment of the present invention.
  • FIG. 13 is a view for explaining an embodiment of the present invention.
  • FIG. 14 is a view for explaining an embodiment of the present invention.
  • FIG. 15 is a view showing an example of a facial description according to the fifth embodiment of the present invention.
  • FIG. 16 is a view showing an example of a rule when a binary representation syntax is used in the fifth embodiment of the present invention.
  • FIG. 17 is a view for explaining how to extract a Fourier feature (FourierFeature) in the fifth embodiment of the present invention.
  • FIG. 18 is a view showing an example of a Fourier spectrum scanning method in the fifth embodiment of the present invention.
  • FIG. 19 is a table showing an example of a Fourier spectrum scanning rule in the fifth embodiment of the present invention.
  • FIG. 20 is a table showing an example of scanning regions in a Fourier space for CentralFourierFeature elements in the fifth embodiment of the present invention.
  • FIG. 21 is a view showing an example of a block diagram in the fifth embodiment of the present invention.
  • FIG. 1 is a block diagram showing a pattern feature extraction device using a pattern feature extraction device according to the present invention.
  • the pattern feature extraction device will be described in detail below.
  • the pattern feature extraction device includes a first linear transformation means 11 for linearly transforming an input feature vector x 1 , a second linear transformation means 12 for linearly transforming an input feature vector x 2 , and a third linear transformation means 13 for receiving feature vectors which are transformed and dimension-reduced by the linear transformation means 11 and 12 and linearly transforming them.
  • the respective linear transformation means perform basis transformation based on discriminant analysis by using discriminant matrices obtained in advance by learning and stored in discriminant matrix storage means 14 , 15 , and 16 .
  • the input feature vectors x 1 and x 2 are feature amounts which are extracted in accordance with purposes in character recognition, face verification, and the like, and include, for example, directional features calculated from the gradient features of an image, and density features which are nothing less than the pixel values of an image.
  • Each vector includes a plurality of elements.
  • N 1 directional features are input as one feature vector x 1
  • other N 2 density values as the feature vector x 2 .
  • the discriminant matrix storage means 14 and 15 store discriminant matrices W 1 and W 2 obtained by performing linear discriminant analysis on the feature vectors x 1 and x 2 .
  • discriminant matrices may be obtained by calculating a within-class covariance matrix ⁇ W (equation (2)) and between-class covariance matrix ⁇ B (equation (3)) with respect to feature vectors in prepared learning samples in accordance with their classes, as described above.
  • Discriminant matrices can be obtained in advance by selecting eigenvectors W i corresponding to the large eigenvalues in an eigenvalue problem expressed by equation (6) with respect to these covariance matrices.
  • M 1 -dimensional and M 2 -dimensional bases smaller than input feature dimensions N 1 and N 2 are selected with respect to the feature vectors x 1 and x 2
  • the sizes of the matrices W 1 and W 2 are M 1 ⁇ N 1 and M 2 ⁇ N 2 , respectively.
  • the numbers of feature dimensions can be efficiently reduced by greatly reducing dimension counts M 1 and M 2 of feature spaces to be projected. This can effectively decrease the data amount and increase the processing speed. If, however, the number of feature dimensions is reduced too much, the discriminant performance deteriorates. This is because as the number of feature dimensions is reduced, feature amounts effective for discrimination are lost.
  • the dimension counts M 1 and M 2 of feature vectors are amounts which are easily influenced by the number of learning samples, and are preferably determined on the basis of experiments.
  • the third linear transformation means 13 projects the vectors y 1 and y 2 calculated by the first and second linear transformation means as an input feature vector y to a discriminant space.
  • a discriminant matrix W 3 to be registered in the discriminant matrix storage means 16 is obtained from learning samples as in the case wherein the first and second discriminant matrices are calculated.
  • the Lth-dimensional feature vector y is projected according to equation (9) by using the basis matrix W 3 (the size of the matrix is L ⁇ (M 1 +M 2 )), and a feature vector z to be output is obtained.
  • Z W 3 T y (9)
  • each feature vector is divided, and linear discriminant analysis is performed on learning samples of feature vectors with small dimension counts, thereby suppressing estimation errors, which tend to occur in high-dimensional feature components, and obtaining features effective for discrimination.
  • the three linear transformation means are provided to perform processing concurrently and stepwisely.
  • a linear discrimination means can be basically realized by a product-sum computing unit
  • one linear transformation means can be commonly used by switching discriminant matrices to be read out in accordance with the input feature vector to be linearly transformed.
  • the size of a necessary computing unit can be reduced by using one linear transformation means in this manner.
  • linear transformations using the respective discriminant matrices can be integrated into linear transformation using one matrix.
  • stepwise computation processing the number of times of product-sum computation is L ⁇ (M 1 +M 2 )+M 1 N 1 +M 2 N 2 .
  • the computation amount in batch computation processing as in the latter case is smaller than that in the former case, and hence high-speed computation can be realized.
  • the batch computation method can reduce the computation amount and hence is effective.
  • discriminant analysis is repeatedly performed on a feature vector having undergone discriminant analysis for each feature.
  • a plurality of elements corresponding to one feature may be divided into a plurality of vectors, discriminant analysis may be performed on each element set as an input feature, and the corresponding projected vector may be further subjected to discriminant analysis.
  • the facial image feature extraction device includes an image feature decomposition means 41 for decomposing the density feature of an input facial image, a linear transformation means 42 for projecting a feature vector in accordance with a discriminant matrix corresponding to the feature vector, and a discriminant matrix group storage means 43 for storing the respective discriminant matrices described above.
  • Techniques of extracting features from facial images include a method of positioning facial images at the eye position or the like and setting their density values as vector features, as disclosed in the above article by W. Zhao et al.
  • image feature dimensions it is difficult to perform high-precision feature extraction by directly performing linear discriminant analysis using limited learning samples. Therefore, a deterioration in feature which is caused when the principal component analysis or the like is applied is suppressed by decomposing image feature elements, performing discriminant analysis on the decomposed features, and obtaining discriminant matrices.
  • a basis matrix based on discriminant analysis can be calculated with high precision by sampling several images of each of several hundred individuals, i.e., a total of several thousand facial images. If the number of feature dimensions is as large as that of the original feature (2352 dimensions), in order to obtain similar performance with features based on discriminant analysis, it is expected that facial images of several thousand individuals must be sampled. In practice, however, it is difficult to collect such a large amount of image data, and hence this technique cannot be realized.
  • This second-stage discriminant matrix is also stored in the discriminant matrix group storage means 43 , and discriminant analysis is performed again by the linear transformation means 42 upon receiving the 180-dimensional vector of the first-stage discriminant feature.
  • the first-stage discriminant matrix and second-stage discriminant matrix may be calculated in advance as indicated by equation (10). However, when 252 dimensions ⁇ 9 regions are to be compressed to 20 dimensions ⁇ 9 regions, and the 180 dimensions are to be transformed into 50 dimensions, the calculation in two stages will reduce the memory to be used and the computation amount to 1 ⁇ 2 or less and hence is efficient.
  • a facial feature with high identification performance can be extracted.
  • z, 1 and z, 2 are to be identified.
  • the principal component analysis is performed on each entire character image to extract components having large eigenvalues, the feature “ ⁇ grave over () ⁇ ” that helps to identify and tends to be lost (for this reason, similar character identification is sometimes performed by using a specific high-order feature instead of a feature of a portion with a large eigenvalue obtained by the principal component analysis).
  • the effectiveness of segmenting an image into local regions and extracting discriminant features is similar to a phenomenon in similar character identification in character recognition. It can be thought that spatially limiting a feature that is easy to identify can ensure higher precision per unit dimension than performing discriminant analysis on principal components as a whole.
  • the image feature decomposition means 41 may sample images from an entire image and segment the sampled images instead of segmenting an image and forming a feature vector for each local region.
  • a primary feature is to be divided by nine into nine 252-dimensional vectors
  • sampling is performed from 3 ⁇ 3 regions, as shown in FIG. 6 . That is, the sampled images become reduced images with slight positional differences.
  • These reduced images are raster-scanned to be transformed into nine feature vectors.
  • Such feature vectors are used as primary vectors to calculate discriminant components. These discriminant components may be integrated to perform discriminant analysis again.
  • FIG. 7 is a block diagram showing a facial image matching system using a facial metadata creating device according to the present invention.
  • the facial image matching system will be described in detail below.
  • the facial image matching system includes a facial image input unit 71 which inputs facial images, a facial metadata creating unit 72 which creates facial metadata, a facial metadata storage unit 73 which stores extracted facial metadata, a facial similarity calculation unit 74 which calculates a facial similarity from facial metadata, a facial image database 75 which stores facial images, a control unit 76 which controls the input of images, the creation of metadata, the storage of metadata, and the calculation of facial similarities in accordance with an image registration request/retrieval request, and a display unit 77 of a display which displays facial images and other information.
  • the facial metadata creating unit 72 is comprised of a region cutting means 721 for cutting a facial region from an input facial image, and a facial image feature extraction means 722 which extracts a facial feature of the cut region.
  • the facial metadata creating unit 72 creates metadata about a facial image by extracting facial feature vectors.
  • a facial photo or the like is input upon adjustment of the size and position of the face by using the facial image input unit 71 such as a scanner or video camera.
  • a human face may be directly input from a video camera or the like.
  • the face position of the input image be detected by using a face detection technique like that disclosed in the above reference by Mohaddam and the size and the like of the facial image be automatically normalized.
  • the input facial image is registered in the facial image database 75 as needed.
  • the facial metadata creating unit 72 creates facial metadata and stores it in the facial metadata storage unit 73 .
  • the facial image input unit 71 inputs a facial image
  • the facial metadata creating unit 72 creates facial metadata as in the case of registration.
  • the created facial metadata is either registered in the facial metadata storage unit 73 or directly sent to the facial similarity calculation unit 74 .
  • the control unit 76 selects a facial image from the facial image database 75 on the basis of the result exhibiting the highest similarity, and displays the facial image on the display unit 77 or the like. An operator then checks the coincidence between the faces in the retrieved image and the registered image.
  • the facial similarity calculation unit 74 makes calculation to check whether or not the facial image specified by the ID number coincides with the retrieved image. If the calculated similarity is lower than a predetermined similarity, it is determined that the two images do not coincide with each other, and the result is displayed on the display unit 77 . Assume that this system is used for room access management. In this case, room access management can be performed by causing the control unit 76 to send an opening/closing control signal to an automatic door so as to control the automatic door, instead of displaying a facial image.
  • the facial image matching system operates in the above manner. Such operation can be implemented on a computer system.
  • facial image matching can be realized by storing a metadata creation program for executing metadata creation to be described in detail next and a similarity calculation program in a memory and executing these programs using a program control processor.
  • these programs may be recorded on a computer-readable recording medium.
  • the facial metadata creating unit 72 extracts a facial feature amount by using an image I(x, y) whose position and size have been normalized.
  • the image is preferably normalized to set the eye positions to ( 16 , 24 ) and ( 31 , 24 ) and the size to 46 ⁇ 56 pixels. In the following case, the image has been normalized to this size.
  • the region cutting means 721 then cuts a plurality of preset local regions of the facial image.
  • one region is the entire normalized facial image (f(x, y)) and the other is a central region g(x, y) of 32 ⁇ 32 pixels centered on the face. This region may be cut such that the positions of the two eyes are set to ( 9 , 12 ) and ( 24 , 12 ).
  • a central region of a face is cut in the above manner is that a stable feature can be extracted by cutting a range free from the influences of a hair style and the like even if the hair style changes (for example, when facial verification is used in a home robot, verification can be done even if the hair style changes before and after bathing). If a hair style and the like do not change (for example, personal identification within scenes in a video clip), since an improvement in verification performance can be expected by performing verification using images including hair styles, a large facial image including a hair style and a small facial image of a central portion of the face are cut.
  • the facial image feature extraction means 722 then performs two-dimensional discrete Fourier transforms for the two cut regions f(x, y) to extract a facial image feature.
  • FIG. 8 shows the more detailed arrangement of the facial image feature extraction means 722 .
  • the facial image feature extraction means includes a Fourier transform means 81 for performing a discrete Fourier transform for a normalized cut image, a Fourier power calculation means 82 for calculating the power spectrum of a Fourier-transformed Fourier frequency component, a linear transformation means 83 for regarding a feature vector obtained by raster-scanning the real and imaginary components of the Fourier frequency component calculated by the Fourier transform means 81 as a one-dimensional feature vector and extracting a discriminant feature from the principal components of the feature vector, a basis matrix storage means 84 for storing a basis matrix for the transformation, a linear transformation means 85 for extracting a discriminant feature of principal components from a power spectrum in the same manner as described above, and a basis matrix storage means 86 for storing a basis matrix for the transformation.
  • the facial image feature extraction means 722 further includes a linear transformation means 88 for normalizing each of the discriminant feature of the real and imaginary components of the Fourier feature and the discriminant feature of the power spectrum to a vector with a size of 1 , and calculating a discriminant feature of a vector obtained by combining the two feature vectors, and a discriminant matrix storage means 89 for storing a discriminant matrix for the discriminant feature.
  • discriminant features of principal components are calculated for a feature vector including the real and imaginary parts of the Fourier frequency component as elements and a feature vector including a power spectrum as an element, and a discriminant feature is calculated again for a feature vector obtained by combining the above vectors, thereby calculating the feature amount of the face.
  • This method is widely known and described in, for example, Rosenfeld et al., “Digital Picture Processing”, Kindai Kagaku Sha, pp. 20-26, and hence a description thereof will be omitted.
  • the Fourier power calculation means calculates a Fourier power spectrum
  • ⁇ square root ⁇ right arrow over (
  • the linear transformation means 83 then handles the feature amount extracted as a frequency feature as a vector.
  • a partial space to be defined in advance is set by a basis vector (eigenvector) obtained by preparing a facial image set for learning and performing discriminant analysis on the principal components of a frequency feature vector in a corresponding cut region. Since this basis vector is obtained by a widely known method described in various references including the reference by W. Zhao, a description thereof will be omitted.
  • the reason why discriminant analysis is not directly performed is that the number of dimensions of a feature vector obtained by a Fourier transform is too large to directly handle discriminant analysis.
  • this technique is one choice as a technique of extracting a first-stage feature vector.
  • a basis matrix obtained by the method of repeating discriminant analysis may be used.
  • a discriminant matrix ⁇ 1 of principal components which is to be stored in the basis matrix storage means 84 can be obtained from learning samples in advance by performing discriminant analysis on the principal components of a one-dimensional feature vector x 1 obtained by raster-scanning the real and imaginary components of a frequency feature.
  • a Fourier feature need not always be handled as a complex number, and may be handled as a real number with an imaginary component being handled as another feature element.
  • ⁇ 1 be a basis matrix for principal 25 components
  • W 1 be a discriminant matrix obtained by discriminant analysis on the vector of the principal components
  • the number of dimensions to be reduced by the principal component analysis is set to about 1/10 (about 200 dimensions) of the original feature Fourier feature. Thereafter, the number of dimensions is reduced to about 70 by this discriminant matrix.
  • This basis matrix is calculated in advance from learning samples and is used as information to be stored in the basis matrix storage means 84 .
  • ⁇ 2 T ⁇ 2 T W 2 T , which is obtained by discriminant analysis on the principal components of the feature vector, is obtained in advance by learning samples.
  • Calculating a principal component discriminant feature for each component of a Fourier feature in this manner makes it possible to obtain a discriminant feature y 1 of the principal components of the feature vector x 1 of the real and imaginary components of Fourier components, and a discriminant feature y 2 of the principal components of the feature vector x 2 of a power spectrum.
  • a normalization means 87 normalizes the size of each obtained feature vector to a unit vector with a size of 1.
  • the vector length varies depending on the position of an origin for the measurement of a vector, and hence its reference position must also be determined in advance.
  • a reference point is set by using a mean vector m i obtained from a learning sample of a projected feature vector y i .
  • a mean vector By setting a mean vector as a reference point, feature vectors are distributed around the reference point. In the case of a Gaussian distribution, in particular, feature vectors are isotropically distributed. This makes it easy to limit a distribution region in a case wherein a feature vector is quantized in the end.
  • the normalization means is provided to normalize the feature vector y 1 associated with the real and imaginary numbers of Fourier power and the feature vector y 2 associated with the power to unit vectors in advance. This makes it possible to normalize the sizes of two different kinds of feature amounts and stabilize the distribution features of feature vectors.
  • the feature vectors y 1 0 and y 2 0 normalized in this manner are combined into one feature vector y in the same manner as (equation 8), and the combined feature vector y is projected to a discriminant space by using the discriminant matrix W 3 obtained by performing linear discriminant analysis, thereby obtaining an output feature vector z.
  • the discriminant matrix W 3 for this purpose is stored in the discriminant matrix storage means 89 , and the linear transformation means 88 performs projection computation for this purpose to calculate, for example, a 24-dimensional feature vector z.
  • the size of each element must be normalized in advance.
  • the size of each element is normalized in advance in accordance with the variance value of each element.
  • the size is five bits. In this case, it suffices if the size is quantized to a value falling within the range of ⁇ 16 to 15.
  • normalization is the computation of multiplying each element by the reciprocal of the standard deviation.
  • W 3 0 T ⁇ W 3 T (15)
  • Performing normalization in this manner can perform range correction necessary for quantization.
  • computation based on the Mahalanobis distance can be performed by only calculating a simple L 2 norm in computing the norm of an inter-pattern distance at the time of collation, thereby reducing the computation amount at the time of collation.
  • the facial image feature extraction means 722 extracts a feature vector z f from the normalized image f(x, y) in this manner.
  • a feature vector z g is extracted by the facial image feature extraction means 722 in the same manner as described above.
  • the two feature vectors z f and z g are extracted by using the facial metadata creating unit as a facial feature amount z.
  • a computer may be caused to execute the above facial metadata creation sequence by a computer program.
  • this program may be recorded on a computer-readable recording medium.
  • the facial similarity calculation unit 74 calculates a similarity d(z 1 , z 2 ) by using K-dimensional feature vectors z 1 and z 2 obtained from two facial metadata.
  • one facial image is registered, and a retrieval is performed by using one facial image.
  • a similarity may be calculated for each of a plurality of facial metadata on the registration side.
  • calculating a similarity by obtaining the mean or minimum value of similarity for each combination makes it possible to calculate a similarity for one facial data.
  • the matching system of the present invention can be applied to face verification in an image sequence by regarding the image sequence as a plurality of images.
  • this program may be recorded on a computer-readable recording medium.
  • the present invention is directed to an improvement in the facial metadata creating unit 72 according to the third invention.
  • the discriminant features of the principal components of a feature vector having the real and imaginary parts of a Fourier frequency component obtained by performing a Fourier transform for an input facial image and a feature vector having a power spectrum as an element are calculated, and the discriminant feature of a feature vector obtained by combining the respective vectors is calculated again, thereby calculating the feature amount of the face.
  • a Fourier power spectrum reflects the overall feature amount of an input image
  • components of the input pixels which contain much noise e.g., pixels around the mouth which tend to change in relative position
  • the input image is segmented into regions, and a Fourier transform is performed for each local region.
  • Discriminant analysis is then performed by using a power spectrum for each local region as a feature amount. This can reduce the influences of the feature amount of a region which locally exhibits poor discriminant performance (large within-class variance) by discriminant analysis.
  • FIG. 9 is a view for explaining an embodiment and shows the flow of feature extraction processing.
  • a 32 ⁇ 32 pixel region is segmented into four 16 ⁇ 16 pixel regions, 16 8 ⁇ 8 pixel regions, 64 4 ⁇ 4 pixel regions, 256 2 ⁇ 2 pixel regions, and 1024 1 ⁇ 1 pixel regions (which are substantially the same as the input image, and hence the input image can be used without segmentation) (S 1001 ).
  • a Fourier transform is performed in each segmented region (S 1002 ).
  • a power spectrum is then calculated (S 1003 ).
  • the above calculation is performed all the segmented regions (S 1004 ).
  • the size of a region is changed (S 1005 ).
  • the sizes of all the regions are changed (S 1006 ).
  • the principal component analysis is performed in advance to obtain in advance the basis of the principal component analysis which reduces the number of dimensions. For example, an appropriate number of dimensions is about 300. Discriminant analysis is further performed on the feature vector of this dimension count to obtain a basis which reduces the number of dimensions and corresponds to a feature axis exhibiting good discriminant performance. A basis corresponding to the principal component analysis and discriminant analysis is calculated in advance (this basis will be referred to as a PCLDA projection basis ⁇ ).
  • a discriminant feature z can be obtained by projecting the 5120-dimensional feature by linear computation using the projection basis ⁇ using this PCLDA basis.
  • the feature amount of the face can be obtained by further performing quantization and the like for this feature.
  • the 5120-dimensional feature amount can be reduced in the number of dimensions by considering the symmetry of the Fourier power spectrum and removing and not using high-frequency components. This can realize high-speed learning, reduce the amount of data required, and realize high-speed feature extraction. Therefore, the number of dimensions is preferably reduced as needed.
  • Segmenting a region into blocks and multiplexing Fourier spectra in this manner can sequentially obtain multiple expressions of feature amounts having translation universality and local feature amounts from a feature amount equivalent to an image feature (in the case of 1024 segmentations).
  • a feature amount effective for identification is selected from the multiple, redundant feature expressions by discriminant analysis, thereby obtaining a compact feature amount which provides good identification performance.
  • a Fourier power spectrum is obtained by nonlinear computation for an image, which can calculate an effective feature amount which cannot be obtained by simply applying discriminant analysis based on linear computation to the image.
  • second-stage feature extraction may be performed by using kernel discriminant analysis (discriminant analysis using a kernel technique called Kernel Fisher Discriminant Analysis, KFDA, Kernel Discriminant Analysis: KDA, or Generalized Discriminant Analysis: GDA).
  • kernel discriminant analysis discriminant analysis using a kernel technique called Kernel Fisher Discriminant Analysis, KFDA, Kernel Discriminant Analysis: KDA, or Generalized Discriminant Analysis: GDA.
  • kernel discriminant analysis For a detailed description of kernel discriminant analysis, see the reference by Q. Liu et al. (non-patent reference 3: “Kernel-based Optimized Feature Vectors Selection and Discriminant Analysis for Face Recognition”, Processing of IAPR International Conference on Pattern Recognition (ICPR), Vol. II, pp. 362-365, 2002) or the reference by G. Baudat (non-patent reference 4: Generalized Discriminant Analysis Using a Kernel Approach”, Neural Computation, Vol. 12, pp. 2385-2404, 2000).
  • Each feature vector is then normalized by using its mean value, and second-stage LDA projection is performed.
  • FIG. 12 is a view for explaining still another embodiment and shows the flow of feature extraction processing.
  • region segmentation is performed in a plurality of stages (two stages in FIG. 12 ) to extract multiple power spectra in multiple resolutions as feature amounts for discriminant analysis in consideration of the translation universality of Fourier power spectra in local regions and the reliability of the local regions.
  • Feature extraction is then performed using the optimal feature space obtained by discriminant analysis.
  • of the entire image power spectra
  • of 16 8 ⁇ 8 pixel regions obtained by segmenting the entire image into 16 regions are extracted as feature vectors.
  • a feature vector may be formed without sampling any high-frequency components for discrimination. If, for example, a feature vector is formed by sampling 1 ⁇ 4 of spectra which correspond to low-frequency components, the number of learning samples required can be reduced or the processing time required for learning and recognition can be shortened. If the number of learning data is small, discriminant analysis may be performed after the number of feature dimensions is reduced by the principal component analysis in advance.
  • Discriminant analysis is performed by using a feature vector x 2 f extracted in this manner and a learning set prepared in advance to obtain a basis matrix ⁇ 2 f in advance.
  • FIG. 9 shows an example of projection for the extraction of a discriminant feature from principal components (Principal Component Linear Discriminant Analysis; PCLDA).
  • the feature vector x 2 f is projected by using the basis matrix ⁇ 2 f , and the mean and size of the projected feature vector are normalized, thereby calculating a feature vector y 2 f .
  • the feature vector x 2 f obtained by combining the real and imaginary components of a Fourier frequency is projected by linear computation processing using a basis matrix ⁇ 1 f to obtain a feature vector with a reduced number of dimensions, and the mean and size of the vector are normalized to calculate a feature vector y 1 f .
  • a feature vector obtained by combining these vectors is projected again by using a discriminant basis ⁇ 1 f to obtain a feature vector z f .
  • This vector is quantized in, for example, five bits to extract a facial feature amount.
  • facial feature amounts are also extracted from multiple segmented regions of the 44 ⁇ 56 pixel region of the entire face, including the entire 44 ⁇ 56 pixel region, four 22 ⁇ 28 pixel regions, and 16 11 ⁇ 14 pixel pixels.
  • FIG. 13 shows another embodiment, in which PCLDA projection of a combination of a real component, an imaginary component, and a power spectrum is performed for each local region, or PCLDA projection of a feature obtained by combining a real component and imaginary component and PCLDA projection of a power spectrum are separately performed, and LDA projection is finally performed, as shown in FIG. 14 .
  • FIG. 15 shows a. description of a facial feature amount, as an example of a facial feature description, which uses a DDL representation syntax (Description Definition Language Representation Syntax) in ISO/IEC FDIS 15938-3, “Information technology Multimedia content description interface—Part 3: Visual”.
  • DDL representation syntax (Description Definition Language Representation Syntax) in ISO/IEC FDIS 15938-3, “Information technology Multimedia content description interface—Part 3: Visual”.
  • FIG. 16 shows a rule in a case wherein a binary representation syntax is used for data representation.
  • the sizes of the array components of FourierFeature and CentralFourierFeature are stored in the field of 6-bit integers without signs in numOfFourierFeature and numOfCentralFourier, and each component of FourierFeature and CentralFourierFeature is stored in the form of a 5-bit integer without a sign.
  • This field specifies the number of components of FourierFeature.
  • the allowable range is from 24 to 63.
  • This field specifies the number of components of CentralFourierFeature.
  • the allowable range is from 24 to 63.
  • This element represents a facial feature based on the cascaded LDA of the Fourier characteristics of a normalized face image.
  • the normalized face image is obtained by scaling an original image into 56 lines with 46 luminance values in each line.
  • the center positions of two eyes in the normalized face image shall be located on the 24th row and the 16th and 31st columns for the right and left eyes respectively.
  • the FourierFeature element is derived from two feature vectors; one is a Fourier Spectrum Vector x 1 f , and the other is a Multi-block Fourier Amplitude Vector x 2 f .
  • FIG. 17 illustrates the extraction process of FourierFeature. Given a normalized face image, five steps should be performed to extract the element;
  • a multi-block Fourier Amplitude Vector is extracted from the Fourier amplitudes of partial images in the normalized face image.
  • the partial images three types of images are used; (a) a holistic image, (b) quarter images, and (c) 1/16 images.
  • Multi-block Fourier Amplitude Vectors is obtained by scanning low frequency regions of each amplitude
  • the scan regions are defined in FIG. 19 .
  • the Multi-block Fourier Amplitude Vector x 2 f ( ⁇ F 1 0 ⁇ ( 0 , 0 ) ⁇ ⁇ ⁇ F 1 0 ⁇ ( 43 , 13 ) ⁇ ⁇ F 1 1 ⁇ ( 0 , 0 ) ⁇ ⁇ ⁇ F 1 1 ⁇ ( 21 , 6 ) ⁇ ⁇ F 2 1 ⁇ ( 0 , 0 ) ⁇ ⁇ F 3 1 ⁇ ( 0 , 0 ) ⁇ ⁇ ⁇ F 4 1 ⁇ ( 21 , 6 ) ⁇ ⁇ F 1 2 ⁇ ( 0 , 0 ) ⁇ ⁇ ⁇ F 16 2 ⁇ ( 10 , 2 ) ⁇ ) ( 24 )
  • the dimension of x 2 f is 856.
  • the Fourier Spectrum Vector x 1 f and Multi-block Fourier Amplitude Vector x 2 f shall be respectively projected using the PCLDA basis matrices ⁇ 1 f and ⁇ 2 f , and normalized to unit vectors y 1 f and y 2 f .
  • the PCLDA basis matrix ⁇ k f and the mean vector m k f are a basis matrix obtained by performing linear discriminant analysis on the principal components of x k f and a mean of projected vectors, respectively. The values of them are given by referring to a look-up-table calculated in advance.
  • the dimensions of y 1 f and y 2 f are 70 and 80, respectively.
  • the normalized vectors y 1 f and Y 2 f are combined to form a 150-dimensional Joint Fourier vector y 3 f , and projected using the LDA basis matrix ⁇ 3 f .
  • the quantized elements are stored as FourierFeature.
  • FourierFeature[0] represents the first quantized element w o f
  • FourierFeature[numOfFourierFeature-1] corresponds to the (numOfFourierFeature)th element w f numOfFourierFeature ⁇ 1 .
  • This element represents a facial feature based on the cascaded LDA of the Fourier characteristics of the central part in the normalized face image.
  • CentralFourierFeature is extracted in the similar way as FourierFeature.
  • a 256-dimensional Central Fourier Spectrum Vector x 1 g is produced by scanning the Fourier spectrum G(u, v) as defined in FIG. 20 .
  • STEP 2 Extraction of Multi-block Central Fourier Amplitude Vector
  • a multi-block Central Fourier Amplitude Vector x 2 g is obtained
  • the processing in STEP 3-5) are the same as those FourierFeature, for example, the Joint Central Fourier Vector y 3 g consists of the normalized vectors y 1 g and y 2 g .
  • the basis matrices ⁇ 1 g , ⁇ 2 g , and ⁇ 3 g and the mean vectors m 1 g and m 2 g for CentralFourierFeature are calculated in advance, and prepared in the form of a look-up table.
  • the size of CentralFourierFeature is indicated by numOfCentralFourierFeature.
  • Facial feature description data obtained in this manner is compact in description length but exhibits high recognition performance, and hence is an expression efficient for the storage and transmission of data.
  • the present invention may be implemented by a computer-executable program.
  • the present invention can be implemented by describing the functions indicated by steps 1 to 5 in FIG. 17 in a computer-readable program and making the program function on the computer.
  • this program may be recorded on a computer-readable recording medium.
  • FIG. 17 When the example shown in FIG. 17 is to be implemented as a device, all or some of the functions written in the block diagram of FIG. 21 may be implemented. More specifically, all or some of a normalized face image output means 211 , Fourier spectrum vector extraction means 212 , multiblock Fourier amplitude vector extraction means 213 and PCLDA projection/vector normalization means 214 may be implemented.
  • a feature vector effective for discrimination by discriminant analysis is extracted from an input pattern feature vector for each element vector, and feature extraction is performed again for the obtained feature vector by using a discriminant matrix by discriminant analysis.
  • Each embodiment described above is effective for a case wherein the number of learning samples required for discriminant analysis is limited in spite of a large pattern feature amount. That is, the number of feature dimensions can be reduced, while a loss of features effective for identification is suppressed, without necessarily using the principal component analysis.
  • the image feature extraction method, the image feature extraction device, and the recording medium storing the corresponding program in the field of pattern recognition according to the present invention are suitable for the use in a feature vector transformation technique for compressing feature dimension by extracting feature vectors effective for recognition from input feature vectors.

Abstract

An input pattern feature amount is decomposed into element vectors. For each of the feature vectors, a discriminant matrix obtained by discriminant analysis is prepared in advance. Each of the feature vectors is projected into a discriminant space defined by the discriminant matrix and the dimensions are compressed. According to the feature vector obtained, projection is performed again by the discriminant matrix to calculate the feature vector, thereby suppressing reduction of the feature amount effective for the discrimination and performing effective feature extraction.

Description

    BACKGROUND OF THE INVENTION
  • Conventionally, in the field of pattern recognition, the similarity between patterns such as characters or human faces has been determined by extracting feature vectors from input patterns, extracting feature vectors effective for identification from the feature vectors, and comparing the feature vectors obtained from the respective patterns.
  • In the case of face verification, for example, pixel values of a facial image normalized with the positions of the eyes or the like are raster-scanned to transform the pixel values into a one-dimensional feature vector, and the principal component analysis is performed by using this feature vector as an input feature vector (non-patent reference 1: Moghaddam et al., “Probabilistic Visual Learning for Object Detection”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 17, No. 7, pp. 696-710, 1997) or linear discriminant analysis is performed on the principal components of the feature vector (non-patent reference 2: W. Zhao et al., “Discriminant Analysis of Principal Components for Face Recognition”, Proceedings of the IEEE Third International Conference on Automatic Face and Gesture Recognition, pp. 336-341, 1998), thereby reducing dimensions and performing personal identification or the like based on faces by using obtained feature vectors.
  • In these methods, covariance matrices, within-class covariance matrices, and between-class covariance matrices are calculated with respect to prepared learning samples, and basis vectors are obtained as solutions to the eigenvalue problems in the covariance matrices. The features of input feature vectors are then transformed by using these basis vectors.
  • Linear discriminant analysis will be described in more detail below.
  • Linear discriminant analysis is a method of obtaining a transformation matrix W which maximizes the ratio of a between-class covariance matrix SB to a within-class covariance matrix SW of an M-dimensional vector y (=WTx) obtained when an N-dimensional feature vector x is transformed by the transformation matrix W. As such a covariance evaluation function, equation (1) as an evaluation expression is defined: J ( W ) = S B S W = W T B W W T w W ( 1 )
  • In this equation, the within-class covariance matrix ΣW and between-class covariance matrix ΣB are respectively a covariance matrix Σi within C classes ωi (i=1, 2, . . . , C; their data count ni) in a set of feature vectors x in a learning sample and a covariance matrix between the classes, and are respectively represented by: w = i = 1 C P ( ω i ) i ( 2 ) = i = 1 C ( P ( ω i ) 1 n i x x i ( x - m i ) ( x - m i ) T ) B = i = 1 C P ( ω i ) ( m i - m ) ( m i - m ) T ( 3 )
    where mi is a mean vector of a class ωi (equation (4)), and m is a mean vector of x in total (equation (5)): m i = 1 n i x x i x ( 4 ) m = i = 1 C P ( ω i ) m i ( 5 )
  • If a priori probability P(ωi) of each class ωi reflects a sample count ni in advance, it suffices to assume P(ωi)=ni/n. If each probability can be assumed to be equal, it suffice to set P(ωi)=1/C.
  • The transformation matrix W which maximizes equation (1) can be obtained as a set of generalized eigenvectors corresponding to M large eigenvalues of equation (6) as the eigenvalue problem of a column vector wi. The transformation matrix W obtained in this manner will be referred to as a discriminant matrix.
    ΣB w iiΣW w i  (6)
  • Note that a conventional linear discriminant analysis method is disclosed in, for example, non-patent reference 5: Richard O. Duda et al., “Pattern Recognition” (supervised/translated by Morio Onoue, Shingijutu Communications, 2001, pp. 113-122).
  • Assume that the number of dimensions of the input feature vector x is especially large. In this case, if small learning data is used, ΣW becomes singular. As a consequence, the eigenvalue problem of equation (6) cannot be solved by a general method.
  • As described in patent reference 1: Japanese Patent Laid-Open No. 7-296169, it is known that a high-order component with a small eigenvalue in a covariance matrix includes a large parameter estimation error, which adversely affects recognition precision.
  • According to the above article by W. Zhao et al., the principal component analysis is performed on input feature vectors, and discriminant analysis is applied to principal components with large eigenvalues. More specifically, as shown in FIG. 2, after principal components are extracted by projecting an input feature vector by using a basis matrix obtained by the principal component analysis, a feature vector effective for identification is extracted by projecting principal components by using the discriminant matrix obtained by discriminant analysis as a basis matrix.
  • According to the computation scheme for feature transformation matrices described in patent reference 1: Japanese Patent Laid-Open No. 7-296169, the number of dimensions is reduced by deleting high-order eigenvalues of total covariance matrix ΣT and corresponding eigenvectors, and discriminant analysis is applied to a reduced feature space. Deleting high-order eigenvalues of total covariance matrix and corresponding eigenvectors is equivalent to performing discriminant analysis in a space of only principal components with large eigenvalues by the principal component analysis. In this sense, this technique, like the method by W. Zhao, provides stable parameter estimation by removing high-order features.
  • The principal component analysis using the total covariance matrix ΣT, however, is no more than sequentially selecting orthogonal axes within a feature space in the axial direction in which large covariances appear. For this reason, a feature axis effective for pattern identification is lost.
  • Assume that the feature vector x is comprised of three elements (x=(x1, x2, x3)T), x1 and x2 are features which have large variances but are irrelevant to pattern identification, and x3 is effective for pattern identification but has a small variance (between-class variance/within-class variance, i.e., Fisher's ratio, is large, but the variance value itself is sufficiently smaller than those of x1 and x2). In this case, if the principal component analysis is performed and only two-dimensional values are selected, a feature space associated with x1 and x2 is selected, and the contribution of x3 effective for identification is neglected.
  • This phenomenon will be described with reference to the accompanying drawings. Assume that FIG. 3A is the distribution of data viewed from a direction almost perpendicular to the plane defined by x1 and x2, with the black circles and white circles representing data points in different classes. When viewed in the space defined by x1 and x2 (plane in FIG. 3A), black and white circles cannot be identified. When, however, viewed from a feature axis of x3 perpendicular to this plane as shown in FIG. 3B, black and white circles can be separated from each other. If, however, an axis with a large variance is selected, the plane defined by x1 and x2 is selected as a feature space, which is equivalent to performing discrimination by seeing FIG. 3A. This makes it difficult to perform discrimination.
  • In the prior art, this is a phenomenon which cannot be avoided by the principal component analysis and the technique of deleting spaces with small eigenvalues in (total) covariance matrices.
  • DISCLOSURE OF INVENTION
  • The present invention has been made in consideration of the above problems in the prior art, and has as its object to provide a feature vector transformation technique for suppressing a reduction in feature amount effective for discrimination and performing efficient feature extraction when a feature vector effective for discrimination is to be extracted from an input pattern feature vector and feature dimensions are to be compressed.
  • A pattern feature extraction method according the present invention is characterized by comprising the step of expressing one of a pattern feature and a feature from an image by using a plurality of feature vectors x1, the step of obtaining a discriminant matrix Wi of each feature vector by linear discriminant analysis with respect to each of the plurality of feature vectors x1, the step of obtaining in advance a discriminant matrix WT by linear discriminant analysis with respect to a feature vector y obtained by arraying vectors yi obtained by linearly transforming the vectors xi using the discriminant matrix Wi, and the step of performing linear transformation specified by the discriminant matrix Wi and the discriminant matrix WT.
  • This pattern feature extraction method is characterized in that the step of performing linear transformation comprises the step of compressing the number of feature dimensions by transforming a feature vector of a pattern.
  • In addition, the method is characterized in that the step of expressing comprises the step of dividing a pattern feature into a plurality of feature vectors xi, the step of obtaining the discriminant matrix WT comprises the step of calculating a feature vector yi by performing linear transformation yi=Wi Txi with respect to the feature vector xi using the discriminant matrix Wi, and the step of performing linear transformation comprises the step of compressing the number of dimensions of a pattern feature by calculating a feature vector z by calculating linear transformation z=WT Ty with respect to a vector y obtained by combining calculated feature vector yi using the discriminant matrix WT.
  • In addition, the method is characterized by further comprising the step of calculating in advance a matrix W specified by the discriminant matrices Wi and WT, wherein the step of performing linear transformation comprises the step of compressing the number of dimensions of a pattern feature by calculating a feature vector z by calculating linear transformation z=WTX with respect a feature vector x obtained by combining input feature vectors xi and the matrix W using the matrix W.
  • The above pattern feature extraction method is characterized in that the step of expressing comprises the step of extracting a feature vector xi formed from pixel values obtained from a plurality of sample points with respect to a plurality of sample point sets Si preset in an image, and the step of performing linear transformation comprises the step of extracting a feature amount from the image by transforming a feature vector for each image sample set.
  • This pattern feature extraction method is characterized in that the step of obtaining in advance the discriminant matrix WT comprises the step of calculating a feature vector yi by performing linear transformation yi=Wi Txi with respect to a plurality of feature vectors xi formed from a plurality of sample points by using the discriminant matrix Wi, and the step of performing linear transformation comprises the step of extracting a feature amount from an image by calculating a feature vector z by calculating linear transformation z=WT Ty with respect to a vector y obtained by combining calculated feature vectors yi by using the discriminant matrix WT.
  • The method is characterized by further comprising the step of calculating in advance a matrix W specified by the discriminant matrices Wi and WT, wherein the step of performing the linear transformation comprises the step of extracting a feature amount from an image by calculating a feature vector z by calculating linear transformation z=WTX of a vector x obtained by combining feature vectors xi and the matrix W by using the matrix W.
  • The above pattern feature extraction method is characterized in that the step of expressing comprises the step of segmenting an image into a plurality of preset local regions, and expressing a feature amount as a feature vector xi extracted for each of the plurality of local regions, and the step of performing linear transformation comprises the step of extracting a feature amount from an image by transforming a feature vector of a local region.
  • This pattern feature extraction method is characterized in that the step of obtaining in advance the discriminant matrix WT comprises the step of calculating a feature vector yi by performing linear transformation yi=Wi Txi with respect to a feature vector xi by using the discriminant matrix Wi, and the step of performing linear transformation comprises the step of extracting a feature amount from an image by calculating a feature vector z by calculating linear transformation z=WT Ty with respect to a vector y obtained by combining calculated feature vector yi using the discriminant matrix WT.
  • The method is characterized by further comprising the step of calculating in advance a matrix W specified by the discriminant matrices Wi and WT, wherein the step of performing linear transformation comprises the step of extracting a feature amount from an image by calculating a feature vector z by calculating linear transformation z=WTX with respect a feature vector x obtained by combining input feature vectors xi and the matrix W using the matrix W.
  • The above pattern feature extraction method is characterized by further comprising the step of performing a two-dimensional Fourier transform for an image, wherein the step of expressing comprises the step of extracting a real component and an imaginary component of a two-dimensional Fourier transform as a feature vector xi, and the step of calculating a power spectrum of a two-dimensional Fourier transform, and extracting the power spectrum as a feature vector x2, and in the step of performing linear transformation, a feature amount is extracted from an image by transforming a feature vector.
  • This pattern feature extraction method is characterized in that in the step of performing linear transformation, a feature amount is extracted from an image by transforming a feature vector x1 corresponding to a real component and an imaginary component of a Fourier component and a feature vector x2 corresponding to a power spectrum of the Fourier component by linear transformation specified by a discriminant matrix Wi and a discriminant matrix WT corresponding to principal components of a feature vector xi in such a manner that dimension reduction is realized.
  • This pattern feature extraction method is characterized by further comprising the step of calculating a discriminant feature of principal components of a feature vector x1 formed from a real component and an imaginary component based on a Fourier transform by linear transformation y11 Tx1 using a transformation matrix Ψ1 for transforming the principal components of the feature vector x1 and a basis matrix Φ1 (=(W1 TΨ1 T)T) represented by a discriminant matrix W1 corresponding to the principal components, the step of normalizing a size of an obtained feature vector y1 to a predetermined size, the step of calculating a discriminant feature of principal components of a feature vector x2 formed from a power spectrum based on Fourier transformation by using a transformation matrix Ψ2 for transforming the feature vector x2 to principal components and a basis matrix Φ2 (=(W2 TΨ2 T)T) represented by a discriminant matrix W2 corresponding to the principal components, the step of normalizing a size of an obtained feature vector y2 to a predetermined size, and the step of extracting a feature amount from an image by calculating a feature vector z by calculating linear transformation z=WT Ty using a discriminant matrix WT with respect to a feature vector y obtained by combining two feature vectors y1 and y2.
  • This pattern feature extraction method is characterized in that the step of expressing further comprises the step of segmenting an image into a plurality of regions, and in the step of extracting the feature vector x2, a two-dimensional Fourier power spectrum is calculated in each of the segmented regions.
  • In addition, the method is characterized in that in the step of segmenting, a region is segmented into regions having different sizes in a multiple manner.
  • In addition, the method is characterized in by further comprising the step of reducing feature dimensions by performing feature extraction by kernel discriminant analysis on an obtained two-dimensional Fourier power spectrum and extracting an effective feature amount.
  • The method is characterized by further comprising the step of reducing feature dimensions by performing linear transformation using a discriminant matrix obtained in advance by linear discriminant analysis with respect to an obtained two-dimensional Fourier power spectrum.
  • The method is characterized in that the step of obtaining in advance the discriminant matrix Wi comprises the step of obtaining the discriminant matrix Wi of feature vectors obtained by linear discriminant analysis on principal components of a feature vector xi (i=1, 2), and in the step of performing linear transformation, a feature amount is extracted from an image by transforming a feature vector x1 corresponding to a real component and an imaginary component of a Fourier component and a feature vector x2 corresponding to a power spectrum of the Fourier component by linear transformation specified so as to reduce dimensions by a discriminant matrix Wi and a discriminant matrix WT corresponding to principal components of the feature vector xi.
  • This pattern feature extraction method is characterized in that the step of expressing further comprises the step of calculating a power spectrum of a two-dimensional Fourier transform, the step of segmenting an image into a plurality of regions and calculating a power spectrum of a two-dimensional Fourier transform for each of the regions, and the step of extracting a vector obtained combining the respective power spectra as a feature vector x2.
  • A pattern feature extraction device according to the present invention is a pattern feature extraction device for compressing feature dimensions of a pattern feature by using linear transformation, characterized by comprising basis matrix storage means for storing a basis matrix specified by a discriminant matrix Wi of feature vectors obtained by linear discriminant analysis on a plurality of feature vectors xi representing a pattern feature and a discriminant matrix WT obtained in advance by linear discriminant analysis on a feature vector y obtained by combining vectors yi obtained by performing linear transformation of the vectors xi using a discriminant matrix, and linear transformation means for compressing feature dimension by transforming a feature vector of a pattern by using the basis matrix stored by the basis matrix storage means.
  • A computer-readable storage medium according to the present invention is a computer-readable storage medium which records a program for causing a computer to execute pattern feature extraction to compress feature dimensions of a pattern feature by using linear transformation, the program being characterized by including a program for executing a function of expressing a pattern feature by a plurality of feature vectors xi, obtaining in advance a discriminant matrix Wi of feature vectors obtained by performing linear discriminant analysis on each of the feature vectors xi, and obtaining in advance a discriminant matrix WT by linear discriminant analysis on a feature vector y obtained by combining vectors yi obtained by linear transformation of the vectors xi, and a function of compressing feature dimensions by transforming a feature vector of a pattern by linear transformation specified by the discriminant matrix Wi and the discriminant matrix WT.
  • An image feature extraction method according to the present invention is characterized by comprising the step of obtaining a Fourier spectrum vector by calculating a Fourier spectrum of an input normalized image by using a predetermined mathematic expression, the step of extracting a multiblock Fourier amplitude vector from a Fourier amplitude of a partial image of the normalized image, the step of obtaining normalized vectors of a Fourier spectrum vector and the multiblock intensity vector by performing projection of feature vectors with respect to the Fourier spectrum vector and the multiblock intensity vector by using a basis matrix, the step of combining the normalized vectors to form a combined Fourier vector and obtaining a projection vector of the coupled value by using a second basis matrix, and the step of extracting a Fourier feature by quantizing the projection vector.
  • BRIEF DESCRIPTION OF DRAWINGS
  • FIG. 1 is a block diagram showing the arrangement of a pattern feature extraction device according to an embodiment of the present invention;
  • FIG. 2 is a view for explaining the prior art;
  • FIG. 3 is a view for explaining the distribution of pattern features;
  • FIG. 4 is a block diagram showing the arrangement of a pattern feature extraction device according to the second embodiment of the present invention;
  • FIG. 5 is a view for explaining an embodiment of the present invention;
  • FIG. 6 is a view for explaining an embodiment of the present invention;
  • FIG. 7 is a block diagram showing the arrangement of a facial image matching system according to the third embodiment of the present invention;
  • FIG. 8 is a view for explaining an embodiment of the present invention;
  • FIG. 9 is a view for explaining an embodiment of the present invention;
  • FIG. 10 is a view for explaining an embodiment of the present invention;
  • FIG. 11 is a view for explaining an embodiment of the present invention;
  • FIG. 12 is a view for explaining an embodiment of the present invention;
  • FIG. 13 is a view for explaining an embodiment of the present invention;
  • FIG. 14 is a view for explaining an embodiment of the present invention;
  • FIG. 15 is a view showing an example of a facial description according to the fifth embodiment of the present invention;
  • FIG. 16 is a view showing an example of a rule when a binary representation syntax is used in the fifth embodiment of the present invention;
  • FIG. 17 is a view for explaining how to extract a Fourier feature (FourierFeature) in the fifth embodiment of the present invention;
  • FIG. 18 is a view showing an example of a Fourier spectrum scanning method in the fifth embodiment of the present invention;
  • FIG. 19 is a table showing an example of a Fourier spectrum scanning rule in the fifth embodiment of the present invention;
  • FIG. 20 is a table showing an example of scanning regions in a Fourier space for CentralFourierFeature elements in the fifth embodiment of the present invention; and
  • FIG. 21 is a view showing an example of a block diagram in the fifth embodiment of the present invention.
  • DETAILED DESCRIPTION OF EMBODIMENTS
  • (First Embodiment)
  • An embodiment of the present invention will be described in detail with reference to the accompanying drawings. FIG. 1 is a block diagram showing a pattern feature extraction device using a pattern feature extraction device according to the present invention.
  • The pattern feature extraction device will be described in detail below.
  • As shown in FIG. 1, the pattern feature extraction device according to the present invention includes a first linear transformation means 11 for linearly transforming an input feature vector x1, a second linear transformation means 12 for linearly transforming an input feature vector x2, and a third linear transformation means 13 for receiving feature vectors which are transformed and dimension-reduced by the linear transformation means 11 and 12 and linearly transforming them. The respective linear transformation means perform basis transformation based on discriminant analysis by using discriminant matrices obtained in advance by learning and stored in discriminant matrix storage means 14, 15, and 16.
  • The input feature vectors x1 and x2 are feature amounts which are extracted in accordance with purposes in character recognition, face verification, and the like, and include, for example, directional features calculated from the gradient features of an image, and density features which are nothing less than the pixel values of an image. Each vector includes a plurality of elements. In this case, for example, N1 directional features are input as one feature vector x1, and other N2 density values as the feature vector x2.
  • The discriminant matrix storage means 14 and 15 store discriminant matrices W1 and W2 obtained by performing linear discriminant analysis on the feature vectors x1 and x2.
  • As described above, discriminant matrices may be obtained by calculating a within-class covariance matrix ΣW (equation (2)) and between-class covariance matrix ΣB (equation (3)) with respect to feature vectors in prepared learning samples in accordance with their classes, as described above. A priori probability P(ωi) of each class ωi may be given by P(ωi)=ni/n with sample count sample count ni being reflected.
  • Discriminant matrices can be obtained in advance by selecting eigenvectors Wi corresponding to the large eigenvalues in an eigenvalue problem expressed by equation (6) with respect to these covariance matrices.
  • When M1-dimensional and M2-dimensional bases smaller than input feature dimensions N1 and N2 are selected with respect to the feature vectors x1 and x2, M1-dimensional and M2-dimensional feature vectors y1 and y2 can be obtained by projective transformation to discriminant bases,
    y 1 =W 1 T x 1
    y 2 =W 2 T x 2  (7)
  • In this case, the sizes of the matrices W1 and W2 are M1×N1 and M2×N2, respectively.
  • The numbers of feature dimensions can be efficiently reduced by greatly reducing dimension counts M1 and M2 of feature spaces to be projected. This can effectively decrease the data amount and increase the processing speed. If, however, the number of feature dimensions is reduced too much, the discriminant performance deteriorates. This is because as the number of feature dimensions is reduced, feature amounts effective for discrimination are lost.
  • For this reason, the dimension counts M1 and M2 of feature vectors are amounts which are easily influenced by the number of learning samples, and are preferably determined on the basis of experiments.
  • The third linear transformation means 13 projects the vectors y1 and y2 calculated by the first and second linear transformation means as an input feature vector y to a discriminant space. A discriminant matrix W3 to be registered in the discriminant matrix storage means 16 is obtained from learning samples as in the case wherein the first and second discriminant matrices are calculated. The input feature vector y is a vector having elements arranged as expressed by equation (8) given below: y = ( y 1 y 2 ) ( 8 )
  • As in the case of equation (7), the Lth-dimensional feature vector y is projected according to equation (9) by using the basis matrix W3 (the size of the matrix is L×(M1+M2)), and a feature vector z to be output is obtained.
    Z=W 3 T y  (9)
  • In this manner, each feature vector is divided, and linear discriminant analysis is performed on learning samples of feature vectors with small dimension counts, thereby suppressing estimation errors, which tend to occur in high-dimensional feature components, and obtaining features effective for discrimination.
  • In the above case, the three linear transformation means are provided to perform processing concurrently and stepwisely. However, since a linear discrimination means can be basically realized by a product-sum computing unit, one linear transformation means can be commonly used by switching discriminant matrices to be read out in accordance with the input feature vector to be linearly transformed.
  • The size of a necessary computing unit can be reduced by using one linear transformation means in this manner.
  • As is obvious from equations (7), (8), and (9), the computation of an output feature vector z can be expressed as: Z = W 3 T ( y 1 y 2 ) = W 3 T ( W 1 T x 1 W 2 T x 2 ) = W 3 T ( W 1 T 0 0 W 2 T ) ( x 1 x 2 ) = W T ( x 1 x 2 ) ( 10 )
  • That is, linear transformations using the respective discriminant matrices can be integrated into linear transformation using one matrix. In stepwise computation processing, the number of times of product-sum computation is L×(M1+M2)+M1N1+M2N2. When matrices are integrated into one matrix, the number of times of product-sum computation is L×(N1+N2). If, for example, N1=N2=500, M1=M2=200, and L=100, 240,000 product-sum computations are required in stepwise computation processing. In the latter computation processing, 100,000 product-sum computations are required. The computation amount in batch computation processing as in the latter case is smaller than that in the former case, and hence high-speed computation can be realized. As is obvious from the mathematical expressions, when a final dimension count L is to be reduced, the batch computation method can reduce the computation amount and hence is effective.
  • (Second Embodiment)
  • According to the above case, when different kinds of features, e.g., directional features and density features, are to be joined together, discriminant analysis is repeatedly performed on a feature vector having undergone discriminant analysis for each feature. However, a plurality of elements corresponding to one feature may be divided into a plurality of vectors, discriminant analysis may be performed on each element set as an input feature, and the corresponding projected vector may be further subjected to discriminant analysis.
  • In the second embodiment, a facial image feature extraction device will be described.
  • As shown in FIG. 4, the facial image feature extraction device according to the second invention includes an image feature decomposition means 41 for decomposing the density feature of an input facial image, a linear transformation means 42 for projecting a feature vector in accordance with a discriminant matrix corresponding to the feature vector, and a discriminant matrix group storage means 43 for storing the respective discriminant matrices described above.
  • Techniques of extracting features from facial images include a method of positioning facial images at the eye position or the like and setting their density values as vector features, as disclosed in the above article by W. Zhao et al.
  • In the second invention as well, pixel density values of an image are handled as an input feature, i.e., an original feature. However, an image feature has a large image size, for example, 42×54 pixels=2352 dimensions with the central positions of the left and right eyes being normalized to the coordinates (14, 23) and (29, 23). With such large feature dimensions, it is difficult to perform high-precision feature extraction by directly performing linear discriminant analysis using limited learning samples. Therefore, a deterioration in feature which is caused when the principal component analysis or the like is applied is suppressed by decomposing image feature elements, performing discriminant analysis on the decomposed features, and obtaining discriminant matrices.
  • One of the methods of decomposing image features is to segment an image. For example, as shown in FIG. 5, an image is divided into nine parts each having a size of 14×18 pixels (=252 dimensions), local images having different sizes are set as feature vectors x: (i=1, 2, 3, . . . , 9), and discriminant analysis is performed on the respective partial images by using learning samples, thereby obtaining in advance discriminant matrices Wi corresponding to the respective feature vectors.
  • Note that letting regions have overlaps when an image is segmented makes it possible to reflect, in feature vectors, feature amounts based on the correlations between pixels in the boundary regions. Therefore, the respective regions may be sampled after being overlapped.
  • Since the number of feature dimensions is greatly reduced to 252 as compared with the original image, a basis matrix based on discriminant analysis can be calculated with high precision by sampling several images of each of several hundred individuals, i.e., a total of several thousand facial images. If the number of feature dimensions is as large as that of the original feature (2352 dimensions), in order to obtain similar performance with features based on discriminant analysis, it is expected that facial images of several thousand individuals must be sampled. In practice, however, it is difficult to collect such a large amount of image data, and hence this technique cannot be realized.
  • Assume that the feature in each local region is compressed to a 20-dimensional feature by a first-stage discriminant feature. In this case, the resultant output feature vectors become a feature vector of 9 regions×20 dimensions=180 dimensions. By further performing discriminant analysis on this feature vector, the number of dimensions can be efficiently reduced to about 50 dimensions. This second-stage discriminant matrix is also stored in the discriminant matrix group storage means 43, and discriminant analysis is performed again by the linear transformation means 42 upon receiving the 180-dimensional vector of the first-stage discriminant feature. Note that the first-stage discriminant matrix and second-stage discriminant matrix may be calculated in advance as indicated by equation (10). However, when 252 dimensions×9 regions are to be compressed to 20 dimensions×9 regions, and the 180 dimensions are to be transformed into 50 dimensions, the calculation in two stages will reduce the memory to be used and the computation amount to ½ or less and hence is efficient.
  • By applying discriminant analysis locally and stepwisely, a facial feature with high identification performance can be extracted. Assume that in character recognition, for example, z,1 and z,2 are to be identified. In this case, if the principal component analysis is performed on each entire character image to extract components having large eigenvalues, the feature “{grave over ()}” that helps to identify
    Figure US20050201595A1-20050915-P00001
    and
    Figure US20050201595A1-20050915-P00002
    tends to be lost (for this reason, similar character identification is sometimes performed by using a specific high-order feature instead of a feature of a portion with a large eigenvalue obtained by the principal component analysis). The effectiveness of segmenting an image into local regions and extracting discriminant features is similar to a phenomenon in similar character identification in character recognition. It can be thought that spatially limiting a feature that is easy to identify can ensure higher precision per unit dimension than performing discriminant analysis on principal components as a whole.
  • In addition, the image feature decomposition means 41 may sample images from an entire image and segment the sampled images instead of segmenting an image and forming a feature vector for each local region. When, for example, a primary feature is to be divided by nine into nine 252-dimensional vectors, sampling is performed from 3×3 regions, as shown in FIG. 6. That is, the sampled images become reduced images with slight positional differences. These reduced images are raster-scanned to be transformed into nine feature vectors. Such feature vectors are used as primary vectors to calculate discriminant components. These discriminant components may be integrated to perform discriminant analysis again.
  • (Third Embodiment)
  • Another embodiment of the present invention will be described in detail with reference to the accompanying drawings. FIG. 7 is a block diagram showing a facial image matching system using a facial metadata creating device according to the present invention.
  • The facial image matching system will be described in detail below.
  • As shown in FIG. 7, the facial image matching system according to the present invention includes a facial image input unit 71 which inputs facial images, a facial metadata creating unit 72 which creates facial metadata, a facial metadata storage unit 73 which stores extracted facial metadata, a facial similarity calculation unit 74 which calculates a facial similarity from facial metadata, a facial image database 75 which stores facial images, a control unit 76 which controls the input of images, the creation of metadata, the storage of metadata, and the calculation of facial similarities in accordance with an image registration request/retrieval request, and a display unit 77 of a display which displays facial images and other information.
  • The facial metadata creating unit 72 is comprised of a region cutting means 721 for cutting a facial region from an input facial image, and a facial image feature extraction means 722 which extracts a facial feature of the cut region. The facial metadata creating unit 72 creates metadata about a facial image by extracting facial feature vectors.
  • When a facial image is to be registered, a facial photo or the like is input upon adjustment of the size and position of the face by using the facial image input unit 71 such as a scanner or video camera. Alternatively, a human face may be directly input from a video camera or the like. In this case, it is preferable that the face position of the input image be detected by using a face detection technique like that disclosed in the above reference by Mohaddam and the size and the like of the facial image be automatically normalized.
  • The input facial image is registered in the facial image database 75 as needed. At the same time with facial image registration, the facial metadata creating unit 72 creates facial metadata and stores it in the facial metadata storage unit 73.
  • At the time of retrieval, the facial image input unit 71 inputs a facial image, and the facial metadata creating unit 72 creates facial metadata as in the case of registration. The created facial metadata is either registered in the facial metadata storage unit 73 or directly sent to the facial similarity calculation unit 74.
  • In retrieval operation, when it is to be checked whether or not data identical to a pre-input facial image exists in the database (facial identification), the similarity between the input facial image and each data registered in the facial metadata storage unit 73 is calculated. The control unit 76 selects a facial image from the facial image database 75 on the basis of the result exhibiting the highest similarity, and displays the facial image on the display unit 77 or the like. An operator then checks the coincidence between the faces in the retrieved image and the registered image.
  • When it is to be checked whether or not a facial image specified by an ID number or the like in advance coincides with a retrieved facial image (face verification), the facial similarity calculation unit 74 makes calculation to check whether or not the facial image specified by the ID number coincides with the retrieved image. If the calculated similarity is lower than a predetermined similarity, it is determined that the two images do not coincide with each other, and the result is displayed on the display unit 77. Assume that this system is used for room access management. In this case, room access management can be performed by causing the control unit 76 to send an opening/closing control signal to an automatic door so as to control the automatic door, instead of displaying a facial image.
  • The facial image matching system operates in the above manner. Such operation can be implemented on a computer system. For example, facial image matching can be realized by storing a metadata creation program for executing metadata creation to be described in detail next and a similarity calculation program in a memory and executing these programs using a program control processor.
  • In addition, these programs may be recorded on a computer-readable recording medium.
  • The operation of this facial image matching system, and more specifically, the operations of the facial metadata creating unit 72 and facial similarity calculation unit 74, will be described in detail next.
  • (1) Creation of Facial Metadata
  • The facial metadata creating unit 72 extracts a facial feature amount by using an image I(x, y) whose position and size have been normalized. In normalizing the position and size, the image is preferably normalized to set the eye positions to (16, 24) and (31, 24) and the size to 46×56 pixels. In the following case, the image has been normalized to this size.
  • The region cutting means 721 then cuts a plurality of preset local regions of the facial image. In the case of the above image, for example, one region is the entire normalized facial image (f(x, y)) and the other is a central region g(x, y) of 32×32 pixels centered on the face. This region may be cut such that the positions of the two eyes are set to (9, 12) and (24, 12).
  • The reason why a central region of a face is cut in the above manner is that a stable feature can be extracted by cutting a range free from the influences of a hair style and the like even if the hair style changes (for example, when facial verification is used in a home robot, verification can be done even if the hair style changes before and after bathing). If a hair style and the like do not change (for example, personal identification within scenes in a video clip), since an improvement in verification performance can be expected by performing verification using images including hair styles, a large facial image including a hair style and a small facial image of a central portion of the face are cut.
  • The facial image feature extraction means 722 then performs two-dimensional discrete Fourier transforms for the two cut regions f(x, y) to extract a facial image feature.
  • FIG. 8 shows the more detailed arrangement of the facial image feature extraction means 722. The facial image feature extraction means includes a Fourier transform means 81 for performing a discrete Fourier transform for a normalized cut image, a Fourier power calculation means 82 for calculating the power spectrum of a Fourier-transformed Fourier frequency component, a linear transformation means 83 for regarding a feature vector obtained by raster-scanning the real and imaginary components of the Fourier frequency component calculated by the Fourier transform means 81 as a one-dimensional feature vector and extracting a discriminant feature from the principal components of the feature vector, a basis matrix storage means 84 for storing a basis matrix for the transformation, a linear transformation means 85 for extracting a discriminant feature of principal components from a power spectrum in the same manner as described above, and a basis matrix storage means 86 for storing a basis matrix for the transformation. The facial image feature extraction means 722 further includes a linear transformation means 88 for normalizing each of the discriminant feature of the real and imaginary components of the Fourier feature and the discriminant feature of the power spectrum to a vector with a size of 1, and calculating a discriminant feature of a vector obtained by combining the two feature vectors, and a discriminant matrix storage means 89 for storing a discriminant matrix for the discriminant feature.
  • After a Fourier frequency feature is extracted with this arrangement, discriminant features of principal components are calculated for a feature vector including the real and imaginary parts of the Fourier frequency component as elements and a feature vector including a power spectrum as an element, and a discriminant feature is calculated again for a feature vector obtained by combining the above vectors, thereby calculating the feature amount of the face.
  • Each operation will be described in more detail below.
  • The Fourier transform means 81 performs a two-dimensional Fourier transform for the input image f(x, y) (x=0, 1, 2, . . . , M−1, y=0, 1, 2, . . . , N−1) to calculate a Fourier feature F(u, v) according to equation (11). This method is widely known and described in, for example, Rosenfeld et al., “Digital Picture Processing”, Kindai Kagaku Sha, pp. 20-26, and hence a description thereof will be omitted. F ( u , v ) = x = 0 M - 1 y = 0 N - 1 f ( x , y ) exp ( - 2 Π i ( xu M + yv N ) ) ( 11 )
  • The Fourier power calculation means calculates a Fourier power spectrum |F(u, v)| by obtaining the size of the Fourier feature F(u, v) according to equation (12).
    |F(u, v)|={square root}{right arrow over (|Re(F(u, v))|2 +|Im(Fu, v))|2)}  (12)
  • The two-dimensional Fourier spectra F(u, v) and |F(u, v)| obtained in this manner are obtained by transforming only the images of two-dimensional real parts, the obtained Fourier frequency components become symmetrical. For this reason, these spectrum images F(u, v) and |F(u, v) | have M×N components (u=0, 1, . . . , M−1; v=0, 1, . . . , N−1), and half of the components, i.e., M×N/2 components (u=0, 1, . . . , M−1; v=0, 1, . . . , N−1) and the remaining half components are substantially equivalent. Therefore, the subsequent processing may be performed by using half components as a feature vector. Obviously, computation can be simplified by omitting computation for components which are not used as elements of a feature vector in the Fourier transform means 81 and Fourier power calculation means 82.
  • The linear transformation means 83 then handles the feature amount extracted as a frequency feature as a vector. A partial space to be defined in advance is set by a basis vector (eigenvector) obtained by preparing a facial image set for learning and performing discriminant analysis on the principal components of a frequency feature vector in a corresponding cut region. Since this basis vector is obtained by a widely known method described in various references including the reference by W. Zhao, a description thereof will be omitted. The reason why discriminant analysis is not directly performed is that the number of dimensions of a feature vector obtained by a Fourier transform is too large to directly handle discriminant analysis. Although the already indicated problem in principal component discriminant analysis remains unsolved, this technique is one choice as a technique of extracting a first-stage feature vector. Alternatively, a basis matrix obtained by the method of repeating discriminant analysis may be used.
  • That is, a discriminant matrix Φ1 of principal components which is to be stored in the basis matrix storage means 84 can be obtained from learning samples in advance by performing discriminant analysis on the principal components of a one-dimensional feature vector x1 obtained by raster-scanning the real and imaginary components of a frequency feature. In this case, a Fourier feature need not always be handled as a complex number, and may be handled as a real number with an imaginary component being handled as another feature element.
  • Letting Ψ1 be a basis matrix for principal 25 components, and W1 be a discriminant matrix obtained by discriminant analysis on the vector of the principal components, the discriminant matrix Φ1 of the principal components can be expressed by
    Φ1 T =W 1 TΨ1 T  (13)
  • It suffices if the number of dimensions to be reduced by the principal component analysis is set to about 1/10 (about 200 dimensions) of the original feature Fourier feature. Thereafter, the number of dimensions is reduced to about 70 by this discriminant matrix. This basis matrix is calculated in advance from learning samples and is used as information to be stored in the basis matrix storage means 84.
  • In the case of the Fourier spectrum |F(u, v) | as well, a spectrum is expressed as a one-dimensional feature vector x2 by raster scanning, and basis matrix. Φ2 T2 TW2 T, which is obtained by discriminant analysis on the principal components of the feature vector, is obtained in advance by learning samples.
  • Calculating a principal component discriminant feature for each component of a Fourier feature in this manner makes it possible to obtain a discriminant feature y1 of the principal components of the feature vector x1 of the real and imaginary components of Fourier components, and a discriminant feature y2 of the principal components of the feature vector x2 of a power spectrum.
  • A normalization means 87 normalizes the size of each obtained feature vector to a unit vector with a size of 1. In this case, the vector length varies depending on the position of an origin for the measurement of a vector, and hence its reference position must also be determined in advance. In this case, it suffices if a reference point is set by using a mean vector mi obtained from a learning sample of a projected feature vector yi. By setting a mean vector as a reference point, feature vectors are distributed around the reference point. In the case of a Gaussian distribution, in particular, feature vectors are isotropically distributed. This makes it easy to limit a distribution region in a case wherein a feature vector is quantized in the end.
  • That is, a vector yi 0 obtained by normalizing the feature vector yi to a unit vector by using the mean vector mi can be expressed by y i 0 = y i - m i y i - m i ( 14 )
  • In this manner, the normalization means is provided to normalize the feature vector y1 associated with the real and imaginary numbers of Fourier power and the feature vector y2 associated with the power to unit vectors in advance. This makes it possible to normalize the sizes of two different kinds of feature amounts and stabilize the distribution features of feature vectors.
  • In addition, since the sizes of these vectors have already been normalized within a feature space necessary for discrimination in the process of dimension reduction, normalization robust against noise can be realized as compared with a case wherein normalization is performed in a feature space containing more deleted noise. This normalization can remove the influences of variation elements such as variation components which are proportional to the overall illumination intensity which is difficult to remove by simple linear transformation.
  • The feature vectors y1 0 and y2 0 normalized in this manner are combined into one feature vector y in the same manner as (equation 8), and the combined feature vector y is projected to a discriminant space by using the discriminant matrix W3 obtained by performing linear discriminant analysis, thereby obtaining an output feature vector z. The discriminant matrix W3 for this purpose is stored in the discriminant matrix storage means 89, and the linear transformation means 88 performs projection computation for this purpose to calculate, for example, a 24-dimensional feature vector z.
  • When the output feature vector z is to be quantized in five bits per element, the size of each element must be normalized in advance. For example, the size of each element is normalized in advance in accordance with the variance value of each element.
  • That is, a standard deviation value σi in a learning sample of each element zi of the feature vector z is obtained in advance, and normalization is performed to satisfy z0=16Zi/3σi. Assume that the size is five bits. In this case, it suffices if the size is quantized to a value falling within the range of −16 to 15.
  • In this case, normalization is the computation of multiplying each element by the reciprocal of the standard deviation. In consideration of a matrix Σ having σi as a diagonal element, a normalized vector z0 becomes z0=Σz. That is, since simple linear transformation is performed, Σ may be applied to the discriminant matrix W3 in advance as indicated by equation (15).
    W 3 0 T =ΣW 3 T  (15)
  • Performing normalization in this manner can perform range correction necessary for quantization. In addition, since normalization is performed by using the standard deviation, computation based on the Mahalanobis distance can be performed by only calculating a simple L2 norm in computing the norm of an inter-pattern distance at the time of collation, thereby reducing the computation amount at the time of collation.
  • As described above, the facial image feature extraction means 722 extracts a feature vector zf from the normalized image f(x, y) in this manner. With respect to an image g(x, y) obtained by cutting only a central portion of a face, a feature vector zg is extracted by the facial image feature extraction means 722 in the same manner as described above. The two feature vectors zf and zg are extracted by using the facial metadata creating unit as a facial feature amount z.
  • Note that a computer may be caused to execute the above facial metadata creation sequence by a computer program. In addition, this program may be recorded on a computer-readable recording medium.
  • (2) Facial Similarity Calculation
  • The operation of the facial similarity calculation unit 74 will be described next.
  • The facial similarity calculation unit 74 calculates a similarity d(z1, z2) by using K-dimensional feature vectors z1 and z2 obtained from two facial metadata.
  • For example, a similarity is calculated by the square distance of equation (16): d ( z 1 , z 2 ) = i = 1 K α i z 1 , i - z 2 , i 2 ( 16 )
    where αi is a weighting factor. If, for example, the reciprocal of the standard deviation of each feature dimension zi is used, calculation based on the Mahalanobis distance is performed. If feature vectors are normalized in advance by equation (15) or the like, since a basis matrix is normalized in advance with variance values, the Mahalanobis distance is set. Alternatively, a similarity may be calculated by the cosine of each feature vector to be compared which is expressed by equation (3). d ( Z 1 , Z 2 ) = Z 1 · Z 2 Z 1 Z 2 ( 17 )
  • Note that when a distance is used, a larger value indicates a lower similarity (the faces do not resemble each other), whereas when a cosine is used, a larger value indicates a higher similarity (the faces resemble each other).
  • According to the above description, one facial image is registered, and a retrieval is performed by using one facial image. When, however, a plurality of images are registered for the face of one individual and a retrieval is to be performed by using one facial image, a similarity may be calculated for each of a plurality of facial metadata on the registration side.
  • Likewise, when a plurality of images are to be registered for the face of one individual and a retrieval is to be performed by using a plurality of images, calculating a similarity by obtaining the mean or minimum value of similarity for each combination makes it possible to calculate a similarity for one facial data. This indicates that the matching system of the present invention can be applied to face verification in an image sequence by regarding the image sequence as a plurality of images.
  • The embodiments of the present invention have been described above by referring to the accompanying drawings as need. Obviously, however, the present invention can be implemented by a computer-executable program.
  • In addition, this program may be recorded on a computer-readable recording medium.
  • (Fourth Embodiment)
  • Another embodiment of the present invention will be described in detail with reference to the accompanying drawings. The present invention is directed to an improvement in the facial metadata creating unit 72 according to the third invention. According to the third invention, the discriminant features of the principal components of a feature vector having the real and imaginary parts of a Fourier frequency component obtained by performing a Fourier transform for an input facial image and a feature vector having a power spectrum as an element are calculated, and the discriminant feature of a feature vector obtained by combining the respective vectors is calculated again, thereby calculating the feature amount of the face. In this case, since a Fourier power spectrum reflects the overall feature amount of an input image, components of the input pixels which contain much noise (e.g., pixels around the mouth which tend to change in relative position) are reflected in the power spectrum in the same manner as the remaining pixels. As a consequence, even if an effective feature amount is selected by discriminant analysis, sufficient performance may not be obtained. In such a case, the input image is segmented into regions, and a Fourier transform is performed for each local region. Discriminant analysis is then performed by using a power spectrum for each local region as a feature amount. This can reduce the influences of the feature amount of a region which locally exhibits poor discriminant performance (large within-class variance) by discriminant analysis.
  • FIG. 9 is a view for explaining an embodiment and shows the flow of feature extraction processing. In this embodiment, for example, a 32×32 pixel region is segmented into four 16×16 pixel regions, 16 8×8 pixel regions, 64 4×4 pixel regions, 256 2×2 pixel regions, and 1024 1×1 pixel regions (which are substantially the same as the input image, and hence the input image can be used without segmentation) (S1001). A Fourier transform is performed in each segmented region (S1002). A power spectrum is then calculated (S1003). The above calculation is performed all the segmented regions (S1004). The size of a region is changed (S1005). The sizes of all the regions are changed (S1006). FIG. 10 summarizes this processing flow. 1024×5 dimensions=5120-dimensional feature amount of all the power spectra of the respective regions obtained in this manner is extracted.
  • Since this number of dimensions is too large in general when learning data is small in amount, the principal component analysis is performed in advance to obtain in advance the basis of the principal component analysis which reduces the number of dimensions. For example, an appropriate number of dimensions is about 300. Discriminant analysis is further performed on the feature vector of this dimension count to obtain a basis which reduces the number of dimensions and corresponds to a feature axis exhibiting good discriminant performance. A basis corresponding to the principal component analysis and discriminant analysis is calculated in advance (this basis will be referred to as a PCLDA projection basis Ψ).
  • A discriminant feature z can be obtained by projecting the 5120-dimensional feature by linear computation using the projection basis Ψ using this PCLDA basis. The feature amount of the face can be obtained by further performing quantization and the like for this feature.
  • Note that the 5120-dimensional feature amount can be reduced in the number of dimensions by considering the symmetry of the Fourier power spectrum and removing and not using high-frequency components. This can realize high-speed learning, reduce the amount of data required, and realize high-speed feature extraction. Therefore, the number of dimensions is preferably reduced as needed.
  • Segmenting a region into blocks and multiplexing Fourier spectra in this manner can sequentially obtain multiple expressions of feature amounts having translation universality and local feature amounts from a feature amount equivalent to an image feature (in the case of 1024 segmentations). A feature amount effective for identification is selected from the multiple, redundant feature expressions by discriminant analysis, thereby obtaining a compact feature amount which provides good identification performance.
  • A Fourier power spectrum is obtained by nonlinear computation for an image, which can calculate an effective feature amount which cannot be obtained by simply applying discriminant analysis based on linear computation to the image.
  • Although the application of linear discriminant analysis to principal components has been described above, second-stage feature extraction may be performed by using kernel discriminant analysis (discriminant analysis using a kernel technique called Kernel Fisher Discriminant Analysis, KFDA, Kernel Discriminant Analysis: KDA, or Generalized Discriminant Analysis: GDA).
  • For a detailed description of kernel discriminant analysis, see the reference by Q. Liu et al. (non-patent reference 3: “Kernel-based Optimized Feature Vectors Selection and Discriminant Analysis for Face Recognition”, Processing of IAPR International Conference on Pattern Recognition (ICPR), Vol. II, pp. 362-365, 2002) or the reference by G. Baudat (non-patent reference 4: Generalized Discriminant Analysis Using a Kernel Approach”, Neural Computation, Vol. 12, pp. 2385-2404, 2000).
  • By extracting a feature using kernel discriminant analysis, the effect of nonlinear feature extraction can be enhanced to allow extraction of an effective feature.
  • In this case, however, since a large feature vector of 5120 dimensions is to be processed, a large amount of memory and a large amount of learning data are required even for the principal component analysis. Referring to FIG. 11, in order to avoid such a problem, the principal component analysis/discriminant analysis is individually performed for each block. Thereafter, two-stage discriminant analysis (Linear Discriminant Analysis: LDA) is performed. This makes it possible to reduce the computation amount.
  • In this case, the principal component analysis and discriminant analysis are performed for each region by using a 1024-dimensional feature amount (512 dimensions if the number of dimensions is reduced to half in consideration of symmetry) to obtain a basis matrix Ψi (i=0, 1, 2, . . . , 5) in advance. Each feature vector is then normalized by using its mean value, and second-stage LDA projection is performed.
  • By performing processing for each block in this manner, the number of data and computer resources required for learning can be reduced. This makes it possible to shorten the time required for the optimization of learning.
  • Note that high-speed computation can be realized by omitting the vector normalization processing and calculating a basis matrix for PCLDA projection and a basis matrix for LDA projection in advance.
  • FIG. 12 is a view for explaining still another embodiment and shows the flow of feature extraction processing. In this embodiment, such region segmentation is performed in a plurality of stages (two stages in FIG. 12) to extract multiple power spectra in multiple resolutions as feature amounts for discriminant analysis in consideration of the translation universality of Fourier power spectra in local regions and the reliability of the local regions. Feature extraction is then performed using the optimal feature space obtained by discriminant analysis.
  • Assume that an input image f(x, y) has 32×32 pixels. In this case, as shown in FIG. 10, a power spectrum |F(u, v)| of the entire image, power spectra |F1 1(u, v)|, |F1 2(u, v)|, |F1 3(u, v)|, and |F1 4(u, v) | of four 16×16 pixel regions obtained by segmenting the entire image into four regions, and power spectra |F2 1(u, v)|, |F2 1(u, v), . . . , |F2 16(u, v)| of 16 8×8 pixel regions obtained by segmenting the entire image into 16 regions are extracted as feature vectors.
  • In consideration of the symmetry of the Fourier power spectrum of the real image, it suffices to extract ½ of them. Alternatively, in order to avoid an increase in the size of a feature vector for discriminant analysis, a feature vector may be formed without sampling any high-frequency components for discrimination. If, for example, a feature vector is formed by sampling ¼ of spectra which correspond to low-frequency components, the number of learning samples required can be reduced or the processing time required for learning and recognition can be shortened. If the number of learning data is small, discriminant analysis may be performed after the number of feature dimensions is reduced by the principal component analysis in advance.
  • Discriminant analysis is performed by using a feature vector x2 f extracted in this manner and a learning set prepared in advance to obtain a basis matrix Ψ2 f in advance. FIG. 9 shows an example of projection for the extraction of a discriminant feature from principal components (Principal Component Linear Discriminant Analysis; PCLDA). The feature vector x2 f is projected by using the basis matrix Ψ2 f, and the mean and size of the projected feature vector are normalized, thereby calculating a feature vector y2 f.
  • Likewise, the feature vector x2 f obtained by combining the real and imaginary components of a Fourier frequency is projected by linear computation processing using a basis matrix Ψ1 f to obtain a feature vector with a reduced number of dimensions, and the mean and size of the vector are normalized to calculate a feature vector y1 f. A feature vector obtained by combining these vectors is projected again by using a discriminant basis Ψ1 f to obtain a feature vector zf. This vector is quantized in, for example, five bits to extract a facial feature amount.
  • Assume that the input is a facial image normalized to a size of 44×56 pixels. In this case, the above processing is applied to the 32×32 pixels of a central portion to extract a facial feature amount. In addition, facial feature amounts are also extracted from multiple segmented regions of the 44×56 pixel region of the entire face, including the entire 44×56 pixel region, four 22×28 pixel regions, and 16 11×14 pixel pixels.
  • FIG. 13 shows another embodiment, in which PCLDA projection of a combination of a real component, an imaginary component, and a power spectrum is performed for each local region, or PCLDA projection of a feature obtained by combining a real component and imaginary component and PCLDA projection of a power spectrum are separately performed, and LDA projection is finally performed, as shown in FIG. 14.
  • (Fifth Embodiment)
  • Another embodiment of the present invention will be described in detail with reference to the accompanying drawings.
  • This embodiment is an embodiment of a facial feature description method using the present invention and descriptors of facial features. FIG. 15 shows a. description of a facial feature amount, as an example of a facial feature description, which uses a DDL representation syntax (Description Definition Language Representation Syntax) in ISO/IEC FDIS 15938-3, “Information technology Multimedia content description interface—Part 3: Visual”.
  • In this case, for a description of a facial feature named “AdvancedFaceRecognition”, elements named “FourierFeature” and “CentralFourieFeature” are provided. Each of “FourierFeature” and “CentralFourieFeature” is a 5-bit integer without a sign, representing that it can have 24-dimensional components to 63-dimensional components.
  • FIG. 16 shows a rule in a case wherein a binary representation syntax is used for data representation. According to this rule, the sizes of the array components of FourierFeature and CentralFourierFeature are stored in the field of 6-bit integers without signs in numOfFourierFeature and numOfCentralFourier, and each component of FourierFeature and CentralFourierFeature is stored in the form of a 5-bit integer without a sign.
  • Descriptors of such facial features using the present invention will be described in more detail.
    • numofFourierFeature
  • This field specifies the number of components of FourierFeature. The allowable range is from 24 to 63.
    • numOfCentralFourierFeature
  • This field specifies the number of components of CentralFourierFeature. The allowable range is from 24 to 63.
    • FourierFeature
  • This element represents a facial feature based on the cascaded LDA of the Fourier characteristics of a normalized face image. The normalized face image is obtained by scaling an original image into 56 lines with 46 luminance values in each line. The center positions of two eyes in the normalized face image shall be located on the 24th row and the 16th and 31st columns for the right and left eyes respectively.
  • The FourierFeature element is derived from two feature vectors; one is a Fourier Spectrum Vector x1 f, and the other is a Multi-block Fourier Amplitude Vector x2 f. FIG. 17 illustrates the extraction process of FourierFeature. Given a normalized face image, five steps should be performed to extract the element;
      • (1) Extraction of a Fourier Spectrum Vector x1 f,
      • (2) Extraction of a Multi-block Fourier Amplitude Vector x2 f,
      • (3) Projections of feature vectors using PCLDA basis matrices Ψ1 f, Ψ2 f, and their normalization to unit vectors y1 f, Ψ2 f,
      • (4) Projection of a Joint Fourier Vector y3 f of the unit vectors using an LDA basis matrix Ψ3 f,
      • (5) Quantization of the projected vector Zf.
        STEP-1) Extraction of Fourier Spectrum Vector Given a normalized face image f(x, y), the Fourier spectrum F(u, v) of f(x, y) is calculated by F ( u , v ) = x = 0 M - 1 y = 0 N - 1 f ( x , y ) exp ( - 2 π i ( xu M + yv N ) ) ( u = 0 , , M - 1 ; v = 0 , , N - 1 ) ( 18 )
        where, M=46 and N=56. A Fourier Spectrum Vector x1 f is defined as a set of scanned components of the Fourier spectrum. FIG. 18 shows the scanning method of the Fourier spectrum. The scanning shall be performed only on two rectangle regions, regions A and region B, in the Fourier domain. The scanning rule is concluded in FIG. 19. Here, SR(u, v) denotes the top-left coordinate of region R, and ER(u, v) does the bottom-right point of region R. Therefore, the Fourier Spectrum Vector x1 f is expressed by x 1 f = ( Re [ F ( 0 , 0 ) ] Re [ F ( 11 , 0 ) ] Re [ F ( 35 , 0 ) ] Re [ F ( 45 , 0 ) ] Re [ F ( 45 , 13 ) ] Im [ F ( 0 , 0 ) ] Im [ F ( 11 , 0 ) ] Im [ F ( 35 , 0 ) ] Im [ F ( 45 , 0 ) ] Im [ F ( 45 , 13 ) ] ) ( 19 )
        The dimension of x1 f is 644.
        STEP 2) Extraction of Multi-block Fourier Amplitude Vector
  • A multi-block Fourier Amplitude Vector is extracted from the Fourier amplitudes of partial images in the normalized face image. As the partial images, three types of images are used; (a) a holistic image, (b) quarter images, and (c) 1/16 images.
  • (a) holistic image
  • A holistic image f1 0(x, y) is obtained by clipping the normalized image f(x, y) in 44×56 image size removing boundary columns in both sides. It is given by
    f 1 0(x, y)=f(x+1, y) (x=0, 1, . . . , 43; y=0,1, . . . , 55)  (20)
    (b) quarter images
  • Quarter images are obtained by dividing the holistic image f1 0(x, y) equally into 4 blocks fk 1(x, y) (k=1, 2, 3, 4) given by
    f k 1(x, y)=f10(x+22s k 1 , y+28t k 1) (x=0, 1, . . . , 21; y=0, 1, . . . , 27)  (21)
    where sk 1=(k−1)%2, tk 1=(k−1)/2.
    (c) one-sixteenth images
  • One-sixteenth images are obtained by dividing f1 0(x, y) equally into 16 equal blocks fk 2(x, y) (k=1, 2, 3, . . . , 16) given by
    f k 2(x, y)=f 1 0(x+11s k 2 , y+14t k 2) (x=0, 1, . . . , 10; y=0, 1, . . . , 13)  (22)
    where sk 2=(k−1)%4, tk 2=(k−1)/4.
  • From these images, Fourier amplitudes |Fk j(u, v)| are calculated follows: F k j ( u , v ) = x = 0 M j - 1 y = 0 N j - 1 f k j ( x , y ) exp ( - 2 π i ( xu M j + yv N j ) ) , F k j ( u , v ) = Re [ F k j ( u , v ) ] 2 + Im [ F k j ( u , v ) ] 2 ( 23 )
    where Mj is the width of each partial image, that is, M0=44, M1=22, and M2=11. NJ denotes the height of each partial image, that is, N0=56, N1=28, and N2=14.
  • Multi-block Fourier Amplitude Vectors is obtained by scanning low frequency regions of each amplitude |Fk j(u, v)| of 1) the holistic image (k=1), 2) the quarter images (k=1, 2, 3, 4), and 3) the one-sixteenth images (k=1, 2, . . . , 16). The scan regions are defined in FIG. 19.
  • Therefore, the Multi-block Fourier Amplitude Vector x2 f is expressed as follows: x x f = ( F 1 0 ( 0 , 0 ) F 1 0 ( 43 , 13 ) F 1 1 ( 0 , 0 ) F 1 1 ( 21 , 6 ) F 2 1 ( 0 , 0 ) F 2 1 ( 21 , 6 ) F 3 1 ( 0 , 0 ) F 4 1 ( 21 , 6 ) F 1 2 ( 0 , 0 ) F 16 2 ( 10 , 2 ) ) ( 24 )
  • The dimension of x2 f is 856.
  • STEP 3) PCLDA Projection and Vector Normalization
  • The Fourier Spectrum Vector x1 f and Multi-block Fourier Amplitude Vector x2 f shall be respectively projected using the PCLDA basis matrices Ψ1 f and Ψ2 f, and normalized to unit vectors y1 f and y2 f. The normalized vector yk f (k=1, 2) is given by y k f = Ψ k f T x k f - m k f Ψ k f T x k f - m k f ( 25 )
    where, the PCLDA basis matrix Ψk f and the mean vector mk f are a basis matrix obtained by performing linear discriminant analysis on the principal components of xk f and a mean of projected vectors, respectively. The values of them are given by referring to a look-up-table calculated in advance. The dimensions of y1 f and y2 f are 70 and 80, respectively.
    STEP 4) LDA Projection of Joint Fourier Vector
  • The normalized vectors y1 f and Y2 f are combined to form a 150-dimensional Joint Fourier vector y3 f, and projected using the LDA basis matrix Ψ3 f. The projected vector zf is given by z f = Ψ k f T y 3 f = Ψ k f T ( y 1 f y 2 f ) ( 26 )
    STEP 5) Quantization
  • Each elements of zf is clipped in the range of 5bit-unsigned integer using the following equation: w i f = { 0 if z 1 f < - 16 31 , if z 1 f > - 15 floor ( z i t + 16 ) others ( 27 )
  • The quantized elements are stored as FourierFeature. FourierFeature[0] represents the first quantized element wo f, and FourierFeature[numOfFourierFeature-1] corresponds to the (numOfFourierFeature)th element wf numOfFourierFeature−1.
    • CentralFourierFeature
  • This element represents a facial feature based on the cascaded LDA of the Fourier characteristics of the central part in the normalized face image. CentralFourierFeature is extracted in the similar way as FourierFeature.
  • The central portion g(x, y) is obtained by clipping the image f(x, y) into 32×32 image starting at (7, 12) as follows:
    g(x, y)=f(x+7, y+12) (x=0, 1, . . . , 31; y=0, 1, . . . , 31)  (28)
    STEP 1) Extraction of Central Fourier Spectrum Vector
  • The Fourier spectrum G(u, v) of g(x, y) is calculated by G ( u , v ) = x = 0 M - 1 y = 0 N - 1 g ( x , y ) exp ( - 2 π i ( xu M + yv N ) ) ( u = 0 , , M - 1 ; v = 0 , , N - 1 ) ( 29 )
    where, M=32 and N=32. A 256-dimensional Central Fourier Spectrum Vector x1 g is produced by scanning the Fourier spectrum G(u, v) as defined in FIG. 20. STEP 2) Extraction of Multi-block Central Fourier Amplitude Vector
  • A Central Multi-block Fourier Amplitude Vectors x2 g is extracted from the Fourier amplitudes of (a) the central part g1 0(x, y), (b) quarter images gk 1(x, y) (k=1, 2, 3, 4), and (c) 1/16 images gk 2(x, y) (k=1, 2, 3, . . . , 16).
  • (a) central part
    g 1 0(x, y)=g(x, y) (x=0, 1, . . . , 31; y=0, 1, . . . ,31)  (31)
    (b) quarter images
    g k 1(x, y)=g(x+16s k 1 , y+16t k 1) (x=0, 1, . . . , 15; y=0, 1, . . . , 15)  (31)
    where Sk 1=(k−1)%2, tk 1(k−1)/2
    (c) one-sixteenth images
    g k 2(x, y)=g 1 0(x+8s k 2 , y+8t k 2) (x=0, 1, . . . , 7; y=0, 1, . . . , 7)  (32)
    where sk 2=(k−1)%4 and tk 2=(k−1)/4.
  • A Fourier amplitude |Gk j(u, v)| of each image is calculated as follows: G k j ( u , v ) = x = 0 M j - 1 y = 0 N j - 1 g k j ( x , y ) exp ( - 2 π i ( xu M j + yv N j ) ) G k j ( u , v ) = Re [ G k j ( u , v ) ] 2 + Im [ G k j ( u , v ) ] 2 ( 33 )
    where M0=32, M1=16, M2=8, N0=32, N1=16, and N2=8. A multi-block Central Fourier Amplitude Vector x2 g is obtained by scanning each amplitude |Gk j(u, v)| as defined in FIG. 20.
  • The processing in STEP 3-5) are the same as those FourierFeature, for example, the Joint Central Fourier Vector y3g consists of the normalized vectors y1 g and y2 g. The basis matrices Ψ1 g, Ψ2 g, and Ψ3 g and the mean vectors m1 g and m2 g for CentralFourierFeature are calculated in advance, and prepared in the form of a look-up table.
  • The size of CentralFourierFeature is indicated by numOfCentralFourierFeature.
  • Facial feature description data obtained in this manner is compact in description length but exhibits high recognition performance, and hence is an expression efficient for the storage and transmission of data.
  • Note that the present invention may be implemented by a computer-executable program. In the case of the fifth embodiment, the present invention can be implemented by describing the functions indicated by steps 1 to 5 in FIG. 17 in a computer-readable program and making the program function on the computer.
  • In addition, this program may be recorded on a computer-readable recording medium.
  • When the example shown in FIG. 17 is to be implemented as a device, all or some of the functions written in the block diagram of FIG. 21 may be implemented. More specifically, all or some of a normalized face image output means 211, Fourier spectrum vector extraction means 212, multiblock Fourier amplitude vector extraction means 213 and PCLDA projection/vector normalization means 214 may be implemented.
  • According to each embodiment described above, a feature vector effective for discrimination by discriminant analysis is extracted from an input pattern feature vector for each element vector, and feature extraction is performed again for the obtained feature vector by using a discriminant matrix by discriminant analysis. This makes it possible to suppress a reduction in feature amount effective for discrimination when feature dimension reduction is performed, and to transform a feature vector for efficient feature extraction.
  • Each embodiment described above is effective for a case wherein the number of learning samples required for discriminant analysis is limited in spite of a large pattern feature amount. That is, the number of feature dimensions can be reduced, while a loss of features effective for identification is suppressed, without necessarily using the principal component analysis.
  • As has been described above, the image feature extraction method, the image feature extraction device, and the recording medium storing the corresponding program in the field of pattern recognition according to the present invention are suitable for the use in a feature vector transformation technique for compressing feature dimension by extracting feature vectors effective for recognition from input feature vectors.

Claims (34)

1-22. (canceled)
23. A pattern feature extraction method comprising the steps of
extracting a plurality of input vectors from an input pattern,
projecting the input vectors to obtain projection vectors by using basis matrices respectively corresponding to the input vectors, and
projecting, using a discriminant matrix corresponding to a joint vector, the joint vector obtained by combining a plurality of projection vectors, thereby extracting a feature of the input pattern.
24. A pattern feature extraction method comprising the steps of extracting a plurality of input vectors from an input pattern, projecting the input vectors to obtain projection vectors by using basis matrices respectively corresponding to the input vectors, normalizing the projection vectors to obtain normalized vectors, and projecting, using a discriminant matrix corresponding to a joint vector, the joint vector obtained by combining a plurality of normalized vectors, thereby extracting a feature of the input pattern.
25. A pattern feature extraction method including the steps of extracting a plurality of input vectors from an input pattern and projecting the input vectors to obtain projection vectors, thereby extracting a feature of the input pattern, characterized in that in the step of projecting the input vectors to obtain projection vectors, the input vectors are projected using a transformation matrix specified by basis matrices respectively corresponding to the input vectors and by a discriminant matrix corresponding to a joint vector obtained by combining projection vectors respectively obtained by projecting the input vectors using the basis matrices.
26. A pattern feature extraction method according to claim 23, characterized in that the basis matrices corresponding to the input vectors serve as discriminant matrices for the input vectors.
27. A pattern feature extraction method according to claim 24, characterized in that the basis matrices corresponding to the input vectors serve as discriminant matrices for the input vectors.
28. A pattern feature extraction method according to claim 25, characterized in that the basis matrices corresponding to the input vectors serve as discriminant matrices for the input vectors.
29. A pattern feature extraction method according to claim 23, characterized in that the basis matrices corresponding to the input vectors are basis matrices specified by transformation matrices for extracting principal component vectors of the input vectors and by discriminant matrices for the principal component vectors.
30. A pattern feature extraction method according to claim 24, characterized in that the basis matrices corresponding to the input vectors are basis matrices specified by transformation matrices for extracting principal component vectors of the input vectors and by discriminant matrices for the principal component vectors.
31. A pattern feature extraction method according to claim 25, characterized in that the basis matrices corresponding to the input vectors are basis matrices specified by transformation matrices for extracting principal component vectors of the input vectors and by discriminant matrices for the principal component vectors.
32. A pattern feature extraction method according to claim 23, characterized in that the step of extracting input vectors comprises the step of extracting vectors whose elements are pixel values obtained from sample points in each sample point set for each of a plurality of predetermined sample point sets in an image serving as an input pattern.
33. A pattern feature extraction method according to claim 24, characterized in that the step of extracting input vectors comprises the step of extracting vectors whose elements are pixel values obtained from sample points in each sample point set for each of a plurality of predetermined sample point sets in an image serving as an input pattern.
34. A pattern feature extraction method according to claim 25, characterized in that the step of extracting input vectors comprises the step of extracting vectors whose elements are pixel values obtained from sample points in each sample point set for each of a plurality of predetermined sample point sets in an image serving as an input pattern.
35. A pattern feature extraction method according to claim 32, characterized in that the sample point set comprises a set having as sample points pixels in partial images obtained from a local region of the image, thereby extracting a feature of the image.
36. A pattern feature extraction method according to claim 32, characterized in that the sample point set comprises a set having as sample points pixels in each reduced image obtained from the image, thereby extracting a feature of the image.
37. A pattern feature extraction method according to claim 23, characterized in that the step of extracting input vectors comprises the step of extracting as input vectors feature amounts calculated from each local region for each of a plurality of local regions of the image serving as the input pattern.
38. A pattern feature extraction method according to claim 24, characterized in that the step of extracting input vectors comprises the step of extracting as input vectors feature amounts calculated from each local region for each of a plurality of local regions of the image serving as the input pattern.
39. A pattern feature extraction method according to claim 25, characterized in that the step of extracting input vectors comprises the step of extracting as input vectors feature amounts calculated from each local region for each of a plurality of local regions of the image serving as the input pattern.
40. A pattern feature extraction method according to claim 23, characterized in that the step of extracting input vectors comprises the steps of Fourier-transforming the image serving as the input pattern, extracting Fourier spectrum vectors as the input vectors from a Fourier spectrum of the image, and extracting Fourier amplitude vectors as the input vectors from a Fourier amplitude spectrum of the image, thereby extracting a feature of the image.
41. A pattern feature extraction method according to claim 24, characterized in that the step of extracting input vectors comprises the steps of Fourier-transforming the image serving as the input pattern, extracting Fourier spectrum vectors as the input vectors from a Fourier spectrum of the image, and extracting Fourier amplitude vectors as the input vectors from a Fourier amplitude spectrum of the image, thereby extracting a feature of the image.
42. A pattern feature extraction method according to claim 25, characterized in that the step of extracting input vectors comprises the steps of Fourier-transforming the image serving as the input pattern, extracting Fourier spectrum vectors as the input vectors from a Fourier spectrum of the image, and extracting Fourier amplitude vectors as the input vectors from a Fourier amplitude spectrum of the image, thereby extracting a feature of the image.
43. A pattern feature extraction method according to claim 40, characterized in that a plurality of partial images or reduced images are extracted from the image, and Fourier spectrum vectors or Fourier amplitude vectors of the partial images or reduced images are extracted to extract a feature of the image.
44. A pattern feature extraction apparatus comprising vector extraction means for extracting a plurality of input vectors from an input pattern, basis matrix storage means for storing basis matrices respectively corresponding to the input vectors, linear transformation means for projecting the input vectors to obtain projection vectors using the basis matrices stored in said basis matrix storage means, discriminant matrix storage means for storing a discriminant matrix corresponding to a joint vector obtained by combining a plurality of projection vectors obtained by said linear transformation means, and second linear transformation means for projecting, using the discriminant matrix stored in said discriminant matrix storage means, the joint vector obtained by combining the plurality of projection vectors, thereby extracting a feature of the input pattern.
45. A pattern feature extraction apparatus comprising vector extraction means for extracting a plurality of input vectors from an input pattern, basis matrix storage means for storing basis matrices respectively corresponding to the input vectors, linear transformation means for projecting the input vectors to obtain projection vectors using the basis matrices stored in said basis matrix storage means, normalization means for normalizing the projection vectors to obtain normalized vectors, discriminant matrix storage means for storing a discriminant matrix corresponding to a joint vector obtained by combining a plurality of normalized vectors obtained by said normalization means, and second linear transformation means for projecting, using the discriminant matrix stored in said discriminant matrix storage means, the joint vector obtained by combining the plurality of normalized vectors, thereby extracting a feature of the input pattern.
46. A pattern feature extraction apparatus comprising vector extraction means for extracting a plurality of input vectors from an input pattern, basis matrix storage means for storing basis matrices respectively corresponding to the input vectors, and linear transformation means for projecting the input vectors using the transformation matrices stored in said transformation matrix storage means, thereby extracting a feature of the input pattern, characterized in that the transformation matrices stored in said transformation matrix storage means comprise transformation matrices specified by basis matrices respectively corresponding to the input vectors and the discriminant matrix corresponding to the joint vector obtained by combining the plurality of projection vectors obtained by projecting the input vectors using the basis matrices.
47. A computer-readable storage medium which stores a program for allowing a computer to execute pattern feature extraction for extracting a feature of an input pattern, characterized in that the program comprises a program which executes a function of extracting a plurality of input vectors from an input pattern, a function of projecting the input vectors to obtain projection vectors using basis matrices respectively corresponding to the input vectors, and a function of projecting, using a discriminant matrix corresponding to a joint vector, the joint vector obtained by combining the projection vectors.
48. A computer-readable storage medium which stores a program for allowing a computer to execute pattern feature extraction for extracting a feature of an input pattern, characterized in that the program comprises a program which executes a function of extracting a plurality of input vectors from an input pattern, a function of projecting the input vectors to obtain projection vectors using basis matrices respectively corresponding to the input vectors, a function of normalizing the projection vectors to obtain normalized vectors, and a function of projecting, using a discriminant matrix corresponding to a joint vector, the joint vector obtained by combining the normalized vectors.
49. A computer-readable storage medium which stores a program for allowing a computer to execute pattern feature extraction for extracting a feature of an input pattern by executing a function of extracting a plurality of input vectors from the input pattern and a function of projecting the input vectors, characterized in that the function of projecting the input vectors comprises a function of projecting the input vectors by using a transformation matrix specified by basis matrices respectively corresponding to the input vectors and by a discriminant matrix corresponding to a joint vector obtained by combining the plurality of projection vectors obtained by projecting the input vectors using the basis matrices.
50. A pattern feature extraction method characterized by comprising the steps of segmenting an input image using different segmentation numbers to obtain a plurality of block images and the step of extracting Fourier amplitudes of the block images, thereby extracting a feature amount of the input image.
51. A pattern feature extraction method according to claim 50, characterized by comprising the steps of scanning the Fourier amplitudes to extract multiblock Fourier amplitude vectors, and projecting the multiblock Fourier amplitude vectors using basis matrices to obtain projection vectors.
52. A pattern feature extraction method according to claim 51, characterized by further comprising the step of normalizing the projection vectors to obtain normalized vectors.
53. A pattern feature extraction method according to claim 51, characterized in that the basis matrices comprise basis matrices specified by transformation matrices for extracting principal component vectors of the multiblock Fourier amplitude vectors and by discriminant matrices corresponding to the principal component vectors.
54. A pattern feature extraction method according to claim 50, characterized in that in the step of obtaining the plurality of block images, at least one entire image having the entire input image as one block image, four block images obtained by segmenting the entire input image into four blocks, and 16 block images obtained by segmenting the input image into 16 blocks are obtained.
55. A pattern feature extraction method characterized by comprising the steps of
obtaining a Fourier spectrum vector by calculating a Fourier spectrum for an input normalized image by using a predetermined calculation expression,
extracting a multiblock Fourier amplitude vector from a Fourier amplitude of a partial image of the normalized image,
performing feature vector projection of the Fourier spectrum vector and the multiblock intensity vector by using a basis matrix, thereby obtaining respective normalized vectors,
combining the normalized vectors to obtain a coupled Fourier vector and using a second basis matrix to transform the coupled value into a projection vector, and
extracting a Fourier feature by quantizing the projection vector.
US10/512,194 2002-07-16 2003-07-04 Pattern characteristic extraction method and device for the same Abandoned US20050201595A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/133,084 US8116571B2 (en) 2002-07-16 2008-06-04 Pattern feature extraction via fourier amplitudes of a block image

Applications Claiming Priority (7)

Application Number Priority Date Filing Date Title
JP2002-207022 2002-07-16
JP2002207022 2002-07-16
JP2002300594 2002-10-15
JP2002-300594 2002-10-15
JP2003-68916 2003-03-13
JP2003068916A JP4292837B2 (en) 2002-07-16 2003-03-13 Pattern feature extraction method and apparatus
PCT/JP2003/008556 WO2004008391A1 (en) 2002-07-16 2003-07-04 Pattern characteristic extraction method and device for the same

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US12/133,084 Division US8116571B2 (en) 2002-07-16 2008-06-04 Pattern feature extraction via fourier amplitudes of a block image

Publications (1)

Publication Number Publication Date
US20050201595A1 true US20050201595A1 (en) 2005-09-15

Family

ID=30118927

Family Applications (2)

Application Number Title Priority Date Filing Date
US10/512,194 Abandoned US20050201595A1 (en) 2002-07-16 2003-07-04 Pattern characteristic extraction method and device for the same
US12/133,084 Expired - Fee Related US8116571B2 (en) 2002-07-16 2008-06-04 Pattern feature extraction via fourier amplitudes of a block image

Family Applications After (1)

Application Number Title Priority Date Filing Date
US12/133,084 Expired - Fee Related US8116571B2 (en) 2002-07-16 2008-06-04 Pattern feature extraction via fourier amplitudes of a block image

Country Status (7)

Country Link
US (2) US20050201595A1 (en)
EP (2) EP1522962B1 (en)
JP (1) JP4292837B2 (en)
KR (1) KR100578614B1 (en)
AT (1) ATE511680T1 (en)
HK (1) HK1090157A1 (en)
WO (1) WO2004008391A1 (en)

Cited By (46)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040170305A1 (en) * 2002-10-15 2004-09-02 Samsung Electronics Co., Ltd. Method and apparatus for extracting feature vector used for face recognition and retrieval
US20050180610A1 (en) * 2004-02-18 2005-08-18 Fuji Xerox Co., Ltd. Image processing apparatus
US20060034517A1 (en) * 2004-05-17 2006-02-16 Mitsubishi Denki Kabushiki Kaisha Method and apparatus for face description and recognition
US20060148323A1 (en) * 2004-12-03 2006-07-06 Ulrich Canzler Facial feature analysis system
US20060153430A1 (en) * 2004-12-03 2006-07-13 Ulrich Canzler Facial feature analysis system for users with physical disabilities
US20070046662A1 (en) * 2005-08-23 2007-03-01 Konica Minolta Holdings, Inc. Authentication apparatus and authentication method
US20070088548A1 (en) * 2005-10-19 2007-04-19 Kabushiki Kaisha Toshiba Device, method, and computer program product for determining speech/non-speech
US20080037841A1 (en) * 2006-08-02 2008-02-14 Sony Corporation Image-capturing apparatus and method, expression evaluation apparatus, and program
US20080077400A1 (en) * 2006-09-27 2008-03-27 Kabushiki Kaisha Toshiba Speech-duration detector and computer program product therefor
US20080304750A1 (en) * 2002-07-16 2008-12-11 Nec Corporation Pattern feature extraction method and device for the same
US20090087036A1 (en) * 2005-05-31 2009-04-02 Nec Corporation Pattern Matching Method, Pattern Matching System, and Pattern Matching Program
US20090123077A1 (en) * 2005-10-24 2009-05-14 Nec Corporation Coefficient determining method, feature extracting method, system, and program, and pattern checking method, system, and program
US20090136121A1 (en) * 2007-11-14 2009-05-28 Ryo Nakagaki Defect review method and apparatus
US20090180671A1 (en) * 2007-10-19 2009-07-16 Samsung Electronics Co., Ltd. Multi-view face recognition method and system
US20090254341A1 (en) * 2008-04-03 2009-10-08 Kabushiki Kaisha Toshiba Apparatus, method, and computer program product for judging speech/non-speech
US20100076723A1 (en) * 2008-09-23 2010-03-25 Microsoft Corporation Tensor linear laplacian discrimination for feature extraction
US7724960B1 (en) 2006-09-08 2010-05-25 University Of Central Florida Research Foundation Inc. Recognition and classification based on principal component analysis in the transform domain
US7734097B1 (en) * 2006-08-01 2010-06-08 Mitsubishi Electric Research Laboratories, Inc. Detecting objects in images with covariance matrices
US20100150452A1 (en) * 2007-04-23 2010-06-17 Nec Corporation Matching method for two-dimensional pattern, feature extracting method, appartus used for the methods, and programs
US20100166266A1 (en) * 2008-12-30 2010-07-01 Michael Jeffrey Jones Method for Identifying Faces in Images with Improved Accuracy Using Compressed Feature Vectors
US20110058733A1 (en) * 2008-04-30 2011-03-10 Osaka Prefecture University Public Corporation Method of compiling three-dimensional object identifying image database, processing apparatus and processing program
US20110135167A1 (en) * 2008-07-10 2011-06-09 Nec Corporation Personal authentication system and personal authentication method
US20110194776A1 (en) * 2008-10-10 2011-08-11 Nec Corporation Collating device, collating method, and program
US7999846B2 (en) * 2005-12-06 2011-08-16 Hitachi Kokusai Electric Inc. Image processing apparatus, image processing system, and recording medium for programs therefor
US20120023134A1 (en) * 2009-03-27 2012-01-26 Nec Corporation Pattern matching device, pattern matching method, and pattern matching program
US20120027309A1 (en) * 2009-04-14 2012-02-02 Nec Corporation Image signature extraction device
US8218880B2 (en) 2008-05-29 2012-07-10 Microsoft Corporation Linear laplacian discrimination for feature extraction
US20120237134A1 (en) * 2008-04-07 2012-09-20 Microsoft Corporation Image descriptor quantization
US20130142399A1 (en) * 2011-12-04 2013-06-06 King Saud University Face recognition using multilayered discriminant analysis
US20130243271A1 (en) * 2012-03-14 2013-09-19 Kabushiki Kaisha Toshiba Collation apparatus, collation method, and computer program product
WO2014001610A1 (en) * 2012-06-25 2014-01-03 Nokia Corporation Method, apparatus and computer program product for human-face features extraction
US8805013B2 (en) 2011-06-16 2014-08-12 Shinkawa Ltd. Pattern position detecting method
US20150030216A1 (en) * 2013-07-26 2015-01-29 Fujitsu Limited Image processing apparatus and method
US20150234861A1 (en) * 2006-02-01 2015-08-20 Sony Corporation System, apparatus, method, program and recording medium for processing image
US20160042223A1 (en) * 2014-08-08 2016-02-11 Samsung Electronics Co., Ltd. Method and apparatus for facial recognition
US9317909B2 (en) * 2013-02-27 2016-04-19 Samsung Electronics Co., Ltd. Image subsystem including image feature detection hardware component and image processing system including the same
US9576222B2 (en) 2013-04-09 2017-02-21 Hitachi Kokusai Electric Inc. Image retrieval apparatus, image retrieval method, and recording medium
US20170069112A1 (en) * 2014-05-22 2017-03-09 Huawei Technologies Co., Ltd. Image Characteristic Estimation Method and Device
US9652664B1 (en) * 2014-12-30 2017-05-16 Morphotrust Usa, Llc Facial recognition using fractal features
US9704065B2 (en) 2014-03-20 2017-07-11 Kabushiki Kaisha Toshiba Dimension reduction apparatus, dimension reduction method, and computer program product
US10366279B2 (en) 2017-08-29 2019-07-30 Bank Of America Corporation System for execution of multiple events based on image data extraction and evaluation
US20200226772A1 (en) * 2019-01-13 2020-07-16 Hunan Agricultural Information And Engineering Institute Anti-counterfeiting method based on feature of surface texture image of products
CN112115881A (en) * 2020-09-21 2020-12-22 国网重庆市电力公司长寿供电分公司 Image feature extraction method based on robust identification feature learning
US11222199B2 (en) * 2018-12-05 2022-01-11 International Business Machines Corporation Automatically suggesting behavioral adjustments during video conferences
US11244169B2 (en) 2020-06-15 2022-02-08 Bank Of America Corporation System for executing multiple events based on video data extraction and evaluation
CN116347104A (en) * 2023-05-22 2023-06-27 宁波康达凯能医疗科技有限公司 Method, device and storage medium for encoding intra-frame image based on efficient discriminant analysis

Families Citing this family (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003081574A1 (en) * 2002-03-27 2003-10-02 Nokia Corporation Pattern recognition
JP4770932B2 (en) * 2002-07-16 2011-09-14 日本電気株式会社 Pattern feature extraction method and apparatus
KR100723406B1 (en) * 2005-06-20 2007-05-30 삼성전자주식회사 Face image verification method and apparatus using LBPLocal Binary Pattern discriminant method
KR100682987B1 (en) * 2005-12-08 2007-02-15 한국전자통신연구원 Apparatus and method for three-dimensional motion recognition using linear discriminant analysis
WO2007066953A1 (en) * 2005-12-08 2007-06-14 Electronics And Telecommunications Research Institute Apparatus for recognizing three-dimensional motion using linear discriminant analysis
KR100723417B1 (en) * 2005-12-23 2007-05-30 삼성전자주식회사 Apparatuses and methods for recognizing face, and apparatus and method for extracting face from multiple face images
WO2007088926A1 (en) 2006-02-01 2007-08-09 Nec Corporation Image processing, image characteristic extracting and image checking devices, methods and programs, and image checking system
JP4631809B2 (en) * 2006-06-09 2011-02-16 富士ゼロックス株式会社 Defect classification system, image forming apparatus, and defect classification program
KR100771244B1 (en) * 2006-06-12 2007-10-29 삼성전자주식회사 Method and apparatus for processing video data
JP5151522B2 (en) * 2007-02-09 2013-02-27 株式会社デンソー Arithmetic apparatus and program
JP4925120B2 (en) * 2007-07-02 2012-04-25 独立行政法人産業技術総合研究所 Object recognition apparatus and object recognition method
US8068551B2 (en) * 2007-09-06 2011-11-29 Sharp Laboratories Of America, Inc. Systems and methods for designing a reference signal to be transmitted in a multiplexed cellular system
JP4985264B2 (en) * 2007-09-25 2012-07-25 株式会社豊田中央研究所 Object identification device
JP4992046B2 (en) * 2007-12-07 2012-08-08 Jfeスチール株式会社 Method and apparatus for abnormal prediction of steel process
KR101247891B1 (en) * 2008-04-28 2013-03-26 고리츠다이가쿠호징 오사카후리츠다이가쿠 Method for creating image database for object recognition, processing device, and processing program
US8170341B2 (en) 2009-01-23 2012-05-01 Nec Corporation Image signature extraction device
JP5726850B2 (en) * 2009-03-20 2015-06-03 ザ ジョンズ ホプキンス ユニバーシティ Method and system for quantifying technical skills
JP5548508B2 (en) * 2010-04-21 2014-07-16 日本放送協会 Image classification apparatus and program
JP2012027687A (en) * 2010-07-23 2012-02-09 Casio Comput Co Ltd Image processing apparatus and program
US8838147B2 (en) * 2011-08-31 2014-09-16 Nokia Corporation Method and apparatus for determining environmental context utilizing features obtained by multiple radio receivers
US8593452B2 (en) * 2011-12-20 2013-11-26 Apple Inc. Face feature vector construction
JP5701787B2 (en) * 2012-02-17 2015-04-15 日本電信電話株式会社 Data classification prediction apparatus, method, and program
KR101351132B1 (en) * 2012-12-27 2014-01-14 조선대학교산학협력단 Image segmentation apparatus and method based on anisotropic wavelet transform
CN103473544A (en) * 2013-04-28 2013-12-25 南京理工大学 Robust human body feature rapid extraction method
JP6167733B2 (en) * 2013-07-30 2017-07-26 富士通株式会社 Biometric feature vector extraction device, biometric feature vector extraction method, and biometric feature vector extraction program
CN108171216A (en) * 2018-01-22 2018-06-15 扬州大学 A kind of sparse noise matrix decomposition method for recognition of face
US11521460B2 (en) 2018-07-25 2022-12-06 Konami Gaming, Inc. Casino management system with a patron facial recognition system and methods of operating same
AU2019208182B2 (en) 2018-07-25 2021-04-08 Konami Gaming, Inc. Casino management system with a patron facial recognition system and methods of operating same
WO2022244250A1 (en) * 2021-05-21 2022-11-24 日本電気株式会社 Information generation device, determination device, information generation method, determination method, and storage medium
CN113743485A (en) * 2021-08-23 2021-12-03 江苏大学 Data dimension reduction method based on Fourier domain principal component analysis

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4259661A (en) * 1978-09-01 1981-03-31 Burroughs Corporation Apparatus and method for recognizing a pattern
US6021219A (en) * 1997-01-07 2000-02-01 Lucent Technologies Inc. Methods and apparatus for distinguishing among several visual patterns

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH04292837A (en) 1991-03-20 1992-10-16 Fujitsu Ltd Plasma display panel
US20020009208A1 (en) * 1995-08-09 2002-01-24 Adnan Alattar Authentication of physical and electronic media objects using digital watermarks
WO1995020296A1 (en) 1994-01-19 1995-07-27 University Of Strathclyde Fractal data compression
US5684890A (en) * 1994-02-28 1997-11-04 Nec Corporation Three-dimensional reference image segmenting method and apparatus
JP3022946B2 (en) 1994-04-20 2000-03-21 株式会社エヌ・ティ・ティ・データ Calculation method of feature transformation matrix for dimension reduction of feature vector for pattern recognition
WO1997005566A1 (en) * 1995-07-26 1997-02-13 Identification Technologies International, Inc. System for object verification and identification
JP3305551B2 (en) 1995-11-30 2002-07-22 日本電信電話株式会社 Specific symmetric object judgment method
JP2739856B2 (en) * 1995-12-18 1998-04-15 日本電気株式会社 Finger and palm print image processing device
JPH11196262A (en) * 1997-11-07 1999-07-21 Matsushita Electric Ind Co Ltd Digital information imbedding extracting device/method, and medium recording program to execute the method
JP2000222572A (en) 1999-01-28 2000-08-11 Toshiba Tec Corp Sex discrimination method
JP3809305B2 (en) * 1999-07-27 2006-08-16 キヤノン株式会社 Image search apparatus, image search method, and computer-readable storage medium
GB0013016D0 (en) 2000-05-26 2000-07-19 Univ Surrey Personal identity authentication process and system
JP4292837B2 (en) 2002-07-16 2009-07-08 日本電気株式会社 Pattern feature extraction method and apparatus

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4259661A (en) * 1978-09-01 1981-03-31 Burroughs Corporation Apparatus and method for recognizing a pattern
US6021219A (en) * 1997-01-07 2000-02-01 Lucent Technologies Inc. Methods and apparatus for distinguishing among several visual patterns

Cited By (94)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8116571B2 (en) 2002-07-16 2012-02-14 Nec Corporation Pattern feature extraction via fourier amplitudes of a block image
US20080304750A1 (en) * 2002-07-16 2008-12-11 Nec Corporation Pattern feature extraction method and device for the same
US20040170305A1 (en) * 2002-10-15 2004-09-02 Samsung Electronics Co., Ltd. Method and apparatus for extracting feature vector used for face recognition and retrieval
US7391889B2 (en) * 2002-10-15 2008-06-24 Samsung Electronics Co., Ltd. Method and apparatus for extracting feature vector used for face recognition and retrieval
US20050180610A1 (en) * 2004-02-18 2005-08-18 Fuji Xerox Co., Ltd. Image processing apparatus
US7590266B2 (en) * 2004-02-18 2009-09-15 Fuji Xerox Co., Ltd. Image processing apparatus
US20060034517A1 (en) * 2004-05-17 2006-02-16 Mitsubishi Denki Kabushiki Kaisha Method and apparatus for face description and recognition
US7630526B2 (en) * 2004-05-17 2009-12-08 Mitsubishi Denki Kabushiki Kaisha Method and apparatus for face description and recognition
US20060153430A1 (en) * 2004-12-03 2006-07-13 Ulrich Canzler Facial feature analysis system for users with physical disabilities
US20060148323A1 (en) * 2004-12-03 2006-07-06 Ulrich Canzler Facial feature analysis system
US7689010B2 (en) * 2004-12-03 2010-03-30 Invacare International Sarl Facial feature analysis system
US20090087036A1 (en) * 2005-05-31 2009-04-02 Nec Corporation Pattern Matching Method, Pattern Matching System, and Pattern Matching Program
US20070046662A1 (en) * 2005-08-23 2007-03-01 Konica Minolta Holdings, Inc. Authentication apparatus and authentication method
US20070088548A1 (en) * 2005-10-19 2007-04-19 Kabushiki Kaisha Toshiba Device, method, and computer program product for determining speech/non-speech
US8121357B2 (en) 2005-10-24 2012-02-21 Nec Corporation Coefficient determining method, feature extracting method, system, and program, and pattern checking method, system, and program
US20090123077A1 (en) * 2005-10-24 2009-05-14 Nec Corporation Coefficient determining method, feature extracting method, system, and program, and pattern checking method, system, and program
US7999846B2 (en) * 2005-12-06 2011-08-16 Hitachi Kokusai Electric Inc. Image processing apparatus, image processing system, and recording medium for programs therefor
US9805064B2 (en) * 2006-02-01 2017-10-31 Sony Corporation System, apparatus, method, program and recording medium for processing image
US20170083543A1 (en) * 2006-02-01 2017-03-23 Sony Corporation System, apparatus, method, program and recording medium for processing image
US20150234861A1 (en) * 2006-02-01 2015-08-20 Sony Corporation System, apparatus, method, program and recording medium for processing image
US11281712B2 (en) * 2006-02-01 2022-03-22 Sony Group Corporation System, apparatus, method, program and recording medium for processing image
US9710490B2 (en) * 2006-02-01 2017-07-18 Sony Corporation System, apparatus, method, program and recording medium for processing image
US20180004773A1 (en) * 2006-02-01 2018-01-04 Sony Corporation System, apparatus, method, program and recording medium for processing image
US20200151210A1 (en) * 2006-02-01 2020-05-14 Sony Corporation System, Apparatus, Method, Program And Recording Medium For Processing Image
US9965493B2 (en) * 2006-02-01 2018-05-08 Sony Corporation System, apparatus, method, program and recording medium for processing image
US10579669B2 (en) * 2006-02-01 2020-03-03 Sony Corporation System, apparatus, method, program and recording medium for processing image
US20190205335A1 (en) * 2006-02-01 2019-07-04 Sony Corporation System, Apparatus, Method, Program And Recording Medium For Processing Image
US10289698B2 (en) * 2006-02-01 2019-05-14 Sony Corporation System, apparatus, method, program and recording medium for processing image
US7734097B1 (en) * 2006-08-01 2010-06-08 Mitsubishi Electric Research Laboratories, Inc. Detecting objects in images with covariance matrices
US20110216943A1 (en) * 2006-08-02 2011-09-08 Sony Corporation Image-capturing apparatus and method, expression evaluation apparatus, and program
US20110216218A1 (en) * 2006-08-02 2011-09-08 Sony Corporation Image-capturing apparatus and method, expression evaluation apparatus, and program
US20110216217A1 (en) * 2006-08-02 2011-09-08 Sony Corporation Image-capturing apparatus and method, expression evaluation apparatus, and program
US20110216216A1 (en) * 2006-08-02 2011-09-08 Sony Corporation Image-capturing apparatus and method, expression evaluation apparatus, and program
US8416996B2 (en) 2006-08-02 2013-04-09 Sony Corporation Image-capturing apparatus and method, expression evaluation apparatus, and program
US8260041B2 (en) 2006-08-02 2012-09-04 Sony Corporation Image-capturing apparatus and method, expression evaluation apparatus, and program
US8416999B2 (en) * 2006-08-02 2013-04-09 Sony Corporation Image-capturing apparatus and method, expression evaluation apparatus, and program
US8406485B2 (en) 2006-08-02 2013-03-26 Sony Corporation Image-capturing apparatus and method, expression evaluation apparatus, and program
US20080037841A1 (en) * 2006-08-02 2008-02-14 Sony Corporation Image-capturing apparatus and method, expression evaluation apparatus, and program
US8260012B2 (en) 2006-08-02 2012-09-04 Sony Corporation Image-capturing apparatus and method, expression evaluation apparatus, and program
US20110216942A1 (en) * 2006-08-02 2011-09-08 Sony Corporation Image-capturing apparatus and method, expression evaluation apparatus, and program
US8238618B2 (en) 2006-08-02 2012-08-07 Sony Corporation Image-capturing apparatus and method, facial expression evaluation apparatus, and program
US7724960B1 (en) 2006-09-08 2010-05-25 University Of Central Florida Research Foundation Inc. Recognition and classification based on principal component analysis in the transform domain
US8099277B2 (en) 2006-09-27 2012-01-17 Kabushiki Kaisha Toshiba Speech-duration detector and computer program product therefor
US20080077400A1 (en) * 2006-09-27 2008-03-27 Kabushiki Kaisha Toshiba Speech-duration detector and computer program product therefor
US8606019B2 (en) 2007-04-23 2013-12-10 Nec Corporation Matching method for two-dimensional pattern, feature extracting method, apparatus used for the methods, and programs
US20100150452A1 (en) * 2007-04-23 2010-06-17 Nec Corporation Matching method for two-dimensional pattern, feature extracting method, appartus used for the methods, and programs
US20090180671A1 (en) * 2007-10-19 2009-07-16 Samsung Electronics Co., Ltd. Multi-view face recognition method and system
US8615113B2 (en) * 2007-10-19 2013-12-24 Samsung Electronics Co., Ltd. Multi-view face recognition method and system
US20090136121A1 (en) * 2007-11-14 2009-05-28 Ryo Nakagaki Defect review method and apparatus
US8526710B2 (en) * 2007-11-14 2013-09-03 Hitachi High-Technologies Corporation Defect review method and apparatus
US8380500B2 (en) 2008-04-03 2013-02-19 Kabushiki Kaisha Toshiba Apparatus, method, and computer program product for judging speech/non-speech
US20090254341A1 (en) * 2008-04-03 2009-10-08 Kabushiki Kaisha Toshiba Apparatus, method, and computer program product for judging speech/non-speech
US20120237134A1 (en) * 2008-04-07 2012-09-20 Microsoft Corporation Image descriptor quantization
US8712159B2 (en) * 2008-04-07 2014-04-29 Microsoft Corporation Image descriptor quantization
US20110058733A1 (en) * 2008-04-30 2011-03-10 Osaka Prefecture University Public Corporation Method of compiling three-dimensional object identifying image database, processing apparatus and processing program
US8306315B2 (en) 2008-04-30 2012-11-06 Osaka Prefecture University Public Corporation Method of compiling three-dimensional object identifying image database, processing apparatus and processing program
US8218880B2 (en) 2008-05-29 2012-07-10 Microsoft Corporation Linear laplacian discrimination for feature extraction
US20110135167A1 (en) * 2008-07-10 2011-06-09 Nec Corporation Personal authentication system and personal authentication method
US8553983B2 (en) * 2008-07-10 2013-10-08 Nec Corporation Personal authentication system and personal authentication method
US20100076723A1 (en) * 2008-09-23 2010-03-25 Microsoft Corporation Tensor linear laplacian discrimination for feature extraction
US8024152B2 (en) * 2008-09-23 2011-09-20 Microsoft Corporation Tensor linear laplacian discrimination for feature extraction
US8712190B2 (en) * 2008-10-10 2014-04-29 Nec Corporation Collating device, collating method, and program
US20110194776A1 (en) * 2008-10-10 2011-08-11 Nec Corporation Collating device, collating method, and program
US8213691B2 (en) * 2008-12-30 2012-07-03 Mitsubishi Electric Research Laboratories, Inc. Method for identifying faces in images with improved accuracy using compressed feature vectors
US20100166266A1 (en) * 2008-12-30 2010-07-01 Michael Jeffrey Jones Method for Identifying Faces in Images with Improved Accuracy Using Compressed Feature Vectors
US20120023134A1 (en) * 2009-03-27 2012-01-26 Nec Corporation Pattern matching device, pattern matching method, and pattern matching program
US8861871B2 (en) * 2009-04-14 2014-10-14 Nec Corporation Image signature extraction device
US20120027309A1 (en) * 2009-04-14 2012-02-02 Nec Corporation Image signature extraction device
US8805013B2 (en) 2011-06-16 2014-08-12 Shinkawa Ltd. Pattern position detecting method
US9355303B2 (en) * 2011-12-04 2016-05-31 King Saud University Face recognition using multilayered discriminant analysis
US20130142399A1 (en) * 2011-12-04 2013-06-06 King Saud University Face recognition using multilayered discriminant analysis
US9471830B2 (en) * 2012-03-14 2016-10-18 Kabushiki Kaisha Toshiba Collation apparatus, collation method, and computer program product
US20130243271A1 (en) * 2012-03-14 2013-09-19 Kabushiki Kaisha Toshiba Collation apparatus, collation method, and computer program product
WO2014001610A1 (en) * 2012-06-25 2014-01-03 Nokia Corporation Method, apparatus and computer program product for human-face features extraction
US9710698B2 (en) 2012-06-25 2017-07-18 Nokia Technologies Oy Method, apparatus and computer program product for human-face features extraction
US9317909B2 (en) * 2013-02-27 2016-04-19 Samsung Electronics Co., Ltd. Image subsystem including image feature detection hardware component and image processing system including the same
US9576222B2 (en) 2013-04-09 2017-02-21 Hitachi Kokusai Electric Inc. Image retrieval apparatus, image retrieval method, and recording medium
US20150030216A1 (en) * 2013-07-26 2015-01-29 Fujitsu Limited Image processing apparatus and method
US9483710B1 (en) * 2013-07-26 2016-11-01 Fujitsu Limited Image processing apparatus and method
US9405956B2 (en) * 2013-07-26 2016-08-02 Fujitsu Limited Image processing apparatus and method
US9704065B2 (en) 2014-03-20 2017-07-11 Kabushiki Kaisha Toshiba Dimension reduction apparatus, dimension reduction method, and computer program product
US20170069112A1 (en) * 2014-05-22 2017-03-09 Huawei Technologies Co., Ltd. Image Characteristic Estimation Method and Device
US10115208B2 (en) * 2014-05-22 2018-10-30 Huawei Technologies Co., Ltd. Image characteristic estimation method and device
US10083343B2 (en) * 2014-08-08 2018-09-25 Samsung Electronics Co., Ltd. Method and apparatus for facial recognition
US20160042223A1 (en) * 2014-08-08 2016-02-11 Samsung Electronics Co., Ltd. Method and apparatus for facial recognition
US9652664B1 (en) * 2014-12-30 2017-05-16 Morphotrust Usa, Llc Facial recognition using fractal features
US10095916B1 (en) 2014-12-30 2018-10-09 Morphotrust Usa, Llc Facial recognition using fractal features
US10402629B1 (en) 2014-12-30 2019-09-03 Morphotrust Usa, Llc Facial recognition using fractal features
US10366279B2 (en) 2017-08-29 2019-07-30 Bank Of America Corporation System for execution of multiple events based on image data extraction and evaluation
US11222199B2 (en) * 2018-12-05 2022-01-11 International Business Machines Corporation Automatically suggesting behavioral adjustments during video conferences
US20200226772A1 (en) * 2019-01-13 2020-07-16 Hunan Agricultural Information And Engineering Institute Anti-counterfeiting method based on feature of surface texture image of products
US11244169B2 (en) 2020-06-15 2022-02-08 Bank Of America Corporation System for executing multiple events based on video data extraction and evaluation
CN112115881A (en) * 2020-09-21 2020-12-22 国网重庆市电力公司长寿供电分公司 Image feature extraction method based on robust identification feature learning
CN116347104A (en) * 2023-05-22 2023-06-27 宁波康达凯能医疗科技有限公司 Method, device and storage medium for encoding intra-frame image based on efficient discriminant analysis

Also Published As

Publication number Publication date
HK1090157A1 (en) 2006-12-15
ATE511680T1 (en) 2011-06-15
US20080304750A1 (en) 2008-12-11
EP2149852A1 (en) 2010-02-03
EP1522962A1 (en) 2005-04-13
EP1522962B1 (en) 2011-06-01
KR100578614B1 (en) 2006-05-10
EP2149852B1 (en) 2016-01-06
JP2004192603A (en) 2004-07-08
US8116571B2 (en) 2012-02-14
KR20050009698A (en) 2005-01-25
JP4292837B2 (en) 2009-07-08
EP1522962A4 (en) 2007-12-26
WO2004008391A1 (en) 2004-01-22

Similar Documents

Publication Publication Date Title
US8116571B2 (en) Pattern feature extraction via fourier amplitudes of a block image
JP4589625B2 (en) Face recognition using kernel fisher face
Learned-Miller Data driven image models through continuous joint alignment
US7254256B2 (en) Method and computer program product for locating facial features
JP4770932B2 (en) Pattern feature extraction method and apparatus
US9053388B2 (en) Image processing apparatus and method, and computer-readable storage medium
US8897503B2 (en) Face recognition with combined PCA-based datasets
US20120121142A1 (en) Ultra-low dimensional representation for face recognition under varying expressions
EP1758059A1 (en) Parameter estimation method, parameter estimation device, and collation method
JP2005512201A5 (en)
US20080267519A1 (en) Method for Identifying Images After Cropping
US20080279424A1 (en) Method of Identifying Faces from Face Images and Corresponding Device and Computer Program
Laparra et al. Principal polynomial analysis
CN114445904A (en) Iris segmentation method, apparatus, medium, and device based on full convolution neural network
CN110321935B (en) Method and device for determining business event relation and computer readable storage medium
JPH10289320A (en) Pattern recognition device and method for sorting by means of candidate table
El Meslouhi et al. Unimodal multi-feature fusion and one-dimensional hidden Markov models for low-resolution face recognition
CN101082955B (en) Pattern characteristic extraction method and device for the same
CN116310462B (en) Image clustering method and device based on rank constraint self-expression
Tuffour Enhancing Facial Recognition: A Comparative Analysis of Image Preprocessing Techniques in PCA, KL Expansion and SVD
Mahanta Unified framework for matrix-variate linear discriminant analysis
Manfredi et al. Face detection in a compressed domain
Altowairqi Master of Information Security & Biometrics, School of Engineering and Digital Arts, the University of Kent
Zhi et al. Improved neighborhood preserving embedding approach
Pang et al. Compression of pre-computed per-pixel texture features using MDS

Legal Events

Date Code Title Description
AS Assignment

Owner name: NEC CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KAMEI, TOSHIO;REEL/FRAME:016595/0974

Effective date: 20041004

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION