IMAGE PROCESSING APPARATUS AND IMAGE PROCESSING METHOD
1. Field of the Invention The present invention relates to an image processing apparatus and an image processing method, and more particularly, to a technology preferably used to detect a face of a person from image data and perform recognition processing for identifying a person of the detected face. 2. Description of the Related Art As discussed in Japanese Patent Application Laid-Open No. 2008-271310, a configuration is known for detecting a face of a person from image data and performing recognition processing for identifying a person of the detected face. In addition, there is a known image processing apparatus configured to segment high resolution image data into a plurality of image data and to process the segmented image data in parallel. For example, as discussed in Japanese Patent Application Laid-Open No. 2000-253397, a configuration is known for controlling parameters in parallel in an encoding process. By combining such techniques, a configuration may be conceivable, in which the high resolution image data is segmented, and recognition processing for identifying a person is performed in parallel for the respective segmented image data. In such an image processing apparatus, recognition processing for different persons can be made among the respective segmented image data. Therefore, it is possible to reduce a time taken for the recognition processing for all persons included in the image data before the segmentation. However, if the recognition processing is performed on overall segmented image data at the same timing, it may be sometimes difficult to effectively reduce the time taken for the recognition processing. For example, in a moving picture containing image data of a plurality of frames, it is assumed that a person appears in any segmented image data at the timing of the face detection processing just before the start of the recognition processing. In addition, it is assumed that a person appears in another segmented image data slightly later than the timing of the face detection processing just before the start of the recognition processing. In this case, since the face detection processing and the recognition processing are performed for the first person as soon as he/she appears on a screen, it is possible to identify the person with a small time loss. However, in the case of the later person, the person is identified in the next recognition processing after the image is displayed, and then the first recognition processing is completed. Therefore, a much time loss is produced in waiting for completion of the first recognition processing. In this manner, if the recognition processing is applied to a plurality of segmented image data at the same timing, the time taken until the recognition processing is completed after the person appears may increase in comparison with the person who appears in other segmented image data. The invention provides an image processing apparatus and an image processing method capable of suitably controlling timing for executing face detection processing and recognition processing for identifying a person of the detected face for segmented image data. According to an aspect of the invention, there is provided an image processing apparatus including a segmentation unit configured to segment image data into a plurality of segments to create a plurality of segmented image data, a detection unit configured to execute face detection processing for detecting a face area from image data before being segmented and each of the plurality of segmented image data, a recognition unit configured to execute recognition processing for determining whether the face detected by the detection unit is a face of a registered person, in each of the plurality of segmented image data, and a control unit configured to control whether to cause the recognition unit to execute the recognition processing is as to each of the plurality of segmented image data depending on a result of detection of a face in the image data before being segmented. Further features and aspects of the present invention will become apparent from the following detailed description of exemplary embodiments with reference to the attached drawings. The accompanying drawings, which are incorporated in and constitute a part of the identification, illustrate exemplary embodiments, features, and aspects of the invention and, together with the description, serve to explain the principles of the invention. Various exemplary embodiments, features, and aspects of the invention will be described in detail below with reference to the drawings. Hereinafter, exemplary embodiments of the invention will be described. Referring to The image processing unit 105 performs various processing such as noise reduction, gamma correction, luminance signal correction, or color-difference signal correction for the input image data. The face detection unit 106 performs face detection processing for detecting faces in the image data, and the characteristic extraction unit 107 extracts characteristics from the face detected by the face detection unit 106. The recognition unit 108 compares the characteristics extracted by the characteristic extraction unit 107 and characteristics registered in the temporary memory 109 in advance to determine whether a person of the detected face is a registered person. More specifically, the characteristic extraction unit 107, the recognition unit 108, and the temporary memory 109 perform recognition processing for identifying a person of the face detected by the face detection unit 106. The temporary memory 109 may temporarily store the characteristics used by the recognition unit 108, the image data necessary in each processing performed by the image processing unit 105, the face detection unit 106, and the characteristic extraction unit 107, or information created during the processing. The camera signal processing units 104 The codec unit 111 performs an encoding processing for encoding the image data into recording data having a predetermined format and a decoding processing for decoding the recording data into the image data. The temporary memory 112 temporarily stores the data used in the encoding processing and the decoding processing. The recording signal processing units 110 The display data creation unit 113 creates various kinds of information indicating device states or display data for implementing a graphical user interface and outputs it to the mixer 114. The mixer 114 combines the image data output from the recording signal processing group and the display data from the display data creation unit 113. The display device 115 includes a display panel such as an LCD or an organic EL and displays the image data output from the mixer 114. The recording data stored in the recording media 117 is read/written by the recording signal processing units 110 Next, a signal flow during the imaging process will be described with reference to The light passing through the lens 101 is photoelectrically converted in the image sensor 102. The image sensor 102 has a number of pixels necessary to output four or more than four times as much an amount of the image data as the high definition television (HDTV) and outputs the created image data to the multiplexer unit 103. In the multiplexer unit 103, the processing is performed for every single frame of the image data. In the multiplexer unit 103, among the input image data, first, the image including 4098 horizontal pixels and 2196 vertical pixels as effective pixels in The segmented image data A, B, C, and D are input to the camera signal processing units 104 The camera signal processing group performs various signal processing such as noise reduction, gamma correction, luminance signal correction, or color-difference signal correction on the segmented image data or the reduced image data. The segmented image data and the reduced image data are stored in the temporary memory 109, and the image processing is performed by the image processing unit 105. The image data processed by the image processing unit 105 is input to the recording signal processing group. The segmented image data and the reduced image data after the image processing are temporarily stored in the temporary memory 112 and are encoded into a predetermined format using the codec unit 111. The encoded image data is written to the recording media 117 through the recording media interface 116. The connection between the recording signal processing units 110 In addition, each image data output from the recording signal processing group is input the mixer 114. In the mixer 114, the four segmented image data are combined in synchronization with the recording signal processing group to create combined image data having the same angle of view as the image data before the segmentation. The combined image data becomes an image of 4098 horizontal pixels and 2196 vertical pixels. The combined image data is converted into a size and a format suitable for the display resolution of the display device 115 and is used to display an image on the display device 115. The system controller 118 notifies the display data creation unit 113 of the information regarding the status of the imaging apparatus. The display data creation unit 113 creates display data including texts or icons for informing a user of the status of the apparatus based on the notified information. The mixer 114 overlaps the display data with the combined image data, and the overlapped combined image data is output to the display device 115. Next, the face detection processing and the recognition processing will be described. The face detection unit 106 detects a face of the person included in the image signal input to the camera signal processing units 104 The recognition unit 108 compares the characteristic information of the created face information and the registered characteristic information stored in the memory 119 and generates recognition result indicating whether the detected face is a registered face based on similarity between the information. The recognition processing is performed on the image data of a plurality of frames, and it is determined that the person of the corresponding face is a registered person if the similarity to the same registered characteristic information exceeds a predetermined threshold predetermined times. The system controller 118 also controls operations of the face detection unit 106, the characteristic extraction unit 107, and the recognition unit 108. For example, the system controller 118 switches the threshold for identifying a registered person depending on the imaging condition. Further, the system controller 118 switches a person whose characteristic information is compared if the registered characteristic information regarding a plurality of registered persons is provided. For example, since it takes a long time if the recognition processing is performed using all pieces of the registered characteristic information, the system controller 118 causes a user in advance to select characteristic information which is used and performs the recognition processing using the characteristic information of the selected person. If the information is not selected by a user, the recognition processing may be performed using all pieces of the registered characteristic information. In addition, the system controller 118 manages the registered characteristic information. If it is determined that the characteristic information of the detected face is not similar to any registered characteristic information, the system controller 118 performs control such that the characteristic information extracted from this face is newly stored in the memory 119 as the registered characteristic information for this person. In addition, the system controller 118 automatically deletes the registered characteristic information which has not been used for a long time and is determined as unnecessary. In addition, for the purpose of information protection, the characteristic information may be encrypted when it is written and decrypted at the time of reading. Next, the operation of the system controller 118 when the object is captured will be described by way of example. The image in First, the camera signal processing unit 104 Next, the system controller 118 instructs the camera signal processing unit 104 The system controller 118 instructs each of the camera signal processing units 104 In this manner, the system controller 118 repeatedly controls the camera signal processing unit, which receives the segmented image data corresponding to the area in which the face is not detected, from the reduced image data E, to perform the face detection processing and then the recognition processing when the face is detected. If the face is not detected even when the face detection processing is performed, the face detection processing is performed again. In addition, since a person does not exist in the segmented image data C and D, the camera signal processing units 104 If a face having a detectable size appears in the area corresponding to any one of the segmented image data B to D of the reduced image data E, the camera signal processing unit which receives the segmented image data corresponding to the position of the corresponding face does not perform the face detection processing and performs the recognition processing. If the recognition unit 108 performs the recognition processing and successfully identifies a person, the system controller 118 obtains name information corresponding to the personal identification information for the identified person. Then, the system controller 118 causes the display data creation unit 113 to process the obtained information, and the display device 115 displays the name along with the image data. At the same time, the face frame is displayed as information indicating a fact that the face is detected. As a result, a user can recognize a person who is detected and know who he is based on the scene currently displayed on the screen. In this manner, the system controller 118 performs the face detection processing for the reduced image data E and then performs control to decide whether the face detection processing or the recognition processing is performed for each of the segmented image data A to D based on the detection result. Next, description will be made about switching control between the face detection processing and the recognition processing performed for the camera signal processing group of the system controller 118 according to the present exemplary embodiment with reference to the flowcharts in According to the present exemplary embodiment, it is assumed that each of the camera signal processing units 104 In step S501 in In step S502, the system controller 118 determines whether the face detection processing for the reduced image data E is completed by the camera signal processing unit 104 In step S503, the system controller 118 stores the frame identification information (FID) for identifying the frame for which the face detection processing has been completed by the camera signal processing unit 104 In step S504, the system controller 118 obtains information regarding an execution flag indicating whether the face detection processing or the recognition processing for each segmented image corresponding to the frame indicated by the LFID is being executed from each of the camera signal processing units 104 In step S505, the system controller 118 determines whether a determination flag of the face detection processing or the recognition processing is set to be ON in step S508 described below for the camera signal processing units 104 In step S506, the system controller 118 performs control to execute the face detection processing or the recognition processing for the camera signal processing units 104 In step S507, out of the camera signal processing units 104 In step S508, the system controller 118 determines whether the face detection processing or the recognition processing is executed for the camera signal processing unit in which both the determination flag and the execution flag are set to OFF and sets the determination flag of the corresponding camera signal processing unit to be ON. Then, the process advances to step S504. In step S509, the system controller 118 causes the camera signal processing unit to continue the execution of the face detection processing or the recognition processing being executed, and the process advances to step S504. Then, the processing of step S508 will be described with reference to the flowchart in In step S601, the system controller 118 receives information indicating the size and the position of the face from the camera signal processing unit 104 In step S602, the system controller 118 selects information indicating the size of the face of the area corresponding to the segmented image data A out of the information indicating the sizes of faces obtained from the reduced image data E. The system controller 118 obtains the size of the face in the segmented image data A by multiplying the size of the selected face by a reciprocal number of the reduction ratio of the reduced image data E, and the process advances to step S603. In addition, if a plurality of faces is detected from the segmented image data A, the system controller 118 multiplies a sum of the sizes of faces by a reciprocal number of the reduction ratio of the reduced image data E. In step S603, the system controller 118 computes a ratio between the area of the face included in the segmented image data A computed in step S602 and the area of the segmented image data A and obtains a ratio of the face area in the segmented image data A. If the ratio of the face area is equal to or higher than a threshold, in other words, if the ratio occupied by the face within the screen is high, the system controller 118 determines that there is a low possibility that other detectable faces exist in the remaining areas, and the process advances to step S604. Otherwise, the process advances to step S605. In step S604, the system controller 118 instructs the camera signal processing unit 104 If the process advances from step S603 to step S604, the camera signal processing unit 104 In step S605, the system controller 118 determines whether the face detection processing has been performed just previously, using the segmented image data A. If it is determined that the face detection processing is performed just previously, the process advances to step S604. Otherwise, the process advances to step S606. In step S606, the system controller 118 instructs the camera signal processing unit 104 The flowchart in Next, a recognition rate according to the first exemplary embodiment of the invention will be described. In the method of the conventional art, as illustrated in In In comparison, according to the first exemplary embodiment described above, the system controller 118 determines whether the face detection processing or the recognition processing is executed for each of the camera signal processing units 104 In addition, for the segmented image data B where a small-sized face exists, the face is not detected from the area corresponding to the segmented image data B of the reduced image data E. Therefore, the camera signal processing unit 104 Further, since a face does not exist in the segmented image data C and D, the camera signal processing units 104 With respect to the person A who is close to the imaging apparatus, and whose face is large in size, a frequency of the recognition processing increases compared to the conventional art. For the person B who is distant from the imaging apparatus, and whose face is small in size, the process waits for the face detection processing as to the reduced image data E only in an initial single cycle. In the subsequent cycles, the recognition processing is executed with the same frequency as the conventional art. If this person B gets closer to the imaging apparatus, and the size of the detected face increases, the frequency of the recognition processing for the person B increases similar to the person A. In the area where a face does not exist, the face detection processing is performed in every cycle. Therefore, in comparison with the conventional art, a frequency of the face detection processing increases. Therefore, when a face appears in such an area, it is possible to detect the face immediately. In the conventional art, as to the segmented image data corresponding to the area where a face does not exist, there is a cycle in which both the face detection processing and the recognition processing are not performed. However, according to the present exemplary embodiment, since one of the processing is always performed, an efficient control can be carried out compared to the conventional art. Hereinafter, a second exemplary embodiment of the invention will be described. The imaging apparatus according to the second exemplary embodiment of the invention has a configuration similar to the imaging apparatus according to the first exemplary embodiment. The imaging apparatus according to the second exemplary embodiment is different from the imaging apparatus according to the first exemplary embodiment in that each face detection unit 106 of the camera signal processing units 104 Each face detection unit 106 of the camera signal processing units 104 The system controller 118 causes the camera signal processing units 104 The system controller 118 controls the face detection processing and the recognition processing according to the flowchart in The processing of step S508 according to the second exemplary embodiment will be described with reference to the flowchart in In step S901, the system controller 118 receives information indicating the size and the position of the face from the camera signal processing unit 104 In step S902, the system controller 118 selects information indicating the size of the face in the area corresponding to the segmented image data A out of the information indicating the size of the face obtained from the reduced image data E. The system controller 118 obtains the size of the face in the segmented image data A by multiplying the size of the selected face by a reciprocal number of the reduction ratio of the reduced image data E, and the process advances to step S903. If a plurality of faces is detected in the segmented image data A, the system controller 118 multiplies a sum of the sizes of the corresponding faces by a reciprocal number of the reduction ratio of the reduced image data E. In step S903, the system controller 118 computes a ratio between the area of the face included in the segmented image data A computed in step S902 and the area of the segmented image data A to obtain a ratio of the face area in the segmented image data A. If the ratio of the face area is equal to or higher than a threshold, in other words, if the ratio of the area occupied by the face in the screen is high, the system controller 118 determines that a possibility is low that there is a face having the size to be recognized by the recognition unit 108 in a rest of the area, and the process advances to step S904. Otherwise, the process advances to step S905. In step S904, the system controller 118 instructs the camera signal processing unit 104 If the process advances from step S903 to step S904, the camera signal processing unit 104 In step S905, the system controller 118 determines whether the face detection processing has been performed just previously on the segmented image data A. If it is determined that the face detection processing has been performed just previously, the process advances to step S904. Otherwise, the process advances to step S906. In step S906, the system controller 118 instructs the camera signal processing unit 104 Next, the recognition rate according to the second exemplary embodiment of the invention will be described with reference to According to the second exemplary embodiment, the system controller 118 determines whether only the face detection processing is executed or both the face detection processing and the recognition processing are executed in parallel by each of the camera signal processing units 104 As to the segmented image data B where a small-sized face exists, a face is not detected from the area corresponding to the segmented image data B of the reduced image data E. Therefore, the camera signal processing unit 104 Thus, similar to the first exemplary embodiment, as to the person A whose face size is large and who is close to the imaging apparatus, the frequency of the recognition processing increases compared to the conventional art. As to the person B whose face size is small and who is far distant from the imaging apparatus, the process waits for the face detection processing in the reduced image data E only in an initial single cycle. However, in the subsequent cycles, the recognition processing is executed with the same frequency as the conventional art. If the person B is close to the imaging apparatus, and the size of the detected face increases, the frequency of the recognition processing increases similar to the person A. Since the face detection processing is performed in every cycle regardless whether the recognition processing is performed, the frequency of the face detection processing increases compared to the conventional art. Accordingly, it is possible to detect a face immediately after a face appears in the corresponding area. Each face detection unit 106 of the camera signal processing units 104 In addition, according to the present exemplary embodiment, the recognition processing can be initiated at independent timing as to each segmented image data. Therefore, the process is completed without performing the unnecessary recognition processing. As a result, it is possible to reduce a time lag elapsed until the recognition result is obtained after a person appears in the segmented image data. The exemplary embodiments of the invention have been described in detail hereinbefore. However, it should be understood that the invention includes not only the face detection processing and the recognition processing at the time of the imaging but also a case where the face detection processing and the personal recognition processing are executed at the time of image data reproduction after the recording. According to the exemplary embodiments described above, whether the recognition processing is executed is determined based on the information on whether a face exists and information on the size of the face. However, the invention is not limited thereto. For example, whether the face detection processing or the recognition processing is performed may be determined considering other kinds of information regarding the face obtained through the face detection, such as information on the distance to the face or a difference of the characteristic information of the face coming from the age. In addition, the invention may be realized by executing the following processing. Software (computer program) capable of realizing the functions of the aforementioned exemplary embodiments may be supplied to a system or an apparatus through various kinds of computer readable storage medium or network, and a computer (CPU or MPU) of the system or the apparatus may read and execute the program. While the present invention has been described with reference to exemplary embodiments, it is to be understood that the invention is not limited to the disclosed exemplary embodiments. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all modifications, equivalent structures, and functions. This application claims priority from Japanese Patent Application No. 2011-060547 filed Mar. 18, 2011, and No. 2012-035594 filed Feb. 21, 2012 which are hereby incorporated by reference herein in their entirety. Provided is an image processing apparatus including a segmentation unit configured to segment image data into a plurality of segments to create a plurality of segmented image data, a detection unit configured to execute face detection processing for detecting a face area from image data before being segmented and each of the plurality of segmented image data, a recognition unit configured to execute recognition processing for determining whether the face detected by the detection unit is a face of a registered person, in each of the plurality of segmented image data, and a control unit configured to control whether to cause the recognition unit to execute the recognition processing as to each of the plurality of segmented image data depending on a result of the face detection in the image data before being segmented. 1. An image processing apparatus comprising:
a segmentation unit configured to segment image data into a plurality of segments to create a plurality of segmented image data; a detection unit configured to execute face detection processing for detecting a face area from image data before being segmented and each of the plurality of segmented image data; a recognition unit configured to execute recognition processing for determining whether the face detected by the detection unit is a face of a registered person, in each of the plurality of segmented image data; and a control unit configured to control whether to cause the recognition unit to execute the recognition processing as to each of the plurality of segmented image data depending on a result of detection of a face in the image data before being segmented. 2. The image processing apparatus according to wherein the detection unit detects the face area from the image data before the segmentation reduced by the reduction unit. 3. The image processing apparatus according to wherein the control unit determines the segmented image data which includes a face based on a result of the face detection in the image data before the segmentation, the control unit causes the recognition unit to execute the recognition processing as to the segmented image data determined to include the face, and the control unit causes the detection unit to execute the face detection processing as to the segmented image data determined not to include a face. 4. The image processing apparatus according to wherein the control unit determines the segmented image data which includes the face based on a result of the face detection in the image data before the segmentation, the control unit causes the recognition unit to execute the recognition processing and causes the detection unit to execute the face detection processing as to the segmented image data determined to include a face, and the control unit causes the detection unit to execute the face detection processing and does not cause the recognition unit to execute the recognition processing as to the segmented image data determined not to include a face. 5. The image processing apparatus according to wherein the control unit determines the segmented image data which includes the face based on a result of the face detection in the image data before the segmentation, the control unit obtains a ratio of the face area in the segmented image data as to the segmented image data determined to include the face, the control unit causes the recognition unit to execute the recognition processing if the ratio is equal to or higher than a threshold, and the control unit causes the detection unit to execute the face detection processing if the ratio is lower than the threshold. 6. The image processing apparatus according to wherein the control unit determines the segmented image data which includes the face based on a result of the face detection in the image data before the segmentation, the control unit obtains a ratio of the face area in the segmented image data as to the segmented image data determined to include a face, the control unit causes the recognition unit to execute the recognition processing and causes the detection unit to execute the face detection processing if the ratio is equal to or higher than a threshold, and the control unit does not cause the recognition unit to execute the recognition processing and causes the detection unit to execute the face detection processing if the ratio is lower than the threshold. 7. An image processing method comprising:
segmenting image data into a plurality of segments to create a plurality of segmented image data; executing face detection processing for detecting a face area from the image data before being segmented and each of the plurality of segmented image data; executing recognition processing for determining whether the detected face is a face of a registered person, in each of the plurality of segmented image data; and controlling whether the recognition processing is executed as to each of the plurality of segmented image data based on a result of the face detection in the image data before being segmented. BACKGROUND OF THE INVENTION
SUMMARY OF THE INVENTION
BRIEF DESCRIPTION OF THE DRAWINGS
DESCRIPTION OF THE EMBODIMENTS
Other Embodiments