EP2703949A1 - Information processing device, information processing method, and recording medium - Google Patents

Information processing device, information processing method, and recording medium Download PDF

Info

Publication number
EP2703949A1
EP2703949A1 EP12776429.8A EP12776429A EP2703949A1 EP 2703949 A1 EP2703949 A1 EP 2703949A1 EP 12776429 A EP12776429 A EP 12776429A EP 2703949 A1 EP2703949 A1 EP 2703949A1
Authority
EP
European Patent Office
Prior art keywords
reference data
subject
distance
camera
hand
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
EP12776429.8A
Other languages
German (de)
French (fr)
Other versions
EP2703949A4 (en
EP2703949B1 (en
Inventor
Takafumi Kurokawa
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NEC Solution Innovators Ltd
Original Assignee
NEC System Technologies Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NEC System Technologies Ltd filed Critical NEC System Technologies Ltd
Publication of EP2703949A1 publication Critical patent/EP2703949A1/en
Publication of EP2703949A4 publication Critical patent/EP2703949A4/en
Application granted granted Critical
Publication of EP2703949B1 publication Critical patent/EP2703949B1/en
Not-in-force legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/002Specific input/output arrangements not covered by G06F3/01 - G06F3/16
    • G06F3/005Input arrangements through a video camera
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language

Definitions

  • the present invention relates to an information processing device, an information processing method, and a recording medium.
  • Patent Literature 1 discloses an interface device that obtains the outline of a user's hand in a captured image, and moves a cursor to a virtual switch on a display and selects the virtual switch, in accordance with the movement and change in shape of the hand.
  • Patent Literature 2 discloses an information processing device that recognizes the shape of a user's hand from a captured image, and displays a menu of a plurality of executable functions when a first shape is recognized. The information processing device selects an executable function in accordance with the position of the shape recognized in the image, and executes a selected function when a second shape is recognized.
  • Patent Literature 3 discloses an information processing device that masks a part of an image of the palm of a user's hand captured by a camera, detects movement of the palm of the hand by means of the masked image, and recognizes a command corresponding to movement of the hand.
  • All of the information processing devices disclosed in the Patent Literatures 1 to 3 capture an image of a user's hand, extracts an image of the hand from the captured image, and in addition, recognizes the movement or shape of the hand.
  • extracting an image of a user's hand and also recognizing the movement or shape of the hand are not easy. For example, as the distance between the user and the camera becomes further, the image of the hand becomes smaller, the time for extraction processing takes longer, and also the recognition rate becomes declined. Therefore, inputting a suitable operation command becomes difficult.
  • the present invention takes the aforementioned circumstances into account, and an objective thereof is to provide an information processing device, an information processing method, and a recording medium, which make it possible for more suitable data to be input on the basis of the state of a captured subject.
  • An information processing device comprises:
  • An information processing method comprises:
  • a program recorded on a recording medium causes a computer to function as:
  • An information processing device 100 is provided with a gesture input function that specifies a gesture (state) that appears in the movement, shape, inclination and the like of a hand 200 serving as a subject, and inputs input data such as a command corresponding to the specified gesture.
  • the information processing device 100 is provided with an operator 1, a camera 2 serving as an image capturer, and a display 3.
  • the information processing device 100 is a commonly used computer.
  • the information processing device 100 processes input data corresponding to the gesture of the hand 200 and data input by the operation of the operator 1.
  • the operator 1 is provided with a keyboard, a mouse, and the like.
  • the operator 1 inputs, to the information processing device 100, data corresponding to operational input of the user.
  • the camera 2 is provided with imaging elements such as a CCD (charge-coupled device) or CMOS (complementary metal-oxide semiconductor) sensor.
  • the camera 2 for example, captures an image of the hand 200 of the user for the information processing device 100, as a subject positioned within the field of vision.
  • the user inputs various data by moving the hand 200 and altering the shape of the hand 200, within the field of vision of the camera 2.
  • the display 3 is provided with a LCD (liquid crystal display) or the like.
  • the display 3 displays an image relating to an application being executed by the information processing device 100. Furthermore, the display 3 displays a gesture input support window 3W constantly or in response to a predetermined operation. Selection targets S such as menu items, files, or icons, and a selection cursor C are displayed in the gesture input support window 3W.
  • FIG. 2 depicts a functional configuration of the information processing device 100.
  • the information processing device 100 is provided with a storer 4 and a controller 5 as well as the operator 1, the camera 2, and the display 3.
  • the storer 4 is provided with a storage device or the like such as a RAM (random-access memory) or a hard disk.
  • the storer 4 stores various types of data.
  • the storer 4 stores reference data for specifying a user's gesture by the hand 200 and input data corresponding to the gesture, for each distance between the camera 2 and the hand 200.
  • the storer 4 is provided with a long-distance reference data storer 41, a middle-distance reference data storer 42, and a short-distance reference data storer 43.
  • the reference data is data for specifying a gesture (also referred to as a hand gesture operation pattern) expressed by the movement, shape, inclination and the like of the hand 200, such as those depicted in FIG. 3 to FIG. 5 .
  • the long-distance reference data storer 41 stores reference data specifying a user's simple gesture by the hand 200 that is able to be specified even from a long distance.
  • the long-distance reference data storer 41 for example, stores reference data for specifying the shape of the hand 200 and the leftward, rightward, upward, and downward movements of the hand 200 as exemplified in FIG. 3 .
  • the reference data for specifying the shape of the hand 200 is, for example, pattern data or the like for which the shape of the hand is converted to data.
  • the reference data for specifying the leftward, rightward, upward, and downward movements of the hand 200 is, for example, data relating to the amount of movement of the hand's center of gravity per unit time, or pattern data, which is converted data of the locus with respect to the center of gravity when the hand is moved, or the like.
  • the reference data stored in the long-distance reference data storer 41 is used for specifying a user's gesture by the hand 200, in the case where the distance between the camera 2 and the hand 200 is far apart and the region of the hand 200 in the image data is small.
  • the middle-distance reference data storer 42 stores reference data for specifying a user's gesture by the hand 200 that is relatively difficult to specify.
  • the middle-distance reference data storer 42 stores reference data for specifying shapes of the hand 200, such as those exemplified in FIG. 4A to FIG. 4D.
  • FIG. 4A is a shape of a hand in which all fingers are extended.
  • FIG. 4B is a shape of a hand in which the index finger is extended and the remaining fingers are bent.
  • FIG. 4C is a shape of a hand in which the index finger and the middle finger are extended and the remaining fingers are bent.
  • FIG. 4D is a shape of a hand in which the index finger, the middle finger, and the third finger are extended and the remaining fingers are bent.
  • the short-distance reference data storer 43 stores reference data for specifying a user's gesture by the hand 200 that is even more relatively difficult to specify.
  • the short-distance reference data storer 43 stores reference data for specifying shapes of the hand 200, such as those exemplified in FIG. 5A to FIG. 5D.
  • FIG. 5A to FIG. 5D are examples in which the shapes of the hand itself are all the same but the inclinations of the hand 200 are different.
  • the reference data stored in the short-distance reference data storer 43 is used for specifying a gesture by the hand 200, in the case where the distance between the camera 2 and the hand 200 is close and the hand 200 in the image data is large.
  • the middle-distance reference data storer 42 also stores reference data that is stored by the long-distance reference data storer 41.
  • the short-distance reference data storer 43 also stores reference data that is stored by the long-distance reference data storer 41 and reference data that is stored by the middle-distance reference data storer 42.
  • the reference data is ranked in accordance with the difficulty of specifying the hand gestures, and if the short-distance reference data storer 43 is taken as having the top rank and the long-distance reference data storer 41 is taken as having the bottom rank, top-rank reference data includes bottom-rank reference data.
  • the storer 4 stores reference data for specifying a larger number of gestures as the distance between the camera 2 and the hand 200 becomes closer.
  • the storer 4 stores input data associated with a user's gesture by the hand 200, for each distance between the camera 2 and the hand 200. It is arbitrary as to what kinds of data are associated with a user's gesture by the hand 200.
  • the long-distance reference data storer 41 associates and stores the direction of the movement of the hand 200 depicted in FIG. 3 specified by stored reference data, and a command for moving the cursor C within the gesture input support window 3W in the direction.
  • the middle-distance reference data storer 42 associates and stores the number of extended fingers of the hand 200 depicted in FIG. 4A to FIG. 4D specified by the stored reference data with a command for designating a menu number that matches the number of fingers within the gesture input support window 3W. Besides this, it is permissible for the middle-distance reference data storer 42 to associate and store the number of extended fingers of the hand 200, and a command for inputting numerical data that matches the number of fingers.
  • the short-distance reference data storer 43 associates and stores the shape of the hand 200 depicted in FIG. 5A to FIG. 5D specified by stored reference data with a command for scrolling an image within the gesture input support window 3W in the direction indicated by the index finger in that shape. Besides this, it is permissible for the short-distance reference data storer 43 to associate and store the shape of the hand 200 with a command such as for moving a cursor in the direction indicated by the index finger in the shape.
  • the controller 5 is provided with a processor (processing unit) or the like.
  • the processor has an auxiliary storage device such as a flash memory, an EEPROM (electrically erasable program read-only memory), or a hard disk device, and has RAM or the like that is working memory.
  • the processor in accordance with a program stored in the EEPROM, temporarily stores image data and other data in the RAM, and executes an information processing algorithm.
  • the controller 5 is provided with an inputter 51, an acquirer 52, a selector 53, a specifier 54, and an executer 55.
  • the controller 5 operates in accordance with a program stored in the storer 4, and executes a variety of processing in accordance with input data such as a command input by a user.
  • the inputter 51 inputs a captured image of the hand 200 captured by the camera 2.
  • the inputter 51 converts an analog signal of the captured image into a digital signal and generates image data on a frame-by-frame basis.
  • the inputter 51 outputs the image data on a frame-by-frame basis to the display 3 and displays this image data as a live view image.
  • the acquirer 52 acquires information indicating the distance between the camera 2 (image capturer) and the hand 200.
  • the acquirer 52 analyzes image data captured by the camera 2 and acquires the information indicating the distance between the camera 2 and the hand 200.
  • the acquirer 52 acquires the length L1 between the temples in a captured user face region, the length L2 of the distance between the eyes, the width L3 of the back of the hand 200, or the width L4 of the wrist (hereafter, referred to as the temple length L1 or the like), such as those depicted in FIG. 6 , as information indicating the distance between the camera 2 and the hand 200.
  • the acquirer 52 acquires the temple length L1 or the like from image analysis data generated by the specifier 54 that is described later.
  • the temple length L1 or the like correlates with the distance between the camera 2 and the user without being much affected by individual differences. Therefore, it is able to be said that the shorter the distance between the camera 2 and the hand 200 becomes, the shorter the temple length L1 or the like becomes.
  • the selector 53 selects reference data corresponding to the distances indicated by L1 to L4 or the like acquired by the acquirer 52.
  • the selector 53 detects the distance D between the camera 2 and the hand 200 of the user on the basis of L1 to L4 acquired by the acquirer 52. For example, in the detection of the distance D, first, the selector 53 obtains the distance between the camera 2 and the user by substituting L1 to L4 into a relational expression of these distances and the distance between the camera 2 and the user. Because the distance between the camera 2 and the user approximately matches the distance D between the camera 2 and the hand 200, the selector 53 detects the obtained distance as the distance D between the camera 2 and the hand 200.
  • the selector 53 selects reference data stored in the short-distance reference data storer 43. If the distance D is greater than the reference value D1 and equal to or less than a reference value D2, the selector 53 selects reference data stored in the middle-distance reference data storer 42. If the distance D is greater than the reference value D2, the selector 53 selects reference data stored in the long-distance reference data storer 41.
  • the specifier 54 refers to the reference data selected by the selector 53, specifies a user's gesture by the hand 200 in the captured image input to the inputter 51, and specifies input data corresponding to the specified gesture.
  • the specifier 54 extracts the hand 200 from image data on a frame-by-frame basis generated by the inputter 51 and generates image analysis data.
  • the specifier 54 is provided with a memory that stores image data of a plurality of frames.
  • the specifier 54 obtains a difference image of an image of the current captured frame and an image of the preceding captured frame.
  • the specifier 54 binarizes the difference image obtained, and if the eight neighboring pixel values surrounding each white pixel in the binarized image are greater than a preset threshold value, the specifier 54 expands the white pixel region. By doing this, the specifier 54 executes expansion processing for connecting and expanding white regions.
  • the specifier 54 reduces the black pixel region. By doing this, reduction processing for removing black dots that constitute noise is executed. In this way, the specifier 54 generates image analysis data by extracting, from the image data, an image of the hand 200 as an object having movement.
  • the specifier 54 uses the above-described difference image so as to include, in the image analysis data, data relating to the amount of movement per unit time of the center of gravity of the hand 200, or pattern data which is converted data of the locus with respect to the center of gravity when the hand 200 is moved, or the like.
  • Image analysis data generated in this way is used in the acquisition of information indicating the distance between the camera 2 and the user by the acquirer 52. It should be noted that it is permissible for the specifier 54 to extract an image of the hand 200 by analyzing skin-color degree information or brightness information or the like with respect to the image data, and generate image analysis data.
  • the specifier 54 extracts an image of the hand 200 in the captured image that is input by the inputter 51, and on the basis of the reference data selected by the selector 53, specifies a user's gesture by the hand 200 depicted by the extracted image of the hand 200. More specifically, the specifier 54 extracts data corresponding to the hand 200 from the generated analysis image data, and compares the extracted data with the reference data selected by the selector 53. If a user's gesture by the hand 200 included in the analysis image data matches the reference data, the specifier 54 specifies the gesture of the hand 200 corresponding to the matching reference data. The specifier 54 reads out input data associated with the specified gesture from the storer 4, and inputs the input data to the executer 55.
  • the executer 55 executes processing corresponding to the input data associated with the user's gesture by the hand 200 specified by the specifier 54.
  • the executer 55 refers to the long-distance reference data storer 41, the middle-distance reference data storer 42, and the short-distance reference data storer 43, and executes a command associated with the gesture of the hand 200 specified by the specifier 54.
  • the executer 55 outputs a movement command for cursor C to the display 3.
  • the cursor C within the gesture input support window 3W of the display 3 moves.
  • the information processing device 100 is able to execute a variety of application software.
  • the information processing device 100 performs input processing in the case where a user carries out input by means of a gesture, in addition to input processing from the operator 1.
  • the controller 5 executes input processing depicted in FIG. 7 in, for example, a time-divided manner. It should be noted that it is permissible for the controller 5 to display the gesture input support window 3W along with the start of the input processing, or to switch between display and non-display of the gesture input support window 3W by means of a gesture or input from the operator 1.
  • the specifier 54 stores image data of a frame output from the inputter 51 (step S1).
  • step S2 the specifier 54 generates image analysis data (step S2). If an image of the hand 200 is not extracted upon analysis of the image analysis data (step S3: NO), the controller 5 returns to step S 1. On the other hand, if an image of the hand 200 is extracted upon analysis of the image analysis data (step S3: YES), the acquirer 52 acquires the temple length L1 or the like (step S4). Next, the selector 53 detects the distance D on the basis of the temple length L1 or the like (step S5).
  • step S6 If the detected distance D is equal to or less than the reference value D1 (step S6: YES), the selector 53 selects reference data stored in the short-distance reference data storer 43 (step S7).
  • step S6 If the detected distance D is greater than the reference value D1 (step S6: NO), and the detected distance D is equal to or less than the reference value D2 (step S8: YES), the selector 53 selects reference data stored in the middle-distance reference data storer 42 (step S9).
  • step S8 NO
  • the selector 53 selects reference data stored in the long-distance reference data storer 41 (step S10).
  • the specifier 54 compares the generated analysis image data with the reference data selected by the selector 53, and determines whether or not the gesture of the hand 200 matches the reference data (step S11).
  • step S11: NO If the gesture of the hand 200 does not match the reference data (step S11: NO), the controller 5 returns to step S1. On the other hand, if the gesture of the hand 200 matches the reference data (step S11: YES), the specifier 54 specifies the gesture of the hand 200 and a command associated with the specified gesture (step S12).
  • the executer 55 executes the command associated with the gesture of the hand 200 specified by the specifier 54 (step S 13). Then, the controller 5 terminates the input processing.
  • reference data for specifying a user's gesture by the hand 200 is selected in accordance with the distance between the camera 2 and the hand 200.
  • a user's gesture by the hand 200 that is specifiable at the distance between the camera 2 and the hand 200 is specified, it is possible for more suitable data to be input on the basis of the user's gesture by the hand 200 captured by the camera 2.
  • the storer 4 is configured so as to store reference data for specifying a larger number of gestures as the distance between the camera 2 and the hand 200 becomes closer.
  • the camera 2 and the hand 200 are positioned at a middle distance, although the number of specifiable user's gestures by the hand 200 decreases compared to the case of a short distance, it is possible to reduce the processing load and erroneous specifying because the reference patterns of unspecifiable gestures and analysis image data are not compared to each other.
  • the distance between the camera 2 and the hand 200 is far apart, although the number of specifiable user's gestures by the hand 200 further decreases compared to the cases of a short distance and a middle distance, it is possible to further reduce the processing load and erroneous specifying because the reference patterns of unspecifiable gestures and analysis image data are not compared to each other. Furthermore, because the reference patterns of specifiable gestures and analysis image data are compared even in the case where the distance between the camera 2 and the hand 200 is far apart, it is possible for input data associated with a user gesture by the hand 200 to be input reliably.
  • the storer 4 is configured so as to store reference data for specifying, as a state, the movement, shape, and inclination of the hand 200 serving as a subject.
  • reference data for specifying, as a state, the movement, shape, and inclination of the hand 200 serving as a subject.
  • an image of the hand 200 in a captured image input by the inputter 51 is extracted, and a user's gesture by the hand 200 depicted by the extracted image of the hand 200 is specified on the basis of reference data selected by the selector 53.
  • the acquirer 52 is configured so as to analyze image data captured by the camera 2 and acquire information indicating the distance between the camera 2 and the user.
  • the device configuration it is possible for the device configuration to be implemented in a simple manner compared to the case where an optical sensor or the like is used.
  • the display 3 is configured so as to display an image corresponding to the image data captured by the camera 2.
  • an image of the hand 200 extracted from the image data is displayed in the gesture input support window 3W.
  • the display 3 it is permissible for the display 3 to display, in a window 3i, the detected distance D and a message reporting the types of commands that are able to be specified at the distance.
  • the user is able to know what kinds of commands are able to be input by means of a gesture.
  • the display 3 is permissible for the display 3 to be configured so as to display a gesture of the hand 200 specified on the basis of reference data selected by the selector 53, and a command or the like associated with the gesture of the hand 200.
  • the display 3 displays the gesture input support window 3W that displays an image of the hand 200.
  • the display 3 displays, in the window 3i, gestures of the hand 200 that correspond to reference data stored in the middle-distance reference data storer 42 and are able to be specified at the current distance between the hand 200 and the camera 2, and displays a command specifying a menu number that is input when the gesture is specified.
  • the user it is possible for the user to know the gestures by the hand 200 that are able to be specified and what kinds of commands are able to be input with respect to those gestures.
  • the acquirer 52 it is permissible, for example, for the acquirer 52 to be configured so as to be provided with an infrared-light emitter, and to acquire information indicating the distance between the camera 2 and the user by means of reflection light in which infrared rays emitted from the infrared-light emitter are reflected by the hand 200. Furthermore, it is permissible for a plurality of cameras to be arranged, and for the acquirer 52 to be configured so as to acquire stereoscopic images from the plurality of cameras to obtain information indicating the distance between the camera 2 and the user on the basis of a parallax. It is also permissible for the acquirer 52 to use various types of sensors such as an optical sensor for the acquisition of information indicating the distance between the camera 2 and the user.
  • the information processing device 100 may use an infrared camera or the like as the camera 2 to capture an image of the hand 200. Still further, it is permissible for the information processing device 100 to use, for example, a depth sensor or the like to capture an image of the hand 200, without being restricted to an optical camera or an infrared camera or the like.
  • the distance between the camera 2 and the hand 200 is divided into three stages; however, it is permissible for this distance to be divided into two stages, or to be divided into four or more stages.
  • the user's gestures that appear in the movement, shape, inclination and the like of a hand 200 depicted in FIG. 3 to FIG. 5 are examples, and are not restricted to these examples.
  • the specifier 54 it is permissible for the specifier 54 to specify gestures by left hand or right hand, and to specify gestures using both hands.
  • gestures that accompany high-speed movements are difficult to specify at a long distance compared to gestures with no movement. Therefore, for example, it is permissible for the specifier 54 to be configured so as to specify the shape of the hand only at the state where the hand is stationary when the distance D between the camera 2 and the hand 200 is farther than a predetermined distance. In this regard, it is permissible for the specifier 54 to be configured so as to specify a gesture that combines the shape and movement of the hand 200 when the distance D is closer than the predetermined distance.
  • an example has been given in which input is carried out by means of a gesture by the hand 200 in the present embodiment; however, for example, it is permissible for the information processing device 100 to be configured so as to be able to capture an image of the eyes, mouth, eyelids, tongue and the like, and to carry out input on the basis of a gesture or state (for example, relative position, movement, direction, degree of opening, degree of closing, and the like) implemented thereby.
  • a gesture or state for example, relative position, movement, direction, degree of opening, degree of closing, and the like
  • the storer 4 stores reference data for each distance in the present embodiment
  • the present invention is not restricted to this.
  • the information processing device 100 and the camera 2 are configured as a single unit in the present embodiment, it is also possible for the camera 2 to be arranged independently from the information processing device 100, for example.
  • the user carries out operational input implemented by a gesture of the hand 200 to the information processing device 100, by means of the camera 2 transmitting image data to the information processing device 100 via a communication network.
  • the user is able to input suitable data even in a remote operation specification such as when the camera 2 is placed away from a main body of the information processing device 100.
  • the information processing device 100 of the present embodiment is implemented by means of a dedicated system, or by means of a normal computer system.
  • a recording medium that records the aforementioned program it is possible to use a computer-readable recording medium such as a USB memory, flexible disk, CD, DVD, Blu-ray Disc (registered trademark), MO, SD card, Memory Stick (registered trademark), as well as a magnetic disk, optical disk, magneto-optical disk, semiconductor memory, or magnetic tape. Furthermore, it is possible to use a recording medium such as a hard disk or a SSD (solid state drive) or the like that is normally used fixedly in a system or device.
  • a computer-readable recording medium such as a USB memory, flexible disk, CD, DVD, Blu-ray Disc (registered trademark), MO, SD card, Memory Stick (registered trademark), as well as a magnetic disk, optical disk, magneto-optical disk, semiconductor memory, or magnetic tape.
  • a recording medium such as a hard disk or a SSD (solid state drive) or the like that is normally used fixedly in a system or device.
  • An information processing device comprising:
  • the information processing device according to Appendix 1 or 2, wherein the storer stores reference data for specifying, as the state, a movement, shape, and inclination of a hand serving as the subject.
  • the information processing device according to any one of Appendices 1 to 3, wherein the specifier extracts an image of the subject in the captured image input by the inputter, and specifies the state of the subject depicted by the extracted image of the subject, based on the reference data selected by the selector.
  • the information processing device according to any one of Appendices 1 to 4, wherein the acquirer analyzes image data captured by the image capturer, and acquires the information indicating the distance between the camera and the subject.
  • the information processing device further comprises a display that displays the captured image of the subject captured by the camera.
  • the information processing device wherein the display displays the state of the subject specified based on the reference data selected by the selector, and the input data corresponding to the state.
  • An information processing method comprising:
  • the present invention is suitable for an information processing device, an information processing method, and a recording medium with which data is input by means of a user's gesture.

Abstract

An inputter (51) inputs a captured image of a hand captured by a camera (2). An acquirer (52) acquires information indicating the distance between the camera (2) and the hand. A storer (4) stores reference data for specifying a hand gesture and a command corresponding to the gesture, for each distance between the camera (2) and the hand. A selector (53) selects, from among the reference data stored in the storer (4), reference data corresponding to the distance indicated by the information acquired by the acquirer (52). A specifier (54) refers to the reference data selected by the selector (53), specifies a hand gesture in the captured image input by the inputter (51), and specifies a command corresponding to the specified gesture.

Description

    Technical Field
  • The present invention relates to an information processing device, an information processing method, and a recording medium.
  • Background Art
  • Information processing devices that capture a user by means of a camera, and input an operation command by means of a captured user's gesture are proposed.
  • For example, Patent Literature 1 discloses an interface device that obtains the outline of a user's hand in a captured image, and moves a cursor to a virtual switch on a display and selects the virtual switch, in accordance with the movement and change in shape of the hand.
  • Patent Literature 2 discloses an information processing device that recognizes the shape of a user's hand from a captured image, and displays a menu of a plurality of executable functions when a first shape is recognized. The information processing device selects an executable function in accordance with the position of the shape recognized in the image, and executes a selected function when a second shape is recognized.
  • Furthermore, Patent Literature 3 discloses an information processing device that masks a part of an image of the palm of a user's hand captured by a camera, detects movement of the palm of the hand by means of the masked image, and recognizes a command corresponding to movement of the hand.
  • Citation List Patent Literature
    • Patent Literature 1: Unexamined Japanese Patent Application Kokai Publication No. 2004-78977
    • Patent Literature 2: Unexamined Japanese Patent Application Kokai Publication No. 2008-146243
    • Patent Literature 3: Unexamined Japanese Patent Application Kokai Publication No. 2002-83302
    Summary of Invention Technical Problem
  • All of the information processing devices disclosed in the Patent Literatures 1 to 3 capture an image of a user's hand, extracts an image of the hand from the captured image, and in addition, recognizes the movement or shape of the hand. However, extracting an image of a user's hand and also recognizing the movement or shape of the hand are not easy. For example, as the distance between the user and the camera becomes further, the image of the hand becomes smaller, the time for extraction processing takes longer, and also the recognition rate becomes declined. Therefore, inputting a suitable operation command becomes difficult.
  • The present invention takes the aforementioned circumstances into account, and an objective thereof is to provide an information processing device, an information processing method, and a recording medium, which make it possible for more suitable data to be input on the basis of the state of a captured subject.
  • Solution to Problem
  • An information processing device according to a first aspect of the present invention comprises:
    • an inputter that inputs a captured image of a subject captured by a camera;
    • an acquirer that acquires information indicating a distance between the camera and the subject;
    • a storer that stores reference data for specifying a state of the subject and input data corresponding to the state, for each distance between the camera and the subject;
    • a selector that selects, from among the reference data stored in the storer, reference data corresponding to the distance indicated by the information acquired by the acquirer; and
    • a specifier that refers to the reference data selected by the selector, specifies the state of the subject in the captured image input by the inputter, and specifies data corresponding to the specified state.
  • An information processing method according to a second aspect of the present invention comprises:
    • a storing step for storing reference data for specifying a state of a subject captured by a camera and input data corresponding to the state, for each distance between the camera and the subject;
    • an inputting step for inputting a captured image of the subject captured by the camera;
    • an acquirer that acquires information indicating a distance between the camera and the subject;
    • a selecting step for selecting, from among the reference data stored in the storing step, reference data corresponding to the distance indicated by the information acquired in the acquiring step; and
    • a specifying step for referring to the reference data selected in the selecting step, specifying the state of the subject in the captured image input in the inputting step, and specifying data corresponding to the specified state.
  • A program recorded on a recording medium according to a third aspect of the present invention causes a computer to function as:
    • an inputter that inputs a captured image of a subject captured by a camera;
    • an acquirer that acquires information indicating a distance between the camera and the subject;
    • a storer that stores reference data for specifying a state of the subject and input data corresponding to the state for each distance between the camera and the subject;
    • a selector that selects, from among the reference data stored in the storer, reference data corresponding to the distance indicated by the information acquired by the acquirer; and
    • a specifier that refers to the reference data selected by the selector, specifies the state of the subject in the captured image input by the inputter, and specifies data corresponding to the specified state.
    Advantageous Effects of Invention
  • According to the present invention, it is possible for more suitable data to be input on the basis of the state of a captured subject.
  • Brief Description of Drawings
    • FIG. 1 is a perspective view of an information processing device according to an embodiment of the present invention;
    • FIG. 2 is a functional block diagram of the information processing device of FIG. 1;
    • FIG. 3 is a drawing depicting an example of hand movements as long-distance gestures;
    • FIGS. 4A, 4B, 4C, and 4D are drawings depicting examples of hand shapes as middle-distance gestures;
    • FIGS. 5A, 5B, 5C, and 5D are drawings depicting examples of hand inclinations as short-distance gestures;
    • FIG. 6 is a drawing for illustrating a method for acquiring information indicating the distance between a camera and a user;
    • FIG. 7 is a flowchart for illustrating input processing in the embodiment;
    • FIG. 8 is a drawing depicting an exemplary display of an image of a hand and a message, on a display; and
    • FIG. 9 is a drawing depicting an exemplary display of an image of a hand and specifiable gestures on a display.
    Description of Embodiments
  • An information processing device according to an embodiment of the present invention is hereafter described with reference to the drawings. It should be noted that the same reference signs are appended to the same or corresponding portions in the drawings.
  • An information processing device 100 according to the present embodiment is provided with a gesture input function that specifies a gesture (state) that appears in the movement, shape, inclination and the like of a hand 200 serving as a subject, and inputs input data such as a command corresponding to the specified gesture. As depicted in FIG. 1, the information processing device 100 is provided with an operator 1, a camera 2 serving as an image capturer, and a display 3.
  • The information processing device 100 is a commonly used computer. The information processing device 100 processes input data corresponding to the gesture of the hand 200 and data input by the operation of the operator 1.
  • The operator 1 is provided with a keyboard, a mouse, and the like. The operator 1 inputs, to the information processing device 100, data corresponding to operational input of the user.
  • The camera 2 is provided with imaging elements such as a CCD (charge-coupled device) or CMOS (complementary metal-oxide semiconductor) sensor. The camera 2, for example, captures an image of the hand 200 of the user for the information processing device 100, as a subject positioned within the field of vision. The user inputs various data by moving the hand 200 and altering the shape of the hand 200, within the field of vision of the camera 2.
  • The display 3 is provided with a LCD (liquid crystal display) or the like. The display 3 displays an image relating to an application being executed by the information processing device 100. Furthermore, the display 3 displays a gesture input support window 3W constantly or in response to a predetermined operation. Selection targets S such as menu items, files, or icons, and a selection cursor C are displayed in the gesture input support window 3W.
  • FIG. 2 depicts a functional configuration of the information processing device 100. The information processing device 100 is provided with a storer 4 and a controller 5 as well as the operator 1, the camera 2, and the display 3.
  • The storer 4 is provided with a storage device or the like such as a RAM (random-access memory) or a hard disk. The storer 4 stores various types of data. For example, the storer 4 stores reference data for specifying a user's gesture by the hand 200 and input data corresponding to the gesture, for each distance between the camera 2 and the hand 200. For this reason, the storer 4 is provided with a long-distance reference data storer 41, a middle-distance reference data storer 42, and a short-distance reference data storer 43. The reference data is data for specifying a gesture (also referred to as a hand gesture operation pattern) expressed by the movement, shape, inclination and the like of the hand 200, such as those depicted in FIG. 3 to FIG. 5.
  • The long-distance reference data storer 41 stores reference data specifying a user's simple gesture by the hand 200 that is able to be specified even from a long distance. The long-distance reference data storer 41, for example, stores reference data for specifying the shape of the hand 200 and the leftward, rightward, upward, and downward movements of the hand 200 as exemplified in FIG. 3. The reference data for specifying the shape of the hand 200 is, for example, pattern data or the like for which the shape of the hand is converted to data. The reference data for specifying the leftward, rightward, upward, and downward movements of the hand 200 is, for example, data relating to the amount of movement of the hand's center of gravity per unit time, or pattern data, which is converted data of the locus with respect to the center of gravity when the hand is moved, or the like. The reference data stored in the long-distance reference data storer 41 is used for specifying a user's gesture by the hand 200, in the case where the distance between the camera 2 and the hand 200 is far apart and the region of the hand 200 in the image data is small.
  • The middle-distance reference data storer 42 stores reference data for specifying a user's gesture by the hand 200 that is relatively difficult to specify. The middle-distance reference data storer 42, for example, stores reference data for specifying shapes of the hand 200, such as those exemplified in FIG. 4A to FIG. 4D. FIG. 4A is a shape of a hand in which all fingers are extended. FIG. 4B is a shape of a hand in which the index finger is extended and the remaining fingers are bent. FIG. 4C is a shape of a hand in which the index finger and the middle finger are extended and the remaining fingers are bent. FIG. 4D is a shape of a hand in which the index finger, the middle finger, and the third finger are extended and the remaining fingers are bent.
  • The short-distance reference data storer 43 stores reference data for specifying a user's gesture by the hand 200 that is even more relatively difficult to specify. The short-distance reference data storer 43, for example, stores reference data for specifying shapes of the hand 200, such as those exemplified in FIG. 5A to FIG. 5D. FIG. 5A to FIG. 5D are examples in which the shapes of the hand itself are all the same but the inclinations of the hand 200 are different. The reference data stored in the short-distance reference data storer 43 is used for specifying a gesture by the hand 200, in the case where the distance between the camera 2 and the hand 200 is close and the hand 200 in the image data is large.
  • It should be noted that the middle-distance reference data storer 42 also stores reference data that is stored by the long-distance reference data storer 41. Furthermore, the short-distance reference data storer 43 also stores reference data that is stored by the long-distance reference data storer 41 and reference data that is stored by the middle-distance reference data storer 42. In other words, the reference data is ranked in accordance with the difficulty of specifying the hand gestures, and if the short-distance reference data storer 43 is taken as having the top rank and the long-distance reference data storer 41 is taken as having the bottom rank, top-rank reference data includes bottom-rank reference data. In this way, the storer 4 stores reference data for specifying a larger number of gestures as the distance between the camera 2 and the hand 200 becomes closer.
  • Furthermore, the storer 4 stores input data associated with a user's gesture by the hand 200, for each distance between the camera 2 and the hand 200. It is arbitrary as to what kinds of data are associated with a user's gesture by the hand 200.
  • For example, the long-distance reference data storer 41 associates and stores the direction of the movement of the hand 200 depicted in FIG. 3 specified by stored reference data, and a command for moving the cursor C within the gesture input support window 3W in the direction.
  • For example, the middle-distance reference data storer 42 associates and stores the number of extended fingers of the hand 200 depicted in FIG. 4A to FIG. 4D specified by the stored reference data with a command for designating a menu number that matches the number of fingers within the gesture input support window 3W. Besides this, it is permissible for the middle-distance reference data storer 42 to associate and store the number of extended fingers of the hand 200, and a command for inputting numerical data that matches the number of fingers.
  • For example, the short-distance reference data storer 43 associates and stores the shape of the hand 200 depicted in FIG. 5A to FIG. 5D specified by stored reference data with a command for scrolling an image within the gesture input support window 3W in the direction indicated by the index finger in that shape. Besides this, it is permissible for the short-distance reference data storer 43 to associate and store the shape of the hand 200 with a command such as for moving a cursor in the direction indicated by the index finger in the shape.
  • The controller 5 is provided with a processor (processing unit) or the like. The processor has an auxiliary storage device such as a flash memory, an EEPROM (electrically erasable program read-only memory), or a hard disk device, and has RAM or the like that is working memory. The processor, in accordance with a program stored in the EEPROM, temporarily stores image data and other data in the RAM, and executes an information processing algorithm.
  • The controller 5 is provided with an inputter 51, an acquirer 52, a selector 53, a specifier 54, and an executer 55. The controller 5 operates in accordance with a program stored in the storer 4, and executes a variety of processing in accordance with input data such as a command input by a user.
  • The inputter 51 inputs a captured image of the hand 200 captured by the camera 2. The inputter 51 converts an analog signal of the captured image into a digital signal and generates image data on a frame-by-frame basis. The inputter 51 outputs the image data on a frame-by-frame basis to the display 3 and displays this image data as a live view image.
  • The acquirer 52 acquires information indicating the distance between the camera 2 (image capturer) and the hand 200. The acquirer 52 analyzes image data captured by the camera 2 and acquires the information indicating the distance between the camera 2 and the hand 200. The acquirer 52 acquires the length L1 between the temples in a captured user face region, the length L2 of the distance between the eyes, the width L3 of the back of the hand 200, or the width L4 of the wrist (hereafter, referred to as the temple length L1 or the like), such as those depicted in FIG. 6, as information indicating the distance between the camera 2 and the hand 200. For example, the acquirer 52 acquires the temple length L1 or the like from image analysis data generated by the specifier 54 that is described later. The temple length L1 or the like correlates with the distance between the camera 2 and the user without being much affected by individual differences. Therefore, it is able to be said that the shorter the distance between the camera 2 and the hand 200 becomes, the shorter the temple length L1 or the like becomes.
  • The selector 53 selects reference data corresponding to the distances indicated by L1 to L4 or the like acquired by the acquirer 52. The selector 53 detects the distance D between the camera 2 and the hand 200 of the user on the basis of L1 to L4 acquired by the acquirer 52. For example, in the detection of the distance D, first, the selector 53 obtains the distance between the camera 2 and the user by substituting L1 to L4 into a relational expression of these distances and the distance between the camera 2 and the user. Because the distance between the camera 2 and the user approximately matches the distance D between the camera 2 and the hand 200, the selector 53 detects the obtained distance as the distance D between the camera 2 and the hand 200.
  • For example, if the detected distance D is equal to or less than a reference value D1, the selector 53 selects reference data stored in the short-distance reference data storer 43. If the distance D is greater than the reference value D1 and equal to or less than a reference value D2, the selector 53 selects reference data stored in the middle-distance reference data storer 42. If the distance D is greater than the reference value D2, the selector 53 selects reference data stored in the long-distance reference data storer 41.
  • The specifier 54 refers to the reference data selected by the selector 53, specifies a user's gesture by the hand 200 in the captured image input to the inputter 51, and specifies input data corresponding to the specified gesture. In specifying the user's gesture by the hand 200, for example, the specifier 54 extracts the hand 200 from image data on a frame-by-frame basis generated by the inputter 51 and generates image analysis data.
  • Here, the generation of image analysis data will be described in detail. The specifier 54 is provided with a memory that stores image data of a plurality of frames. The specifier 54 obtains a difference image of an image of the current captured frame and an image of the preceding captured frame. In addition, the specifier 54 binarizes the difference image obtained, and if the eight neighboring pixel values surrounding each white pixel in the binarized image are greater than a preset threshold value, the specifier 54 expands the white pixel region. By doing this, the specifier 54 executes expansion processing for connecting and expanding white regions. Furthermore, if the eight neighboring pixel values surrounding each black pixel in the binarized image are less than the preset threshold value, the specifier 54 reduces the black pixel region. By doing this, reduction processing for removing black dots that constitute noise is executed. In this way, the specifier 54 generates image analysis data by extracting, from the image data, an image of the hand 200 as an object having movement.
  • Furthermore, it is permissible for the specifier 54 to use the above-described difference image so as to include, in the image analysis data, data relating to the amount of movement per unit time of the center of gravity of the hand 200, or pattern data which is converted data of the locus with respect to the center of gravity when the hand 200 is moved, or the like. Image analysis data generated in this way is used in the acquisition of information indicating the distance between the camera 2 and the user by the acquirer 52. It should be noted that it is permissible for the specifier 54 to extract an image of the hand 200 by analyzing skin-color degree information or brightness information or the like with respect to the image data, and generate image analysis data.
  • The specifier 54 extracts an image of the hand 200 in the captured image that is input by the inputter 51, and on the basis of the reference data selected by the selector 53, specifies a user's gesture by the hand 200 depicted by the extracted image of the hand 200. More specifically, the specifier 54 extracts data corresponding to the hand 200 from the generated analysis image data, and compares the extracted data with the reference data selected by the selector 53. If a user's gesture by the hand 200 included in the analysis image data matches the reference data, the specifier 54 specifies the gesture of the hand 200 corresponding to the matching reference data. The specifier 54 reads out input data associated with the specified gesture from the storer 4, and inputs the input data to the executer 55.
  • The executer 55 executes processing corresponding to the input data associated with the user's gesture by the hand 200 specified by the specifier 54. For example, the executer 55 refers to the long-distance reference data storer 41, the middle-distance reference data storer 42, and the short-distance reference data storer 43, and executes a command associated with the gesture of the hand 200 specified by the specifier 54. For example, in the case of a command for moving the cursor C, the executer 55 outputs a movement command for cursor C to the display 3. Thus, the cursor C within the gesture input support window 3W of the display 3 moves.
  • Next, the flow of input processing by the information processing device 100 will be described. The information processing device 100 is able to execute a variety of application software. In order to input to an application software, the information processing device 100, as described hereafter, performs input processing in the case where a user carries out input by means of a gesture, in addition to input processing from the operator 1.
  • The controller 5 executes input processing depicted in FIG. 7 in, for example, a time-divided manner. It should be noted that it is permissible for the controller 5 to display the gesture input support window 3W along with the start of the input processing, or to switch between display and non-display of the gesture input support window 3W by means of a gesture or input from the operator 1.
  • When input processing is started, first, the specifier 54 stores image data of a frame output from the inputter 51 (step S1).
  • Next, the specifier 54 generates image analysis data (step S2). If an image of the hand 200 is not extracted upon analysis of the image analysis data (step S3: NO), the controller 5 returns to step S 1. On the other hand, if an image of the hand 200 is extracted upon analysis of the image analysis data (step S3: YES), the acquirer 52 acquires the temple length L1 or the like (step S4). Next, the selector 53 detects the distance D on the basis of the temple length L1 or the like (step S5).
  • If the detected distance D is equal to or less than the reference value D1 (step S6: YES), the selector 53 selects reference data stored in the short-distance reference data storer 43 (step S7).
  • If the detected distance D is greater than the reference value D1 (step S6: NO), and the detected distance D is equal to or less than the reference value D2 (step S8: YES), the selector 53 selects reference data stored in the middle-distance reference data storer 42 (step S9).
  • If the detected distance D is greater than D2 (step S8: NO), the selector 53 selects reference data stored in the long-distance reference data storer 41 (step S10).
  • When reference data is selected by the selector 53, the specifier 54 compares the generated analysis image data with the reference data selected by the selector 53, and determines whether or not the gesture of the hand 200 matches the reference data (step S11).
  • If the gesture of the hand 200 does not match the reference data (step S11: NO), the controller 5 returns to step S1. On the other hand, if the gesture of the hand 200 matches the reference data (step S11: YES), the specifier 54 specifies the gesture of the hand 200 and a command associated with the specified gesture (step S12).
  • Next, the executer 55 executes the command associated with the gesture of the hand 200 specified by the specifier 54 (step S 13). Then, the controller 5 terminates the input processing.
  • According to the present embodiment as described above in detail, reference data for specifying a user's gesture by the hand 200 is selected in accordance with the distance between the camera 2 and the hand 200. Thus, because a user's gesture by the hand 200 that is specifiable at the distance between the camera 2 and the hand 200 is specified, it is possible for more suitable data to be input on the basis of the user's gesture by the hand 200 captured by the camera 2.
  • Furthermore, in the above-described embodiment, the storer 4 is configured so as to store reference data for specifying a larger number of gestures as the distance between the camera 2 and the hand 200 becomes closer. By doing this, in the case where the distance between the camera 2 and the hand 200 is close, it is possible for input data of a wide variety of commands or the like to be input since a larger number of gestures are able to be specified.
  • In the case where the camera 2 and the hand 200 are positioned at a middle distance, although the number of specifiable user's gestures by the hand 200 decreases compared to the case of a short distance, it is possible to reduce the processing load and erroneous specifying because the reference patterns of unspecifiable gestures and analysis image data are not compared to each other.
  • In the case where the distance between the camera 2 and the hand 200 is far apart, although the number of specifiable user's gestures by the hand 200 further decreases compared to the cases of a short distance and a middle distance, it is possible to further reduce the processing load and erroneous specifying because the reference patterns of unspecifiable gestures and analysis image data are not compared to each other. Furthermore, because the reference patterns of specifiable gestures and analysis image data are compared even in the case where the distance between the camera 2 and the hand 200 is far apart, it is possible for input data associated with a user gesture by the hand 200 to be input reliably.
  • It should be noted that, in the above-described embodiment, the storer 4 is configured so as to store reference data for specifying, as a state, the movement, shape, and inclination of the hand 200 serving as a subject. Thus, it is possible for the user to input data of a large number of categories because it is possible for data to be input by means of the movement (hand gesture) of the hand 200 that is able to express a variety of gestures.
  • Furthermore, in the above-described embodiment, an image of the hand 200 in a captured image input by the inputter 51 is extracted, and a user's gesture by the hand 200 depicted by the extracted image of the hand 200 is specified on the basis of reference data selected by the selector 53. By doing this, it is possible for the information processing device 100 to specify a user's gesture in real time, and input data in a prompt manner.
  • Furthermore, in the above-described embodiment, the acquirer 52 is configured so as to analyze image data captured by the camera 2 and acquire information indicating the distance between the camera 2 and the user. Thus, it is possible for the device configuration to be implemented in a simple manner compared to the case where an optical sensor or the like is used.
  • Furthermore, it is permissible for the display 3 to be configured so as to display an image corresponding to the image data captured by the camera 2. In this case, for example, as depicted in FIG. 8, an image of the hand 200 extracted from the image data is displayed in the gesture input support window 3W. By doing this, it is possible for the user to carry out operational input while confirming the gesture of the hand 200 to input data more reliably. In addition, it is permissible for the display 3 to display, in a window 3i, the detected distance D and a message reporting the types of commands that are able to be specified at the distance. Thus, the user is able to know what kinds of commands are able to be input by means of a gesture.
  • Furthermore, it is permissible for the display 3 to be configured so as to display a gesture of the hand 200 specified on the basis of reference data selected by the selector 53, and a command or the like associated with the gesture of the hand 200. For example, as depicted in FIG. 9, the display 3 displays the gesture input support window 3W that displays an image of the hand 200. In addition, the display 3 displays, in the window 3i, gestures of the hand 200 that correspond to reference data stored in the middle-distance reference data storer 42 and are able to be specified at the current distance between the hand 200 and the camera 2, and displays a command specifying a menu number that is input when the gesture is specified. Thus, it is possible for the user to know the gestures by the hand 200 that are able to be specified and what kinds of commands are able to be input with respect to those gestures.
  • It should be noted that, in the present embodiment, an example has been given in which information indicating the distance between the camera 2 and the user is obtained from image data; however, it is permissible, for example, for the acquirer 52 to be configured so as to be provided with an infrared-light emitter, and to acquire information indicating the distance between the camera 2 and the user by means of reflection light in which infrared rays emitted from the infrared-light emitter are reflected by the hand 200. Furthermore, it is permissible for a plurality of cameras to be arranged, and for the acquirer 52 to be configured so as to acquire stereoscopic images from the plurality of cameras to obtain information indicating the distance between the camera 2 and the user on the basis of a parallax. It is also permissible for the acquirer 52 to use various types of sensors such as an optical sensor for the acquisition of information indicating the distance between the camera 2 and the user.
  • Furthermore, it is permissible for the information processing device 100 to use an infrared camera or the like as the camera 2 to capture an image of the hand 200. Still further, it is permissible for the information processing device 100 to use, for example, a depth sensor or the like to capture an image of the hand 200, without being restricted to an optical camera or an infrared camera or the like.
  • It should be noted that, in the present embodiment, the distance between the camera 2 and the hand 200 is divided into three stages; however, it is permissible for this distance to be divided into two stages, or to be divided into four or more stages.
  • Furthermore, the user's gestures that appear in the movement, shape, inclination and the like of a hand 200 depicted in FIG. 3 to FIG. 5 are examples, and are not restricted to these examples. For example, it is permissible for the specifier 54 to specify gestures by left hand or right hand, and to specify gestures using both hands.
  • Furthermore, gestures that accompany high-speed movements are difficult to specify at a long distance compared to gestures with no movement. Therefore, for example, it is permissible for the specifier 54 to be configured so as to specify the shape of the hand only at the state where the hand is stationary when the distance D between the camera 2 and the hand 200 is farther than a predetermined distance. In this regard, it is permissible for the specifier 54 to be configured so as to specify a gesture that combines the shape and movement of the hand 200 when the distance D is closer than the predetermined distance.
  • Furthermore, an example has been given in which input is carried out by means of a gesture by the hand 200 in the present embodiment; however, for example, it is permissible for the information processing device 100 to be configured so as to be able to capture an image of the eyes, mouth, eyelids, tongue and the like, and to carry out input on the basis of a gesture or state (for example, relative position, movement, direction, degree of opening, degree of closing, and the like) implemented thereby.
  • It should be noted that, although the storer 4 stores reference data for each distance in the present embodiment, the present invention is not restricted to this. For example, it is permissible for the storer 4 to store reference data by appending identifiers to each reference data for identifying all distances, short distances, middle distances, and long distances. In this case, in accordance with the distance D detected by the selector 53, it is permissible for reference data having an identifier corresponding to the distance appended thereto to be read out, and to be determined whether or not the reference data matches the gesture of the hand 200 included in the image analysis data.
  • Furthermore, although the information processing device 100 and the camera 2 are configured as a single unit in the present embodiment, it is also possible for the camera 2 to be arranged independently from the information processing device 100, for example. In this case, the user carries out operational input implemented by a gesture of the hand 200 to the information processing device 100, by means of the camera 2 transmitting image data to the information processing device 100 via a communication network. By doing this, the user is able to input suitable data even in a remote operation specification such as when the camera 2 is placed away from a main body of the information processing device 100.
  • It is permissible for the information processing device 100 of the present embodiment to be implemented by means of a dedicated system, or by means of a normal computer system. For example, it is permissible to store and distribute a program for executing the aforementioned operations on a computer-readable recording medium, and to configure an input device 100 by installing this program on a computer to execute the aforementioned processing. Furthermore, it is permissible to store on a disk device provided in a server device on a network such as the Internet, and make it possible to download and the like to a computer, for example. Furthermore, it is permissible to implement the aforementioned functions by collaboration between an OS (operating system) and application software. In this case, it is permissible for only the portions other than the OS to be stored and distributed on a medium, and, furthermore, for those portions to be downloaded and the like to a computer.
  • As a recording medium that records the aforementioned program, it is possible to use a computer-readable recording medium such as a USB memory, flexible disk, CD, DVD, Blu-ray Disc (registered trademark), MO, SD card, Memory Stick (registered trademark), as well as a magnetic disk, optical disk, magneto-optical disk, semiconductor memory, or magnetic tape. Furthermore, it is possible to use a recording medium such as a hard disk or a SSD (solid state drive) or the like that is normally used fixedly in a system or device.
  • It is possible for the present invention to have various types of embodiments and modifications without departing from the broad spirit and scope of the present invention. Furthermore, the aforementioned embodiment is for describing the present invention, and does not restrict the scope of the present invention. In other words, the scope of the present invention is indicated not in the embodiment, but in Claims. Various types of modifications implemented within the claims and the meaning of the invention that is equivalent to the claims are deemed to be within the scope of the present invention.
  • It is possible for some or all of the aforementioned embodiment to be also described as in the following appendices without restriction thereto.
  • (Appendix 1)
  • An information processing device comprising:
    • an inputter that inputs a captured image of a subject captured by a camera;
    • an acquirer that acquires information indicating a distance between the camera and the subject;
    • a storer that stores reference data for specifying a state of the subject and input data corresponding to the state, for each distance between the camera and the subject;
    • a selector that selects, from among the reference data stored in the storer, reference data corresponding to the distance indicated by the information acquired by the acquirer; and
    • a specifier that refers to the reference data selected by the selector, specifies the state of the subject in the captured image input by the inputter, and specifies data corresponding to the specified state.
    (Appendix 2)
  • The information processing device according to Appendix 1, wherein the storer stores reference data for specifying a larger number of states as the distance between the camera and the subject becomes closer.
  • (Appendix 3)
  • The information processing device according to Appendix 1 or 2, wherein the storer stores reference data for specifying, as the state, a movement, shape, and inclination of a hand serving as the subject.
  • (Appendix 4)
  • The information processing device according to any one of Appendices 1 to 3, wherein the specifier extracts an image of the subject in the captured image input by the inputter, and specifies the state of the subject depicted by the extracted image of the subject, based on the reference data selected by the selector.
  • (Appendix 5)
  • The information processing device according to any one of Appendices 1 to 4, wherein the acquirer analyzes image data captured by the image capturer, and acquires the information indicating the distance between the camera and the subject.
  • (Appendix 6)
  • The information processing device according to any one of Appendices 1 to 5, further comprises a display that displays the captured image of the subject captured by the camera.
  • (Appendix 7)
  • The information processing device according to Appendix 6, wherein the display displays the state of the subject specified based on the reference data selected by the selector, and the input data corresponding to the state.
  • (Appendix 8)
  • An information processing method comprising:
    • a storing step for storing reference data for specifying a state of a subject captured by a camera and input data corresponding to the state, for each distance between the camera and the subject;
    • an inputting step for inputting a captured image of the subject captured by the camera;
    • an acquirer that acquires information indicating a distance between the camera and the subject;
    • a selecting step for selecting, from among the reference data stored in the storing step, reference data corresponding to the distance indicated by the information acquired in the acquiring step; and
    • a specifying step for referring to the reference data selected in the selecting step, specifying the state of the subject in the captured image input in the inputting step, and specifying data corresponding to the specified state.
    (Appendix 9)
  • A recording medium having a program recorded thereon to allow a computer to function as:
    • an inputter that inputs a captured image of a subject captured by a camera;
    • an acquirer that acquires information indicating a distance between the camera and the subject;
    • a storer that stores reference data for specifying a state of the subject and input data corresponding to the state, for each distance between the camera and the subject;
    • a selector that selects, from among the reference data stored in the storer, reference data corresponding to the distance indicated by the information acquired by the acquirer; and
    • a specifier that refers to the reference data selected by the selector, specifies the state of the subject in the captured image input by the inputter, and specifies data corresponding to the specified state.
  • The present invention is based upon Japanese Patent Application No. 2011-101894 filed on April 28, 2011 . The entirety of the specification, scope of the patent claims, and drawings of Japanese Patent Application No. 2011-101894 is incorporated herein by reference.
  • Industrial Applicability
  • The present invention is suitable for an information processing device, an information processing method, and a recording medium with which data is input by means of a user's gesture.
  • Reference Signs List
  • 1
    Operator
    2
    Camera
    3
    Display
    3W
    Gesture input support window
    3i
    Window
    4
    Storer
    5
    Controller
    41
    Long-distance reference data storer
    42
    Middle-distance reference data storer
    43
    Short-distance reference data storer
    51
    Inputter
    52
    Acquirer
    53
    Selector
    54
    Specifier
    55
    Executer
    100
    Information processing device
    200
    Hand

Claims (9)

  1. An information processing device comprising:
    an inputter that inputs a captured image of a subject captured by a camera;
    an acquirer that acquires information indicating a distance between the camera and the subject;
    a storer that stores reference data for specifying a state of the subject and input data corresponding to the state, for each distance between the camera and the subject;
    a selector that selects, from among the reference data stored in the storer, reference data corresponding to the distance indicated by the information acquired by the acquirer; and
    a specifier that refers to the reference data selected by the selector, specifies the state of the subject in the captured image input by the inputter, and specifies data corresponding to the specified state.
  2. The information processing device according to Claim 1, wherein the storer stores reference data for specifying a larger number of states as the distance between the camera and the subject becomes closer.
  3. The information processing device according to Claim 1 or 2, wherein the storer stores reference data for specifying, as the state, a movement, shape, and inclination of a hand serving as the subject.
  4. The information processing device according to any one of Claims 1 to 3, wherein the specifier extracts an image of the subject in the captured image input by the inputter, and specifies the state of the subject depicted by the extracted image of the subject, based on the reference data selected by the selector.
  5. The information processing device according to any one of Claims 1 to 4, wherein the acquirer analyzes image data captured by the image capturer, and acquires the information indicating the distance between the camera and the subject.
  6. The information processing device according to any one of Claims 1 to 5, further comprising a display that displays the captured image of the subject captured by the camera.
  7. The information processing device according to Claim 6, wherein the display displays the state of the subject specified based on the reference data selected by the selector, and the input data corresponding to the state.
  8. An information processing method comprising:
    a storing step for storing reference data for specifying a state of a subject captured by a camera and input data corresponding to the state, for each distance between the camera and the subject;
    an inputting step for inputting a captured image of the subject captured by the camera;
    an acquirer that acquires information indicating a distance between the camera and the subject;
    a selecting step for selecting, from among the reference data stored in the storing step, reference data corresponding to the distance indicated by the information acquired in the acquiring step; and
    a specifying step for referring to the reference data selected in the selecting step, specifying the state of the subject in the captured image input in the inputting step, and specifying data corresponding to the specified state.
  9. A recording medium having a program recorded thereon that causes a computer to function as:
    an inputter that inputs a captured image of a subject captured by a camera;
    an acquirer that acquires information indicating a distance between the camera and the subject;
    a storer that stores reference data for specifying a state of the subject and input data corresponding to the state, for each distance between the camera and the subject;
    a selector that selects, from among the reference data stored in the storer, reference data corresponding to the distance indicated by the information acquired by the acquirer; and
    a specifier that refers to the reference data selected by the selector, specifies the state of the subject in the captured image input by the inputter, and specifies data corresponding to the specified state.
EP12776429.8A 2011-04-28 2012-04-27 Information processing device, information processing method, and recording medium Not-in-force EP2703949B1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2011101894 2011-04-28
PCT/JP2012/061470 WO2012147960A1 (en) 2011-04-28 2012-04-27 Information processing device, information processing method, and recording medium

Publications (3)

Publication Number Publication Date
EP2703949A1 true EP2703949A1 (en) 2014-03-05
EP2703949A4 EP2703949A4 (en) 2014-10-22
EP2703949B1 EP2703949B1 (en) 2017-10-25

Family

ID=47072471

Family Applications (1)

Application Number Title Priority Date Filing Date
EP12776429.8A Not-in-force EP2703949B1 (en) 2011-04-28 2012-04-27 Information processing device, information processing method, and recording medium

Country Status (7)

Country Link
US (1) US9329673B2 (en)
EP (1) EP2703949B1 (en)
JP (1) JP5709227B2 (en)
KR (1) KR101514168B1 (en)
CN (1) CN103562821B (en)
IL (1) IL229043B (en)
WO (1) WO2012147960A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3690605A4 (en) * 2017-09-26 2021-12-15 BOE Technology Group Co., Ltd. Gesture recognition method and electronic device

Families Citing this family (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10600235B2 (en) 2012-02-23 2020-03-24 Charles D. Huston System and method for capturing and sharing a location based experience
CA2864003C (en) 2012-02-23 2021-06-15 Charles D. Huston System and method for creating an environment and for sharing a location based experience in an environment
JP2014097112A (en) * 2012-11-13 2014-05-29 Takashi Yoshimine Operation device and operation method
JP6110731B2 (en) * 2013-05-31 2017-04-05 株式会社第一興商 Command input recognition system by gesture
JP6344032B2 (en) * 2013-09-26 2018-06-20 富士通株式会社 Gesture input device, gesture input method, and gesture input program
US10057483B2 (en) * 2014-02-12 2018-08-21 Lg Electronics Inc. Mobile terminal and method thereof
JP6341755B2 (en) * 2014-05-26 2018-06-13 キヤノン株式会社 Information processing apparatus, method, program, and recording medium
KR101636460B1 (en) * 2014-11-05 2016-07-05 삼성전자주식회사 Electronic device and method for controlling the same
CN107533366B (en) * 2015-04-20 2020-07-03 三菱电机株式会社 Information display device and information display method
MX2017015362A (en) 2015-06-03 2018-04-20 Koninklijke Philips Nv Safety improvements for uv radiation in aquatic applications.
KR20170068049A (en) * 2015-12-09 2017-06-19 원광대학교산학협력단 Application auto activating apparatus
CN108369451B (en) * 2015-12-18 2021-10-29 索尼公司 Information processing apparatus, information processing method, and computer-readable storage medium
US10437342B2 (en) 2016-12-05 2019-10-08 Youspace, Inc. Calibration systems and methods for depth-based interfaces with disparate fields of view
US10303259B2 (en) * 2017-04-03 2019-05-28 Youspace, Inc. Systems and methods for gesture-based interaction
US10303417B2 (en) 2017-04-03 2019-05-28 Youspace, Inc. Interactive systems for depth-based input
US10325184B2 (en) * 2017-04-12 2019-06-18 Youspace, Inc. Depth-value classification using forests
US11221683B2 (en) * 2019-05-09 2022-01-11 Dell Products, L.P. Graphical user interface (GUI) manipulation using hand gestures over a hovering keyboard
KR102269466B1 (en) * 2019-05-21 2021-06-28 이진우 Method and apparatus for inputting character based on motion recognition
CN112822388B (en) * 2019-11-15 2022-07-22 北京小米移动软件有限公司 Shooting mode triggering method, device, equipment and storage medium
CN113630556A (en) * 2021-09-26 2021-11-09 北京市商汤科技开发有限公司 Focusing method, focusing device, electronic equipment and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2005003948A1 (en) * 2003-07-08 2005-01-13 Sony Computer Entertainment Inc. Control system and control method
WO2005057921A2 (en) * 2003-12-09 2005-06-23 Reactrix Systems, Inc. Self-contained interactive video display system
WO2009035705A1 (en) * 2007-09-14 2009-03-19 Reactrix Systems, Inc. Processing of gesture-based user interactions

Family Cites Families (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5594469A (en) 1995-02-21 1997-01-14 Mitsubishi Electric Information Technology Center America Inc. Hand gesture machine control system
JP3777650B2 (en) * 1995-04-28 2006-05-24 松下電器産業株式会社 Interface equipment
JP2001216069A (en) 2000-02-01 2001-08-10 Toshiba Corp Operation inputting device and direction detecting method
JP2002083302A (en) 2000-09-07 2002-03-22 Sony Corp Information processing device, action recognition processing method, and program storage medium
JP3725460B2 (en) 2000-10-06 2005-12-14 株式会社ソニー・コンピュータエンタテインメント Image processing apparatus, image processing method, recording medium, computer program, semiconductor device
JP2003131785A (en) * 2001-10-22 2003-05-09 Toshiba Corp Interface device, operation control method and program product
JP2004356819A (en) * 2003-05-28 2004-12-16 Sharp Corp Remote control apparatus
JP2004078977A (en) 2003-09-19 2004-03-11 Matsushita Electric Ind Co Ltd Interface device
JP2005190283A (en) 2003-12-26 2005-07-14 Nec Access Technica Ltd Information input apparatus
JP4692159B2 (en) 2004-08-31 2011-06-01 パナソニック電工株式会社 Gesture switch
JP2007219713A (en) 2006-02-15 2007-08-30 Sony Corp Inquiry system, imaging apparatus, inquiry device, information processing method, and program
JP2008146243A (en) 2006-12-07 2008-06-26 Toshiba Corp Information processor, information processing method and program
JP4318056B1 (en) 2008-06-03 2009-08-19 島根県 Image recognition apparatus and operation determination method
US20100251171A1 (en) * 2009-03-31 2010-09-30 Parulski Kenneth A Graphical user interface which adapts to viewing distance
JP5256109B2 (en) 2009-04-23 2013-08-07 株式会社日立製作所 Display device
KR101567785B1 (en) 2009-05-28 2015-11-11 삼성전자주식회사 Apparatus and method for controlling zoom function of a portable terminal
JP5614014B2 (en) 2009-09-04 2014-10-29 ソニー株式会社 Information processing apparatus, display control method, and display control program
US8564534B2 (en) * 2009-10-07 2013-10-22 Microsoft Corporation Human tracking system
JP2011192081A (en) 2010-03-15 2011-09-29 Canon Inc Information processing apparatus and method of controlling the same

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2005003948A1 (en) * 2003-07-08 2005-01-13 Sony Computer Entertainment Inc. Control system and control method
WO2005057921A2 (en) * 2003-12-09 2005-06-23 Reactrix Systems, Inc. Self-contained interactive video display system
WO2009035705A1 (en) * 2007-09-14 2009-03-19 Reactrix Systems, Inc. Processing of gesture-based user interactions

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
None *
See also references of WO2012147960A1 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3690605A4 (en) * 2017-09-26 2021-12-15 BOE Technology Group Co., Ltd. Gesture recognition method and electronic device

Also Published As

Publication number Publication date
IL229043A0 (en) 2013-12-31
WO2012147960A1 (en) 2012-11-01
JPWO2012147960A1 (en) 2014-07-28
CN103562821B (en) 2016-11-09
EP2703949A4 (en) 2014-10-22
KR20140002007A (en) 2014-01-07
US20140043232A1 (en) 2014-02-13
IL229043B (en) 2018-04-30
US9329673B2 (en) 2016-05-03
JP5709227B2 (en) 2015-04-30
KR101514168B1 (en) 2015-04-21
EP2703949B1 (en) 2017-10-25
CN103562821A (en) 2014-02-05

Similar Documents

Publication Publication Date Title
EP2703949B1 (en) Information processing device, information processing method, and recording medium
US9367732B2 (en) Information processing device, information processing method, and recording medium
EP3644165B1 (en) Information processing device, information processing method, and recording medium
US11048333B2 (en) System and method for close-range movement tracking
US9910498B2 (en) System and method for close-range movement tracking
CN106845335B (en) Gesture recognition method and device for virtual reality equipment and virtual reality equipment
US9734393B2 (en) Gesture-based control system
US11782514B2 (en) Wearable device and control method thereof, gesture recognition method, and control system
US9904372B2 (en) Method by which eyeglass-type display device recognizes and inputs movement
US10372229B2 (en) Information processing system, information processing apparatus, control method, and program
KR101281461B1 (en) Multi-touch input method and system using image analysis
US20210109600A1 (en) Methods and apparatuses for controlling a system via a sensor
CN109101173A (en) Screen layout's control method, device, equipment and computer readable storage medium
JP2015184906A (en) Skin color detection condition determination device, skin color detection condition determination method and skin color detection condition determination computer program
JP5558899B2 (en) Information processing apparatus, processing method thereof, and program
KR20210003515A (en) Augmented Reality Implementation Device Supporting Interactive Mode
US20230305634A1 (en) Gesture recognition apparatus, head-mounted-type display apparatus, gesture recognition method, and non-transitory computer readable medium
CA3229530A1 (en) Electronic apparatus and program
JPWO2023031988A5 (en)

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20131120

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

DAX Request for extension of the european patent (deleted)
RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: NEC SOLUTION INNOVATORS, LTD.

A4 Supplementary search report drawn up and despatched

Effective date: 20140919

RIC1 Information provided on ipc code assigned before grant

Ipc: G06F 3/01 20060101AFI20140915BHEP

Ipc: G06F 3/03 20060101ALI20140915BHEP

17Q First examination report despatched

Effective date: 20160412

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

INTG Intention to grant announced

Effective date: 20170509

R17C First examination report despatched (corrected)

Effective date: 20160412

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

REG Reference to a national code

Ref country code: AT

Ref legal event code: REF

Ref document number: 940483

Country of ref document: AT

Kind code of ref document: T

Effective date: 20171115

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: DE

Ref legal event code: R096

Ref document number: 602012038970

Country of ref document: DE

REG Reference to a national code

Ref country code: NL

Ref legal event code: MP

Effective date: 20171025

REG Reference to a national code

Ref country code: LT

Ref legal event code: MG4D

REG Reference to a national code

Ref country code: AT

Ref legal event code: MK05

Ref document number: 940483

Country of ref document: AT

Kind code of ref document: T

Effective date: 20171025

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: NL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: NO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180125

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

Ref country code: SE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

Ref country code: LT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: HR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

Ref country code: BG

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180125

Ref country code: LV

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

Ref country code: IS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180225

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180126

Ref country code: RS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

REG Reference to a national code

Ref country code: DE

Ref legal event code: R097

Ref document number: 602012038970

Country of ref document: DE

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

Ref country code: EE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

Ref country code: SK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

Ref country code: CZ

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SM

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

Ref country code: IT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

Ref country code: RO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

Ref country code: PL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

26N No opposition filed

Effective date: 20180726

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

Ref country code: SI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

REG Reference to a national code

Ref country code: BE

Ref legal event code: MM

Effective date: 20180430

GBPC Gb: european patent ceased through non-payment of renewal fee

Effective date: 20180427

REG Reference to a national code

Ref country code: IE

Ref legal event code: MM4A

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20180427

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20180430

Ref country code: BE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20180430

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20180430

Ref country code: GB

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20180427

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20180427

Ref country code: FR

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20180430

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: FI

Payment date: 20190409

Year of fee payment: 8

Ref country code: DE

Payment date: 20190416

Year of fee payment: 8

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MT

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20180427

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: TR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: HU

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO

Effective date: 20120427

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MK

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20171025

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: AL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20171025

REG Reference to a national code

Ref country code: DE

Ref legal event code: R119

Ref document number: 602012038970

Country of ref document: DE

REG Reference to a national code

Ref country code: FI

Ref legal event code: MAE

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: FI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20200427

Ref country code: DE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20201103