WO2004010370A2 - Hand-held computer interactive device - Google Patents

Hand-held computer interactive device Download PDF

Info

Publication number
WO2004010370A2
WO2004010370A2 PCT/US2003/021769 US0321769W WO2004010370A2 WO 2004010370 A2 WO2004010370 A2 WO 2004010370A2 US 0321769 W US0321769 W US 0321769W WO 2004010370 A2 WO2004010370 A2 WO 2004010370A2
Authority
WO
WIPO (PCT)
Prior art keywords
input device
user
central body
hand
computing
Prior art date
Application number
PCT/US2003/021769
Other languages
French (fr)
Other versions
WO2004010370A3 (en
Inventor
Tyler Jon Daniel
Original Assignee
Sony Computer Entertainment Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sony Computer Entertainment Inc. filed Critical Sony Computer Entertainment Inc.
Priority to EP03765543A priority Critical patent/EP1523725B1/en
Priority to AT03765543T priority patent/ATE480814T1/en
Priority to KR1020057000857A priority patent/KR100742029B1/en
Priority to JP2005505516A priority patent/JP4044114B2/en
Priority to DE60334101T priority patent/DE60334101D1/en
Priority to AU2003249062A priority patent/AU2003249062A1/en
Publication of WO2004010370A2 publication Critical patent/WO2004010370A2/en
Publication of WO2004010370A3 publication Critical patent/WO2004010370A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/014Hand-worn input/output arrangements, e.g. data gloves
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/213Input arrangements for video game devices characterised by their sensors, purposes or types comprising photodetecting means, e.g. cameras, photodiodes or infrared cells
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/218Input arrangements for video game devices characterised by their sensors, purposes or types using pressure sensors, e.g. generating a signal proportional to the pressure applied by the player
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/22Setup operations, e.g. calibration, key configuration or button assignment
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/23Input arrangements for video game devices for interfacing with the game device, e.g. specific interfaces between game controller and console
    • A63F13/235Input arrangements for video game devices for interfacing with the game device, e.g. specific interfaces between game controller and console using a wireless connection, e.g. infrared or piconet
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/25Output arrangements for video game devices
    • A63F13/28Output arrangements for video game devices responding to control signals received from the game device for affecting ambient conditions, e.g. for vibrating players' seats, activating scent dispensers or affecting temperature or light
    • A63F13/285Generating tactile feedback signals via the game input device, e.g. force feedback
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/40Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
    • A63F13/42Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
    • A63F13/428Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle involving motion or position input signals, e.g. signals representing the rotation of an input controller or a player's arm motions sensed by accelerometers or gyroscopes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/016Input arrangements with force or tactile feedback as computer generated output to the user
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/212Input arrangements for video game devices characterised by their sensors, purposes or types using sensors worn by the player, e.g. for measuring heart beat or leg activity
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/10Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
    • A63F2300/1018Calibration; Key and button assignment
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/10Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
    • A63F2300/1025Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals details of the interface with the game device, e.g. USB version detection
    • A63F2300/1031Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals details of the interface with the game device, e.g. USB version detection using a wireless connection, e.g. Bluetooth, infrared connections
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/10Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
    • A63F2300/1037Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals being specially adapted for converting control signals received from the game device into a haptic signal, e.g. using force feedback
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/10Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
    • A63F2300/1056Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals involving pressure sensitive buttons
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/10Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
    • A63F2300/1087Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals comprising photodetecting means, e.g. a camera
    • A63F2300/1093Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals comprising photodetecting means, e.g. a camera using visible light

Definitions

  • This invention relates generally to input devices for computer systems, and more particularly to an input device that is held by a user.
  • an input device configured to be held by a user.
  • the input device includes a central body.
  • a response and communications system is incorporated with the central body.
  • the response and communication system enables communication between the input device and a computing device.
  • a plurality of protrusions extending from the central body are included.
  • the central body and the plurality of protrusions are configured to capture a user's hand movements.
  • the input device is configured to be held by the user.
  • an input detection program executed at a computing device includes program instructions for determining when to trigger input commands of a main program and tactile feedback responses to an external input device configured to be held by a user.
  • the input detection program includes program instmctions for detecting a signal from the external input device. The signal is generated by causing a physical change to the external input device.
  • Program instructions for triggering an input command at the program run on the computing device are provided. The triggering results from the detected signal from the external input object.
  • program instructions for generating a tactile feedback signal at the external input device in response to the detected signal are provided.
  • a computing system enabling tactile feedback to a user.
  • the computing system includes a computing unit.
  • the computing unit includes circuitry configured to receive and transmit commands.
  • a display monitor in communication with the computing unit is included.
  • An input device configured to be held in a user's hand is included.
  • the input device is enabled to provide tactile feedback to a user in response to hand movements of the user.
  • a computing system is provided.
  • the computing system includes a computing unit.
  • the computing unit includes circuitry configured to receive and transmit commands.
  • a display monitor in communication with the computing unit is included.
  • a hand-held input device is provided. The input device is enabled to manipulate objects displayed on the display monitor.
  • a computing system capable of providing vibrational feedback to a user.
  • the computing system includes means for capturing hand movements of a user through an external input device configured to be held by the user. Means for transmitting the captured hand movements to a computing device and means for transmitting commands to the external input device in response to the transmitted hand movements of the user are included. Means for providing tactile feedback to the user in response to the transmitted commands to the external input device are provided. Means for displaying graphic images associated with the commands received from the external input device are included.
  • a method for communicating commands to a computer system from an external input device is provided. The method initiates with providing a receiver configured to communicate with an external input device configured to be held by a user. Then, hand movements of the user are captured through the deformation of the external input device. Next, the hand movements are translated to a command to be communicated to the receiver for causing input control at the receiver.
  • a method for communicating commands to a computer system from a hand-held input device begins with providing a game console having a receiver configured to communicate with a handheld input device. Then, hand movements of a user are captured through the deformation of the input device by a user's hand. Next, the hand movements are translated to a command to be communicated to the receiver. Then, in response to the command being communicated to the receiver, the user's hand is stimulated with tactile feedback provided through the input device. The user's hand is stimulated by controlling a hardness and a pattern of a vibration applied to the user's hand.
  • computer program instmctions for managing computer input commands are provided.
  • the computer program instmctions are embodied in a computer readable media.
  • the computer program instmctions include program instmctions for capturing hand movements applied to a wireless hand-held input device. Program instructions for receiving specific commands associated with the captured hand movements are included. Program instructions for applying the specific commands to a program being directed by the hand-held input device are provided also.
  • an input device is provided.
  • the input device includes a central body and at least one protrusion extending from the central body. The central body and the at least one protrusion are configured to capture a user's finger movements.
  • a response and communications system is disposed within the input device.
  • an input device in still yet another embodiment, includes a central body. At least one protrasion extending from the central body is included. The central body and the at least one protrasion are configured to capture a user's hand movements.
  • a communications system incorporated with the central body enables communication between the input device and a computing device.
  • Figure 1 is a simplified schematic diagram of the external input device configured to be held by a user to provide input to a receiver in accordance with one embodiment of the invention.
  • Figure 2 is a simplified schematic diagram of a pair of input devices configured to communicate with a computing device in communication with a display monitor in accordance with one embodiment of the invention.
  • Figure 3 is a bottom view of the external input device illustrating the embedded electronics in accordance with one embodiment of the invention.
  • Figure 4 is a bottom perspective view of the input device in an open position in accordance with one embodiment of the invention.
  • Figure 5A is a schematic diagram of the input device being held by a user while the input device is in an open position in accordance with one embodiment of the invention.
  • Figure 5B is a schematic diagram of the input device being held by a user while the input device is in a closed position in accordance with one embodiment of the invention.
  • Figure 6 is a top perspective view of an input device in accordance with one embodiment of the invention.
  • Figure 7 is a bottom perspective view of an alternative embodiment of the input device of Figure 4.
  • Figure 8 is a flowchart diagram of the method operations of method for communicating commands to a computer system from an external input device in accordance with one embodiment of the invention.
  • the embodiments of the present invention provide a user input device that is held by a user, as opposed to being worn like a glove.
  • the disclosed device is designed to capture data regarding a user's hand movements.
  • hand movements include finger movements, e.g., bending of fingers and thumbs, flexing of fingers and thumbs, etc.
  • Hand movements also includes the movement of the whole hand by the arm.
  • the input device may be used to capture finger and thumb movements and at the same time a computing device can track the hand movements through space so that a click and drag operation can be accomplished with the input device as described in more detail below.
  • the captured data is in the form of digital data, which allows the input device to be used as an interactive device with a computer program.
  • the movement of particular fingers is thus transferred to particular protrusions of the input device. These movements are translated and then transferred to a computing system to initiate a command, respond to a query, provide a command, maneuver objects in an interactive video game, etc.
  • the input device provides vibrational feedback to the user so that the user can experience a sense of touch associated with the commands being transferred to the computing system.
  • FIG. 1 is a simplified schematic diagram of the external input device configured to be held by a user to provide input to a receiver in accordance with one embodiment of the invention.
  • External input device 100 is configured to be held within the palm of a user's hand 102.
  • electronics embedded in the input device sense the change in position of the protrusions and generate commands transmitted to receiver 106.
  • Receiver 106 is in communication with device 108 which is controlled by the signals transmitted to the receiver from input device 100.
  • device 108 can be a computer, or a game console which can be manufactured by any company.
  • the game console may be a PLAYSTATIONTM game console.
  • the embedded electronics may generate short range radio signals. These signals can be processed in accordance with public or proprietary processing circuitry and/or software. For exemplary purposes, communication of the radio signals can be done using standards such as BLUETOOTH, or other suitable wireless technology (e.g., such as IEEE 802.11). While it is preferred to employ a technology not requiring a line of sight, the embodiments described herein can also be applied to technologies requiring a line of sight such as infrared signals.
  • FIG. 2 is a simplified schematic diagram of a pair of input devices configured to communicate with a computing device in communication with a display monitor in accordance with one embodiment of the invention.
  • External input devices 100 are configured to fit within the palm of each of a user's hands 102.
  • each of input devices 100 is enabled to communicate with computing device 108 through receiver 106. While receiver 106 is shown outside of the frame of computing device 108, it should be appreciated that the receiver can be integrated into the computing device in one embodiment of the invention.
  • Input device 100 transmits and receives signals through a response and communication system embedded within the input device in one embodiment.
  • Computing device 108 is in communication with display monitor 110.
  • Computing device 108 transmits digital data to display monitor 110 so that the digital data can be viewed.
  • Display monitor 110 may display text 110a, menu 110b and/or graphics 110c. In one embodiment, where computing device 108 is a game console, display monitor 110 displays graphics or a user interface associated with a game being played.
  • Each of input devices 100 is configured to generate commands 104 in response to a user's hand movements.
  • Commands 104 are captured by computing device 108 through receiver 106.
  • the captured data is in the form of digital data, which enables input devices 100 to be used as an interactive device with a computer program.
  • the movement of a particular finger or fingers is thus transferred to particular protrusions of input device 100.
  • these movements are then transferred to computing device 108 to initiate a command, respond to a query, maneuver objects in an interactive video game, etc.
  • input devices 100 are configured to fit comfortably within a user's hand. Thus different sizes can be made available based upon a user's age, gender, etc.
  • input devices 100 are configured to remain in a user's hand even if the user is holding his hands in a typing position. That is the natural bend of the user's fingers will support the input device in order to prevent the input device form falling out of the user's hand.
  • finger straps such as VELCRO straps, may be incorporated to secure the input device to the user's hand in one embodiment of the invention.
  • the material of the protrusions is selected such that, the material does not easily slide off of the fingers while a user is holding the input device.
  • FIG. 3 is a bottom view of the external input device illustrating the embedded electronics in accordance with one embodiment of the invention.
  • External input device 100 is composed of central body 116 and protrasions 112a-l 12e.
  • Protrusions 112a-l 12e may also be referred to as legs.
  • Each of protrasions 112a-112e are available for each finger and thumb of a user's hand.
  • Protrasions 112a-l 12e are configured to bend as each respective finger or thumb bends.
  • the configuration of external input device 100 is analogous to a spider, where protrasions 112a-l 12e are the legs in one embodiment.
  • Central body 116 is configured to rest in the palm area of a user's hand.
  • central body 116 includes electronics for translating finger and thumb movements and provides an anchor point for each of protrasions 112a-l 12e.
  • the electronics for translating finger and thumb movements may be disposed anywhere within input device 100 or even attached to the input device.
  • transmitter/receiver 118 contains circuitry for transmitting signals to a computing device, such as device 108 of Figure 2.
  • Transmitter/receiver 118 also contains circuitry for receiving signals from an external device, such as computing device 108 of Figure
  • External input device 100 of Figure 3 is enabled to provide low latency tactile/haptic feedback to the fingers, thumb and palm of a user's hand through vibrators 120a-120g positioned within the external input device.
  • the haptic/tactile feedback is provided through a TOUCHENGINETM tactile/haptic interface component consisting of an actuator and controlling unit.
  • any tactile/haptic interface component either manufactured especially for the present embodiment, or one that can be engineered from outside supplier hardware and/or software may be used.
  • each protrusion 112a-112e and central body 116 include at least one vibrator for providing low latency haptic/tactile feedback.
  • Each protrusion is enabled to measure pressure at: (a) the fingertip, (b) calculate the flex of the finger, and (c) stimulate the finger with patterned vibrations.
  • the surface of the protrasion that contacts the user's hand is, in one embodiment, rubberized to give the user a secure and comfortable grip.
  • the number of protrusions can be varied so that, for example, only the thumb and first two fingers are sensed and the user could grip the bottom part of the input device with the remaining two fingers. It should be appreciated that since the hand makes up a large part of expressive ability, it is well suited to communicate information to a computing device.
  • the protrasions may be composed of any suitable material that can be deformed and subsequently returned to a natural position once the force causing the deformation is released.
  • the terms tactile and haptic are interchangeable as both terms refer to a sense of touch.
  • Line extensions 114a-114e are configured to sense the flex or degree of bend of a finger or thumb.
  • the flex or degree of bend is communicated to transmitter/receiver 118 where the signal is translated into a command that is subsequently transmitted to a computing device in one embodiment.
  • the signal can be communicated directly to transmitter/receiver 118 without being translated.
  • Input device 100 is constructed of lightweight material in one embodiment.
  • central body 116 is composed polystyrene, rubber foam, plastic or some other suitable lightweight material capable of holding the above mentioned electronics and acting as an anchor for protrasions 112a-112e.
  • Protrusions 112a-l 12e are composed of a material that can be gripped by a user's hand.
  • the material for protrasions 112a-112e offers enough resistance so that input device 100 can be held comfortably and without much effort while a user's hand.
  • the material allows protrasions 112a-112e to be squeezed into a closed position, as illustrated with reference to Figure 5B, and then returned to an open position, once the force causing input device 100 to close is removed.
  • Electronics such as receiver/transmitter 118, line extensions 114a-114e, and vibrators 120a-120g are embedded within central body 116 and protrasions 112a-112e.
  • pressure sensors are included at the fingertip positions of protrasions 112a-l 12e to monitor fingertip pressure.
  • the above described electronics are included in the response and communications system, also referred to as the communications system, of input device 100.
  • power is supplied for the embedded electronics of the response and communications system through a battery located within central body 116.
  • FIG. 4 is a bottom perspective view of the input device in an open position in accordance with one embodiment of the invention.
  • Central body 116 is configured to fit inside a user's palm, while protrusions 112a-l 12e conform to a user's fingers and thumb.
  • the input device is of a lightweight construction that allows a user as much freedom of movement as possible.
  • the electronics embedded in central body 116 allow for unrestricted placement of the input device during operation as there is no line-of-sight requirement. That is, a short range radio technology, rather than an infrared signal, is used to communicate through the input device and the computing device.
  • each of protrusions 112a-l 12e include a sensor configured to detect the pressure exerted by the fingertips of a user.
  • Figure 5A is a schematic diagram of the input device being held by a user while the input device is in an open position in accordance with one embodiment of the invention.
  • Central body 116 of input device 100 fits within the palm of a user's hand 102.
  • Each of protrasions 112a-112e are configured to accommodate a finger or thumb of the user's hand 102.
  • Figure 5B is a schematic diagram of the input device being held by a user while the input device is in a closed position in accordance with one embodiment of the invention.
  • input device 100 is brought to a closed position as the user squeezes protrusions 112a-112e while holding the input device in the user's hand 102.
  • Central body 116 remains against the user's palm while input device 100 is closed.
  • a latching mechanism allows a user to latch input device 100 in a closed position, h the latched position, protrusions 112a- 112e essentially wrap around central body 116 to define a cylinder.
  • input device 100 can easily fit inside a pocket or can conveniently be stored by the user as the area for storage is minimized in the latched position.
  • a button can be depressed or moved to release the latching mechanism so that the input device opens, i.e., the protrusions are released from the latched position and "open up".
  • the input device 100 may be utilized for numerous input functions. Some exemplary functions are described below.
  • the input device can act as a virtual keyboard in one embodiment.
  • low latency vibrators such as TOUCHENGINETM vibrators, mounted on each finger to give a tactile "click" when moving from one virtual key to another. A different click can be used to indicate when a key is pressed (via pressure sensors on fingertips), and a continuous vibration can be used whenever the finger is out of the "home" row.
  • Positional left/right tracking of the hand could be used to switch between different zones of the virtual keyboard if, for example, only a single hand is being used.
  • Such a virtual keyboard has the advantage that any surface can be typed on, in any orientation. For instance, a user might be standing and can type on the sides of their legs.
  • the input device can act as a virtual piano with feedback similar to the keyboard embodiment described above.
  • a camera pointed at the user could show live video of the user combined with a three dimensional computer graphic piano responding to the users hands.
  • the input device can act as a control for a game in one embodiment.
  • a folded sheet of plastic can be used to simulate the cockpit controls (e.g., the instrument panel can be printed on the plastic or other material), and the user can be wearing the input devices on each hand to enable piloting of the airplane or other vehicle or device.
  • a display monitor connected to a computer, or game unit allows the user to view the interactions.
  • the user can touch the four comers of the sheet, which calibrates the positional tracking.
  • the four comers of the sheet can be marked with red crosses to indicate the areas to be touched.
  • a combination of fingertip pressure, positional tracking, and tactile feedback can be used to interact with the game.
  • the game may also include a plastic yoke or joystick that contains no electronics and only serves to constrain or steady the users hands.
  • the game detects the user gripping the controls by a combination of fingertip pressure and finger flex information and then tracks the motion and/or orientation of the hand to interpret the manipulation of the physical controls.
  • the plastic yoke might be designed with the shape of the input device in mind and include, for example, depressions for the central body of the input device to fit in.
  • the degree of pressure applied when a user squeezes their fingers together is correlated to a tightness of a grip exerted on an object in a video game application being displayed on a monitor.
  • FIG. 6 is a top perspective view of an input device in accordance with one embodiment of the invention.
  • Input device 100 consists of central body 116 and protrasions
  • central body 116 includes embedded electronics for transmitting signals to and receiving signals from a computing device such as a game console.
  • Protrusions 112a-112e include extension lines for determining the degree of bend of a user's fingers. Pressure sensors may also be provided to obtain fingertip pressure.
  • Vibrational feedback is provided to a user through vibrators located in protrusions 112a-112e and central body 116.
  • the vibrators are configured to control the hardness and the pattern of the vibration to provide the user with tactile feedback. It should be appreciated that by varying the hardness and the pattern of the vibration the sensation perceived by the user may differ. For example, a user can be provided a sense of gripping a steering wheel, typing, striking a piano key, etc.
  • FIG. 7 is a bottom perspective view of an alternative embodiment of the input device of Figure 4.
  • Input device 100 includes central body 116 and protrusions 112a, 112b and 112c.
  • the thumb, forefinger and middle finger of the user are accommodated by protrusions 112a, 112b and 112c.
  • Button 130 on the bottom surface of central body 116 releases protrusions 112a, 112b and 112c from a closed position. That is, depressing button 130 releases the latching mechanism holding protrasions 112a, 112b and 112c in a closed position.
  • button 130 may be incorporated into the embodiments of the input device having protrasions for each finger.
  • latching mechanism may be spring activated to release to an open position upon depression of button 130.
  • input device 100 may be configured as having one protrusion.
  • one protrusion for an index fmger extending from central body 116 may be provided.
  • the movement of the index finger is captured, although the single protrusion can accommodate any one finger or thumb.
  • Vibrators may also be located in the single protrasion as well as the central core to provide tactile feedback to the user's hand through the finger and the user's palm.
  • Figure 8 is a flowchart diagram of the method operations of method for communicating commands to a computer system from an external input device in accordance with one embodiment of the invention. The method initiates with operation 150 where a receiver is provided.
  • the receiver is configured to communicate with an external input device configured to be held within a user's palm.
  • the receiver is integrated with a computing device, such as a game console.
  • An exemplary input device is shown with reference to Figures 3-7.
  • the method then advances to operation 152 where the hand movements of a user are captured through deformation of the input device.
  • the user applies and releases pressure to the input device.
  • Electronics embedded within the input device capture the change in position of the protrasions, which yields information on the degree of flex of the fingers as well as the pressure at the fingertips.
  • the method of Figure 8 then proceeds to operation 154 where the captured hand movements are translated to a command to be communicated to a receiver.
  • the movements are translated and then transferred to a computing system to initiate a command, respond to a query, maneuver objects in an interactive video game, etc.
  • the method then moves to operation 156 where the user's hand is stimulated with tactile feedback provided through the input device.
  • tactile feedback provided through the input device.
  • low latency vibrators as illustrated in Figure 3 provide the tactile feedback.
  • the hardness and the pattern of the vibration is controlled to provide a certain feel to the user. That is, the type of vibrational feedback is dependent on the application being executed.
  • the invention may be a computing system capable of providing vibrational feedback to a user consisting of various components defined in terms of the functions of each of the components.
  • the computing system can include a means for capturing hand movements.
  • the means for capturing hand movements of a user are provided through an external input device configured to be held by the user, such as the hand-held input device illustrated in Figures 3-7.
  • Means for transmitting the captured hand movements are included.
  • the means for transmitting the captured hand movements to a computing device may be the embedded electronics discussed above with reference to Figure 3.
  • Means for transmitting commands to the external input device in response to the transmitted hand movements of the user are provided.
  • a transmitter associated with a computing device transmits the commands to the external input device.
  • Means for providing tactile feedback to the user in response to the transmitted commands to the external input device are provided by the vibrators illustrated in Figure 3.
  • Means for displaying graphic images associated with the commands received from the external input device are provided by a display monitor in communication with a computing device.
  • a hand-held input device is provided.
  • the hand-held input device is configured to be non-surface relational. That is, the input device does not require movement along a surface to function, as is required for a mouse.
  • a user's finger movements generate signals for controlling objects displayed on a monitor associated with a computing unit that receives the generated signals.
  • the objects may be associated with a video game application where the computing unit is a video game console.
  • the computing unit can be a game console or a server accessed through a distributed network, such as the Internet.
  • the invention may employ various computer-implemented operations involving data stored in computer systems. These operations are those requiring physical manipulation of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated. Further, the manipulations performed are often referred to in terms, such as producing, identifying, determining, or comparing.
  • the invention also relates to a device or an apparatus for performing these operations.
  • the apparatus may be specially constructed for the required purposes, or it may be a general purpose computer selectively activated or configured by a computer program stored in the computer.
  • various general purpose machines may be used with computer programs written in accordance with the teachings herein, or it may be more convenient to construct a more specialized apparatus to perform the required operations.
  • the invention can also be embodied as computer readable code on a computer readable medium.
  • the computer readable medium is any data storage device that can store data which can be thereafter be read by a computer system. Examples of the computer readable medium include hard drives, network attached storage (NAS), read-only memory, random-access memory, CD-ROMs, CD-Rs, CD-RWs, magnetic tapes, and other optical and non-optical data storage devices.
  • the computer readable medium can also be distributed over a network coupled computer systems so that the computer readable code is stored and executed in a distributed fashion.

Abstract

In one embodiment an input devic (100)e is provided. The input device (100) includes a central body. A response and communications system is included in the input device. The response and communication system enables communication between the input device (100) and a computing device (110). At least one protrusion extending from the central body are included. The central body and the least one protrusion are configured to capture a user's hand movements. The input device (100) is configured to be held by the user. An input detection program, a computing system, computer program instructions and a method for communicating commands to a computer system from an external input device are provided.

Description

HAND-HELD COMPUTER INTERACTIVE DEVICE by Inventor: Tyler Jon Daniel
BACKGROUND OF THE INVENTION 1. Field of the Invention
This invention relates generally to input devices for computer systems, and more particularly to an input device that is held by a user.
2. Description of the Related Art
There has been a trend to transition the personal computer from a work environment to an entertainment environment. As part of the trend, alternative input devices are being evaluated. The alternative input devices are geared toward a "living room interface" focusing more on entertainment rather than a work environment. For example, gloves that capture hand motions have been evaluated for virtual reality, telerobotics, medicine, video games, music generation, etc. applications. One of the shortcomings with the gloves is that manufacturers are finding that consumers do not like to wear a glove for entertainment applications. The gloves require the user to constantly remove the gloves and put the gloves back on, especially if the user is moving between multiple applications. For example, where the user may have to perform typing or some other function requiring a sense of feeling in the hands, consumers tend to remove the gloves. In addition, the data glove hinders the movement of the hand and has substantial weight. Thus, gloves have not been embraced by users for entertainment applications.
Another solution being offered as an alternative input device is computer vision systems. However, computer vision systems are limited in their applications for the entertainment environment. Computer vision systems do not provide for tactile feedback and are limited by lighting conditions. Furthermore, computer vision systems have the problem of "address", i.e., the question of whether the gesture is intended for the computer. Also, parameters are inferred and hence are potentially inaccurate. In addition, the input device must be within view of a camera. Still another disadvantage of computer vision systems is that some gestures are self-occluding and are difficult to interpret. Thus, there is a need to solve the problems of the prior art to provide an input device ' convenient for a user to use that provides tactile/haptic feedback and does not have to be worn by a user.
SUMMARY OF THE INVENTION Broadly speaking, the present invention fills these needs by providing an input device configured to be held by a user. It should be appreciated that the present invention can be implemented in numerous ways, including as an apparatus, a method, a process, a system, program instructions or a device. Several inventive embodiments of the present invention are described below. In one embodiment an input device is provided. The input device includes a central body. A response and communications system is incorporated with the central body. The response and communication system enables communication between the input device and a computing device. A plurality of protrusions extending from the central body are included. The central body and the plurality of protrusions are configured to capture a user's hand movements. The input device is configured to be held by the user.
In another embodiment, an input detection program executed at a computing device is provided. The input detection program includes program instructions for determining when to trigger input commands of a main program and tactile feedback responses to an external input device configured to be held by a user. The input detection program includes program instmctions for detecting a signal from the external input device. The signal is generated by causing a physical change to the external input device. Program instructions for triggering an input command at the program run on the computing device are provided. The triggering results from the detected signal from the external input object. In one embodiment program instructions for generating a tactile feedback signal at the external input device in response to the detected signal are provided.
In yet another embodiment, a computing system enabling tactile feedback to a user is provided. The computing system includes a computing unit. The computing unit includes circuitry configured to receive and transmit commands. A display monitor in communication with the computing unit is included. An input device configured to be held in a user's hand is included. The input device is enabled to provide tactile feedback to a user in response to hand movements of the user. In still yet another embodiment, a computing system is provided. The computing system includes a computing unit. The computing unit includes circuitry configured to receive and transmit commands. A display monitor in communication with the computing unit is included. A hand-held input device is provided. The input device is enabled to manipulate objects displayed on the display monitor.
In another embodiment, a computing system capable of providing vibrational feedback to a user is provided. The computing system includes means for capturing hand movements of a user through an external input device configured to be held by the user. Means for transmitting the captured hand movements to a computing device and means for transmitting commands to the external input device in response to the transmitted hand movements of the user are included. Means for providing tactile feedback to the user in response to the transmitted commands to the external input device are provided. Means for displaying graphic images associated with the commands received from the external input device are included. In yet another embodiment, a method for communicating commands to a computer system from an external input device is provided. The method initiates with providing a receiver configured to communicate with an external input device configured to be held by a user. Then, hand movements of the user are captured through the deformation of the external input device. Next, the hand movements are translated to a command to be communicated to the receiver for causing input control at the receiver.
In still yet another embodiment, a method for communicating commands to a computer system from a hand-held input device is provided. The method initiates with providing a game console having a receiver configured to communicate with a handheld input device. Then, hand movements of a user are captured through the deformation of the input device by a user's hand. Next, the hand movements are translated to a command to be communicated to the receiver. Then, in response to the command being communicated to the receiver, the user's hand is stimulated with tactile feedback provided through the input device. The user's hand is stimulated by controlling a hardness and a pattern of a vibration applied to the user's hand. In another embodiment, computer program instmctions for managing computer input commands are provided. The computer program instmctions are embodied in a computer readable media. The computer program instmctions include program instmctions for capturing hand movements applied to a wireless hand-held input device. Program instructions for receiving specific commands associated with the captured hand movements are included. Program instructions for applying the specific commands to a program being directed by the hand-held input device are provided also. In yet another embodiment, an input device is provided. The input device includes a central body and at least one protrusion extending from the central body. The central body and the at least one protrusion are configured to capture a user's finger movements. A response and communications system is disposed within the input device.
In still yet another embodiment, an input device is provided. The input device includes a central body. At least one protrasion extending from the central body is included. The central body and the at least one protrasion are configured to capture a user's hand movements. A communications system incorporated with the central body enables communication between the input device and a computing device.
Other aspects and advantages of the invention will become apparent from the following detailed description, taken in conjunction with the accompanying drawings, illustrating by way of example the principles of the invention.
BRIEF DESCRIPTION OF THE DRAWINGS
The invention, together with further advantages thereof, may best be understood by reference to the following description taken in conjunction with the accompanying drawings in which:
Figure 1 is a simplified schematic diagram of the external input device configured to be held by a user to provide input to a receiver in accordance with one embodiment of the invention.
Figure 2 is a simplified schematic diagram of a pair of input devices configured to communicate with a computing device in communication with a display monitor in accordance with one embodiment of the invention.
Figure 3 is a bottom view of the external input device illustrating the embedded electronics in accordance with one embodiment of the invention. Figure 4 is a bottom perspective view of the input device in an open position in accordance with one embodiment of the invention.
Figure 5A is a schematic diagram of the input device being held by a user while the input device is in an open position in accordance with one embodiment of the invention.
Figure 5B is a schematic diagram of the input device being held by a user while the input device is in a closed position in accordance with one embodiment of the invention.
Figure 6 is a top perspective view of an input device in accordance with one embodiment of the invention.
Figure 7 is a bottom perspective view of an alternative embodiment of the input device of Figure 4.
Figure 8 is a flowchart diagram of the method operations of method for communicating commands to a computer system from an external input device in accordance with one embodiment of the invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS An invention is disclosed for an input device that is held by a user and provides tactile feedback to a user to simulate a hand-based activity. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. It will be apparent, however, to one skilled in the art that the present invention may be practiced without some or all of these specific details. In other instances, well known process steps have not been described in detail in order not to unnecessarily obscure the present invention.
The embodiments of the present invention provide a user input device that is held by a user, as opposed to being worn like a glove. The disclosed device is designed to capture data regarding a user's hand movements. As used herein, hand movements include finger movements, e.g., bending of fingers and thumbs, flexing of fingers and thumbs, etc. Hand movements also includes the movement of the whole hand by the arm. Thus, the input device may be used to capture finger and thumb movements and at the same time a computing device can track the hand movements through space so that a click and drag operation can be accomplished with the input device as described in more detail below. The captured data is in the form of digital data, which allows the input device to be used as an interactive device with a computer program. The movement of particular fingers is thus transferred to particular protrusions of the input device. These movements are translated and then transferred to a computing system to initiate a command, respond to a query, provide a command, maneuver objects in an interactive video game, etc. In addition, the input device provides vibrational feedback to the user so that the user can experience a sense of touch associated with the commands being transferred to the computing system.
Figure 1 is a simplified schematic diagram of the external input device configured to be held by a user to provide input to a receiver in accordance with one embodiment of the invention. External input device 100 is configured to be held within the palm of a user's hand 102. As the user squeezes and relaxes the pressure applied on the protrusions of input device 100, through pressure applied by the user's fingers and thumb, electronics embedded in the input device sense the change in position of the protrusions and generate commands transmitted to receiver 106. Receiver 106 is in communication with device 108 which is controlled by the signals transmitted to the receiver from input device 100. In one embodiment device 108 can be a computer, or a game console which can be manufactured by any company. For purposes of example, the game console may be a PLAYSTATION™ game console. It will be apparent to one skilled in the art that the embedded electronics may generate short range radio signals. These signals can be processed in accordance with public or proprietary processing circuitry and/or software. For exemplary purposes, communication of the radio signals can be done using standards such as BLUETOOTH, or other suitable wireless technology (e.g., such as IEEE 802.11). While it is preferred to employ a technology not requiring a line of sight, the embodiments described herein can also be applied to technologies requiring a line of sight such as infrared signals.
Figure 2 is a simplified schematic diagram of a pair of input devices configured to communicate with a computing device in communication with a display monitor in accordance with one embodiment of the invention. External input devices 100 are configured to fit within the palm of each of a user's hands 102. As shown, each of input devices 100 is enabled to communicate with computing device 108 through receiver 106. While receiver 106 is shown outside of the frame of computing device 108, it should be appreciated that the receiver can be integrated into the computing device in one embodiment of the invention. Input device 100 transmits and receives signals through a response and communication system embedded within the input device in one embodiment. Computing device 108 is in communication with display monitor 110. Computing device 108 transmits digital data to display monitor 110 so that the digital data can be viewed. Display monitor 110 may display text 110a, menu 110b and/or graphics 110c. In one embodiment, where computing device 108 is a game console, display monitor 110 displays graphics or a user interface associated with a game being played.
Each of input devices 100, of Figure 2, is configured to generate commands 104 in response to a user's hand movements. Commands 104 are captured by computing device 108 through receiver 106. The captured data is in the form of digital data, which enables input devices 100 to be used as an interactive device with a computer program. The movement of a particular finger or fingers is thus transferred to particular protrusions of input device 100. As noted above, these movements are then transferred to computing device 108 to initiate a command, respond to a query, maneuver objects in an interactive video game, etc. It should be appreciated that input devices 100 are configured to fit comfortably within a user's hand. Thus different sizes can be made available based upon a user's age, gender, etc. Additionally, input devices 100 are configured to remain in a user's hand even if the user is holding his hands in a typing position. That is the natural bend of the user's fingers will support the input device in order to prevent the input device form falling out of the user's hand. Of course, finger straps, such as VELCRO straps, may be incorporated to secure the input device to the user's hand in one embodiment of the invention. The material of the protrusions is selected such that, the material does not easily slide off of the fingers while a user is holding the input device.
Figure 3 is a bottom view of the external input device illustrating the embedded electronics in accordance with one embodiment of the invention. External input device 100 is composed of central body 116 and protrasions 112a-l 12e. Protrusions 112a-l 12e may also be referred to as legs. Each of protrasions 112a-112e are available for each finger and thumb of a user's hand. Protrasions 112a-l 12e are configured to bend as each respective finger or thumb bends. In essence, the configuration of external input device 100 is analogous to a spider, where protrasions 112a-l 12e are the legs in one embodiment. Central body 116 is configured to rest in the palm area of a user's hand. Additionally, central body 116 includes electronics for translating finger and thumb movements and provides an anchor point for each of protrasions 112a-l 12e. Of course, the electronics for translating finger and thumb movements may be disposed anywhere within input device 100 or even attached to the input device. In one embodiment, transmitter/receiver 118 contains circuitry for transmitting signals to a computing device, such as device 108 of Figure 2. Transmitter/receiver 118 also contains circuitry for receiving signals from an external device, such as computing device 108 of Figure
2. External input device 100 of Figure 3 is enabled to provide low latency tactile/haptic feedback to the fingers, thumb and palm of a user's hand through vibrators 120a-120g positioned within the external input device. In one embodiment, the haptic/tactile feedback is provided through a TOUCHENGINE™ tactile/haptic interface component consisting of an actuator and controlling unit. Of course, any tactile/haptic interface component, either manufactured especially for the present embodiment, or one that can be engineered from outside supplier hardware and/or software may be used. In another embodiment, each protrusion 112a-112e and central body 116 include at least one vibrator for providing low latency haptic/tactile feedback. Each protrusion is enabled to measure pressure at: (a) the fingertip, (b) calculate the flex of the finger, and (c) stimulate the finger with patterned vibrations. The surface of the protrasion that contacts the user's hand is, in one embodiment, rubberized to give the user a secure and comfortable grip. In yet another embodiment, the number of protrusions can be varied so that, for example, only the thumb and first two fingers are sensed and the user could grip the bottom part of the input device with the remaining two fingers. It should be appreciated that since the hand makes up a large part of expressive ability, it is well suited to communicate information to a computing device. It should be appreciated that the protrasions may be composed of any suitable material that can be deformed and subsequently returned to a natural position once the force causing the deformation is released. As used herein, the terms tactile and haptic are interchangeable as both terms refer to a sense of touch. Line extensions 114a-114e are configured to sense the flex or degree of bend of a finger or thumb. The flex or degree of bend is communicated to transmitter/receiver 118 where the signal is translated into a command that is subsequently transmitted to a computing device in one embodiment. Of course, the signal can be communicated directly to transmitter/receiver 118 without being translated. Input device 100 is constructed of lightweight material in one embodiment. For example, central body 116 is composed polystyrene, rubber foam, plastic or some other suitable lightweight material capable of holding the above mentioned electronics and acting as an anchor for protrasions 112a-112e. Protrusions 112a-l 12e are composed of a material that can be gripped by a user's hand. In one embodiment, the material for protrasions 112a-112e offers enough resistance so that input device 100 can be held comfortably and without much effort while a user's hand. As mentioned above, the material allows protrasions 112a-112e to be squeezed into a closed position, as illustrated with reference to Figure 5B, and then returned to an open position, once the force causing input device 100 to close is removed. Electronics, such as receiver/transmitter 118, line extensions 114a-114e, and vibrators 120a-120g are embedded within central body 116 and protrasions 112a-112e. In one embodiment, pressure sensors are included at the fingertip positions of protrasions 112a-l 12e to monitor fingertip pressure. The above described electronics are included in the response and communications system, also referred to as the communications system, of input device 100. In another embodiment, power is supplied for the embedded electronics of the response and communications system through a battery located within central body 116.
Figure 4 is a bottom perspective view of the input device in an open position in accordance with one embodiment of the invention. Central body 116 is configured to fit inside a user's palm, while protrusions 112a-l 12e conform to a user's fingers and thumb. The input device is of a lightweight construction that allows a user as much freedom of movement as possible. In one embodiment, the electronics embedded in central body 116 allow for unrestricted placement of the input device during operation as there is no line-of-sight requirement. That is, a short range radio technology, rather than an infrared signal, is used to communicate through the input device and the computing device. In another embodiment, each of protrusions 112a-l 12e include a sensor configured to detect the pressure exerted by the fingertips of a user.
Figure 5A is a schematic diagram of the input device being held by a user while the input device is in an open position in accordance with one embodiment of the invention. Central body 116 of input device 100 fits within the palm of a user's hand 102. Each of protrasions 112a-112e are configured to accommodate a finger or thumb of the user's hand 102. Figure 5B is a schematic diagram of the input device being held by a user while the input device is in a closed position in accordance with one embodiment of the invention. Here, input device 100 is brought to a closed position as the user squeezes protrusions 112a-112e while holding the input device in the user's hand 102. Central body 116 remains against the user's palm while input device 100 is closed. In one embodiment, a latching mechanism allows a user to latch input device 100 in a closed position, h the latched position, protrusions 112a- 112e essentially wrap around central body 116 to define a cylinder. Thus, input device 100 can easily fit inside a pocket or can conveniently be stored by the user as the area for storage is minimized in the latched position. A button can be depressed or moved to release the latching mechanism so that the input device opens, i.e., the protrusions are released from the latched position and "open up".
Input device 100 may be utilized for numerous input functions. Some exemplary functions are described below. The input device can act as a virtual keyboard in one embodiment. Here, low latency vibrators, such as TOUCHENGINE™ vibrators, mounted on each finger to give a tactile "click" when moving from one virtual key to another. A different click can be used to indicate when a key is pressed (via pressure sensors on fingertips), and a continuous vibration can be used whenever the finger is out of the "home" row. Positional left/right tracking of the hand could be used to switch between different zones of the virtual keyboard if, for example, only a single hand is being used. Such a virtual keyboard has the advantage that any surface can be typed on, in any orientation. For instance, a user might be standing and can type on the sides of their legs.
In another embodiment, the input device can act as a virtual piano with feedback similar to the keyboard embodiment described above. For example, a camera pointed at the user could show live video of the user combined with a three dimensional computer graphic piano responding to the users hands.
The input device can act as a control for a game in one embodiment. For instance, for a game where the user is a pilot, a folded sheet of plastic can be used to simulate the cockpit controls (e.g., the instrument panel can be printed on the plastic or other material), and the user can be wearing the input devices on each hand to enable piloting of the airplane or other vehicle or device. A display monitor connected to a computer, or game unit, allows the user to view the interactions. To begin, the user can touch the four comers of the sheet, which calibrates the positional tracking. For example, the four comers of the sheet can be marked with red crosses to indicate the areas to be touched. A combination of fingertip pressure, positional tracking, and tactile feedback can be used to interact with the game. The game may also include a plastic yoke or joystick that contains no electronics and only serves to constrain or steady the users hands. The game detects the user gripping the controls by a combination of fingertip pressure and finger flex information and then tracks the motion and/or orientation of the hand to interpret the manipulation of the physical controls. The plastic yoke might be designed with the shape of the input device in mind and include, for example, depressions for the central body of the input device to fit in. In another embodiment, the degree of pressure applied when a user squeezes their fingers together is correlated to a tightness of a grip exerted on an object in a video game application being displayed on a monitor.
It will be apparent to one skilled in the art that through the addition of a camera and computer vision techniques, the exemplary embodiments described herein could be enhanced by adding graphical representations into live video of the user. Fingertip menus with selections above each fmger, selected by touching the thumb and finger together, can also be accommodated through the input device. Here, the input device may also be used to perform click and drag operations similar to a mouse.
Figure 6 is a top perspective view of an input device in accordance with one embodiment of the invention. Input device 100 consists of central body 116 and protrasions
112a-112e. As mentioned with reference to Figure 3, central body 116 includes embedded electronics for transmitting signals to and receiving signals from a computing device such as a game console. Protrusions 112a-112e include extension lines for determining the degree of bend of a user's fingers. Pressure sensors may also be provided to obtain fingertip pressure. Vibrational feedback is provided to a user through vibrators located in protrusions 112a-112e and central body 116. In one embodiment, the vibrators are configured to control the hardness and the pattern of the vibration to provide the user with tactile feedback. It should be appreciated that by varying the hardness and the pattern of the vibration the sensation perceived by the user may differ. For example, a user can be provided a sense of gripping a steering wheel, typing, striking a piano key, etc.
Figure 7 is a bottom perspective view of an alternative embodiment of the input device of Figure 4. Input device 100 includes central body 116 and protrusions 112a, 112b and 112c. Thus, in this embodiment, the thumb, forefinger and middle finger of the user are accommodated by protrusions 112a, 112b and 112c. Button 130 on the bottom surface of central body 116 releases protrusions 112a, 112b and 112c from a closed position. That is, depressing button 130 releases the latching mechanism holding protrasions 112a, 112b and 112c in a closed position. Of course, button 130 may be incorporated into the embodiments of the input device having protrasions for each finger. One skilled in the art will appreciate that the latching mechanism may be spring activated to release to an open position upon depression of button 130. In another embodiment, input device 100 may be configured as having one protrusion. For example, one protrusion for an index fmger extending from central body 116 may be provided. Here, the movement of the index finger is captured, although the single protrusion can accommodate any one finger or thumb. Vibrators may also be located in the single protrasion as well as the central core to provide tactile feedback to the user's hand through the finger and the user's palm. Figure 8 is a flowchart diagram of the method operations of method for communicating commands to a computer system from an external input device in accordance with one embodiment of the invention. The method initiates with operation 150 where a receiver is provided. The receiver is configured to communicate with an external input device configured to be held within a user's palm. In one embodiment, the receiver is integrated with a computing device, such as a game console. An exemplary input device is shown with reference to Figures 3-7. The method then advances to operation 152 where the hand movements of a user are captured through deformation of the input device. Here, the user applies and releases pressure to the input device. Electronics embedded within the input device capture the change in position of the protrasions, which yields information on the degree of flex of the fingers as well as the pressure at the fingertips.
The method of Figure 8 then proceeds to operation 154 where the captured hand movements are translated to a command to be communicated to a receiver. In one embodiment, the movements are translated and then transferred to a computing system to initiate a command, respond to a query, maneuver objects in an interactive video game, etc. The method then moves to operation 156 where the user's hand is stimulated with tactile feedback provided through the input device. Here, low latency vibrators as illustrated in Figure 3 provide the tactile feedback. In one embodiment, the hardness and the pattern of the vibration is controlled to provide a certain feel to the user. That is, the type of vibrational feedback is dependent on the application being executed. In one embodiment, the invention may be a computing system capable of providing vibrational feedback to a user consisting of various components defined in terms of the functions of each of the components. For example, the computing system can include a means for capturing hand movements. The means for capturing hand movements of a user are provided through an external input device configured to be held by the user, such as the hand-held input device illustrated in Figures 3-7. Means for transmitting the captured hand movements are included. The means for transmitting the captured hand movements to a computing device may be the embedded electronics discussed above with reference to Figure 3. Means for transmitting commands to the external input device in response to the transmitted hand movements of the user are provided. Here, a transmitter associated with a computing device transmits the commands to the external input device. Means for providing tactile feedback to the user in response to the transmitted commands to the external input device are provided by the vibrators illustrated in Figure 3. Means for displaying graphic images associated with the commands received from the external input device are provided by a display monitor in communication with a computing device.
In summary, a hand-held input device is provided. The hand-held input device is configured to be non-surface relational. That is, the input device does not require movement along a surface to function, as is required for a mouse. In one embodiment, a user's finger movements generate signals for controlling objects displayed on a monitor associated with a computing unit that receives the generated signals. For example, the objects may be associated with a video game application where the computing unit is a video game console. One skilled in the art will appreciate that above mentioned embodiments are applicable to a game console running the video games as well as on-line video gaming where a user accesses the video games from a server. Thus, the computing unit can be a game console or a server accessed through a distributed network, such as the Internet.
The invention may employ various computer-implemented operations involving data stored in computer systems. These operations are those requiring physical manipulation of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated. Further, the manipulations performed are often referred to in terms, such as producing, identifying, determining, or comparing.
Any of the operations described herein that form part of the invention are useful machine operations. The invention also relates to a device or an apparatus for performing these operations. The apparatus may be specially constructed for the required purposes, or it may be a general purpose computer selectively activated or configured by a computer program stored in the computer. In particular, various general purpose machines may be used with computer programs written in accordance with the teachings herein, or it may be more convenient to construct a more specialized apparatus to perform the required operations.
The invention can also be embodied as computer readable code on a computer readable medium. The computer readable medium is any data storage device that can store data which can be thereafter be read by a computer system. Examples of the computer readable medium include hard drives, network attached storage (NAS), read-only memory, random-access memory, CD-ROMs, CD-Rs, CD-RWs, magnetic tapes, and other optical and non-optical data storage devices. The computer readable medium can also be distributed over a network coupled computer systems so that the computer readable code is stored and executed in a distributed fashion.
Although the foregoing invention has been described in some detail for purposes of clarity of understanding, it will be apparent that certain changes and modifications may be practiced. Accordingly, the present embodiments are to be considered as illustrative and not restrictive, and the invention is not to be limited to the details given herein, but may be modified within the scope and equivalents of the description.
What is claimed is:

Claims

Claims 1. An input device, comprising: a central body; a response and communications system incorporated with the central body, the response and communication system enabling communication between the input device and a computing device; and a plurality of protrusions extending from the central body, wherein the central body and the plurality of protrusions are configured to capture a user's hand movements.
2. The input device of claim 1, wherein the plurality of protrusions and the central body are configured to be held in a user's hand.
3. The input device of claim 1 , wherein the central body and each of the plurality of protrusions include, a vibrator configured to provide tactile feedback to a user.
4. The input device of claim 3, wherein a hardness and a pattern of a vibration generated by the vibrator are controlled.
5. The input device of claim 1, wherein the response and communication system transmits signals to a device to be controlled, the signals defined by the user's hand movements.
6. An input detection program executed at a computing device, the input detection program having program instructions for determining when to trigger input commands of a program and tactile feedback responses to an external input device configured to be held by a user, the input detection program, comprising: program instructions for detecting a signal from the external input device that is held by a user, the signal generated by detecting a physical change to the external input device; and program instructions for triggering an input command at the program run on the computing device, the triggering resulting from the detected signal from the external mput object.
7. The input detection program of claim 6, further including program instructions for transmitting a tactile feedback signal to the external input device in response to the detected signal.
8. The input detection program of claim 7, wherein the program instructions for generating a tactile feedback signal at the external input device in response to the detected signal, include, program instructions for controlling a vibration hardness and a vibration pattern of a vibrator embedded in the external input device:
9. The input detection program of claim 6, wherein the external input device is configured to be held in a user's hand.
10. The input detection program of claim 6, wherein the computing device is selected from the group consisting of a game console and a server.
11. A computing system enabling tactile feedback to a user, comprising: a computing unit, the computing unit including circuitry configured to receive and transmit commands; a display monitor in communication with the computing unit; and an input device configured to be held in a user's hand, the input device enabled to provide tactile feedback to a user in response to hand movements of the user.
12. The computing system of claim 11, wherein the computing unit is a game console.
13. The computing system of claim 11, wherein the input device includes a central body with a plurality of protrusions extending therefrom.
14. The computing system of claim 11, wherein the tactile feedback is vibrational feedback generated by a vibrator.
15. The computing system of claim 14, wherein a hardness and a pattern of the vibrational feedback are controlled.
PCT/US2003/021769 2002-07-18 2003-07-11 Hand-held computer interactive device WO2004010370A2 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
EP03765543A EP1523725B1 (en) 2002-07-18 2003-07-11 Hand-held computer interactive device
AT03765543T ATE480814T1 (en) 2002-07-18 2003-07-11 INTERACTIVE PORTABLE INPUT DEVICE FOR A COMPUTER
KR1020057000857A KR100742029B1 (en) 2002-07-18 2003-07-11 Hand-held computer interactive device
JP2005505516A JP4044114B2 (en) 2002-07-18 2003-07-11 Grasping computer interactive device
DE60334101T DE60334101D1 (en) 2002-07-18 2003-07-11 INTERACTIVE PORTABLE INPUT DEVICE FOR A COMPUTER
AU2003249062A AU2003249062A1 (en) 2002-07-18 2003-07-11 Hand-held computer interactive device

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US39709202P 2002-07-18 2002-07-18
US60/397,092 2002-07-18
US10/290,789 US7161579B2 (en) 2002-07-18 2002-11-07 Hand-held computer interactive device
US10/290,789 2002-11-07

Publications (2)

Publication Number Publication Date
WO2004010370A2 true WO2004010370A2 (en) 2004-01-29
WO2004010370A3 WO2004010370A3 (en) 2004-11-18

Family

ID=30448122

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2003/021769 WO2004010370A2 (en) 2002-07-18 2003-07-11 Hand-held computer interactive device

Country Status (9)

Country Link
US (3) US7161579B2 (en)
EP (1) EP1523725B1 (en)
JP (1) JP4044114B2 (en)
KR (1) KR100742029B1 (en)
CN (1) CN100397304C (en)
AT (1) ATE480814T1 (en)
AU (1) AU2003249062A1 (en)
DE (1) DE60334101D1 (en)
WO (1) WO2004010370A2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8059089B2 (en) 2004-05-25 2011-11-15 Sony Computer Entertainment Inc. Input device and method, and character input method

Families Citing this family (116)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8797260B2 (en) 2002-07-27 2014-08-05 Sony Computer Entertainment Inc. Inertially trackable hand-held controller
US7102615B2 (en) * 2002-07-27 2006-09-05 Sony Computer Entertainment Inc. Man-machine interface using a deformable device
KR100590526B1 (en) * 2003-04-18 2006-06-15 삼성전자주식회사 Apparatus and method for detecting finger-motion
US6937227B2 (en) * 2003-07-14 2005-08-30 Iowa State University Research Foundation, Inc. Hand-held pointing device
JP4611667B2 (en) * 2003-11-25 2011-01-12 健爾 西 Information input device, storage device, information input device, and information processing device
GB2409596A (en) * 2003-12-23 2005-06-29 Nokia Corp Handset with squeezable body
US20050257150A1 (en) * 2004-05-11 2005-11-17 Universite Des Sciences Et Technologies De Lille Ground-based haptic interface comprising at least two decoupled rotary finger actuators
US20060164383A1 (en) * 2004-12-16 2006-07-27 Media Lab Europe (In Voluntary Liquidation) Remote controller ring for user interaction
US20060221059A1 (en) * 2005-04-01 2006-10-05 Samsung Electronics Co., Ltd. Portable terminal having display buttons and method of inputting functions using display buttons
JP4788246B2 (en) * 2005-08-31 2011-10-05 ソニー株式会社 Input device and input method
US20070158911A1 (en) * 2005-11-07 2007-07-12 Torre Gabriel D L Interactive role-play toy apparatus
US7810504B2 (en) * 2005-12-28 2010-10-12 Depuy Products, Inc. System and method for wearable user interface in computer assisted surgery
JP4892671B2 (en) * 2006-08-21 2012-03-07 学校法人立命館 Motion detection sensor and actuator system
KR100826872B1 (en) * 2006-08-30 2008-05-06 한국전자통신연구원 Wearable computer system and method controlling information/service in wearable computer system
US9032336B2 (en) * 2006-09-07 2015-05-12 Osaka Electro-Communication University Gesture input system, method and program
US20090017910A1 (en) 2007-06-22 2009-01-15 Broadcom Corporation Position and motion tracking of an object
JP2008305198A (en) * 2007-06-07 2008-12-18 Fujitsu Component Ltd Input system and input device
US20090002325A1 (en) * 2007-06-27 2009-01-01 Think/Thing System and method for operating an electronic device
US20090146947A1 (en) * 2007-12-07 2009-06-11 James Ng Universal wearable input and authentication device
US20090160666A1 (en) * 2007-12-21 2009-06-25 Think/Thing System and method for operating and powering an electronic device
KR100907411B1 (en) * 2008-01-10 2009-07-10 에스케이 텔레콤주식회사 Tactile feedback system utilizing multi-touch input device and method of the same
US8358277B2 (en) * 2008-03-18 2013-01-22 Microsoft Corporation Virtual keyboard based activation and dismissal
US20100030089A1 (en) * 2008-04-24 2010-02-04 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Methods and systems for monitoring and modifying a combination treatment
US20100081861A1 (en) * 2008-04-24 2010-04-01 Searete Llc Computational System and Method for Memory Modification
US20090312668A1 (en) * 2008-04-24 2009-12-17 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Computational system and method for memory modification
US20100100036A1 (en) * 2008-04-24 2010-04-22 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Computational System and Method for Memory Modification
US7801686B2 (en) * 2008-04-24 2010-09-21 The Invention Science Fund I, Llc Combination treatment alteration methods and systems
US8876688B2 (en) * 2008-04-24 2014-11-04 The Invention Science Fund I, Llc Combination treatment modification methods and systems
US9649469B2 (en) 2008-04-24 2017-05-16 The Invention Science Fund I Llc Methods and systems for presenting a combination treatment
US20100041958A1 (en) * 2008-04-24 2010-02-18 Searete Llc Computational system and method for memory modification
US20100042578A1 (en) * 2008-04-24 2010-02-18 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Computational system and method for memory modification
US9064036B2 (en) * 2008-04-24 2015-06-23 The Invention Science Fund I, Llc Methods and systems for monitoring bioactive agent use
US20090271122A1 (en) * 2008-04-24 2009-10-29 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Methods and systems for monitoring and modifying a combination treatment
US9560967B2 (en) 2008-04-24 2017-02-07 The Invention Science Fund I Llc Systems and apparatus for measuring a bioactive agent effect
US20090271347A1 (en) * 2008-04-24 2009-10-29 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Methods and systems for monitoring bioactive agent use
US20100041964A1 (en) * 2008-04-24 2010-02-18 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Methods and systems for monitoring and modifying a combination treatment
US20090270687A1 (en) * 2008-04-24 2009-10-29 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Methods and systems for modifying bioactive agent use
US8930208B2 (en) * 2008-04-24 2015-01-06 The Invention Science Fund I, Llc Methods and systems for detecting a bioactive agent effect
US20100125561A1 (en) * 2008-04-24 2010-05-20 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Computational system and method for memory modification
US7974787B2 (en) 2008-04-24 2011-07-05 The Invention Science Fund I, Llc Combination treatment alteration methods and systems
US20100022820A1 (en) * 2008-04-24 2010-01-28 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Computational system and method for memory modification
US20100076249A1 (en) * 2008-04-24 2010-03-25 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Computational system and method for memory modification
US9026369B2 (en) * 2008-04-24 2015-05-05 The Invention Science Fund I, Llc Methods and systems for presenting a combination treatment
US20100081860A1 (en) * 2008-04-24 2010-04-01 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Computational System and Method for Memory Modification
US20090271009A1 (en) * 2008-04-24 2009-10-29 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Combination treatment modification methods and systems
US20090270694A1 (en) * 2008-04-24 2009-10-29 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Methods and systems for monitoring and modifying a combination treatment
US9282927B2 (en) * 2008-04-24 2016-03-15 Invention Science Fund I, Llc Methods and systems for modifying bioactive agent use
US20100015583A1 (en) * 2008-04-24 2010-01-21 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Computational System and method for memory modification
US8615407B2 (en) * 2008-04-24 2013-12-24 The Invention Science Fund I, Llc Methods and systems for detecting a bioactive agent effect
US9662391B2 (en) 2008-04-24 2017-05-30 The Invention Science Fund I Llc Side effect ameliorating combination therapeutic products and systems
US20100004762A1 (en) * 2008-04-24 2010-01-07 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Computational system and method for memory modification
US8606592B2 (en) * 2008-04-24 2013-12-10 The Invention Science Fund I, Llc Methods and systems for monitoring bioactive agent use
US9449150B2 (en) * 2008-04-24 2016-09-20 The Invention Science Fund I, Llc Combination treatment selection methods and systems
US9239906B2 (en) * 2008-04-24 2016-01-19 The Invention Science Fund I, Llc Combination treatment selection methods and systems
US20090269329A1 (en) * 2008-04-24 2009-10-29 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Combination Therapeutic products and systems
US8682687B2 (en) * 2008-04-24 2014-03-25 The Invention Science Fund I, Llc Methods and systems for presenting a combination treatment
US20100017001A1 (en) * 2008-04-24 2010-01-21 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Computational system and method for memory modification
US20090271375A1 (en) * 2008-04-24 2009-10-29 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Combination treatment selection methods and systems
US20090270688A1 (en) * 2008-04-24 2009-10-29 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Methods and systems for presenting a combination treatment
US8605041B2 (en) * 2009-06-08 2013-12-10 Kevin Theodore Johnson Data input device
US8947360B2 (en) * 2009-08-07 2015-02-03 Vivek Gupta Set of handheld adjustable panels of ergonomic keys and mouse
WO2011032295A1 (en) * 2009-09-20 2011-03-24 Tungle Corporation Mobile application for calendar sharing and scheduling
CN102741787A (en) * 2009-10-22 2012-10-17 乔舒亚·迈克尔·扬 Man-machine interface device
WO2011070554A2 (en) * 2009-12-13 2011-06-16 Ringbow Ltd. Finger-worn input devices and methods of use
WO2012051664A1 (en) * 2010-10-22 2012-04-26 Joshua Michael Young Methods devices and systems for creating control signals
KR101871359B1 (en) 2010-12-21 2018-08-02 삼성전자주식회사 Method, apparatus, and system for providing sensitivity information and sensitivity
JP5536237B2 (en) * 2011-01-13 2014-07-02 エンパイア テクノロジー ディベロップメント エルエルシー Tactile feedback device using electrorheological fluid environment
EP2613223A1 (en) 2012-01-09 2013-07-10 Softkinetic Software System and method for enhanced gesture-based interaction
US9050529B2 (en) 2012-01-30 2015-06-09 Microsoft Technology Licensing, Llc Force feedback triggers
RU2493580C1 (en) * 2012-04-13 2013-09-20 Евгений Маркович Чепурных Information input device
US8992324B2 (en) * 2012-07-16 2015-03-31 Wms Gaming Inc. Position sensing gesture hand attachment
GB2508137A (en) * 2012-11-07 2014-05-28 Stuart George Wilson Handheld haptic feedback device
NL1040096C2 (en) * 2012-11-14 2014-09-16 Holland Haptics B V Haptic communication system, method and device.
WO2014194163A2 (en) * 2013-05-30 2014-12-04 Homayoon Kazerooni User-coupled human-machine interface
US9542027B2 (en) 2014-04-16 2017-01-10 At&T Intellectual Property I, L.P. Pressure-based input method for user devices
CN103984411B (en) * 2014-04-25 2017-07-18 深圳超多维光电子有限公司 Tactile force feedback system and tactile force feedback network system realization
CN103949053B (en) * 2014-05-23 2017-07-07 无锡梵天信息技术股份有限公司 The online electronic game communication system of many people
EP3193238A4 (en) 2014-09-10 2018-07-18 Sony Corporation Detection device, detection method, control device, and control method
US9727138B2 (en) * 2014-10-27 2017-08-08 Cherif Algreatly Nanotechnology clothing for human-computer interaction
US9174134B1 (en) * 2014-11-12 2015-11-03 Immersion Corporation Peripheral device with haptic diminishment prevention component
JP6566028B2 (en) 2015-05-11 2019-08-28 富士通株式会社 Simulation system
JPWO2016181469A1 (en) * 2015-05-11 2018-03-08 富士通株式会社 Simulation system
KR101653146B1 (en) 2015-09-04 2016-09-01 홍유정 Drone controller
EP3361350A4 (en) * 2015-10-05 2018-10-17 Alps Electric Co., Ltd. Haptic reproduction device
CN106648043A (en) * 2015-11-02 2017-05-10 广东虚拟现实科技有限公司 Signal acquisition method for controller and controller
KR101921376B1 (en) 2016-02-24 2018-11-22 홍유정 Object controller
CN105892754A (en) * 2016-05-27 2016-08-24 北京小鸟看看科技有限公司 Finger action recognition method and system
CN106227339A (en) * 2016-08-16 2016-12-14 西安中科比奇创新科技有限责任公司 wearable device, virtual reality human-computer interaction system and method
US10013069B2 (en) * 2016-09-13 2018-07-03 Intel Corporation Methods and apparatus to detect vibration inducing hand gestures
CN106371604B (en) * 2016-09-18 2020-03-20 Tcl集团股份有限公司 Interactive control glove, virtual reality system and application method thereof
US10649583B1 (en) 2016-10-11 2020-05-12 Valve Corporation Sensor fusion algorithms for a handheld controller that includes a force sensing resistor (FSR)
US10888773B2 (en) 2016-10-11 2021-01-12 Valve Corporation Force sensing resistor (FSR) with polyimide substrate, systems, and methods thereof
US10307669B2 (en) 2016-10-11 2019-06-04 Valve Corporation Electronic controller with finger sensing and an adjustable hand retainer
US11625898B2 (en) 2016-10-11 2023-04-11 Valve Corporation Holding and releasing virtual objects
US10549183B2 (en) * 2016-10-11 2020-02-04 Valve Corporation Electronic controller with a hand retainer, outer shell, and finger sensing
US10391400B1 (en) * 2016-10-11 2019-08-27 Valve Corporation Electronic controller with hand retainer and finger motion sensing
US10691233B2 (en) 2016-10-11 2020-06-23 Valve Corporation Sensor fusion algorithms for a handheld controller that includes a force sensing resistor (FSR)
US10987573B2 (en) 2016-10-11 2021-04-27 Valve Corporation Virtual reality hand gesture generation
US11185763B2 (en) 2016-10-11 2021-11-30 Valve Corporation Holding and releasing virtual objects
CN106621323B (en) * 2016-12-09 2023-04-14 深圳趣感科技有限公司 Signal input device and method
JP6832005B2 (en) * 2017-02-24 2021-02-24 公立大学法人公立諏訪東京理科大学 Subject judgment device, method, and program
US10183217B2 (en) * 2017-04-13 2019-01-22 Facebook Technologies, Llc Hand-held controller using segmented capacitive touch trigger
US11501552B2 (en) 2017-04-27 2022-11-15 Sony Interactive Entertainment Inc. Control apparatus, information processing system, control method, and program
US11237642B2 (en) * 2017-05-14 2022-02-01 Mark Stouffer Single-hand interdigit-coordinated mobile keyboard input device
US10528134B2 (en) * 2017-05-14 2020-01-07 Mark Stouffer One-handed input chording user input device for converting finger movements into digital input
DE102017113103A1 (en) * 2017-06-14 2018-12-20 RUD. IBACH SOHN GmbH & Co. KG Data glove and system for simulating a musical instrument
EP3652619A4 (en) 2017-06-16 2021-02-24 Valve Corporation Electronic controller with finger motion sensing
US10444840B2 (en) 2017-08-30 2019-10-15 Disney Enterprises, Inc. Systems and methods to synchronize visual effects and haptic feedback for interactive experiences
RU184063U1 (en) * 2017-09-12 2018-10-15 Василий Львович Беспрозванный Information input device
US11130050B2 (en) * 2017-10-16 2021-09-28 Sony Interactive Entertainment Inc. Information processing system, controller device, and information processing apparatus
CN111565808A (en) * 2017-12-07 2020-08-21 威尔乌集团 Electronic controller with hand holder, housing and finger sensing
TWI632533B (en) * 2018-01-11 2018-08-11 和碩聯合科技股份有限公司 Learning assistant system
WO2019222698A1 (en) 2018-05-18 2019-11-21 Valve Corporation Sensor fusion algorithms for a handheld controller that includes a force sensing resistor (fsr)
KR102183827B1 (en) * 2018-08-03 2020-11-27 홍유정 Object controller
US11353959B2 (en) * 2018-08-21 2022-06-07 Sony Interactive Entertainment Inc. Controller device
WO2020251566A1 (en) 2019-06-12 2020-12-17 Hewlett-Packard Development Company, L.P. Extended reality grasp controller

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5568928A (en) * 1995-02-01 1996-10-29 Exertron, Inc. Video game controller for use with an exercise apparatus
US5583478A (en) * 1995-03-01 1996-12-10 Renzi; Ronald Virtual environment tactile system
US20020021277A1 (en) * 2000-04-17 2002-02-21 Kramer James F. Interface for controlling a graphical image
WO2002027456A1 (en) * 2000-09-29 2002-04-04 Senseboard Technologies Ab Wearable data input interface

Family Cites Families (312)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3943277A (en) 1969-02-20 1976-03-09 The United States Of America As Represented By The Secretary Of The Navy Digital memory area correlation tracker
US4313227A (en) 1979-01-29 1982-01-26 Texas Instruments Incorporated Light energy information transmission system
US4263504A (en) 1979-08-01 1981-04-21 Ncr Corporation High density matrix code
US6772057B2 (en) 1995-06-07 2004-08-03 Automotive Technologies International, Inc. Vehicular monitoring systems using image processing
US4565999A (en) 1983-04-01 1986-01-21 Prime Computer, Inc. Light pencil
US4558864A (en) 1984-06-13 1985-12-17 Medwedeff Marion C Handgrip exercising, computer game controller
JPS6244827A (en) * 1985-08-20 1987-02-26 ブィ・ピィ・エル・リサーチ・インコーポレイテッド Apparatus and method for generating control signal accordingto action and position of hand
JPH06102980B2 (en) 1985-12-12 1994-12-14 スズキ株式会社 Two-cycle engine control method
US5195179A (en) 1986-01-29 1993-03-16 Hitachi, Ltd. Coordinate input apparatus
US4843568A (en) 1986-04-11 1989-06-27 Krueger Myron W Real time perception of and response to the actions of an unencumbered participant/user
JPS6347616A (en) 1986-08-15 1988-02-29 Ricoh Co Ltd Measurement of moving quantity
US4802227A (en) 1987-04-03 1989-01-31 American Telephone And Telegraph Company Noise reduction processing arrangement for microphone arrays
GB2206716A (en) 1987-07-06 1989-01-11 Gen Electric Plc Apparatus for monitoring the presence or movement of an object
JP2720455B2 (en) * 1988-05-12 1998-03-04 ヤマハ株式会社 Finger switch
IT1219405B (en) 1988-06-27 1990-05-11 Fiat Ricerche PROCEDURE AND DEVICE FOR INSTRUMENTAL VISION IN POOR CONDITIONS VISIBILITY IN PARTICULAR FOR DRIVING IN THE MIST
EP0449889A4 (en) 1988-12-20 1992-12-02 Australian Meat And Live-Stock Research And Development Corporation An optically readable coded target
US5034986A (en) 1989-03-01 1991-07-23 Siemens Aktiengesellschaft Method for detecting and tracking moving objects in a digital image sequence having a stationary background
JPH0816861B2 (en) * 1989-09-22 1996-02-21 東京電力株式会社 3D graphic tactile feedback device
US5055840A (en) 1990-01-16 1991-10-08 Carroll Touch Incorporated Infrared touch input device and light emitted activation circuit
AU654118B2 (en) 1990-05-01 1994-10-27 Wang Laboratories, Inc. Keyboard access with alternate input device
US5111401A (en) 1990-05-19 1992-05-05 The United States Of America As Represented By The Secretary Of The Navy Navigational control system for an autonomous vehicle
US5662111A (en) 1991-01-28 1997-09-02 Cosman; Eric R. Process of stereotactic optical navigation
US5485273A (en) 1991-04-22 1996-01-16 Litton Systems, Inc. Ring laser gyroscope enhanced resolution system
JPH04324516A (en) * 1991-04-24 1992-11-13 Akemasa Abe Handy keyboard
US5534917A (en) * 1991-05-09 1996-07-09 Very Vivid, Inc. Video image based control system
US5144594A (en) 1991-05-29 1992-09-01 Cyber Scientific Acoustic mouse system
US5455685A (en) 1991-09-04 1995-10-03 Fuji Photo Film Co., Ltd. Video camera exposure control apparatus for controlling iris diaphragm and automatic gain control operating speed
JPH0566908A (en) * 1991-09-05 1993-03-19 Nec Corp Object hold display method
US5889670A (en) * 1991-10-24 1999-03-30 Immersion Corporation Method and apparatus for tactilely responsive user interface
US5444462A (en) 1991-12-16 1995-08-22 Wambach; Mark L. Computer mouse glove with remote communication
US5296871A (en) * 1992-07-27 1994-03-22 Paley W Bradford Three-dimensional mouse with tactile feedback
US5790834A (en) 1992-08-31 1998-08-04 Intel Corporation Apparatus and method using an ID instruction to identify a computer microprocessor
JP3244798B2 (en) 1992-09-08 2002-01-07 株式会社東芝 Moving image processing device
US7098891B1 (en) 1992-09-18 2006-08-29 Pryor Timothy R Method for providing human input to a computer
JPH06102980A (en) * 1992-09-24 1994-04-15 Olympus Optical Co Ltd Tactual sense presentation device
US5394168A (en) 1993-01-06 1995-02-28 Smith Engineering Dual-mode hand-held game controller
US5335011A (en) 1993-01-12 1994-08-02 Bell Communications Research, Inc. Sound localization system for teleconferencing using self-steering microphone arrays
DE69414153T2 (en) 1993-02-24 1999-06-10 Matsushita Electric Ind Co Ltd Device for gradation correction and image recording device with such a device
US5435554A (en) 1993-03-08 1995-07-25 Atari Games Corporation Baseball simulation system
JP3679426B2 (en) 1993-03-15 2005-08-03 マサチューセッツ・インスティチュート・オブ・テクノロジー A system that encodes image data into multiple layers, each representing a coherent region of motion, and motion parameters associated with the layers.
US5677710A (en) 1993-05-10 1997-10-14 Apple Computer, Inc. Recognition keypad
US5297061A (en) 1993-05-19 1994-03-22 University Of Maryland Three dimensional pointing device monitored by computer vision
JPH06337630A (en) * 1993-05-27 1994-12-06 Hitachi Ltd Portable type sign language input device
US5959596A (en) 1993-06-24 1999-09-28 Nintendo Co., Ltd. Airline-based video game and communications system
US5473701A (en) 1993-11-05 1995-12-05 At&T Corp. Adaptive microphone array
JP2552427B2 (en) 1993-12-28 1996-11-13 コナミ株式会社 Tv play system
FR2714502A1 (en) 1993-12-29 1995-06-30 Philips Laboratoire Electroniq An image processing method and apparatus for constructing from a source image a target image with perspective change.
US5611000A (en) 1994-02-22 1997-03-11 Digital Equipment Corporation Spline-based image registration
JPH086708A (en) 1994-04-22 1996-01-12 Canon Inc Display device
US5543818A (en) 1994-05-13 1996-08-06 Sony Corporation Method and apparatus for entering text using an input device having a small number of keys
US5846086A (en) 1994-07-01 1998-12-08 Massachusetts Institute Of Technology System for human trajectory learning in virtual environments
US5563988A (en) 1994-08-01 1996-10-08 Massachusetts Institute Of Technology Method and system for facilitating wireless, full-body, real-time user interaction with a digitally represented visual environment
JP3270643B2 (en) 1994-12-22 2002-04-02 キヤノン株式会社 Pointed position detection method and device
US5929444A (en) 1995-01-31 1999-07-27 Hewlett-Packard Company Aiming device using radiated energy
US5638228A (en) 1995-02-14 1997-06-10 Iomega Corporation Retroreflective marker for data storage cartridge
US5930741A (en) 1995-02-28 1999-07-27 Virtual Technologies, Inc. Accurate, rapid, reliable position sensing using multiple sensing technologies
US5900863A (en) 1995-03-16 1999-05-04 Kabushiki Kaisha Toshiba Method and apparatus for controlling computer without touching input device
DE69634913T2 (en) 1995-04-28 2006-01-05 Matsushita Electric Industrial Co., Ltd., Kadoma INTERFACE DEVICE
US5706364A (en) 1995-04-28 1998-01-06 Xerox Corporation Method of producing character templates using unsegmented samples
US5913727A (en) 1995-06-02 1999-06-22 Ahdoot; Ned Interactive movement and contact simulation game
US5649021A (en) 1995-06-07 1997-07-15 David Sarnoff Research Center, Inc. Method and system for object detection for instrument control
WO1997001112A2 (en) 1995-06-22 1997-01-09 3Dv Systems Ltd. Telecentric 3d camera and method of rangefinding
IL114278A (en) 1995-06-22 2010-06-16 Microsoft Internat Holdings B Camera and method
DE69607472T2 (en) 1995-07-26 2000-08-24 Tegic Communications Inc SYSTEM FOR SUPPRESSING AMBIANCE IN A REDUCED KEYBOARD
US6311214B1 (en) 1995-07-27 2001-10-30 Digimarc Corporation Linking of computers based on optical sensing of digital data
US5768415A (en) 1995-09-08 1998-06-16 Lucent Technologies Inc. Apparatus and methods for performing electronic scene analysis and enhancement
US5611731A (en) 1995-09-08 1997-03-18 Thrustmaster, Inc. Video pinball machine controller having an optical accelerometer for detecting slide and tilt
US5850222A (en) 1995-09-13 1998-12-15 Pixel Dust, Inc. Method and system for displaying a graphic image of a person modeling a garment
US5818424A (en) 1995-10-19 1998-10-06 International Business Machines Corporation Rod shaped device and data acquisition apparatus for determining the position and orientation of an object in space
US6281930B1 (en) 1995-10-20 2001-08-28 Parkervision, Inc. System and method for controlling the field of view of a camera
US5963250A (en) 1995-10-20 1999-10-05 Parkervision, Inc. System and method for controlling the field of view of a camera
US5719561A (en) * 1995-10-25 1998-02-17 Gilbert R. Gonzales Tactile communication device and method
US5870100A (en) 1995-11-22 1999-02-09 Compaq Computer Corporation Filling of graphical regions
AU1328597A (en) * 1995-11-30 1997-06-19 Virtual Technologies, Inc. Tactile feedback man-machine interface device
JP4079463B2 (en) 1996-01-26 2008-04-23 ソニー株式会社 Subject detection apparatus and subject detection method
US6049619A (en) 1996-02-12 2000-04-11 Sarnoff Corporation Method and apparatus for detecting moving objects in two- and three-dimensional scenes
JP2891159B2 (en) 1996-02-14 1999-05-17 日本電気株式会社 Object detection method from multi-view images
RU2069885C1 (en) 1996-03-01 1996-11-27 Йелстаун Корпорейшн Н.В. Method and device for observing objects at low illumination intensity
JPH09244793A (en) 1996-03-04 1997-09-19 Alps Electric Co Ltd Input device
KR100501145B1 (en) 1996-03-05 2005-07-18 가부시키가이샤 세가 Manipulation controller and electronic device using the same
EP0895696B1 (en) 1996-04-02 2004-01-14 Cognex Corporation Image formation apparatus for viewing indicia on a planar specular substrate
US5937081A (en) 1996-04-10 1999-08-10 O'brill; Michael R. Image composition system and method of using same
US5923318A (en) 1996-04-12 1999-07-13 Zhai; Shumin Finger manipulatable 6 degree-of-freedom input device
US5917493A (en) 1996-04-17 1999-06-29 Hewlett-Packard Company Method and apparatus for randomly generating information for subsequent correlating
JPH1039996A (en) * 1996-07-27 1998-02-13 Nobutaka Kosaka Pointing device mounted to hand
US6151009A (en) 1996-08-21 2000-11-21 Carnegie Mellon University Method and apparatus for merging real and synthetic images
US6400374B2 (en) 1996-09-18 2002-06-04 Eyematic Interfaces, Inc. Video superposition system and method
US5930383A (en) 1996-09-24 1999-07-27 Netzer; Yishay Depth sensing camera systems and methods
US5978772A (en) 1996-10-11 1999-11-02 Mold; Jeffrey W. Merchandise checkout system
US5832931A (en) 1996-10-30 1998-11-10 Photogen, Inc. Method for improved selectivity in photo-activation and detection of molecular diagnostic agents
NL1004648C2 (en) 1996-11-11 1998-05-14 Johan Michiel Schaaij Computer game system.
JPH10171591A (en) * 1996-12-06 1998-06-26 Seiko Epson Corp Pointing device, moving method of pointer and portable information processor
JPH10176919A (en) * 1996-12-18 1998-06-30 Olympus Optical Co Ltd Shape-inputting device
US5914723A (en) 1996-12-30 1999-06-22 Sun Microsystems, Inc. Method and system for converting images in computer systems
US6243491B1 (en) 1996-12-31 2001-06-05 Lucent Technologies Inc. Methods and apparatus for controlling a video system with visually recognized props
US5850473A (en) 1997-01-07 1998-12-15 Lucent Technologies Inc. Method and apparatus for compensating for color variation in a video system
US6021219A (en) 1997-01-07 2000-02-01 Lucent Technologies Inc. Methods and apparatus for distinguishing among several visual patterns
US5796354A (en) * 1997-02-07 1998-08-18 Reality Quest Corp. Hand-attachable controller with direction sensing
US5993314A (en) 1997-02-10 1999-11-30 Stadium Games, Ltd. Method and apparatus for interactive audience participation by audio command
US6009210A (en) * 1997-03-05 1999-12-28 Digital Equipment Corporation Hands-free interface to a virtual reality environment using head tracking
JP2001515646A (en) 1997-03-07 2001-09-18 スリーディーヴィー システムズ リミテッド Optical shutter
US6061055A (en) 1997-03-21 2000-05-09 Autodesk, Inc. Method of tracking objects with an imaging device
US6144367A (en) 1997-03-26 2000-11-07 International Business Machines Corporation Method and system for simultaneous operation of multiple handheld control devices in a data processing system
US8120652B2 (en) 1997-04-02 2012-02-21 Gentex Corporation System for controlling vehicle equipment
US6587573B1 (en) 2000-03-20 2003-07-01 Gentex Corporation System for controlling exterior vehicle lights
US6215898B1 (en) 1997-04-15 2001-04-10 Interval Research Corporation Data processing system and method
US6809776B1 (en) 1997-04-23 2004-10-26 Thomson Licensing S.A. Control of video level by region and content of information displayed
US6428411B1 (en) 1997-05-02 2002-08-06 Konami Co., Ltd. Volleyball video game system
JPH10308134A (en) * 1997-05-06 1998-11-17 Hosiden Corp Multiple switch
US6075895A (en) 1997-06-20 2000-06-13 Holoplex Methods and apparatus for gesture recognition based on templates
DE69811050T2 (en) 1997-07-29 2003-11-06 Koninkl Philips Electronics Nv Reconstruction method, device and decoding system for three-dimensional scenes.
US6044181A (en) 1997-08-01 2000-03-28 Microsoft Corporation Focal length estimation method and apparatus for construction of panoramic mosaic images
US6720949B1 (en) 1997-08-22 2004-04-13 Timothy R. Pryor Man machine interfaces and applications
US20020036617A1 (en) 1998-08-21 2002-03-28 Timothy R. Pryor Novel man machine interfaces and applications
US6243074B1 (en) 1997-08-29 2001-06-05 Xerox Corporation Handedness detection for a physical manipulatory grammar
US6297838B1 (en) * 1997-08-29 2001-10-02 Xerox Corporation Spinning as a morpheme for a physical manipulatory grammar
US6031934A (en) 1997-10-15 2000-02-29 Electric Planet, Inc. Computer vision system for subject characterization
US6101289A (en) 1997-10-15 2000-08-08 Electric Planet, Inc. Method and apparatus for unencumbered capture of an object
US6072494A (en) 1997-10-15 2000-06-06 Electric Planet, Inc. Method and apparatus for real-time gesture recognition
US6411744B1 (en) 1997-10-15 2002-06-25 Electric Planet, Inc. Method and apparatus for performing a clean background subtraction
JPH11154240A (en) 1997-11-20 1999-06-08 Nintendo Co Ltd Image producing device to produce image by using fetched image
US6677987B1 (en) 1997-12-03 2004-01-13 8×8, Inc. Wireless user-interface arrangement and method
US6195104B1 (en) 1997-12-23 2001-02-27 Philips Electronics North America Corp. System and method for permitting three-dimensional navigation through a virtual reality environment using camera-based gesture inputs
US6160540A (en) 1998-01-12 2000-12-12 Xerox Company Zoomorphic computer user interface
US6134346A (en) 1998-01-16 2000-10-17 Ultimatte Corp Method for removing from an image the background surrounding a selected object
WO1999040478A1 (en) 1998-02-08 1999-08-12 3Dv Systems Ltd. Large aperture optical image shutter
US6115052A (en) 1998-02-12 2000-09-05 Mitsubishi Electric Information Technology Center America, Inc. (Ita) System for reconstructing the 3-dimensional motions of a human figure from a monocularly-viewed image sequence
US6037942A (en) 1998-03-10 2000-03-14 Magellan Dis, Inc. Navigation system character input device
JPH11265249A (en) 1998-03-17 1999-09-28 Toshiba Corp Information input device, information input method and storage medium
US6411392B1 (en) 1998-04-15 2002-06-25 Massachusetts Institute Of Technology Method and apparatus for data hiding in printed images
US6393142B1 (en) 1998-04-22 2002-05-21 At&T Corp. Method and apparatus for adaptive stripe based patch matching for depth estimation
US6173059B1 (en) 1998-04-24 2001-01-09 Gentner Communications Corporation Teleconferencing system with visual feedback
US6593956B1 (en) 1998-05-15 2003-07-15 Polycom, Inc. Locating an audio source
US6473516B1 (en) 1998-05-22 2002-10-29 Asa Systems, Inc. Large capacity steganography
JP3646969B2 (en) 1998-05-25 2005-05-11 富士通株式会社 3D image display device
JP3841132B2 (en) 1998-06-01 2006-11-01 株式会社ソニー・コンピュータエンタテインメント Input position detection device and entertainment system
US6513160B2 (en) 1998-06-17 2003-01-28 Opentv, Inc. System and method for promoting viewer interaction in a television system
FR2780176B1 (en) 1998-06-17 2001-01-26 Gabriel Guary SHOOTING GUN FOR VIDEO GAME
IL125221A0 (en) 1998-07-06 1999-03-12 Toy Control Ltd Motion activation using passive sound source
US6256398B1 (en) 1998-08-22 2001-07-03 Kenneth H. P. Chang Encoding and decoding a message within an image
US6970183B1 (en) 2000-06-14 2005-11-29 E-Watch, Inc. Multimedia surveillance and monitoring system including network configuration
US6621938B1 (en) 1998-09-18 2003-09-16 Fuji Photo Film Co., Ltd. Image capture apparatus and method
US6184863B1 (en) 1998-10-13 2001-02-06 The George Washington University Direct pointing apparatus and method therefor
JP2000122791A (en) * 1998-10-16 2000-04-28 Sony Corp Kinesthetic sense presenting device
DE19849515C1 (en) 1998-10-19 2000-12-28 Gerhard Wergen Method for transferring characters in particular to a computer and input device using this method
US6307568B1 (en) 1998-10-28 2001-10-23 Imaginarix Ltd. Virtual dressing over the internet
US6712703B2 (en) 1998-11-19 2004-03-30 Nintendo Co., Ltd. Video game apparatus and information storage medium for video game
JP2000172431A (en) 1998-12-09 2000-06-23 Sony Corp Information input device and game device
US6533420B1 (en) 1999-01-22 2003-03-18 Dimension Technologies, Inc. Apparatus and method for generating and projecting autostereoscopic images
US7139767B1 (en) 1999-03-05 2006-11-21 Canon Kabushiki Kaisha Image processing apparatus and database
JP2000261752A (en) 1999-03-05 2000-09-22 Hewlett Packard Co <Hp> Device for recording and reproducing image having sound
JP4332254B2 (en) * 1999-04-28 2009-09-16 株式会社アミテック Intelligent glove-type hand input device
US6323942B1 (en) 1999-04-30 2001-11-27 Canesta, Inc. CMOS-compatible three-dimensional image sensor IC
US7164413B2 (en) 1999-05-19 2007-01-16 Digimarc Corporation Enhanced input peripheral
US6791531B1 (en) 1999-06-07 2004-09-14 Dot On, Inc. Device and method for cursor motion control calibration and object selection
JP2000350865A (en) 1999-06-11 2000-12-19 Mr System Kenkyusho:Kk Game device for composite real space, image processing method therefor and program storage medium
US6952198B2 (en) 1999-07-06 2005-10-04 Hansen Karl C System and method for communication with enhanced optical pointer
US6819318B1 (en) 1999-07-23 2004-11-16 Z. Jason Geng Method and apparatus for modeling via a three-dimensional image mosaic system
US6545706B1 (en) 1999-07-30 2003-04-08 Electric Planet, Inc. System, method and article of manufacture for tracking a head of a camera-generated image of a person
US6417836B1 (en) 1999-08-02 2002-07-09 Lucent Technologies Inc. Computer input device having six degrees of freedom for controlling movement of a three-dimensional object
JP2001056742A (en) 1999-08-19 2001-02-27 Alps Electric Co Ltd Input device
US6556704B1 (en) 1999-08-25 2003-04-29 Eastman Kodak Company Method for forming a depth image from digital image data
EP1214609B1 (en) 1999-09-08 2004-12-15 3DV Systems Ltd. 3d imaging system
KR20020032595A (en) 1999-09-17 2002-05-03 기타무라 가즈코 Image pickup system, image processor, and camera
JP3847058B2 (en) 1999-10-04 2006-11-15 任天堂株式会社 GAME SYSTEM AND GAME INFORMATION STORAGE MEDIUM USED FOR THE SAME
US6441825B1 (en) 1999-10-04 2002-08-27 Intel Corporation Video token tracking system for animation
US6753849B1 (en) 1999-10-27 2004-06-22 Ken Curran & Associates Universal remote TV mouse
US6519359B1 (en) 1999-10-28 2003-02-11 General Electric Company Range camera controller for acquiring 3D models
US20050037844A1 (en) 2002-10-30 2005-02-17 Nike, Inc. Sigils for use with apparel
JP4403474B2 (en) * 1999-12-09 2010-01-27 ソニー株式会社 Tactile sense presentation mechanism and force-tactile sense presentation device using the same
US6785329B1 (en) 1999-12-21 2004-08-31 Microsoft Corporation Automatic video object extraction
KR100384406B1 (en) 2000-01-24 2003-05-22 (주) 베스트소프트 A program drive divice for computer
US6587835B1 (en) 2000-02-09 2003-07-01 G. Victor Treyz Shopping assistance with handheld computing device
TW522732B (en) 2000-02-15 2003-03-01 Sorceron Inc Method and system for distributing captured motion data over a network
JP2001242986A (en) * 2000-02-29 2001-09-07 Kiyohide Maekawa Information input device
WO2001064481A2 (en) 2000-03-02 2001-09-07 Donnelly Corporation Video mirror systems incorporating an accessory module
JP2001265275A (en) 2000-03-16 2001-09-28 Olympus Optical Co Ltd Picture display device
JP3409160B2 (en) * 2000-04-26 2003-05-26 独立行政法人産業技術総合研究所 Grasping data input device
CN2432987Y (en) * 2000-06-15 2001-06-06 刘崇文 Gloves-type game machine handle
TW527518B (en) 2000-07-14 2003-04-11 Massachusetts Inst Technology Method and system for high resolution, ultra fast, 3-D imaging
US6795068B1 (en) 2000-07-21 2004-09-21 Sony Computer Entertainment Inc. Prop input device and method for mapping an object from a two-dimensional camera image to a three-dimensional space for controlling action in a game program
US7227526B2 (en) 2000-07-24 2007-06-05 Gesturetek, Inc. Video-based image control system
WO2002008849A2 (en) * 2000-07-24 2002-01-31 Herman Ehrenburg Computer-compatible, color-coded manual input system
AUPQ896000A0 (en) 2000-07-24 2000-08-17 Seeing Machines Pty Ltd Facial image processing system
US6873747B2 (en) 2000-07-25 2005-03-29 Farid Askary Method for measurement of pitch in metrology and imaging systems
AUPQ952400A0 (en) 2000-08-18 2000-09-14 Telefonaktiebolaget Lm Ericsson (Publ) Improved method and system of effecting a financial transaction
JP4815661B2 (en) 2000-08-24 2011-11-16 ソニー株式会社 Signal processing apparatus and signal processing method
FR2814965A1 (en) 2000-10-11 2002-04-12 Janick Simeray Remote communicating game personal computer having game transducers sound/ultrasonic signals transmitting and computer with microphone processing detections/identifying signals/computer programme interacting.
US7016532B2 (en) 2000-11-06 2006-03-21 Evryx Technologies Image capture and identification system and process
US7061507B1 (en) 2000-11-12 2006-06-13 Bitboys, Inc. Antialiasing method and apparatus for video applications
KR100385563B1 (en) 2000-12-01 2003-05-27 한국과학기술원 Spectrophotometer With Driving Means And Intensity Of Light Measurement Method
US6751338B1 (en) 2000-12-15 2004-06-15 Cognex Corporation System and method of using range image data with machine vision tools
DE10103922A1 (en) 2001-01-30 2002-08-01 Physoptics Opto Electronic Gmb Interactive data viewing and operating system
US6741741B2 (en) 2001-02-01 2004-05-25 Xerox Corporation System and method for automatically detecting edges of scanned documents
US6789967B1 (en) 2001-02-02 2004-09-14 George Forester Distal chording keyboard
US20020134151A1 (en) 2001-02-05 2002-09-26 Matsushita Electric Industrial Co., Ltd. Apparatus and method for measuring distances
US6746124B2 (en) 2001-02-06 2004-06-08 Robert E. Fischer Flashlight producing uniform high brightness
US7116330B2 (en) 2001-02-28 2006-10-03 Intel Corporation Approximating motion using a three-dimensional model
US20020123381A1 (en) * 2001-03-01 2002-09-05 Tasesa Akeripa Game system
US6724366B2 (en) * 2001-04-03 2004-04-20 Peter James Crawford Thumb actuated x-y input device
JP2002304246A (en) * 2001-04-04 2002-10-18 Nippon Telegr & Teleph Corp <Ntt> Tactile presenting device, and imaginary space system
GB2376397A (en) 2001-06-04 2002-12-11 Hewlett Packard Co Virtual or augmented reality
JP2003018604A (en) 2001-07-04 2003-01-17 Matsushita Electric Ind Co Ltd Image signal encoding method, device thereof and recording medium
US20030006962A1 (en) * 2001-07-06 2003-01-09 Bajramovic Mark B. Computer mouse on a glove
US6931125B2 (en) 2001-07-10 2005-08-16 Gregory N. Smallwood Telephone equipment compatible, twelve button alphanumeric keypad
KR20030009919A (en) 2001-07-24 2003-02-05 삼성전자주식회사 Inputting device for computer game having inertial sense
US7039253B2 (en) 2001-07-24 2006-05-02 Casio Computer Co., Ltd. Image display device, image display method, program, and projection system
US7148922B2 (en) 2001-08-02 2006-12-12 Olympus Optical Co., Ltd. Electronic camera which detects flash influence on an image and controls white balance in accordance with the flash influence
JP3442754B2 (en) 2001-08-10 2003-09-02 株式会社コナミコンピュータエンタテインメント東京 Gun shooting game apparatus, computer control method and program
MXPA04001463A (en) 2001-08-16 2005-04-29 Trans World New York Llc User-personalized media sampling, recommendation and purchasing system using real-time inventory database.
US6669341B2 (en) 2001-08-31 2003-12-30 Metrologic Instruments, Inc. Ophthalmic instrument having wavefront sensor with multiple imaging devices that simultaneously capture multiple images of an array of spots produced by a lenslet array
US7555157B2 (en) 2001-09-07 2009-06-30 Geoff Davidson System and method for transforming graphical images
KR100846761B1 (en) 2001-09-11 2008-07-16 삼성전자주식회사 Pointer control method, pointing apparatus and host apparatus therefor
JP2003114640A (en) 2001-10-04 2003-04-18 Nec Corp Plasma display panel and its driving method
US20030093591A1 (en) 2001-11-09 2003-05-15 David Hohl System and method for fast data transfer to display driver
FR2832892B1 (en) 2001-11-27 2004-04-02 Thomson Licensing Sa SPECIAL EFFECTS VIDEO CAMERA
US20030100363A1 (en) 2001-11-28 2003-05-29 Ali Guiseppe C. Method and apparatus for inputting appearance of computer operator into a computer program
US20040070565A1 (en) 2001-12-05 2004-04-15 Nayar Shree K Method and apparatus for displaying images
KR20030048570A (en) 2001-12-12 2003-06-25 한국전자통신연구원 A keypad assembly with the supplementary buttons and its operating method
US7106366B2 (en) 2001-12-19 2006-09-12 Eastman Kodak Company Image capture system incorporating metadata to facilitate transcoding
US7305114B2 (en) 2001-12-26 2007-12-04 Cognex Technology And Investment Corporation Human/machine interface for a machine vision sensor and method for installing and operating the same
US7436887B2 (en) 2002-02-06 2008-10-14 Playtex Products, Inc. Method and apparatus for video frame sequence-based object tracking
US6990639B2 (en) 2002-02-07 2006-01-24 Microsoft Corporation System and process for controlling electronic components in a ubiquitous computing environment using multimodal integration
US20030160862A1 (en) 2002-02-27 2003-08-28 Charlier Michael L. Apparatus having cooperating wide-angle digital camera system and microphone array
US20030167211A1 (en) 2002-03-04 2003-09-04 Marco Scibora Method and apparatus for digitally marking media content
US7301547B2 (en) 2002-03-22 2007-11-27 Intel Corporation Augmented reality system
FR2837597A1 (en) 2002-03-25 2003-09-26 Thomson Licensing Sa Three-dimensional scene modeling process, involves calculating point of reference image on basis of set of images, of defined minimum and maximum depth values of point depth corresponding to maximum distortion
US6847311B2 (en) 2002-03-28 2005-01-25 Motorola Inc. Method and apparatus for character entry in a wireless communication device
GB2388418A (en) 2002-03-28 2003-11-12 Marcus James Eales Input or pointing device with a camera
US20030199324A1 (en) 2002-04-23 2003-10-23 Xiaoling Wang Apparatus and a method for more realistic shooting video games on computers or similar devices using visible or invisible light
JP2004021345A (en) 2002-06-12 2004-01-22 Toshiba Corp Image processing device and its method
US20040001082A1 (en) 2002-06-26 2004-01-01 Amir Said System and method of interaction with a computer controlled image display system using a projected light source
CA2390072C (en) 2002-06-28 2018-02-27 Adrian Gh Podoleanu Optical mapping apparatus with adjustable depth resolution and multiple functionality
US7227976B1 (en) 2002-07-08 2007-06-05 Videomining Corporation Method and system for real-time facial image enhancement
CN1167996C (en) * 2002-07-11 2004-09-22 庄胜雄 Glove mouse capable of forming virtual tracing ball
US7970147B2 (en) 2004-04-07 2011-06-28 Sony Computer Entertainment Inc. Video game controller with noise canceling logic
US7623115B2 (en) 2002-07-27 2009-11-24 Sony Computer Entertainment Inc. Method and apparatus for light input device
US7613310B2 (en) 2003-08-27 2009-11-03 Sony Computer Entertainment Inc. Audio input system
US7646372B2 (en) 2003-09-15 2010-01-12 Sony Computer Entertainment Inc. Methods and systems for enabling direction detection when interfacing with a computer program
US7697700B2 (en) 2006-05-04 2010-04-13 Sony Computer Entertainment Inc. Noise removal for electronic device with far field microphone on console
US8797260B2 (en) 2002-07-27 2014-08-05 Sony Computer Entertainment Inc. Inertially trackable hand-held controller
US7102615B2 (en) 2002-07-27 2006-09-05 Sony Computer Entertainment Inc. Man-machine interface using a deformable device
US7809145B2 (en) 2006-05-04 2010-10-05 Sony Computer Entertainment Inc. Ultra small microphone array
US8073157B2 (en) 2003-08-27 2011-12-06 Sony Computer Entertainment Inc. Methods and apparatus for targeted sound detection and characterization
US7783061B2 (en) 2003-08-27 2010-08-24 Sony Computer Entertainment Inc. Methods and apparatus for the targeted sound detection
US7545926B2 (en) 2006-05-04 2009-06-09 Sony Computer Entertainment Inc. Echo and noise cancellation
US20040017355A1 (en) 2002-07-24 2004-01-29 Youngtack Shim Cursor control systems and methods
US20070260517A1 (en) 2006-05-08 2007-11-08 Gary Zalewski Profile detection
US8139793B2 (en) 2003-08-27 2012-03-20 Sony Computer Entertainment Inc. Methods and apparatus for capturing audio signals based on a visual image
US20070061413A1 (en) 2005-09-15 2007-03-15 Larsen Eric J System and method for obtaining user information from voices
US7391409B2 (en) 2002-07-27 2008-06-24 Sony Computer Entertainment America Inc. Method and system for applying gearing effects to multi-channel mixed input
US7918733B2 (en) 2002-07-27 2011-04-05 Sony Computer Entertainment America Inc. Multi-input game control mixer
US9393487B2 (en) 2002-07-27 2016-07-19 Sony Interactive Entertainment Inc. Method for mapping movements of a hand-held controller to game commands
US8160269B2 (en) 2003-08-27 2012-04-17 Sony Computer Entertainment Inc. Methods and apparatuses for adjusting a listening area for capturing sounds
US9474968B2 (en) 2002-07-27 2016-10-25 Sony Interactive Entertainment America Llc Method and system for applying gearing effects to visual tracking
US20060256081A1 (en) 2002-07-27 2006-11-16 Sony Computer Entertainment America Inc. Scheme for detecting and tracking user manipulation of a game controller body
US8233642B2 (en) 2003-08-27 2012-07-31 Sony Computer Entertainment Inc. Methods and apparatuses for capturing an audio signal based on a location of the signal
US20060264260A1 (en) 2002-07-27 2006-11-23 Sony Computer Entertainment Inc. Detectable and trackable hand-held controller
US7352359B2 (en) 2002-07-27 2008-04-01 Sony Computer Entertainment America Inc. Method and system for applying gearing effects to inertial tracking
US10086282B2 (en) 2002-07-27 2018-10-02 Sony Interactive Entertainment Inc. Tracking device for use in obtaining information for controlling game program execution
US7760248B2 (en) 2002-07-27 2010-07-20 Sony Computer Entertainment Inc. Selective sound source listening in conjunction with computer interactive processing
US7850526B2 (en) 2002-07-27 2010-12-14 Sony Computer Entertainment America Inc. System for tracking user manipulations within an environment
US7627139B2 (en) 2002-07-27 2009-12-01 Sony Computer Entertainment Inc. Computer image and audio processing of intensity and input devices for interfacing with a computer program
US20070261077A1 (en) 2006-05-08 2007-11-08 Gary Zalewski Using audio/visual environment to select ads on game platform
US20070015559A1 (en) 2002-07-27 2007-01-18 Sony Computer Entertainment America Inc. Method and apparatus for use in determining lack of user activity in relation to a system
US7803050B2 (en) 2002-07-27 2010-09-28 Sony Computer Entertainment Inc. Tracking device with sound emitter for use in obtaining information for controlling game program execution
US20060282873A1 (en) 2002-07-27 2006-12-14 Sony Computer Entertainment Inc. Hand-held controller having detectable elements for tracking purposes
US8313380B2 (en) 2002-07-27 2012-11-20 Sony Computer Entertainment America Llc Scheme for translating movements of a hand-held controller into inputs for a system
US7854655B2 (en) 2002-07-27 2010-12-21 Sony Computer Entertainment America Inc. Obtaining input for controlling execution of a game program
US8686939B2 (en) 2002-07-27 2014-04-01 Sony Computer Entertainment Inc. System, method, and apparatus for three-dimensional input control
GB2392286B (en) 2002-08-19 2004-07-07 Chunghwa Telecom Co Ltd Personal identification system based on the reading of multiple one-dimensional barcodes scanned from scanned from PDA/cell phone screen
US7039199B2 (en) 2002-08-26 2006-05-02 Microsoft Corporation System and process for locating a speaker using 360 degree sound source localization
US6917688B2 (en) 2002-09-11 2005-07-12 Nanyang Technological University Adaptive noise cancelling microphone system
US20040063480A1 (en) 2002-09-30 2004-04-01 Xiaoling Wang Apparatus and a method for more realistic interactive video games on computers or similar devices
EP1411461A1 (en) 2002-10-14 2004-04-21 STMicroelectronics S.r.l. User controlled device for sending control signals to an electric appliance, in particular user controlled pointing device such as mouse or joystick, with 3D-motion detection
US6995666B1 (en) 2002-10-16 2006-02-07 Luttrell Clyde K Cellemetry-operated railroad switch heater
JP2004145448A (en) 2002-10-22 2004-05-20 Toshiba Corp Terminal device, server device, and image processing method
US8206219B2 (en) 2002-10-30 2012-06-26 Nike, Inc. Interactive gaming apparel for interactive gaming
EP2039403B1 (en) 2002-10-30 2016-08-24 NIKE Innovate C.V. Clothes with concealable motion capture markers for computer games
US20040095327A1 (en) 2002-11-14 2004-05-20 Lo Fook Loong Alphanumeric data input system and method
US8012025B2 (en) 2002-12-13 2011-09-06 Applied Minds, Llc Video game controller hub with control input reduction and combination schemes
US7212308B2 (en) 2002-12-18 2007-05-01 Morgan Carol L Interactive photo kiosk
US20040140955A1 (en) 2003-01-21 2004-07-22 Metz Kristofer Erik Input device for a computer and method of operation
JP2004261236A (en) 2003-02-20 2004-09-24 Konami Co Ltd Game system
GB2398690B (en) 2003-02-21 2006-05-10 Sony Comp Entertainment Europe Control of data processing
GB2398691B (en) 2003-02-21 2006-05-31 Sony Comp Entertainment Europe Control of data processing
US7161634B2 (en) 2003-03-06 2007-01-09 Huaya Microelectronics, Ltd. Encoding system for error diffusion dithering
US7665041B2 (en) 2003-03-25 2010-02-16 Microsoft Corporation Architecture for controlling a computer using hand gestures
US7519186B2 (en) 2003-04-25 2009-04-14 Microsoft Corporation Noise reduction systems and methods for voice applications
US7379559B2 (en) 2003-05-28 2008-05-27 Trw Automotive U.S. Llc Method and apparatus for determining an occupant's head location in an actuatable occupant restraining system
US6881147B2 (en) 2003-06-06 2005-04-19 Nyko Technologies, Inc. Video game controller with integrated microphone and speaker
TW571812U (en) 2003-06-11 2004-01-11 Vision Electronics Co Ltd Audio device for TV game machine
EP1489596B1 (en) 2003-06-17 2006-09-13 Sony Ericsson Mobile Communications AB Device and method for voice activity detection
US7156311B2 (en) 2003-07-16 2007-01-02 Scanbuy, Inc. System and method for decoding and analyzing barcodes using a mobile device
DE602004020505D1 (en) 2003-10-28 2009-05-20 Welch Allyn Inc DIGITAL DOCUMENT OPHTHALMOSCOPE
KR20050047329A (en) * 2003-11-17 2005-05-20 한국전자통신연구원 Input information device and method using finger motion
US20050105777A1 (en) 2003-11-18 2005-05-19 Kozlowski William J.Jr. Personalized multimedia summary
US7296007B1 (en) 2004-07-06 2007-11-13 Ailive, Inc. Real time context learning by software agents
US7263462B2 (en) 2004-07-30 2007-08-28 Ailive, Inc. Non-disruptive embedding of specialized elements
US7613610B1 (en) 2005-03-14 2009-11-03 Escription, Inc. Transcription data extraction
US7918732B2 (en) 2005-05-06 2011-04-05 Milton Charles Van Noland Manifold compatibility electronic omni axis human interface
US7927216B2 (en) 2005-09-15 2011-04-19 Nintendo Co., Ltd. Video game system with wireless modular handheld controller
US7620316B2 (en) 2005-11-28 2009-11-17 Navisense Method and device for touchless control of a camera
US7834850B2 (en) 2005-11-29 2010-11-16 Navisense Method and system for object control
US7636645B1 (en) 2007-06-18 2009-12-22 Ailive Inc. Self-contained inertial navigation system for interactive control using movable controllers
JP4481280B2 (en) 2006-08-30 2010-06-16 富士フイルム株式会社 Image processing apparatus and image processing method
US8277316B2 (en) 2006-09-14 2012-10-02 Nintendo Co., Ltd. Method and apparatus for using a common pointing input to control 3D viewpoint and object targeting
US7636697B1 (en) 2007-01-29 2009-12-22 Ailive Inc. Method and system for rapid evaluation of logical expressions
US7937243B2 (en) 2007-08-03 2011-05-03 Ailive, Inc. Method and apparatus for non-disruptive embedding of specialized elements
US20090221368A1 (en) 2007-11-28 2009-09-03 Ailive Inc., Method and system for creating a shared game space for a networked game
US8419545B2 (en) 2007-11-28 2013-04-16 Ailive, Inc. Method and system for controlling movements of objects in a videogame
US8655622B2 (en) 2008-07-05 2014-02-18 Ailive, Inc. Method and apparatus for interpreting orientation invariant motion

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5568928A (en) * 1995-02-01 1996-10-29 Exertron, Inc. Video game controller for use with an exercise apparatus
US5583478A (en) * 1995-03-01 1996-12-10 Renzi; Ronald Virtual environment tactile system
US20020021277A1 (en) * 2000-04-17 2002-02-21 Kramer James F. Interface for controlling a graphical image
WO2002027456A1 (en) * 2000-09-29 2002-04-04 Senseboard Technologies Ab Wearable data input interface

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
SHIMOGA K B ET AL: "Touch and force reflection for telepresence surgery" ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY, 1994. ENGINEERING ADVANCES: NEW OPPORTUNITIES FOR BIOMEDICAL ENGINEERS., PROCEEDINGS OF THE 16TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE BALTIMORE, MD, USA 3-6 NOV. 1994, NEW YORK, NY, USA,IEEE, US, 3 November 1994 (1994-11-03), pages 1049-1050, XP010145559 ISBN: 0-7803-2050-6 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8059089B2 (en) 2004-05-25 2011-11-15 Sony Computer Entertainment Inc. Input device and method, and character input method

Also Published As

Publication number Publication date
US20040012557A1 (en) 2004-01-22
US20110269544A1 (en) 2011-11-03
US7161579B2 (en) 2007-01-09
JP2005537596A (en) 2005-12-08
JP4044114B2 (en) 2008-02-06
EP1523725B1 (en) 2010-09-08
US20070075966A1 (en) 2007-04-05
AU2003249062A8 (en) 2004-02-09
US8035629B2 (en) 2011-10-11
WO2004010370A3 (en) 2004-11-18
CN100397304C (en) 2008-06-25
ATE480814T1 (en) 2010-09-15
DE60334101D1 (en) 2010-10-21
KR20050021500A (en) 2005-03-07
AU2003249062A1 (en) 2004-02-09
CN1682238A (en) 2005-10-12
KR100742029B1 (en) 2007-07-23
EP1523725A2 (en) 2005-04-20

Similar Documents

Publication Publication Date Title
US8035629B2 (en) Hand-held computer interactive device
US10838495B2 (en) Devices for controlling computers based on motions and positions of hands
US10564730B2 (en) Non-collocated haptic cues in immersive environments
US9134797B2 (en) Systems and methods for providing haptic feedback to touch-sensitive input devices
US10579151B2 (en) Controller for finger gesture recognition and method for recognizing finger gesture
Douglas et al. The ergonomics of computer pointing devices
US7131073B2 (en) Force feedback applications based on cursor engagement with graphical targets
US6448977B1 (en) Textures and other spatial sensations for a relative haptic interface device
US20110148607A1 (en) System,device and method for providing haptic technology
CN110325947A (en) Haptic interaction method, tool and system
CN107209582A (en) The method and apparatus of high intuitive man-machine interface
US10617942B2 (en) Controller with haptic feedback
EP1160651A1 (en) Wireless cursor control
Chang et al. TanGo: Exploring Expressive Tangible Interactions on Head-Mounted Displays
US11307671B2 (en) Controller for finger gesture recognition and method for recognizing finger gesture
Wolf Overcoming challenges of controller-based and free-handed interaction in cross-reality
Douglas et al. Factors in Applying Psychomotor Studies to Pointing Devices
Lim 3D interaction design and application development

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ OM PH PL PT RO RU SC SD SE SG SK SL TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 1020057000857

Country of ref document: KR

WWE Wipo information: entry into national phase

Ref document number: 2005505516

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 2003765543

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 1020057000857

Country of ref document: KR

WWE Wipo information: entry into national phase

Ref document number: 20038218194

Country of ref document: CN

WWP Wipo information: published in national office

Ref document number: 2003765543

Country of ref document: EP