US20110182519A1 - Gesture recognition with principal component anaysis - Google Patents
Gesture recognition with principal component anaysis Download PDFInfo
- Publication number
- US20110182519A1 US20110182519A1 US13/013,676 US201113013676A US2011182519A1 US 20110182519 A1 US20110182519 A1 US 20110182519A1 US 201113013676 A US201113013676 A US 201113013676A US 2011182519 A1 US2011182519 A1 US 2011182519A1
- Authority
- US
- United States
- Prior art keywords
- motion
- sensors
- electrical signal
- components
- sensor
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 claims abstract description 36
- 238000012549 training Methods 0.000 claims abstract description 31
- 230000033001 locomotion Effects 0.000 claims description 70
- 238000000513 principal component analysis Methods 0.000 claims description 24
- 238000012545 processing Methods 0.000 claims description 22
- 238000012360 testing method Methods 0.000 claims description 8
- 238000004364 calculation method Methods 0.000 claims description 3
- 230000004044 response Effects 0.000 claims description 2
- 238000007781 pre-processing Methods 0.000 abstract description 14
- 230000003287 optical effect Effects 0.000 description 25
- 238000004422 calculation algorithm Methods 0.000 description 9
- 230000000875 corresponding effect Effects 0.000 description 6
- 239000000969 carrier Substances 0.000 description 5
- 238000001514 detection method Methods 0.000 description 5
- 230000009466 transformation Effects 0.000 description 4
- 230000009471 action Effects 0.000 description 3
- 238000013461 design Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 230000006872 improvement Effects 0.000 description 3
- 239000004973 liquid crystal related substance Substances 0.000 description 3
- 238000005259 measurement Methods 0.000 description 3
- 239000003990 capacitor Substances 0.000 description 2
- 230000001413 cellular effect Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 230000035945 sensitivity Effects 0.000 description 2
- 238000007619 statistical method Methods 0.000 description 2
- 230000004075 alteration Effects 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000001276 controlling effect Effects 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000008878 coupling Effects 0.000 description 1
- 238000010168 coupling process Methods 0.000 description 1
- 238000005859 coupling reaction Methods 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 230000004069 differentiation Effects 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000006266 hibernation Effects 0.000 description 1
- 230000001939 inductive effect Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000010363 phase shift Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/17—Systems in which incident light is modified in accordance with the properties of the material investigated
- G01N21/55—Specular reflectivity
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01J—MEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
- G01J1/00—Photometry, e.g. photographic exposure meter
- G01J1/42—Photometry, e.g. photographic exposure meter using electric radiation detectors
- G01J1/44—Electric circuits
- G01J1/46—Electric circuits using a capacitor
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01J—MEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
- G01J9/00—Measuring optical phase difference; Determining degree of coherence; Measuring optical wavelength
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
- G01S17/08—Systems determining position data of a target for measuring distance only
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L31/00—Semiconductor devices sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation; Processes or apparatus specially adapted for the manufacture or treatment thereof or of parts thereof; Details thereof
- H01L31/08—Semiconductor devices sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation; Processes or apparatus specially adapted for the manufacture or treatment thereof or of parts thereof; Details thereof in which radiation controls flow of current through the device, e.g. photoresistors
- H01L31/10—Semiconductor devices sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation; Processes or apparatus specially adapted for the manufacture or treatment thereof or of parts thereof; Details thereof in which radiation controls flow of current through the device, e.g. photoresistors characterised by at least one potential-jump barrier or surface barrier, e.g. phototransistors
- H01L31/101—Devices sensitive to infrared, visible or ultraviolet radiation
Definitions
- FIG. 1 illustrates an exemplary system that detects motion with an arbitrary spatial arrangement of optical sensors.
- FIG. 2 illustrates an exemplary system that utilizes preprocessing to detect motion with an arbitrary spatial arrangement of optical sensors.
- FIG. 3 illustrates an exemplary methodology for developing a motion recognition system.
- FIG. 4 illustrates an exemplary methodology for preprocessing data from a motion recognition system.
- FIG. 5 illustrates an exemplary methodology for determining the effectiveness of a spatial configuration of one or more sensors in a motion recognition system.
- FIG. 6 illustrates an exemplary functional block diagram for the architecture of the subject disclosure.
- a category of monolithic devices is emerging that allows electronic products to sense their environment. These include diverse devices, such as, accelerometers, monolithic gyroscopes, light sensors and imagers.
- light sensors are one of the simplest and cheapest, allowing their inclusion in multitudes of consumer products, for example, nightlights, cameras, cell phones, laptops etc.
- light sensors can be employed in a wide variety of applications related to proximity sensing, such as, but not limited to, detecting the presence and/or distance of a user to the product for the purpose of controlling power, displays, or other interface options.
- Infrared (IR) detectors utilize IR light to detect objects within the sense area of the IR sensor. Moreover, IR light is transmitted by an IR Light emitting diode (LED) emitter, which reflects off of objects in the surrounding area and the reflections are sensed by a detector. Moreover, the detector can be a diode, e.g., a PIN diode, and/or any other type of apparatus that converts IR light into an electric signal. The sensed signal is analyzed to determine whether an object is present in the sense area and/or to detect motion within the sense area. In determining whether an object is present in the sense area and/or to detect motion within the sense area, conventional systems can employ an algorithm that depends on a specific spatial arrangement of one or more detectors.
- the systems and methods disclosed herein allow an arbitrary spatial arrangement of the detectors.
- PCA principal component analysis
- the system can be trained with any arbitrary spatial arrangement of detectors.
- the system can be trained by demonstrating distinct motions within the sense area.
- algorithms that depend on a specific spatial arrangement of the detectors are handicapped by the specific placement requirement.
- the subject specification is described with respect to IR light, the systems and methods disclosed herein can utilize most any wavelength.
- the subject system and/or methodology can be employed for acoustical proximity detection and/or ultrasonic range finding applications.
- the word “exemplary” is used herein to mean serving as an example, instance, or illustration. Any aspect or design described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects or designs. Rather, use of the word “exemplary” is intended to present concepts in a concrete fashion.
- the term “or” is intended to mean an inclusive “or” rather than an exclusive “or”. That is, unless specified otherwise, or clear from context, “X employs A or B” is intended to mean any of the natural inclusive permutations. That is, if X employs A; X employs B; or X employs both A and B, then “X employs A or B” is satisfied under any of the foregoing instances.
- the articles “a” and “an” as used in this application and the appended claims should generally be construed to mean “one or more” unless specified otherwise or clear from context to be directed to a singular form.
- the word “coupled” is used herein to mean direct or indirect electrical or mechanical coupling.
- the terms “sense area,” “vision field,” “optical field,” and similar terminology are utilized interchangeably in the subject application, unless context warrants particular distinction(s) among the terms.
- the terms “sensor,” “detector,” and similar terminology are utilized interchangeably in the subject application, unless context warrants particular distinction(s) among the terms.
- the optical sensors 102 can include one or more infrared sensors (e.g., photodiodes, such as tuned positive-intrinsic-negative (PIN) diodes) or any sensors that can convert a light signal into an electrical signal.
- infrared sensors e.g., photodiodes, such as tuned positive-intrinsic-negative (PIN) diodes
- PIN tuned positive-intrinsic-negative
- the optical sensors 102 are not limited to utilizing IR light, and, instead, can be any sensor, detector, or combination of sensors and detectors that can utilize light signals of most any wavelength.
- the optical sensors 102 can be arranged in any arbitrary spatial configuration. Although three optical sensors 102 are shown for simplicity, a system can employ more optical sensors 102 or fewer optical sensors 102 in any spatial arrangement. It can be appreciated that the optical sensors 102 can be an array of optical sensors 102 .
- system 100 can be employed in most any light sensing/optical proximity application.
- a laptop computer can detect a gesture (e.g., a tap or a swipe) on a track pad utilizing optical sensors arbitrarily arranged on the track pad.
- a cellular phone or a personal digital assistant (PDA) can detect a gesture (e.g., a tap or a swipe) on a screen utilizing optical sensors arbitrarily arranged on the screen.
- PDA personal digital assistant
- the optical sensors 102 can be coupled to a signal processing circuit 104 to transmit sensed outputs 106 , 108 , 110 .
- the sensed outputs are electrical signals corresponding to sensed light. These electrical signals can vary over time.
- the signal processing circuit 104 can analyze the sensed outputs 106 , 108 , 110 and determine whether an object 112 is present in the sense area and/or detect and identify motion(s) made by the object 112 within the sense area.
- the object 112 can be most any entity of interest, such as, but not limited to, a human entity, an automated component, a device, an item, an animal, etc.
- Sensor outputs (arbitrary units) can be captured over time. The following table shows exemplary sensor outputs values for three sensors captured at four different points in times.
- Peaks can then be recognized from the captured sensors data.
- the system can be programmed such that a sensor output value must be above a certain minimum threshold for the system to detect a peak.
- the threshold can be set to avoid confusing noise with signal detection. For example, as shown below, Sensors 1 and 3 each record a peak above a threshold, while Sensor 2 does not recognize a peak because no value it records is above a threshold.
- the system 100 employs a signal processing circuit 104 that simplifies the sensor data 106 , 108 , 112 collection process.
- the signal processing circuit 104 can be embedded on a single integrated circuit (IC) chip (e.g., a microcontroller).
- IC integrated circuit
- the signal processing circuit 104 need not be embedded on a single IC chip, and, instead, components of the signal processing circuit 104 can be distributed among several IC chips.
- Signal processing circuit 104 can provide automatic, time-correlated peak detection for the optical sensors 102 .
- the signal processing circuit 104 can automatically record time and value of peaks in magnitude and store the values (e.g., in a database). The time and value of the peaks in magnitude can be stored in a first in first out (FIFO) configuration.
- One arrangement of optical sensors 102 that can facilitate automated, time-correlated peak detection is a serial chain (e.g., daisy chain) arrangement of optical sensors 102 . When the optical sensors 102 are in a serial chain, the signal processing circuit 104 can automatically record the time and value of peaks in magnitude and store the peaks in a first in first out (FIFO) configuration.
- the optical sensors 102 can wake up or interrupt the signal processing circuit 104 .
- the signal processing circuit 104 can read out the time and value of peaks from all of the optical sensors 102 that record a peak and reconstruct an image of the event (e.g., motion trajectory) that occurred with the filled in boxes representing peak values. For example:
- System 100 processes a reduced quantity of data because only the peaks are recorded, thus reducing the required computational power. This allows system 100 to be implemented on a low-power IC (e.g., a microcontroller) or any other such low-power device.
- a low-power IC e.g., a microcontroller
- FIG. 2 there illustrated is an example system 200 that utilizes a preprocessing circuit 202 that can preprocess signals 106 , 108 and 110 from optical sensors arranged in an arbitrary spatial consideration before signals 106 , 108 and 110 reach the signal processing circuit 104 .
- the preprocessing circuit 202 and the signal processing circuit 104 can be embodied on a single IC chip 204 .
- the preprocessing circuit 202 and the signal processing circuit 104 are not limited to a single IC chip.
- the preprocessing circuit 202 and the signal processing circuit 104 can, for example, be distributed between one or more IC chips.
- the IC chip 204 can be a low power microcontroller.
- a microcontroller can have a processor core, a memory and programmable input/output (I/O) peripherals.
- the I/O peripherals can include the optical sensors.
- the preprocessing circuit 202 allows the IC chip 204 to operate the signal processing circuit 104 in a low power mode. For example, the preprocessing circuit 202 takes the task of training the sensors away from the processor 104 so that the processor 104 only needs to match a sensor output to training data in order to recognize a gesture.
- the preprocessing circuit 202 employs an algorithm to train the system 200 to recognize peaks in the image data.
- the algorithm can be a training algorithm that can create a set of training images that can facilitate recognition of a specific type(s) of motion(s) sensed in terms of distance, motion trajectory and/or ambient light.
- the algorithm can be one or more of a large class of image processing algorithms to classify types of motions sensed in terms of distance, motion trajectory and/or ambient light.
- System 200 can be programmed to identify and classify distinct types of motions.
- the types of motions can be gestures on a surface, for example, a screen of a cellular phone or a track pad of a laptop computer.
- the system 200 can classify the corresponding data from the optical sensors into one of the distinct types of motion.
- the algorithm can be Principal Component Analysis (PCA) (also known as the “eigenface method”), a statistical method utilized to reduce the dimensionality of a data set.
- PCA applies an orthogonal transformation to an input dataset that includes data corresponding to the distinct motions to be recognized, and creates a set of training images.
- the set of training images can be formulated from principal components, an algorithm that utilizes a smaller set variables than the entire trajectory for the motion.
- the principal components can span a space of reduced dimensionality and can capture variations in the training images in an efficient manner.
- PCA is utilized as a pre-processing computation to reduce the training images to a smallest set of variables necessary to accurately classify new motions. With the training images, the classification problem can reduce to a set of simple vector projections. This is a task simple enough to be carried out in basic microcontroller firmware.
- the motion recognition system can include one or more sensors and/or light emitting diodes arranged in any arbitrary spatial configuration.
- This methodology can be utilized, for example, in connection with liquid crystal display (LCD) screens or other devices where sensor placement is not necessarily known a priori.
- LCD liquid crystal display
- Methodology 300 begins at element 302 where one or more sensors can be placed on a device, like an LCD screen. Placement of the sensors need not be constrained by any specific placement restrictions. Rather, since methodology 300 does not depend on the placement of the sensors, the sensors can be arranged in any arbitrary location on the device.
- the motion recognition system can be trained to recognize one or more specific motions.
- the one or more specific motions can be specific types of gestures.
- the specific types of gestures can correspond to specific actions taken by the device. For example, one type of training can be demonstrating each specific type of gesture. This allows the sensors to be placed in an arbitrary special configuration on the device.
- the motion recognition system can utilize the training data to recognize the specific gestures without requiring significant computational power.
- the motion recognition system can include one or more sensors and/or light emitting diodes arranged in any arbitrary spatial configuration.
- This methodology can be utilized, for example, in connection with liquid crystal display (LCD) screens or other devices where sensor placement is not necessarily known a priori.
- the preprocessing can employ PCA to reduce the dimensionality of the dataset.
- Methodology 400 begins at element 402 where one or more sensors can be placed on a device, like an LCD screen. Placement of the sensors need not be constrained by any specific placement restrictions. Rather, since methodology 300 does not depend on the placement of the sensors, the sensors can be arranged in any arbitrary location on the device.
- motions can include, but are not limited to, gestures (e.g., taping and/or swiping) on or near a surface of a device.
- the one or more specific motions can be specific types of gestures.
- the specific types of gestures can correspond to specific actions taken by the device. For example, each specific type of gesture that will be utilized by the device can be demonstrated.
- the motion control system can record image data corresponding to each of the one or more specific types of motions. Then, at element 408 , the motion control system can employ PCA to preprocess the image data to principal components.
- PCA is a statistical method that can be utilized to reduce the dimensionality of the image data. More specifically, PCA can apply an orthogonal transformation to image data, including examples of each distinct motion to be recognized, creating a set of principal components corresponding to each specific type of gesture.
- PCA is a procedure that can use an orthogonal transformation to convert a set of observations of possibly correlated variables into a set of values of uncorrelated variables called principal components. The number of principal components is less than or equal to the number of original variables. This transformation is defined in such a way that the first principal component has a variance that accounts for as much of the variability in the data as possible, and each succeeding component in turn has a variance with the constraint that it be orthogonal to the preceding components.
- the principal components can be a smaller set of variables than the image data spanning a space of reduced dimensionality.
- the principal components can, therefore, capture variations in the training images in an efficient manner, reducing the training images to a smallest set of variables necessary to accurately classify new motions.
- the motion recognition system can utilize the training data to recognize the specific gestures without requiring significant computational power. Since the training images are already reduced to a smallest set of variables necessary to accurately classify gestures (e.g., a low dimensional vector), the classification problem can reduce to a set of simple vector projections and one Euclidian distance calculation to find the closest training image, a task simple enough to be carried out in basic microcontroller program and/or firmware. For example, the Euclidian distance between the incoming gesture's vector and each training image's vector can be computed and the best-matching training image can be determined as the training image with a smallest distance to the incoming gesture image.
- the Euclidian distance between the incoming gesture's vector and each training image's vector can be computed and the best-matching training image can be determined as the training image with a smallest distance to the incoming gesture image.
- methodology 500 for determining the effectiveness of a spatial configuration of one or more sensors in a motion recognition system.
- methodology 500 can demonstrate when one or more sensors are redundant and/or when too few sensors are utilized for reliable computation. This can maximize the quality of data obtained from the smallest possible number of sensors.
- Methodology 500 begins at element 502 where one or more sensors can be placed on a device.
- the device can be an LCD screen.
- the motion recognition system can include one or more sensors and/or light emitting diodes arranged in any arbitrary spatial configuration on the LCD screen. Since sensor placement is not known a priori, any number of sensors can be placed on the device in any configuration. This can lead to too many or too few sensors to be utilized, minimizing data quality. For example, if too many sensors are utilized, one or more sensors can be redundant. If too few sensors are utilized, computation may not be reliable.
- one or more specific types of motions can be made on the device.
- the one or more specific motions can be specific types of gestures.
- the specific types of gestures can correspond to specific actions taken by the device. For example, each specific type of gesture that will be utilized by the device can be demonstrated.
- the motion control system can record image data corresponding to each of the one or more specific types of motions. Then, at element 508 , the motion control system can reduce the image data to a smaller set of variables.
- the motion recognition system can employ PCA to preprocess the image data to principal components, thereby reducing the dimensionality of the image data to the smallest set of variables (principle components) necessary to accurately classify new motions.
- the motion recognition system can utilize the training data to recognize the specific gestures without requiring significant computational power. Since the training images are already reduced to a smallest set of variables necessary to accurately classify gestures, the classification problem can reduce to a set of simple vector projections and one Euclidian distance calculation to find the closest training image, a task simple enough to be carried out in basic microcontroller program and/or firmware.
- the motion recognition system can determine the effectiveness of the spatial configuration of the sensors.
- the motion recognition system can employ PCA to quantitatively measure the effectiveness of the spatial configuration of the sensors.
- the PCA can demonstrate when one or more of the sensors is redundant or when too few sensors are used for reliable classifications. This allows a maximization of the quality of data obtained from the smallest possible number of sensors. If the training images have weak principal components (e.g., if the resulting vector elements are all about the same magnitude), then the training images do not capture enough variability to achieve good classification. Such weak principal components can imply a poor spatial arrangement of sensors.
- the motion recognition system can employ PCA to determine the best configuration for the one or more sensors.
- the motion recognition system can test one or more potential configurations, apply PCA, and subsequently determine a most effective configuration of sensors.
- the most effective configuration of sensors can be utilized at element 502 to begin method 500 .
- FIG. 6 illustrates an exemplary functional block diagram for the architecture 600 of the subject disclosure.
- the systems e.g., 100 - 200
- the systems can be employed in a reflection based proximity and motion detector with an integrated ambient light sensor (ALS) depicted in FIG. 6 .
- the architecture 600 includes a LED and associated driver circuitry, a photodiode sensor, an analog front end and signal processing, data conversion circuitry, digital control and signal processing, interface circuitry and results display.
- the architecture 600 adaptively optimizes sensitivity and power for a given environment.
- the architecture 600 derives significant performance improvements from its novel ALS structure, and its light emitting diode (LED) driver circuitry is much more efficient than the conventional resistive drive.
- LED light emitting diode
- the architecture 600 includes a Resonant Front End 602 , which includes a Trans-Impedance Resonator (TIR).
- TIR Trans-Impedance Resonator
- the TIR 602 is used in place of the Trans-Inductance Amplifier (TIA), which is conventionally used.
- TIA Trans-Inductance Amplifier
- the TIR 602 plays the same role as a conventional TIA, the TIR 602 gives an order of magnitude improvement in achievable Signal-to-Noise-Ratio (SNR) due to its band-pass nature (e.g., TIR 602 includes an inductor and a capacitor), which allows for an increased range of sensing.
- the capacitor of the TIR can include the capacitance of the photodiode that is being resonated.
- the band-pass nature of the TIR 602 causes the architecture 600 to operate over a narrow band of frequencies, which allows for little noise compared to the wide band TIA.
- the ALS 610 uses a light to frequency converter based on a relaxation oscillator instead of the conventional TIA.
- a relaxation oscillator is an oscillator based upon the relaxation behavior of a physical system.
- An exemplary implementation for the relaxation oscillator of the subject disclosure can be done by connecting the inverting input of an Operational Amplifier (Op Amp) to a fixed bias voltage via a switch and also the photodiode, with the non-inverting input connected to ground. When the switch to the fixed bias voltage is opened, the photodiode will discharge towards ground. The rate of discharge will depend on the photodiode current, which is a measure of the incident ambient light.
- Operational Amp Operational Amplifier
- the Computer Programmable Logic Device When the photodiode is discharged to ground, the Computer Programmable Logic Device (CPLD) resets the oscillator by switching back in the bias voltage.
- the CPLD counts the number of cycles that the photodiode takes to discharge, and thus can estimate the ambient light intensity incident on the photodiode.
- the ALS 610 can be used for ambient light sensing applications and the TIR 602 can be used for proximity and motion sensing applications.
- the output of the Front end 602 is subjected to multiple stages of voltage gain 616 to maximize the SNR of the output signal.
- the voltage gain is adaptively set based on the magnitude of the signal received from the Front end 602 , which is potentially made up of both measureable interferers such as a backscatter and a crosstalk from the LED, and also the desired signal to be measured.
- the interferers are dynamically calibrated out of the measurement to improve the sensitivity.
- the LED drive circuitry 656 uses an inductive drive, which results in a significant efficiency improvement over the conventional resistive drive.
- the architecture 600 also includes a Quad Demodulator with low pass filters (LPFs) 620 , dual [I & Q] Analog to Digital Converters (ADCs) 626 , Digital to Analog Converters (DACs) 630 driven by the bias voltage provided by the Automatic Gain Control module, Oscillator DACs 644 for I and Q carriers, the Universal Serial Bus (USB) processor for Control Interface, and the Computer Programmable Logic Device (CPLD) that include several modules.
- the I and Q relate to In-Phase and Quadrature demodulation components.
- the USB processor can include one or more USB processors.
- the pre-processor as described above, can be associated with a first USB processor and the processor, as described above, can be associated with a second USB processor. Additionally or alternatively, the preprocessor can be associated with a dedicated part of the USB processor and/or a co-processor that can take over a portion of the load of the processor.
- Quadrature amplitude modulation is both an analog and a digital modulation scheme.
- QAM is a modulation scheme in which two sinusoidal carriers, one exactly 90 degrees out of phase with respect to the other, are used to transmit data over a given physical channel. Since the orthogonal carriers occupy the same frequency band and differ by a 90 degree phase shift, each can be modulated independently, transmitted over the same frequency band, and separated by demodulation at the receiver.
- PAM pulse amplitude modulation
- a numerically controlled oscillator can be employed to design a dual-output oscillator that accurately generates the in-phase and quadrature carriers used by a QAM modulator and/or demodulator.
- a filter for example, a raised cosine finite impulse response (FIR) filter can be utilized to filter the data streams before modulation onto the quadrature carriers.
- FIR raised cosine finite impulse response
- the in-phase and quadrature demodulated components are created by multiplying the signal by both a carrier signal, and also a signal 90 degrees out of phase of that carrier, and low pass filtering the result ( 620 in FIG. 6 ).
- the resultant I and Q are a baseband representation of the received signal.
- the phase of the derivative of the I and Q channels can be obtained, which is indicative of the distance of the target to be calculated. Further, the position of a moving object can be accurately identified based on the phase data.
- the resultant phase information can be used as a direct output of the system as a measure of distance/position, and/or can be used to reconstruct the static component of the signal and allow the calibration of a non-derivative TOF measurement.
- the architecture 600 of the subject disclosure can be used in many applications including computers, automotive, industrial, television displays and others.
- the architecture 600 can be used to detect that a user has entered the room and automatically cause a laptop computer in hibernation mode to wake up and enter into the active mode so that the user can use it.
- the architecture 600 of the subject disclosure can be used to automatically and adaptively adjust the intensity of a liquid crystal display (LCD) based on the ambient lighting conditions.
- the architecture 600 can perform motion and proximity sensing at a range of up to 1-2 meters.
- the architecture 600 of the subject disclosure can perform its operations by using less than twenty milli-watts (mW) of power.
- mW milli-watts
- the entire architecture 600 can be implemented in a single integrated circuit chip (IC).
- all components of the architecture 600 can be implemented in the IC except for the two inductors for the TIR 602 and the LED driver circuitry 656 and the LED, which can be implemented outside the IC.
- all components of the architecture 600 can be implemented in the IC except for the TIR 602 inductor, the LED and the inductor and the resistor for the LED driver circuitry, which can be implemented outside the IC.
- various components of the architecture 600 can be located inside or outside the IC.
- the terms (including a reference to a “means”) used to describe such components are intended to correspond, unless otherwise indicated, to any component which performs the specified function of the described component (e.g., a functional equivalent), even though not structurally equivalent to the disclosed structure, which performs the function in the herein illustrated exemplary aspects of the claimed subject matter.
- the disclosure includes a system as well as a computer-readable medium having computer-executable instructions for performing the acts and/or events of the various methods of the claimed subject matter.
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Power Engineering (AREA)
- Electromagnetism (AREA)
- Pathology (AREA)
- Radar, Positioning & Navigation (AREA)
- Analytical Chemistry (AREA)
- Biochemistry (AREA)
- General Health & Medical Sciences (AREA)
- Immunology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Computer Networks & Wireless Communication (AREA)
- Chemical & Material Sciences (AREA)
- Remote Sensing (AREA)
- Optical Radar Systems And Details Thereof (AREA)
- Measurement Of Optical Distance (AREA)
- User Interface Of Digital Computer (AREA)
- Length Measuring Devices By Optical Means (AREA)
- Position Input By Displaying (AREA)
- Photometry And Measurement Of Optical Pulse Characteristics (AREA)
- Amplifiers (AREA)
Abstract
A system and method for identifying a position of a moving object, utilizing sensors arranged in any arbitrary configuration, is provided. A pre-processing method is applied to permit implementation on a low power computing device, such as a microcontroller. The preprocessing creates a set of training data corresponding to different gestures, based on different positions of moving objects. Accordingly, utilizing the training data, different types of gestures can be classified by comparing a sensed signal to the set of training gestures.
Description
- This application claims priority to U.S. Provisional Patent Application Ser. No. 61/298,895, filed on Jan. 27, 2010, and entitled “ARCHITECTURE FOR A REFLECTION BASED LONG RANGE PROXIMITY AND MOTION DETECTOR HAVING AN INTEGRATED AMBIENT LIGHT SENSOR,” the entirety of which is incorporated by reference herein. Further, this application is related to co-pending U.S. patent application Ser. No. 12/979,726, filed on Dec. 28, 2010 (Attorney docket number SE-2773/INTEP105USA), entitled “DISTANCE SENSING BY IQ DOMAIN DIFFERENTIATION OF TIME OF FLIGHT (TOF) MEASUREMENTS,” co-pending U.S. patent application Ser. No. ______, filed on ______ (Attorney docket number SE-2874-AN/INTEP105USB), entitled “DIRECT CURRENT (DC) CORRECTION CIRCUIT FOR A TIME OF FLIGHT (TOF) PHOTODIODE FRONT END”, co-pending U.S. patent application Ser. No. ______, filed on ______ (Attorney docket number SE-2785-AN/INTEP105USC), entitled “PHOTODIODE FRONT END WITH IMPROVED POWER SUPPLY REJECTION RATIO (PSRR),” co-pending U.S. patent application Ser. No. ______, filed on ______ (Attorney docket number SE-2877-AN/INTEP105USD), entitled “AUTOMATIC CALIBRATION TECHNIQUE FOR TIME OF FLIGHT (TOF) TRANSCEIVERS,” and co-pending U.S. patent application Ser. No. ______, filed on ______ (Attorney docket number SE-2877-AN/INTEP105USE), entitled “SERIAL-CHAINING PROXIMITY SENSORS FOR GESTURE RECOGNITION.” The entireties of each of the foregoing applications are incorporated herein by reference.
-
FIG. 1 illustrates an exemplary system that detects motion with an arbitrary spatial arrangement of optical sensors. -
FIG. 2 illustrates an exemplary system that utilizes preprocessing to detect motion with an arbitrary spatial arrangement of optical sensors. -
FIG. 3 illustrates an exemplary methodology for developing a motion recognition system. -
FIG. 4 illustrates an exemplary methodology for preprocessing data from a motion recognition system. -
FIG. 5 illustrates an exemplary methodology for determining the effectiveness of a spatial configuration of one or more sensors in a motion recognition system. -
FIG. 6 illustrates an exemplary functional block diagram for the architecture of the subject disclosure. - A category of monolithic devices is emerging that allows electronic products to sense their environment. These include diverse devices, such as, accelerometers, monolithic gyroscopes, light sensors and imagers. In particular, light sensors are one of the simplest and cheapest, allowing their inclusion in multitudes of consumer products, for example, nightlights, cameras, cell phones, laptops etc. Typically, light sensors can be employed in a wide variety of applications related to proximity sensing, such as, but not limited to, detecting the presence and/or distance of a user to the product for the purpose of controlling power, displays, or other interface options.
- Infrared (IR) detectors utilize IR light to detect objects within the sense area of the IR sensor. Moreover, IR light is transmitted by an IR Light emitting diode (LED) emitter, which reflects off of objects in the surrounding area and the reflections are sensed by a detector. Moreover, the detector can be a diode, e.g., a PIN diode, and/or any other type of apparatus that converts IR light into an electric signal. The sensed signal is analyzed to determine whether an object is present in the sense area and/or to detect motion within the sense area. In determining whether an object is present in the sense area and/or to detect motion within the sense area, conventional systems can employ an algorithm that depends on a specific spatial arrangement of one or more detectors.
- The systems and methods disclosed herein allow an arbitrary spatial arrangement of the detectors. Applying principal component analysis (PCA), the system can be trained with any arbitrary spatial arrangement of detectors. For example, the system can be trained by demonstrating distinct motions within the sense area. In contrast, algorithms that depend on a specific spatial arrangement of the detectors are handicapped by the specific placement requirement. It can be appreciated that although the subject specification is described with respect to IR light, the systems and methods disclosed herein can utilize most any wavelength. As an example, the subject system and/or methodology can be employed for acoustical proximity detection and/or ultrasonic range finding applications.
- The subject matter is described with reference to the drawings, wherein like reference numerals are used to refer to like elements throughout. In the following description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the subject disclosure. It may be evident, however, that the subject matter may be practiced without these specific details. In other instances, well-known structures and devices are shown in block diagram form in order to facilitate describing the subject disclosure. Many modifications may be made to this configuration without departing from the scope or spirit of the claimed subject matter.
- Moreover, the word “exemplary” is used herein to mean serving as an example, instance, or illustration. Any aspect or design described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects or designs. Rather, use of the word “exemplary” is intended to present concepts in a concrete fashion. As used in this application, the term “or” is intended to mean an inclusive “or” rather than an exclusive “or”. That is, unless specified otherwise, or clear from context, “X employs A or B” is intended to mean any of the natural inclusive permutations. That is, if X employs A; X employs B; or X employs both A and B, then “X employs A or B” is satisfied under any of the foregoing instances. In addition, the articles “a” and “an” as used in this application and the appended claims should generally be construed to mean “one or more” unless specified otherwise or clear from context to be directed to a singular form. In addition, the word “coupled” is used herein to mean direct or indirect electrical or mechanical coupling. Further, the terms “sense area,” “vision field,” “optical field,” and similar terminology are utilized interchangeably in the subject application, unless context warrants particular distinction(s) among the terms. Moreover, the terms “sensor,” “detector,” and similar terminology are utilized interchangeably in the subject application, unless context warrants particular distinction(s) among the terms.
- Referring initially to
FIG. 1 , there illustrated is anexample system 100 that employsoptical sensors 102 to sense distance, motion and/or ambient light, according to an aspect of the subject specification. In one aspect, theoptical sensors 102 can include one or more infrared sensors (e.g., photodiodes, such as tuned positive-intrinsic-negative (PIN) diodes) or any sensors that can convert a light signal into an electrical signal. It can be appreciated that theoptical sensors 102 are not limited to utilizing IR light, and, instead, can be any sensor, detector, or combination of sensors and detectors that can utilize light signals of most any wavelength. - The
optical sensors 102 can be arranged in any arbitrary spatial configuration. Although threeoptical sensors 102 are shown for simplicity, a system can employ moreoptical sensors 102 or feweroptical sensors 102 in any spatial arrangement. It can be appreciated that theoptical sensors 102 can be an array ofoptical sensors 102. - In general,
system 100 can be employed in most any light sensing/optical proximity application. For example, a laptop computer can detect a gesture (e.g., a tap or a swipe) on a track pad utilizing optical sensors arbitrarily arranged on the track pad. In another example, a cellular phone or a personal digital assistant (PDA) can detect a gesture (e.g., a tap or a swipe) on a screen utilizing optical sensors arbitrarily arranged on the screen. - The
optical sensors 102 can be coupled to asignal processing circuit 104 to transmitsensed outputs - The
signal processing circuit 104 can analyze thesensed outputs object 112 is present in the sense area and/or detect and identify motion(s) made by theobject 112 within the sense area. Theobject 112 can be most any entity of interest, such as, but not limited to, a human entity, an automated component, a device, an item, an animal, etc. Sensor outputs (arbitrary units) can be captured over time. The following table shows exemplary sensor outputs values for three sensors captured at four different points in times. -
t = 0 t = 1 t = 2 t = 3 Sensor 1 200 2500 500 100 Sensor 2 0 0 100 100 Sensor 3 100 200 500 2200 - Peaks can then be recognized from the captured sensors data. A peak refers to a sensor output value that is larger than other sensor output values, for a set of captured sensor output values. For example, referring to the table above, Sensor 1 registered a peak magnitude at t=1 and Sensor 3 registered a peak magnitude at t=4. When viewed in this manner, the sensor outputs can effectively form a two-dimensional image with the filled in boxes of the chart below representing the peak values. The system can be programmed such that a sensor output value must be above a certain minimum threshold for the system to detect a peak. The threshold can be set to avoid confusing noise with signal detection. For example, as shown below, Sensors 1 and 3 each record a peak above a threshold, while Sensor 2 does not recognize a peak because no value it records is above a threshold.
- The
system 100 employs asignal processing circuit 104 that simplifies thesensor data signal processing circuit 104 can be embedded on a single integrated circuit (IC) chip (e.g., a microcontroller). However, thesignal processing circuit 104 need not be embedded on a single IC chip, and, instead, components of thesignal processing circuit 104 can be distributed among several IC chips. -
Signal processing circuit 104 can provide automatic, time-correlated peak detection for theoptical sensors 102. For example, thesignal processing circuit 104 can automatically record time and value of peaks in magnitude and store the values (e.g., in a database). The time and value of the peaks in magnitude can be stored in a first in first out (FIFO) configuration. One arrangement ofoptical sensors 102 that can facilitate automated, time-correlated peak detection is a serial chain (e.g., daisy chain) arrangement ofoptical sensors 102. When theoptical sensors 102 are in a serial chain, thesignal processing circuit 104 can automatically record the time and value of peaks in magnitude and store the peaks in a first in first out (FIFO) configuration. When an appropriate event occurs (e.g., detection of distance, motion and/or ambient light), theoptical sensors 102 can wake up or interrupt thesignal processing circuit 104. Thesignal processing circuit 104 can read out the time and value of peaks from all of theoptical sensors 102 that record a peak and reconstruct an image of the event (e.g., motion trajectory) that occurred with the filled in boxes representing peak values. For example: -
System 100 processes a reduced quantity of data because only the peaks are recorded, thus reducing the required computational power. This allowssystem 100 to be implemented on a low-power IC (e.g., a microcontroller) or any other such low-power device. - Referring now to
FIG. 2 , there illustrated is anexample system 200 that utilizes apreprocessing circuit 202 that can preprocesssignals signals signal processing circuit 104. - In an aspect, the
preprocessing circuit 202 and thesignal processing circuit 104 can be embodied on asingle IC chip 204. Thepreprocessing circuit 202 and thesignal processing circuit 104 are not limited to a single IC chip. For example, thepreprocessing circuit 202 and thesignal processing circuit 104 can, for example, be distributed between one or more IC chips. - According to an aspect, the
IC chip 204 can be a low power microcontroller. A microcontroller can have a processor core, a memory and programmable input/output (I/O) peripherals. According to an aspect, the I/O peripherals can include the optical sensors. Thepreprocessing circuit 202 allows theIC chip 204 to operate thesignal processing circuit 104 in a low power mode. For example, thepreprocessing circuit 202 takes the task of training the sensors away from theprocessor 104 so that theprocessor 104 only needs to match a sensor output to training data in order to recognize a gesture. - The
preprocessing circuit 202 employs an algorithm to train thesystem 200 to recognize peaks in the image data. For example, the algorithm can be a training algorithm that can create a set of training images that can facilitate recognition of a specific type(s) of motion(s) sensed in terms of distance, motion trajectory and/or ambient light. - The algorithm can be one or more of a large class of image processing algorithms to classify types of motions sensed in terms of distance, motion trajectory and/or ambient light.
System 200 can be programmed to identify and classify distinct types of motions. For example, the types of motions can be gestures on a surface, for example, a screen of a cellular phone or a track pad of a laptop computer. Each time a distinct type of motion occurs, thesystem 200 can classify the corresponding data from the optical sensors into one of the distinct types of motion. - According to an aspect, the algorithm can be Principal Component Analysis (PCA) (also known as the “eigenface method”), a statistical method utilized to reduce the dimensionality of a data set. PCA applies an orthogonal transformation to an input dataset that includes data corresponding to the distinct motions to be recognized, and creates a set of training images. The set of training images can be formulated from principal components, an algorithm that utilizes a smaller set variables than the entire trajectory for the motion. The principal components can span a space of reduced dimensionality and can capture variations in the training images in an efficient manner. PCA is utilized as a pre-processing computation to reduce the training images to a smallest set of variables necessary to accurately classify new motions. With the training images, the classification problem can reduce to a set of simple vector projections. This is a task simple enough to be carried out in basic microcontroller firmware.
- Referring now to
FIG. 3 , there illustrated is anexemplary methodology 300 for developing a motion recognition system. The motion recognition system can include one or more sensors and/or light emitting diodes arranged in any arbitrary spatial configuration. This methodology can be utilized, for example, in connection with liquid crystal display (LCD) screens or other devices where sensor placement is not necessarily known a priori. -
Methodology 300 begins atelement 302 where one or more sensors can be placed on a device, like an LCD screen. Placement of the sensors need not be constrained by any specific placement restrictions. Rather, sincemethodology 300 does not depend on the placement of the sensors, the sensors can be arranged in any arbitrary location on the device. - At
element 304, the motion recognition system can be trained to recognize one or more specific motions. According to an aspect, the one or more specific motions can be specific types of gestures. The specific types of gestures can correspond to specific actions taken by the device. For example, one type of training can be demonstrating each specific type of gesture. This allows the sensors to be placed in an arbitrary special configuration on the device. Atelement 306, the motion recognition system can utilize the training data to recognize the specific gestures without requiring significant computational power. - Referring now to
FIG. 4 , there illustrated is amethodology 400 for preprocessing data from a motion recognition system. The motion recognition system can include one or more sensors and/or light emitting diodes arranged in any arbitrary spatial configuration. This methodology can be utilized, for example, in connection with liquid crystal display (LCD) screens or other devices where sensor placement is not necessarily known a priori. The preprocessing can employ PCA to reduce the dimensionality of the dataset. -
Methodology 400 begins atelement 402 where one or more sensors can be placed on a device, like an LCD screen. Placement of the sensors need not be constrained by any specific placement restrictions. Rather, sincemethodology 300 does not depend on the placement of the sensors, the sensors can be arranged in any arbitrary location on the device. - At
element 404, one or more specific types of motions can be made. For example, motions can include, but are not limited to, gestures (e.g., taping and/or swiping) on or near a surface of a device. According to an aspect, the one or more specific motions can be specific types of gestures. The specific types of gestures can correspond to specific actions taken by the device. For example, each specific type of gesture that will be utilized by the device can be demonstrated. - At
element 406, the motion control system can record image data corresponding to each of the one or more specific types of motions. Then, atelement 408, the motion control system can employ PCA to preprocess the image data to principal components. - PCA is a statistical method that can be utilized to reduce the dimensionality of the image data. More specifically, PCA can apply an orthogonal transformation to image data, including examples of each distinct motion to be recognized, creating a set of principal components corresponding to each specific type of gesture. PCA is a procedure that can use an orthogonal transformation to convert a set of observations of possibly correlated variables into a set of values of uncorrelated variables called principal components. The number of principal components is less than or equal to the number of original variables. This transformation is defined in such a way that the first principal component has a variance that accounts for as much of the variability in the data as possible, and each succeeding component in turn has a variance with the constraint that it be orthogonal to the preceding components. Accordingly, the principal components can be a smaller set of variables than the image data spanning a space of reduced dimensionality. The principal components can, therefore, capture variations in the training images in an efficient manner, reducing the training images to a smallest set of variables necessary to accurately classify new motions.
- Most of the computation complexity associated with
methodology 400 is due to the reduction of the training images to principal components with PCA. This computation can be performed, for example, on a computer. The resulting table of pre-computed principal components can be made small enough to be easily stored in a microcontroller program and/or firmware. - At
element 410, the motion recognition system can utilize the training data to recognize the specific gestures without requiring significant computational power. Since the training images are already reduced to a smallest set of variables necessary to accurately classify gestures (e.g., a low dimensional vector), the classification problem can reduce to a set of simple vector projections and one Euclidian distance calculation to find the closest training image, a task simple enough to be carried out in basic microcontroller program and/or firmware. For example, the Euclidian distance between the incoming gesture's vector and each training image's vector can be computed and the best-matching training image can be determined as the training image with a smallest distance to the incoming gesture image. - Since data from the sensors of the motion control system has a sparse nature (e.g., most data points are zero), very few mathematical operations are required for each classification operation. Further reduction of problem complexity can be achieved by using 1 for peaks and 0 for all other data points.
- Referring now to
FIG. 5 , illustrated is amethodology 500 for determining the effectiveness of a spatial configuration of one or more sensors in a motion recognition system. For example, utilizing PCA,methodology 500 can demonstrate when one or more sensors are redundant and/or when too few sensors are utilized for reliable computation. This can maximize the quality of data obtained from the smallest possible number of sensors. -
Methodology 500 begins atelement 502 where one or more sensors can be placed on a device. For example, the device can be an LCD screen. The motion recognition system can include one or more sensors and/or light emitting diodes arranged in any arbitrary spatial configuration on the LCD screen. Since sensor placement is not known a priori, any number of sensors can be placed on the device in any configuration. This can lead to too many or too few sensors to be utilized, minimizing data quality. For example, if too many sensors are utilized, one or more sensors can be redundant. If too few sensors are utilized, computation may not be reliable. - At
element 504, one or more specific types of motions can be made on the device. According to an aspect, the one or more specific motions can be specific types of gestures. The specific types of gestures can correspond to specific actions taken by the device. For example, each specific type of gesture that will be utilized by the device can be demonstrated. - At
element 506, the motion control system can record image data corresponding to each of the one or more specific types of motions. Then, atelement 508, the motion control system can reduce the image data to a smaller set of variables. For example, the motion recognition system can employ PCA to preprocess the image data to principal components, thereby reducing the dimensionality of the image data to the smallest set of variables (principle components) necessary to accurately classify new motions. - At
element 510, the motion recognition system can utilize the training data to recognize the specific gestures without requiring significant computational power. Since the training images are already reduced to a smallest set of variables necessary to accurately classify gestures, the classification problem can reduce to a set of simple vector projections and one Euclidian distance calculation to find the closest training image, a task simple enough to be carried out in basic microcontroller program and/or firmware. - At
element 512, the motion recognition system can determine the effectiveness of the spatial configuration of the sensors. According to an aspect, the motion recognition system can employ PCA to quantitatively measure the effectiveness of the spatial configuration of the sensors. - The PCA can demonstrate when one or more of the sensors is redundant or when too few sensors are used for reliable classifications. This allows a maximization of the quality of data obtained from the smallest possible number of sensors. If the training images have weak principal components (e.g., if the resulting vector elements are all about the same magnitude), then the training images do not capture enough variability to achieve good classification. Such weak principal components can imply a poor spatial arrangement of sensors.
- In another embodiment, the motion recognition system can employ PCA to determine the best configuration for the one or more sensors. The motion recognition system can test one or more potential configurations, apply PCA, and subsequently determine a most effective configuration of sensors. The most effective configuration of sensors can be utilized at
element 502 to beginmethod 500. - In order to provide additional context for various aspects of the subject specification,
FIG. 6 illustrates an exemplary functional block diagram for thearchitecture 600 of the subject disclosure. In one aspect, the systems (e.g., 100-200) disclosed herein can be employed in a reflection based proximity and motion detector with an integrated ambient light sensor (ALS) depicted inFIG. 6 . Thearchitecture 600 includes a LED and associated driver circuitry, a photodiode sensor, an analog front end and signal processing, data conversion circuitry, digital control and signal processing, interface circuitry and results display. Thearchitecture 600 adaptively optimizes sensitivity and power for a given environment. Moreover, thearchitecture 600 derives significant performance improvements from its novel ALS structure, and its light emitting diode (LED) driver circuitry is much more efficient than the conventional resistive drive. - According to an aspect of the subject disclosure, the
architecture 600 includes aResonant Front End 602, which includes a Trans-Impedance Resonator (TIR). In thearchitecture 600, theTIR 602 is used in place of the Trans-Inductance Amplifier (TIA), which is conventionally used. Although theTIR 602 plays the same role as a conventional TIA, theTIR 602 gives an order of magnitude improvement in achievable Signal-to-Noise-Ratio (SNR) due to its band-pass nature (e.g.,TIR 602 includes an inductor and a capacitor), which allows for an increased range of sensing. The capacitor of the TIR can include the capacitance of the photodiode that is being resonated. The band-pass nature of theTIR 602 causes thearchitecture 600 to operate over a narrow band of frequencies, which allows for little noise compared to the wide band TIA. - According to another aspect of the subject disclosure, the
ALS 610 uses a light to frequency converter based on a relaxation oscillator instead of the conventional TIA. A relaxation oscillator is an oscillator based upon the relaxation behavior of a physical system. An exemplary implementation for the relaxation oscillator of the subject disclosure can be done by connecting the inverting input of an Operational Amplifier (Op Amp) to a fixed bias voltage via a switch and also the photodiode, with the non-inverting input connected to ground. When the switch to the fixed bias voltage is opened, the photodiode will discharge towards ground. The rate of discharge will depend on the photodiode current, which is a measure of the incident ambient light. When the photodiode is discharged to ground, the Computer Programmable Logic Device (CPLD) resets the oscillator by switching back in the bias voltage. The CPLD counts the number of cycles that the photodiode takes to discharge, and thus can estimate the ambient light intensity incident on the photodiode. TheALS 610 can be used for ambient light sensing applications and theTIR 602 can be used for proximity and motion sensing applications. - The output of the
Front end 602 is subjected to multiple stages ofvoltage gain 616 to maximize the SNR of the output signal. The voltage gain is adaptively set based on the magnitude of the signal received from theFront end 602, which is potentially made up of both measureable interferers such as a backscatter and a crosstalk from the LED, and also the desired signal to be measured. The interferers are dynamically calibrated out of the measurement to improve the sensitivity. According to another aspect of the subject disclosure, theLED drive circuitry 656 uses an inductive drive, which results in a significant efficiency improvement over the conventional resistive drive. - The
architecture 600 also includes a Quad Demodulator with low pass filters (LPFs) 620, dual [I & Q] Analog to Digital Converters (ADCs) 626, Digital to Analog Converters (DACs) 630 driven by the bias voltage provided by the Automatic Gain Control module,Oscillator DACs 644 for I and Q carriers, the Universal Serial Bus (USB) processor for Control Interface, and the Computer Programmable Logic Device (CPLD) that include several modules. The I and Q relate to In-Phase and Quadrature demodulation components. - The USB processor can include one or more USB processors. For example, the pre-processor, as described above, can be associated with a first USB processor and the processor, as described above, can be associated with a second USB processor. Additionally or alternatively, the preprocessor can be associated with a dedicated part of the USB processor and/or a co-processor that can take over a portion of the load of the processor.
- Quadrature amplitude modulation (QAM) is both an analog and a digital modulation scheme. Moreover, QAM is a modulation scheme in which two sinusoidal carriers, one exactly 90 degrees out of phase with respect to the other, are used to transmit data over a given physical channel. Since the orthogonal carriers occupy the same frequency band and differ by a 90 degree phase shift, each can be modulated independently, transmitted over the same frequency band, and separated by demodulation at the receiver. Thus, QAM enables data transmission at twice the rate of standard pulse amplitude modulation (PAM) without any degradation in the bit error rate (BER). In one example a numerically controlled oscillator (NCO) can be employed to design a dual-output oscillator that accurately generates the in-phase and quadrature carriers used by a QAM modulator and/or demodulator. A filter, for example, a raised cosine finite impulse response (FIR) filter can be utilized to filter the data streams before modulation onto the quadrature carriers.
- The in-phase and quadrature demodulated components are created by multiplying the signal by both a carrier signal, and also a signal 90 degrees out of phase of that carrier, and low pass filtering the result (620 in
FIG. 6 ). The resultant I and Q are a baseband representation of the received signal. In one example, the phase of the derivative of the I and Q channels can be obtained, which is indicative of the distance of the target to be calculated. Further, the position of a moving object can be accurately identified based on the phase data. Typically, the resultant phase information can be used as a direct output of the system as a measure of distance/position, and/or can be used to reconstruct the static component of the signal and allow the calibration of a non-derivative TOF measurement. - The
architecture 600 of the subject disclosure can be used in many applications including computers, automotive, industrial, television displays and others. For example, thearchitecture 600 can be used to detect that a user has entered the room and automatically cause a laptop computer in hibernation mode to wake up and enter into the active mode so that the user can use it. In another example, thearchitecture 600 of the subject disclosure can be used to automatically and adaptively adjust the intensity of a liquid crystal display (LCD) based on the ambient lighting conditions. According to an aspect of the subject disclosure, thearchitecture 600 can perform motion and proximity sensing at a range of up to 1-2 meters. According to another aspect of the subject disclosure, thearchitecture 600 of the subject disclosure can perform its operations by using less than twenty milli-watts (mW) of power. - In one embodiment of the subject disclosure, the
entire architecture 600 can be implemented in a single integrated circuit chip (IC). In another embodiment of the subject disclosure, all components of thearchitecture 600 can be implemented in the IC except for the two inductors for theTIR 602 and theLED driver circuitry 656 and the LED, which can be implemented outside the IC. In yet another embodiment of the subject disclosure, all components of thearchitecture 600 can be implemented in the IC except for theTIR 602 inductor, the LED and the inductor and the resistor for the LED driver circuitry, which can be implemented outside the IC. In still another embodiment of the subject disclosure, various components of thearchitecture 600 can be located inside or outside the IC. - What has been described above includes examples of the subject disclosure. It is, of course, not possible to describe every conceivable combination of components or methodologies for purposes of describing the claimed subject matter, but many further combinations and permutations of the subject disclosure are possible. Accordingly, the claimed subject matter is intended to embrace all such alterations, modifications, and variations that fall within the spirit and scope of the appended claims.
- In particular and in regard to the various functions performed by the above described components, devices, circuits, systems and the like, the terms (including a reference to a “means”) used to describe such components are intended to correspond, unless otherwise indicated, to any component which performs the specified function of the described component (e.g., a functional equivalent), even though not structurally equivalent to the disclosed structure, which performs the function in the herein illustrated exemplary aspects of the claimed subject matter. In this regard, it will also be recognized that the disclosure includes a system as well as a computer-readable medium having computer-executable instructions for performing the acts and/or events of the various methods of the claimed subject matter.
- The aforementioned systems/circuits/modules have been described with respect to interaction between several components. It can be appreciated that such systems/circuits/modules and components can include those components or specified sub-components, some of the specified components or sub-components, and/or additional components, and according to various permutations and combinations of the foregoing. Sub-components can also be implemented as components communicatively coupled to other components rather than included within parent components (hierarchical). Additionally, it should be noted that one or more components may be combined into a single component providing aggregate functionality or divided into several separate sub-components, and any one or more middle layers, such as a management layer, may be provided to communicatively couple to such sub-components in order to provide integrated functionality. Any components described herein may also interact with one or more other components not specifically described herein but generally known by those of skill in the art.
- In addition, while a particular feature of the subject disclosure may have been disclosed with respect to only one of several implementations, such feature may be combined with one or more other features of the other implementations as may be desired and advantageous for any given or particular application. Furthermore, to the extent that the terms “includes,” “including,” “has,” “contains,” variants thereof, and other similar words are used in either the detailed description or the claims, these terms are intended to be inclusive in a manner similar to the term “comprising” as an open transition word without precluding any additional or other elements.
Claims (20)
1. An apparatus, comprising:
a processor including a motion detector circuit;
a preprocessor communicably coupled to the processor;
at least one detector that generates an electrical signal in response to detecting at least one predetermined motion;
the preprocessor for identifying one or more components of the electrical signal; and
a motion detector circuit for correlating the one or more components of the electrical signal with the predetermined motion for storage in a database.
2. The apparatus of claim 1 , wherein the at least one detector is arranged in an arbitrarily spatial configuration on the apparatus.
3. The apparatus of claim 1 , wherein the at least one detector is at least one light sensor.
4. The apparatus of claim 3 , wherein the at least one light sensor comprises at least one infrared (IR) sensor.
5. The apparatus of claim 1 , wherein the preprocessor employs Principal Component Analysis (PCA) to identify the one or more components of the electrical signal.
6. The apparatus of claim 5 , wherein preprocessor employs the PCA to reduce the electrical signal into principal components of the electrical signal.
7. The apparatus of claim 5 , wherein the processor is operating in a low power mode.
8. The apparatus of claim 1 , wherein the database is used for motion recognition.
9. The apparatus of claim 8 , further comprising: the motion detector circuit for retrieving the correlation information from the database.
10. The apparatus of claim 9 , further comprising: the motion detector circuit compares an electrical signal corresponding to a detected motion to the database by utilizing at least one Euclidian distance calculation.
11. A method, comprising:
arranging one or more sensors in an arbitrary spatial configuration on a device;
sensing image data corresponding to at least one test motion on the device;
extracting test data from the image data;
correlating the test data with the test motion;
comparing data related to a motion to the test data; and
determining an effectiveness of the arbitrarily spatial configuration.
12. The method of claim 11 , further comprising: rearranging the one or more sensors in a second arbitrary spatial configuration on the device.
13. The method of claim 12 , wherein the rearranging further comprises reducing the number of sensors.
14. The method of claim 12 , wherein the rearranging further comprises increasing the number of sensors.
15. The method of claim 11 , wherein the processing further comprises applying Principal Component Analysis (PCA) to extract the test data from the image data.
16. The method of claim 15 , wherein the applying further comprises reducing a dimensionality of the image data to a smaller dimensionality of the test data.
17. The method of claim 11 , wherein the applying further comprises utilizing at least one computer to apply the PCA.
18. The method of claim 11 , wherein the correlating further comprises calculating at least one Euclidian distance.
19. A system, comprising:
at least one light emitting diode (LED) that emits a frequency modulated signal, wherein at least a portion of the frequency modulated signal reflects back from a moving object;
at least one sensor that generates an electric signal based on the reflected portion of the frequency modulated signal;
a memory that stores training data corresponding to at least one gesture, wherein the training is generated from principal components of an electrical signal generated by the at least one sensor corresponding to the gesture; and
a signal processing circuit that compares the electric signal based on the reflected portion of the frequency modulated signal with the training data, wherein, the signal processing circuit for determining if the electrical signal is indicative of the at least one gesture.
20. The system of claim 19 , wherein the at least one sensor is arranged in an arbitrary spatial location with respect to the at least one LED.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/013,676 US20110182519A1 (en) | 2010-01-27 | 2011-01-25 | Gesture recognition with principal component anaysis |
TW100103046A TW201145075A (en) | 2010-01-27 | 2011-01-27 | Gesture recognition with principal component anaysis |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US29889510P | 2010-01-27 | 2010-01-27 | |
US13/013,676 US20110182519A1 (en) | 2010-01-27 | 2011-01-25 | Gesture recognition with principal component anaysis |
Publications (1)
Publication Number | Publication Date |
---|---|
US20110182519A1 true US20110182519A1 (en) | 2011-07-28 |
Family
ID=44308255
Family Applications (7)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/979,726 Active 2032-09-24 US8760631B2 (en) | 2010-01-27 | 2010-12-28 | Distance sensing by IQ domain differentiation of time of flight (TOF) measurements |
US13/013,173 Active 2031-06-12 US8274037B2 (en) | 2010-01-27 | 2011-01-25 | Automatic calibration technique for time of flight (TOF) transceivers |
US13/013,199 Abandoned US20110180693A1 (en) | 2010-01-27 | 2011-01-25 | Photodiode front end with improved power supply rejection ratio (psrr) |
US13/013,640 Abandoned US20110180709A1 (en) | 2010-01-27 | 2011-01-25 | Serial-chaining proximity sensors for gesture recognition |
US13/013,146 Active 2032-01-13 US8530819B2 (en) | 2010-01-27 | 2011-01-25 | Direct current (DC) correction circuit for a time of flight (TOF) photodiode front end |
US13/013,676 Abandoned US20110182519A1 (en) | 2010-01-27 | 2011-01-25 | Gesture recognition with principal component anaysis |
US14/296,277 Active 2033-06-28 US10031078B2 (en) | 2010-01-27 | 2014-06-04 | Distance sensing by IQ domain differentiation of time of flight (TOF) measurements |
Family Applications Before (5)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/979,726 Active 2032-09-24 US8760631B2 (en) | 2010-01-27 | 2010-12-28 | Distance sensing by IQ domain differentiation of time of flight (TOF) measurements |
US13/013,173 Active 2031-06-12 US8274037B2 (en) | 2010-01-27 | 2011-01-25 | Automatic calibration technique for time of flight (TOF) transceivers |
US13/013,199 Abandoned US20110180693A1 (en) | 2010-01-27 | 2011-01-25 | Photodiode front end with improved power supply rejection ratio (psrr) |
US13/013,640 Abandoned US20110180709A1 (en) | 2010-01-27 | 2011-01-25 | Serial-chaining proximity sensors for gesture recognition |
US13/013,146 Active 2032-01-13 US8530819B2 (en) | 2010-01-27 | 2011-01-25 | Direct current (DC) correction circuit for a time of flight (TOF) photodiode front end |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/296,277 Active 2033-06-28 US10031078B2 (en) | 2010-01-27 | 2014-06-04 | Distance sensing by IQ domain differentiation of time of flight (TOF) measurements |
Country Status (4)
Country | Link |
---|---|
US (7) | US8760631B2 (en) |
CN (3) | CN102822691A (en) |
TW (6) | TWI445993B (en) |
WO (6) | WO2011094361A1 (en) |
Cited By (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120158302A1 (en) * | 2010-12-17 | 2012-06-21 | Hon Hai Precision Industry Co., Ltd. | Light intensity measurement apparatus |
US20120166585A1 (en) * | 2010-12-23 | 2012-06-28 | Electronics And Telecommunications Research Institute | Apparatus and method for accelerating virtual desktop |
US20130241888A1 (en) * | 2012-03-14 | 2013-09-19 | Texas Instruments Incorporated | Detecting Wave Gestures Near an Illuminated Surface |
CN104346043A (en) * | 2014-11-03 | 2015-02-11 | 武汉麦塔威科技有限公司 | Bar code scanning positioning based interaction display method of mobile screen and device of interaction display method |
US20150168553A1 (en) * | 2013-12-16 | 2015-06-18 | Samsung Electronics Co., Ltd. | Event filtering device and motion recognition device thereof |
US9194741B2 (en) | 2013-09-06 | 2015-11-24 | Blackberry Limited | Device having light intensity measurement in presence of shadows |
US9256290B2 (en) | 2013-07-01 | 2016-02-09 | Blackberry Limited | Gesture detection using ambient light sensors |
US9304596B2 (en) | 2013-07-24 | 2016-04-05 | Blackberry Limited | Backlight for touchless gesture detection |
US9323336B2 (en) | 2013-07-01 | 2016-04-26 | Blackberry Limited | Gesture detection using ambient light sensors |
US9342671B2 (en) | 2013-07-01 | 2016-05-17 | Blackberry Limited | Password by touch-less gesture |
US9367137B2 (en) | 2013-07-01 | 2016-06-14 | Blackberry Limited | Alarm operation by touch-less gesture |
US9398221B2 (en) | 2013-07-01 | 2016-07-19 | Blackberry Limited | Camera control using ambient light sensors |
US9405461B2 (en) | 2013-07-09 | 2016-08-02 | Blackberry Limited | Operating a device using touchless and touchscreen gestures |
US9423913B2 (en) | 2013-07-01 | 2016-08-23 | Blackberry Limited | Performance control of ambient light sensors |
US9465448B2 (en) | 2013-07-24 | 2016-10-11 | Blackberry Limited | Backlight for touchless gesture detection |
US9489051B2 (en) | 2013-07-01 | 2016-11-08 | Blackberry Limited | Display navigation using touch-less gestures |
US10031078B2 (en) | 2010-01-27 | 2018-07-24 | Intersil Americas LLC | Distance sensing by IQ domain differentiation of time of flight (TOF) measurements |
US20200200898A1 (en) * | 2018-12-19 | 2020-06-25 | Semiconductor Components Industries, Llc | Acoustic distance measuring circuit and method for low frequency modulated (lfm) chirp signals |
US11240048B2 (en) * | 2019-03-06 | 2022-02-01 | Marvell Asia Pte, Ltd. | Systems and methods for waking a network interface device in a low power mode |
Families Citing this family (107)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
USRE46672E1 (en) | 2006-07-13 | 2018-01-16 | Velodyne Lidar, Inc. | High definition LiDAR system |
US8384559B2 (en) * | 2010-04-13 | 2013-02-26 | Silicon Laboratories Inc. | Sensor device with flexible interface and updatable information store |
EP2418512A1 (en) * | 2010-07-30 | 2012-02-15 | Mechaless Systems GmbH | Optoelectronic measuring assembly with compensation for external light sources |
JP5566934B2 (en) * | 2011-03-23 | 2014-08-06 | 株式会社東芝 | Voltage output circuit and active cable |
US8604436B1 (en) * | 2011-03-24 | 2013-12-10 | Maxim Integrated Products, Inc. | Proximity sensor device |
GB2490108B (en) * | 2011-04-13 | 2018-01-17 | Nokia Technologies Oy | A method, apparatus and computer program for user control of a state of an apparatus |
US9063704B2 (en) * | 2011-05-05 | 2015-06-23 | Net Power And Light, Inc. | Identifying gestures using multiple sensors |
US8988660B2 (en) * | 2011-06-29 | 2015-03-24 | Silicon Laboratories Inc. | Optical detector |
US9298333B2 (en) | 2011-12-22 | 2016-03-29 | Smsc Holdings S.A.R.L. | Gesturing architecture using proximity sensing |
US8994926B2 (en) * | 2012-02-14 | 2015-03-31 | Intersil Americas LLC | Optical proximity sensors using echo cancellation techniques to detect one or more objects |
TWI470485B (en) * | 2012-03-29 | 2015-01-21 | Wistron Corp | Stylus capable of detecting pressure on a tip |
US20130257583A1 (en) * | 2012-03-29 | 2013-10-03 | Motorola Mobility, Inc. | Wireless communication device and method with ultrasonic detection |
US9619036B2 (en) * | 2012-05-11 | 2017-04-11 | Comcast Cable Communications, Llc | System and methods for controlling a user experience |
TWI484379B (en) * | 2012-06-01 | 2015-05-11 | Pixart Imaging Inc | Optical detecting device |
TWI465753B (en) * | 2012-08-15 | 2014-12-21 | Generalplus Technology Inc | Position identification system and method and system and method for gesture identification thereof |
JP6109943B2 (en) * | 2012-09-13 | 2017-04-05 | エムビーディーエー・ユーケー・リミテッド | Apparatus and method for sensing room occupancy |
KR101338459B1 (en) * | 2012-10-19 | 2013-12-10 | 고려대학교 산학협력단 | Sensor unit |
US9081571B2 (en) | 2012-11-29 | 2015-07-14 | Amazon Technologies, Inc. | Gesture detection management for an electronic device |
US20140346361A1 (en) * | 2013-05-23 | 2014-11-27 | Yibing M. WANG | Time-of-flight pixels also sensing proximity and/or detecting motion in imaging devices & methods |
KR102102702B1 (en) * | 2013-06-19 | 2020-04-21 | 삼성전자주식회사 | Unit pixel of image sensor and image sensor having the same |
KR20150010230A (en) * | 2013-07-18 | 2015-01-28 | 삼성전자주식회사 | Method and apparatus for generating color image and depth image of an object using singular filter |
JP6207321B2 (en) * | 2013-09-26 | 2017-10-04 | ローム株式会社 | Optical sensor device |
US9250714B2 (en) * | 2013-11-27 | 2016-02-02 | Intersil Americas LLC | Optical proximity detectors |
JP6146295B2 (en) * | 2013-12-26 | 2017-06-14 | 株式会社豊田中央研究所 | Radar apparatus and speed direction measuring method |
US9681123B2 (en) | 2014-04-04 | 2017-06-13 | Microsoft Technology Licensing, Llc | Time-of-flight phase-offset calibration |
US9523765B2 (en) | 2014-07-14 | 2016-12-20 | Omnivision Technologies, Inc. | Pixel-level oversampling for a time of flight 3D image sensor with dual range measurements |
WO2016032048A1 (en) * | 2014-08-28 | 2016-03-03 | 엘지전자 주식회사 | Proximity sensor and method for controlling same |
US11284808B2 (en) | 2014-10-11 | 2022-03-29 | Linet Spol. S.R.O. | Device and method for measurement of vital functions, including intracranial pressure, and system and method for collecting data |
US9977512B2 (en) * | 2014-10-24 | 2018-05-22 | Intersil Americas LLC | Open loop correction for optical proximity detectors |
US10795005B2 (en) * | 2014-12-09 | 2020-10-06 | Intersil Americas LLC | Precision estimation for optical proximity detectors |
JP2016150130A (en) * | 2015-02-18 | 2016-08-22 | セイコーエプソン株式会社 | Information acquisition device and information acquisition method |
US9983680B2 (en) | 2015-03-23 | 2018-05-29 | Intel Corporation | Gesture recognition mechanism |
JP6386410B2 (en) * | 2015-03-31 | 2018-09-05 | 新明和工業株式会社 | Motion detection device |
US9667349B1 (en) * | 2015-04-15 | 2017-05-30 | The United States Of America As Represented By The Secretary Of The Navy | Dynamic range extension of heterodyne fiber-optic interferometers via instantaneous carrier measurement |
CN105034802B (en) * | 2015-05-26 | 2017-12-05 | 李洪金 | E-Gas based on infrared distance measurement |
US10048357B2 (en) * | 2015-06-15 | 2018-08-14 | Microsoft Technology Licensing, Llc | Time-of-flight (TOF) system calibration |
US10422870B2 (en) | 2015-06-15 | 2019-09-24 | Humatics Corporation | High precision time of flight measurement system for industrial automation |
CA2989708A1 (en) * | 2015-06-15 | 2016-12-22 | Humatics Corporation | High-precision time of flight measurement system for industrial automation |
US10591592B2 (en) | 2015-06-15 | 2020-03-17 | Humatics Corporation | High-precision time of flight measurement systems |
CA2988658C (en) * | 2015-06-25 | 2023-07-18 | Fresenius Medical Care Holdings, Inc. | Direct light differential measurement system |
US9952676B2 (en) | 2015-06-25 | 2018-04-24 | Intel Corporation | Wearable device with gesture recognition mechanism |
US11143750B2 (en) * | 2015-10-22 | 2021-10-12 | Ams Sensors Singapore Pte. Ltd. | Optical crosstalk calibration for ranging systems |
US10324494B2 (en) | 2015-11-25 | 2019-06-18 | Intel Corporation | Apparatus for detecting electromagnetic field change in response to gesture |
US20170179570A1 (en) | 2015-12-17 | 2017-06-22 | Humatics Corporation | Dual-band antenna on a substrate |
US10627490B2 (en) | 2016-01-31 | 2020-04-21 | Velodyne Lidar, Inc. | Multiple pulse, LIDAR based 3-D imaging |
WO2017164989A1 (en) | 2016-03-19 | 2017-09-28 | Velodyne Lidar, Inc. | Integrated illumination and detection for lidar based 3-d imaging |
WO2017180688A1 (en) | 2016-04-15 | 2017-10-19 | Spectrum Brands, Inc. | Wireless lockset with integrated angle of arrival (aoa) detection |
US10139341B2 (en) | 2016-05-31 | 2018-11-27 | Tt Electronics Plc | Self-calibrating optical detector |
CA3024510C (en) | 2016-06-01 | 2022-10-04 | Velodyne Lidar, Inc. | Multiple pixel scanning lidar |
US10298282B2 (en) | 2016-06-16 | 2019-05-21 | Intel Corporation | Multi-modal sensing wearable device for physiological context measurement |
EP3474733A1 (en) * | 2016-06-22 | 2019-05-01 | Linet Spol. S.R.O. | Medical data collection system and method of use thereof |
US10598783B2 (en) | 2016-07-07 | 2020-03-24 | Microsoft Technology Licensing, Llc | Multi-frequency unwrapping |
CN106405524A (en) * | 2016-08-24 | 2017-02-15 | 上海兰宝传感科技股份有限公司 | Infrared distance measuring sensor |
CN106405565A (en) * | 2016-08-24 | 2017-02-15 | 上海兰宝传感科技股份有限公司 | Unmanned aerial vehicle barrier-avoiding/height-determining system |
US10620295B2 (en) * | 2016-09-14 | 2020-04-14 | Htc Corporation | Positioning signal receiver, positioning system and positioning method of the same |
CN106405567B (en) * | 2016-10-14 | 2018-03-02 | 海伯森技术(深圳)有限公司 | A kind of range-measurement system and its bearing calibration based on TOF |
US10291895B2 (en) | 2016-10-25 | 2019-05-14 | Omnivision Technologies, Inc. | Time of flight photosensor |
CN106546993B (en) * | 2016-11-04 | 2019-04-16 | 武汉万集信息技术有限公司 | A kind of range unit and distance measuring method improving pulse type laser range accuracy |
CA3057988A1 (en) | 2017-03-31 | 2018-10-04 | Velodyne Lidar, Inc. | Integrated lidar illumination power control |
WO2018208843A1 (en) | 2017-05-08 | 2018-11-15 | Velodyne Lidar, Inc. | Lidar data acquisition and control |
TWI621868B (en) * | 2017-06-21 | 2018-04-21 | Univ Kun Shan | System and method for guiding brain waves to blind people |
CN109213385B (en) * | 2017-07-05 | 2021-07-20 | 光宝科技新加坡私人有限公司 | Mobile device and proximity sensing module thereof |
US10575384B2 (en) * | 2017-10-23 | 2020-02-25 | Infineon Technologies Ag | Adaptive transmit light control |
US10944486B2 (en) * | 2017-12-06 | 2021-03-09 | Elenion Technologies, Llc | DC current cancellation scheme for an optical receiver |
US11294041B2 (en) * | 2017-12-08 | 2022-04-05 | Velodyne Lidar Usa, Inc. | Systems and methods for improving detection of a return signal in a light ranging and detection system |
US10393875B2 (en) | 2017-12-19 | 2019-08-27 | Nortek Security & Control Llc | Time of flight based sensor |
USD850946S1 (en) | 2018-02-01 | 2019-06-11 | Tyco Fire & Security Gmbh | Position detector |
US10718147B2 (en) | 2018-04-06 | 2020-07-21 | Tyco Fire & Security Gmbh | Optical displacement detector with adjustable pattern direction |
US20190317196A1 (en) * | 2018-04-17 | 2019-10-17 | Continental Automotive Systems, Inc. | Crosstalk mitigation circuit for lidar pixel receivers |
TWI660591B (en) * | 2018-05-24 | 2019-05-21 | 立積電子股份有限公司 | Phase information extraction circuit and phase information extraction method for object movement |
CN109031192B (en) * | 2018-06-26 | 2020-11-06 | 北京永安信通科技有限公司 | Object positioning method, object positioning device and electronic equipment |
TWI697845B (en) | 2018-07-13 | 2020-07-01 | 緯創資通股份有限公司 | Multi-object tracking method and system |
CN109035345A (en) * | 2018-07-20 | 2018-12-18 | 齐鲁工业大学 | The TOF camera range correction method returned based on Gaussian process |
CN109061609B (en) * | 2018-08-01 | 2021-09-07 | 歌尔光学科技有限公司 | TOF module calibration device and method |
US11073615B2 (en) * | 2018-08-20 | 2021-07-27 | Lite-On Singapore Pte. Ltd. | Proximity sensor module with two sensors |
US10712434B2 (en) | 2018-09-18 | 2020-07-14 | Velodyne Lidar, Inc. | Multi-channel LIDAR illumination driver |
TWI693421B (en) * | 2018-10-24 | 2020-05-11 | 精準基因生物科技股份有限公司 | Time-of-flight ranging device and time-of-flight ranging method |
US11082010B2 (en) | 2018-11-06 | 2021-08-03 | Velodyne Lidar Usa, Inc. | Systems and methods for TIA base current detection and compensation |
US11423572B2 (en) * | 2018-12-12 | 2022-08-23 | Analog Devices, Inc. | Built-in calibration of time-of-flight depth imaging systems |
DE102019134142A1 (en) * | 2018-12-12 | 2020-06-18 | Analog Devices, Inc. | BUILT-IN CALIBRATION OF RUN TIME DEPTH IMAGING SYSTEMS |
US11885958B2 (en) | 2019-01-07 | 2024-01-30 | Velodyne Lidar Usa, Inc. | Systems and methods for a dual axis resonant scanning mirror |
US11500100B2 (en) * | 2019-04-15 | 2022-11-15 | Microsoft Technology Licensing, Llc | Time-of-flight measurements using linear inverse function |
US10819920B1 (en) | 2019-05-22 | 2020-10-27 | Dell Products L.P. | Augmented information handling system user presence detection |
CN112135071B (en) * | 2019-06-25 | 2023-07-28 | 上海耕岩智能科技有限公司 | Image sensing system |
US11317040B2 (en) * | 2019-06-25 | 2022-04-26 | Shanghai Harvest Intelligence Technology Co., Ltd. | Image sensing apparatus |
US10613203B1 (en) | 2019-07-01 | 2020-04-07 | Velodyne Lidar, Inc. | Interference mitigation for light detection and ranging |
TWI739204B (en) * | 2019-07-22 | 2021-09-11 | 財團法人工業技術研究院 | System and method for signal sensing |
TWI717842B (en) * | 2019-09-18 | 2021-02-01 | 茂達電子股份有限公司 | Optical proximity sensor with digital correction circuit and digital correction method thereof |
US11435475B2 (en) | 2019-10-11 | 2022-09-06 | Dell Products L.P. | Information handling system infrared proximity detection with frequency domain modulation |
US11294054B2 (en) | 2019-10-11 | 2022-04-05 | Dell Products L.P. | Information handling system infrared proximity detection with ambient light management |
US11662695B2 (en) | 2019-10-11 | 2023-05-30 | Dell Products L.P. | Information handling system infrared proximity detection with distance reduction detection |
US11435447B2 (en) | 2019-10-11 | 2022-09-06 | Dell Products L.P. | Information handling system proximity sensor with mechanically adjusted field of view |
CN112860053B (en) | 2019-11-28 | 2024-02-20 | 京东方科技集团股份有限公司 | Gesture recognition apparatus, gesture recognition method, computer device, and storage medium |
WO2021109138A1 (en) | 2019-12-06 | 2021-06-10 | 深圳市汇顶科技股份有限公司 | Three-dimensional image sensing system and related electronic device, and time-of-flight ranging method |
US11334146B2 (en) | 2020-01-31 | 2022-05-17 | Dell Products L.P. | Information handling system peripheral enhanced user presence detection |
US11663343B2 (en) | 2020-01-31 | 2023-05-30 | Dell Products L.P. | Information handling system adaptive user presence detection |
US11513813B2 (en) | 2020-01-31 | 2022-11-29 | Dell Products L.P. | Information handling system notification presentation based upon user presence detection |
JP2023518457A (en) * | 2020-03-17 | 2023-05-01 | ベクトン・ディキンソン・アンド・カンパニー | Gain-matched amplifier for photodetection |
US11796715B2 (en) | 2020-06-24 | 2023-10-24 | Sloan Valve Company | Hybrid time-of-flight sensor and IR sensor |
TWI759213B (en) * | 2020-07-10 | 2022-03-21 | 大陸商廣州印芯半導體技術有限公司 | Light sensor and sensing method thereof |
US11280847B1 (en) * | 2020-10-30 | 2022-03-22 | Taiwan Semiconductor Manufacturing Company Ltd. | Circuit, semiconductor device and method for parameter PSRR measurement |
US11808895B2 (en) | 2020-12-11 | 2023-11-07 | Stmicroelectronics (Research & Development) Limited | Methods and devices for crosstalk compensation |
KR102354158B1 (en) | 2021-01-14 | 2022-01-21 | 박천수 | Multi Phase correlation Vector Synthesis Ranging Method and apparatus |
TWI761047B (en) * | 2021-01-22 | 2022-04-11 | 明泰科技股份有限公司 | Method for verifying detection range of image detector |
EP4151958A1 (en) * | 2021-09-17 | 2023-03-22 | Melexis Bulgaria EOOD | Sensing system comprising a chain of sensors |
CN114489229B (en) * | 2021-12-24 | 2023-01-31 | 芯海科技(深圳)股份有限公司 | Drift voltage correction circuit, integrated circuit, measuring device, and electronic apparatus |
TWI805425B (en) * | 2022-06-29 | 2023-06-11 | 中國鋼鐵股份有限公司 | System and method for determining time of light of ultrasound wave |
Citations (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3996590A (en) * | 1961-02-02 | 1976-12-07 | Hammack Calvin M | Method and apparatus for automatically detecting and tracking moving objects and similar applications |
US4458212A (en) * | 1981-12-30 | 1984-07-03 | Mostek Corporation | Compensated amplifier having pole zero tracking |
US4542475A (en) * | 1982-11-23 | 1985-09-17 | Rca Corporation | Sampled data filter system as for a digital TV |
US4551710A (en) * | 1983-04-29 | 1985-11-05 | Cerberus Ag | Method and apparatus for reporting dangerous conditions |
US4648364A (en) * | 1985-07-08 | 1987-03-10 | Wills William H | Engine protection apparatus |
US4942561A (en) * | 1988-06-29 | 1990-07-17 | Kabushikikaisha Topcon | Delay time measuring device |
US5055671A (en) * | 1990-10-03 | 1991-10-08 | Spacelabs, Inc. | Apparatus for detecting transducer movement using a first and second light detector |
US5563701A (en) * | 1993-12-29 | 1996-10-08 | Hyundai Electronics Industries Co., Ltd. | Apparatus and method for measuring distance optically using phase variation |
US5593430A (en) * | 1995-01-27 | 1997-01-14 | Pacesetter, Inc. | Bus system for interconnecting an implantable medical device with a plurality of sensors |
US5828899A (en) * | 1996-01-04 | 1998-10-27 | Compaq Computer Corporation | System for peripheral devices recursively generating unique addresses based on the number of devices connected dependent upon the relative position to the port |
US5892540A (en) * | 1996-06-13 | 1999-04-06 | Rockwell International Corporation | Low noise amplifier for passive pixel CMOS imager |
US5990409A (en) * | 1997-12-26 | 1999-11-23 | Roland Kabushiki Kaisha | Musical apparatus detecting maximum values and/or peak values of reflected light beams to control musical functions |
US6111256A (en) * | 1997-04-10 | 2000-08-29 | Shmuel Hershkovitz & Pinhas Shpater | Infrared motion detection signal sampler |
US6392539B1 (en) * | 1998-07-13 | 2002-05-21 | Honda Giken Kogyo Kabushiki Kaisha | Object detection apparatus |
US20020097743A1 (en) * | 1993-03-09 | 2002-07-25 | Ertugrul Baydar | Integrated digital loop carrier system with virtual tributary mapper circuit |
US6462726B1 (en) * | 1998-02-19 | 2002-10-08 | Matsushita Electric Industrial Co., Ltd. | Video signal processor |
US20030234341A1 (en) * | 2002-06-20 | 2003-12-25 | Osborn Jon V. | Microelectromechanical system optical sun sensor |
US6744248B2 (en) * | 2001-12-08 | 2004-06-01 | Koninklijke Philips Electronics N.V. | Arrangement for detecting motion of an encoder |
US6753950B2 (en) * | 2000-01-26 | 2004-06-22 | Instro Precision Limited | Optical distance measurement |
US20040140961A1 (en) * | 2003-01-17 | 2004-07-22 | Eastman Kodak Company | Oled display and touch screen |
US6803555B1 (en) * | 2001-09-07 | 2004-10-12 | Indigo Systems Corporation | Two-stage auto-zero amplifier circuit for electro-optical arrays |
US6819782B1 (en) * | 1999-06-08 | 2004-11-16 | Matsushita Electric Industrial Co., Ltd. | Device and method for recognizing hand shape and position, and recording medium having program for carrying out the method recorded thereon |
US6836212B2 (en) * | 2002-10-10 | 2004-12-28 | Motorola, Inc. | Method and apparatus for reducing the likelihood of losing a portable electronic device |
US6888938B2 (en) * | 1999-05-11 | 2005-05-03 | Agere Systems Inc. | Dynamically adjustable digital gyrator having extendable feedback for stable DC load line |
US20060120621A1 (en) * | 2000-01-06 | 2006-06-08 | Canon Kabushiki Kaisha | Demodulation and phase estimation of two-dimensional patterns |
US20070013791A1 (en) * | 2005-07-05 | 2007-01-18 | Koichi Kinoshita | Tracking apparatus |
US7212655B2 (en) * | 2000-09-15 | 2007-05-01 | Tumey David M | Fingerprint verification system |
US20070121095A1 (en) * | 2005-11-28 | 2007-05-31 | Robert Lewis | Distance measurement device with short range optics |
US20080119716A1 (en) * | 2006-05-17 | 2008-05-22 | Olga Boric-Lubecke | Determining presence and/or physiological motion of one or more subjects with quadrature doppler radar receiver systems |
US20080205820A1 (en) * | 2007-02-27 | 2008-08-28 | Intersil Americas Inc. | Dynamically configurable multiple wavelength photodetector array for optical storage applications |
US20080256494A1 (en) * | 2007-04-16 | 2008-10-16 | Greenfield Mfg Co Inc | Touchless hand gesture device controller |
US20080266128A1 (en) * | 2007-04-27 | 2008-10-30 | Sensormatic Electronics Corporation | Handheld data capture system with power and safety monitor and method therefore |
US20090006730A1 (en) * | 2007-06-26 | 2009-01-01 | International Business Machines Corporation | Data eye monitor method and apparatus |
US20090027529A1 (en) * | 2007-07-16 | 2009-01-29 | Jung Sang-Il | Image sensor with wide operating range |
US7486386B1 (en) * | 2007-09-21 | 2009-02-03 | Silison Laboratories Inc. | Optical reflectance proximity sensor |
US7532870B2 (en) * | 2004-04-13 | 2009-05-12 | Maxlinear, Inc. | Method and apparatus for DC offset removal |
US7616032B2 (en) * | 2003-12-30 | 2009-11-10 | Hynix Semiconductor Inc. | Internal voltage initializing circuit for use in semiconductor memory device and driving method thereof |
US7620202B2 (en) * | 2003-06-12 | 2009-11-17 | Honda Motor Co., Ltd. | Target orientation estimation using depth sensing |
US7619293B2 (en) * | 2005-06-03 | 2009-11-17 | Sanyo Electric Co., Ltd. | Pin photodiode with improved blue light sensitivity |
US20090295729A1 (en) * | 2008-06-03 | 2009-12-03 | Asustek Computer Inc. | Input device and operation method of computer system |
US20100150399A1 (en) * | 2008-12-12 | 2010-06-17 | Miroslav Svajda | Apparatus and method for optical gesture recognition |
US20100295773A1 (en) * | 2009-05-22 | 2010-11-25 | Rachid Alameh | Electronic device with sensing assembly and method for interpreting offset gestures |
US20110176069A1 (en) * | 2010-01-21 | 2011-07-21 | Intersil Americas Inc. | Systems and methods for projector light beam alignment |
US20110310005A1 (en) * | 2010-06-17 | 2011-12-22 | Qualcomm Incorporated | Methods and apparatus for contactless gesture recognition |
US8086971B2 (en) * | 2006-06-28 | 2011-12-27 | Nokia Corporation | Apparatus, methods and computer program products providing finger-based and hand-based gesture commands for portable electronic device applications |
Family Cites Families (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB1343298A (en) | 1971-07-30 | 1974-01-10 | Mullard Ltd | Crt display systems |
US4644341A (en) | 1984-07-05 | 1987-02-17 | Tri-Tronics Company, Inc. | Contrast indicating arrangement in photoelectric sensing systems |
EP0250534B2 (en) | 1985-12-24 | 1994-09-28 | British Broadcasting Corporation | Bandwidth compression for television signals |
JP2945423B2 (en) | 1989-12-01 | 1999-09-06 | 株式会社アドバンテスト | Color image signal evaluation method |
US5469748A (en) * | 1994-07-20 | 1995-11-28 | Micro Motion, Inc. | Noise reduction filter system for a coriolis flowmeter |
US5742379A (en) * | 1995-11-29 | 1998-04-21 | Reifer; Michael H. | Device and method for electronically measuring distances |
JPH09236662A (en) * | 1996-02-29 | 1997-09-09 | Ushikata Shokai:Kk | Electronic distance meter |
US6127884A (en) | 1999-05-24 | 2000-10-03 | Philips Electronics North America Corp. | Differentiate and multiply based timing recovery in a quadrature demodulator |
US6891440B2 (en) | 2000-10-02 | 2005-05-10 | A. Michael Straub | Quadrature oscillator with phase error correction |
US6677759B2 (en) * | 2001-05-02 | 2004-01-13 | Microchip Technology Incorporated | Method and apparatus for high-voltage battery array monitoring sensors network |
US7016537B2 (en) | 2001-07-02 | 2006-03-21 | Trw Inc. | Vehicle occupant sensor apparatus and method including scanned, phased beam transmission for occupant characteristic determination |
US7288755B1 (en) | 2001-10-19 | 2007-10-30 | Brian P. Platner | Portable handheld artificial light detector |
US6590520B1 (en) * | 2002-02-04 | 2003-07-08 | Lockheed Martin Corporation | Method and system for determining air turbulence using bi-static measurements |
US7184951B2 (en) * | 2002-02-15 | 2007-02-27 | Radiodetection Limted | Methods and systems for generating phase-derivative sound |
DE10305861A1 (en) * | 2003-02-13 | 2004-08-26 | Adam Opel Ag | Motor vehicle device for spatial measurement of a scene inside or outside the vehicle, combines a LIDAR system with an image sensor system to obtain optimum 3D spatial image data |
DE602004031749D1 (en) * | 2003-09-29 | 2011-04-21 | Photosense L L C | FREQUENZDOMÄNENLUMINESZENZINSTRUMENTIERUNG |
JP2005249764A (en) * | 2004-03-02 | 2005-09-15 | Katsumi Mori | Object color measuring system |
JP4199144B2 (en) * | 2004-03-11 | 2008-12-17 | 株式会社東芝 | Weight function generation device, reference signal generation device, transmission signal generation device, signal processing device, and antenna device |
JP2006010506A (en) * | 2004-06-25 | 2006-01-12 | Sharp Corp | Optical ranging sensor and self-running cleaner |
CN1619295A (en) * | 2004-12-10 | 2005-05-25 | 南京农业大学 | Pork colour grading instrument |
US7735037B2 (en) * | 2005-04-15 | 2010-06-08 | Rambus, Inc. | Generating interface adjustment signals in a device-to-device interconnection system |
WO2007029191A2 (en) | 2005-09-08 | 2007-03-15 | Koninklijke Philips Electronics N. V. | Determination of low currents with high dynamic range for optical imaging |
ATE438950T1 (en) * | 2005-10-21 | 2009-08-15 | Nxp Bv | POLAR MODULATION APPARATUS AND METHOD USING FM MODULATION |
JP4799216B2 (en) * | 2006-03-03 | 2011-10-26 | 富士通株式会社 | Imaging device having distance measuring function |
JP4116053B2 (en) | 2006-09-20 | 2008-07-09 | 北陽電機株式会社 | Ranging device |
JP4971744B2 (en) * | 2006-10-18 | 2012-07-11 | パナソニック株式会社 | Spatial information detector using intensity-modulated light |
EP2075935A1 (en) | 2007-12-31 | 2009-07-01 | Motorola, Inc. | A method and apparatus for providing uninterrupted media to a user |
JP5507053B2 (en) * | 2008-03-27 | 2014-05-28 | パナソニック株式会社 | Distance measuring device |
CN101581783B (en) | 2008-05-16 | 2013-04-17 | 深圳市迈测科技有限公司 | Calibration method for phase measurement, device and distance measuring apparatus |
US20100066442A1 (en) * | 2008-09-15 | 2010-03-18 | Fenghao Mu | Method and Apparatus for Tunable Current-Mode Filtering |
JP2010127739A (en) * | 2008-11-27 | 2010-06-10 | Toppan Printing Co Ltd | Spectral sensitivity characteristic measurement apparatus and spectral sensitivity characteristic measurement method |
US8760631B2 (en) | 2010-01-27 | 2014-06-24 | Intersil Americas Inc. | Distance sensing by IQ domain differentiation of time of flight (TOF) measurements |
-
2010
- 2010-12-28 US US12/979,726 patent/US8760631B2/en active Active
-
2011
- 2011-01-25 US US13/013,173 patent/US8274037B2/en active Active
- 2011-01-25 US US13/013,199 patent/US20110180693A1/en not_active Abandoned
- 2011-01-25 US US13/013,640 patent/US20110180709A1/en not_active Abandoned
- 2011-01-25 US US13/013,146 patent/US8530819B2/en active Active
- 2011-01-25 US US13/013,676 patent/US20110182519A1/en not_active Abandoned
- 2011-01-26 CN CN2011800159510A patent/CN102822691A/en active Pending
- 2011-01-26 CN CN201180015570.2A patent/CN102822759B/en active Active
- 2011-01-26 WO PCT/US2011/022646 patent/WO2011094361A1/en active Application Filing
- 2011-01-26 WO PCT/US2011/022650 patent/WO2011094365A1/en active Application Filing
- 2011-01-26 WO PCT/US2011/022644 patent/WO2011094360A1/en active Application Filing
- 2011-01-26 WO PCT/US2011/022647 patent/WO2011094362A1/en active Application Filing
- 2011-01-26 CN CN201610559617.6A patent/CN106052862B/en active Active
- 2011-01-26 WO PCT/US2011/022649 patent/WO2011094364A1/en active Application Filing
- 2011-01-26 WO PCT/US2011/022651 patent/WO2011094366A1/en active Application Filing
- 2011-01-27 TW TW100103043A patent/TWI445993B/en active
- 2011-01-27 TW TW100103040A patent/TWI439716B/en active
- 2011-01-27 TW TW103115498A patent/TWI531809B/en active
- 2011-01-27 TW TW100103042A patent/TW201205047A/en unknown
- 2011-01-27 TW TW100103045A patent/TW201203008A/en unknown
- 2011-01-27 TW TW100103046A patent/TW201145075A/en unknown
-
2014
- 2014-06-04 US US14/296,277 patent/US10031078B2/en active Active
Patent Citations (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3996590A (en) * | 1961-02-02 | 1976-12-07 | Hammack Calvin M | Method and apparatus for automatically detecting and tracking moving objects and similar applications |
US4458212A (en) * | 1981-12-30 | 1984-07-03 | Mostek Corporation | Compensated amplifier having pole zero tracking |
US4542475A (en) * | 1982-11-23 | 1985-09-17 | Rca Corporation | Sampled data filter system as for a digital TV |
US4551710A (en) * | 1983-04-29 | 1985-11-05 | Cerberus Ag | Method and apparatus for reporting dangerous conditions |
US4648364A (en) * | 1985-07-08 | 1987-03-10 | Wills William H | Engine protection apparatus |
US4942561A (en) * | 1988-06-29 | 1990-07-17 | Kabushikikaisha Topcon | Delay time measuring device |
US5055671A (en) * | 1990-10-03 | 1991-10-08 | Spacelabs, Inc. | Apparatus for detecting transducer movement using a first and second light detector |
US20020097743A1 (en) * | 1993-03-09 | 2002-07-25 | Ertugrul Baydar | Integrated digital loop carrier system with virtual tributary mapper circuit |
US5563701A (en) * | 1993-12-29 | 1996-10-08 | Hyundai Electronics Industries Co., Ltd. | Apparatus and method for measuring distance optically using phase variation |
US5593430A (en) * | 1995-01-27 | 1997-01-14 | Pacesetter, Inc. | Bus system for interconnecting an implantable medical device with a plurality of sensors |
US5828899A (en) * | 1996-01-04 | 1998-10-27 | Compaq Computer Corporation | System for peripheral devices recursively generating unique addresses based on the number of devices connected dependent upon the relative position to the port |
US5892540A (en) * | 1996-06-13 | 1999-04-06 | Rockwell International Corporation | Low noise amplifier for passive pixel CMOS imager |
US6111256A (en) * | 1997-04-10 | 2000-08-29 | Shmuel Hershkovitz & Pinhas Shpater | Infrared motion detection signal sampler |
US5990409A (en) * | 1997-12-26 | 1999-11-23 | Roland Kabushiki Kaisha | Musical apparatus detecting maximum values and/or peak values of reflected light beams to control musical functions |
US6462726B1 (en) * | 1998-02-19 | 2002-10-08 | Matsushita Electric Industrial Co., Ltd. | Video signal processor |
US6392539B1 (en) * | 1998-07-13 | 2002-05-21 | Honda Giken Kogyo Kabushiki Kaisha | Object detection apparatus |
US6888938B2 (en) * | 1999-05-11 | 2005-05-03 | Agere Systems Inc. | Dynamically adjustable digital gyrator having extendable feedback for stable DC load line |
US6819782B1 (en) * | 1999-06-08 | 2004-11-16 | Matsushita Electric Industrial Co., Ltd. | Device and method for recognizing hand shape and position, and recording medium having program for carrying out the method recorded thereon |
US20060120621A1 (en) * | 2000-01-06 | 2006-06-08 | Canon Kabushiki Kaisha | Demodulation and phase estimation of two-dimensional patterns |
US6753950B2 (en) * | 2000-01-26 | 2004-06-22 | Instro Precision Limited | Optical distance measurement |
US7212655B2 (en) * | 2000-09-15 | 2007-05-01 | Tumey David M | Fingerprint verification system |
US6803555B1 (en) * | 2001-09-07 | 2004-10-12 | Indigo Systems Corporation | Two-stage auto-zero amplifier circuit for electro-optical arrays |
US6744248B2 (en) * | 2001-12-08 | 2004-06-01 | Koninklijke Philips Electronics N.V. | Arrangement for detecting motion of an encoder |
US20030234341A1 (en) * | 2002-06-20 | 2003-12-25 | Osborn Jon V. | Microelectromechanical system optical sun sensor |
US6836212B2 (en) * | 2002-10-10 | 2004-12-28 | Motorola, Inc. | Method and apparatus for reducing the likelihood of losing a portable electronic device |
US20040140961A1 (en) * | 2003-01-17 | 2004-07-22 | Eastman Kodak Company | Oled display and touch screen |
US7620202B2 (en) * | 2003-06-12 | 2009-11-17 | Honda Motor Co., Ltd. | Target orientation estimation using depth sensing |
US7616032B2 (en) * | 2003-12-30 | 2009-11-10 | Hynix Semiconductor Inc. | Internal voltage initializing circuit for use in semiconductor memory device and driving method thereof |
US7532870B2 (en) * | 2004-04-13 | 2009-05-12 | Maxlinear, Inc. | Method and apparatus for DC offset removal |
US7619293B2 (en) * | 2005-06-03 | 2009-11-17 | Sanyo Electric Co., Ltd. | Pin photodiode with improved blue light sensitivity |
US20070013791A1 (en) * | 2005-07-05 | 2007-01-18 | Koichi Kinoshita | Tracking apparatus |
US20070121095A1 (en) * | 2005-11-28 | 2007-05-31 | Robert Lewis | Distance measurement device with short range optics |
US20080119716A1 (en) * | 2006-05-17 | 2008-05-22 | Olga Boric-Lubecke | Determining presence and/or physiological motion of one or more subjects with quadrature doppler radar receiver systems |
US8086971B2 (en) * | 2006-06-28 | 2011-12-27 | Nokia Corporation | Apparatus, methods and computer program products providing finger-based and hand-based gesture commands for portable electronic device applications |
US20080205820A1 (en) * | 2007-02-27 | 2008-08-28 | Intersil Americas Inc. | Dynamically configurable multiple wavelength photodetector array for optical storage applications |
US20080256494A1 (en) * | 2007-04-16 | 2008-10-16 | Greenfield Mfg Co Inc | Touchless hand gesture device controller |
US20080266128A1 (en) * | 2007-04-27 | 2008-10-30 | Sensormatic Electronics Corporation | Handheld data capture system with power and safety monitor and method therefore |
US20090006730A1 (en) * | 2007-06-26 | 2009-01-01 | International Business Machines Corporation | Data eye monitor method and apparatus |
US20090027529A1 (en) * | 2007-07-16 | 2009-01-29 | Jung Sang-Il | Image sensor with wide operating range |
US7486386B1 (en) * | 2007-09-21 | 2009-02-03 | Silison Laboratories Inc. | Optical reflectance proximity sensor |
US20090295729A1 (en) * | 2008-06-03 | 2009-12-03 | Asustek Computer Inc. | Input device and operation method of computer system |
US20100150399A1 (en) * | 2008-12-12 | 2010-06-17 | Miroslav Svajda | Apparatus and method for optical gesture recognition |
US20100295773A1 (en) * | 2009-05-22 | 2010-11-25 | Rachid Alameh | Electronic device with sensing assembly and method for interpreting offset gestures |
US20110176069A1 (en) * | 2010-01-21 | 2011-07-21 | Intersil Americas Inc. | Systems and methods for projector light beam alignment |
US20110310005A1 (en) * | 2010-06-17 | 2011-12-22 | Qualcomm Incorporated | Methods and apparatus for contactless gesture recognition |
Cited By (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10031078B2 (en) | 2010-01-27 | 2018-07-24 | Intersil Americas LLC | Distance sensing by IQ domain differentiation of time of flight (TOF) measurements |
US20120158302A1 (en) * | 2010-12-17 | 2012-06-21 | Hon Hai Precision Industry Co., Ltd. | Light intensity measurement apparatus |
US20120166585A1 (en) * | 2010-12-23 | 2012-06-28 | Electronics And Telecommunications Research Institute | Apparatus and method for accelerating virtual desktop |
US9122354B2 (en) * | 2012-03-14 | 2015-09-01 | Texas Instruments Incorporated | Detecting wave gestures near an illuminated surface |
US20130241888A1 (en) * | 2012-03-14 | 2013-09-19 | Texas Instruments Incorporated | Detecting Wave Gestures Near an Illuminated Surface |
US9928356B2 (en) | 2013-07-01 | 2018-03-27 | Blackberry Limited | Password by touch-less gesture |
US9423913B2 (en) | 2013-07-01 | 2016-08-23 | Blackberry Limited | Performance control of ambient light sensors |
US9256290B2 (en) | 2013-07-01 | 2016-02-09 | Blackberry Limited | Gesture detection using ambient light sensors |
US9865227B2 (en) | 2013-07-01 | 2018-01-09 | Blackberry Limited | Performance control of ambient light sensors |
US9323336B2 (en) | 2013-07-01 | 2016-04-26 | Blackberry Limited | Gesture detection using ambient light sensors |
US9342671B2 (en) | 2013-07-01 | 2016-05-17 | Blackberry Limited | Password by touch-less gesture |
US9367137B2 (en) | 2013-07-01 | 2016-06-14 | Blackberry Limited | Alarm operation by touch-less gesture |
US9398221B2 (en) | 2013-07-01 | 2016-07-19 | Blackberry Limited | Camera control using ambient light sensors |
US9489051B2 (en) | 2013-07-01 | 2016-11-08 | Blackberry Limited | Display navigation using touch-less gestures |
US9405461B2 (en) | 2013-07-09 | 2016-08-02 | Blackberry Limited | Operating a device using touchless and touchscreen gestures |
US9465448B2 (en) | 2013-07-24 | 2016-10-11 | Blackberry Limited | Backlight for touchless gesture detection |
US9304596B2 (en) | 2013-07-24 | 2016-04-05 | Blackberry Limited | Backlight for touchless gesture detection |
US9194741B2 (en) | 2013-09-06 | 2015-11-24 | Blackberry Limited | Device having light intensity measurement in presence of shadows |
US20150168553A1 (en) * | 2013-12-16 | 2015-06-18 | Samsung Electronics Co., Ltd. | Event filtering device and motion recognition device thereof |
US9927523B2 (en) * | 2013-12-16 | 2018-03-27 | Samsung Electronics Co., Ltd. | Event filtering device and motion recognition device thereof |
CN104346043A (en) * | 2014-11-03 | 2015-02-11 | 武汉麦塔威科技有限公司 | Bar code scanning positioning based interaction display method of mobile screen and device of interaction display method |
US20200200898A1 (en) * | 2018-12-19 | 2020-06-25 | Semiconductor Components Industries, Llc | Acoustic distance measuring circuit and method for low frequency modulated (lfm) chirp signals |
US11885874B2 (en) * | 2018-12-19 | 2024-01-30 | Semiconductor Components Industries, Llc | Acoustic distance measuring circuit and method for low frequency modulated (LFM) chirp signals |
US11240048B2 (en) * | 2019-03-06 | 2022-02-01 | Marvell Asia Pte, Ltd. | Systems and methods for waking a network interface device in a low power mode |
Also Published As
Publication number | Publication date |
---|---|
WO2011094361A1 (en) | 2011-08-04 |
WO2011094364A1 (en) | 2011-08-04 |
WO2011094365A1 (en) | 2011-08-04 |
US20110180709A1 (en) | 2011-07-28 |
CN106052862A (en) | 2016-10-26 |
US20140327900A1 (en) | 2014-11-06 |
TW201202730A (en) | 2012-01-16 |
CN102822691A (en) | 2012-12-12 |
US10031078B2 (en) | 2018-07-24 |
TW201203008A (en) | 2012-01-16 |
CN106052862B (en) | 2018-08-28 |
US20110181861A1 (en) | 2011-07-28 |
US8760631B2 (en) | 2014-06-24 |
WO2011094362A1 (en) | 2011-08-04 |
TW201432283A (en) | 2014-08-16 |
TWI445993B (en) | 2014-07-21 |
TW201145075A (en) | 2011-12-16 |
US20110180693A1 (en) | 2011-07-28 |
CN102822759A (en) | 2012-12-12 |
TWI439716B (en) | 2014-06-01 |
TW201202731A (en) | 2012-01-16 |
CN102822759B (en) | 2014-08-20 |
WO2011094360A1 (en) | 2011-08-04 |
US20110181254A1 (en) | 2011-07-28 |
US20110181892A1 (en) | 2011-07-28 |
TWI531809B (en) | 2016-05-01 |
US8530819B2 (en) | 2013-09-10 |
WO2011094366A1 (en) | 2011-08-04 |
US8274037B2 (en) | 2012-09-25 |
TW201205047A (en) | 2012-02-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20110182519A1 (en) | Gesture recognition with principal component anaysis | |
US8787632B2 (en) | Apparatus and method for reducing noise in fingerprint sensing circuits | |
US20200394386A1 (en) | Optical biometric sensor with automatic gain and exposure control | |
KR101634939B1 (en) | Apparatus and method for optical gesture recognition | |
US9103658B2 (en) | Optical navigation module with capacitive sensor | |
US20100245289A1 (en) | Apparatus and method for optical proximity sensing and touch input control | |
US10121051B2 (en) | Optical apparatus and a method for identifying an object | |
KR102331920B1 (en) | Sensor for motion information, illumination information and proximity information, and method for operating processor using the sensor | |
CN104620207A (en) | Low power operation of an optical touch-sensitive device for detecting multitouch events | |
WO2015054419A1 (en) | Devices, systems, and methods for controlling devices using gestures | |
WO2012103693A1 (en) | Multiple-input touch panel and method for gesture recognition | |
CN114088644A (en) | Sensing device, sensing method and sensing operation assembly | |
CN113302624A (en) | Monitoring activity using depth and multispectral cameras | |
US9377366B2 (en) | Navigation device including thermal sensor | |
US20180373380A1 (en) | Optical control key, operating method thereof, and image sensor | |
US20180307880A1 (en) | Image sensor with range and light-level detection | |
US8462114B2 (en) | Computer navigation devices | |
CN105796051A (en) | Three-dimensional physiological detection system and operation method thereof | |
US9946917B2 (en) | Efficient determination of biometric attribute for fast rejection of enrolled templates and other applications | |
US20220053155A1 (en) | Presence detection in time of flight sensing systems | |
US9897484B1 (en) | Measuring wideband spectrum information in mobile devices via an integrated optical system that uses multiple spectral sensors, multiple light sources and MEMS actuation | |
US9098144B1 (en) | Adaptive ambient light auto-movement blocking in optical navigation modules |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: INTERSIL AMERICAS INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CRADDOCK, C. WARREN;RITTER, DAVID W.;GOLDEN, PHILIP;REEL/FRAME:025696/0142 Effective date: 20110121 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |