US20110180709A1 - Serial-chaining proximity sensors for gesture recognition - Google Patents
Serial-chaining proximity sensors for gesture recognition Download PDFInfo
- Publication number
- US20110180709A1 US20110180709A1 US13/013,640 US201113013640A US2011180709A1 US 20110180709 A1 US20110180709 A1 US 20110180709A1 US 201113013640 A US201113013640 A US 201113013640A US 2011180709 A1 US2011180709 A1 US 2011180709A1
- Authority
- US
- United States
- Prior art keywords
- sensor
- detector
- serial chain
- sensors
- emitter
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 claims abstract description 33
- 238000012545 processing Methods 0.000 claims description 18
- 230000004044 response Effects 0.000 claims description 2
- 238000001514 detection method Methods 0.000 abstract description 13
- 238000012163 sequencing technique Methods 0.000 abstract description 3
- 230000003287 optical effect Effects 0.000 description 55
- 239000000969 carrier Substances 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 4
- 238000013461 design Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 230000006872 improvement Effects 0.000 description 3
- 238000005259 measurement Methods 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 230000006399 behavior Effects 0.000 description 2
- 239000003990 capacitor Substances 0.000 description 2
- 230000001413 cellular effect Effects 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000035945 sensitivity Effects 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000008878 coupling Effects 0.000 description 1
- 238000010168 coupling process Methods 0.000 description 1
- 238000005859 coupling reaction Methods 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 230000004069 differentiation Effects 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000006266 hibernation Effects 0.000 description 1
- 230000001939 inductive effect Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 230000010363 phase shift Effects 0.000 description 1
- 238000000513 principal component analysis Methods 0.000 description 1
- 230000011664 signaling Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
- 239000002699 waste material Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/17—Systems in which incident light is modified in accordance with the properties of the material investigated
- G01N21/55—Specular reflectivity
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01J—MEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
- G01J1/00—Photometry, e.g. photographic exposure meter
- G01J1/42—Photometry, e.g. photographic exposure meter using electric radiation detectors
- G01J1/44—Electric circuits
- G01J1/46—Electric circuits using a capacitor
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01J—MEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
- G01J9/00—Measuring optical phase difference; Determining degree of coherence; Measuring optical wavelength
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
- G01S17/08—Systems determining position data of a target for measuring distance only
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L31/00—Semiconductor devices sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation; Processes or apparatus specially adapted for the manufacture or treatment thereof or of parts thereof; Details thereof
- H01L31/08—Semiconductor devices sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation; Processes or apparatus specially adapted for the manufacture or treatment thereof or of parts thereof; Details thereof in which radiation controls flow of current through the device, e.g. photoresistors
- H01L31/10—Semiconductor devices sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation; Processes or apparatus specially adapted for the manufacture or treatment thereof or of parts thereof; Details thereof in which radiation controls flow of current through the device, e.g. photoresistors characterised by at least one potential-jump barrier or surface barrier, e.g. phototransistors
- H01L31/101—Devices sensitive to infrared, visible or ultraviolet radiation
Definitions
- FIG. 1 illustrates an exemplary system that detects motion utilizing a serial chain of optical sensors.
- FIG. 2 illustrates an exemplary topology for a serial chain utilized by a motion detection system.
- FIG. 3 illustrates an exemplary datagram sent by optical sensors in the serial chain.
- FIG. 4 illustrates an exemplary methodology for assigning addresses to optical sensors in a motion detection system.
- FIG. 5 illustrates an exemplary methodology for automatic emitter sharing between optical sensors in a motion detection system.
- FIG. 6 illustrates an exemplary methodology for automatic time correlation between optical sensors in a motion detection system.
- FIG. 7 illustrates an exemplary functional block diagram for the architecture of the subject disclosure.
- a category of monolithic devices is emerging that allows electronic products to sense their environment. These include diverse devices, such as, accelerometers, monolithic gyroscopes, light sensors and imagers.
- light sensors are one of the simplest and cheapest, allowing their inclusion in multitudes of consumer products, for example, nightlights, cameras, cell phones, laptops etc.
- light sensors can be employed in a wide variety of applications related to proximity sensing, such as, but not limited to, detecting the presence and/or distance of a user to the product for the purpose of controlling power, displays, or other interface options.
- Infrared (IR) detectors utilize IR light to detect objects within the sense area of the IR sensor. Moreover, IR light is transmitted by an IR Light emitting diode (LED) emitter, which reflects off of objects in the surrounding area and the reflections are sensed by a detector. Moreover, the detector can be a diode, e.g., a PIN diode, and/or any other type of apparatus that converts IR light into an electric signal. The sensed signal is analyzed to determine whether an object is present in the sense area and/or to detect motion within the sense area. It can be appreciated that although the subject specification is described with respect to IR light, the systems and methods disclosed herein can utilize most any wavelength. As an example, the subject system and/or methodology can be employed for acoustical proximity detection and/or ultrasonic range finding applications.
- the sensor can be one of multiple proximity sensors arranged on a device to detect motion.
- multiple sensors can be arranged to detect motion, for example, a gesture on a screen.
- Each of the multiple sensors that can sense the gesture, for example, at different times at different locations on the screen.
- Each of the multiple sensors can transmit a sensed signal corresponding to the gesture to be analyzed across a system bus.
- addresses of the multiple sensors are assigned through dedicated pins.
- the systems and methods disclosed herein allow for automatic address assignment for multiple sensors without dedicated pins. Through a serial chain of sensors, access to an emitter shared by the sensors can be automatically sequenced and data from the sensors can be automatically time-correlated without involvement of a controller, conserving power because these capabilities can be done in firmware. In contrast, conventional systems waste power because they cannot perform these actions in firmware, and, instead, must invoke substantial processing power of corresponding devices.
- the word “exemplary” is used herein to mean serving as an example, instance, or illustration. Any aspect or design described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects or designs. Rather, use of the word “exemplary” is intended to present concepts in a concrete fashion.
- the term “or” is intended to mean an inclusive “or” rather than an exclusive “or”. That is, unless specified otherwise, or clear from context, “X employs A or B” is intended to mean any of the natural inclusive permutations. That is, if X employs A; X employs B; or X employs both A and B, then “X employs A or B” is satisfied under any of the foregoing instances.
- the articles “a” and “an” as used in this application and the appended claims should generally be construed to mean “one or more” unless specified otherwise or clear from context to be directed to a singular form.
- the word “coupled” is used herein to mean direct or indirect electrical or mechanical coupling.
- the terms “sense area,” “vision field,” “optical field,” and similar terminology are utilized interchangeably in the subject application, unless context warrants particular distinction(s) among the terms.
- the terms “sensor,” “detector,” and similar terminology are utilized interchangeably in the subject application, unless context warrants particular distinction(s) among the terms.
- the optical sensors 102 can include one or more proximity sensors.
- the optical sensors 102 can include one or more infrared sensors (e.g., photodiodes, such as tuned positive-intrinsic-negative (PIN) diodes).
- the optical sensors 102 can include any sensors that can convert a light signal to an electric signal.
- the optical sensors 102 are not limited to utilizing IR light, and, instead, can be any sensor, detector, or combination of sensors and detectors that can utilize light signals of most any wavelength.
- three optical sensors 102 are shown for simplicity in system 100 , a system can employ more optical sensors 102 or fewer optical sensors 102 . It can be appreciated that the optical sensors 102 can be an array of optical sensors 102 .
- system 100 can be employed in most any light sensing/optical proximity application.
- a laptop computer can detect a gesture (e.g., a tap or a swipe) on a track pad utilizing optical sensors arbitrarily arranged on the track pad.
- a cellular phone or a personal digital assistant (PDA) can detect a gesture (e.g., a tap or a swipe) on a screen utilizing optical sensors arbitrarily arranged on the screen.
- PDA personal digital assistant
- the optical sensors 102 can be coupled to a signal processing circuit 104 to transmit sensed outputs 106 , 108 , 110 .
- the sensed outputs 106 , 108 , 110 are electrical signals corresponding to sensed light. These electrical signals can vary over time.
- the signal processing circuit 104 can analyze the sensed outputs 106 , 108 , 110 and determine whether an object 112 is present in the sense area and/or to detect motion by the object 112 within the sense area.
- the object 112 can be most any entity of interest, such as, but not limited to, a human entity, an automated component, a device, an item, an animal, etc.
- signal processing circuit 104 can be embedded on a single integrated circuit (IC) chip (e.g., a microcontroller). However, the signal processing circuit 104 need not be embedded on a single IC chip, and, instead, components of the signal processing circuit 104 can be distributed among several IC chips.
- the sensed outputs 106 , 108 , 110 can be transmitted to the signal processing circuit 104 via a bus 114 (e.g. a serial bus, such as an Inter Integrated Circuit (“I 2 C”) bus).
- I 2 C Inter Integrated Circuit
- the bus 114 can be an I 2 C bus.
- the I 2 C bus includes two wires: a clock line and a data line.
- the clock line can synchronize data transfers across the I 2 C bus.
- the clock line and the data line are connected to each optical sensor 102 on the I 2 C bus.
- the signal processing circuit 104 a master on the I 2 C bus to the optical sensors 102 , which are slaves.
- the master signal processing circuit 104 drives the clock line.
- the slave optical sensors 102 are unable to initiate transmission or data (e.g. sensed outputs 106 , 108 , 110 ). Instead, data transmission from the slave optical sensors 102 is controlled by the master signal processing circuit 104 and can only begin when initiated by the master signal processing circuit 104 .
- each optical sensor 102 requires a unique address on the bus (for example, the address can be an I 2 C address) for data transfer to take place.
- the address can be a seven bit address.
- addresses are assigned by hardware on the signal processing circuit 104 .
- the addresses can be assigned according to dedicated hardware (e.g., dedicated pins).
- System 100 in contrast, can automatically assign addresses (e.g., I 2 C addresses) to each optical sensor 102 without requiring dedicated hardware.
- System 100 does not require involvement by the signal processing circuit 104 to assign addresses to the optical sensors 102 .
- system 100 configures the sensors 102 in accordance with a serial chain.
- the serial chain can be a daisy chain.
- the serial chain can be utilized to automatically assign unique addresses to each optical sensor 102 . This eliminates the need for the signal processor 104 to be involved with assigning addresses. Eliminating the need for the signal processor 104 to be involved with assigning addresses can conserve power, allowing system 100 to be implemented on a low power a microcontroller.
- serial chain 200 employed by a motion detection system.
- the serial chain 200 can be implemented within system 100 as described above.
- serial chain 200 can be constructed between many more optical sensors 102 .
- serial chain 200 can be constructed between more than 255 optical sensors 102 .
- the number of optical sensors 102 supported by serial chain 200 can be limited only by capabilities of the serial bus.
- Serial chain 200 can be utilized to automatically assign addresses (e.g., I 2 C addresses) to each optical sensor 102 a , 102 b , and 102 c on the serial chain 200 .
- Optical sensors 102 can include a digital input 202 a , 202 b , 202 c and a digital output 204 a , 204 b , 204 c , which are utilized to construct serial chain 200 between the optical sensors 102 a , 102 b , 102 c .
- the digital input 202 a , 202 b , 202 c can be an ENABLE (“EN”) pin and the digital output 204 a , 204 b , 204 c can be an INT pin.
- ENABLE pin and the INT pin can have other purposes including enabling/disabling the associated device and/or signaling an interrupt based on an activity in at least one sensor's 102 sense area.
- Serial chain 200 can be utilized to relay ENABLE input from a first optical sensor to one or more other optical sensors, thereby preserving ENABLE functionality. Additionally, serial chain 200 can relay INT output from one or more optical sensors to a last optical sensor on the serial chain 200 , thereby preserving INT functionality. It is noted that sensors 102 a , 102 b , 102 c can have additional and/or different pins that are not illustrated for the sake of simplicity.
- Serial chain 200 is a unidirectional interface between the output of a first optical sensor and the input of a second optical sensor.
- Serial chain 200 can transmit one or more datagrams of specific format from the first optical sensor to the second optical sensor.
- output 204 c from optical sensor 102 a can be connected to input 202 b of sensor 102 b .
- Output 204 b from optical sensor 102 b can be connected to input 202 c of sensor 102 c .
- Serial chain 200 can automatically sequence access to a single, shared emitter (not shown), like a light emitting diode (LED), without involvement from a controller 206 (e.g., a microcontroller).
- a controller 206 e.g., a microcontroller
- Serial chain 200 can be asynchronous and/or self-clocking, allowing serial chain 200 to automatically time-correlate data from the optical sensors 102 a , 102 b , 102 c .
- One example of a self clocking methodology employed by sensors 102 a , 102 b , 102 c in the serial chain 200 is Manchester coding.
- Manchester coding also known as Phase Encoding, can encode data from one or more of sensors 102 a , 102 b , 102 c with a clock signal that can be recovered from the encoded data.
- each of the optical sensors 102 a , 102 b , 102 c can include a four-sample first-in-first-out (FIFO) that can operate in a “peak detect” mode to store a time and a magnitude of peaks in magnitude of an electrical signal.
- FIFO first-in-first-out
- the peaks in magnitude can correspond to a gesture.
- the magnitude value for the peak and a time code can be pushed into the FIFO.
- the time code can be the last Sequence Value received over the serial chain 200 . Accordingly, time can be sub-divided into 8 millisecond periods, each labeled with a unique 8-bit code.
- the unique 8-bit code can be agreed upon by all sensors 102 a , 102 b , 102 c within the serial chain 200 .
- An interrupt can be sent to a microcontroller indicating the peaks.
- the microcontroller can read the peaks (magnitude values and time code pairs) from each sensor and reconstruct the gesture associated with the peak in magnitude at times corresponding to the peaks.
- the first sensor is responsible for sending datagrams with appropriate time codes down the serial chain.
- the first sensor can include a register that stores the number of sensors in the serial chain for this purpose.
- All of the behaviors are automatic and do not require any interaction with the microcontroller.
- the device e.g., laptop, control panel, cellular phone, etc.
- the device associated with motion sensing system 100 can remain in its lowest power mode until a gesture occurs.
- a datagram 300 that can be utilized by a serial chain in a motion detection system.
- one or more optical sensors can transmit datagrams (e.g., packets) formatted in accordance with datagram 300 through serial chain 200 .
- a datagram is a basic transfer unit in which the delivery arrival time and order of packets of data (e.g., from different sources) are not guaranteed.
- Datagram 300 can be formatted in a specific manner for transmission between sensors.
- datagram 300 can be formatted with four distinct fields: Sequence Value 302 , INT 304 , EN 306 and COMMAND 308 .
- the Sequence Value 302 field can be an 8-bit field that is incremented by one each time a sensor communicates with the next sensor in the serial chain.
- the INT 304 field can be a 2-bit field that indicates a status of an interrupt signal from previous sensors. Any optical sensor on the serial chain can send an interrupt that can propagate to the end of the serial chain to the controller, depending on register settings.
- the EN 306 field can be a 1-bit field that can indicate a status of a first sensor's ENABLE pin. The enable signal from the first device in the serial chain can, depending on register settings, propagate through the entire serial chain.
- the COMMAND 308 field can be a 2-bit field containing a command being transmitted.
- the value of the COMMAND 308 field can be, for example, 00, indicating that the Sequence Value should be used as an address offset or 01, indicating that the Sequence Value should be used as a time code for time correlation.
- Other vales for the COMMAND 308 field can exist.
- methodology 400 for automatic address assignment in a motion detection system can address one or more optical sensors in the motion detection system via a serial chain. Unlike conventional addressing methodologies, methodology 400 does not utilize dedicated pins. Accordingly, this methodology leads to less power consumption than conventional methods for assigning addresses to the optical sensors.
- the methodology 400 begins at element 402 where the sensors are powered up. At power up, INT pins on every sensor are driven low. A first sensor on the serial chain has its ENABLE pin driven high, while the other sensors have their ENABLE pins driven low. At element 404 , the sensors can determine which sensor is the first sensor on the serial chain. For example, each sensor can determine whether it is the first sensor on the serial chain by determining whether it is the sensor with its ENABLE pin driven high.
- the methodology moves to element 406 .
- the first sensor can send a datagram (e.g., a datagram in accordance with datagram 300 as described above) to the next sensor (e.g., a second sensor) on the serial chain.
- the datagram can be configured with a COMMAND field sent to 00, indicating that a Sequence Value should be used as an address offset, and the Sequence Value set to 1.
- the second sensor can receive the datagram from the first sensor and set its address (e.g., I 2 C address) with the given address offset from the Sequence Value.
- the second sensor can increment the sequence value (e.g. the Sequence Value can be set to 2, indicating that a sensor has communicated twice with the next sensor on the serial chain).
- the second sensor can send the datagram to the next sensor (third sensor on the serial chain).
- an exemplary methodology 500 for automatic emitter (e.g., LED) sharing between optical sensors of a motion control system For example, access to the shared emitter by the optical sensors can be facilitated through automatic sequencing of the access to the shared emitter.
- automatic emitter e.g., LED
- the methodology 500 begins at element 502 , where all sensors on a serial chain are configured to share an emitter.
- the sensors can be configured to share the emitter through a register map.
- a first sensor is able to drive the emitter.
- the first sensor can drive the emitter through an assigned pin (such as an IRDR pin, not illustrated).
- the first sensor can drive the emitter, for example, for 128 samples, each 62.5 microseconds long, for a total of 8 milliseconds.
- the first sensor After driving the emitter for a certain time (e.g., 8 milliseconds), at element 506 the first sensor stops driving the emitter. For example, the first sensor can turn off its assigned pin. At element 508 , the first sensor sends a datagram (e.g., a datagram 300 as defined in FIG. 3 ) to a next sensor. At element 510 , the next sensor can drive the emitter for a certain time (e.g., 8 milliseconds). This methodology can repeat until all sensors on the serial chain have an opportunity to drive the emitter.
- a certain time e.g. 8 milliseconds
- FIG. 6 illustrated is an exemplary methodology 600 for automatic time correlation between optical sensors in a motion control system. This time correlation can be done in firmware. Accordingly, this methodology leads to greater power savings when compared to conventional methodologies.
- Methodology 600 begins at element 602 , where sensors are configured to store a time and a magnitude of peaks in magnitude of an electrical signal.
- Each of the sensors can include a four-sample first-in-first-out (FIFO) that can be configured to operate in a “peak detect” mode, where the FIFO can store the time and magnitude of peaks in magnitude.
- FIFO first-in-first-out
- a peak in magnitude can be detected when a user's hand makes its closest approach to the optical sensor (e.g., when making a gesture).
- a sensor can detect a peak.
- the magnitude value for the peak and a time code can be pushed into the FIFO.
- the time code can be the last Sequence Value received over the serial chain. Accordingly, time can be sub-divided into 8 millisecond periods, each labeled with a unique 8-bit code.
- the unique 8-bit code can be agreed upon by all sensors within the serial chain.
- an interrupt can be sent to a microcontroller.
- the microcontroller can read the peaks (magnitude values and time code pairs) from each sensor and reconstruct the gesture associated with the peak in magnitude.
- FIG. 7 illustrates an exemplary functional block diagram for the architecture 700 of the subject disclosure.
- the systems e.g., 100-200
- the systems e.g., 100-200
- ALS ambient light sensor
- the architecture 700 includes one or more LEDs and associated driver circuitry, one or more photodiode sensors (arranged in a serial chain, for example the serial chain as illustrated in FIG. 2 ), an analog front end and signal processing (e.g., of signals from the photodiodes arranged in the serial chain, for example elements 106 , 108 , 110 as illustrated in FIG.
- analog front end and signal processing e.g., of signals from the photodiodes arranged in the serial chain, for example elements 106 , 108 , 110 as illustrated in FIG.
- the architecture 700 adaptively optimizes sensitivity and power for a given environment. Moreover, the architecture 700 derives significant performance improvements from its novel ALS structure, and its light emitting diode (LED) driver circuitry is much more efficient than the conventional resistive drive.
- LED light emitting diode
- the architecture 700 includes a Resonant Front End 702 , which includes a Trans-Impedance Resonator (TIR).
- TIR 702 is used in place of the Trans-Inductance Amplifier (TIA), which is conventionally used.
- TIA can also be employed in signal processing applications of signals received from the serial chained photodiodes (e.g., signals 106 , 108 , 110 as described above.).
- the TIR 702 plays the same role as a conventional TIA, the TIR 702 gives an order of magnitude improvement in achievable Signal-to-Noise-Ratio (SNR) due to its band-pass nature (e.g., TIR 702 includes an inductor and a capacitor), which allows for an increased range of sensing.
- the capacitor of the TIR can include the capacitance of the photodiode that is being resonated.
- the band-pass nature of the TIR 702 causes the architecture 700 to operate over a narrow band of frequencies, which allows for little noise compared to the wide band TIA.
- the ALS 710 uses a light to frequency converter based on a relaxation oscillator instead of the conventional TIA.
- a relaxation oscillator is an oscillator based upon the relaxation behavior of a physical system.
- An exemplary implementation for the relaxation oscillator of the subject disclosure can be done by connecting the inverting input of an Operational Amplifier (Op Amp) to a fixed bias voltage via a switch and also the photodiode, with the non-inverting input connected to ground. When the switch to the fixed bias voltage is opened, the photodiode will discharge towards ground. The rate of discharge will depend on the photodiode current, which is a measure of the incident ambient light.
- Operational Amp Operational Amplifier
- the Computer Programmable Logic Device When the photodiode is discharged to ground, the Computer Programmable Logic Device (CPLD) resets the oscillator by switching back in the bias voltage.
- the CPLD counts the number of cycles that the photodiode takes to discharge, and thus can estimate the ambient light intensity incident on the photodiode.
- the ALS 710 can be used for ambient light sensing applications and the TIR 702 can be used for proximity and motion sensing applications.
- the output of the Front end 702 is subjected to multiple stages of voltage gain 716 to maximize the SNR of the output signal.
- the voltage gain is adaptively set based on the magnitude of the signal received from the Front end 702 , which is potentially made up of both measureable interferers such as a backscatter and a crosstalk from the LED, and also the desired signal to be measured.
- the interferers are dynamically calibrated out of the measurement to improve the sensitivity.
- the LED drive circuitry 756 uses an inductive drive, which results in a significant efficiency improvement over the conventional resistive drive.
- the architecture 700 also includes a Quad Demodulator with low pass filters (LPFs) 720 , dual [I & Q] Analog to Digital Converters (ADCs) 726 , Digital to Analog Converters (DACs) 730 driven by the bias voltage provided by the Automatic Gain Control module, Oscillator DACs 744 for 1 and Q carriers, the Universal Serial Bus (USB) processor for Control Interface, and the Computer Programmable Logic Device (CPLD) that include several modules.
- the CPLD is merely exemplary. One or more different parts can be employed in lieu of a CPLD. For example, one or more Application Specific IC(s), a digital signal processor and memory.
- the I and Q relate to In-Phase and Quadrature demodulation components.
- QAM is both an analog and a digital modulation scheme.
- QAM is a modulation scheme in which two sinusoidal carriers, one exactly 90 degrees out of phase with respect to the other, are used to transmit data over a given physical channel. Since the orthogonal carriers occupy the same frequency band and differ by a 90 degree phase shift, each can be modulated independently, transmitted over the same frequency band, and separated by demodulation at the receiver.
- PAM pulse amplitude modulation
- a numerically controlled oscillator can be employed to design a dual-output oscillator that accurately generates the in-phase and quadrature carriers used by a QAM modulator and/or demodulator.
- a filter for example, a raised cosine finite impulse response (FIR) filter can be utilized to filter the data streams before modulation onto the quadrature carriers.
- FIR raised cosine finite impulse response
- the in-phase and quadrature demodulated components are created by multiplying the signal by both a carrier signal, and also a signal 90 degrees out of phase of that carrier, and low pass filtering the result ( 720 in FIG. 7 ).
- the resultant I and Q are a baseband representation of the received signal.
- the phase of the derivative of I and Q channels can be obtained, which is indicative of the distance of the target to be calculated. Further, the position of a moving object can be accurately identified based on the phase data.
- the resultant phase information can be used as a direct output of the system as a measure of distance/position, and/or can be used to reconstruct the static component of the signal and allow the calibration of a non-derivative time of flight (TOF) measurement.
- TOF time of flight
- the architecture 700 of the subject disclosure can be used in many applications including computers, automotive, industrial, television displays and others.
- the architecture 700 can be used to detect that a user has entered the room and automatically cause a laptop computer in hibernation mode to wake up and enter into the active mode so that the user can use it.
- the architecture 700 of the subject disclosure can be used to automatically and adaptively adjust the intensity of a liquid crystal display (LCD) based on the ambient lighting conditions.
- the architecture 700 can perform motion and proximity sensing at a range of up to 1-2 meters.
- the architecture 700 of the subject disclosure can perform its operations by using less than twenty milli-watts (mW) of power.
- mW milli-watts
- the entire architecture 700 can be implemented in a single integrated circuit chip (IC).
- all components of the architecture 700 can be implemented in the IC except for the two inductors for the TIR 702 and the LED driver circuitry 756 and the LED, which can be implemented outside the IC.
- all components of the architecture 700 can be implemented in the IC except for the TIR 702 inductor, the LED and the inductor and the resistor for the LED driver circuitry, which can be implemented outside the IC.
- various components of the architecture 700 can be located inside or outside the IC.
- the terms (including a reference to a “means”) used to describe such components are intended to correspond, unless otherwise indicated, to any component which performs the specified function of the described component (e.g., a functional equivalent), even though not structurally equivalent to the disclosed structure, which performs the function in the herein illustrated exemplary aspects of the claimed subject matter.
- the disclosure includes a system as well as a computer-readable medium having computer-executable instructions for performing the acts and/or events of the various methods of the claimed subject matter.
Abstract
A system and method for identifying a position of a moving object, utilizing a serial chain of sensors is provided. The serial chain reduces the power needed for the motion detection system, allowing implementation of the motion detection system on a low-power computing device, such as a microcontroller. The serial chain can provide automatic address assignment without dedicated pins. The serial chain can also provide automatic sequencing of access to a shared LED. The serial chain can also provide automatic time correlation of data from multiple sensors in the motion detection system.
Description
- This application claims priority to U.S. Provisional Patent Application Ser. No. 61/298,895, filed on Jan. 27, 2010, and entitled “ARCHITECTURE FOR A REFLECTION BASED LONG RANGE PROXIMITY AND MOTION DETECTOR HAVING AN INTEGRATED AMBIENT LIGHT SENSOR,” the entirety of which is incorporated by reference herein. Further, this application is related to co-pending U.S. patent application Ser. No. 12/979,726, filed on Dec. 28, 2010 (Attorney docket number SE-2773/INTEP105USA), entitled “DISTANCE SENSING BY IQ DOMAIN DIFFERENTIATION OF TIME OF FLIGHT (TOF) MEASUREMENTS,” co-pending U.S. patent application Ser. No. ______, filed on ______ (Attorney docket number SE-2874-AN/INTEP105USB), entitled “DIRECT CURRENT (DC) CORRECTION CIRCUIT FOR A TIME OF FLIGHT (TOF) PHOTODIODE FRONT END”, co-pending U.S. patent application Ser. No. ______, filed on ______ (Attorney docket number SE-2785-AN/INTEP105USC), entitled “PHOTODIODE FRONT END WITH IMPROVED POWER SUPPLY REJECTION RATIO (PSRR),” co-pending U.S. patent application Ser. No. ______, filed on ______ (Attorney docket number SE-2877-AN/INTEP105USD), entitled “AUTOMATIC CALIBRATION TECHNIQUE FOR TIME OF FLIGHT (TOF) TRANSCEIVERS,” and co-pending U.S. patent application Ser. No. ______, filed on ______ (Attorney docket number SE-2878-AN/INTEP105USF), entitled “GESTURE RECOGNITION WITH PRINCIPAL COMPONENT ANALYSIS.” The entireties of each of the foregoing applications are incorporated herein by reference.
-
FIG. 1 illustrates an exemplary system that detects motion utilizing a serial chain of optical sensors. -
FIG. 2 illustrates an exemplary topology for a serial chain utilized by a motion detection system. -
FIG. 3 illustrates an exemplary datagram sent by optical sensors in the serial chain. -
FIG. 4 illustrates an exemplary methodology for assigning addresses to optical sensors in a motion detection system. -
FIG. 5 illustrates an exemplary methodology for automatic emitter sharing between optical sensors in a motion detection system. -
FIG. 6 illustrates an exemplary methodology for automatic time correlation between optical sensors in a motion detection system. -
FIG. 7 illustrates an exemplary functional block diagram for the architecture of the subject disclosure. - A category of monolithic devices is emerging that allows electronic products to sense their environment. These include diverse devices, such as, accelerometers, monolithic gyroscopes, light sensors and imagers. In particular, light sensors are one of the simplest and cheapest, allowing their inclusion in multitudes of consumer products, for example, nightlights, cameras, cell phones, laptops etc. Typically, light sensors can be employed in a wide variety of applications related to proximity sensing, such as, but not limited to, detecting the presence and/or distance of a user to the product for the purpose of controlling power, displays, or other interface options.
- Infrared (IR) detectors utilize IR light to detect objects within the sense area of the IR sensor. Moreover, IR light is transmitted by an IR Light emitting diode (LED) emitter, which reflects off of objects in the surrounding area and the reflections are sensed by a detector. Moreover, the detector can be a diode, e.g., a PIN diode, and/or any other type of apparatus that converts IR light into an electric signal. The sensed signal is analyzed to determine whether an object is present in the sense area and/or to detect motion within the sense area. It can be appreciated that although the subject specification is described with respect to IR light, the systems and methods disclosed herein can utilize most any wavelength. As an example, the subject system and/or methodology can be employed for acoustical proximity detection and/or ultrasonic range finding applications.
- For example, the sensor can be one of multiple proximity sensors arranged on a device to detect motion. In a motion detection system, multiple sensors can be arranged to detect motion, for example, a gesture on a screen. Each of the multiple sensors that can sense the gesture, for example, at different times at different locations on the screen. Each of the multiple sensors can transmit a sensed signal corresponding to the gesture to be analyzed across a system bus. In conventional systems, addresses of the multiple sensors are assigned through dedicated pins. The systems and methods disclosed herein allow for automatic address assignment for multiple sensors without dedicated pins. Through a serial chain of sensors, access to an emitter shared by the sensors can be automatically sequenced and data from the sensors can be automatically time-correlated without involvement of a controller, conserving power because these capabilities can be done in firmware. In contrast, conventional systems waste power because they cannot perform these actions in firmware, and, instead, must invoke substantial processing power of corresponding devices.
- The subject matter is described with reference to the drawings, wherein like reference numerals are used to refer to like elements throughout. In the following description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the subject disclosure. It may be evident, however, that the subject matter may be practiced without these specific details. In other instances, well-known structures and devices are shown in block diagram form in order to facilitate describing the subject disclosure. Many modifications may be made to this configuration without departing from the scope or spirit of the claimed subject matter.
- Moreover, the word “exemplary” is used herein to mean serving as an example, instance, or illustration. Any aspect or design described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects or designs. Rather, use of the word “exemplary” is intended to present concepts in a concrete fashion. As used in this application, the term “or” is intended to mean an inclusive “or” rather than an exclusive “or”. That is, unless specified otherwise, or clear from context, “X employs A or B” is intended to mean any of the natural inclusive permutations. That is, if X employs A; X employs B; or X employs both A and B, then “X employs A or B” is satisfied under any of the foregoing instances. In addition, the articles “a” and “an” as used in this application and the appended claims should generally be construed to mean “one or more” unless specified otherwise or clear from context to be directed to a singular form. In addition, the word “coupled” is used herein to mean direct or indirect electrical or mechanical coupling. Further, the terms “sense area,” “vision field,” “optical field,” and similar terminology are utilized interchangeably in the subject application, unless context warrants particular distinction(s) among the terms. Moreover, the terms “sensor,” “detector,” and similar terminology are utilized interchangeably in the subject application, unless context warrants particular distinction(s) among the terms.
- Referring initially to
FIG. 1 , there illustrated is anexample system 100 that employsoptical sensors 102 to sense distance, motion and/or ambient light, according to an aspect of the subject specification. For example, theoptical sensors 102 can include one or more proximity sensors. In one aspect, theoptical sensors 102 can include one or more infrared sensors (e.g., photodiodes, such as tuned positive-intrinsic-negative (PIN) diodes). However, theoptical sensors 102 can include any sensors that can convert a light signal to an electric signal. It can be appreciated that theoptical sensors 102 are not limited to utilizing IR light, and, instead, can be any sensor, detector, or combination of sensors and detectors that can utilize light signals of most any wavelength. Although threeoptical sensors 102 are shown for simplicity insystem 100, a system can employ moreoptical sensors 102 or feweroptical sensors 102. It can be appreciated that theoptical sensors 102 can be an array ofoptical sensors 102. - In general,
system 100 can be employed in most any light sensing/optical proximity application. For example, a laptop computer can detect a gesture (e.g., a tap or a swipe) on a track pad utilizing optical sensors arbitrarily arranged on the track pad. In another example, a cellular phone or a personal digital assistant (PDA) can detect a gesture (e.g., a tap or a swipe) on a screen utilizing optical sensors arbitrarily arranged on the screen. - The
optical sensors 102 can be coupled to asignal processing circuit 104 to transmitsensed outputs sensed outputs signal processing circuit 104 can analyze the sensedoutputs object 112 is present in the sense area and/or to detect motion by theobject 112 within the sense area. Theobject 112 can be most any entity of interest, such as, but not limited to, a human entity, an automated component, a device, an item, an animal, etc. - According to one aspect,
signal processing circuit 104 can be embedded on a single integrated circuit (IC) chip (e.g., a microcontroller). However, thesignal processing circuit 104 need not be embedded on a single IC chip, and, instead, components of thesignal processing circuit 104 can be distributed among several IC chips. The sensed outputs 106, 108, 110 can be transmitted to thesignal processing circuit 104 via a bus 114 (e.g. a serial bus, such as an Inter Integrated Circuit (“I2C”) bus). Although an I2C bus is described herein, it should be understood that the I2C bus is only utilized as an example and the systems and methods described herein can be utilized with any serial bus. - According to an aspect, the
bus 114 can be an I2C bus. The I2C bus includes two wires: a clock line and a data line. The clock line can synchronize data transfers across the I2C bus. The clock line and the data line are connected to eachoptical sensor 102 on the I2C bus. The signal processing circuit 104 a master on the I2C bus to theoptical sensors 102, which are slaves. The mastersignal processing circuit 104 drives the clock line. The slaveoptical sensors 102 are unable to initiate transmission or data (e.g. sensedoutputs optical sensors 102 is controlled by the mastersignal processing circuit 104 and can only begin when initiated by the mastersignal processing circuit 104. - In order to communicate across the bus, each
optical sensor 102 requires a unique address on the bus (for example, the address can be an I2C address) for data transfer to take place. For example, the address can be a seven bit address. In some conventional systems, addresses are assigned by hardware on thesignal processing circuit 104. For example, the addresses can be assigned according to dedicated hardware (e.g., dedicated pins). -
System 100, in contrast, can automatically assign addresses (e.g., I2C addresses) to eachoptical sensor 102 without requiring dedicated hardware.System 100 does not require involvement by thesignal processing circuit 104 to assign addresses to theoptical sensors 102. Instead,system 100 configures thesensors 102 in accordance with a serial chain. For example, the serial chain can be a daisy chain. - The serial chain can be utilized to automatically assign unique addresses to each
optical sensor 102. This eliminates the need for thesignal processor 104 to be involved with assigning addresses. Eliminating the need for thesignal processor 104 to be involved with assigning addresses can conserve power, allowingsystem 100 to be implemented on a low power a microcontroller. - Referring now to
FIG. 2 , illustrated is a topology for aserial chain 200 employed by a motion detection system. For example, theserial chain 200 can be implemented withinsystem 100 as described above. Although only threeoptical sensors serial chain 200 can be constructed between many moreoptical sensors 102. For example,serial chain 200 can be constructed between more than 255optical sensors 102. The number ofoptical sensors 102 supported byserial chain 200 can be limited only by capabilities of the serial bus.Serial chain 200 can be utilized to automatically assign addresses (e.g., I2C addresses) to eachoptical sensor serial chain 200. -
Optical sensors 102 can include adigital input digital output serial chain 200 between theoptical sensors digital input digital output Serial chain 200 can be utilized to relay ENABLE input from a first optical sensor to one or more other optical sensors, thereby preserving ENABLE functionality. Additionally,serial chain 200 can relay INT output from one or more optical sensors to a last optical sensor on theserial chain 200, thereby preserving INT functionality. It is noted thatsensors -
Serial chain 200 is a unidirectional interface between the output of a first optical sensor and the input of a second optical sensor.Serial chain 200 can transmit one or more datagrams of specific format from the first optical sensor to the second optical sensor. For example,output 204 c fromoptical sensor 102 a can be connected to input 202 b ofsensor 102 b.Output 204 b fromoptical sensor 102 b can be connected to input 202 c ofsensor 102 c.Serial chain 200 can automatically sequence access to a single, shared emitter (not shown), like a light emitting diode (LED), without involvement from a controller 206 (e.g., a microcontroller). -
Serial chain 200 can be asynchronous and/or self-clocking, allowingserial chain 200 to automatically time-correlate data from theoptical sensors sensors serial chain 200 is Manchester coding. Manchester coding, also known as Phase Encoding, can encode data from one or more ofsensors - For example, each of the
optical sensors - When a peak is detected, the magnitude value for the peak and a time code can be pushed into the FIFO. The time code can be the last Sequence Value received over the
serial chain 200. Accordingly, time can be sub-divided into 8 millisecond periods, each labeled with a unique 8-bit code. The unique 8-bit code can be agreed upon by allsensors serial chain 200. An interrupt can be sent to a microcontroller indicating the peaks. The microcontroller can read the peaks (magnitude values and time code pairs) from each sensor and reconstruct the gesture associated with the peak in magnitude at times corresponding to the peaks. - Since the serial chain is unidirectional, the first sensor is responsible for sending datagrams with appropriate time codes down the serial chain. The first sensor can include a register that stores the number of sensors in the serial chain for this purpose.
- All of the behaviors (addressing, sequencing access to a shared emitter and time correlation) are automatic and do not require any interaction with the microcontroller. In this way, the device (e.g., laptop, control panel, cellular phone, etc.) associated with
motion sensing system 100 can remain in its lowest power mode until a gesture occurs. - Referring now to
FIG. 3 , illustrated is adatagram 300 that can be utilized by a serial chain in a motion detection system. According to an aspect, one or more optical sensors can transmit datagrams (e.g., packets) formatted in accordance withdatagram 300 throughserial chain 200. A datagram is a basic transfer unit in which the delivery arrival time and order of packets of data (e.g., from different sources) are not guaranteed. -
Datagram 300 can be formatted in a specific manner for transmission between sensors. For example,datagram 300 can be formatted with four distinct fields:Sequence Value 302,INT 304,EN 306 andCOMMAND 308. TheSequence Value 302 field can be an 8-bit field that is incremented by one each time a sensor communicates with the next sensor in the serial chain. TheINT 304 field can be a 2-bit field that indicates a status of an interrupt signal from previous sensors. Any optical sensor on the serial chain can send an interrupt that can propagate to the end of the serial chain to the controller, depending on register settings. TheEN 306 field can be a 1-bit field that can indicate a status of a first sensor's ENABLE pin. The enable signal from the first device in the serial chain can, depending on register settings, propagate through the entire serial chain. - The
COMMAND 308 field can be a 2-bit field containing a command being transmitted. For example, the value of theCOMMAND 308 field can be, for example, 00, indicating that the Sequence Value should be used as an address offset or 01, indicating that the Sequence Value should be used as a time code for time correlation. Other vales for theCOMMAND 308 field can exist. - Referring now to
FIG. 4 , illustrated is anexemplary methodology 400 for automatic address assignment in a motion detection system. The methodology can address one or more optical sensors in the motion detection system via a serial chain. Unlike conventional addressing methodologies,methodology 400 does not utilize dedicated pins. Accordingly, this methodology leads to less power consumption than conventional methods for assigning addresses to the optical sensors. - The
methodology 400 begins atelement 402 where the sensors are powered up. At power up, INT pins on every sensor are driven low. A first sensor on the serial chain has its ENABLE pin driven high, while the other sensors have their ENABLE pins driven low. Atelement 404, the sensors can determine which sensor is the first sensor on the serial chain. For example, each sensor can determine whether it is the first sensor on the serial chain by determining whether it is the sensor with its ENABLE pin driven high. - After the power-on process (or power-on-reset process) of 402 and 404 completes, the methodology moves to
element 406. Atelement 406, the first sensor can send a datagram (e.g., a datagram in accordance withdatagram 300 as described above) to the next sensor (e.g., a second sensor) on the serial chain. For example, the datagram can be configured with a COMMAND field sent to 00, indicating that a Sequence Value should be used as an address offset, and the Sequence Value set to 1. - At
element 408, the second sensor can receive the datagram from the first sensor and set its address (e.g., I2C address) with the given address offset from the Sequence Value. Atelement 410, the second sensor can increment the sequence value (e.g. the Sequence Value can be set to 2, indicating that a sensor has communicated twice with the next sensor on the serial chain). Atelement 412, the second sensor can send the datagram to the next sensor (third sensor on the serial chain). - Referring now to
FIG. 5 , illustrated is anexemplary methodology 500 for automatic emitter (e.g., LED) sharing between optical sensors of a motion control system. For example, access to the shared emitter by the optical sensors can be facilitated through automatic sequencing of the access to the shared emitter. - The
methodology 500 begins atelement 502, where all sensors on a serial chain are configured to share an emitter. For example, the sensors can be configured to share the emitter through a register map. Atelement 504, a first sensor is able to drive the emitter. For example, the first sensor can drive the emitter through an assigned pin (such as an IRDR pin, not illustrated). The first sensor can drive the emitter, for example, for 128 samples, each 62.5 microseconds long, for a total of 8 milliseconds. - After driving the emitter for a certain time (e.g., 8 milliseconds), at
element 506 the first sensor stops driving the emitter. For example, the first sensor can turn off its assigned pin. Atelement 508, the first sensor sends a datagram (e.g., adatagram 300 as defined inFIG. 3 ) to a next sensor. Atelement 510, the next sensor can drive the emitter for a certain time (e.g., 8 milliseconds). This methodology can repeat until all sensors on the serial chain have an opportunity to drive the emitter. - Referring now to
FIG. 6 , illustrated is anexemplary methodology 600 for automatic time correlation between optical sensors in a motion control system. This time correlation can be done in firmware. Accordingly, this methodology leads to greater power savings when compared to conventional methodologies. -
Methodology 600 begins atelement 602, where sensors are configured to store a time and a magnitude of peaks in magnitude of an electrical signal. Each of the sensors can include a four-sample first-in-first-out (FIFO) that can be configured to operate in a “peak detect” mode, where the FIFO can store the time and magnitude of peaks in magnitude. For example, a peak in magnitude can be detected when a user's hand makes its closest approach to the optical sensor (e.g., when making a gesture). - At
element 604, a sensor can detect a peak. When a peak is detected, the magnitude value for the peak and a time code can be pushed into the FIFO. The time code can be the last Sequence Value received over the serial chain. Accordingly, time can be sub-divided into 8 millisecond periods, each labeled with a unique 8-bit code. The unique 8-bit code can be agreed upon by all sensors within the serial chain. - At
element 606, an interrupt can be sent to a microcontroller. The microcontroller can read the peaks (magnitude values and time code pairs) from each sensor and reconstruct the gesture associated with the peak in magnitude. - In order to provide additional context for various aspects of the subject specification,
FIG. 7 illustrates an exemplary functional block diagram for thearchitecture 700 of the subject disclosure. In one aspect, the systems (e.g., 100-200) disclosed herein can be employed in a reflection based proximity and motion detector with an integrated ambient light sensor (ALS) depicted inFIG. 7 . Thearchitecture 700 includes one or more LEDs and associated driver circuitry, one or more photodiode sensors (arranged in a serial chain, for example the serial chain as illustrated inFIG. 2 ), an analog front end and signal processing (e.g., of signals from the photodiodes arranged in the serial chain, forexample elements FIG. 1 above.), data conversion circuitry, digital control and signal processing, interface circuitry and results display. Thearchitecture 700 adaptively optimizes sensitivity and power for a given environment. Moreover, thearchitecture 700 derives significant performance improvements from its novel ALS structure, and its light emitting diode (LED) driver circuitry is much more efficient than the conventional resistive drive. - According to an aspect of the subject disclosure, the
architecture 700 includes aResonant Front End 702, which includes a Trans-Impedance Resonator (TIR). In thearchitecture 700, theTIR 702 is used in place of the Trans-Inductance Amplifier (TIA), which is conventionally used. However, TIA can also be employed in signal processing applications of signals received from the serial chained photodiodes (e.g., signals 106, 108, 110 as described above.). - Although the
TIR 702 plays the same role as a conventional TIA, theTIR 702 gives an order of magnitude improvement in achievable Signal-to-Noise-Ratio (SNR) due to its band-pass nature (e.g.,TIR 702 includes an inductor and a capacitor), which allows for an increased range of sensing. The capacitor of the TIR can include the capacitance of the photodiode that is being resonated. The band-pass nature of theTIR 702 causes thearchitecture 700 to operate over a narrow band of frequencies, which allows for little noise compared to the wide band TIA. - According to another aspect of the subject disclosure, the
ALS 710 uses a light to frequency converter based on a relaxation oscillator instead of the conventional TIA. A relaxation oscillator is an oscillator based upon the relaxation behavior of a physical system. An exemplary implementation for the relaxation oscillator of the subject disclosure can be done by connecting the inverting input of an Operational Amplifier (Op Amp) to a fixed bias voltage via a switch and also the photodiode, with the non-inverting input connected to ground. When the switch to the fixed bias voltage is opened, the photodiode will discharge towards ground. The rate of discharge will depend on the photodiode current, which is a measure of the incident ambient light. When the photodiode is discharged to ground, the Computer Programmable Logic Device (CPLD) resets the oscillator by switching back in the bias voltage. The CPLD counts the number of cycles that the photodiode takes to discharge, and thus can estimate the ambient light intensity incident on the photodiode. TheALS 710 can be used for ambient light sensing applications and theTIR 702 can be used for proximity and motion sensing applications. - The output of the
Front end 702 is subjected to multiple stages ofvoltage gain 716 to maximize the SNR of the output signal. The voltage gain is adaptively set based on the magnitude of the signal received from theFront end 702, which is potentially made up of both measureable interferers such as a backscatter and a crosstalk from the LED, and also the desired signal to be measured. The interferers are dynamically calibrated out of the measurement to improve the sensitivity. According to another aspect of the subject disclosure, theLED drive circuitry 756 uses an inductive drive, which results in a significant efficiency improvement over the conventional resistive drive. - The
architecture 700 also includes a Quad Demodulator with low pass filters (LPFs) 720, dual [I & Q] Analog to Digital Converters (ADCs) 726, Digital to Analog Converters (DACs) 730 driven by the bias voltage provided by the Automatic Gain Control module,Oscillator DACs 744 for 1 and Q carriers, the Universal Serial Bus (USB) processor for Control Interface, and the Computer Programmable Logic Device (CPLD) that include several modules. The CPLD is merely exemplary. One or more different parts can be employed in lieu of a CPLD. For example, one or more Application Specific IC(s), a digital signal processor and memory. The I and Q relate to In-Phase and Quadrature demodulation components. - QAM is both an analog and a digital modulation scheme. Moreover, QAM is a modulation scheme in which two sinusoidal carriers, one exactly 90 degrees out of phase with respect to the other, are used to transmit data over a given physical channel. Since the orthogonal carriers occupy the same frequency band and differ by a 90 degree phase shift, each can be modulated independently, transmitted over the same frequency band, and separated by demodulation at the receiver. Thus, QAM enables data transmission at twice the rate of standard pulse amplitude modulation (PAM) without any degradation in the bit error rate (BER). In one example a numerically controlled oscillator (NCO) can be employed to design a dual-output oscillator that accurately generates the in-phase and quadrature carriers used by a QAM modulator and/or demodulator. A filter, for example, a raised cosine finite impulse response (FIR) filter can be utilized to filter the data streams before modulation onto the quadrature carriers.
- The in-phase and quadrature demodulated components are created by multiplying the signal by both a carrier signal, and also a signal 90 degrees out of phase of that carrier, and low pass filtering the result (720 in
FIG. 7 ). The resultant I and Q are a baseband representation of the received signal. In one example, the phase of the derivative of I and Q channels can be obtained, which is indicative of the distance of the target to be calculated. Further, the position of a moving object can be accurately identified based on the phase data. Typically, the resultant phase information can be used as a direct output of the system as a measure of distance/position, and/or can be used to reconstruct the static component of the signal and allow the calibration of a non-derivative time of flight (TOF) measurement. - The
architecture 700 of the subject disclosure can be used in many applications including computers, automotive, industrial, television displays and others. For example, thearchitecture 700 can be used to detect that a user has entered the room and automatically cause a laptop computer in hibernation mode to wake up and enter into the active mode so that the user can use it. In another example, thearchitecture 700 of the subject disclosure can be used to automatically and adaptively adjust the intensity of a liquid crystal display (LCD) based on the ambient lighting conditions. According to an aspect of the subject disclosure, thearchitecture 700 can perform motion and proximity sensing at a range of up to 1-2 meters. According to another aspect of the subject disclosure, thearchitecture 700 of the subject disclosure can perform its operations by using less than twenty milli-watts (mW) of power. - In one embodiment of the subject disclosure, the
entire architecture 700 can be implemented in a single integrated circuit chip (IC). In another embodiment of the subject disclosure, all components of thearchitecture 700 can be implemented in the IC except for the two inductors for theTIR 702 and theLED driver circuitry 756 and the LED, which can be implemented outside the IC. In yet another embodiment of the subject disclosure, all components of thearchitecture 700 can be implemented in the IC except for theTIR 702 inductor, the LED and the inductor and the resistor for the LED driver circuitry, which can be implemented outside the IC. In still another embodiment of the subject disclosure, various components of thearchitecture 700 can be located inside or outside the IC. - What has been described above includes examples of the subject disclosure. It is, of course, not possible to describe every conceivable combination of components or methodologies for purposes of describing the claimed subject matter, but many further combinations and permutations of the subject disclosure are possible. Accordingly, the claimed subject matter is intended to embrace all such alterations, modifications, and variations that fall within the spirit and scope of the appended claims.
- In particular and in regard to the various functions performed by the above described components, devices, circuits, systems and the like, the terms (including a reference to a “means”) used to describe such components are intended to correspond, unless otherwise indicated, to any component which performs the specified function of the described component (e.g., a functional equivalent), even though not structurally equivalent to the disclosed structure, which performs the function in the herein illustrated exemplary aspects of the claimed subject matter. In this regard, it will also be recognized that the disclosure includes a system as well as a computer-readable medium having computer-executable instructions for performing the acts and/or events of the various methods of the claimed subject matter.
- The aforementioned systems/circuits/modules have been described with respect to interaction between several components. It can be appreciated that such systems/circuits/modules and components can include those components or specified sub-components, some of the specified components or sub-components, and/or additional components, and according to various permutations and combinations of the foregoing. Sub-components can also be implemented as components communicatively coupled to other components rather than included within parent components (hierarchical). Additionally, it should be noted that one or more components may be combined into a single component providing aggregate functionality or divided into several separate sub-components, and any one or more middle layers, such as a management layer, may be provided to communicatively couple to such sub-components in order to provide integrated functionality. Any components described herein may also interact with one or more other components not specifically described herein but generally known by those of skill in the art.
- In addition, while a particular feature of the subject disclosure may have been disclosed with respect to only one of several implementations, such feature may be combined with one or more other features of the other implementations as may be desired and advantageous for any given or particular application. Furthermore, to the extent that the terms “includes,” “including,” “has,” “contains,” variants thereof, and other similar words are used in either the detailed description or the claims, these terms are intended to be inclusive in a manner similar to the term “comprising” as an open transition word without precluding any additional or other elements.
Claims (20)
1. An apparatus, comprising:
at least one detector that generates an electrical signal in response to detecting at least one motion; and
a motion detector circuit that determines that the electrical signal corresponds to the at least one motion; wherein,
the at least one detector sends the electrical signal to the motion detector circuit across a bus; and
the at least one detector receives an address on the bus according to a serial chain.
2. The apparatus of claim 1 , wherein the at least one detector is at least one light sensor.
3. The apparatus of claim 3 , wherein the at least one light sensor comprises at least one infrared (IR) sensor.
4. The apparatus of claim 1 , wherein the electrical signal comprises a datagram and the address is assigned for the at least one detector based at least in part on the datagram
5. The apparatus of claim 4 , wherein the datagram comprises a first field that indicates that an address should be incremented by a number in a second field.
6. The apparatus of claim 5 , wherein the first data field is a COMMAND field set to 00.
7. The apparatus of claim 5 , wherein the second data field is an 8-bit field.
8. The apparatus sensor of claim 7 , wherein the second data field is initially set to 1.
9. The apparatus of claim 8 , wherein the at least one detector increments the second data field by one before sending the datagram to a next detector.
10. The apparatus of claim 1 , wherein the at least one detector comprises an input pin that is driven high and an output pin that is driven low on power up.
11. The apparatus of claim 7 , further comprising a next detector coupled to the first detector, wherein the next detector comprises an input pin that is driven low and an output pin that is driven low on power up.
12. A method, comprising:
configuring at least a first sensor and a next sensor on a serial chain to share an emitter;
driving the emitter by the first sensor for a predetermined time period;
sending a datagram from the first sensor to the next sensor; and
driving the emitter by the second sensor for the predetermined time period.
13. The method of claim 12 , wherein the pre-determined time period is 8 milliseconds.
14. The method of claim 12 , wherein the configuring further comprises configuring the at least the first sensor and the second sensor on the serial chain to share the emitter through a register map.
15. The method of claim 12 , wherein the driving the emitter by the first sensor further comprises driving the emitter for 128 samples, each 62.5 microseconds long.
16. The method of claim 12 , wherein the driving the emitter by using the first sensor further comprises driving the emitter by using the first sensor through a driving pin of the first sensor.
17. A system, comprising:
at least one light emitting diode (LED) that emits a signal, wherein at least a portion of the signal reflects back from a moving object;
one or more sensors for generating an electric signal based on a peak in the reflected portion of the frequency modulated signal, wherein the one or more sensors are arranged in a serial chain and transmit the electric signal across the serial chain; and
a signal processing circuit that reconstructs the at least one gesture based at least in part on the electric signal.
18. The system of claim 17 , wherein the serial chain facilitates automatic time correlation between the one or more sensors.
19. The system of claim 17 , wherein the one or more sensors comprise one or more four-sample first-in-first-outs (FIFOs) that store a time code of the peak and a magnitude of the peak.
20. The system of claim 19 , wherein the signal processing circuit reads the peak from the one or more sensors and reconstructs the at least one gesture based at least in part on the magnitude of the peak and the time code of the peak.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/013,640 US20110180709A1 (en) | 2010-01-27 | 2011-01-25 | Serial-chaining proximity sensors for gesture recognition |
TW100103045A TW201203008A (en) | 2010-01-27 | 2011-01-27 | Serial-chaining proximity sensors for gesture recognition |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US29889510P | 2010-01-27 | 2010-01-27 | |
US13/013,640 US20110180709A1 (en) | 2010-01-27 | 2011-01-25 | Serial-chaining proximity sensors for gesture recognition |
Publications (1)
Publication Number | Publication Date |
---|---|
US20110180709A1 true US20110180709A1 (en) | 2011-07-28 |
Family
ID=44308255
Family Applications (7)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/979,726 Active 2032-09-24 US8760631B2 (en) | 2010-01-27 | 2010-12-28 | Distance sensing by IQ domain differentiation of time of flight (TOF) measurements |
US13/013,173 Active 2031-06-12 US8274037B2 (en) | 2010-01-27 | 2011-01-25 | Automatic calibration technique for time of flight (TOF) transceivers |
US13/013,640 Abandoned US20110180709A1 (en) | 2010-01-27 | 2011-01-25 | Serial-chaining proximity sensors for gesture recognition |
US13/013,199 Abandoned US20110180693A1 (en) | 2010-01-27 | 2011-01-25 | Photodiode front end with improved power supply rejection ratio (psrr) |
US13/013,146 Active 2032-01-13 US8530819B2 (en) | 2010-01-27 | 2011-01-25 | Direct current (DC) correction circuit for a time of flight (TOF) photodiode front end |
US13/013,676 Abandoned US20110182519A1 (en) | 2010-01-27 | 2011-01-25 | Gesture recognition with principal component anaysis |
US14/296,277 Active 2033-06-28 US10031078B2 (en) | 2010-01-27 | 2014-06-04 | Distance sensing by IQ domain differentiation of time of flight (TOF) measurements |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/979,726 Active 2032-09-24 US8760631B2 (en) | 2010-01-27 | 2010-12-28 | Distance sensing by IQ domain differentiation of time of flight (TOF) measurements |
US13/013,173 Active 2031-06-12 US8274037B2 (en) | 2010-01-27 | 2011-01-25 | Automatic calibration technique for time of flight (TOF) transceivers |
Family Applications After (4)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/013,199 Abandoned US20110180693A1 (en) | 2010-01-27 | 2011-01-25 | Photodiode front end with improved power supply rejection ratio (psrr) |
US13/013,146 Active 2032-01-13 US8530819B2 (en) | 2010-01-27 | 2011-01-25 | Direct current (DC) correction circuit for a time of flight (TOF) photodiode front end |
US13/013,676 Abandoned US20110182519A1 (en) | 2010-01-27 | 2011-01-25 | Gesture recognition with principal component anaysis |
US14/296,277 Active 2033-06-28 US10031078B2 (en) | 2010-01-27 | 2014-06-04 | Distance sensing by IQ domain differentiation of time of flight (TOF) measurements |
Country Status (4)
Country | Link |
---|---|
US (7) | US8760631B2 (en) |
CN (3) | CN102822759B (en) |
TW (6) | TWI439716B (en) |
WO (6) | WO2011094365A1 (en) |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110181892A1 (en) * | 2010-01-27 | 2011-07-28 | Intersil Americas Inc. | Automatic calibration technique for time of flight (tof) transceivers |
US20110248172A1 (en) * | 2010-04-13 | 2011-10-13 | Rueger Timothy T | Apparatus with optical functionality and power management and associated methods |
US20120280905A1 (en) * | 2011-05-05 | 2012-11-08 | Net Power And Light, Inc. | Identifying gestures using multiple sensors |
US20130241888A1 (en) * | 2012-03-14 | 2013-09-19 | Texas Instruments Incorporated | Detecting Wave Gestures Near an Illuminated Surface |
US20130300644A1 (en) * | 2012-05-11 | 2013-11-14 | Comcast Cable Communications, Llc | System and Methods for Controlling a User Experience |
US20140033141A1 (en) * | 2011-04-13 | 2014-01-30 | Nokia Corporation | Method, apparatus and computer program for user control of a state of an apparatus |
WO2014085269A1 (en) * | 2012-11-29 | 2014-06-05 | Amazon Technologies, Inc. | Gesture detection management for an electronic device |
US9194741B2 (en) | 2013-09-06 | 2015-11-24 | Blackberry Limited | Device having light intensity measurement in presence of shadows |
US9256290B2 (en) | 2013-07-01 | 2016-02-09 | Blackberry Limited | Gesture detection using ambient light sensors |
US9298333B2 (en) | 2011-12-22 | 2016-03-29 | Smsc Holdings S.A.R.L. | Gesturing architecture using proximity sensing |
US9304596B2 (en) | 2013-07-24 | 2016-04-05 | Blackberry Limited | Backlight for touchless gesture detection |
US9323336B2 (en) | 2013-07-01 | 2016-04-26 | Blackberry Limited | Gesture detection using ambient light sensors |
US9342671B2 (en) | 2013-07-01 | 2016-05-17 | Blackberry Limited | Password by touch-less gesture |
US9367137B2 (en) | 2013-07-01 | 2016-06-14 | Blackberry Limited | Alarm operation by touch-less gesture |
US9398221B2 (en) | 2013-07-01 | 2016-07-19 | Blackberry Limited | Camera control using ambient light sensors |
US9405461B2 (en) | 2013-07-09 | 2016-08-02 | Blackberry Limited | Operating a device using touchless and touchscreen gestures |
US9423913B2 (en) | 2013-07-01 | 2016-08-23 | Blackberry Limited | Performance control of ambient light sensors |
US9465448B2 (en) | 2013-07-24 | 2016-10-11 | Blackberry Limited | Backlight for touchless gesture detection |
US9489051B2 (en) | 2013-07-01 | 2016-11-08 | Blackberry Limited | Display navigation using touch-less gestures |
US11076789B2 (en) * | 2016-06-22 | 2021-08-03 | Linet Spol. S.R.O | Medical data collection system and method of use thereof |
US11284808B2 (en) | 2014-10-11 | 2022-03-29 | Linet Spol. S.R.O. | Device and method for measurement of vital functions, including intracranial pressure, and system and method for collecting data |
US11599199B2 (en) | 2019-11-28 | 2023-03-07 | Boe Technology Group Co., Ltd. | Gesture recognition apparatus, gesture recognition method, computer device and storage medium |
US20230089705A1 (en) * | 2021-09-17 | 2023-03-23 | Melexis Bulgaria EOOD | Sensing system comprising a chain of sensors |
Families Citing this family (103)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
USRE46672E1 (en) | 2006-07-13 | 2018-01-16 | Velodyne Lidar, Inc. | High definition LiDAR system |
EP2418512A1 (en) * | 2010-07-30 | 2012-02-15 | Mechaless Systems GmbH | Optoelectronic measuring assembly with compensation for external light sources |
CN102564576B (en) * | 2010-12-17 | 2013-11-06 | 鸿富锦精密工业(深圳)有限公司 | Light intensity testing device |
KR20120072134A (en) * | 2010-12-23 | 2012-07-03 | 한국전자통신연구원 | Apparatus and method for accelerating virtual desktop |
JP5566934B2 (en) * | 2011-03-23 | 2014-08-06 | 株式会社東芝 | Voltage output circuit and active cable |
US8604436B1 (en) * | 2011-03-24 | 2013-12-10 | Maxim Integrated Products, Inc. | Proximity sensor device |
US8988660B2 (en) * | 2011-06-29 | 2015-03-24 | Silicon Laboratories Inc. | Optical detector |
US8994926B2 (en) * | 2012-02-14 | 2015-03-31 | Intersil Americas LLC | Optical proximity sensors using echo cancellation techniques to detect one or more objects |
TWI470485B (en) * | 2012-03-29 | 2015-01-21 | Wistron Corp | Stylus capable of detecting pressure on a tip |
US20130257583A1 (en) * | 2012-03-29 | 2013-10-03 | Motorola Mobility, Inc. | Wireless communication device and method with ultrasonic detection |
TWI484379B (en) * | 2012-06-01 | 2015-05-11 | Pixart Imaging Inc | Optical detecting device |
TWI465753B (en) * | 2012-08-15 | 2014-12-21 | Generalplus Technology Inc | Position identification system and method and system and method for gesture identification thereof |
IN2015DN02057A (en) * | 2012-09-13 | 2015-08-14 | Mbda Uk Ltd | |
KR101338459B1 (en) * | 2012-10-19 | 2013-12-10 | 고려대학교 산학협력단 | Sensor unit |
US20140346361A1 (en) * | 2013-05-23 | 2014-11-27 | Yibing M. WANG | Time-of-flight pixels also sensing proximity and/or detecting motion in imaging devices & methods |
KR102102702B1 (en) * | 2013-06-19 | 2020-04-21 | 삼성전자주식회사 | Unit pixel of image sensor and image sensor having the same |
KR20150010230A (en) * | 2013-07-18 | 2015-01-28 | 삼성전자주식회사 | Method and apparatus for generating color image and depth image of an object using singular filter |
JP6207321B2 (en) * | 2013-09-26 | 2017-10-04 | ローム株式会社 | Optical sensor device |
US9250714B2 (en) | 2013-11-27 | 2016-02-02 | Intersil Americas LLC | Optical proximity detectors |
KR102159996B1 (en) * | 2013-12-16 | 2020-09-25 | 삼성전자주식회사 | Event filtering device and motion recognition device thereof |
JP6146295B2 (en) * | 2013-12-26 | 2017-06-14 | 株式会社豊田中央研究所 | Radar apparatus and speed direction measuring method |
US9681123B2 (en) | 2014-04-04 | 2017-06-13 | Microsoft Technology Licensing, Llc | Time-of-flight phase-offset calibration |
US9523765B2 (en) | 2014-07-14 | 2016-12-20 | Omnivision Technologies, Inc. | Pixel-level oversampling for a time of flight 3D image sensor with dual range measurements |
WO2016032048A1 (en) * | 2014-08-28 | 2016-03-03 | 엘지전자 주식회사 | Proximity sensor and method for controlling same |
US9977512B2 (en) * | 2014-10-24 | 2018-05-22 | Intersil Americas LLC | Open loop correction for optical proximity detectors |
CN104346043A (en) * | 2014-11-03 | 2015-02-11 | 武汉麦塔威科技有限公司 | Bar code scanning positioning based interaction display method of mobile screen and device of interaction display method |
US10795005B2 (en) * | 2014-12-09 | 2020-10-06 | Intersil Americas LLC | Precision estimation for optical proximity detectors |
JP2016150130A (en) * | 2015-02-18 | 2016-08-22 | セイコーエプソン株式会社 | Information acquisition device and information acquisition method |
US9983680B2 (en) * | 2015-03-23 | 2018-05-29 | Intel Corporation | Gesture recognition mechanism |
JP6386410B2 (en) * | 2015-03-31 | 2018-09-05 | 新明和工業株式会社 | Motion detection device |
US9667349B1 (en) * | 2015-04-15 | 2017-05-30 | The United States Of America As Represented By The Secretary Of The Navy | Dynamic range extension of heterodyne fiber-optic interferometers via instantaneous carrier measurement |
CN105034802B (en) * | 2015-05-26 | 2017-12-05 | 李洪金 | E-Gas based on infrared distance measurement |
US10422870B2 (en) | 2015-06-15 | 2019-09-24 | Humatics Corporation | High precision time of flight measurement system for industrial automation |
CN108024513A (en) * | 2015-06-15 | 2018-05-11 | 修麦提克斯公司 | High-precision time-of-flight measurement system for industrial automation |
US10591592B2 (en) | 2015-06-15 | 2020-03-17 | Humatics Corporation | High-precision time of flight measurement systems |
US10048357B2 (en) * | 2015-06-15 | 2018-08-14 | Microsoft Technology Licensing, Llc | Time-of-flight (TOF) system calibration |
CN107683109B (en) * | 2015-06-25 | 2021-06-08 | 费森尤斯医疗控股股份有限公司 | Direct light differential measurement system |
US9952676B2 (en) | 2015-06-25 | 2018-04-24 | Intel Corporation | Wearable device with gesture recognition mechanism |
TWI708070B (en) * | 2015-10-22 | 2020-10-21 | 新加坡商海特根微光學公司 | Optical crosstalk calibration for ranging systems |
US10324494B2 (en) | 2015-11-25 | 2019-06-18 | Intel Corporation | Apparatus for detecting electromagnetic field change in response to gesture |
JP2019508920A (en) | 2015-12-17 | 2019-03-28 | ヒューマティクス コーポレイション | Radio frequency localization technology and related systems, devices, and methods |
US10627490B2 (en) | 2016-01-31 | 2020-04-21 | Velodyne Lidar, Inc. | Multiple pulse, LIDAR based 3-D imaging |
CN109154661A (en) | 2016-03-19 | 2019-01-04 | 威力登激光雷达有限公司 | Integrated irradiation and detection for the 3-D imaging based on LIDAR |
US11401733B2 (en) | 2016-04-15 | 2022-08-02 | Spectrum Brands, Inc. | Wireless lockset with integrated angle of arrival (AOA) detection |
US10139341B2 (en) | 2016-05-31 | 2018-11-27 | Tt Electronics Plc | Self-calibrating optical detector |
CA3024510C (en) | 2016-06-01 | 2022-10-04 | Velodyne Lidar, Inc. | Multiple pixel scanning lidar |
US10298282B2 (en) | 2016-06-16 | 2019-05-21 | Intel Corporation | Multi-modal sensing wearable device for physiological context measurement |
US10598783B2 (en) | 2016-07-07 | 2020-03-24 | Microsoft Technology Licensing, Llc | Multi-frequency unwrapping |
CN106405524A (en) * | 2016-08-24 | 2017-02-15 | 上海兰宝传感科技股份有限公司 | Infrared distance measuring sensor |
CN106405565A (en) * | 2016-08-24 | 2017-02-15 | 上海兰宝传感科技股份有限公司 | Unmanned aerial vehicle barrier-avoiding/height-determining system |
US10620295B2 (en) * | 2016-09-14 | 2020-04-14 | Htc Corporation | Positioning signal receiver, positioning system and positioning method of the same |
CN106405567B (en) * | 2016-10-14 | 2018-03-02 | 海伯森技术(深圳)有限公司 | A kind of range-measurement system and its bearing calibration based on TOF |
US10291895B2 (en) | 2016-10-25 | 2019-05-14 | Omnivision Technologies, Inc. | Time of flight photosensor |
CN106546993B (en) * | 2016-11-04 | 2019-04-16 | 武汉万集信息技术有限公司 | A kind of range unit and distance measuring method improving pulse type laser range accuracy |
CN110914705A (en) | 2017-03-31 | 2020-03-24 | 威力登激光雷达有限公司 | Integrated LIDAR lighting power control |
CA3062701A1 (en) | 2017-05-08 | 2018-11-15 | Velodyne Lidar, Inc. | Lidar data acquisition and control |
TWI621868B (en) * | 2017-06-21 | 2018-04-21 | Univ Kun Shan | System and method for guiding brain waves to blind people |
CN109213385B (en) * | 2017-07-05 | 2021-07-20 | 光宝科技新加坡私人有限公司 | Mobile device and proximity sensing module thereof |
US10575384B2 (en) * | 2017-10-23 | 2020-02-25 | Infineon Technologies Ag | Adaptive transmit light control |
US10944486B2 (en) * | 2017-12-06 | 2021-03-09 | Elenion Technologies, Llc | DC current cancellation scheme for an optical receiver |
US11294041B2 (en) | 2017-12-08 | 2022-04-05 | Velodyne Lidar Usa, Inc. | Systems and methods for improving detection of a return signal in a light ranging and detection system |
US10393875B2 (en) | 2017-12-19 | 2019-08-27 | Nortek Security & Control Llc | Time of flight based sensor |
USD850946S1 (en) | 2018-02-01 | 2019-06-11 | Tyco Fire & Security Gmbh | Position detector |
US10718147B2 (en) | 2018-04-06 | 2020-07-21 | Tyco Fire & Security Gmbh | Optical displacement detector with adjustable pattern direction |
US20190317196A1 (en) * | 2018-04-17 | 2019-10-17 | Continental Automotive Systems, Inc. | Crosstalk mitigation circuit for lidar pixel receivers |
TWI660591B (en) | 2018-05-24 | 2019-05-21 | 立積電子股份有限公司 | Phase information extraction circuit and phase information extraction method for object movement |
CN109031192B (en) * | 2018-06-26 | 2020-11-06 | 北京永安信通科技有限公司 | Object positioning method, object positioning device and electronic equipment |
TWI697845B (en) * | 2018-07-13 | 2020-07-01 | 緯創資通股份有限公司 | Multi-object tracking method and system |
CN109035345A (en) * | 2018-07-20 | 2018-12-18 | 齐鲁工业大学 | The TOF camera range correction method returned based on Gaussian process |
CN109061609B (en) * | 2018-08-01 | 2021-09-07 | 歌尔光学科技有限公司 | TOF module calibration device and method |
US11073615B2 (en) * | 2018-08-20 | 2021-07-27 | Lite-On Singapore Pte. Ltd. | Proximity sensor module with two sensors |
US10712434B2 (en) | 2018-09-18 | 2020-07-14 | Velodyne Lidar, Inc. | Multi-channel LIDAR illumination driver |
TWI693421B (en) * | 2018-10-24 | 2020-05-11 | 精準基因生物科技股份有限公司 | Time-of-flight ranging device and time-of-flight ranging method |
US11082010B2 (en) | 2018-11-06 | 2021-08-03 | Velodyne Lidar Usa, Inc. | Systems and methods for TIA base current detection and compensation |
DE102019134142A1 (en) * | 2018-12-12 | 2020-06-18 | Analog Devices, Inc. | BUILT-IN CALIBRATION OF RUN TIME DEPTH IMAGING SYSTEMS |
US11423572B2 (en) * | 2018-12-12 | 2022-08-23 | Analog Devices, Inc. | Built-in calibration of time-of-flight depth imaging systems |
US11885874B2 (en) * | 2018-12-19 | 2024-01-30 | Semiconductor Components Industries, Llc | Acoustic distance measuring circuit and method for low frequency modulated (LFM) chirp signals |
US11885958B2 (en) | 2019-01-07 | 2024-01-30 | Velodyne Lidar Usa, Inc. | Systems and methods for a dual axis resonant scanning mirror |
CN111669278A (en) * | 2019-03-06 | 2020-09-15 | 马维尔亚洲私人有限公司 | Method and device for waking up physical layer of first node in low-power mode |
US11500100B2 (en) * | 2019-04-15 | 2022-11-15 | Microsoft Technology Licensing, Llc | Time-of-flight measurements using linear inverse function |
US10819920B1 (en) | 2019-05-22 | 2020-10-27 | Dell Products L.P. | Augmented information handling system user presence detection |
CN112135071B (en) * | 2019-06-25 | 2023-07-28 | 上海耕岩智能科技有限公司 | Image sensing system |
US11317040B2 (en) * | 2019-06-25 | 2022-04-26 | Shanghai Harvest Intelligence Technology Co., Ltd. | Image sensing apparatus |
US10613203B1 (en) | 2019-07-01 | 2020-04-07 | Velodyne Lidar, Inc. | Interference mitigation for light detection and ranging |
TWI739204B (en) * | 2019-07-22 | 2021-09-11 | 財團法人工業技術研究院 | System and method for signal sensing |
TWI717842B (en) * | 2019-09-18 | 2021-02-01 | 茂達電子股份有限公司 | Optical proximity sensor with digital correction circuit and digital correction method thereof |
US11294054B2 (en) | 2019-10-11 | 2022-04-05 | Dell Products L.P. | Information handling system infrared proximity detection with ambient light management |
US11435447B2 (en) | 2019-10-11 | 2022-09-06 | Dell Products L.P. | Information handling system proximity sensor with mechanically adjusted field of view |
US11435475B2 (en) | 2019-10-11 | 2022-09-06 | Dell Products L.P. | Information handling system infrared proximity detection with frequency domain modulation |
US11662695B2 (en) | 2019-10-11 | 2023-05-30 | Dell Products L.P. | Information handling system infrared proximity detection with distance reduction detection |
CN111095914B (en) | 2019-12-06 | 2022-04-29 | 深圳市汇顶科技股份有限公司 | Three-dimensional image sensing system, related electronic device and time-of-flight distance measurement method |
US11334146B2 (en) | 2020-01-31 | 2022-05-17 | Dell Products L.P. | Information handling system peripheral enhanced user presence detection |
US11513813B2 (en) | 2020-01-31 | 2022-11-29 | Dell Products L.P. | Information handling system notification presentation based upon user presence detection |
US11663343B2 (en) | 2020-01-31 | 2023-05-30 | Dell Products L.P. | Information handling system adaptive user presence detection |
WO2021188435A1 (en) * | 2020-03-17 | 2021-09-23 | Becton, Dickinson And Company | Gain matched amplifiers for light detection |
US11796715B2 (en) | 2020-06-24 | 2023-10-24 | Sloan Valve Company | Hybrid time-of-flight sensor and IR sensor |
TWI759213B (en) * | 2020-07-10 | 2022-03-21 | 大陸商廣州印芯半導體技術有限公司 | Light sensor and sensing method thereof |
US11280847B1 (en) * | 2020-10-30 | 2022-03-22 | Taiwan Semiconductor Manufacturing Company Ltd. | Circuit, semiconductor device and method for parameter PSRR measurement |
US11808895B2 (en) | 2020-12-11 | 2023-11-07 | Stmicroelectronics (Research & Development) Limited | Methods and devices for crosstalk compensation |
KR102354158B1 (en) | 2021-01-14 | 2022-01-21 | 박천수 | Multi Phase correlation Vector Synthesis Ranging Method and apparatus |
TWI761047B (en) * | 2021-01-22 | 2022-04-11 | 明泰科技股份有限公司 | Method for verifying detection range of image detector |
CN114489229B (en) * | 2021-12-24 | 2023-01-31 | 芯海科技(深圳)股份有限公司 | Drift voltage correction circuit, integrated circuit, measuring device, and electronic apparatus |
TWI805425B (en) * | 2022-06-29 | 2023-06-11 | 中國鋼鐵股份有限公司 | System and method for determining time of light of ultrasound wave |
Citations (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3996590A (en) * | 1961-02-02 | 1976-12-07 | Hammack Calvin M | Method and apparatus for automatically detecting and tracking moving objects and similar applications |
US4458212A (en) * | 1981-12-30 | 1984-07-03 | Mostek Corporation | Compensated amplifier having pole zero tracking |
US4542475A (en) * | 1982-11-23 | 1985-09-17 | Rca Corporation | Sampled data filter system as for a digital TV |
US4551710A (en) * | 1983-04-29 | 1985-11-05 | Cerberus Ag | Method and apparatus for reporting dangerous conditions |
US4648364A (en) * | 1985-07-08 | 1987-03-10 | Wills William H | Engine protection apparatus |
US4942561A (en) * | 1988-06-29 | 1990-07-17 | Kabushikikaisha Topcon | Delay time measuring device |
US5055671A (en) * | 1990-10-03 | 1991-10-08 | Spacelabs, Inc. | Apparatus for detecting transducer movement using a first and second light detector |
US5563701A (en) * | 1993-12-29 | 1996-10-08 | Hyundai Electronics Industries Co., Ltd. | Apparatus and method for measuring distance optically using phase variation |
US5593430A (en) * | 1995-01-27 | 1997-01-14 | Pacesetter, Inc. | Bus system for interconnecting an implantable medical device with a plurality of sensors |
US5828899A (en) * | 1996-01-04 | 1998-10-27 | Compaq Computer Corporation | System for peripheral devices recursively generating unique addresses based on the number of devices connected dependent upon the relative position to the port |
US5892540A (en) * | 1996-06-13 | 1999-04-06 | Rockwell International Corporation | Low noise amplifier for passive pixel CMOS imager |
US5990409A (en) * | 1997-12-26 | 1999-11-23 | Roland Kabushiki Kaisha | Musical apparatus detecting maximum values and/or peak values of reflected light beams to control musical functions |
US6111256A (en) * | 1997-04-10 | 2000-08-29 | Shmuel Hershkovitz & Pinhas Shpater | Infrared motion detection signal sampler |
US6392539B1 (en) * | 1998-07-13 | 2002-05-21 | Honda Giken Kogyo Kabushiki Kaisha | Object detection apparatus |
US20020097743A1 (en) * | 1993-03-09 | 2002-07-25 | Ertugrul Baydar | Integrated digital loop carrier system with virtual tributary mapper circuit |
US6462726B1 (en) * | 1998-02-19 | 2002-10-08 | Matsushita Electric Industrial Co., Ltd. | Video signal processor |
US20030234341A1 (en) * | 2002-06-20 | 2003-12-25 | Osborn Jon V. | Microelectromechanical system optical sun sensor |
US6744248B2 (en) * | 2001-12-08 | 2004-06-01 | Koninklijke Philips Electronics N.V. | Arrangement for detecting motion of an encoder |
US6753950B2 (en) * | 2000-01-26 | 2004-06-22 | Instro Precision Limited | Optical distance measurement |
US20040140961A1 (en) * | 2003-01-17 | 2004-07-22 | Eastman Kodak Company | Oled display and touch screen |
US6803555B1 (en) * | 2001-09-07 | 2004-10-12 | Indigo Systems Corporation | Two-stage auto-zero amplifier circuit for electro-optical arrays |
US6819782B1 (en) * | 1999-06-08 | 2004-11-16 | Matsushita Electric Industrial Co., Ltd. | Device and method for recognizing hand shape and position, and recording medium having program for carrying out the method recorded thereon |
US6836212B2 (en) * | 2002-10-10 | 2004-12-28 | Motorola, Inc. | Method and apparatus for reducing the likelihood of losing a portable electronic device |
US6888938B2 (en) * | 1999-05-11 | 2005-05-03 | Agere Systems Inc. | Dynamically adjustable digital gyrator having extendable feedback for stable DC load line |
US20060120621A1 (en) * | 2000-01-06 | 2006-06-08 | Canon Kabushiki Kaisha | Demodulation and phase estimation of two-dimensional patterns |
US20070013791A1 (en) * | 2005-07-05 | 2007-01-18 | Koichi Kinoshita | Tracking apparatus |
US7212655B2 (en) * | 2000-09-15 | 2007-05-01 | Tumey David M | Fingerprint verification system |
US20070121095A1 (en) * | 2005-11-28 | 2007-05-31 | Robert Lewis | Distance measurement device with short range optics |
US20080119716A1 (en) * | 2006-05-17 | 2008-05-22 | Olga Boric-Lubecke | Determining presence and/or physiological motion of one or more subjects with quadrature doppler radar receiver systems |
US20080205820A1 (en) * | 2007-02-27 | 2008-08-28 | Intersil Americas Inc. | Dynamically configurable multiple wavelength photodetector array for optical storage applications |
US20080256494A1 (en) * | 2007-04-16 | 2008-10-16 | Greenfield Mfg Co Inc | Touchless hand gesture device controller |
US20080266128A1 (en) * | 2007-04-27 | 2008-10-30 | Sensormatic Electronics Corporation | Handheld data capture system with power and safety monitor and method therefore |
US20090006730A1 (en) * | 2007-06-26 | 2009-01-01 | International Business Machines Corporation | Data eye monitor method and apparatus |
US20090027529A1 (en) * | 2007-07-16 | 2009-01-29 | Jung Sang-Il | Image sensor with wide operating range |
US7486386B1 (en) * | 2007-09-21 | 2009-02-03 | Silison Laboratories Inc. | Optical reflectance proximity sensor |
US7532870B2 (en) * | 2004-04-13 | 2009-05-12 | Maxlinear, Inc. | Method and apparatus for DC offset removal |
US7616032B2 (en) * | 2003-12-30 | 2009-11-10 | Hynix Semiconductor Inc. | Internal voltage initializing circuit for use in semiconductor memory device and driving method thereof |
US7619293B2 (en) * | 2005-06-03 | 2009-11-17 | Sanyo Electric Co., Ltd. | Pin photodiode with improved blue light sensitivity |
US7620202B2 (en) * | 2003-06-12 | 2009-11-17 | Honda Motor Co., Ltd. | Target orientation estimation using depth sensing |
US20090295729A1 (en) * | 2008-06-03 | 2009-12-03 | Asustek Computer Inc. | Input device and operation method of computer system |
US20100150399A1 (en) * | 2008-12-12 | 2010-06-17 | Miroslav Svajda | Apparatus and method for optical gesture recognition |
US20100295773A1 (en) * | 2009-05-22 | 2010-11-25 | Rachid Alameh | Electronic device with sensing assembly and method for interpreting offset gestures |
US20110176069A1 (en) * | 2010-01-21 | 2011-07-21 | Intersil Americas Inc. | Systems and methods for projector light beam alignment |
US20110310005A1 (en) * | 2010-06-17 | 2011-12-22 | Qualcomm Incorporated | Methods and apparatus for contactless gesture recognition |
US8086971B2 (en) * | 2006-06-28 | 2011-12-27 | Nokia Corporation | Apparatus, methods and computer program products providing finger-based and hand-based gesture commands for portable electronic device applications |
Family Cites Families (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB1343298A (en) | 1971-07-30 | 1974-01-10 | Mullard Ltd | Crt display systems |
US4644341A (en) * | 1984-07-05 | 1987-02-17 | Tri-Tronics Company, Inc. | Contrast indicating arrangement in photoelectric sensing systems |
EP0250534B2 (en) | 1985-12-24 | 1994-09-28 | British Broadcasting Corporation | Bandwidth compression for television signals |
JP2945423B2 (en) | 1989-12-01 | 1999-09-06 | 株式会社アドバンテスト | Color image signal evaluation method |
US5469748A (en) * | 1994-07-20 | 1995-11-28 | Micro Motion, Inc. | Noise reduction filter system for a coriolis flowmeter |
US5742379A (en) * | 1995-11-29 | 1998-04-21 | Reifer; Michael H. | Device and method for electronically measuring distances |
JPH09236662A (en) * | 1996-02-29 | 1997-09-09 | Ushikata Shokai:Kk | Electronic distance meter |
US6127884A (en) | 1999-05-24 | 2000-10-03 | Philips Electronics North America Corp. | Differentiate and multiply based timing recovery in a quadrature demodulator |
US6891440B2 (en) | 2000-10-02 | 2005-05-10 | A. Michael Straub | Quadrature oscillator with phase error correction |
EP1384305A4 (en) * | 2001-05-02 | 2005-11-09 | Microchip Tech Inc | Method and apparatus for high-voltage battery array monitoring sensors network |
US7016537B2 (en) * | 2001-07-02 | 2006-03-21 | Trw Inc. | Vehicle occupant sensor apparatus and method including scanned, phased beam transmission for occupant characteristic determination |
US7288755B1 (en) * | 2001-10-19 | 2007-10-30 | Brian P. Platner | Portable handheld artificial light detector |
US6590520B1 (en) * | 2002-02-04 | 2003-07-08 | Lockheed Martin Corporation | Method and system for determining air turbulence using bi-static measurements |
US7184951B2 (en) * | 2002-02-15 | 2007-02-27 | Radiodetection Limted | Methods and systems for generating phase-derivative sound |
DE10305861A1 (en) * | 2003-02-13 | 2004-08-26 | Adam Opel Ag | Motor vehicle device for spatial measurement of a scene inside or outside the vehicle, combines a LIDAR system with an image sensor system to obtain optimum 3D spatial image data |
US20090283699A1 (en) * | 2003-09-29 | 2009-11-19 | Baltz Nathan T | Frequency domain luminescence instrumentation |
JP2005249764A (en) * | 2004-03-02 | 2005-09-15 | Katsumi Mori | Object color measuring system |
JP4199144B2 (en) * | 2004-03-11 | 2008-12-17 | 株式会社東芝 | Weight function generation device, reference signal generation device, transmission signal generation device, signal processing device, and antenna device |
JP2006010506A (en) * | 2004-06-25 | 2006-01-12 | Sharp Corp | Optical ranging sensor and self-running cleaner |
CN1619295A (en) * | 2004-12-10 | 2005-05-25 | 南京农业大学 | Pork colour grading instrument |
US7735037B2 (en) * | 2005-04-15 | 2010-06-08 | Rambus, Inc. | Generating interface adjustment signals in a device-to-device interconnection system |
JP5572315B2 (en) * | 2005-09-08 | 2014-08-13 | コーニンクレッカ フィリップス エヌ ヴェ | Low current measurement with high dynamic range for optical imaging |
DE602006008342D1 (en) * | 2005-10-21 | 2009-09-17 | Nxp Bv | POLAR MODULATION DEVICE AND METHOD USING FM MODULATION |
JP4799216B2 (en) * | 2006-03-03 | 2011-10-26 | 富士通株式会社 | Imaging device having distance measuring function |
JP4116053B2 (en) * | 2006-09-20 | 2008-07-09 | 北陽電機株式会社 | Ranging device |
JP4971744B2 (en) * | 2006-10-18 | 2012-07-11 | パナソニック株式会社 | Spatial information detector using intensity-modulated light |
EP2075935A1 (en) | 2007-12-31 | 2009-07-01 | Motorola, Inc. | A method and apparatus for providing uninterrupted media to a user |
JP5507053B2 (en) * | 2008-03-27 | 2014-05-28 | パナソニック株式会社 | Distance measuring device |
CN101581783B (en) | 2008-05-16 | 2013-04-17 | 深圳市迈测科技有限公司 | Calibration method for phase measurement, device and distance measuring apparatus |
US20100066442A1 (en) * | 2008-09-15 | 2010-03-18 | Fenghao Mu | Method and Apparatus for Tunable Current-Mode Filtering |
JP2010127739A (en) * | 2008-11-27 | 2010-06-10 | Toppan Printing Co Ltd | Spectral sensitivity characteristic measurement apparatus and spectral sensitivity characteristic measurement method |
US8760631B2 (en) | 2010-01-27 | 2014-06-24 | Intersil Americas Inc. | Distance sensing by IQ domain differentiation of time of flight (TOF) measurements |
-
2010
- 2010-12-28 US US12/979,726 patent/US8760631B2/en active Active
-
2011
- 2011-01-25 US US13/013,173 patent/US8274037B2/en active Active
- 2011-01-25 US US13/013,640 patent/US20110180709A1/en not_active Abandoned
- 2011-01-25 US US13/013,199 patent/US20110180693A1/en not_active Abandoned
- 2011-01-25 US US13/013,146 patent/US8530819B2/en active Active
- 2011-01-25 US US13/013,676 patent/US20110182519A1/en not_active Abandoned
- 2011-01-26 CN CN201180015570.2A patent/CN102822759B/en active Active
- 2011-01-26 WO PCT/US2011/022650 patent/WO2011094365A1/en active Application Filing
- 2011-01-26 CN CN2011800159510A patent/CN102822691A/en active Pending
- 2011-01-26 WO PCT/US2011/022649 patent/WO2011094364A1/en active Application Filing
- 2011-01-26 WO PCT/US2011/022644 patent/WO2011094360A1/en active Application Filing
- 2011-01-26 WO PCT/US2011/022647 patent/WO2011094362A1/en active Application Filing
- 2011-01-26 CN CN201610559617.6A patent/CN106052862B/en active Active
- 2011-01-26 WO PCT/US2011/022651 patent/WO2011094366A1/en active Application Filing
- 2011-01-26 WO PCT/US2011/022646 patent/WO2011094361A1/en active Application Filing
- 2011-01-27 TW TW100103040A patent/TWI439716B/en active
- 2011-01-27 TW TW103115498A patent/TWI531809B/en active
- 2011-01-27 TW TW100103043A patent/TWI445993B/en active
- 2011-01-27 TW TW100103042A patent/TW201205047A/en unknown
- 2011-01-27 TW TW100103046A patent/TW201145075A/en unknown
- 2011-01-27 TW TW100103045A patent/TW201203008A/en unknown
-
2014
- 2014-06-04 US US14/296,277 patent/US10031078B2/en active Active
Patent Citations (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3996590A (en) * | 1961-02-02 | 1976-12-07 | Hammack Calvin M | Method and apparatus for automatically detecting and tracking moving objects and similar applications |
US4458212A (en) * | 1981-12-30 | 1984-07-03 | Mostek Corporation | Compensated amplifier having pole zero tracking |
US4542475A (en) * | 1982-11-23 | 1985-09-17 | Rca Corporation | Sampled data filter system as for a digital TV |
US4551710A (en) * | 1983-04-29 | 1985-11-05 | Cerberus Ag | Method and apparatus for reporting dangerous conditions |
US4648364A (en) * | 1985-07-08 | 1987-03-10 | Wills William H | Engine protection apparatus |
US4942561A (en) * | 1988-06-29 | 1990-07-17 | Kabushikikaisha Topcon | Delay time measuring device |
US5055671A (en) * | 1990-10-03 | 1991-10-08 | Spacelabs, Inc. | Apparatus for detecting transducer movement using a first and second light detector |
US20020097743A1 (en) * | 1993-03-09 | 2002-07-25 | Ertugrul Baydar | Integrated digital loop carrier system with virtual tributary mapper circuit |
US5563701A (en) * | 1993-12-29 | 1996-10-08 | Hyundai Electronics Industries Co., Ltd. | Apparatus and method for measuring distance optically using phase variation |
US5593430A (en) * | 1995-01-27 | 1997-01-14 | Pacesetter, Inc. | Bus system for interconnecting an implantable medical device with a plurality of sensors |
US5828899A (en) * | 1996-01-04 | 1998-10-27 | Compaq Computer Corporation | System for peripheral devices recursively generating unique addresses based on the number of devices connected dependent upon the relative position to the port |
US5892540A (en) * | 1996-06-13 | 1999-04-06 | Rockwell International Corporation | Low noise amplifier for passive pixel CMOS imager |
US6111256A (en) * | 1997-04-10 | 2000-08-29 | Shmuel Hershkovitz & Pinhas Shpater | Infrared motion detection signal sampler |
US5990409A (en) * | 1997-12-26 | 1999-11-23 | Roland Kabushiki Kaisha | Musical apparatus detecting maximum values and/or peak values of reflected light beams to control musical functions |
US6462726B1 (en) * | 1998-02-19 | 2002-10-08 | Matsushita Electric Industrial Co., Ltd. | Video signal processor |
US6392539B1 (en) * | 1998-07-13 | 2002-05-21 | Honda Giken Kogyo Kabushiki Kaisha | Object detection apparatus |
US6888938B2 (en) * | 1999-05-11 | 2005-05-03 | Agere Systems Inc. | Dynamically adjustable digital gyrator having extendable feedback for stable DC load line |
US6819782B1 (en) * | 1999-06-08 | 2004-11-16 | Matsushita Electric Industrial Co., Ltd. | Device and method for recognizing hand shape and position, and recording medium having program for carrying out the method recorded thereon |
US20060120621A1 (en) * | 2000-01-06 | 2006-06-08 | Canon Kabushiki Kaisha | Demodulation and phase estimation of two-dimensional patterns |
US6753950B2 (en) * | 2000-01-26 | 2004-06-22 | Instro Precision Limited | Optical distance measurement |
US7212655B2 (en) * | 2000-09-15 | 2007-05-01 | Tumey David M | Fingerprint verification system |
US6803555B1 (en) * | 2001-09-07 | 2004-10-12 | Indigo Systems Corporation | Two-stage auto-zero amplifier circuit for electro-optical arrays |
US6744248B2 (en) * | 2001-12-08 | 2004-06-01 | Koninklijke Philips Electronics N.V. | Arrangement for detecting motion of an encoder |
US20030234341A1 (en) * | 2002-06-20 | 2003-12-25 | Osborn Jon V. | Microelectromechanical system optical sun sensor |
US6836212B2 (en) * | 2002-10-10 | 2004-12-28 | Motorola, Inc. | Method and apparatus for reducing the likelihood of losing a portable electronic device |
US20040140961A1 (en) * | 2003-01-17 | 2004-07-22 | Eastman Kodak Company | Oled display and touch screen |
US7620202B2 (en) * | 2003-06-12 | 2009-11-17 | Honda Motor Co., Ltd. | Target orientation estimation using depth sensing |
US7616032B2 (en) * | 2003-12-30 | 2009-11-10 | Hynix Semiconductor Inc. | Internal voltage initializing circuit for use in semiconductor memory device and driving method thereof |
US7532870B2 (en) * | 2004-04-13 | 2009-05-12 | Maxlinear, Inc. | Method and apparatus for DC offset removal |
US7619293B2 (en) * | 2005-06-03 | 2009-11-17 | Sanyo Electric Co., Ltd. | Pin photodiode with improved blue light sensitivity |
US20070013791A1 (en) * | 2005-07-05 | 2007-01-18 | Koichi Kinoshita | Tracking apparatus |
US20070121095A1 (en) * | 2005-11-28 | 2007-05-31 | Robert Lewis | Distance measurement device with short range optics |
US20080119716A1 (en) * | 2006-05-17 | 2008-05-22 | Olga Boric-Lubecke | Determining presence and/or physiological motion of one or more subjects with quadrature doppler radar receiver systems |
US8086971B2 (en) * | 2006-06-28 | 2011-12-27 | Nokia Corporation | Apparatus, methods and computer program products providing finger-based and hand-based gesture commands for portable electronic device applications |
US20080205820A1 (en) * | 2007-02-27 | 2008-08-28 | Intersil Americas Inc. | Dynamically configurable multiple wavelength photodetector array for optical storage applications |
US20080256494A1 (en) * | 2007-04-16 | 2008-10-16 | Greenfield Mfg Co Inc | Touchless hand gesture device controller |
US20080266128A1 (en) * | 2007-04-27 | 2008-10-30 | Sensormatic Electronics Corporation | Handheld data capture system with power and safety monitor and method therefore |
US20090006730A1 (en) * | 2007-06-26 | 2009-01-01 | International Business Machines Corporation | Data eye monitor method and apparatus |
US20090027529A1 (en) * | 2007-07-16 | 2009-01-29 | Jung Sang-Il | Image sensor with wide operating range |
US7486386B1 (en) * | 2007-09-21 | 2009-02-03 | Silison Laboratories Inc. | Optical reflectance proximity sensor |
US20090295729A1 (en) * | 2008-06-03 | 2009-12-03 | Asustek Computer Inc. | Input device and operation method of computer system |
US20100150399A1 (en) * | 2008-12-12 | 2010-06-17 | Miroslav Svajda | Apparatus and method for optical gesture recognition |
US20100295773A1 (en) * | 2009-05-22 | 2010-11-25 | Rachid Alameh | Electronic device with sensing assembly and method for interpreting offset gestures |
US20110176069A1 (en) * | 2010-01-21 | 2011-07-21 | Intersil Americas Inc. | Systems and methods for projector light beam alignment |
US20110310005A1 (en) * | 2010-06-17 | 2011-12-22 | Qualcomm Incorporated | Methods and apparatus for contactless gesture recognition |
Non-Patent Citations (1)
Title |
---|
IEEE 802.5v-2001 (Amendment to IEEE Std 802.5, 1998 Edition and IEEE Stds 802.5r and 802.5j, 1998 Edition) 16 Nov. 2001. * |
Cited By (36)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8274037B2 (en) * | 2010-01-27 | 2012-09-25 | Intersil Americas Inc. | Automatic calibration technique for time of flight (TOF) transceivers |
US10031078B2 (en) | 2010-01-27 | 2018-07-24 | Intersil Americas LLC | Distance sensing by IQ domain differentiation of time of flight (TOF) measurements |
US20110181892A1 (en) * | 2010-01-27 | 2011-07-28 | Intersil Americas Inc. | Automatic calibration technique for time of flight (tof) transceivers |
US8912496B2 (en) * | 2010-04-13 | 2014-12-16 | Silicon Laboratories Inc. | Apparatus with optical functionality and power management and associated methods |
US20110248172A1 (en) * | 2010-04-13 | 2011-10-13 | Rueger Timothy T | Apparatus with optical functionality and power management and associated methods |
US11112872B2 (en) * | 2011-04-13 | 2021-09-07 | Nokia Technologies Oy | Method, apparatus and computer program for user control of a state of an apparatus |
US20140033141A1 (en) * | 2011-04-13 | 2014-01-30 | Nokia Corporation | Method, apparatus and computer program for user control of a state of an apparatus |
US9063704B2 (en) * | 2011-05-05 | 2015-06-23 | Net Power And Light, Inc. | Identifying gestures using multiple sensors |
US20120280905A1 (en) * | 2011-05-05 | 2012-11-08 | Net Power And Light, Inc. | Identifying gestures using multiple sensors |
US9298333B2 (en) | 2011-12-22 | 2016-03-29 | Smsc Holdings S.A.R.L. | Gesturing architecture using proximity sensing |
US9122354B2 (en) * | 2012-03-14 | 2015-09-01 | Texas Instruments Incorporated | Detecting wave gestures near an illuminated surface |
US20130241888A1 (en) * | 2012-03-14 | 2013-09-19 | Texas Instruments Incorporated | Detecting Wave Gestures Near an Illuminated Surface |
US10664062B2 (en) | 2012-05-11 | 2020-05-26 | Comcast Cable Communications, Llc | System and method for controlling a user experience |
US20130300644A1 (en) * | 2012-05-11 | 2013-11-14 | Comcast Cable Communications, Llc | System and Methods for Controlling a User Experience |
US11093047B2 (en) | 2012-05-11 | 2021-08-17 | Comcast Cable Communications, Llc | System and method for controlling a user experience |
US9619036B2 (en) * | 2012-05-11 | 2017-04-11 | Comcast Cable Communications, Llc | System and methods for controlling a user experience |
US9081571B2 (en) | 2012-11-29 | 2015-07-14 | Amazon Technologies, Inc. | Gesture detection management for an electronic device |
WO2014085269A1 (en) * | 2012-11-29 | 2014-06-05 | Amazon Technologies, Inc. | Gesture detection management for an electronic device |
US9367137B2 (en) | 2013-07-01 | 2016-06-14 | Blackberry Limited | Alarm operation by touch-less gesture |
US9398221B2 (en) | 2013-07-01 | 2016-07-19 | Blackberry Limited | Camera control using ambient light sensors |
US9423913B2 (en) | 2013-07-01 | 2016-08-23 | Blackberry Limited | Performance control of ambient light sensors |
US9342671B2 (en) | 2013-07-01 | 2016-05-17 | Blackberry Limited | Password by touch-less gesture |
US9489051B2 (en) | 2013-07-01 | 2016-11-08 | Blackberry Limited | Display navigation using touch-less gestures |
US9323336B2 (en) | 2013-07-01 | 2016-04-26 | Blackberry Limited | Gesture detection using ambient light sensors |
US9865227B2 (en) | 2013-07-01 | 2018-01-09 | Blackberry Limited | Performance control of ambient light sensors |
US9928356B2 (en) | 2013-07-01 | 2018-03-27 | Blackberry Limited | Password by touch-less gesture |
US9256290B2 (en) | 2013-07-01 | 2016-02-09 | Blackberry Limited | Gesture detection using ambient light sensors |
US9405461B2 (en) | 2013-07-09 | 2016-08-02 | Blackberry Limited | Operating a device using touchless and touchscreen gestures |
US9465448B2 (en) | 2013-07-24 | 2016-10-11 | Blackberry Limited | Backlight for touchless gesture detection |
US9304596B2 (en) | 2013-07-24 | 2016-04-05 | Blackberry Limited | Backlight for touchless gesture detection |
US9194741B2 (en) | 2013-09-06 | 2015-11-24 | Blackberry Limited | Device having light intensity measurement in presence of shadows |
US11284808B2 (en) | 2014-10-11 | 2022-03-29 | Linet Spol. S.R.O. | Device and method for measurement of vital functions, including intracranial pressure, and system and method for collecting data |
US11076789B2 (en) * | 2016-06-22 | 2021-08-03 | Linet Spol. S.R.O | Medical data collection system and method of use thereof |
US11599199B2 (en) | 2019-11-28 | 2023-03-07 | Boe Technology Group Co., Ltd. | Gesture recognition apparatus, gesture recognition method, computer device and storage medium |
US20230089705A1 (en) * | 2021-09-17 | 2023-03-23 | Melexis Bulgaria EOOD | Sensing system comprising a chain of sensors |
US11933639B2 (en) * | 2021-09-17 | 2024-03-19 | Melexis Bulgaria EOOD | Sensing system comprising a chain of sensors |
Also Published As
Publication number | Publication date |
---|---|
TW201203008A (en) | 2012-01-16 |
TWI445993B (en) | 2014-07-21 |
WO2011094364A1 (en) | 2011-08-04 |
WO2011094365A1 (en) | 2011-08-04 |
CN102822691A (en) | 2012-12-12 |
TW201202731A (en) | 2012-01-16 |
US10031078B2 (en) | 2018-07-24 |
US20110181861A1 (en) | 2011-07-28 |
US20110182519A1 (en) | 2011-07-28 |
US8530819B2 (en) | 2013-09-10 |
US8274037B2 (en) | 2012-09-25 |
TW201202730A (en) | 2012-01-16 |
CN106052862B (en) | 2018-08-28 |
TWI439716B (en) | 2014-06-01 |
CN106052862A (en) | 2016-10-26 |
US8760631B2 (en) | 2014-06-24 |
US20110180693A1 (en) | 2011-07-28 |
WO2011094362A1 (en) | 2011-08-04 |
US20110181254A1 (en) | 2011-07-28 |
TWI531809B (en) | 2016-05-01 |
CN102822759A (en) | 2012-12-12 |
US20140327900A1 (en) | 2014-11-06 |
TW201145075A (en) | 2011-12-16 |
US20110181892A1 (en) | 2011-07-28 |
WO2011094361A1 (en) | 2011-08-04 |
WO2011094366A1 (en) | 2011-08-04 |
WO2011094360A1 (en) | 2011-08-04 |
CN102822759B (en) | 2014-08-20 |
TW201432283A (en) | 2014-08-16 |
TW201205047A (en) | 2012-02-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20110180709A1 (en) | Serial-chaining proximity sensors for gesture recognition | |
CN101944897B (en) | Proximity sensors with improved ambient light rejection | |
US9507051B2 (en) | Photoelectric switch for detection of an object and method for operating the same | |
CN105474147A (en) | Capacitive touch system | |
US9959128B2 (en) | Digital sensor system | |
US10895921B2 (en) | Touch sensitive processing apparatus, system and operating method thereof for receiving electrical signals carrying pressure information | |
US11392221B2 (en) | Touch sensitive processing apparatus, system and operating method thereof for receiving electrical signals carrying pressure information | |
US10758134B2 (en) | Sensor, sensor apparatus, and electronic device | |
US8013714B2 (en) | RFID sensor using pulse processing | |
US10890987B2 (en) | Stylus and operating method thereof for transmitting electrical signals carrying pressure information | |
US8026795B2 (en) | RFID sensor array and sensor group based on pulse-processing | |
JP2008204234A (en) | Rfid tag built-in saw (surface acoustic wave) sensor and measurement system | |
CN206584312U (en) | A kind of key mapping and keyboard realized every sky reaction | |
US11409379B2 (en) | Stylus and operating method thereof for transmitting electrical signals carrying pressure information | |
US20180123688A1 (en) | Methods and devices for monitoring optical signals | |
Jeeva et al. | Study of inter-integrated circuit color sensor with microcontroller | |
CN211234464U (en) | Grating device | |
KR20040085628A (en) | Wireless mouse for having Power saving function and power saving method thereof | |
US10305520B2 (en) | Removing RF interference through scan rate locking | |
US9773200B2 (en) | Miniature integrated sensor circuit | |
KR102344783B1 (en) | Touch screen for interlocking with electronic pen | |
Salameh et al. | A personal computer based general purpose control system using general remote control |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: INTERSIL AMERICAS INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CRADDOCK, C. WARREN;RITTER, DAVID W.;GOLDEN, PHILIP;REEL/FRAME:025696/0097 Effective date: 20110121 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |