US20050063564A1 - Hand pattern switch device - Google Patents

Hand pattern switch device Download PDF

Info

Publication number
US20050063564A1
US20050063564A1 US10/915,952 US91595204A US2005063564A1 US 20050063564 A1 US20050063564 A1 US 20050063564A1 US 91595204 A US91595204 A US 91595204A US 2005063564 A1 US2005063564 A1 US 2005063564A1
Authority
US
United States
Prior art keywords
hand
pattern
hand pattern
switch device
finger
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10/915,952
Inventor
Keiichi Yamamoto
Hiromitsu Sato
Shinji Ozawa
Hideo Saito
Hiroya Igarashi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Mitsubishi Fuso Truck and Bus Corp
Keio University
Original Assignee
Mitsubishi Fuso Truck and Bus Corp
Keio University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mitsubishi Fuso Truck and Bus Corp, Keio University filed Critical Mitsubishi Fuso Truck and Bus Corp
Assigned to MITSUBISHI FUSO TRUCK AND BUS CORPORATION, KEIO UNIVERSITY reassignment MITSUBISHI FUSO TRUCK AND BUS CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: IGARASHI, HIROYA, OZAWA, SHINJI, SAITO, HIDEO, SATO, HIROMITSU, YAMAMOTO, KEIICHI
Publication of US20050063564A1 publication Critical patent/US20050063564A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R25/00Fittings or systems for preventing or indicating unauthorised use or theft of vehicles
    • B60R25/20Means to switch the anti-theft system on or off
    • B60R25/2045Means to switch the anti-theft system on or off by hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • G06V10/12Details of acquisition arrangements; Constructional details thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language
    • B60K2360/21
    • EFIXED CONSTRUCTIONS
    • E05LOCKS; KEYS; WINDOW OR DOOR FITTINGS; SAFES
    • E05FDEVICES FOR MOVING WINGS INTO OPEN OR CLOSED POSITION; CHECKS FOR WINGS; WING FITTINGS NOT OTHERWISE PROVIDED FOR, CONCERNED WITH THE FUNCTIONING OF THE WING
    • E05F15/00Power-operated mechanisms for wings
    • EFIXED CONSTRUCTIONS
    • E05LOCKS; KEYS; WINDOW OR DOOR FITTINGS; SAFES
    • E05YINDEXING SCHEME RELATING TO HINGES OR OTHER SUSPENSION DEVICES FOR DOORS, WINDOWS OR WINGS AND DEVICES FOR MOVING WINGS INTO OPEN OR CLOSED POSITION, CHECKS FOR WINGS AND WING FITTINGS NOT OTHERWISE PROVIDED FOR, CONCERNED WITH THE FUNCTIONING OF THE WING
    • E05Y2400/00Electronic control; Power supply; Power or signal transmission; User interfaces
    • E05Y2400/80User interfaces
    • E05Y2400/85User input means
    • E05Y2400/852Sensors
    • E05Y2400/856Actuation thereof
    • E05Y2400/858Actuation thereof by body parts
    • E05Y2400/86Actuation thereof by body parts by hand

Definitions

  • the present invention relates to a hand pattern switch device suitable for a driver to easily operate vehicle-mounted equipment such as air conditioner equipment and audio equipment and ancillary vehicle equipment such as side mirrors, without his/her driving being affected and without the need of touching an operation panel of the vehicle-mounted equipment.
  • vehicle-mounted equipment such as air conditioner equipment and audio equipment
  • ancillary vehicle equipment such as side mirrors
  • This kind of art realized by the pattern recognition to recognize a hand pattern from a picked-up image of a hand or realized by the motion detection to detect a hand motion by tracing a positional change of a recognized hand, is called as a hand pattern switch device in the present specification for the sake of convenience.
  • the pattern or motion of the driver's (operator's) hand must be detected reliably and accurately. To this end, it is necessary to accurately recognize which part of the picked-up image corresponds to the driver's (operator's) hand.
  • the driver (operator) sometimes wears a long sleeve shirt, a wrist watch, or the like. In that case, a wrist portion in the input image is detected to be extraordinary large, or detected to be disconnected due to the presence of an image component corresponding to the wrist watch or the like.
  • palm a portion corresponding to the driver's palm or the back of his/her hand (hereinafter collectively referred to as palm) being unable to be detected with reliability, despite that such portion is to be detected for pattern recognition.
  • the prior art poses a further problem of the processing load being increased, since it generally uses a complicated image processing technique, such as region segmentation, or a matching technique in which a predetermined standard hand pattern is referred to.
  • the object of this invention is to provide a hand pattern switch device capable of easily and reliably detecting a hand pattern or a hand motion of a driver (operator) observed when the driver operates various vehicle-mounted equipment and ancillary vehicle equipment, thereby properly inputting information used for operations of these equipment.
  • a hand pattern switch device which has image pickup means for picking up an image of a distal arm that is within a predetermined image pickup zone and in which a hand pattern and/or a motion of a finger of a hand is detected from the image picked up by the image pickup means to obtain predetermined switch operation information.
  • the hand pattern switch device comprises first image processing means for determining a central axis passing through a center of the arm based on the picked-up image, scanning line setting means for setting at least either a first scanning line extending perpendicular to the central axis or a second scanning line extending along the central axis, and determination means for determining whether or not any finger of the hand is extended based on the at least either the first or second scanning line set by the scanning line setting means.
  • FIG. 1 is a view showing the outline of structure of a hand pattern switch device according to an embodiment of this invention
  • FIG. 2 is a view showing a hand/finger image pickup zone in the hand pattern switch device shown in FIG. 1 ;
  • FIG. 3 is a flowchart showing an example of processing procedures for recognition of hand pattern and palm center
  • FIG. 4 is a conceptual view for explaining the processing for recognition of hand pattern and palm center shown in FIG. 3 ;
  • FIG. 5A is a view for explaining drawbacks of conventional typical processing for hand pattern recognition
  • FIG. 5B is a view similar to FIG. 5A ;
  • FIG. 6A is a view showing hand pattern 1 used in the embodiment of this invention.
  • FIG. 6B is a view showing hand pattern 2 ;
  • FIG. 6C is a view showing hand pattern 3 ;
  • FIG. 6D is a view showing hand pattern 4 ;
  • FIG. 7 is a flowchart showing an example of processing procedures for hand pattern recognition performed by an instructed-operation recognizing section in the hand pattern switch device shown in FIG. 1 ;
  • FIG. 8 is a flowchart showing an example of processing procedures for operation amount detection
  • FIG. 9 is a flowchart showing an example of processing procedures for operation amount detection in a time mode
  • FIG. 10 is a flowchart showing an example of processing procedures for operation amount detection in a distance/time mode
  • FIG. 11 is a view showing input modes of inputting switch-operation information to the hand pattern switch device with use of hand/fingers.
  • FIG. 12 is a view showing an example of systemized selection of controlled objects.
  • FIG. 1 is a view of a general construction of essential part of the hand pattern switch device according to the present embodiment, showing a state around a driver's seat of a vehicle and functions of the hand pattern switch device realized for example by a microcomputer (ECU) and the like.
  • ECU microcomputer
  • a steering wheel 1 adapted to be steered by a driver, a combination switch (not shown), etc. are provided, whereas an operating section 2 for audio equipment, air conditioner equipment, etc. is provided on a console panel.
  • a video camera 3 is disposed for picking up an image of a hand of the driver who extends his/her arm to an image pickup zone located laterally to the steering wheel 1 .
  • the camera 3 is comprised of a small-sized CCD camera or the like.
  • the camera 3 may be the one which obtains a visible light image under predetermined illumination (daytime).
  • a so-called infrared camera which emits near-infrared light to the pickup zone to obtain an infrared image may be used, when the illumination for the pickup zone is insufficient, as in nighttime.
  • the hand pattern switch device the hand pattern is changed by selectively flexing desired one or ones of the fingers, with the palm positioned horizontally in the pickup zone, and the palm position is displaced (moved) back and forth and left and right.
  • the term “palm” is used in the description here that represents not only the palm but also the back of hand whose image is to be picked up.
  • the hand pattern switch device performs the processing to recognize a driver's hand pattern or a hand motion on the basis of an image picked up by and input from the camera 3 , and based on results of the recognition, acquires predetermined corresponding switch-operation information.
  • the hand pattern switch device serves, instead of the operating section 2 , to provide switch-operation information to the audio equipment, air conditioner equipment, etc.
  • the hand pattern switch device comprises a binarization processing section 11 for binarizing an input image picked up by the camera 3 so that background image components are removed to extract image components corresponding to the distal arm, mainly the palm and fingers of the hand, from the picked-up image; a centroid detecting section 12 for determining a centroid position of the hand based on the image of the palm and fingers of the hand extracted by the binarization processing; and a pattern recognition section 13 for recognizing a hand/finger pattern.
  • a binarization processing section 11 for binarizing an input image picked up by the camera 3 so that background image components are removed to extract image components corresponding to the distal arm, mainly the palm and fingers of the hand, from the picked-up image
  • a centroid detecting section 12 for determining a centroid position of the hand based on the image of the palm and fingers of the hand extracted by the binarization processing
  • a pattern recognition section 13 for recognizing a hand/finger pattern.
  • the hand pattern switch device further comprises an instructed-operation recognizing section 14 for recognizing a switch operation given by the driver by the hand pattern or hand motion, based on results of recognition performed by the pattern recognition section 13 and the centroid position of the hand detected by the centroid detecting section 12 .
  • This instructed-operation recognizing section 14 generally comprises a function determination section 16 for determining (identifying) a type of operation intended by the hand pattern recognized as mentioned above, referring to a relation between hand patterns registered beforehand in a memory 15 and their functions, a displacement detecting section 17 for tracing a motion of centroid position of palm with a particular finger pattern or a motion of fingertip to thereby detect a displacement thereof from its reference position, and a timer 18 for monitoring the palm motion or fingertip motion in terms of elapsed time during the palm or fingertip is moved.
  • the instructed-operation recognizing section 14 determines predetermined switch-operation information specified by the driver's hand pattern and palm motion, and outputs this switch-operation information by way of example to the audio equipment, air conditioner equipment, or the like.
  • the instructed-operation recognizing section 14 is further provided with a guidance section 19 that provides a predetermined guidance to the driver according to results of the aforementioned determination, etc.
  • the driver is notified of the guidance form a speaker 20 in the form of a speech message that specifies for example the audio equipment or air conditioner equipment (controlled object equipment), or volume/channel setting, wind volume/temperature, or the like (controlled object function), or in the form of confirmation sound such as pip tone or beep tone that identifies a switch operation (operation amount) having been made.
  • the instructed-operation recognizing section 14 i.e., control of output of switch-operation information in respect of controlled objects such as audio equipment, air conditioner equipment, and the like, explanations will be given later.
  • the image pickup zone 3 a of the camera 3 located laterally to the steering wheel 1 is at least 50 mm, preferably about 100 mm, apart from the outer periphery of the steering wheel 1 .
  • the image pickup zone is at a position to which the driver can extend the arm without changing a driving posture, while resting the arm on an arm rest 5 that is provided laterally to the driver's seat and which is located away from the operating section 2 for audio equipment, etc., so that the hand extended to the image pickup zone does not touch the operating section 2 .
  • the image pickup zone 3 a is rectangle in shape and has a size of about 600 mm in a fingertip direction of and about 350 mm in a width direction of the driver's hand extended laterally to the steering wheel 1 .
  • the image pickup zone 3 a is a zone that is set such that an image of the driver's hand is not picked up when the driver holds the steering wheel 1 or operates the combination switch (not shown) provided at the steering column shaft and such that the driver can move his/her hand into the zone without largely moving the arm.
  • a hand motion for a driving operation or a hand/finger motion for a direct operation of the operating section 2 of the audio equipment, etc. is prevented from being erroneously detected as a motion for providing switch-operation information.
  • a pressure-sensitive sensor for example may be provided in the gearshift lever to make a detection as to whether the gearshift lever is grasped by the driver.
  • the provision of such sensor makes it possible to easily determine which of the gearshift lever or the hand pattern switch device is operated by the driver's hand extended to the image pickup zone 3 a, whereby a driving operation is prevented from being erroneously detected as a switch operation.
  • a height of driver's hand may be detected by using a stereoscopic camera serving as the camera 3 , to determine whether the driver's hand extended to the image pickup zone 3 a operates the gearshift lever or is present in a space above the gearshift lever.
  • the setting of the image pickup zone 3 a is made based on a range (displacement width) of arm/hand motion to which the driver can naturally extend the arm without changing a driving posture while resting the arm (elbow) on the arm rest 5 and to which the driver can comfortably and naturally move the arm/hand when making the imaginary switch operation.
  • the image pickup zone 3 a is determined to be a rectangle in shape and to have a 600 mm length and a 350 mm width, as mentioned above.
  • the driver's hand coming off the steering wheel 1 and then naturally moved without a sense of incompatibility can be captured without fail and without a hand/arm motion for a driving operation being erroneously detected. It is also possible to reliably grasp a change in hand position or a hand motion for switch operation in the image pickup zone 3 a, so that the hand pattern recognition and the detection of an amount of hand motion (deviation) can easily be made with relatively simplified image processing.
  • the driver For the driver, he/she can perform a desired switch operation by simply moving the hand after forming a corresponding one of predetermined hand patterns, while extending the arm laterally to the steering wheel 1 without changing a driving posture and without directly touching the operating section 2 for audio equipment, etc. This reduces a load of the driver performing the switch operation.
  • a hand motion and/or an arm motion for a driving operation cannot erroneously be detected as instruction for switch operation, there are advantages that the driver can concentrate on driving without paying attention to the hand pattern switch device, and can, where required, easily give instruction for switch operation by simply moving his/her hand (palm) to the image pickup zone 3 a.
  • an input image of the pickup zone 3 a picked up by the camera 3 is subject to binarization processing using a predetermined threshold value, whereby the background of the image is discriminated as black from other portions of the image, as white, corresponding to the arm and palm (step S 1 ).
  • a binarized image of the pickup zone 3 a is obtained as exemplarily shown in FIG. 4 .
  • the centroid G of the white region of the binarized image, corresponding to image components of the arm and palm, is determined.
  • a central axis B is determined from the centroid G and a longitudinal moment that is determined by analyzing the direction in which white elements are present passing through the centroid G (step S 2 ).
  • a plurality of first scanning lines Si extending at right angles with respect to the central axis B are set at equal intervals between an upper side or fingertip side of the binarized image and the centroid G (step S 3 ).
  • widths W of the white image on the respective scanning lines S 1 are determined in sequence from the upper side of the binarized image, to thereby detect the scanning line Si which is maximum in the white image width W (step S 4 ).
  • the detection of the scanning line S 1 that is maximum in white image width W a determination is made whether or not the white image width W detected sequentially from the upper side of the image is larger than that detected on the immediately preceding scanning line S 1 , and when a peak appears in the detected widths, the associated scanning line S 1 is detected as the one having the maximum white image width.
  • a point of intersection of the thus detected scanning line S 1 having the maximum white image width W and the central axis B is detected as a palm center position C (step S 5 ).
  • a number of scanning lines S 1 for each of which a width equal to or larger than the predetermined width w has been detected is determined (step S 3 ). More specifically, an examination is made whether the width detected on each scanning line S 1 is equal to or larger than the predetermined width w that is set to a value of ⁇ fraction (1/7) ⁇ to 1 ⁇ 4 of the maximum width W at the palm center position C.
  • the thumb finger can be detected in a similar manner. Since the detection object is the left hand and the thumb finger is extended in a direction different from the direction in which the forefinger is extended, second scanning lines S 2 used to detect the thumb finger are set on the right side of the palm center position C so as to be inclined at an angle of about 10 degrees relative to the central axis B (step S 7 ). This setting is based on the fact that the thumb finger extends slightly obliquely with respect to the central axis B when it is extended to be opened at the maximum. By setting the second scanning lines S 2 such that the thumb finger opened at the maximum extends substantially perpendicular to the scanning lines S 2 , a reliable detection of the thumb finger can be achieved.
  • the predetermined number or more of scanning lines S 2 is detected, it is detected that the thumb finger is extended from the palm to the right side.
  • information can be determined that represents the hand pattern in the pickup zone 3 a and the palm center position C. Then, determinations are made whether or not the forefinger is detected and whether or not the thumb finger is detected, thereby determining which pattern is formed among the following: a clenched-fist pattern (hand pattern 1 ) in which all the fingers are bent into the palm; a finger-up pattern (hand pattern 2 ) in which only the forefinger is extended; an acceptance (OK) pattern (hand pattern 3 ) in which only the thumb finger is extended horizontally; and an L-shaped pattern (hand pattern 4 ) in which the forefinger and the thumb finger are extended. These patterns are shown in FIGS. 6A-6D , respectively.
  • the L-shaped pattern (hand pattern 4 ) is used to instruct the start of operation to the hand pattern switch device.
  • the hand pattern 3 is used in combination of the clenched-fist pattern (hand pattern 1 ) to express image of depressing a push button, by changing the hand pattern by putting the thumb finger in and out (flexing).
  • the hand pattern 3 is used to input information for selection of controlled objects.
  • the finger-up pattern (hand pattern 2 ) is to express image of an indicating needle of an analog meter, and is used to instruct an amount of operation to the controlled object by changing the-position of fingertip (or palm).
  • the clenched-fist pattern (hand pattern 1 ) is also used to instruct completion of operation of the hand pattern switch device.
  • the hand pattern and the change in palm position recognized as mentioned above are subject to the recognition processing that is performed in accordance with procedures exemplarily shown in FIG. 7 , whereby switch operations by means of the driver's (switch operator's) hand are interpreted and switch-operation information is output to the controlled objects.
  • step S 18 whether or not the hand pattern is the clenched-fist pattern (hand pattern 1 ) is then determined (step S 18 ). If the hand pattern 1 is detected, whether the immediately precedingly detected hand pattern was the hand pattern 3 is determined (step S 19 ).
  • the controlled object is changed, considering that the change in hand pattern is the instruction to make changeover of the controlled objects (step S 20 ).
  • the change in controlled object there may be a case where controlled objects are three, one for sound volume in audio equipment, one for temperature in air conditioner, and one for wind amount in air conditioner. In such a case, these controlled objects may be cyclically changed over as mentioned later.
  • step S 16 If the hand pattern 3 is not detected in a state where the controlled object selection mode is not set (step S 16 ), or if the controlled object selection mode is released (step S 22 ), whether or not the hand pattern is the finger-up pattern (hand pattern 2 ) with only the forefinger extended is determined (step S 23 ). When the hand pattern 2 is detected, the below-mentioned processing to detect the switch operation amount is carried out (step S 24 ). If the hand pattern 2 is not detected at step S 23 , whether or not the hand pattern is the clenched-fist pattern (hand pattern 1 ) is determined (step S 25 ).
  • step S 27 If the hand pattern 1 is not maintained for the predetermined time T or more, that is, if the hand pattern is changed to the hand pattern 2 again within the predetermined time T (step S 27 ), the processing from step S 11 is resumed, making it possible to perform reoperation.
  • step S 31 As for data subsequently input, it is determined that the flag K is set (step S 31 ), and therefore, the distance of deviation between the palm center position C determined at that time and the reference position C 0 , i.e., a moved distance D from the reference position C 0 , is determined (step S 35 ). In order to calculate the moved distance D, it is enough to determine a distance between picture elements in the input image. In accordance with the moved distance thus determined and predetermined modes for the detection of operation amount (S 36 ), processing to detect the operation amount is selectively carried out in a time mode (step S 37 ) or in a distance/time mode (step S 38 ).
  • the time mode is a mode in which switch-operation information is output in accordance with a stop time for which the hand displaced from the reference position C 0 is kept stopped, and is suitable for example for adjustment of sound volume in audio equipment and for adjustment of temperature in air conditioner.
  • the distance/time mode is a mode in which the switch-operation information determined according to an amount of hand motion is output when the hand is moved slightly, whereas the information determined according to a stop time of the hand at a moved position is output when the hand has been moved by a predetermined distance or more to the moved position.
  • the distance/time mode is suitable for example for controlled objects that are subject to a fine adjustment after being roughly adjusted.
  • the instruction to input the switch operation amount by means of the finger-up pattern is carried out by moving the palm to the right and left around the arm on the arm rest 5 or by moving the hand right and left around the wrist as fulcrum.
  • Such palm/hand motion to the right and left is performed within a range not falling outside of the pickup zone 3 a, for instance, within an angular range of about ⁇ 45 degrees.
  • the amount of palm motion with the hand pattern 2 is detected n steps.
  • a preset value H or ⁇ H used for determination of maximum motion amount
  • step S 40 If it is determined at step S 40 that the moved distance D exceeds the preset value (threshold value) H, the timer value t is counted up (step S 42 ). When the counted-up timer value t reaches a reference time T (step S 43 ), the setting (switch-operation information) at that time is increased by one stage (step S 44 ). Then, the timer value t is reset to 0 (step S 45 ), and the processing from step S 11 is resumed.
  • step S 40 If it is determined at step S 40 that the moved distance D exceeds the threshold value ⁇ H in the opposite direction, the timer value t is counted up (step S 46 ). When it is determined that the counted-up timer value t reaches the reference time T (step S 47 ), the setting (switch-operation information) at that time is decreased by one stage (step S 48 ). Then, the timer value t is reset to 0 (step S 49 ). Whereupon, the processing from step S 11 is resumed.
  • the operation information (set value) for the controlled object is increased or decreased one stage by one stage in accordance with the stop time, and the switch-operation information is output.
  • step S 50 If it is determined at step S 50 that the moved distance D of the palm from the reference position C 0 does not reach the maximum motion amount, the currently and immediately precedingly detected moved distances D and D′ are compared with each other, to thereby determine the direction of motion of the palm with forefinger-up (step S 51 ). If the direction of motion is the increasing direction, whether or not condition 1 is satisfied is determined. Specifically, a determination is made as to whether the currently detected moved distance D from the reference position C 0 is larger than a detection distance [h*(n+1)] defined as integral multiple of a predetermined unit distance h and at the same time the immediately precedingly detected moved distance D′ is equal to or less than the just-mentioned detection distance (step S 52 ).
  • the n is a parameter used for setting the detection distance. If the palm currently moves in the increasing direction beyond the detection distance [h*(n+1)] used for determination and if the preceding moved distance D′ is equal to or less than the detection distance, that is, if the palm moves by a predetermined distance or more in the increasing direction from the preceding cycle to the present cycle so that condition 1 of D >h*(n+1) and D′ I h*(n+1) is fulfilled, the parameter n is incremented to set the detection distance for the next determination (step S 54 ), whereby the operation information (setting) for the controlled object is increased by one stage (step S 56 ).
  • step S 51 If the direction of motion determined at step S 51 is the decreasing direction, whether or not condition 2 is satisfied is determined. Specifically, a determination is made as to whether the currently detected moved distance D from the reference position C 0 is smaller than a detection distance [h*(n ⁇ 1)] defined as integral multiple of the predetermined unit distance h and the immediately precedingly detected moved distance D′ is equal to or larger than the detection distance (step S 53 ).
  • the parameter n is decremented to set the detection distance for the next determination (step S 55 ), whereby the operation information (setting) for the controlled object is decreased by one stage (step S 57 ).
  • the switch-operation amount can continuously be changed in accordance with a stop time of the palm at a stop position.
  • the switch-operation amount can be set immediately.
  • the switch-operation amount can be set finely, where required.
  • switch information for various controlled objects can be input by simply forming a predetermined hand pattern and moving the hand and/or palm as exemplarily shown in FIG. 11 , without the need of touching the operating section 2 of audio equipment, air conditioner equipment, etc.
  • driver's hands fall outside the image pickup zone 3 a as shown by initial state P 1 .
  • the input image at that time only includes image components that will be removed as merely representing the background of the vehicle compartment, so that the hand pattern switch device does not operate.
  • a driver's hand coming off the steering wheel 1 and then formed into an L-shaped pattern (hand pattern 4 ) enters the image pickup zone 3 a as shown by operation state P 2 , it is determined that the hand pattern switch device is instructed to start operation. After outputting confirmation sound such as pip tone, the hand pattern switch device enters a standby state.
  • speech guidance may be notified such that “sound volume adjustment mode is set,” “temperature adjustment mode is set,” or “wind amount adjustment mode is set” each time a switch operation such as sound volume, temperature, wind amount is detected as mentioned above. More simply, the word such as “sound volume,” “temperature,” “wind amount” may be notified as speech message.
  • speech message may be notified such that “sound volume adjustment mode is set,” “temperature adjustment mode is set,” or “wind amount adjustment mode is set” each time a switch operation such as sound volume, temperature, wind amount is detected as mentioned above. More simply, the word such as “sound volume,” “temperature,” “wind amount” may be notified as speech message.
  • the finger-up pattern (hand pattern 2 ) is formed as shown in operation state P 4 .
  • the hand pattern 2 is recognized, and the operation amount setting mode is set.
  • the palm with the finger-up pattern (hand pattern 2 ) is moved left and right as shown in operation state P 5 a or P 5 b, whereby switch-operation amount information for the controlled object set as mentioned above is input.
  • the clenched-fist pattern (hand pattern 1 ) is formed as shown in operation state P 6 , whereby instruction to indicate the completion of operation is given to the hand pattern switch device.
  • switch-operation instructions based on the predetermined hand patterns and motions can easily and effectively be detected with reliability and without being affected by hand/finger motions and arm motions for a driving operation, and in accordance with detection results, switch-operation information can properly be provided to the desired vehicle-mounted equipment.
  • the driver's load in operating the hand pattern switch device is reduced or eliminated since the region (image pickup zone 3 a ), in which an image of hand/fingers to give switch-operation instructions is picked up, is located at a position laterally to the steering wheel 1 and is set such that the driver can naturally extend the arm to this region without changing a driving posture.
  • the hand patter switch device can achieve practical advantages such as for example that the driver can easily input instructions or switch-operation information through the use of the hand pattern switch device, with the feeling of directly operating the operating section 2 of audio equipment, etc.
  • the central axis extending toward the fingertip is determined from a binarized image of the palm, finger widths of the hand on scanning lines extending approximately perpendicular to the central axis are sequentially determined, and a point of intersection of the central axis and a scanning line which is maximum in finger width is determined as palm center.
  • scanning lines for finger-detection are set in the direction perpendicular to the extending direction of a finger to be detected, and an image component for which a width equal to or larger than a predetermined width is detected on each scanning line is determined as a finger width. Whether or not the finger is extended from the palm is then determined based on the number of scanning lines on which the predetermined or more finger width is detected. Therefore, a finger pattern can easily and reliably be recognized (detected).
  • by determining extended states of the forefinger and thumb finger from the palm by positively utilizing a difference between directions in which these fingers can be extended, individual features of the hand patterns can be grasped with reliability in the hand pattern recognition. This makes it possible to surely recognize the hand pattern and the palm motion (positional change) even by means of simplified, less costly image processing, resulting in advantages that operations can be simplified, and the like.
  • the present invention is not limited to the foregoing embodiment.
  • explanations have been given under the assumption that this invention is applied to a right-steering-wheel vehicle, but it is of course applicable to a left-steering-wheel vehicle.
  • This invention is also applicable to an ordinary passenger car other than a large-sized car such as truck.
  • the controlled object expansions can be made to operation of wiper on/off control, adjustment of interval of wiper operation, side mirror open/close control, etc., as exemplarily shown in FIG. 12 .
  • the controlled objects are systematically classified in the form of tree structure in advance, so that a desired one of these controlled objects may be selected stepwise.
  • the controlled objects are broadly classified into driving equipment system” and “comfortable equipment system.”
  • the driving equipment system it is divided into medium classes as “direction indicator,” “wiper,” “light” and “mirror.” Functions of each of the controlled objects belonging to the same medium class are further divided into narrow classes.
  • the comfortable equipment system is divided into medium classes as “audio” and “air conditioner.”
  • the audio it is classified into types of equipment such as “radio,” “CD,” “tape,” and “MD.” Further, each type of equipment is classified into functions such as operation mode and sound volume. From the viewpoint of easy operation, in actual, the setting is made such that only the necessity minimum controlled objects are selectable because the selection operation becomes complicated if the setting is made to include a large number of classes.
  • a finger of a hand is detected through the use of first scanning lines set to extend perpendicular to the central axis extending from an arm portion to a fingertip in a binarized image and/or second scanning lines set to extend along the central axis. It is therefore possible to reliably detect whether or not a finger of the hand is extended, without being affected by image components corresponding to a long sleeve shirt, a wrist watch, etc., that are sometimes worn by the operator. Thus, the hand pattern can be determined with accuracy.
  • the central axis is determined as passing through the palm center, and a finger width equal to or larger than ⁇ fraction (1/7) ⁇ to 1 ⁇ 4 of an image width detected on a scanning line passing through the palm center is detected, to thereby make a determination whether or not a forefinger or a thumb finger is extended.
  • the finger pattern can easily and reliably be recognized (detected).
  • the hand pattern and the hand motion can be recognized with reliability, while reducing the load of the recognition processing, in which the hand pattern and/or the palm (fingertip) motion is detected and switch-operation information is given to various controlled objects.

Abstract

A hand pattern switch device capable of easily and reliably detecting a hand pattern or a palm motion of an operator. From a picked-up image of a distal arm, an axis passing through the centroid of a hand is determined as a central axis passing through the center of arm. At least either first scanning lines perpendicular to the central axis or second scanning lines extending along the central axis are set between a fingertip and a palm center. While changing the scanning line to be examined from the fingertip side toward the palm center, a determination is made to determine how many number of scanning lines for each of which a finger width equal to or larger than a predetermined width is detected are present, thereby making a detection of whether or not the finger is extended from the palm.

Description

    CROSS-REFERENCE TO A RELATED APPLICATION
  • This nonprovisional application claims priority under 35 U.S.C. §119(a) on Patent Application No. 2003-291380 filed in Japan on Aug. 11, 2003, the entire contents of which are hereby incorporated by reference.
  • BACKGROUND OF THE INVENTION
  • 1. Technical Field
  • The present invention relates to a hand pattern switch device suitable for a driver to easily operate vehicle-mounted equipment such as air conditioner equipment and audio equipment and ancillary vehicle equipment such as side mirrors, without his/her driving being affected and without the need of touching an operation panel of the vehicle-mounted equipment.
  • 2. Related Art
  • There has been proposed a technical art to operate vehicle-mounted equipment such as air conditioner equipment and audio equipment without touching an operation panel of the vehicle-mounted equipment, in which an image of a body part (for example, a left hand) of a driver is picked up by a camera and subject to pattern recognition to obtain information that is used to operate the vehicle-mounted equipment (refer for example to JP-A-11-134090). Another technical art has also been proposed that detects a driver's gesture such as hand pattern and hand motion, from which information used to operate vehicle-mounted equipment is acquired (refer for example to JP-A-2001-216069).
  • This kind of art, realized by the pattern recognition to recognize a hand pattern from a picked-up image of a hand or realized by the motion detection to detect a hand motion by tracing a positional change of a recognized hand, is called as a hand pattern switch device in the present specification for the sake of convenience.
  • In the case of using the hand pattern switch device in order to operate vehicle-mounted equipment, the pattern or motion of the driver's (operator's) hand must be detected reliably and accurately. To this end, it is necessary to accurately recognize which part of the picked-up image corresponds to the driver's (operator's) hand. However, the driver (operator) sometimes wears a long sleeve shirt, a wrist watch, or the like. In that case, a wrist portion in the input image is detected to be extraordinary large, or detected to be disconnected due to the presence of an image component corresponding to the wrist watch or the like. This results in the fear of a portion corresponding to the driver's palm or the back of his/her hand (hereinafter collectively referred to as palm) being unable to be detected with reliability, despite that such portion is to be detected for pattern recognition. In addition, the prior art poses a further problem of the processing load being increased, since it generally uses a complicated image processing technique, such as region segmentation, or a matching technique in which a predetermined standard hand pattern is referred to.
  • SUMMARY OF THE INVENTION
  • The object of this invention is to provide a hand pattern switch device capable of easily and reliably detecting a hand pattern or a hand motion of a driver (operator) observed when the driver operates various vehicle-mounted equipment and ancillary vehicle equipment, thereby properly inputting information used for operations of these equipment.
  • According to this invention, there is provided a hand pattern switch device which has image pickup means for picking up an image of a distal arm that is within a predetermined image pickup zone and in which a hand pattern and/or a motion of a finger of a hand is detected from the image picked up by the image pickup means to obtain predetermined switch operation information. The hand pattern switch device comprises first image processing means for determining a central axis passing through a center of the arm based on the picked-up image, scanning line setting means for setting at least either a first scanning line extending perpendicular to the central axis or a second scanning line extending along the central axis, and determination means for determining whether or not any finger of the hand is extended based on the at least either the first or second scanning line set by the scanning line setting means.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The present invention will become more fully understood from the detailed description given herein below and the accompanying drawings which are given by way of illustration only, and thus, are not limitative of the present invention, and wherein:
  • FIG. 1 is a view showing the outline of structure of a hand pattern switch device according to an embodiment of this invention;
  • FIG. 2 is a view showing a hand/finger image pickup zone in the hand pattern switch device shown in FIG. 1;
  • FIG. 3 is a flowchart showing an example of processing procedures for recognition of hand pattern and palm center;
  • FIG. 4 is a conceptual view for explaining the processing for recognition of hand pattern and palm center shown in FIG. 3;
  • FIG. 5A is a view for explaining drawbacks of conventional typical processing for hand pattern recognition;
  • FIG. 5B is a view similar to FIG. 5A;
  • FIG. 6A is a view showing hand pattern 1 used in the embodiment of this invention;
  • FIG. 6B is a view showing hand pattern 2;
  • FIG. 6C is a view showing hand pattern 3;
  • FIG. 6D is a view showing hand pattern 4;
  • FIG. 7 is a flowchart showing an example of processing procedures for hand pattern recognition performed by an instructed-operation recognizing section in the hand pattern switch device shown in FIG. 1;
  • FIG. 8 is a flowchart showing an example of processing procedures for operation amount detection;
  • FIG. 9 is a flowchart showing an example of processing procedures for operation amount detection in a time mode;
  • FIG. 10 is a flowchart showing an example of processing procedures for operation amount detection in a distance/time mode;
  • FIG. 11 is a view showing input modes of inputting switch-operation information to the hand pattern switch device with use of hand/fingers; and
  • FIG. 12 is a view showing an example of systemized selection of controlled objects.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • In the following, a hand pattern switch device according to an embodiment of this invention will be explained with reference to the drawings.
  • FIG. 1 is a view of a general construction of essential part of the hand pattern switch device according to the present embodiment, showing a state around a driver's seat of a vehicle and functions of the hand pattern switch device realized for example by a microcomputer (ECU) and the like. At the front of the driver's seat, a steering wheel 1 adapted to be steered by a driver, a combination switch (not shown), etc. are provided, whereas an operating section 2 for audio equipment, air conditioner equipment, etc. is provided on a console panel. At a ceiling located above the driver's seat, a video camera 3 is disposed for picking up an image of a hand of the driver who extends his/her arm to an image pickup zone located laterally to the steering wheel 1. The camera 3 is comprised of a small-sized CCD camera or the like. The camera 3 may be the one which obtains a visible light image under predetermined illumination (daytime). Of course, a so-called infrared camera which emits near-infrared light to the pickup zone to obtain an infrared image may be used, when the illumination for the pickup zone is insufficient, as in nighttime. To operate the hand pattern switch device, the hand pattern is changed by selectively flexing desired one or ones of the fingers, with the palm positioned horizontally in the pickup zone, and the palm position is displaced (moved) back and forth and left and right. Despite that an image of the back of hand is actually picked up by the camera 3, the term “palm” is used in the description here that represents not only the palm but also the back of hand whose image is to be picked up.
  • Basically, the hand pattern switch device performs the processing to recognize a driver's hand pattern or a hand motion on the basis of an image picked up by and input from the camera 3, and based on results of the recognition, acquires predetermined corresponding switch-operation information. Thus, the hand pattern switch device serves, instead of the operating section 2, to provide switch-operation information to the audio equipment, air conditioner equipment, etc. More specifically, the hand pattern switch device comprises a binarization processing section 11 for binarizing an input image picked up by the camera 3 so that background image components are removed to extract image components corresponding to the distal arm, mainly the palm and fingers of the hand, from the picked-up image; a centroid detecting section 12 for determining a centroid position of the hand based on the image of the palm and fingers of the hand extracted by the binarization processing; and a pattern recognition section 13 for recognizing a hand/finger pattern.
  • The hand pattern switch device further comprises an instructed-operation recognizing section 14 for recognizing a switch operation given by the driver by the hand pattern or hand motion, based on results of recognition performed by the pattern recognition section 13 and the centroid position of the hand detected by the centroid detecting section 12. This instructed-operation recognizing section 14 generally comprises a function determination section 16 for determining (identifying) a type of operation intended by the hand pattern recognized as mentioned above, referring to a relation between hand patterns registered beforehand in a memory 15 and their functions, a displacement detecting section 17 for tracing a motion of centroid position of palm with a particular finger pattern or a motion of fingertip to thereby detect a displacement thereof from its reference position, and a timer 18 for monitoring the palm motion or fingertip motion in terms of elapsed time during the palm or fingertip is moved. On the basis of results of determination and monitoring, the instructed-operation recognizing section 14 determines predetermined switch-operation information specified by the driver's hand pattern and palm motion, and outputs this switch-operation information by way of example to the audio equipment, air conditioner equipment, or the like.
  • The instructed-operation recognizing section 14 is further provided with a guidance section 19 that provides a predetermined guidance to the driver according to results of the aforementioned determination, etc. The driver is notified of the guidance form a speaker 20 in the form of a speech message that specifies for example the audio equipment or air conditioner equipment (controlled object equipment), or volume/channel setting, wind volume/temperature, or the like (controlled object function), or in the form of confirmation sound such as pip tone or beep tone that identifies a switch operation (operation amount) having been made. As for concrete operation modes of the instructed-operation recognizing section 14, i.e., control of output of switch-operation information in respect of controlled objects such as audio equipment, air conditioner equipment, and the like, explanations will be given later.
  • As shown in FIG. 2, the image pickup zone 3 a of the camera 3 located laterally to the steering wheel 1 is at least 50 mm, preferably about 100 mm, apart from the outer periphery of the steering wheel 1. In particular, the image pickup zone is at a position to which the driver can extend the arm without changing a driving posture, while resting the arm on an arm rest 5 that is provided laterally to the driver's seat and which is located away from the operating section 2 for audio equipment, etc., so that the hand extended to the image pickup zone does not touch the operating section 2. The image pickup zone 3 a is rectangle in shape and has a size of about 600 mm in a fingertip direction of and about 350 mm in a width direction of the driver's hand extended laterally to the steering wheel 1.
  • Specifically, the image pickup zone 3 a is a zone that is set such that an image of the driver's hand is not picked up when the driver holds the steering wheel 1 or operates the combination switch (not shown) provided at the steering column shaft and such that the driver can move his/her hand into the zone without largely moving the arm. Thus, a hand motion for a driving operation or a hand/finger motion for a direct operation of the operating section 2 of the audio equipment, etc. is prevented from being erroneously detected as a motion for providing switch-operation information.
  • If a gearshift lever (not shown) is located in the image pickup zone 3 a set as described above, a pressure-sensitive sensor for example may be provided in the gearshift lever to make a detection as to whether the gearshift lever is grasped by the driver. The provision of such sensor makes it possible to easily determine which of the gearshift lever or the hand pattern switch device is operated by the driver's hand extended to the image pickup zone 3a, whereby a driving operation is prevented from being erroneously detected as a switch operation. Alternatively, a height of driver's hand (i.e., distance from the camera 3) may be detected by using a stereoscopic camera serving as the camera 3, to determine whether the driver's hand extended to the image pickup zone 3 a operates the gearshift lever or is present in a space above the gearshift lever.
  • The setting of the image pickup zone 3 a is made based on a range (displacement width) of arm/hand motion to which the driver can naturally extend the arm without changing a driving posture while resting the arm (elbow) on the arm rest 5 and to which the driver can comfortably and naturally move the arm/hand when making the imaginary switch operation. In particular, by taking into account of a typical length from the wrist to the fingertip being about 200 mm and a typical hand width being about 120 mm, the image pickup zone 3 a is determined to be a rectangle in shape and to have a 600 mm length and a 350 mm width, as mentioned above.
  • By setting the image pickup zone 3 a as described above, the driver's hand coming off the steering wheel 1 and then naturally moved without a sense of incompatibility can be captured without fail and without a hand/arm motion for a driving operation being erroneously detected. It is also possible to reliably grasp a change in hand position or a hand motion for switch operation in the image pickup zone 3a, so that the hand pattern recognition and the detection of an amount of hand motion (deviation) can easily be made with relatively simplified image processing.
  • For the driver, he/she can perform a desired switch operation by simply moving the hand after forming a corresponding one of predetermined hand patterns, while extending the arm laterally to the steering wheel 1 without changing a driving posture and without directly touching the operating section 2 for audio equipment, etc. This reduces a load of the driver performing the switch operation. In addition, since a hand motion and/or an arm motion for a driving operation cannot erroneously be detected as instruction for switch operation, there are advantages that the driver can concentrate on driving without paying attention to the hand pattern switch device, and can, where required, easily give instruction for switch operation by simply moving his/her hand (palm) to the image pickup zone 3 a.
  • The following is the description of the hand recognition processing, which is one of the features of this invention and which is performed based on the binary image.
  • Referring to processing procedures shown in FIG. 3, at start of the recognition processing, an input image of the pickup zone 3 a picked up by the camera 3 is subject to binarization processing using a predetermined threshold value, whereby the background of the image is discriminated as black from other portions of the image, as white, corresponding to the arm and palm (step S1). By means of the binarization processing, a binarized image of the pickup zone 3 a is obtained as exemplarily shown in FIG. 4. Next, the centroid G of the white region of the binarized image, corresponding to image components of the arm and palm, is determined. Then, a central axis B is determined from the centroid G and a longitudinal moment that is determined by analyzing the direction in which white elements are present passing through the centroid G (step S2). By determining the central axis B passing through the centroid G of the white region in this manner, the position of the central axis B can accurately be set.
  • Subsequently, a plurality of first scanning lines Si extending at right angles with respect to the central axis B are set at equal intervals between an upper side or fingertip side of the binarized image and the centroid G (step S3). Then, widths W of the white image on the respective scanning lines S1 are determined in sequence from the upper side of the binarized image, to thereby detect the scanning line Si which is maximum in the white image width W (step S4). At this time, it is preferable that only the white image, which is continuous on the both sides of the central axis B, be selected as the detection object in detecting the white image width W on each scanning line S1. As for the detection of the scanning line S1 that is maximum in white image width W, a determination is made whether or not the white image width W detected sequentially from the upper side of the image is larger than that detected on the immediately preceding scanning line S1, and when a peak appears in the detected widths, the associated scanning line S1 is detected as the one having the maximum white image width.
  • Next, a point of intersection of the thus detected scanning line S1 having the maximum white image width W and the central axis B is detected as a palm center position C (step S5). By doing the above processing to determine the palm center position C, it is possible to accurately determine the palm center position C since the width on the wrist side is generally narrower than the width of a central portion of palm, even when the binarized image is disconnected at a portion (wrist portion) between the arm and palm as shown in FIG. 5A due to the presence of a watch or a wrist band attached to the wrist or even when the size (width) of arm, especially the size of the wrist portion, is detected to be extraordinarily large as shown in FIG. 5B for the reason that the wrist is hidden by a long-sleeve cloth which the driver wears.
  • Subsequently, on the basis of the palm center position C and the palm width W determined as mentioned above, among the scanning lines S1 located between the upper side of the binarized image and the center position C, a number of scanning lines S1 for each of which a width equal to or larger than the predetermined width w has been detected is determined (step S3). More specifically, an examination is made whether the width detected on each scanning line S1 is equal to or larger than the predetermined width w that is set to a value of {fraction (1/7)} to ¼ of the maximum width W at the palm center position C. Then, a determination is made whether the number of scanning lines S1, for which a white image whose width is equal to or larger than the predetermined width w has been detected, is equal to or larger than a value that is set beforehand in accordance with the spacing between the adjacent scanning lines S1. If there is a white image region extending by a predetermined length or more in the direction of the central axis B, the white image region is detected as a finger (forefinger, for instance) extended from the palm. For example, if the length measured from the palm center position C, which is detected in terms of the number of the scanning lines S1 as mentioned above, is 10 cm or more, it is detected that the forefinger is extended from the palm.
  • The thumb finger can be detected in a similar manner. Since the detection object is the left hand and the thumb finger is extended in a direction different from the direction in which the forefinger is extended, second scanning lines S2 used to detect the thumb finger are set on the right side of the palm center position C so as to be inclined at an angle of about 10 degrees relative to the central axis B (step S7). This setting is based on the fact that the thumb finger extends slightly obliquely with respect to the central axis B when it is extended to be opened at the maximum. By setting the second scanning lines S2 such that the thumb finger opened at the maximum extends substantially perpendicular to the scanning lines S2, a reliable detection of the thumb finger can be achieved.
  • Next, a determination is made as to how many number of scanning lines S2 are present for each of which a width equal to the predetermined width w or more has been detected, among the scanning lines S2 located between the right side of the binarized image and the center position C (step S8). At this time, as in the case of the detection of the forefinger, a determination is made whether the number of scanning lines S2 for which a white image whose width is equal to or larger than the width w set to be {fraction (1/7)} to ¼ of the maximum width W at the palm center position C has been detected is equal to or larger than a predetermined value. When the predetermined number or more of scanning lines S2 is detected, it is detected that the thumb finger is extended from the palm to the right side.
  • By means of the above-mentioned recognition processing, information can be determined that represents the hand pattern in the pickup zone 3 a and the palm center position C. Then, determinations are made whether or not the forefinger is detected and whether or not the thumb finger is detected, thereby determining which pattern is formed among the following: a clenched-fist pattern (hand pattern 1) in which all the fingers are bent into the palm; a finger-up pattern (hand pattern 2) in which only the forefinger is extended; an acceptance (OK) pattern (hand pattern 3) in which only the thumb finger is extended horizontally; and an L-shaped pattern (hand pattern 4) in which the forefinger and the thumb finger are extended. These patterns are shown in FIGS. 6A-6D, respectively.
  • In this embodiment, the L-shaped pattern (hand pattern 4) is used to instruct the start of operation to the hand pattern switch device. The hand pattern 3 is used in combination of the clenched-fist pattern (hand pattern 1) to express image of depressing a push button, by changing the hand pattern by putting the thumb finger in and out (flexing). The hand pattern 3 is used to input information for selection of controlled objects. The finger-up pattern (hand pattern 2) is to express image of an indicating needle of an analog meter, and is used to instruct an amount of operation to the controlled object by changing the-position of fingertip (or palm). The clenched-fist pattern (hand pattern 1) is also used to instruct completion of operation of the hand pattern switch device.
  • In the instructed-operation recognizing section 14, the hand pattern and the change in palm position recognized as mentioned above are subject to the recognition processing that is performed in accordance with procedures exemplarily shown in FIG. 7, whereby switch operations by means of the driver's (switch operator's) hand are interpreted and switch-operation information is output to the controlled objects.
  • More specifically, the instructed-operation recognizing section 14 inputs data representing results of recognition in the pattern recognition section 13 and information of palm center position C (step S11). Then, a flag F used to discriminate whether a switch operation is instructed is checked (step S12). If the flag F is not set (F=0), whether or not the hand pattern is the L-shaped pattern (hand pattern 4) instructing the start of operation is determined (step S13). When the hand pattern 4 is detected, the flag F is set (step S14), whereby the input of switch-operation information is started. If the hand pattern 4 is not detected, the aforementioned processing is repeatedly performed until the hand pattern 4 is detected.
  • If the flag F is set (F=1), it is determined that the input processing for the switch operation is already started (step S12). Thus, a determination is made in respect of a flag M that is used to discriminate whether a function selection mode for specifying a controlled object is set (step S15). If the flag M is not set (M=0), whether the hand pattern is the hand pattern 3 used to select a controlled object is determined (step S16). If the hand pattern 3 is determined, the flag M is set to 1 (M=1), thereby setting the controlled object selection mode (step S17). If the hand pattern 3 is not determined, the later-mentioned processing to input a switch operation amount is executed, determining that a controlled object is already specified.
  • In case that the hand pattern 3 is detected and the controlled object selection mode is set, whether or not the hand pattern is the clenched-fist pattern (hand pattern 1) is then determined (step S18). If the hand pattern 1 is detected, whether the immediately precedingly detected hand pattern was the hand pattern 3 is determined (step S19). When the change from the hand pattern 3 to the hand pattern 1 is detected, the controlled object is changed, considering that the change in hand pattern is the instruction to make changeover of the controlled objects (step S20). As for the change in controlled object, there may be a case where controlled objects are three, one for sound volume in audio equipment, one for temperature in air conditioner, and one for wind amount in air conditioner. In such a case, these controlled objects may be cyclically changed over as mentioned later.
  • Meanwhile, even when the hand pattern 1 is detected in the present cycle, if the hand pattern in the immediately preceding cycle was not the hand pattern 3 (step S19), the processing from step S11 is resumed, considering that the changeover from the hand pattern 3 to the hand pattern 1 is not performed in the controlled object selection mode. If the hand pattern 1 is not detected at step S18, whether or not the detected hand pattern is the hand pattern 3 is then determined (step S12). If the detected hand pattern is the hand pattern 3, the processing from step S11 is resumed, considering that the hand pattern 3 remains unchanged in the controlled object selection mode. If the detected hand pattern is neither the hand pattern 1 nor the hand pattern 3 (steps S18 and S21), the flag M is reset to 0 (M=0), the controlled object selection mode set as mentioned above is released.
  • If the hand pattern 3 is not detected in a state where the controlled object selection mode is not set (step S16), or if the controlled object selection mode is released (step S22), whether or not the hand pattern is the finger-up pattern (hand pattern 2) with only the forefinger extended is determined (step S23). When the hand pattern 2 is detected, the below-mentioned processing to detect the switch operation amount is carried out (step S24). If the hand pattern 2 is not detected at step S23, whether or not the hand pattern is the clenched-fist pattern (hand pattern 1) is determined (step S25). In the case of the hand pattern 1, a timer t is counted up (step S26), and whether or not a predetermined time T has elapsed is determined referring to the counted-up timer t (step S27). If the hand pattern 1 is maintained for the predetermined time T or more, the flags F and M are reset to 0 (F=0, M=0), and the aforementioned series of processing is completed, considering that the completion of switch operation is instructed (step S28). If the hand pattern is neither the hand pattern 2 nor the hand pattern 1 (steps S23 and S25), the processing from step S11 is resumed to await for the next instruction being input. If the hand pattern 1 is not maintained for the predetermined time T or more, that is, if the hand pattern is changed to the hand pattern 2 again within the predetermined time T (step S27), the processing from step S11 is resumed, making it possible to perform reoperation.
  • The following is a concrete explanation on the processing to detect the switch operation amount in response to the finger-up pattern (hand pattern 2). Such processing is generally performed in accordance with processing procedures shown in FIG. 8. At start of the processing to detect the switch operation amount, whether or not a flag K used to identify whether the operation amount setting mode is set is determined (step S31). If the operation amount setting mode is not set (K=0), the palm center position C determined as mentioned above is set as a reference position C0 for the operation amount detection (step S32). Next, the flag K is set to 1 (K=1) to thereby set the operation amount setting mode (S33), and a timer value t used in the operation amount setting mode is reset to 0 (step S34).
  • As for data subsequently input, it is determined that the flag K is set (step S31), and therefore, the distance of deviation between the palm center position C determined at that time and the reference position C0, i.e., a moved distance D from the reference position C0, is determined (step S35). In order to calculate the moved distance D, it is enough to determine a distance between picture elements in the input image. In accordance with the moved distance thus determined and predetermined modes for the detection of operation amount (S36), processing to detect the operation amount is selectively carried out in a time mode (step S37) or in a distance/time mode (step S38).
  • The time mode is a mode in which switch-operation information is output in accordance with a stop time for which the hand displaced from the reference position C0 is kept stopped, and is suitable for example for adjustment of sound volume in audio equipment and for adjustment of temperature in air conditioner. The distance/time mode is a mode in which the switch-operation information determined according to an amount of hand motion is output when the hand is moved slightly, whereas the information determined according to a stop time of the hand at a moved position is output when the hand has been moved by a predetermined distance or more to the moved position. The distance/time mode is suitable for example for controlled objects that are subject to a fine adjustment after being roughly adjusted.
  • In this embodiment, the instruction to input the switch operation amount by means of the finger-up pattern (hand pattern 2) is carried out by moving the palm to the right and left around the arm on the arm rest 5 or by moving the hand right and left around the wrist as fulcrum. Such palm/hand motion to the right and left is performed within a range not falling outside of the pickup zone 3 a, for instance, within an angular range of about ±45 degrees. To be noted, in this embodiment, the amount of palm motion with the hand pattern 2 is detected n steps.
  • Referring to FIG. 9 exemplarily showing the processing procedures for the detection of operation amount in the time mode, a determination is first made to determine whether or not a motion amount D measured from the reference palm position C0 has exceeded a preset value (threshold value) H or −H used for determination of maximum motion amount (step S40). If the moved distance D does not reach the preset value (threshold value) H or −H, a timer value t is set to 0 (step S41). Whereupon the processing starting from step S11 is resumed.
  • If it is determined at step S40 that the moved distance D exceeds the preset value (threshold value) H, the timer value t is counted up (step S42). When the counted-up timer value t reaches a reference time T (step S43), the setting (switch-operation information) at that time is increased by one stage (step S44). Then, the timer value t is reset to 0 (step S45), and the processing from step S11 is resumed.
  • If it is determined at step S40 that the moved distance D exceeds the threshold value −H in the opposite direction, the timer value t is counted up (step S46). When it is determined that the counted-up timer value t reaches the reference time T (step S47), the setting (switch-operation information) at that time is decreased by one stage (step S48). Then, the timer value t is reset to 0 (step S49). Whereupon, the processing from step S11 is resumed. By means of such a series of processing, when the palm with forefinger-up has moved left and right by the predetermined distance and stops there, the operation information (set value) for the controlled object is increased or decreased one stage by one stage in accordance with the stop time, and the switch-operation information is output.
  • Referring to FIG. 10 exemplarily showing the processing procedures for the detection of operation amount in the distance/time mode, a determination is first made to determine whether or not a motion amount D measured from the reference palm position C0 has exceeded a preset value (threshold value) H or −H used for determination of maximum motion amount (step S50). If the determination threshold value H or −H is exceeded, the operation information (set value) for the controlled object is variably set in accordance with the stop time of the palm with forefinger-up at the maximum moved position, as shown in steps S42 a-S45 a and steps 46 a-49 a, as in the case of the time mode.
  • If it is determined at step S50 that the moved distance D of the palm from the reference position C0 does not reach the maximum motion amount, the currently and immediately precedingly detected moved distances D and D′ are compared with each other, to thereby determine the direction of motion of the palm with forefinger-up (step S51). If the direction of motion is the increasing direction, whether or not condition 1 is satisfied is determined. Specifically, a determination is made as to whether the currently detected moved distance D from the reference position C0 is larger than a detection distance [h*(n+1)] defined as integral multiple of a predetermined unit distance h and at the same time the immediately precedingly detected moved distance D′ is equal to or less than the just-mentioned detection distance (step S52). The n is a parameter used for setting the detection distance. If the palm currently moves in the increasing direction beyond the detection distance [h*(n+1)] used for determination and if the preceding moved distance D′ is equal to or less than the detection distance, that is, if the palm moves by a predetermined distance or more in the increasing direction from the preceding cycle to the present cycle so that condition 1 of D >h*(n+1) and D′ I h*(n+1) is fulfilled, the parameter n is incremented to set the detection distance for the next determination (step S54), whereby the operation information (setting) for the controlled object is increased by one stage (step S56).
  • If the direction of motion determined at step S51 is the decreasing direction, whether or not condition 2 is satisfied is determined. Specifically, a determination is made as to whether the currently detected moved distance D from the reference position C0 is smaller than a detection distance [h*(n−1)] defined as integral multiple of the predetermined unit distance h and the immediately precedingly detected moved distance D′ is equal to or larger than the detection distance (step S53). If the palm currently moves in the decreasing direction beyond the detection distance [h*(n−1)] used for determination and if the preceding moved distance D′ is equal to or larger than the detection distance, that is, if the palm moves by the predetermined distance or more in the decreasing direction from the preceding cycle to the present cycle so that condition 2 of D<h*(n−1) and D′≧h*(n−1) is fulfilled, the parameter n is decremented to set the detection distance for the next determination (step S55), whereby the operation information (setting) for the controlled object is decreased by one stage (step S57).
  • With the operation-amount detection processing according to the distance/time mode, it is possible to substantially continuously afford the switch-operation amount in accordance with the moved distance D of the palm with the forefinger up that is measured from the reference position C0. When the palm with the forefinger up is largely moved, the switch-operation amount can continuously be changed in accordance with a stop time of the palm at a stop position. Thus, in accordance with the palm or hand motion, the switch-operation amount can be set immediately. In addition, the switch-operation amount can be set finely, where required.
  • By using the instructed-operation recognizing section 14 which detects the hand/finger pattern and the palm motion to thereby identify the intention of the driver's (switch operator's) switch operation, switch information for various controlled objects can be input by simply forming a predetermined hand pattern and moving the hand and/or palm as exemplarily shown in FIG. 11, without the need of touching the operating section 2 of audio equipment, air conditioner equipment, etc.
  • Specifically, when the driver grasps the steering wheel 1 to drive the vehicle, driver's hands fall outside the image pickup zone 3 a as shown by initial state P1. The input image at that time only includes image components that will be removed as merely representing the background of the vehicle compartment, so that the hand pattern switch device does not operate. On the other hand, when a driver's hand coming off the steering wheel 1 and then formed into an L-shaped pattern (hand pattern 4) enters the image pickup zone 3 a as shown by operation state P2, it is determined that the hand pattern switch device is instructed to start operation. After outputting confirmation sound such as pip tone, the hand pattern switch device enters a standby state.
  • Subsequently, when the pattern (hand pattern 3) in which only the thumb finger is extended horizontally is set as shown by operation state P3, such pattern is detected and the function changeover mode (controlled object selection mode) is set. At this time, in order to notify the driver of the function changeover mode being set, a sound message is sent or a music box is operated. During this time, when the thumb finger is bent into the palm to form the clenched-fist pattern (hand pattern 1), it is determined that a push button switch operation is instructed, and the controlled object is changed over. At the time of controlled object changeover, speech guidance (speech message) may be notified such that “sound volume adjustment mode is set,” “temperature adjustment mode is set,” or “wind amount adjustment mode is set” each time a switch operation such as sound volume, temperature, wind amount is detected as mentioned above. More simply, the word such as “sound volume,” “temperature,” “wind amount” may be notified as speech message. Such guidance makes it possible for the driver to recognize a state of switch operation without the need of visually confirming the switch operation, whereby the driver is enabled to concentrate on driving.
  • After the desired controlled object is set, the finger-up pattern (hand pattern 2) is formed as shown in operation state P4. In response to this, the hand pattern 2 is recognized, and the operation amount setting mode is set. By taking into account of the predetermined operation modes, the palm with the finger-up pattern (hand pattern 2) is moved left and right as shown in operation state P5 a or P5 b, whereby switch-operation amount information for the controlled object set as mentioned above is input. When the desired switch operation is completed, the clenched-fist pattern (hand pattern 1) is formed as shown in operation state P6, whereby instruction to indicate the completion of operation is given to the hand pattern switch device.
  • During the course of moving the palm with finger-up pattern (hand pattern 2) left and right to input the switch-operation amount information, if the pattern (hand pattern 3) in which only the thumb finger is horizontally extended is formed, the processing to input the switch-operation amount information is completed at that time. In this case, processing for controlled object selection/changeover and subsequent processing can be performed again. Therefore, even in a case where a plurality of controlled objects are sequentially operated, continuous and repeated operations for the controlled objects can be carried out, without the recognition processing being interrupted. This makes it easy to perform operation.
  • With the hand pattern switch device constructed as mentioned above, switch-operation instructions based on the predetermined hand patterns and motions can easily and effectively be detected with reliability and without being affected by hand/finger motions and arm motions for a driving operation, and in accordance with detection results, switch-operation information can properly be provided to the desired vehicle-mounted equipment. The driver's load in operating the hand pattern switch device is reduced or eliminated since the region (image pickup zone 3 a), in which an image of hand/fingers to give switch-operation instructions is picked up, is located at a position laterally to the steering wheel 1 and is set such that the driver can naturally extend the arm to this region without changing a driving posture. The hand patter switch device can achieve practical advantages such as for example that the driver can easily input instructions or switch-operation information through the use of the hand pattern switch device, with the feeling of directly operating the operating section 2 of audio equipment, etc.
  • In particular, according to the hand pattern switch device, the central axis extending toward the fingertip is determined from a binarized image of the palm, finger widths of the hand on scanning lines extending approximately perpendicular to the central axis are sequentially determined, and a point of intersection of the central axis and a scanning line which is maximum in finger width is determined as palm center. This makes it possible to detect a palm portion in the binarized image with reliability. Even when the operator wears a long sleeve shirt and/or a wrist watch or the like, the palm center and the palm portion can reliably be detected, without being affected by image components corresponding to the shirt, etc.
  • According to the hand pattern switch device, scanning lines for finger-detection are set in the direction perpendicular to the extending direction of a finger to be detected, and an image component for which a width equal to or larger than a predetermined width is detected on each scanning line is determined as a finger width. Whether or not the finger is extended from the palm is then determined based on the number of scanning lines on which the predetermined or more finger width is detected. Therefore, a finger pattern can easily and reliably be recognized (detected). In particular, by determining extended states of the forefinger and thumb finger from the palm by positively utilizing a difference between directions in which these fingers can be extended, individual features of the hand patterns can be grasped with reliability in the hand pattern recognition. This makes it possible to surely recognize the hand pattern and the palm motion (positional change) even by means of simplified, less costly image processing, resulting in advantages that operations can be simplified, and the like.
  • The present invention is not limited to the foregoing embodiment. In the embodiment, explanations have been given under the assumption that this invention is applied to a right-steering-wheel vehicle, but it is of course applicable to a left-steering-wheel vehicle. This invention is also applicable to an ordinary passenger car other than a large-sized car such as truck. As for the controlled object, expansions can be made to operation of wiper on/off control, adjustment of interval of wiper operation, side mirror open/close control, etc., as exemplarily shown in FIG. 12. In this case, the controlled objects are systematically classified in the form of tree structure in advance, so that a desired one of these controlled objects may be selected stepwise.
  • Specifically, the controlled objects are broadly classified into driving equipment system” and “comfortable equipment system.” As for the driving equipment system, it is divided into medium classes as “direction indicator,” “wiper,” “light” and “mirror.” Functions of each of the controlled objects belonging to the same medium class are further divided into narrow classes. Similarly, the comfortable equipment system is divided into medium classes as “audio” and “air conditioner.” As for the audio, it is classified into types of equipment such as “radio,” “CD,” “tape,” and “MD.” Further, each type of equipment is classified into functions such as operation mode and sound volume. From the viewpoint of easy operation, in actual, the setting is made such that only the necessity minimum controlled objects are selectable because the selection operation becomes complicated if the setting is made to include a large number of classes.
  • Of course, the hand patterns used for information input are not limited to those described by way of example in the above. In other respects, this invention may be modified variously, without departing from the scope of invention.
  • According to the thus constructed hand pattern switch device, a finger of a hand is detected through the use of first scanning lines set to extend perpendicular to the central axis extending from an arm portion to a fingertip in a binarized image and/or second scanning lines set to extend along the central axis. It is therefore possible to reliably detect whether or not a finger of the hand is extended, without being affected by image components corresponding to a long sleeve shirt, a wrist watch, etc., that are sometimes worn by the operator. Thus, the hand pattern can be determined with accuracy. In particular, the central axis is determined as passing through the palm center, and a finger width equal to or larger than {fraction (1/7)} to ¼ of an image width detected on a scanning line passing through the palm center is detected, to thereby make a determination whether or not a forefinger or a thumb finger is extended. Thus, the finger pattern can easily and reliably be recognized (detected).
  • As a consequence, the hand pattern and the hand motion can be recognized with reliability, while reducing the load of the recognition processing, in which the hand pattern and/or the palm (fingertip) motion is detected and switch-operation information is given to various controlled objects.

Claims (15)

1. A hand pattern switch device having image pickup means for picking up an image of a distal arm that is within a predetermined image pickup zone, and detecting a hand pattern and/or a motion of a finger of a hand from the image picked up by the image pickup means to obtain predetermined switch operation information, comprising:
first image processing means for determining a central axis passing through a center of the arm based on the picked-up image;
scanning line setting means for setting at least either a first scanning line extending perpendicular to the central axis or a second scanning line extending along the central axis; and
determination means for determining whether or not any finger of the hand is extended based on the at least either the first or second scanning line set by the scanning line setting means.
2. The hand pattern switch device according to claim 1, wherein said determination means includes function selection means for detecting projection and withdrawal of a particular finger of the hand and for cyclically selecting and specifying one of the controlled objects each time when the projection or withdrawal of the particular finger is detected, and equipment operation means for providing a control amount for the controlled object specified by said function selection means in accordance with a predetermined hand pattern and/or a motion of the hand with such hand pattern.
3. The hand pattern switch device according to claim 2, wherein said equipment operation means varies the control amount to be provided to the controlled object in accordance with an amount of hand motion from a reference position to right and left and/or a stop time at a destination of motion.
4. The hand pattern switch device according to claim 2, wherein said function selection means and said equipment operation means are caused to stop selecting and specifying the one of the controlled objects and to stop providing the control amount to the controlled object when said determination means detects that a clenched-fist pattern in which all fingers are bent into a palm is maintained for a predetermined time or more, and determines that completion of operation is instructed.
5. The hand pattern switch device according to claim 2, wherein said predetermined hand pattern includes a clenched-fist pattern in which all fingers are bent into a palm, a finger-up pattern in which only a forefinger is extended, a pattern in which only a thumb finger is extended horizontally, and an L-shaped pattern in which the forefinger and the thumb finger are extended.
6. The hand pattern switch device according to claim 2, wherein said particular finger is a thumb finger, said predetermined hand pattern is a finger-up pattern in which only a forefinger is extended, and said equipment operation means detects a left and right motion of the finger-up pattern.
7. The hand pattern switch device according to claim 1, wherein said image pickup means is a camera installed at a ceiling above a driver's seat of a vehicle.
8. The hand pattern switch device according to claim 2, further comprising a guidance function to provide confirmation sound when one of the controlled objects is selected by said function selection means.
9. The hand pattern switch device according to claim 7, wherein said image pickup zone is located at a position to which a driver can extend his/her arm without changing a driving posture while resting the arm on an arm rest provided laterally to the driver's seat of the vehicle and without a driver's hand being touched to an operating section of a console provided in the vehicle.
10. The hand pattern switch device according to claim 1, wherein said image processing means includes a binarization processing means for subjecting the picked-up image to binarization processing, and centroid detecting means for determining a centroid of the picked-up image having been subject to the binarization processing, and
the central axis passing through the center of the arm in the image is determined as an axis passing through the centroid of the image.
11. The hand pattern switch device according to claim 10, wherein the first scanning line is set in plural numbers between a side of the distal arm and the centroid, and
second image processing means is provided which determines a scanning line for which a width of the distal arm having been subject to the binarization processing becomes maximum, and determines a point of intersection between the just-mentioned scanning line and the central axis as a palm center.
12. The hand pattern switch device according to claim 1, wherein said scanning line setting means sets both the first and second scanning lines, and
said determination means determines whether or not a forefinger is extended by using the first scanning line, and determines whether or not a thumb finger is extended by using the second scanning line.
13. The hand pattern switch device according to claim 12, wherein said scanning line setting means sets the second scanning line to be inclined at an angle of about 10 degrees with respect to the central axis.
14. The hand pattern switch device according to claim 13, wherein said determination means determines that a finger of the hand is extended from the palm when a number of scanning lines for each of which a finger width equal to or larger than a predetermined width is detected is equal to or larger than a predetermined number.
15. The hand pattern switch device according to claim 14, wherein said predetermined width is {fraction (1/7)} to ¼ or more of the width detected in the scanning line passing through the palm center.
US10/915,952 2003-08-11 2004-08-11 Hand pattern switch device Abandoned US20050063564A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2003-291380 2003-08-11
JP2003291380A JP3752246B2 (en) 2003-08-11 2003-08-11 Hand pattern switch device

Publications (1)

Publication Number Publication Date
US20050063564A1 true US20050063564A1 (en) 2005-03-24

Family

ID=34213317

Family Applications (1)

Application Number Title Priority Date Filing Date
US10/915,952 Abandoned US20050063564A1 (en) 2003-08-11 2004-08-11 Hand pattern switch device

Country Status (5)

Country Link
US (1) US20050063564A1 (en)
JP (1) JP3752246B2 (en)
KR (1) KR100575504B1 (en)
CN (1) CN1313905C (en)
DE (1) DE102004038965B4 (en)

Cited By (60)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040141634A1 (en) * 2002-10-25 2004-07-22 Keiichi Yamamoto Hand pattern switch device
US20050238202A1 (en) * 2004-02-26 2005-10-27 Mitsubishi Fuso Truck And Bus Corporation Hand pattern switching apparatus
WO2006109476A1 (en) * 2005-04-05 2006-10-19 Nissan Motor Co., Ltd. Command input system
WO2007107368A1 (en) * 2006-03-22 2007-09-27 Volkswagen Ag Interactive operating device and method for operating the interactive operating device
US20070283296A1 (en) * 2006-05-31 2007-12-06 Sony Ericsson Mobile Communications Ab Camera based control
WO2008053433A2 (en) * 2006-11-02 2008-05-08 Koninklijke Philips Electronics N.V. Hand gesture recognition by scanning line-wise hand images and by extracting contour extreme points
US20080130953A1 (en) * 2006-12-04 2008-06-05 Denso Corporation Operation estimating apparatus and program
US20080168403A1 (en) * 2007-01-06 2008-07-10 Appl Inc. Detecting and interpreting real-world and security gestures on touch and hover sensitive devices
US20080181456A1 (en) * 2006-12-27 2008-07-31 Takata Corporation Vehicular actuation system
US20080197996A1 (en) * 2007-01-30 2008-08-21 Toyota Jidosha Kabushiki Kaisha Operating device
US20080211832A1 (en) * 2005-09-05 2008-09-04 Toyota Jidosha Kabushiki Kaisha Vehicular Operating Apparatus
DE102007034273A1 (en) * 2007-07-19 2009-01-22 Volkswagen Ag Method for determining the position of a user's finger in a motor vehicle and position determining device
US20090102788A1 (en) * 2007-10-22 2009-04-23 Mitsubishi Electric Corporation Manipulation input device
US20100295782A1 (en) * 2009-05-21 2010-11-25 Yehuda Binder System and method for control based on face ore hand gesture detection
US20110060499A1 (en) * 2009-09-04 2011-03-10 Hyundai Motor Japan R&D Center, Inc. Operation system for vehicle
US20110063425A1 (en) * 2009-09-15 2011-03-17 Delphi Technologies, Inc. Vehicle Operator Control Input Assistance
US20110160933A1 (en) * 2009-12-25 2011-06-30 Honda Access Corp. Operation apparatus for on-board devices in automobile
US20110286676A1 (en) * 2010-05-20 2011-11-24 Edge3 Technologies Llc Systems and related methods for three dimensional gesture recognition in vehicles
US20120062736A1 (en) * 2010-09-13 2012-03-15 Xiong Huaixin Hand and indicating-point positioning method and hand gesture determining method used in human-computer interaction system
WO2012061256A1 (en) * 2010-11-01 2012-05-10 Robert Bosch Gmbh Robust video-based handwriting and gesture recognition for in-car applications
US20130050076A1 (en) * 2011-08-22 2013-02-28 Research & Business Foundation Sungkyunkwan University Method of recognizing a control command based on finger motion and mobile device using the same
CN103049111A (en) * 2012-12-20 2013-04-17 广州视睿电子科技有限公司 Stylus and touch coordinate calculation method
US20130146234A1 (en) * 2011-12-07 2013-06-13 Hyundai Motor Company Apparatus and method for blocking incident rays from entering an interior cabin of vehicle
US20130176232A1 (en) * 2009-12-12 2013-07-11 Christoph WAELLER Operating Method for a Display Device in a Vehicle
US20130321462A1 (en) * 2012-06-01 2013-12-05 Tom G. Salter Gesture based region identification for holograms
EP2703950A1 (en) * 2011-04-28 2014-03-05 Nec System Technologies, Ltd. Information processing device, information processing method, and recording medium
US20140079285A1 (en) * 2012-09-19 2014-03-20 Alps Electric Co., Ltd. Movement prediction device and input apparatus using the same
US20140153774A1 (en) * 2012-12-04 2014-06-05 Alpine Electronics, Inc. Gesture recognition apparatus, gesture recognition method, and recording medium
CN103885572A (en) * 2012-12-19 2014-06-25 原相科技股份有限公司 Switching device
WO2014095070A1 (en) * 2012-12-21 2014-06-26 Harman Becker Automotive Systems Gmbh Input device for a motor vehicle
WO2014108160A2 (en) * 2013-01-08 2014-07-17 Audi Ag User interface for the contactless selection of a device function
US20140240213A1 (en) * 2013-02-25 2014-08-28 Honda Motor Co., Ltd. Multi-resolution gesture recognition
US8896536B2 (en) 2008-06-10 2014-11-25 Mediatek Inc. Methods and systems for contactlessly controlling electronic devices according to signals from a digital camera and a sensor module
JP2014221636A (en) * 2008-06-18 2014-11-27 オブロング・インダストリーズ・インコーポレーテッド Gesture-based control system for vehicle interface
US20140347263A1 (en) * 2013-05-23 2014-11-27 Fastvdo Llc Motion-Assisted Visual Language For Human Computer Interfaces
US20140361989A1 (en) * 2012-01-10 2014-12-11 Daimler Ag Method and Device for Operating Functions in a Vehicle Using Gestures Performed in Three-Dimensional Space, and Related Computer Program Product
US20150026646A1 (en) * 2013-07-18 2015-01-22 Korea Electronics Technology Institute User interface apparatus based on hand gesture and method providing the same
US8942881B2 (en) 2012-04-02 2015-01-27 Google Inc. Gesture-based automotive controls
US20150089455A1 (en) * 2013-09-26 2015-03-26 Fujitsu Limited Gesture input method
US20150097798A1 (en) * 2011-11-16 2015-04-09 Flextronics Ap, Llc Gesture recognition for on-board display
EP2755115A4 (en) * 2011-09-07 2015-05-06 Nitto Denko Corp Method for detecting motion of input body and input device using same
US20150131857A1 (en) * 2013-11-08 2015-05-14 Hyundai Motor Company Vehicle recognizing user gesture and method for controlling the same
US20150143277A1 (en) * 2013-11-18 2015-05-21 Samsung Electronics Co., Ltd. Method for changing an input mode in an electronic device
GB2525840A (en) * 2014-02-18 2015-11-11 Jaguar Land Rover Ltd Autonomous driving system and method for same
US9373026B2 (en) 2012-12-28 2016-06-21 Hyundai Motor Company Method and system for recognizing hand gesture using selective illumination
US9436872B2 (en) 2014-02-24 2016-09-06 Hong Kong Applied Science and Technology Research Institute Company Limited System and method for detecting and tracking multiple parts of an object
US9440537B2 (en) 2012-01-09 2016-09-13 Daimler Ag Method and device for operating functions displayed on a display unit of a vehicle using gestures which are carried out in a three-dimensional space, and corresponding computer program product
US9639323B2 (en) * 2015-04-14 2017-05-02 Hon Hai Precision Industry Co., Ltd. Audio control system and control method thereof
US20170192629A1 (en) * 2014-07-04 2017-07-06 Clarion Co., Ltd. Information processing device
US9754167B1 (en) 2014-04-17 2017-09-05 Leap Motion, Inc. Safety for wearable virtual reality devices via object detection and tracking
US9868449B1 (en) 2014-05-30 2018-01-16 Leap Motion, Inc. Recognizing in-air gestures of a control object to control a vehicular control system
US20180059798A1 (en) * 2015-02-20 2018-03-01 Clarion Co., Ltd. Information processing device
CZ307236B6 (en) * 2016-10-03 2018-04-18 Ĺ KODA AUTO a.s. A device for interactive control of a display device and a method of controlling the device for interactive control of a display device
US10007329B1 (en) 2014-02-11 2018-06-26 Leap Motion, Inc. Drift cancelation for portable object detection and tracking
US10437347B2 (en) 2014-06-26 2019-10-08 Ultrahaptics IP Two Limited Integrated gestural interaction and multi-user collaboration in immersive virtual reality environments
US10642356B1 (en) * 2016-06-26 2020-05-05 Apple Inc. Wearable interactive user interface
US10895918B2 (en) * 2019-03-14 2021-01-19 Igt Gesture recognition system and method
US11307669B2 (en) * 2018-02-14 2022-04-19 Kyocera Corporation Electronic device, moving body, program and control method
US11386711B2 (en) 2014-08-15 2022-07-12 Ultrahaptics IP Two Limited Automotive and industrial motion sensory device
US11487388B2 (en) * 2017-10-09 2022-11-01 Huawei Technologies Co., Ltd. Anti-accidental touch detection method and apparatus, and terminal

Families Citing this family (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006285370A (en) * 2005-03-31 2006-10-19 Mitsubishi Fuso Truck & Bus Corp Hand pattern switch device and hand pattern operation method
US7697827B2 (en) 2005-10-17 2010-04-13 Konicek Jeffrey C User-friendlier interfaces for a camera
CN100428123C (en) * 2005-12-27 2008-10-22 联想(北京)有限公司 Information input device of digital equipment
DE102006009291A1 (en) 2006-03-01 2007-09-06 Audi Ag Method and device for operating at least two functional components of a system, in particular of a vehicle
CN100426200C (en) * 2006-10-13 2008-10-15 广东威创视讯科技股份有限公司 Intelligent code-inputting method based on interaction type input apparatus
DE102007045967A1 (en) * 2007-09-25 2009-04-02 Continental Automotive Gmbh Method and device for contactless input of characters
US9002119B2 (en) * 2008-06-04 2015-04-07 University Of Tsukuba, National University Corporation Device method and program for human hand posture estimation
KR100977443B1 (en) * 2008-10-01 2010-08-24 숭실대학교산학협력단 Apparatus and method for controlling home appliances based on gesture
JP2010258623A (en) * 2009-04-22 2010-11-11 Yamaha Corp Operation detecting apparatus
JP2010277197A (en) * 2009-05-26 2010-12-09 Sony Corp Information processing device, information processing method, and program
JP5416489B2 (en) * 2009-06-17 2014-02-12 日本電信電話株式会社 3D fingertip position detection method, 3D fingertip position detection device, and program
JP5521727B2 (en) 2010-04-19 2014-06-18 ソニー株式会社 Image processing system, image processing apparatus, image processing method, and program
JP5865615B2 (en) * 2011-06-30 2016-02-17 株式会社東芝 Electronic apparatus and control method
DE102011080592A1 (en) * 2011-08-08 2013-02-14 Siemens Aktiengesellschaft Device and method for controlling a rail vehicle
DE102012216181A1 (en) * 2012-09-12 2014-06-12 Bayerische Motoren Werke Aktiengesellschaft System for gesture-based adjustment of seat mounted in vehicle by user, has control unit that controls setting of vehicle seat associated with recognized gesture and gesture area
DE102012021220A1 (en) * 2012-10-27 2014-04-30 Volkswagen Aktiengesellschaft Operating arrangement for detection of gestures in motor vehicle, has gesture detection sensor for detecting gestures and for passing on gesture signals, and processing unit for processing gesture signals and for outputting result signals
JP5459385B2 (en) * 2012-12-26 2014-04-02 株式会社デンソー Image display apparatus and indicator image display method
DE102013001330A1 (en) 2013-01-26 2014-07-31 Audi Ag Method for operating air conveying fan of fan device of motor vehicle, involves determining predetermined gesture in such way that occupant abducts fingers of his hand before clenching his fist
DE102013010018B3 (en) * 2013-06-14 2014-12-04 Volkswagen Ag Motor vehicle with a compartment for storing an object and method for operating a motor vehicle
DE102013214326A1 (en) * 2013-07-23 2015-01-29 Robert Bosch Gmbh Method for operating an input device, input device
DE102013226682A1 (en) * 2013-12-19 2015-06-25 Zf Friedrichshafen Ag Wristband sensor and method of operating a wristband sensor
DE102014224618A1 (en) * 2014-12-02 2016-06-02 Robert Bosch Gmbh Method and device for operating an input device
DE102015201901B4 (en) 2015-02-04 2021-07-22 Volkswagen Aktiengesellschaft Determination of a position of a non-vehicle object in a vehicle
KR101724108B1 (en) * 2015-10-26 2017-04-06 재단법인대구경북과학기술원 Device control method by hand shape and gesture and control device thereby
JP6716897B2 (en) * 2015-11-30 2020-07-01 富士通株式会社 Operation detection method, operation detection device, and operation detection program
CN110333772B (en) * 2018-03-31 2023-05-05 广州卓腾科技有限公司 Gesture control method for controlling movement of object
DE102019204481A1 (en) * 2019-03-29 2020-10-01 Deere & Company System for recognizing an operating intention on a manually operated operating unit
JP7470069B2 (en) 2021-02-17 2024-04-17 株式会社日立製作所 Pointing object detection device, pointing object detection method, and pointing object detection system

Citations (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US1965944A (en) * 1933-03-13 1934-07-10 Dudley L Lea Truck construction
US1984030A (en) * 1932-09-22 1934-12-11 John B Nixon Means for serving cocktails and the like
US5203704A (en) * 1990-12-21 1993-04-20 Mccloud Seth R Method of communication using pointing vector gestures and mnemonic devices to assist in learning point vector gestures
US5454043A (en) * 1993-07-30 1995-09-26 Mitsubishi Electric Research Laboratories, Inc. Dynamic and static hand gesture recognition through low-level image analysis
US5594469A (en) * 1995-02-21 1997-01-14 Mitsubishi Electric Information Technology Center America Inc. Hand gesture machine control system
US5815147A (en) * 1996-06-07 1998-09-29 The Trustees Of The University Of Pennsylvania Virtual play environment for disabled children
US6002808A (en) * 1996-07-26 1999-12-14 Mitsubishi Electric Information Technology Center America, Inc. Hand gesture control system
US6072494A (en) * 1997-10-15 2000-06-06 Electric Planet, Inc. Method and apparatus for real-time gesture recognition
US6128003A (en) * 1996-12-20 2000-10-03 Hitachi, Ltd. Hand gesture recognition system and method
US6236736B1 (en) * 1997-02-07 2001-05-22 Ncr Corporation Method and apparatus for detecting movement patterns at a self-service checkout terminal
US6256400B1 (en) * 1998-09-28 2001-07-03 Matsushita Electric Industrial Co., Ltd. Method and device for segmenting hand gestures
US6359512B1 (en) * 2001-01-18 2002-03-19 Texas Instruments Incorporated Slew rate boost circuitry and method
US6359612B1 (en) * 1998-09-30 2002-03-19 Siemens Aktiengesellschaft Imaging system for displaying image information that has been acquired by means of a medical diagnostic imaging device
US20020036617A1 (en) * 1998-08-21 2002-03-28 Timothy R. Pryor Novel man machine interfaces and applications
US20020041260A1 (en) * 2000-08-11 2002-04-11 Norbert Grassmann System and method of operator control
US6434255B1 (en) * 1997-10-29 2002-08-13 Takenaka Corporation Hand pointing apparatus
US20020118880A1 (en) * 2000-11-02 2002-08-29 Che-Bin Liu System and method for gesture interface
US20020126876A1 (en) * 1999-08-10 2002-09-12 Paul George V. Tracking and gesture recognition system particularly suited to vehicular control applications
US20030138130A1 (en) * 1998-08-10 2003-07-24 Charles J. Cohen Gesture-controlled interfaces for self-service machines and other applications
US20030202219A1 (en) * 2002-04-24 2003-10-30 Chin-Chung Lien Method and structure for changing a scanning resolution
US20040054284A1 (en) * 2002-09-13 2004-03-18 Acuson Corporation Overlapped scanning for multi-directional compounding of ultrasound images
US6766036B1 (en) * 1999-07-08 2004-07-20 Timothy R. Pryor Camera based man machine interfaces
US20040141634A1 (en) * 2002-10-25 2004-07-22 Keiichi Yamamoto Hand pattern switch device
US20040161132A1 (en) * 1998-08-10 2004-08-19 Cohen Charles J. Gesture-controlled interfaces for self-service machines and other applications
US6788809B1 (en) * 2000-06-30 2004-09-07 Intel Corporation System and method for gesture recognition in three dimensions using stereo imaging and color vision
US20040190776A1 (en) * 2003-03-31 2004-09-30 Honda Motor Co., Ltd. Gesture recognition apparatus, gesture recognition method, and gesture recognition program
US20040189720A1 (en) * 2003-03-25 2004-09-30 Wilson Andrew D. Architecture for controlling a computer using hand gestures
US6819782B1 (en) * 1999-06-08 2004-11-16 Matsushita Electric Industrial Co., Ltd. Device and method for recognizing hand shape and position, and recording medium having program for carrying out the method recorded thereon
US20050271279A1 (en) * 2004-05-14 2005-12-08 Honda Motor Co., Ltd. Sign based human-machine interaction
US7006055B2 (en) * 2001-11-29 2006-02-28 Hewlett-Packard Development Company, L.P. Wireless multi-user multi-projector presentation system

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH09102046A (en) * 1995-08-01 1997-04-15 Matsushita Electric Ind Co Ltd Hand shape recognition method/device
EP0905644A3 (en) * 1997-09-26 2004-02-25 Matsushita Electric Industrial Co., Ltd. Hand gesture recognizing device
JPH11134090A (en) * 1997-10-30 1999-05-21 Tokai Rika Co Ltd Operation signal output device
EP0919906B1 (en) * 1997-11-27 2005-05-25 Matsushita Electric Industrial Co., Ltd. Control method
JPH11167455A (en) 1997-12-05 1999-06-22 Fujitsu Ltd Hand form recognition device and monochromatic object form recognition device
US6501515B1 (en) * 1998-10-13 2002-12-31 Sony Corporation Remote control system
JP2000331170A (en) 1999-05-21 2000-11-30 Atr Media Integration & Communications Res Lab Hand motion recognizing device
JP2001216069A (en) 2000-02-01 2001-08-10 Toshiba Corp Operation inputting device and direction detecting method
JP2002236534A (en) 2001-02-13 2002-08-23 Mitsubishi Motors Corp On-vehicle equipment operation device
JP2003141547A (en) 2001-10-31 2003-05-16 Matsushita Electric Ind Co Ltd Sign language translation apparatus and method

Patent Citations (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US1984030A (en) * 1932-09-22 1934-12-11 John B Nixon Means for serving cocktails and the like
US1965944A (en) * 1933-03-13 1934-07-10 Dudley L Lea Truck construction
US5203704A (en) * 1990-12-21 1993-04-20 Mccloud Seth R Method of communication using pointing vector gestures and mnemonic devices to assist in learning point vector gestures
US5454043A (en) * 1993-07-30 1995-09-26 Mitsubishi Electric Research Laboratories, Inc. Dynamic and static hand gesture recognition through low-level image analysis
US5594469A (en) * 1995-02-21 1997-01-14 Mitsubishi Electric Information Technology Center America Inc. Hand gesture machine control system
US5815147A (en) * 1996-06-07 1998-09-29 The Trustees Of The University Of Pennsylvania Virtual play environment for disabled children
US6002808A (en) * 1996-07-26 1999-12-14 Mitsubishi Electric Information Technology Center America, Inc. Hand gesture control system
US6128003A (en) * 1996-12-20 2000-10-03 Hitachi, Ltd. Hand gesture recognition system and method
US6236736B1 (en) * 1997-02-07 2001-05-22 Ncr Corporation Method and apparatus for detecting movement patterns at a self-service checkout terminal
US6072494A (en) * 1997-10-15 2000-06-06 Electric Planet, Inc. Method and apparatus for real-time gesture recognition
US6434255B1 (en) * 1997-10-29 2002-08-13 Takenaka Corporation Hand pointing apparatus
US20040161132A1 (en) * 1998-08-10 2004-08-19 Cohen Charles J. Gesture-controlled interfaces for self-service machines and other applications
US20060013440A1 (en) * 1998-08-10 2006-01-19 Cohen Charles J Gesture-controlled interfaces for self-service machines and other applications
US20030138130A1 (en) * 1998-08-10 2003-07-24 Charles J. Cohen Gesture-controlled interfaces for self-service machines and other applications
US20020036617A1 (en) * 1998-08-21 2002-03-28 Timothy R. Pryor Novel man machine interfaces and applications
US6256400B1 (en) * 1998-09-28 2001-07-03 Matsushita Electric Industrial Co., Ltd. Method and device for segmenting hand gestures
US6359612B1 (en) * 1998-09-30 2002-03-19 Siemens Aktiengesellschaft Imaging system for displaying image information that has been acquired by means of a medical diagnostic imaging device
US6819782B1 (en) * 1999-06-08 2004-11-16 Matsushita Electric Industrial Co., Ltd. Device and method for recognizing hand shape and position, and recording medium having program for carrying out the method recorded thereon
US6766036B1 (en) * 1999-07-08 2004-07-20 Timothy R. Pryor Camera based man machine interfaces
US20070195997A1 (en) * 1999-08-10 2007-08-23 Paul George V Tracking and gesture recognition system particularly suited to vehicular control applications
US7050606B2 (en) * 1999-08-10 2006-05-23 Cybernet Systems Corporation Tracking and gesture recognition system particularly suited to vehicular control applications
US20020126876A1 (en) * 1999-08-10 2002-09-12 Paul George V. Tracking and gesture recognition system particularly suited to vehicular control applications
US6788809B1 (en) * 2000-06-30 2004-09-07 Intel Corporation System and method for gesture recognition in three dimensions using stereo imaging and color vision
US20020041260A1 (en) * 2000-08-11 2002-04-11 Norbert Grassmann System and method of operator control
US20020118880A1 (en) * 2000-11-02 2002-08-29 Che-Bin Liu System and method for gesture interface
US6359512B1 (en) * 2001-01-18 2002-03-19 Texas Instruments Incorporated Slew rate boost circuitry and method
US7006055B2 (en) * 2001-11-29 2006-02-28 Hewlett-Packard Development Company, L.P. Wireless multi-user multi-projector presentation system
US20030202219A1 (en) * 2002-04-24 2003-10-30 Chin-Chung Lien Method and structure for changing a scanning resolution
US20040054284A1 (en) * 2002-09-13 2004-03-18 Acuson Corporation Overlapped scanning for multi-directional compounding of ultrasound images
US20040141634A1 (en) * 2002-10-25 2004-07-22 Keiichi Yamamoto Hand pattern switch device
US20040189720A1 (en) * 2003-03-25 2004-09-30 Wilson Andrew D. Architecture for controlling a computer using hand gestures
US20040190776A1 (en) * 2003-03-31 2004-09-30 Honda Motor Co., Ltd. Gesture recognition apparatus, gesture recognition method, and gesture recognition program
US20050271279A1 (en) * 2004-05-14 2005-12-08 Honda Motor Co., Ltd. Sign based human-machine interaction

Cited By (120)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040141634A1 (en) * 2002-10-25 2004-07-22 Keiichi Yamamoto Hand pattern switch device
US7289645B2 (en) 2002-10-25 2007-10-30 Mitsubishi Fuso Truck And Bus Corporation Hand pattern switch device
US20050238202A1 (en) * 2004-02-26 2005-10-27 Mitsubishi Fuso Truck And Bus Corporation Hand pattern switching apparatus
US7499569B2 (en) 2004-02-26 2009-03-03 Mitsubishi Fuso Truck And Bus Corporation Hand pattern switching apparatus
WO2006109476A1 (en) * 2005-04-05 2006-10-19 Nissan Motor Co., Ltd. Command input system
US20090287361A1 (en) * 2005-04-05 2009-11-19 Nissan Motor Co., Ltd Command Input System
US20080211832A1 (en) * 2005-09-05 2008-09-04 Toyota Jidosha Kabushiki Kaisha Vehicular Operating Apparatus
US8049722B2 (en) * 2005-09-05 2011-11-01 Toyota Jidosha Kabushiki Kaisha Vehicular operating apparatus
WO2007107368A1 (en) * 2006-03-22 2007-09-27 Volkswagen Ag Interactive operating device and method for operating the interactive operating device
US20090327977A1 (en) * 2006-03-22 2009-12-31 Bachfischer Katharina Interactive control device and method for operating the interactive control device
US9671867B2 (en) 2006-03-22 2017-06-06 Volkswagen Ag Interactive control device and method for operating the interactive control device
US7721207B2 (en) 2006-05-31 2010-05-18 Sony Ericsson Mobile Communications Ab Camera based control
WO2007138393A3 (en) * 2006-05-31 2008-04-17 Sony Ericsson Mobile Comm Ab Camera based control
WO2007138393A2 (en) * 2006-05-31 2007-12-06 Sony Ericsson Mobile Communications Ab Camera based control
US20070283296A1 (en) * 2006-05-31 2007-12-06 Sony Ericsson Mobile Communications Ab Camera based control
WO2008053433A3 (en) * 2006-11-02 2009-03-19 Koninkl Philips Electronics Nv Hand gesture recognition by scanning line-wise hand images and by extracting contour extreme points
WO2008053433A2 (en) * 2006-11-02 2008-05-08 Koninklijke Philips Electronics N.V. Hand gesture recognition by scanning line-wise hand images and by extracting contour extreme points
US20080130953A1 (en) * 2006-12-04 2008-06-05 Denso Corporation Operation estimating apparatus and program
US8077970B2 (en) 2006-12-04 2011-12-13 Denso Corporation Operation estimating apparatus and related article of manufacture
US7983475B2 (en) 2006-12-27 2011-07-19 Takata Corporation Vehicular actuation system
US20080181456A1 (en) * 2006-12-27 2008-07-31 Takata Corporation Vehicular actuation system
US20080168403A1 (en) * 2007-01-06 2008-07-10 Appl Inc. Detecting and interpreting real-world and security gestures on touch and hover sensitive devices
US20100192109A1 (en) * 2007-01-06 2010-07-29 Wayne Carl Westerman Detecting and Interpreting Real-World and Security Gestures on Touch and Hover Sensitive Devices
US20100211920A1 (en) * 2007-01-06 2010-08-19 Wayne Carl Westerman Detecting and Interpreting Real-World and Security Gestures on Touch and Hover Sensitive Devices
US9367235B2 (en) * 2007-01-06 2016-06-14 Apple Inc. Detecting and interpreting real-world and security gestures on touch and hover sensitive devices
US7877707B2 (en) 2007-01-06 2011-01-25 Apple Inc. Detecting and interpreting real-world and security gestures on touch and hover sensitive devices
WO2008085788A2 (en) * 2007-01-06 2008-07-17 Apple Inc. Detecting and interpreting real-world and security gestures on touch and hover sensitive devices
US9158454B2 (en) * 2007-01-06 2015-10-13 Apple Inc. Detecting and interpreting real-world and security gestures on touch and hover sensitive devices
WO2008085788A3 (en) * 2007-01-06 2009-03-05 Apple Inc Detecting and interpreting real-world and security gestures on touch and hover sensitive devices
US20080197996A1 (en) * 2007-01-30 2008-08-21 Toyota Jidosha Kabushiki Kaisha Operating device
US8094189B2 (en) 2007-01-30 2012-01-10 Toyota Jidosha Kabushiki Kaisha Operating device
US20110175843A1 (en) * 2007-07-19 2011-07-21 Bachfischer Katharina Method for determining the position of an actuation element, in particular a finger of a user in a motor vehicle and position determination device
DE102007034273A1 (en) * 2007-07-19 2009-01-22 Volkswagen Ag Method for determining the position of a user's finger in a motor vehicle and position determining device
US9001049B2 (en) 2007-07-19 2015-04-07 Volkswagen Ag Method for determining the position of an actuation element, in particular a finger of a user in a motor vehicle and position determination device
US8378970B2 (en) 2007-10-22 2013-02-19 Mitsubishi Electric Corporation Manipulation input device which detects human hand manipulations from captured motion images
US20090102788A1 (en) * 2007-10-22 2009-04-23 Mitsubishi Electric Corporation Manipulation input device
US8681099B2 (en) 2007-10-22 2014-03-25 Mitsubishi Electric Corporation Manipulation input device which detects human hand manipulations from captured motion images
US8896536B2 (en) 2008-06-10 2014-11-25 Mediatek Inc. Methods and systems for contactlessly controlling electronic devices according to signals from a digital camera and a sensor module
JP2014221636A (en) * 2008-06-18 2014-11-27 オブロング・インダストリーズ・インコーポレーテッド Gesture-based control system for vehicle interface
US8614674B2 (en) 2009-05-21 2013-12-24 May Patents Ltd. System and method for control based on face or hand gesture detection
US20100295782A1 (en) * 2009-05-21 2010-11-25 Yehuda Binder System and method for control based on face ore hand gesture detection
US10582144B2 (en) 2009-05-21 2020-03-03 May Patents Ltd. System and method for control based on face or hand gesture detection
US8614673B2 (en) 2009-05-21 2013-12-24 May Patents Ltd. System and method for control based on face or hand gesture detection
US20110060499A1 (en) * 2009-09-04 2011-03-10 Hyundai Motor Japan R&D Center, Inc. Operation system for vehicle
US8849506B2 (en) * 2009-09-04 2014-09-30 Hyundai Motor Japan R&D Center, Inc. Operation system for vehicle
US20110063425A1 (en) * 2009-09-15 2011-03-17 Delphi Technologies, Inc. Vehicle Operator Control Input Assistance
US9395915B2 (en) * 2009-12-12 2016-07-19 Volkswagen Ag Operating method for a display device in a vehicle
US20130176232A1 (en) * 2009-12-12 2013-07-11 Christoph WAELLER Operating Method for a Display Device in a Vehicle
US8639414B2 (en) * 2009-12-25 2014-01-28 Honda Access Corp. Operation apparatus for on-board devices in automobile
US20110160933A1 (en) * 2009-12-25 2011-06-30 Honda Access Corp. Operation apparatus for on-board devices in automobile
US20110286676A1 (en) * 2010-05-20 2011-11-24 Edge3 Technologies Llc Systems and related methods for three dimensional gesture recognition in vehicles
US8396252B2 (en) * 2010-05-20 2013-03-12 Edge 3 Technologies Systems and related methods for three dimensional gesture recognition in vehicles
US8970696B2 (en) * 2010-09-13 2015-03-03 Ricoh Company, Ltd. Hand and indicating-point positioning method and hand gesture determining method used in human-computer interaction system
US20120062736A1 (en) * 2010-09-13 2012-03-15 Xiong Huaixin Hand and indicating-point positioning method and hand gesture determining method used in human-computer interaction system
WO2012061256A1 (en) * 2010-11-01 2012-05-10 Robert Bosch Gmbh Robust video-based handwriting and gesture recognition for in-car applications
US8817087B2 (en) 2010-11-01 2014-08-26 Robert Bosch Gmbh Robust video-based handwriting and gesture recognition for in-car applications
EP2703950A1 (en) * 2011-04-28 2014-03-05 Nec System Technologies, Ltd. Information processing device, information processing method, and recording medium
US9367732B2 (en) 2011-04-28 2016-06-14 Nec Solution Innovators, Ltd. Information processing device, information processing method, and recording medium
EP2703950A4 (en) * 2011-04-28 2015-01-14 Nec Solution Innovators Ltd Information processing device, information processing method, and recording medium
US20130050076A1 (en) * 2011-08-22 2013-02-28 Research & Business Foundation Sungkyunkwan University Method of recognizing a control command based on finger motion and mobile device using the same
EP2755115A4 (en) * 2011-09-07 2015-05-06 Nitto Denko Corp Method for detecting motion of input body and input device using same
US9449516B2 (en) * 2011-11-16 2016-09-20 Autoconnect Holdings Llc Gesture recognition for on-board display
US20150097798A1 (en) * 2011-11-16 2015-04-09 Flextronics Ap, Llc Gesture recognition for on-board display
US20130146234A1 (en) * 2011-12-07 2013-06-13 Hyundai Motor Company Apparatus and method for blocking incident rays from entering an interior cabin of vehicle
US9108492B2 (en) * 2011-12-07 2015-08-18 Hyundai Motor Company Apparatus and method for blocking incident rays from entering an interior cabin of vehicle
US9440537B2 (en) 2012-01-09 2016-09-13 Daimler Ag Method and device for operating functions displayed on a display unit of a vehicle using gestures which are carried out in a three-dimensional space, and corresponding computer program product
US20140361989A1 (en) * 2012-01-10 2014-12-11 Daimler Ag Method and Device for Operating Functions in a Vehicle Using Gestures Performed in Three-Dimensional Space, and Related Computer Program Product
US8942881B2 (en) 2012-04-02 2015-01-27 Google Inc. Gesture-based automotive controls
US20130321462A1 (en) * 2012-06-01 2013-12-05 Tom G. Salter Gesture based region identification for holograms
US9116666B2 (en) * 2012-06-01 2015-08-25 Microsoft Technology Licensing, Llc Gesture based region identification for holograms
US20140079285A1 (en) * 2012-09-19 2014-03-20 Alps Electric Co., Ltd. Movement prediction device and input apparatus using the same
US20140153774A1 (en) * 2012-12-04 2014-06-05 Alpine Electronics, Inc. Gesture recognition apparatus, gesture recognition method, and recording medium
US9256779B2 (en) * 2012-12-04 2016-02-09 Alpine Electronics, Inc. Gesture recognition apparatus, gesture recognition method, and recording medium
EP2741232A3 (en) * 2012-12-04 2016-04-27 Alpine Electronics, Inc. Gesture recognition apparatus, gesture recognition method, and recording medium
CN103885572A (en) * 2012-12-19 2014-06-25 原相科技股份有限公司 Switching device
CN103049111A (en) * 2012-12-20 2013-04-17 广州视睿电子科技有限公司 Stylus and touch coordinate calculation method
WO2014095070A1 (en) * 2012-12-21 2014-06-26 Harman Becker Automotive Systems Gmbh Input device for a motor vehicle
US20150367859A1 (en) * 2012-12-21 2015-12-24 Harman Becker Automotive Systems Gmbh Input device for a motor vehicle
US9373026B2 (en) 2012-12-28 2016-06-21 Hyundai Motor Company Method and system for recognizing hand gesture using selective illumination
WO2014108160A3 (en) * 2013-01-08 2014-11-27 Audi Ag User interface for the contactless selection of a device function
DE102013000081B4 (en) * 2013-01-08 2018-11-15 Audi Ag Operator interface for contactless selection of a device function
WO2014108160A2 (en) * 2013-01-08 2014-07-17 Audi Ag User interface for the contactless selection of a device function
US9158381B2 (en) * 2013-02-25 2015-10-13 Honda Motor Co., Ltd. Multi-resolution gesture recognition
US20140240213A1 (en) * 2013-02-25 2014-08-28 Honda Motor Co., Ltd. Multi-resolution gesture recognition
US10168794B2 (en) * 2013-05-23 2019-01-01 Fastvdo Llc Motion-assisted visual language for human computer interfaces
US9829984B2 (en) * 2013-05-23 2017-11-28 Fastvdo Llc Motion-assisted visual language for human computer interfaces
US20140347263A1 (en) * 2013-05-23 2014-11-27 Fastvdo Llc Motion-Assisted Visual Language For Human Computer Interfaces
US20150026646A1 (en) * 2013-07-18 2015-01-22 Korea Electronics Technology Institute User interface apparatus based on hand gesture and method providing the same
US20150089455A1 (en) * 2013-09-26 2015-03-26 Fujitsu Limited Gesture input method
US9639164B2 (en) * 2013-09-26 2017-05-02 Fujitsu Limited Gesture input method
US20150131857A1 (en) * 2013-11-08 2015-05-14 Hyundai Motor Company Vehicle recognizing user gesture and method for controlling the same
US10545663B2 (en) * 2013-11-18 2020-01-28 Samsung Electronics Co., Ltd Method for changing an input mode in an electronic device
US20150143277A1 (en) * 2013-11-18 2015-05-21 Samsung Electronics Co., Ltd. Method for changing an input mode in an electronic device
US10444825B2 (en) 2014-02-11 2019-10-15 Ultrahaptics IP Two Limited Drift cancelation for portable object detection and tracking
US10007329B1 (en) 2014-02-11 2018-06-26 Leap Motion, Inc. Drift cancelation for portable object detection and tracking
US11099630B2 (en) 2014-02-11 2021-08-24 Ultrahaptics IP Two Limited Drift cancelation for portable object detection and tracking
US11537196B2 (en) 2014-02-11 2022-12-27 Ultrahaptics IP Two Limited Drift cancelation for portable object detection and tracking
US10345806B2 (en) 2014-02-18 2019-07-09 Jaguar Land Rover Limited Autonomous driving system and method for same
GB2525840B (en) * 2014-02-18 2016-09-07 Jaguar Land Rover Ltd Autonomous driving system and method for same
GB2525840A (en) * 2014-02-18 2015-11-11 Jaguar Land Rover Ltd Autonomous driving system and method for same
US9436872B2 (en) 2014-02-24 2016-09-06 Hong Kong Applied Science and Technology Research Institute Company Limited System and method for detecting and tracking multiple parts of an object
US10475249B2 (en) 2014-04-17 2019-11-12 Ultrahaptics IP Two Limited Safety for wearable virtual reality devices via object detection and tracking
US10043320B2 (en) 2014-04-17 2018-08-07 Leap Motion, Inc. Safety for wearable virtual reality devices via object detection and tracking
US9754167B1 (en) 2014-04-17 2017-09-05 Leap Motion, Inc. Safety for wearable virtual reality devices via object detection and tracking
US11538224B2 (en) 2014-04-17 2022-12-27 Ultrahaptics IP Two Limited Safety for wearable virtual reality devices via object detection and tracking
US9868449B1 (en) 2014-05-30 2018-01-16 Leap Motion, Inc. Recognizing in-air gestures of a control object to control a vehicular control system
US10437347B2 (en) 2014-06-26 2019-10-08 Ultrahaptics IP Two Limited Integrated gestural interaction and multi-user collaboration in immersive virtual reality environments
US11226719B2 (en) * 2014-07-04 2022-01-18 Clarion Co., Ltd. Information processing device
US20170192629A1 (en) * 2014-07-04 2017-07-06 Clarion Co., Ltd. Information processing device
US11749026B2 (en) 2014-08-15 2023-09-05 Ultrahaptics IP Two Limited Automotive and industrial motion sensory device
US11386711B2 (en) 2014-08-15 2022-07-12 Ultrahaptics IP Two Limited Automotive and industrial motion sensory device
US10466800B2 (en) * 2015-02-20 2019-11-05 Clarion Co., Ltd. Vehicle information processing device
US20180059798A1 (en) * 2015-02-20 2018-03-01 Clarion Co., Ltd. Information processing device
US9639323B2 (en) * 2015-04-14 2017-05-02 Hon Hai Precision Industry Co., Ltd. Audio control system and control method thereof
US11144121B2 (en) 2016-06-26 2021-10-12 Apple Inc. Wearable interactive user interface
US10642356B1 (en) * 2016-06-26 2020-05-05 Apple Inc. Wearable interactive user interface
CZ307236B6 (en) * 2016-10-03 2018-04-18 Ĺ KODA AUTO a.s. A device for interactive control of a display device and a method of controlling the device for interactive control of a display device
US11487388B2 (en) * 2017-10-09 2022-11-01 Huawei Technologies Co., Ltd. Anti-accidental touch detection method and apparatus, and terminal
US11307669B2 (en) * 2018-02-14 2022-04-19 Kyocera Corporation Electronic device, moving body, program and control method
US10895918B2 (en) * 2019-03-14 2021-01-19 Igt Gesture recognition system and method

Also Published As

Publication number Publication date
CN1595336A (en) 2005-03-16
KR20050019036A (en) 2005-02-28
KR100575504B1 (en) 2006-05-03
DE102004038965A1 (en) 2005-03-17
JP2005063091A (en) 2005-03-10
DE102004038965B4 (en) 2009-04-02
JP3752246B2 (en) 2006-03-08
CN1313905C (en) 2007-05-02

Similar Documents

Publication Publication Date Title
US20050063564A1 (en) Hand pattern switch device
US7289645B2 (en) Hand pattern switch device
US20130204457A1 (en) Interacting with vehicle controls through gesture recognition
US7057505B2 (en) Alarm information providing apparatus for vehicle
US10095313B2 (en) Input device, vehicle having the input device, and method for controlling the vehicle
US20020041260A1 (en) System and method of operator control
JP2009129171A (en) Information processor loaded in mobile body
JP2005242694A (en) Hand pattern switching apparatus
JP2005063092A (en) Hand pattern switch device
KR20150034018A (en) Vehicle operation device
CN110968184B (en) Equipment control device
JP2009252105A (en) Prompter-type operation device
JP2005063090A (en) Hand pattern switch device
JP4266762B2 (en) Operator identification device and multi-function switch
JP4848997B2 (en) Incorrect operation prevention device and operation error prevention method for in-vehicle equipment
JP2006312347A (en) Command input device
JP2006282145A (en) Command input device
JP3742951B2 (en) Hand pattern switch device
JP2006298003A (en) Command input device
JP2004171476A (en) Hand pattern switching unit
JP5261260B2 (en) Vehicle equipment
JP2006312346A (en) Command input device
KR101500412B1 (en) Gesture recognize apparatus for vehicle
JP3867039B2 (en) Hand pattern switch device
JP2005199973A (en) Input apparatus for vehicle

Legal Events

Date Code Title Description
AS Assignment

Owner name: KEIO UNIVERSITY, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:YAMAMOTO, KEIICHI;SATO, HIROMITSU;OZAWA, SHINJI;AND OTHERS;REEL/FRAME:016031/0777;SIGNING DATES FROM 20041015 TO 20041019

Owner name: MITSUBISHI FUSO TRUCK AND BUS CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:YAMAMOTO, KEIICHI;SATO, HIROMITSU;OZAWA, SHINJI;AND OTHERS;REEL/FRAME:016031/0777;SIGNING DATES FROM 20041015 TO 20041019

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO PAY ISSUE FEE