CN103279767A - Human-machine interaction information generation method based on multi-feature-point combination - Google Patents

Human-machine interaction information generation method based on multi-feature-point combination Download PDF

Info

Publication number
CN103279767A
CN103279767A CN2013101751997A CN201310175199A CN103279767A CN 103279767 A CN103279767 A CN 103279767A CN 2013101751997 A CN2013101751997 A CN 2013101751997A CN 201310175199 A CN201310175199 A CN 201310175199A CN 103279767 A CN103279767 A CN 103279767A
Authority
CN
China
Prior art keywords
human
machine interaction
face
interaction information
unique point
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN2013101751997A
Other languages
Chinese (zh)
Inventor
佘青山
杨伟健
昌凤玲
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Dianzi University
Original Assignee
Hangzhou Dianzi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Dianzi University filed Critical Hangzhou Dianzi University
Priority to CN2013101751997A priority Critical patent/CN103279767A/en
Publication of CN103279767A publication Critical patent/CN103279767A/en
Pending legal-status Critical Current

Links

Images

Abstract

The invention relates to a human-machine interaction information generation method based on multi-feature-point combination. At present, the problem of low head posture identification accuracy in a complicated environment still exists in the field of vision-control-based human-machine interfaces, accordingly, incorrect human-machine interaction information is easily generated. According to the invention, human-machine interaction information is generated by positioning and correspondingly combining respective feature points of a user. The human-machine interaction information generation method comprises the following steps of acquiring a head video sequence through a photographing device; preprocessing, namely denoising and enhancing, the acquired image sequence; then detecting a face by an Adaboost algorithm, and positioning the feature points in a face region; defining three feature point combinations, and identifying the three feature point combinations by designing an identification rule; and finally generating the corresponding human-machine interaction information according to identification results of the three feature point combinations. According to the method, the human-machine interaction information can be generated with higher accuracy; and in the field of the vision-control-based human-machine interfaces, the human-machine interaction information generation method has a wide application prospect particularly in the intelligent human-machine interaction.

Description

Human-machine interactive information generation method based on the multi-characteristic points combination
Technical field
The invention belongs to the man-machine interface field, relate to a kind of human-machine interactive information generation method based on the multi-characteristic points combination.
Background technology
Along with the raising of expanding economy and civilization degree, the survival state of this specific group of disabled person just more and more receives the concern of society.2007, national disabled person's sample survey result showed for the second time, and the existing disabled person's sum of China is about 8,296 ten thousand, and the ratio that accounts for country's total population surpasses 6%, and wherein physical disabilities 2,412 ten thousand, are that proportion is the highest in all kinds of deformity.Meanwhile, " Chinese population aging prediction of the development trend research report " estimates that China's elderly population scale in 2014 will reach 200,000,000,2051 and will reach 4.37 hundred million, the elderly is the colony occurred frequently of cerebral apoplexy disease, has physical disabilities in various degree among the patient.Improving disabled person's survival state, improve disabled person's self-care ability and quality of life, has been the emphasis problem that the whole society pays close attention to, becomes an important and pressing task of all-round construction well-to-do level and harmonious society.
Improving patient's mobility, expand the patient moving scope, is fundamental purpose and the approach that the physical disabilities patient is carried out rehabilitation.For serious physical disabilities patients such as amputation, because the damaged or forfeiture of motor function, scope of activities and activity space are restricted, and how to strengthen patient's locomitivity, have become an important topic of biomedical and field of engineering technology.On the one hand, from the source of human motion, analyze brain to the domination function of limb motion and control, by obtaining the motion thinking information of brain, instruct thereby obtain limb action, realize that brain-machine is mutual; On the other hand, from the main body of human motion, by rehabilitation auxiliary implements such as exploitation artificial limb, wheelchairs, the motor function of compensatory realization patient disappearance enlarges scope of activities and self care ability.Break away from the constraint of traditional man-machine interfaces such as operating rod, button, mode and smart machine (comprising the rehabilitation accessory) by the mankind such as voice, body language custom and nature are linked up, allow it that taking the initiative in offering a hand of intelligence is provided, become an important research field of current man-machine interaction, be subjected to correlative study mechanism and scientific research personnel's extensive concern.Yet be pattern-recognition and the application of control technology in the rehabilitation accessory of expression way with voice, body language, also face many difficult problems.Not high as pattern accuracy of identification under the complex environment, man-machine interaction is nature inadequately, particularly be subject to interference such as light variation, complex environment, still there is the key feature point precision and the head pose recognition accuracy is not high and problems such as generation error human-machine interactive information are novel man-machine interface practicability and an intelligentized ubiquity difficult point.Therefore estimate to carry out research and development at head pose, have that the important techniques research and development are worth and market application foreground widely.
Summary of the invention
Purpose of the present invention be exactly at existing based in the man-machine interface field of vision control because of the not high problem that causes the generation error human-machine interactive information of head pose recognition accuracy in the complex environment, a kind of human-machine interactive information generation method based on the multi-characteristic points combination is provided.
The head visual signal can better be passed on effector's intention, has the feasibility as non-contact type natural interaction information source.When the user in the man-machine interaction process, always the variation because of environment produces the mistake location of some unique points or leaks the location, the human-machine interactive information that system may generation error causes maloperation easily.Therefore, the human-machine interactive information generation method based on the multi-characteristic points combination of the present invention's research can improve the accuracy that human-machine interactive information generates under the complex environment, can avoid effectively causing maloperation because of mistake location or the leakage location of unique point.
In order to realize above purpose, the inventive method mainly may further comprise the steps:
Step (1) head video sequence obtains.Head visual information is obtained by optical lens and cmos image sensor assembly.
Step (2) image sequence pre-service.The image sequence that obtains is carried out gray processing, histogram equalization, morphology preprocess method, the head image sequence of obtaining is carried out de-noising, strengthen the effective information of picture headers.
Step (3) people's face and feature point detection.Adopt earlier image sequence after the Adaboost algorithm strengthens step (2) to carry out people's face and detect, then with facial image as input picture, adopt the Adaboost algorithm to carry out eyes, nose, face location equally, thereby obtain unique point.
Step (4) unique point combination definition and identification.According to the oriented unique point of step (3), three kinds of unique point combinations of definition earlier design recognition rule again three kinds of unique point combinations are identified.
Step (5) human-machine interactive information generates.According to the recognition result self-definition design of three kinds of unique points combination in the step (4) and generate human-machine interactive information.
The present invention compares with existing human-machine interactive information generation method, has following characteristics:
1, user's head movement is unrestricted.
In existing intelligent human-machine interaction, the generation of human-machine interactive information needs user's head to remain on a certain position to carry out corresponding operating as far as possible, head and camera head such as the user keep certain distance, and perhaps user's head will remain on video image centre etc.Will make the user feel constraint like this, not nature.And the present invention has carried out corresponding design to recognition rule, user's head can about or seesaw, but can not influence the correct generation of human-machine interactive information, strengthen the handling of user, make the user nature carry out man-machine interaction.
2, the present invention can generate human-machine interactive information in real time and than high-accuracy ground.
Through the repeatedly test statistics under the varying environment, three kinds of unique point combinations can obtain high recognition among the present invention, therefore also can generate human-machine interactive information than high-accuracy ground, can carry out man-machine interaction in real time fully.
Description of drawings
Fig. 1 is implementing procedure figure of the present invention.
Fig. 2 is each characteristic point coordinates Parameter Map.
Fig. 3 changes recognition rule figure about head.
Fig. 4 upwarps recognition rule figure for face.
Fig. 5 is the recognition rule figure that closes one's eyes.
Embodiment
Describe the impetus that the present invention is based on vision in detail below in conjunction with accompanying drawing and have or not the intention method of discrimination, Fig. 1 is implementing procedure figure.
As Fig. 1, the enforcement of the inventive method mainly comprises seven steps: (1) obtains head visual information by optical lens and cmos image sensor assembly; (2) image sequence that obtains is carried out gray processing, histogram equalization, morphology preprocess method, the head image sequence of obtaining is carried out de-noising; (3) adopt earlier the Adaboost algorithm that the image sequence after strengthening in (2) is carried out people's face and detect, then with facial image as input picture, adopt the Adaboost algorithm to carry out eyes, nose, face location equally; (4) according to oriented unique point in (3), the combination of three kinds of unique points of definition earlier designs recognition rule again three kinds of unique point combinations is identified; (5) make up the self-defining settings of recognition result and generate human-machine interactive information according to three kinds of unique points in (4).One by one each step is elaborated below.
Step 1: the head video sequence obtains
Head visual information is obtained by optical lens and cmos image sensor assembly.
Step 2: image sequence pre-service
Because optical system or electron device influence, image unavoidably can be subjected to noise, need carry out denoising Processing.Concrete gray processing, histogram equalization, the morphology preprocess method of adopting carries out de-noising to the header sequence image that obtains, and strengthens the effective information of picture headers.
Step 3: people's face and feature point detection
Earlier by location people's face the sequence of video images of people's face sorter after strengthening that is trained by the Adaboost learning algorithm, then with people's face as area-of-interest, eyes, nose, the face sorter that adopts the Adaboost algorithm to train carries out eyes, nose, face location respectively.
Suppose
Figure DEST_PATH_IMAGE002
Be pattern (sample) space, it comprises Individual pattern
Figure DEST_PATH_IMAGE006
Training set and corresponding class label
Figure DEST_PATH_IMAGE008
, owing to be the two-value classification problem, herein
Figure DEST_PATH_IMAGE010
, the positive and negative of difference representative sample.Then the algorithm steps of Adaboost training strong classifier is described below:
(1) number of initialization Weak Classifier
Figure DEST_PATH_IMAGE012
(2) make the sorter training number of plies
Figure DEST_PATH_IMAGE014
, and the identical weights of initialization sample
Figure DEST_PATH_IMAGE016
(3)
Figure DEST_PATH_IMAGE018
Layer calls the set of training sample and weight Training obtains Weak Classifier
Figure DEST_PATH_IMAGE022
Weak Classifier
Figure 671581DEST_PATH_IMAGE022
Each pattern of giving
Figure 526405DEST_PATH_IMAGE006
Specify a real number value;
(4) calculate the weighting error in classification
Figure DEST_PATH_IMAGE026
Figure DEST_PATH_IMAGE028
(1)
Wherein
Figure DEST_PATH_IMAGE030
It is the empirical probability of observing at training sample.
If
Figure DEST_PATH_IMAGE032
The time, Weak Classifier then
Figure 994558DEST_PATH_IMAGE022
The performance evaluation factor
Figure DEST_PATH_IMAGE034
(2)
If
Figure DEST_PATH_IMAGE036
The time, then delete the epicycle Weak Classifier, and algorithm stops.
(5) upgrade weight
Figure DEST_PATH_IMAGE038
(3)
Wherein, Be normalized factor, make
Figure DEST_PATH_IMAGE042
(6) order
Figure DEST_PATH_IMAGE044
(7) if
Figure DEST_PATH_IMAGE046
, changeed for (3) step.
(8) final strong classifier is defined as
Figure DEST_PATH_IMAGE048
(4)
Error clearly in (4) step
Figure 784176DEST_PATH_IMAGE026
Be with respect to weight distribution Calculate, and
Figure 536231DEST_PATH_IMAGE026
With
Figure 794912DEST_PATH_IMAGE050
Be inversely proportional to, error is characterized as probability
Figure DEST_PATH_IMAGE052
And, wherein need to consider simultaneously weight distribution And sample
Figure 73895DEST_PATH_IMAGE006
The degree of correct classification.But it should be noted, when
Figure 742774DEST_PATH_IMAGE036
The time, algorithm can be deleted the epicycle Weak Classifier, and the algorithm termination, is that the weight update mechanism lost efficacy because in the weight renewal process in (5) step, can be reduced by the weight of mis-classification sample, and the sample weights of correctly being classified can increase.
Final strong classifier is by all
Figure 112969DEST_PATH_IMAGE012
Individual Weak Classifier is being considered its specific performance properties evaluation factor
Figure DEST_PATH_IMAGE054
Rear weight is chosen in a vote.
Step 4: definition and the identification of unique point combination
In order to realize the generation of human-machine interactive information, the present invention has defined three kinds of unique point combinations, is respectively left eye, right eye, nose (LREN), nose, face (NM) and left eye, right eye (LRE).And by designed regular to three kinds of unique points combination identify.
For identifying clearly is described, in Fig. 2, marked each characteristic point coordinates parameter.The concrete identifying of LREN, NM, LRE is as follows:
(1)LREN。When face during towards the dead ahead, the horizontal ordinate of two line mid points equates with the horizontal ordinate of nose, namely
Figure DEST_PATH_IMAGE056
And when head deflection,
Figure DEST_PATH_IMAGE058
(
Figure DEST_PATH_IMAGE060
, head is turned right;
Figure DEST_PATH_IMAGE062
, head turns left), effect is as shown in Figure 3.
Recognition rule is defined as follows: if
Figure 54249DEST_PATH_IMAGE058
Set up, LREN is identified so.
(2)NM。When face towards the dead ahead and in the raw, when remaining unchanged with the distance of camera simultaneously, the distance of nose and face Be normal value
Figure DEST_PATH_IMAGE066
, therefore as shown in Figure 4, when face upwarps slightly, just detect the variation of its distance easily.
Recognition rule is defined as follows: if
Figure DEST_PATH_IMAGE068
Set up, NM is identified so.
According to this recognition rule, when user's head is done when seesawing, the distance between the corresponding nose face also can change, and this state also can be identified, and with the human-machine interactive information of generation error.Therefore, the present invention has done following design:
According to the geometric relationship of people's face and each unique point, no matter the people is bold for a short time, and the relative geometry position of its unique point remains unchanged, and that is to say that the ratio between it remains unchanged.Therefore the nose face that among the present invention face is upwarped front and back carries out size relatively apart from the ratio with people's face height, the nose face distance before the nose face upwarped less than face apart from the ratio with people's face height after if face upwarped and the ratio of people's face height, this state is identified so.Its mathematic(al) representation is:
Figure DEST_PATH_IMAGE070
(5)
Wherein,
Figure DEST_PATH_IMAGE072
For face upwarp before the distance of nose and face;
Figure DEST_PATH_IMAGE074
For face upwarp after the distance of nose and face;
Figure DEST_PATH_IMAGE076
For face upwarps people's face height before; For face upwarps people's face height afterwards.
Therefore, according to above-mentioned design, the user upwarps face or upwarp face when head keeps motionless when rocking before and after head, and system all will generate human-machine interactive information accurately.
(3)LRE。As shown in Figure 5, when being in open configuration for two, can detect two, and mark
Figure DEST_PATH_IMAGE080
And when being in closure state for two, detect less than two i.e. mark
Figure DEST_PATH_IMAGE082
Recognition rule is defined as follows: if
Figure 224549DEST_PATH_IMAGE082
Set up, LRE is identified so.
Step 5: human-machine interactive information generates
According to four kinds of states in the step 4, namely head turns left, head is turned right, face upwarps and the identification of closing one's eyes, and the user can generate four kinds of corresponding human-machine interactive informations at least according to actual conditions.
(1) one to one.Be that a kind of state generates a kind of human-machine interactive information.Be example (down with) with the electric wheelchair, head only turns left that corresponding electric wheelchair turns left.
(2) one-to-many.Be that a kind of state generates multiple human-machine interactive information.As, face upwarps corresponding electric wheelchair and advances, and face upwarps then that corresponding electric wheelchair stops again, and face upwarps again and corresponding electric wheelchair advances.By that analogy, face upwarps this state and can generate alternately that electric wheelchair advances and Stop message.
(3) many-one.Be that various states generates a kind of human-machine interactive information.As, head turns left and the combination of two states of head right-hand rotation can generate the electric wheelchair Stop message.
(4) multi-to-multi.Be that various states generates multiple human-machine interactive information.As, the combination that head left-hand rotation and face upwarp two states can generate the electric wheelchair forward information, and the combination that head right-hand rotation and face upwarp two states can generate the electric wheelchair backward information.

Claims (1)

1. based on the human-machine interactive information generation method of multi-characteristic points combination, it is characterized in that this method comprises the steps:
Step (1) head video sequence obtains, and specifically: head visual information is obtained by optical lens and cmos image sensor assembly;
Step (2) image sequence pre-service, specifically: the image sequence that obtains is carried out gray processing, histogram equalization, morphology preprocess method, the head image sequence of obtaining is carried out de-noising, strengthen the effective information of picture headers;
Step (3) people's face and feature point detection, specifically: adopt the image sequence after the Adaboost algorithm strengthens step (2) to carry out the detection of people's face earlier, then with facial image as input picture, adopt the Adaboost algorithm to carry out eyes, nose, face location equally, thereby obtain unique point;
Step (4) unique point combination definition and identification, specifically: according to the oriented unique point of step (3), three kinds of unique point combinations of definition earlier design recognition rule again three kinds of unique point combinations are identified;
Step (5) human-machine interactive information generates, specifically: according to the recognition result self-definition design of three kinds of unique point combinations in the step (4) and generate human-machine interactive information.
CN2013101751997A 2013-05-10 2013-05-10 Human-machine interaction information generation method based on multi-feature-point combination Pending CN103279767A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2013101751997A CN103279767A (en) 2013-05-10 2013-05-10 Human-machine interaction information generation method based on multi-feature-point combination

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2013101751997A CN103279767A (en) 2013-05-10 2013-05-10 Human-machine interaction information generation method based on multi-feature-point combination

Publications (1)

Publication Number Publication Date
CN103279767A true CN103279767A (en) 2013-09-04

Family

ID=49062282

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2013101751997A Pending CN103279767A (en) 2013-05-10 2013-05-10 Human-machine interaction information generation method based on multi-feature-point combination

Country Status (1)

Country Link
CN (1) CN103279767A (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6492986B1 (en) * 1997-06-02 2002-12-10 The Trustees Of The University Of Pennsylvania Method for human face shape and motion estimation based on integrating optical flow and deformable models
CN101561710A (en) * 2009-05-19 2009-10-21 重庆大学 Man-machine interaction method based on estimation of human face posture
CN102982316A (en) * 2012-11-05 2013-03-20 安维思电子科技(广州)有限公司 Driver abnormal driving behavior recognition device and method thereof

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6492986B1 (en) * 1997-06-02 2002-12-10 The Trustees Of The University Of Pennsylvania Method for human face shape and motion estimation based on integrating optical flow and deformable models
CN101561710A (en) * 2009-05-19 2009-10-21 重庆大学 Man-machine interaction method based on estimation of human face posture
CN102982316A (en) * 2012-11-05 2013-03-20 安维思电子科技(广州)有限公司 Driver abnormal driving behavior recognition device and method thereof

Similar Documents

Publication Publication Date Title
Lian et al. Attention guided U-Net for accurate iris segmentation
CN104463100B (en) Intelligent wheel chair man-machine interactive system and method based on human facial expression recognition pattern
Meena A study on hand gesture recognition technique
CN108294759A (en) A kind of Driver Fatigue Detection based on CNN Eye state recognitions
CN109993068A (en) A kind of contactless human emotion's recognition methods based on heart rate and facial characteristics
Kumar Sign language recognition for hearing impaired people based on hands symbols classification
Khan et al. Nose tracking cursor control for the people with disabilities: An improved HCI
Rupanagudi et al. A video processing based eye gaze recognition algorithm for wheelchair control
Chaskar et al. On a methodology for detecting diabetic presence from iris image analysis
CN103258208A (en) Method for distinguishing whether head has intentions or not based on vision
Fan et al. Nonintrusive driver fatigue detection
CN117274960A (en) Non-driving gesture recognition method and system for L3-level automatic driving vehicle driver
CN103279767A (en) Human-machine interaction information generation method based on multi-feature-point combination
Taher et al. An extended eye movement tracker system for an electric wheelchair movement control
Wang et al. Objective facial paralysis grading based on p face and eigenflow
Viswanatha et al. An Intelligent Camera Based Eye Controlled Wheelchair System: Haar Cascade and Gaze Estimation Algorithms
Huynh A new eye gaze detection algorithm using PCA features and recurrent neural networks
Jerry et al. Convolutional neural networks for eye detection in remote gaze estimation systems
Karthigayan et al. Genetic algorithm and neural network for face emotion recognition
Yalla et al. Wheel chair movement through eyeball recognition using raspberry Pi
Kantharia et al. Facial behavior recognition using soft computing techniques: A survey
Liu et al. An efficient method for driver fatigue state detection based on deep learning
Benoit et al. Fusing bio-inspired vision data for simplified high level scene interpretation: Application to face motion analysis
Huang et al. Driver fatigue expression recognition research based on convolutional neural network
Zhou Eye-Blink Detection under Low-Light Conditions Based on Zero-DCE

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20130904