US20150212598A1 - Dual mode optical navigation device and mode switching method thereof - Google Patents

Dual mode optical navigation device and mode switching method thereof Download PDF

Info

Publication number
US20150212598A1
US20150212598A1 US14/165,831 US201414165831A US2015212598A1 US 20150212598 A1 US20150212598 A1 US 20150212598A1 US 201414165831 A US201414165831 A US 201414165831A US 2015212598 A1 US2015212598 A1 US 2015212598A1
Authority
US
United States
Prior art keywords
navigation device
mode
optical navigation
image
dual mode
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/165,831
Inventor
Sai-Mun Lee
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Pixart Imaging Inc
Original Assignee
Pixart Imaging Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Pixart Imaging Inc filed Critical Pixart Imaging Inc
Priority to US14/165,831 priority Critical patent/US20150212598A1/en
Assigned to PIXART IMAGING INC. reassignment PIXART IMAGING INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: LEE, SAI-MUN
Priority to TW103122621A priority patent/TWI536211B/en
Priority to CN201410333257.9A priority patent/CN104808938B/en
Priority to CN201910116806.XA priority patent/CN110083299A/en
Publication of US20150212598A1 publication Critical patent/US20150212598A1/en
Priority to US15/596,595 priority patent/US9958965B2/en
Priority to US15/919,698 priority patent/US10209787B2/en
Priority to US16/237,786 priority patent/US10558279B2/en
Priority to US16/730,046 priority patent/US11048342B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0354Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
    • G06F3/03543Mice or pucks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0338Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of limited linear or angular displacement of an operating part of the device from a neutral position, e.g. isotonic or isometric joysticks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0346Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0362Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 1D translations or rotations of an operating part of the device, e.g. scroll wheels, sliders, knobs, rollers or belts

Definitions

  • This disclosure generally relates to an optical navigation device and, more particularly, to a dual mode optical navigation device and a mode switching method thereof that may switch the operation mode according to different operating states.
  • the conventional optical navigation device such as an optical mouse, generally includes a light source, an image sensor and a process unit.
  • the light source illuminates the working surface and the image sensor receives light reflected from the working surface.
  • the process unit of the optical mouse calculates a movement value corresponding to the user's operation according to the images successively captured by the image sensor and converts the movement value to an electric signal.
  • a host then relatively controls a cursor movement according to the electric signal.
  • gesture operations users can perform gesture operations via a touch interface such as a touch pad or a touch screen, for example in Microsoft Windows 8, Google Android and Apple iOS.
  • a touch interface such as a touch pad or a touch screen
  • Microsoft Windows 8, Google Android and Apple iOS Compared with the touch interface, conventional optical mice only have limited applications due to only being able to relatively control cursor movements according to the movement value.
  • a roller is generally added to the optical mouse to be operated by the user, e.g. rolling the roller to implement window scrolling or zooming in/out, or a plurality of buttons for performing relative functions are added.
  • the optical mouse can have problems of low accuracy, poor durability (e.g. due to abrasion of the roller) and large volume (e.g. due to ergonomic design of the pressing area of the buttons).
  • the present disclosure further provides a dual mode optical navigation device and a mode switching method thereof that have a cursor control mode and a gesture operation mode simultaneously.
  • the present disclosure provides a dual mode optical navigation device and a mode switching method thereof that may replace the function of the roller in conventional optical navigation devices so as to increase the practicality.
  • the present disclosure further provides a dual mode optical navigation device and a mode switching method thereof that may scroll windows smoothly according to a displacement of the navigation device relative to a working surface.
  • the present disclosure further provides a dual mode optical navigation device and a mode switching method thereof that may perform gesture operations according to a displacement of the navigation device relative to a working surface so as to enhance the user experience.
  • the present disclosure provides a dual mode optical navigation device operated in a first mode or a second mode on a working surface.
  • the dual mode optical navigation device includes a first bottom surface, a second bottom surface, a light source, an image sensor and a process unit. There is an included angle between the second bottom surface and the first bottom surface.
  • the light source is configured to illuminate the working surface through the first bottom surface.
  • the image sensor is configured to capture an image frame of the working surface through the first bottom surface.
  • the process unit is configured to calculate an image feature of the image frame to accordingly enter the first mode or the second mode, wherein the first bottom surface is configured to be contacted with the working surface in the first mode and the second bottom surface is configured to be contacted with the working surface in the second mode.
  • the present disclosure further provides a dual mode optical navigation device operated in a first mode or a second mode on a working surface.
  • the dual mode optical navigation device includes a first bottom surface, a second bottom surface, a light source, an actuator and a process unit. There is an included angle between the second bottom surface and the first bottom surface, wherein the first bottom surface is configured to be contacted with the working surface in the first mode and the second bottom surface is configured to be contacted with the working surface in the second mode.
  • the actuator is configured to generate a detect signal when one of the first bottom surface and the second bottom surface in contact with the working surface changes from the first bottom surface to the second bottom surface or from the second bottom surface to the first bottom surface.
  • the process unit is configured to enter the first mode or the second mode according to the detect signal.
  • the present disclosure further provides a mode switching method of a dual mode optical navigation device, which may be operated on a working surface and has a first bottom surface and a second bottom surface with an included angle therebetween.
  • the mode switching method includes the steps of: determining, using a process unit, whether the first bottom surface or the second bottom surface contacts with the working surface; switching to a first mode when identifying that the first bottom surface contacts with the working surface and capturing, using an image sensor, image frames of the working surface through the first bottom surface; and switching to a second mode when identifying that the second bottom surface contacts with the working surface and capturing, using the image sensor, image frames of the working surface through the first bottom surface.
  • the image feature may include an intensity distribution ratio, an image brightness value or an image quality of the image frame.
  • the process unit may determine whether the first bottom surface or the second bottom surface contacts with the working surface according to a detect signal of a gyroscope, an accelerometer or a switching element.
  • the first mode is configured to control a cursor movement according to image frames captured by the image sensor; and the second mode is configured to perform a gesture operation according to image frames captured by the image sensor.
  • the dual mode optical navigation device may determine whether a first bottom surface or a second bottom surface contacts with a working surface according to an image feature of an image frame associated with the working surface or according to a detect signal of a gyroscope, an accelerometer or a switching element thereby switching the operation mode of the dual mode optical navigation device to control a cursor movement or perform a gesture operation. Accordingly, the problems existing in the conventional optical navigation device mentioned above may be overcome.
  • FIG. 1 shows a schematic diagram of the dual mode optical navigation device operating in the first mode according to the first embodiment of the present disclosure.
  • FIG. 2 shows a schematic diagram of the dual mode optical navigation device operating in the second mode according to the first embodiment of the present disclosure.
  • FIG. 3 a shows a relationship between the operating angle and the image quality of the dual mode optical navigation device according to the first embodiment of the present disclosure.
  • FIG. 3 b shows a schematic diagram of image frames captured by the image sensor of the dual mode optical navigation device according to the embodiment of the present disclosure.
  • FIG. 4 shows a schematic diagram of the dual mode optical navigation device operating in the first mode according to the second embodiment of the present disclosure.
  • FIG. 5 shows a schematic diagram of the dual mode optical navigation device operating in the second mode according to the second embodiment of the present disclosure.
  • FIG. 6 shows a flow chart of the mode switching method of the dual mode optical navigation system according to the embodiment of the present disclosure.
  • FIG. 1 shows a schematic diagram of a dual mode optical navigation device 1 operating in a first mode according to the first embodiment of the present disclosure
  • FIG. 2 shows a schematic diagram of the dual mode optical navigation device 1 operating in a second mode according to the first embodiment of the present disclosure
  • the dual mode optical navigation device 1 includes a first bottom surface 11 , a second bottom surface 12 , a light source 14 , an image sensor 16 and a process unit 18 .
  • the image sensor 16 is electrically connected to the process unit 18 .
  • a user may operate the dual mode optical navigation device 1 on a working surface S in a first mode or a second mode with his/her palm or a plurality of fingers, wherein the first bottom surface 11 of the dual mode optical navigation device 1 is configured to be contacted with the working surface S in the first mode and the second bottom surface 12 of the dual mode optical navigation device 1 is configured to be contacted with the working surface S in the second mode.
  • the dual mode optical navigation device 1 may be an optical mouse device and connected to a mobile device, a smart TV, a computer system or the like through wire/wireless protocols such as PS/ 2 , USB, Bluetooth or Wi-Fi so as to perform corresponding actions accordingly, e.g. controlling a cursor movement when the dual mode optical navigation device 1 operates in the first mode and performing a gesture operation when the dual mode optical navigation device 1 operates in the second mode, or vice versa.
  • controlling a cursor movement herein may be referred to controlling a cursor movement on a display device; and performing a gesture operation herein may include window scrolling, object zooming and volume control.
  • the dual mode navigation device 1 of the present disclosure may further cooperate with an application program to extend the gesture operation thereof.
  • the dual mode optical navigation device 1 may further have a housing for the user to put the palm or finger(s) thereon so as to perform operations, and the housing is also configured to protect the above mentioned elements. It is appreciated that the first bottom surface 11 and the second bottom surface 12 are a part of the bottom surface of the housing respectively, as shown in FIG. 1 .
  • the dual mode optical navigation device 1 may be operated by contacting the working surface S with the first bottom surface 11 in user operation.
  • a contact area of the first bottom surface 11 is preferably larger than a contact area of the second bottom surface 12 so that the dual mode optical navigation device 1 may preset the first bottom surface 11 , but not limited to, to contact with the working surface S.
  • the included angle ⁇ between the second bottom surface 12 and the first bottom surface 11 , wherein the second bottom surface 12 is located at a back end of the first bottom surface 11 under the dual mode optical navigation device 1 , as shown in FIG. 1 , but the present disclosure is not limited thereto.
  • the second bottom surface 12 of the dual mode optical navigation device 1 may be located at a left end or a right end of the first bottom surface 11 under the dual mode optical navigation device 1 .
  • the included angle ⁇ is preferably within 150-180 degrees.
  • the dual mode optical navigation device 1 is described with only one second bottom surface (i.e. the second bottom surface 12 ).
  • the dual mode optical navigation device 1 may have a plurality of second bottom surfaces 12 located at at least one of a back end, a left end or a right end of the first bottom surface 11 under the dual mode optical navigation device 1 .
  • the dual mode optical navigation device 1 may have two second bottom surfaces 12 respectively located at a left end and a right end of the first bottom surface 11 under the dual mode optical navigation device 1 so that the dual mode optical navigation device 1 may be adapted to both left-handed users and right-handed users.
  • the dual mode optical navigation device 1 may be operated by contacting the working surface S with the second bottom surface 12 in user operation. Since the first bottom surface 11 and the second bottom surface 12 have the included angle ⁇ therebetween, the first bottom surface 11 and the second bottom surface 12 of the dual mode optical navigation device 1 do not contact with the working surface S at the same time.
  • a maximum operating angle ⁇ M i.e. a complementary angle of the included angle ⁇
  • ⁇ M there forms the maximum operating angle between the first bottom surface 11 and the working surface S when the second bottom surface 12 contacts the working surface S. That is to say, the dual mode optical navigation device 1 has an operating angle between the angles of 0- ⁇ M .
  • the light source 14 may illuminate the working surface S through an opening or a light transmissive medium (not shown) at the first bottom surface 11 , wherein the light source 14 may be a light emitting diode (LED), a laser diode (LD) or other active light sources.
  • the spectrum of the light source 14 is preferably adapted to a spectrum range that the image sensor 16 can receive.
  • a light guide element may be disposed on the optical path between the light source 14 and the image sensor 16 so that the image sensor 16 can receive reflective light field of the light source 14 effectively.
  • the image sensor 16 is configured to capture image frames of the working surface S through the opening or the light transmissive medium of the first bottom surface 11 , wherein the image sensor may be a complementary metal oxide semiconductor (CMOS) image sensor or charge-coupled device (CCD) image sensor, but not limited thereto. Furthermore, to prevent external light sources from disturbing the image sensor 16 in capturing the image frame, the image sensor 16 may be covered with a film coating configured to filter out spectrums rather than the spectrum of the light source 14 .
  • CMOS complementary metal oxide semiconductor
  • CCD charge-coupled device
  • the image sensor 16 may be embedded in the process unit 18 . In another embodiment, the image sensor 16 may be integrated with the light source 14 as an optical unit.
  • the light source 14 has to illuminate the working surface S to provide reflective light needed by the image sensor 16 while capturing image frames. Therefore, the spatial relationship between the light source 14 , the image sensor 16 , the first bottom surface 11 and the second bottom surface 12 is arranged in a way so that the image sensor 16 can receive reflected light from the working surface S in both the first and second modes.
  • the process unit 18 is configured to calculate an image feature of the image frame and accordingly enter the first mode or the second mode.
  • the process unit 18 may be a digital signal processor (DSP) or other process devices that can be used to process image data according to the image frame captured by the image sensor 16 .
  • DSP digital signal processor
  • the process unit 18 is not limited to software or hardware.
  • the process unit 18 may control the light source 14 to turn on and off.
  • the image feature of the present embodiment may be an image quality.
  • FIG. 3 a shows a relationship between the operating angle and the image quality of the dual mode optical navigation device 1 , wherein when the first bottom surface 11 of the dual mode optical navigation device 1 contacts with the working surface S (e.g. the state of FIG. 1 ), the image quality calculated by the process unit 18 is larger than (or equal to) a quality threshold value (e.g. Q 1 of FIG. 3 a ) and the dual mode optical navigation device 1 enters the first mode. And, when the user applies an external force P 1 onto the dual mode optical navigation device 1 to allow the second bottom surface 12 to contact with the working surface S (e.g. the state of FIG.
  • a quality threshold value e.g. Q 1 of FIG. 3 a
  • the image quality is within a quality threshold range (e.g. a range from Q 2 to Q 3 of FIG. 3 a ) and the dual mode optical navigation device 1 enters the second mode.
  • the image quality may be referred to commonly owned U.S. Pat. Nos. 7,142,695, 7,444,006 and 7,116,801.
  • the quality threshold and the quality threshold range may be previously saved in a memory unit before the dual mode optical navigation device 1 leaves the factory. It is appreciated that reflective light field received by the image sensor 16 becomes weaker when the dual mode optical navigation device 1 switches from the first mode to the second mode, and thus the quality threshold range is preferably smaller than the quality threshold.
  • Q 1 shown in FIG. 3 a may be defined as the quality threshold, and the range from Q 2 to Q 3 may be defined as the quality threshold range; wherein the quality threshold Q 1 and the quality threshold range Q 2 -Q 3 may be set according to actual measurement results.
  • the range from Q 1 to Q 2 may be served as a buffer when the dual mode optical navigation device 1 switches from the first mode to the second mode or from the second mode to the first mode so as to prevent misoperations.
  • the process unit 18 identifies that the image quality of the image frame is within the quality threshold range Q 1 -Q 2 , it means that the dual mode optical navigation device 1 is operated neither in the first mode nor in the second mode, and the process unit 18 may not post-process the image frame.
  • the image feature may be an intensity distribution ratio.
  • the process unit 18 may determine that the intensity distribution ratio of the image frame F 1 is larger than or equal to a ratio threshold so as to enter the first mode.
  • the process unit 18 may determine whether the intensity distribution ratio formed by the bright area and the dark area (e.g. bright area/dark area) is within a ratio threshold range to enter the second mode, wherein the ratio threshold range may be smaller than the ratio threshold.
  • the image feature may be an image brightness value.
  • the process unit 18 may calculate the image brightness value respectively according to two image frames F 1 and F 2 captured by the image sensor 16 , e.g. calculating an average brightness value of two dimensional pixels of the image frame F 1 and an average brightness value of two dimensional pixels of the image frame F 2 and then comparing the average brightness values with a brightness threshold respectively. Similarly, the process unit 18 may determine whether the image brightness value is larger than or equal to a brightness threshold or within a brightness threshold range so as to enter the first mode or the second mode.
  • the image feature may include an intensity distribution ratio, an image brightness value or an image quality.
  • the process unit 18 may further calculate an image profile of the image frame, identify a boundary location in the image frame or process the image frame with other methods to obtain the image feature.
  • the process unit 18 further calculates a displacement according to image frames of the working surface S successively captured by the image sensor 16 .
  • the method of the process unit 18 calculating the displacement is well known, e.g. calculating the displacement according to correlations between image frames, and thus details thereof are not described herein.
  • the first mode of the dual mode optical navigation device 1 is configured to control a cursor movement
  • the second mode is configured to perform a gesture operation
  • FIG. 4 shows a schematic diagram of a dual mode optical navigation device 2 operating in a first mode according to the second embodiment of the present disclosure
  • FIG. 5 shows a schematic diagram of the dual mode optical navigation device 2 operating in a second mode according to the second embodiment of the present disclosure
  • the dual mode optical navigation device 2 includes a first bottom surface 21 , a second bottom surface 22 , a light source 24 , an image sensor 26 , an actuator 27 and a process unit 28 .
  • the image sensor 26 and the actuator 27 are electrically connected to the process unit 28 respectively.
  • the dual mode optical navigation device 2 may be operated in a first mode or a second mode on a working surface S, wherein the first mode is configured to control a cursor movement according to a displacement and the second mode is configured to perform a gesture operation according to the displacement.
  • the first bottom surface 21 is configured to be contacted with the working surface S in the first mode and the second bottom surface 22 is configured to be contacted with the working surface S in the second mode.
  • the light source 24 is configured to illuminate the working surface S through the first bottom surface 21 .
  • the image sensor 26 is configured to capture image frames of the working surface S thought the first bottom surface 21 for the process unit 28 calculating a displacement, wherein the image sensor 26 captures image frames of the working surface S through the first bottom surface 21 in both the first and second modes.
  • the dual mode optical navigation device 2 further has the actuator 27 , and the actuator 27 is configured to generate a detect signal Sd when one of the first bottom surface 21 and the second bottom surface 22 in contact with the working surface S changes from the first bottom surface 21 to the second bottom surface 22 or from the second bottom surface 22 to the first bottom surface 21 .
  • the process unit 28 may determine whether the dual mode optical navigation device 2 enters the first mode or the second mode according to the detect signal Sd.
  • the actuator 27 of the present embodiment may be disposed at the second bottom surface 22 .
  • the actuator 27 When the user applies an external force P 2 onto the dual mode optical navigation device 2 to allow the second bottom surface 22 to contact with the working surface S, the actuator 27 is pressed and generates a pressed signal. Then the process unit 28 determines that the dual mode optical navigation device 2 enters the second mode according to the pressed signal.
  • the dual mode optical navigation device 2 goes back from the state of FIG. 5 to the state of FIG. 4 .
  • the actuator 27 recovers and generates a recovery signal, and the process unit 28 determines that the dual mode optical navigation device 2 enters the first mode according to the recovery signal.
  • both the pressed signal and the recovery signal are belong to the detect signal Sd.
  • the present embodiment exemplarily shows that the second bottom surface 22 is located at a back end of the first bottom surface 21 under the dual mode optical navigation device 2 , and the actuator 27 is disposed at the second bottom surface 22 , but the present invention is not limited to.
  • the dual mode optical navigation device 2 may have two second bottom surfaces 22 respectively located at a left end and a right end of the first bottom surface 21 under the dual mode optical navigation device 2 , and the actuator 27 is disposed at the second bottom surface(s) or at the first bottom surface only.
  • the number and location of the actuator 27 of the dual mode optical navigation device 2 may be determined according to actual applications; that is to say, the actuator 27 may be disposed at the first bottom surface 21 , the second bottom surface 22 or other locations of the dual mode optical navigation device 2 .
  • the actuator 27 of the present embodiment is exemplarily shown as a physical button, but not limited thereto.
  • the actuator 27 may be a capacitive switch, an optical switch or other switch elements that may be configured to detect the switching between the first bottom surface 21 and the second bottom surface 22 and generate a detect signal correspondingly.
  • the actuator 27 may be a gyroscope or an accelerometer.
  • a gyroscope or an accelerometer in the dual mode optical navigation device 2 , when one of the first bottom surface 21 and the second bottom surface 22 in contact with the working surface S changes from the first bottom surface 21 to the second bottom surface 22 or from the second bottom surface 22 to the first bottom surface 21 , the gyroscope or the accelerometer may generate a detect signal Sd for the process unit 28 determining the mode of the dual mode optical navigation device 2 .
  • the actuator 27 may be integrated with the process unit 28 as a control chip when the actuator 27 is a gyroscope or an accelerometer.
  • parameters associated with the actuator 27 may be previously set before the dual mode optical navigation device 2 leaves the factory. For example, momentum parameters associated with one of the first bottom surface 21 and the second bottom surface 22 in contact with the working surface S changing from the first bottom surface 21 to the second bottom surface 22 or from the second bottom surface 22 to the first bottom surface 21 are respectively previously saved in a memory unit. Therefore, the process unit 28 may switch the dual mode optical navigation device 2 to the first mode when the first bottom surface 21 contacts with the working surface S and switch the dual mode optical navigation device 2 to the second mode when the second bottom surface 22 contacts with the working surface S according to the detect signal outputted by the actuator 27 .
  • the present disclosure further provides a dual mode optical navigation device and a mode switching method thereof that may have a cursor control mode and a gesture operation mode simultaneously, and may determine whether a first bottom surface or a second bottom surface contacts with a working surface according to an image feature of an image frame associated with the working surface ( FIGS. 1 , 2 and the first embodiment) or according to a detect signal of a gyroscope, an accelerometer or a switch element ( FIGS. 4 , 5 and the second embodiment) thereby switching the mode of the dual mode optical navigation device to control a cursor movement or perform a gesture operation. Accordingly, the problems existing in the conventional optical navigation device may be overcome.

Abstract

There is provided a mode switching method of a dual mode optical navigation device. The dual mode optical navigation device is configured to be operated on a working surface and has a first bottom surface and a second bottom surface having an included angle with the first bottom surface. The mode switching method includes: determining whether the first bottom surface or the second bottom surface contacts with the working surface; switching the dual mode optical navigation device to a first mode when identifying that the first bottom surface contacts with the working surface; and switching the dual mode optical navigation device to a second mode when identifying that the second bottom surface contacts with the working surface.

Description

    BACKGROUND
  • 1. Field of the Disclosure
  • This disclosure generally relates to an optical navigation device and, more particularly, to a dual mode optical navigation device and a mode switching method thereof that may switch the operation mode according to different operating states.
  • 2. Description of the Related Art
  • The conventional optical navigation device, such as an optical mouse, generally includes a light source, an image sensor and a process unit. When a user operates the optical mouse on a working surface, the light source illuminates the working surface and the image sensor receives light reflected from the working surface. The process unit of the optical mouse calculates a movement value corresponding to the user's operation according to the images successively captured by the image sensor and converts the movement value to an electric signal. A host then relatively controls a cursor movement according to the electric signal.
  • With the popularity of gesture operations, users can perform gesture operations via a touch interface such as a touch pad or a touch screen, for example in Microsoft Windows 8, Google Android and Apple iOS. Compared with the touch interface, conventional optical mice only have limited applications due to only being able to relatively control cursor movements according to the movement value.
  • For increasing functions (or operating modes) of the optical mouse, a roller is generally added to the optical mouse to be operated by the user, e.g. rolling the roller to implement window scrolling or zooming in/out, or a plurality of buttons for performing relative functions are added. However, using such additional elements to improve the user experience, the optical mouse can have problems of low accuracy, poor durability (e.g. due to abrasion of the roller) and large volume (e.g. due to ergonomic design of the pressing area of the buttons).
  • Accordingly, the present disclosure further provides a dual mode optical navigation device and a mode switching method thereof that have a cursor control mode and a gesture operation mode simultaneously.
  • SUMMARY
  • The present disclosure provides a dual mode optical navigation device and a mode switching method thereof that may replace the function of the roller in conventional optical navigation devices so as to increase the practicality.
  • The present disclosure further provides a dual mode optical navigation device and a mode switching method thereof that may scroll windows smoothly according to a displacement of the navigation device relative to a working surface.
  • The present disclosure further provides a dual mode optical navigation device and a mode switching method thereof that may perform gesture operations according to a displacement of the navigation device relative to a working surface so as to enhance the user experience.
  • The present disclosure provides a dual mode optical navigation device operated in a first mode or a second mode on a working surface. The dual mode optical navigation device includes a first bottom surface, a second bottom surface, a light source, an image sensor and a process unit. There is an included angle between the second bottom surface and the first bottom surface. The light source is configured to illuminate the working surface through the first bottom surface. The image sensor is configured to capture an image frame of the working surface through the first bottom surface. The process unit is configured to calculate an image feature of the image frame to accordingly enter the first mode or the second mode, wherein the first bottom surface is configured to be contacted with the working surface in the first mode and the second bottom surface is configured to be contacted with the working surface in the second mode.
  • The present disclosure further provides a dual mode optical navigation device operated in a first mode or a second mode on a working surface. The dual mode optical navigation device includes a first bottom surface, a second bottom surface, a light source, an actuator and a process unit. There is an included angle between the second bottom surface and the first bottom surface, wherein the first bottom surface is configured to be contacted with the working surface in the first mode and the second bottom surface is configured to be contacted with the working surface in the second mode. The actuator is configured to generate a detect signal when one of the first bottom surface and the second bottom surface in contact with the working surface changes from the first bottom surface to the second bottom surface or from the second bottom surface to the first bottom surface. The process unit is configured to enter the first mode or the second mode according to the detect signal.
  • The present disclosure further provides a mode switching method of a dual mode optical navigation device, which may be operated on a working surface and has a first bottom surface and a second bottom surface with an included angle therebetween. The mode switching method includes the steps of: determining, using a process unit, whether the first bottom surface or the second bottom surface contacts with the working surface; switching to a first mode when identifying that the first bottom surface contacts with the working surface and capturing, using an image sensor, image frames of the working surface through the first bottom surface; and switching to a second mode when identifying that the second bottom surface contacts with the working surface and capturing, using the image sensor, image frames of the working surface through the first bottom surface.
  • In one embodiment, the image feature may include an intensity distribution ratio, an image brightness value or an image quality of the image frame.
  • In one embodiment, the process unit may determine whether the first bottom surface or the second bottom surface contacts with the working surface according to a detect signal of a gyroscope, an accelerometer or a switching element.
  • In one embodiment, the first mode is configured to control a cursor movement according to image frames captured by the image sensor; and the second mode is configured to perform a gesture operation according to image frames captured by the image sensor.
  • The dual mode optical navigation device according to the embodiment of the present disclosure may determine whether a first bottom surface or a second bottom surface contacts with a working surface according to an image feature of an image frame associated with the working surface or according to a detect signal of a gyroscope, an accelerometer or a switching element thereby switching the operation mode of the dual mode optical navigation device to control a cursor movement or perform a gesture operation. Accordingly, the problems existing in the conventional optical navigation device mentioned above may be overcome.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Other objects, advantages, and novel features of the present disclosure will become more apparent from the following detailed description when taken in conjunction with the accompanying drawings.
  • FIG. 1 shows a schematic diagram of the dual mode optical navigation device operating in the first mode according to the first embodiment of the present disclosure.
  • FIG. 2 shows a schematic diagram of the dual mode optical navigation device operating in the second mode according to the first embodiment of the present disclosure.
  • FIG. 3 a shows a relationship between the operating angle and the image quality of the dual mode optical navigation device according to the first embodiment of the present disclosure.
  • FIG. 3 b shows a schematic diagram of image frames captured by the image sensor of the dual mode optical navigation device according to the embodiment of the present disclosure.
  • FIG. 4 shows a schematic diagram of the dual mode optical navigation device operating in the first mode according to the second embodiment of the present disclosure.
  • FIG. 5 shows a schematic diagram of the dual mode optical navigation device operating in the second mode according to the second embodiment of the present disclosure.
  • FIG. 6 shows a flow chart of the mode switching method of the dual mode optical navigation system according to the embodiment of the present disclosure.
  • DETAILED DESCRIPTION OF THE EMBODIMENT
  • It should be noted that, wherever possible, the same reference numbers will be used throughout the drawings to refer to the same or like parts.
  • Referring to FIG. 1 and FIG. 2, FIG. 1 shows a schematic diagram of a dual mode optical navigation device 1 operating in a first mode according to the first embodiment of the present disclosure and FIG. 2 shows a schematic diagram of the dual mode optical navigation device 1 operating in a second mode according to the first embodiment of the present disclosure. The dual mode optical navigation device 1 includes a first bottom surface 11, a second bottom surface 12, a light source 14, an image sensor 16 and a process unit 18. The image sensor 16 is electrically connected to the process unit 18. A user (not shown) may operate the dual mode optical navigation device 1 on a working surface S in a first mode or a second mode with his/her palm or a plurality of fingers, wherein the first bottom surface 11 of the dual mode optical navigation device 1 is configured to be contacted with the working surface S in the first mode and the second bottom surface 12 of the dual mode optical navigation device 1 is configured to be contacted with the working surface S in the second mode.
  • The dual mode optical navigation device 1 may be an optical mouse device and connected to a mobile device, a smart TV, a computer system or the like through wire/wireless protocols such as PS/2, USB, Bluetooth or Wi-Fi so as to perform corresponding actions accordingly, e.g. controlling a cursor movement when the dual mode optical navigation device 1 operates in the first mode and performing a gesture operation when the dual mode optical navigation device 1 operates in the second mode, or vice versa. It is appreciated that controlling a cursor movement herein may be referred to controlling a cursor movement on a display device; and performing a gesture operation herein may include window scrolling, object zooming and volume control. In addition, the dual mode navigation device 1 of the present disclosure may further cooperate with an application program to extend the gesture operation thereof.
  • In addition, those skilled in the art are appreciated that the dual mode optical navigation device 1 may further have a housing for the user to put the palm or finger(s) thereon so as to perform operations, and the housing is also configured to protect the above mentioned elements. It is appreciated that the first bottom surface 11 and the second bottom surface 12 are a part of the bottom surface of the housing respectively, as shown in FIG. 1.
  • The dual mode optical navigation device 1 may be operated by contacting the working surface S with the first bottom surface 11 in user operation. In the present embodiment, a contact area of the first bottom surface 11 is preferably larger than a contact area of the second bottom surface 12 so that the dual mode optical navigation device 1 may preset the first bottom surface 11, but not limited to, to contact with the working surface S.
  • There is an included angle θ between the second bottom surface 12 and the first bottom surface 11, wherein the second bottom surface 12 is located at a back end of the first bottom surface 11 under the dual mode optical navigation device 1, as shown in FIG. 1, but the present disclosure is not limited thereto. In other embodiments, the second bottom surface 12 of the dual mode optical navigation device 1 may be located at a left end or a right end of the first bottom surface 11 under the dual mode optical navigation device 1. For conforming to ergonomic design, the included angle θ is preferably within 150-180 degrees.
  • It should be mentioned that in the present disclosure the dual mode optical navigation device 1 is described with only one second bottom surface (i.e. the second bottom surface 12). In other embodiments, the dual mode optical navigation device 1 may have a plurality of second bottom surfaces 12 located at at least one of a back end, a left end or a right end of the first bottom surface 11 under the dual mode optical navigation device 1. For example, the dual mode optical navigation device 1 may have two second bottom surfaces 12 respectively located at a left end and a right end of the first bottom surface 11 under the dual mode optical navigation device 1 so that the dual mode optical navigation device 1 may be adapted to both left-handed users and right-handed users.
  • In addition, the dual mode optical navigation device 1 may be operated by contacting the working surface S with the second bottom surface 12 in user operation. Since the first bottom surface 11 and the second bottom surface 12 have the included angle θ therebetween, the first bottom surface 11 and the second bottom surface 12 of the dual mode optical navigation device 1 do not contact with the working surface S at the same time. When the first bottom surface 11 contacts the working surface S, there forms a maximum operating angle θM (i.e. a complementary angle of the included angle θ) between the second bottom surface 12 and the working surface S. Similarly, there forms the maximum operating angle θM between the first bottom surface 11 and the working surface S when the second bottom surface 12 contacts the working surface S. That is to say, the dual mode optical navigation device 1 has an operating angle between the angles of 0-θM.
  • The light source 14 may illuminate the working surface S through an opening or a light transmissive medium (not shown) at the first bottom surface 11, wherein the light source 14 may be a light emitting diode (LED), a laser diode (LD) or other active light sources. The spectrum of the light source 14 is preferably adapted to a spectrum range that the image sensor 16 can receive. Furthermore, a light guide element may be disposed on the optical path between the light source 14 and the image sensor 16 so that the image sensor 16 can receive reflective light field of the light source 14 effectively.
  • The image sensor 16 is configured to capture image frames of the working surface S through the opening or the light transmissive medium of the first bottom surface 11, wherein the image sensor may be a complementary metal oxide semiconductor (CMOS) image sensor or charge-coupled device (CCD) image sensor, but not limited thereto. Furthermore, to prevent external light sources from disturbing the image sensor 16 in capturing the image frame, the image sensor 16 may be covered with a film coating configured to filter out spectrums rather than the spectrum of the light source 14.
  • In an embodiment, the image sensor 16 may be embedded in the process unit 18. In another embodiment, the image sensor 16 may be integrated with the light source 14 as an optical unit.
  • It should be mentioned that no matter which of the first bottom surface 11 or the second bottom surface 12 of the dual mode optical navigation device 1 contacts with the working surface S, the light source 14 has to illuminate the working surface S to provide reflective light needed by the image sensor 16 while capturing image frames. Therefore, the spatial relationship between the light source 14, the image sensor 16, the first bottom surface 11 and the second bottom surface 12 is arranged in a way so that the image sensor 16 can receive reflected light from the working surface S in both the first and second modes.
  • The process unit 18 is configured to calculate an image feature of the image frame and accordingly enter the first mode or the second mode. The process unit 18 may be a digital signal processor (DSP) or other process devices that can be used to process image data according to the image frame captured by the image sensor 16. The process unit 18 is not limited to software or hardware. Preferably, the process unit 18 may control the light source 14 to turn on and off.
  • It should be mentioned that the image feature of the present embodiment may be an image quality. As shown in FIG. 3 a, it shows a relationship between the operating angle and the image quality of the dual mode optical navigation device 1, wherein when the first bottom surface 11 of the dual mode optical navigation device 1 contacts with the working surface S (e.g. the state of FIG. 1), the image quality calculated by the process unit 18 is larger than (or equal to) a quality threshold value (e.g. Q1 of FIG. 3 a) and the dual mode optical navigation device 1 enters the first mode. And, when the user applies an external force P1 onto the dual mode optical navigation device 1 to allow the second bottom surface 12 to contact with the working surface S (e.g. the state of FIG. 2), the image quality is within a quality threshold range (e.g. a range from Q2 to Q3 of FIG. 3 a) and the dual mode optical navigation device 1 enters the second mode. The image quality may be referred to commonly owned U.S. Pat. Nos. 7,142,695, 7,444,006 and 7,116,801.
  • The quality threshold and the quality threshold range may be previously saved in a memory unit before the dual mode optical navigation device 1 leaves the factory. It is appreciated that reflective light field received by the image sensor 16 becomes weaker when the dual mode optical navigation device 1 switches from the first mode to the second mode, and thus the quality threshold range is preferably smaller than the quality threshold. For example, Q1 shown in FIG. 3 a may be defined as the quality threshold, and the range from Q2 to Q3 may be defined as the quality threshold range; wherein the quality threshold Q1 and the quality threshold range Q2-Q3 may be set according to actual measurement results. In the present embodiment, the range from Q1 to Q2 may be served as a buffer when the dual mode optical navigation device 1 switches from the first mode to the second mode or from the second mode to the first mode so as to prevent misoperations. For example, when the process unit 18 identifies that the image quality of the image frame is within the quality threshold range Q1-Q2, it means that the dual mode optical navigation device 1 is operated neither in the first mode nor in the second mode, and the process unit 18 may not post-process the image frame.
  • In addition to the above mentioned image quality, in another embodiment the image feature may be an intensity distribution ratio. For example, referring to FIG. 3 b, when the first bottom surface H of the dual mode optical navigation device 1 contacts with the working surface S (e.g. the state of FIG. 1), an image frame F1 captured by the image sensor 16 is fully bright, and thus the process unit 18 may determine that the intensity distribution ratio of the image frame F1 is larger than or equal to a ratio threshold so as to enter the first mode. When an image frame F2 captured by the image sensor 16 has a bright area and a dark area, the process unit 18 may determine whether the intensity distribution ratio formed by the bright area and the dark area (e.g. bright area/dark area) is within a ratio threshold range to enter the second mode, wherein the ratio threshold range may be smaller than the ratio threshold.
  • In another embodiment, the image feature may be an image brightness value. Referring to FIG. 3 b, the process unit 18 may calculate the image brightness value respectively according to two image frames F1 and F2 captured by the image sensor 16, e.g. calculating an average brightness value of two dimensional pixels of the image frame F1 and an average brightness value of two dimensional pixels of the image frame F2 and then comparing the average brightness values with a brightness threshold respectively. Similarly, the process unit 18 may determine whether the image brightness value is larger than or equal to a brightness threshold or within a brightness threshold range so as to enter the first mode or the second mode.
  • Therefore, in the present disclosure the image feature may include an intensity distribution ratio, an image brightness value or an image quality. Besides, the process unit 18 may further calculate an image profile of the image frame, identify a boundary location in the image frame or process the image frame with other methods to obtain the image feature.
  • On the other hand, the process unit 18 further calculates a displacement according to image frames of the working surface S successively captured by the image sensor 16. In the present embodiment, the method of the process unit 18 calculating the displacement is well known, e.g. calculating the displacement according to correlations between image frames, and thus details thereof are not described herein.
  • In the present embodiment, the first mode of the dual mode optical navigation device 1 is configured to control a cursor movement, and the second mode is configured to perform a gesture operation.
  • Referring to FIG. 4 and FIG. 5, FIG. 4 shows a schematic diagram of a dual mode optical navigation device 2 operating in a first mode according to the second embodiment of the present disclosure and FIG. 5 shows a schematic diagram of the dual mode optical navigation device 2 operating in a second mode according to the second embodiment of the present disclosure. The dual mode optical navigation device 2 includes a first bottom surface 21, a second bottom surface 22, a light source 24, an image sensor 26, an actuator 27 and a process unit 28. The image sensor 26 and the actuator 27 are electrically connected to the process unit 28 respectively. The dual mode optical navigation device 2 may be operated in a first mode or a second mode on a working surface S, wherein the first mode is configured to control a cursor movement according to a displacement and the second mode is configured to perform a gesture operation according to the displacement.
  • Similar to the first embodiment of the present disclosure, there is an included angle θ between the second bottom surface 22 and the first bottom surface 21, and the first bottom surface 21 is configured to be contacted with the working surface S in the first mode and the second bottom surface 22 is configured to be contacted with the working surface S in the second mode. The light source 24 is configured to illuminate the working surface S through the first bottom surface 21. The image sensor 26 is configured to capture image frames of the working surface S thought the first bottom surface 21 for the process unit 28 calculating a displacement, wherein the image sensor 26 captures image frames of the working surface S through the first bottom surface 21 in both the first and second modes.
  • The difference between the present embodiment and the first embodiment is that the dual mode optical navigation device 2 further has the actuator 27, and the actuator 27 is configured to generate a detect signal Sd when one of the first bottom surface 21 and the second bottom surface 22 in contact with the working surface S changes from the first bottom surface 21 to the second bottom surface 22 or from the second bottom surface 22 to the first bottom surface 21. The process unit 28 may determine whether the dual mode optical navigation device 2 enters the first mode or the second mode according to the detect signal Sd.
  • Referring to FIG. 4 and FIG. 5, for example when the actuator 27 is a mechanical switch, the actuator 27 of the present embodiment may be disposed at the second bottom surface 22. When the user applies an external force P2 onto the dual mode optical navigation device 2 to allow the second bottom surface 22 to contact with the working surface S, the actuator 27 is pressed and generates a pressed signal. Then the process unit 28 determines that the dual mode optical navigation device 2 enters the second mode according to the pressed signal.
  • Similarly, when the user releases the external force P2, the dual mode optical navigation device 2 goes back from the state of FIG. 5 to the state of FIG. 4. Meanwhile the actuator 27 recovers and generates a recovery signal, and the process unit 28 determines that the dual mode optical navigation device 2 enters the first mode according to the recovery signal. In the present embodiment, both the pressed signal and the recovery signal are belong to the detect signal Sd.
  • It should be mentioned that the present embodiment exemplarily shows that the second bottom surface 22 is located at a back end of the first bottom surface 21 under the dual mode optical navigation device 2, and the actuator 27 is disposed at the second bottom surface 22, but the present invention is not limited to. In another embodiment, the dual mode optical navigation device 2 may have two second bottom surfaces 22 respectively located at a left end and a right end of the first bottom surface 21 under the dual mode optical navigation device 2, and the actuator 27 is disposed at the second bottom surface(s) or at the first bottom surface only. Thus the number and location of the actuator 27 of the dual mode optical navigation device 2 may be determined according to actual applications; that is to say, the actuator 27 may be disposed at the first bottom surface 21, the second bottom surface 22 or other locations of the dual mode optical navigation device 2.
  • On the other hand, the actuator 27 of the present embodiment is exemplarily shown as a physical button, but not limited thereto. The actuator 27 may be a capacitive switch, an optical switch or other switch elements that may be configured to detect the switching between the first bottom surface 21 and the second bottom surface 22 and generate a detect signal correspondingly.
  • In other embodiments, the actuator 27 may be a gyroscope or an accelerometer. For example, disposing a gyroscope or an accelerometer in the dual mode optical navigation device 2, when one of the first bottom surface 21 and the second bottom surface 22 in contact with the working surface S changes from the first bottom surface 21 to the second bottom surface 22 or from the second bottom surface 22 to the first bottom surface 21, the gyroscope or the accelerometer may generate a detect signal Sd for the process unit 28 determining the mode of the dual mode optical navigation device 2. The actuator 27 may be integrated with the process unit 28 as a control chip when the actuator 27 is a gyroscope or an accelerometer.
  • Similarly, parameters associated with the actuator 27 may be previously set before the dual mode optical navigation device 2 leaves the factory. For example, momentum parameters associated with one of the first bottom surface 21 and the second bottom surface 22 in contact with the working surface S changing from the first bottom surface 21 to the second bottom surface 22 or from the second bottom surface 22 to the first bottom surface 21 are respectively previously saved in a memory unit. Therefore, the process unit 28 may switch the dual mode optical navigation device 2 to the first mode when the first bottom surface 21 contacts with the working surface S and switch the dual mode optical navigation device 2 to the second mode when the second bottom surface 22 contacts with the working surface S according to the detect signal outputted by the actuator 27.
  • As mentioned above, the conventional optical navigation device only relatively controls cursor movements according to a displacement and have problems of low accuracy, poor durability and large volume. Therefore, the present disclosure further provides a dual mode optical navigation device and a mode switching method thereof that may have a cursor control mode and a gesture operation mode simultaneously, and may determine whether a first bottom surface or a second bottom surface contacts with a working surface according to an image feature of an image frame associated with the working surface (FIGS. 1, 2 and the first embodiment) or according to a detect signal of a gyroscope, an accelerometer or a switch element (FIGS. 4, 5 and the second embodiment) thereby switching the mode of the dual mode optical navigation device to control a cursor movement or perform a gesture operation. Accordingly, the problems existing in the conventional optical navigation device may be overcome.
  • Although the disclosure has been explained in relation to its preferred embodiment, it is not used to limit the disclosure. It is to be understood that many other possible modifications and variations can be made by those skilled in the art without departing from the spirit and scope of the disclosure as hereinafter claimed.

Claims (20)

What is claimed is:
1. A dual mode optical navigation device, operated in a first mode or a second mode on a working surface, the dual mode optical navigation device comprising:
a first bottom surface;
a second bottom surface having an included angle with the first bottom surface;
a light source configured to illuminate the working surface through the first bottom surface;
an image sensor configured to capture an image frame of the working surface through the first bottom surface; and
a process unit configured to calculate an image feature of the image frame to accordingly enter the first mode or the second mode,
wherein the first bottom surface is configured to be contacted with the working surface in the first mode and the second bottom surface is configured to be contacted with the working surface in the second mode.
2. The dual mode optical navigation device as claimed in claim 1, wherein the image feature is an intensity distribution ratio, an image brightness value or an image quality of the image frame.
3. The dual mode optical navigation device as claimed in claim 2, wherein the first mode is entered when the intensity distribution ratio is larger than a ratio threshold, the image brightness value is larger than a brightness threshold or the image quality is larger than a quality threshold; the second mode is entered when the intensity distribution ratio is within a ratio threshold range, the image brightness value is within a brightness threshold range or the image quality is within a quality threshold range; and the ratio threshold range, the brightness threshold range and the quality threshold range are respectively smaller than the ratio threshold, the brightness threshold and the quality threshold.
4. The dual mode optical navigation device as claimed in claim 1, wherein the first mode is configured to control a cursor movement.
5. The dual mode optical navigation device as claimed in claim 1, wherein the second mode is configured to perform a gesture operation.
6. The dual mode optical navigation device as claimed in claim 1, wherein the image sensor comprises a film coating configured to filter out spectrums rather than a spectrum of the light source.
7. The dual mode optical navigation device as claimed in claim 1, wherein the second bottom surface is located at at least one of a back end, a left end or a right end of the first bottom surface under the dual mode optical navigation device.
8. A dual mode optical navigation device, operated in a first mode or a second mode on a working surface, the dual mode optical navigation device comprising:
a first bottom surface;
a second bottom surface having an included angle with the first bottom surface, wherein the first bottom surface is configured to be contacted with the working surface in the first mode and the second bottom surface is configured to be contacted with the working surface in the second mode;
an actuator configured to generate a detect signal when one of the first bottom surface and the second bottom surface in contact with the working surface changes from the first bottom surface to the second bottom surface or from the second bottom surface to the first bottom surface; and
a process unit configured to enter the first mode or the second mode according to the detect signal.
9. The dual mode optical navigation device as claimed in claim 8, further comprising:
a light source configured to illuminate the working surface through the first bottom surface; and
an image sensor configured to capture image frames of the working surface thought the first bottom surface for the process unit calculating a displacement.
10. The dual mode optical navigation device as claimed in claim 9, wherein the first mode is configured to control a cursor movement according to the displacement.
11. The dual mode optical navigation device as claimed in claim 9, wherein the second mode is configured to perform a gesture operation according to the displacement.
12. The dual mode optical navigation device as claimed in claim 8, wherein the actuator is a gyroscope or an accelerometer.
13. The dual mode optical navigation device as claimed in claim 8, wherein the actuator is a switching element disposed at the first bottom surface or the second bottom surface of the dual mode optical navigation device.
14. The dual mode optical navigation device as claimed in claim 8, wherein the second bottom surface is located at at least one of a back end, a left end or a right end of the first bottom surface under the dual mode optical navigation device.
15. A mode switching method of a dual mode optical navigation device, the dual mode optical navigation device being operated on a working surface and having a first bottom surface and a second bottom surface with an included angle therebetween, the mode switching method comprising:
determining, using a process unit, whether the first bottom surface or the second bottom surface contacts with the working surface;
switching to a first mode when identifying that the first bottom surface contacts with the working surface and capturing, using an image sensor, image frames of the working surface through the first bottom surface; and
switching to a second mode when identifying that the second bottom surface contacts with the working surface and capturing, using the image sensor, image frames of the working surface through the first bottom surface.
16. The mode switching method as claimed in claim 15, wherein the process unit determines whether the first bottom surface or the second bottom surface contacts with the working surface according to an image feature of an image frame associated with the working surface.
17. The mode switching method as claimed in claim 16, wherein the image feature comprises an intensity distribution ratio, an image brightness value or an image quality.
18. The mode switching method as claimed in claim 17, wherein the first mode is entered when the intensity distribution ratio is larger than a ratio threshold, the image brightness value is larger than a brightness threshold or the image quality is larger than a quality threshold; the second mode is entered when the intensity distribution ratio is within a ratio threshold range, the image brightness value is within a brightness threshold range or the image quality is within a quality threshold range; and the ratio threshold range, the brightness threshold range and the quality threshold range are respectively smaller than the ratio threshold, the brightness threshold and the quality threshold.
19. The mode switching method as claimed in claim 15, wherein the process unit determines whether the first bottom surface or the second bottom surface contacts with the working surface according to a detect signal of a gyroscope, an accelerometer or a switching element.
20. The mode switching method as claimed in claim 15, further comprising:
calculating, using the process unit, a displacement according to image frames of the working surface captured by the image sensor through the first bottom surface, wherein the first mode is configured to control a cursor movement according to the displacement and the second mode is configured to perform a gesture operation according to the displacement.
US14/165,831 2014-01-28 2014-01-28 Dual mode optical navigation device and mode switching method thereof Abandoned US20150212598A1 (en)

Priority Applications (8)

Application Number Priority Date Filing Date Title
US14/165,831 US20150212598A1 (en) 2014-01-28 2014-01-28 Dual mode optical navigation device and mode switching method thereof
TW103122621A TWI536211B (en) 2014-01-28 2014-06-30 Dual mode optical navigation device and mode switching method thereof
CN201410333257.9A CN104808938B (en) 2014-01-28 2014-07-14 Bimodulus optical navigator and its mode switching method
CN201910116806.XA CN110083299A (en) 2014-01-28 2014-07-14 Bimodulus optical navigator
US15/596,595 US9958965B2 (en) 2014-01-28 2017-05-16 Dual mode optical navigation device and mode switching method thereof
US15/919,698 US10209787B2 (en) 2014-01-28 2018-03-13 Dual mode optical navigation device and mode switching method thereof
US16/237,786 US10558279B2 (en) 2014-01-28 2019-01-02 Dual mode optical navigation device
US16/730,046 US11048342B2 (en) 2014-01-28 2019-12-30 Dual mode optical navigation device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US14/165,831 US20150212598A1 (en) 2014-01-28 2014-01-28 Dual mode optical navigation device and mode switching method thereof

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/596,595 Continuation US9958965B2 (en) 2014-01-28 2017-05-16 Dual mode optical navigation device and mode switching method thereof

Publications (1)

Publication Number Publication Date
US20150212598A1 true US20150212598A1 (en) 2015-07-30

Family

ID=53679002

Family Applications (5)

Application Number Title Priority Date Filing Date
US14/165,831 Abandoned US20150212598A1 (en) 2014-01-28 2014-01-28 Dual mode optical navigation device and mode switching method thereof
US15/596,595 Active US9958965B2 (en) 2014-01-28 2017-05-16 Dual mode optical navigation device and mode switching method thereof
US15/919,698 Active US10209787B2 (en) 2014-01-28 2018-03-13 Dual mode optical navigation device and mode switching method thereof
US16/237,786 Active US10558279B2 (en) 2014-01-28 2019-01-02 Dual mode optical navigation device
US16/730,046 Active US11048342B2 (en) 2014-01-28 2019-12-30 Dual mode optical navigation device

Family Applications After (4)

Application Number Title Priority Date Filing Date
US15/596,595 Active US9958965B2 (en) 2014-01-28 2017-05-16 Dual mode optical navigation device and mode switching method thereof
US15/919,698 Active US10209787B2 (en) 2014-01-28 2018-03-13 Dual mode optical navigation device and mode switching method thereof
US16/237,786 Active US10558279B2 (en) 2014-01-28 2019-01-02 Dual mode optical navigation device
US16/730,046 Active US11048342B2 (en) 2014-01-28 2019-12-30 Dual mode optical navigation device

Country Status (3)

Country Link
US (5) US20150212598A1 (en)
CN (2) CN104808938B (en)
TW (1) TWI536211B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9632657B2 (en) * 2014-12-28 2017-04-25 Sap Se Auxiliary input device
US11307308B2 (en) * 2017-06-02 2022-04-19 Pixart Imaging Inc. Tracking device and electronic device with improved work surface adaptability
US11435837B2 (en) * 2020-03-09 2022-09-06 Dexin Electronic Ltd. Dual mode mouse device

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107037891A (en) * 2015-08-19 2017-08-11 原建桥 A kind of combined type computer mouse
CN106708287A (en) * 2015-11-16 2017-05-24 原相科技股份有限公司 Optical navigation device for compensating unfocused image and compensation circuit thereof
TWI681322B (en) * 2016-08-22 2020-01-01 原相科技股份有限公司 Optical detecting device capable of determining shift of a multi-axis instruction outputting mechanism

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5936612A (en) * 1997-05-30 1999-08-10 Wang; Yanqing Computer input device and method for 3-D direct manipulation of graphic objects
US20040032392A1 (en) * 2002-08-19 2004-02-19 Topseed Technology Corp. Mouse pen device having remote-control function
US6844871B1 (en) * 1999-11-05 2005-01-18 Microsoft Corporation Method and apparatus for computer input using six degrees of freedom
US20050116933A1 (en) * 2003-12-02 2005-06-02 Hsun-Li Huang Dual mode computer mouse
US20080266251A1 (en) * 2007-04-24 2008-10-30 Pixart Imaging Inc. Cursor control device and method for an image display, and image system
US20090153486A1 (en) * 2007-12-18 2009-06-18 Microsoft Corporation Optical mouse with limited wavelength optics
US20090153482A1 (en) * 2007-12-12 2009-06-18 Weinberg Marc S Computer input device with inertial instruments
US20110013001A1 (en) * 2008-01-28 2011-01-20 Thomas Craven-Bartle Digital pens and a method for digital recording of information
US20120020529A1 (en) * 2010-07-23 2012-01-26 Pixart Imaging Inc. Displacement estimation method and displacement estimation device using the same
WO2014009933A1 (en) * 2012-07-12 2014-01-16 Grant Neville Odgers Improvements in devices for use with computers
US20150022449A1 (en) * 2013-07-16 2015-01-22 Amchael Visual Technology Corporation Algorithms, software and an interaction system that support the operation of an on the fly mouse

Family Cites Families (45)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3615462B2 (en) * 2000-05-12 2005-02-02 三洋電機株式会社 Automatic exposure control camera
US7142695B2 (en) 2002-11-01 2006-11-28 Pixart Imaging Inc. Image qualification for optical navigation sensor
US7444006B2 (en) 2002-04-16 2008-10-28 Pixart Imaging Inc. Method of image qualification for optical navigation sensor
CN1206585C (en) * 2003-01-06 2005-06-15 致伸科技股份有限公司 Dual-Purpose input device
US7064744B2 (en) * 2003-04-18 2006-06-20 Forward Electronics Co., Ltd. Optical induction/trackball dual-mode mouse
US7637430B2 (en) * 2003-05-12 2009-12-29 Hand Held Products, Inc. Picture taking optical reader
TWI234733B (en) * 2003-11-17 2005-06-21 Primax Electronics Ltd Cursor control device
CN100359448C (en) * 2004-06-28 2008-01-02 凌阳科技股份有限公司 Method and system for instantly judging pick-up image pixel value abnormality
KR100489959B1 (en) * 2004-10-28 2005-05-17 퍼셉컴(주) Structure of optical pen type pointing device
CN101008874A (en) * 2006-01-24 2007-08-01 达方电子股份有限公司 Mouse and method for controlling cursor thereof
TWI366163B (en) * 2006-09-15 2012-06-11 Au Optronics Corp Apparatus and method for adaptively adjusting backlight
US7514668B2 (en) * 2006-12-19 2009-04-07 Avago Technologies Ecbu Ip (Singapore) Pte. Ltd. Optical navigation device that utilizes a vertical cavity surface emitting laser (VCSEL) configured to emit visible coherent light
US7965278B2 (en) * 2006-12-29 2011-06-21 Avago Technologies Ecbu Ip (Singapore) Pte. Ltd. Optical navigation device adapted for navigation on a transparent plate
US9176598B2 (en) * 2007-05-08 2015-11-03 Thinkoptics, Inc. Free-space multi-dimensional absolute pointer with improved performance
US8730167B2 (en) * 2007-06-28 2014-05-20 Microsoft Corporation Pointing device with optical positioning on low-diffusive surfaces
CN101398723B (en) * 2007-09-25 2010-06-23 昆盈企业股份有限公司 Mouse action mode switching method
CN102609120A (en) * 2007-11-30 2012-07-25 原相科技股份有限公司 Cursor control device and method on image display device, and image system
CN101452349B (en) * 2007-11-30 2012-05-09 原相科技股份有限公司 Cursor controller on image display apparatus, method and image system
US8451226B2 (en) * 2007-12-13 2013-05-28 Microsoft Corporation User input device with optical tracking engine that operates through transparent substrate
CN102473070B (en) * 2009-08-10 2016-06-01 皇家飞利浦电子股份有限公司 For moving the light target system and method on screen
TW201115403A (en) * 2009-10-21 2011-05-01 Kye Systems Corp Selection method for wavelenght section of pen-shaped optical input device, and its pen-shaped optical input
TWI399985B (en) * 2009-12-16 2013-06-21 Micro Star Int Co Ltd Method for adjusting image
CN105718088A (en) * 2010-07-08 2016-06-29 唐建宁 Computer system
TWI414971B (en) * 2010-08-13 2013-11-11 Pixart Imaging Inc Lift detection method for optical mouse and optical mouse using the same
CN102033630A (en) * 2010-12-01 2011-04-27 邓仕林 Optics and acceleration joint location mouse and acceleration correction mode thereof
DE102011104550B4 (en) * 2011-06-17 2014-04-30 Precitec Kg Optical measuring device for monitoring a joint seam, joining head and laser welding head with the same
CN102968177B (en) * 2011-08-31 2015-10-28 敦宏科技股份有限公司 Gesture method for sensing
TWI489325B (en) * 2011-09-21 2015-06-21 Pixart Imaging Inc Optical finger mouse, electronic device and physiological characteristic detection device
US20130088429A1 (en) * 2011-10-05 2013-04-11 Pantech Co., Ltd. Apparatus and method for recognizing user input
TWI522870B (en) * 2011-10-25 2016-02-21 原相科技股份有限公司 Click event detection device
TWI461972B (en) * 2011-11-04 2014-11-21 Pixart Imaging Inc Optical mouse and operating method thereof
TWI442289B (en) * 2011-11-04 2014-06-21 Pixart Imaging Inc Displacement detection device and operation method thereof
TWI474215B (en) * 2011-11-17 2015-02-21 Pixart Imaging Inc Keyboard module and display system
TWI451295B (en) * 2011-11-17 2014-09-01 Pixart Imaging Inc Optical finger mouse, mouse control module and physiological characteristic detection method thereof
TWI475193B (en) * 2011-11-18 2015-03-01 Pixart Imaging Inc Optical distance measurement system and operation method thereof
CN103123542B (en) * 2011-11-21 2015-12-16 原相科技股份有限公司 Optical mouse and method of operating thereof
TWI438658B (en) * 2011-11-22 2014-05-21 Pixart Imaging Inc User interface system and optical finger mouse system
TWI476641B (en) * 2011-11-22 2015-03-11 Pixart Imaging Inc Remote controller and display system
TW201335802A (en) * 2012-02-20 2013-09-01 Pixart Imaging Inc Optical navigation device
TWI439907B (en) * 2012-02-29 2014-06-01 Pixart Imaging Inc Optical touch device and detection method thereof
US9304574B2 (en) * 2012-04-13 2016-04-05 Pixart Imaging Inc. Remote device and power saving method of interactive system
US9743057B2 (en) * 2012-05-31 2017-08-22 Apple Inc. Systems and methods for lens shading correction
JP5662623B1 (en) * 2013-03-27 2015-02-04 富士フイルム株式会社 Image processing apparatus and method for operating endoscope system
US9363446B2 (en) * 2013-04-15 2016-06-07 Htc Corporation Automatic exposure control for sequential images
US9329702B2 (en) * 2013-07-05 2016-05-03 Pixart Imaging Inc. Navigational device with adjustable tracking parameter

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5936612A (en) * 1997-05-30 1999-08-10 Wang; Yanqing Computer input device and method for 3-D direct manipulation of graphic objects
US6844871B1 (en) * 1999-11-05 2005-01-18 Microsoft Corporation Method and apparatus for computer input using six degrees of freedom
US20040032392A1 (en) * 2002-08-19 2004-02-19 Topseed Technology Corp. Mouse pen device having remote-control function
US20050116933A1 (en) * 2003-12-02 2005-06-02 Hsun-Li Huang Dual mode computer mouse
US20080266251A1 (en) * 2007-04-24 2008-10-30 Pixart Imaging Inc. Cursor control device and method for an image display, and image system
US20090153482A1 (en) * 2007-12-12 2009-06-18 Weinberg Marc S Computer input device with inertial instruments
US20090153486A1 (en) * 2007-12-18 2009-06-18 Microsoft Corporation Optical mouse with limited wavelength optics
US20110013001A1 (en) * 2008-01-28 2011-01-20 Thomas Craven-Bartle Digital pens and a method for digital recording of information
US20120020529A1 (en) * 2010-07-23 2012-01-26 Pixart Imaging Inc. Displacement estimation method and displacement estimation device using the same
WO2014009933A1 (en) * 2012-07-12 2014-01-16 Grant Neville Odgers Improvements in devices for use with computers
US20150022449A1 (en) * 2013-07-16 2015-01-22 Amchael Visual Technology Corporation Algorithms, software and an interaction system that support the operation of an on the fly mouse

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9632657B2 (en) * 2014-12-28 2017-04-25 Sap Se Auxiliary input device
US11307308B2 (en) * 2017-06-02 2022-04-19 Pixart Imaging Inc. Tracking device and electronic device with improved work surface adaptability
US20220206151A1 (en) * 2017-06-02 2022-06-30 Pixart Imaging Inc. Tracking device with improved work surface adaptability
US11808853B2 (en) * 2017-06-02 2023-11-07 Pixart Imaging Inc. Tracking device with improved work surface adaptability
US11435837B2 (en) * 2020-03-09 2022-09-06 Dexin Electronic Ltd. Dual mode mouse device

Also Published As

Publication number Publication date
TWI536211B (en) 2016-06-01
US11048342B2 (en) 2021-06-29
US20180203530A1 (en) 2018-07-19
CN104808938A (en) 2015-07-29
US9958965B2 (en) 2018-05-01
US10558279B2 (en) 2020-02-11
CN110083299A (en) 2019-08-02
CN104808938B (en) 2019-03-15
US20200142510A1 (en) 2020-05-07
US20190138119A1 (en) 2019-05-09
TW201530373A (en) 2015-08-01
US20170249027A1 (en) 2017-08-31
US10209787B2 (en) 2019-02-19

Similar Documents

Publication Publication Date Title
US11048342B2 (en) Dual mode optical navigation device
US10528153B2 (en) Keyboard with touch sensitive element
US8666115B2 (en) Computer vision gesture based control of a device
TWI450159B (en) Optical touch device, passive touch system and its input detection method
CN103744542B (en) Hybrid pointing device
US20110095983A1 (en) Optical input device and image system
US9727148B2 (en) Navigation device and image display system with inertial mode
KR100843590B1 (en) Optical pointing apparatus and mobile terminal having the same
US9141234B2 (en) Pressure and position sensing pointing devices and methods
US20130229349A1 (en) Optical touch input by gesture detection from varying images
US10969883B2 (en) Optical navigation device and system with changeable smoothing
US20100207885A1 (en) Optical input device and operating method thereof, and image system
TWI697827B (en) Control system and control method thereof
TWI408574B (en) Shape changeable camera mouse
CN107402654B (en) Touch detection method and touch detection system
TWI489344B (en) Touch control method and touch control apparatus
KR20090122853A (en) Hand mouse
JP5353196B2 (en) Moving measuring device and moving measuring program
KR20140126440A (en) Method of ordering using input device
KR20140072666A (en) Method for interfacing with pen and projection system using the method

Legal Events

Date Code Title Description
AS Assignment

Owner name: PIXART IMAGING INC., TAIWAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:LEE, SAI-MUN;REEL/FRAME:032131/0894

Effective date: 20130725

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION