US20080278455A1 - User-Defined Enablement Protocol - Google Patents
User-Defined Enablement Protocol Download PDFInfo
- Publication number
- US20080278455A1 US20080278455A1 US12/118,047 US11804708A US2008278455A1 US 20080278455 A1 US20080278455 A1 US 20080278455A1 US 11804708 A US11804708 A US 11804708A US 2008278455 A1 US2008278455 A1 US 2008278455A1
- Authority
- US
- United States
- Prior art keywords
- user
- touch
- enablement
- enablement protocol
- function
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0416—Control or interface arrangements specially adapted for digitisers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
Definitions
- the present disclosure relates to user-defined enablement protocols for electronic devices.
- the disclosure has been developed primarily for providing an enhanced security means for enabling or executing functions of electronic devices through an interface such as a touch-sensitive display, and will be described hereinafter with reference to this application. However it will be appreciated that the disclosure is not limited to this particular field of use.
- Electronic devices often require some form of enablement protocol to activate them or to enable a particular function. For example to activate a mobile phone or to log on to a computer system a user may need to enter an alphanumeric code e.g. ‘123’.
- an alphanumeric code e.g. ‘123’.
- a known drawback here is that if an unauthorised person learns the alphanumeric code, it is simple for them to activate the mobile phone or log in to the computer system. If the device is equipped with a touch screen so that the alphanumeric keys can be shown on a display, an additional level of security is available in that the locations of the keys can be scrambled either before or during entry of the code, making it more difficult for another person to learn the code by observation. However once the code is known, it is once again simple for an unauthorised person to enter it.
- US Patent Publication No. 2008/00823015 discloses a non-alphanumeric method for unlocking an electronic device whereby direction keys are used to traverse a path through a displayed maze, and the device is unlocked if the path matches a path preset by the user. Once again, there is no security once the unlock code (i.e. the path) is known, and the restriction to using direction keys limits the range and complexity of possible enablement protocols.
- touch screens offer a much greater range of enablement protocols for an electronic device.
- a user can also provide ‘gestural input’ in the form of one or more swipes or taps for example.
- US Patent Publication No. 2007/0150842 discloses a method for unlocking an electronic device whereby a person touches a displayed image and moves it along an indicated path. Besides unlocking a device, many other functions can be enabled with gestures on a touch screen.
- 2006/0026535 discloses various mechanisms for interacting with one or more GUI (Graphical User Interface) elements based on a user interface mode and in response to one or more detected touches.
- the graphical elements that may be activated include a virtual scroll wheel, a virtual keyboard, a toolbar and a control panel.
- Functions applied include rotating, translating or re-sizing a displayed image, and scrolling through a list.
- U.S. Pat. Nos. 5,825,352 and 5,943,043 disclose various two finger gestural inputs that emulate mouse functions
- US Patent Publication No. 2007/0177803 discloses the possibility of allowing a user to assign an arbitrary function to a predetermined gesture. However in all these cases the user is either guided by displayed material or the gesture is predetermined, so there is limited opportunity for altering the functionality provided by the device or the mechanisms by which those functions are enabled.
- the present disclosure overcomes or ameliorates at least one of the disadvantages of the prior art, and provides a useful alternative.
- the present disclosure provides a user interface method, comprising the steps of:
- defining the enablement protocol for a function comprises user-engagement with an interface, which comprises rises one or more touches or near touches of a touch screen.
- an enablement protocol by engaging other types or configurations of interfaces, for example a rectangular frame surrounding a space wherein adjacent borders of the frame include optical emitters and the opposing borders include photo-detectors, i.e. essentially a touch-sensitive screen apparatus minus the actual touch surface.
- the enablement protocol may comprise voice recognition of pre-defined commands (audible instructions), or the tracking of eye movements, or some other type of gesture of one or more parts of the human body.
- the user will preferably define or record that protocol in the same form.
- the enablement protocol is a sequence of touches with a particular rhythm at the top left corner of a touch screen
- the user will usually define or record that protocol by actually performing the sequence of touches.
- the user could define or record the protocol by other means, such as a spoken or typed description.
- the present disclosure provides user interface method, comprising the steps of:
- an enablement protocol for a function of an electronic device said enablement protocol comprising at least one user-defined touch on a touch-sensitive area operatively associated with said electronic device; and ii. recording and retaining said enablement protocol of said function, such that a user enables said function by substantially reproducing said enablement protocol in the absence of spatial or temporal indication of at least a portion of said enablement protocol.
- the enablement protocol which comprises a plurality of spatial and temporal data points (i.e. touches and/or gestures) may be selected from one or more of the following non-exhaustive list of spatial or temporal indications: a simple touch, movement of touch, direction of touch, timing or rhythm between touches, force of touch, number of sequential or simultaneous touches, position of touch or shape or change of shape of touch, etc.
- the disclosure provides a method for user-defined enablement of a function through an interface, said method comprising the steps of:
- the function can be enabled by one or more touches, and that the user is required to provide the enablement protocol without spatial or temporal indication of where the enablement protocol should be received to execute or enable the function.
- the enablement protocol may comprise a plurality of user-defined touches of the interface, and at least one spatial or temporal indication of the plurality of the user-defined touches of the enablement protocol is not shown or displayed on the interface.
- the enablement protocol may comprise a swipe of a touch screen with a 1 second dwell time at the conclusion of the swipe. The screen may only show where the swipe should be received and not indicate that the user must hold their finger down at the end of the swipe for 1 second. In this case only one of the two spatial/temporal indications is shown to the user.
- the function may be enabled when the user input meets a predefined number of spatial or temporal indications.
- the enablement protocol may comprise a partial swipe of a touch screen, and then a 1 second dwell time, and then a further partial swipe, i.e. swipe+dwell+swipe.
- the function may be enabled if only two out of these three inputs are received, e.g. swipe+dwell, or swipe+swipe, or dwell+swipe.
- the present disclosure provides a method for enablement of a function through an interface, said method comprising the steps of:
- the user may be presented with a choice of suggested enablement protocols from which they can choose and substantially mimic, thereby recording their own specific and unique enablement protocol.
- the user can then assign a function to their unique recorded enablement protocol and execute or enable the function when they input an enablement protocol that substantially coincides with the recorded enablement protocol, wherein the interface is without spatial or temporal indication of where at least a portion of the enablement protocol is to be received to execute or enable the function.
- the present disclosure provides a method for userdefined enablement of a function through an interface, said method comprising the steps of:
- the enablement protocol effectively acts as a ‘gatekeeper’ or security password control over a particular function.
- the user records an enablement protocol, assigns a function to the enablement protocol and then when wishing to execute or enable the function is prompted to input their unique enablement protocol which, if coinciding with the recorded enablement protocol, allows the function to be executed or enabled.
- Prior art devices typically require touches or near touches within one or more pre-existing or specified areas to enable a particular pre-defined function. For instance, a simple keypad on a touch display provides alphanumeric material at specific predefined locations, and touches or near touches at these locations must be accomplished in a specific sequence to enable the function, e.g. to activate a device containing the touch display.
- the user-defined interface method according to the present disclosure has no such spatial or temporal indication to a user of where touches or near touches should be received to enable the function, thereby allowing a user to have complete freedom in regard to defining how the functionality of an electronic device associated with the interface is enabled by touch.
- a simple embodiment may involve the activation of a personal electronic device such as PDA or mobile phone.
- a security code (optionally user-defined) is required to enable the device.
- This security code is normally a series of alphanumeric keystrokes either on a keypad separate from the display or on the display itself.
- the present disclosure envisages no such spatial or temporal indication to a user. Rather a user can define a security code or ‘enablement protocol’ on the device or indeed on an apparatus separate from but operatively associated with the device by defining one or more touches to enable the aforementioned function.
- One such ‘touch’ may be for instance the touch of a thumb in the upper right hand corner of the screen or other touch sensitive area, or both.
- several touches at user-defined locations on the screen may enable the device.
- a sequence of touches either at the user-defined locations or indeed anywhere on the screen can enable the functionality required. The timing or protocol (temporal sequence) of these touches must then be recreated to enable said function.
- both the spatial and temporal nature of the touches are examined to ensure that they meet the user-defined enablement protocol.
- different types of touching bodies could enable different functions by performing otherwise identical enablement protocols. For example a swipe down one side of a touch screen with a stylus enables function A, while the same swipe with a finger enables function B.
- This embodiment expands the range of functions that can be performed with a limited range of gestures, or alternatively provides an additional security feature.
- any ‘touch screen’ input device can have a series of functions which may only be enabled by the user who defined the aforementioned ‘enablement protocol’, since there is no spatial or temporal indication to assist another party who may wish to enable the aforementioned functions.
- enablement protocols may be linked to the profiles of multiple users such that a single device may be enabled according to the profile of a user who defined a specific enablement protocol.
- user A defines touches in the upper and lower left hand corners of a display to initiate a device.
- touches are provided in those upper and lower corners the device is activated according to user A's profile.
- User B may have an entirely separate profile and may wish the device to be set up according to their specific needs. They therefore define an enablement protocol different from that of user A; not only will other parties not be able to access the device without knowing the enablement protocol of either party, but once user B initiates their particular enablement protocol the device will be activated according to their specific profile.
- an enablement protocol such as the locations to be touched or the particular gesture
- other aspects such as the rhythm, pattern or sequence of the touches or their touch signature e.g. finger shape, will be different and therefore they will not be able to replicate the enablement protocol.
- the disclosure as described is particularly useful for users with some form of physical handicap that may prevent them from using conventional touch screen devices.
- Touches and near touches may be detected in several ways, including a touch-down, movement of touch, direction of touch, timing or rhythm (protocol) between touches i.e. temporal sequence, force of touch, number of sequential or simultaneous touches, position of touch, and shape or change of shape of touch approaching, contacting or leaving the touch sensitive area.
- One or more of these detection mechanisms may be used to complete the enablement protocol. For instance, if the touch location is correct but the shape of the touch is incorrect the protocol will not be satisfied (i.e. there is not substantial coincidence between the received and the recorded protocol) and the associated function will not be enabled. Similarly if the timing or rhythm between touches is correct but the force applied by those touches is incorrect, once again the protocol will not be satisfied. It will be appreciated that although only selected touch technologies can directly measure the force of an applied touch, certain other touch technologies, especially those with high spatial resolution, can indirectly measure touch force from deformation of the touch object.
- Touches or near touches may be detected by any known touch sensing technology including analogue resistive, surface capacitive, surface acoustic wave, projected capacitive, optical and bending wave technologies, bearing in mind that some touch technologies are better suited than others for distinguishing multiple simultaneous touches, should these be part of an enablement protocol.
- Enablement protocols that include touch shape or change of shape will generally require high spatial resolution, such as may be offered by optical touch detection technologies disclosed in U.S. Pat. Nos. 5,914,709 and 7,352,940.
- any known touch technology including those mentioned above, can be applied with the present disclosure, with enablement protocols defined within the limitations of that technology.
- Certain touch technologies e.g.
- resistive and bending wave require a direct touch for activation, whereas other touch technologies, such as optical touch where a sheet of light is established in front of a display, can also sense an object hovering close to the display.
- touch sensitive device is activated by a touch or a near touch of an interface.
- the present disclosure provides a user interface device comprising: a touch sensitive area adapted to detect touches; a display device to display one or more functions; and a storage medium including a user-defined enablement protocol for enabling functions operatively associated with said display device, said enablement protocol including one or more user-defined touches on or near said touchsensitive area, there being no spatial or temporal indication to said user of at least a portion of said enablement protocol.
- the enablement protocol be performed on the display device, this is not absolutely necessary.
- a user-defined interface method comprising the steps of:
- said enablement protocol comprises one or more user-defined touches, ii. retaining said enablement protocol of said function, and iii. enabling said function by said one or more user-defined touches wherein, if a user is provided with spatial or temporal indication for a portion of said enablement protocol, said enablement protocol additionally requires detection of an additional touch parameter.
- the additional parameter required for the enablement protocol would include one or more of touch movement, direction of touch, timing or rhythm between touches i.e. temporal sequence, force of touch, number of sequential or simultaneous touches, shape and/or change of shape of touch including in different planes e.g. vertical and horizontal, on approach or withdrawal.
- some devices may have a spatial or temporal indication for a portion of an enablement protocol.
- a mobile phone may have locations shown on the display that must be pressed in a certain sequence for partial enablement of the protocol.
- One mechanism to complete the enablement protocol could be temporal i.e. the locations must be pressed in the precise timing and sequence required.
- the size and shape of the touch may be detected to confirm enablement by an authorised user. This detection of touch size and shape may be accomplished not only in the plane of the display or touch sensitive area but also in planes above the display e.g. the shape of the touch approaching or leaving the touch sensitive area, if the touch sensor can detect near touches. This touch signature is virtually impossible to replicate by another user and therefore provides a secure individual enablement protocol. It will be clear to a person skilled in the art that there are various other options available to confirm enablement.
- the term ‘coincide’ when used herein and when referring to a touch input being compared with a recorded or retained touch input, is meant to define that the inputs should substantially correspond.
- the coincidence is substantial overlap in overall touch area, and in another embodiment may also include one or more other parameters, such as dwell time.
- ‘coincide’ may refer to a predetermined number of matches between the spatial and/or temporal data points of the touch or near touch and those of the enablement protocol.
- touch herein the term is also intended to encompass within its scope the term ‘near touch’.
- the touch-sensitive area is larger than the area of the display device and contains the display device within its borders.
- conventional touch-sensitive input devices include a display device that is at least as large as the touch-sensitive area.
- the touch-sensitive area is larger than the display itself and includes the display device within its borders. This is highly unusual but provides advantages over conventional systems, for example the opportunity of producing a smaller display with consequent power reduction. Also, by including the touch-sensitive area outside the display, the display itself is not obscured while input is applied via the touch-sensitive area.
- the touch portions/locations may be located on a first face of the touch-sensitive area, generally the display side, and/or a second opposite face of the touch-sensitive area. In one particular embodiment the touch portions/locations for enablement of the functions are arranged on the first and second opposite faces of the touch-sensitive area in such a way that they can be simultaneously operated by the thumb and fingers of a user.
- FIGS. 1 and 2 illustrate conventional user interface devices according to the prior art
- FIGS. 3A-3D illustrate a first embodiment of the present disclosure
- FIGS. 4A-4C illustrate a second embodiment of the present disclosure
- FIGS. 5A-5C illustrate a third embodiment of the present disclosure.
- FIGS. 6 and 7 illustrate touch signature recognition according to a fourth embodiment of the present disclosure.
- FIG. 8A-8C illustrate a fifth embodiment of the present disclosure.
- FIGS. 9A-9C illustrate a sixth embodiment of the present disclosure.
- FIGS. 1 and 2 provide illustrations of prior art functionality in conventional devices.
- FIG. 1 shows a mobile phone 10 with an alphanumeric keypad 11 and a display 12 .
- the mobile phone can only be enabled when a user-defined security code is entered via the alphanumeric keypad 11 .
- the alphanumeric keypad may be provided on a touch-sensitive display (i.e. a touch screen) but once again the conventional device comprises a pre-defined arrangement of the enablement protocol such that if an unauthorised user were aware of the code e.g. 123, they could enable the device simply by engaging the predefined locations for numerals 1, 2 and 3.
- FIG. 2 is an illustration of an automatic teller machine 20 with a touch sensitive display 21 .
- This technology is similar to the mobile phone in FIG. 1 , and in this instance various functions and spatial locations for enabling those functions are clearly defined on the display.
- the display includes indicia 22 describing the functions e.g. withdrawal, deposit and account balance, and the spatial locations for those functions are defined by the touch portions 23 . Accordingly any user can engage the functions defined on the display 21 by simply following the spatial arrangement already predefined by the device.
- the enablement protocol of the present disclosure acts to restrict the performing of a user-instigated function. That is, a user attempts to perform (whether by touch or otherwise) a function on an electronic device, and the function is only carried out if the user performs their defined enablement protocol required for that function. Generally this enablement protocol will be performed in response to an instruction, an indication or a request for user input from the electronic device, similar to a conventional ‘password’ dialog box.
- Each user of the device may have defined a single enablement protocol that enables all functions of the device, or they may have defined different enablement protocols for different functions.
- a computer system may provide different levels of access to different classes of user, with one class of users having full access to a given file, a second class of users having read-only access, and other users barred or restricted access.
- the enablement protocol is embedded in the particular touch input, such as a gesture, that instigates the function.
- This instigating touch input may be predetermined by the electronic device or by software running on it (e.g. a ‘double tap’ gesture that opens a file), or it may be user-defined (e.g. tracing a person's initials to instigate a phone call to that person).
- a user may attempt to open a file by double tapping on an icon, but the file will only be opened if the ‘details’ of the double tap gesture, such as the touch object size, the duration of each tap and the delay between the taps, substantially coincide with the pre-recorded gesture (user-defined protocol).
- a user may assign one particular function to a swipe gesture when the gesture is performed with a finger, and a different function when the swipe is performed with a stylus.
- FIGS. 3A-3D illustrate a first embodiment according to the present disclosure.
- FIG. 3A shows a particular device 30 with a touch screen or touch-sensitive area 31 .
- a user has defined the location of a touch or near touch at a location 32 in accordance with an enablement protocol for activation of the device.
- any function of a device 30 may similarly include an enablement protocol comprising one or more user defined touches or near touches 32 .
- the required touch location 32 i.e. the enablement protocol, is not shown on the device at all.
- the dotted line 33 indicates the location for illustrative purposes only, and there is no spatial or temporal indication whatsoever on the device 30 of where or how the touches should be applied to satisfy the user-defined enablement protocol. In fact there may be no display associated with the touch-sensitive area.
- FIG. 3B a user commences application of the enablement protocol, which requires a thumb to be placed at the location 32 .
- FIG. 3C it can be seen that the user places their thumb 34 at the required location 32 , thereby satisfying the enablement protocol that results in activation of the device.
- the enablement protocol comprises a single touch at the user-defined location 32 on the touch-sensitive area 31 of the particular device 30 . It will be clear, however, that the enablement protocol may comprise other forms of touch.
- FIGS. 4A-4C displays a swipe by the user of their thumb to satisfy the enablement protocol.
- This embodiment again comprises a device 40 with a touch-sensitive area 41 .
- the spatial location of the enablement protocol for a certain function is shown by dotted lines 42 again for illustrative purposes only.
- the enablement protocol requires movement of the thumb along the arc 43 .
- the enablement protocol may also include the timing or speed of the swipe from the lower right hand corner to the upper right hand corner, the contact dwell time at the start or end of the swipe, and the shape of the thumb on the surface of the touch-sensitive area. It will be appreciated that this offers an enhanced degree of security in that even if another person knew the correct gesture, say by observation, it would be extremely difficult for them to replicate all other aspects of the enablement protocol.
- FIG. 4B the user holds the device, swipes their thumb 44 across the area 42 in the predefined fashion required by the enablement protocol i.e. speed, timing, area covered etc such that, as shown in FIG. 4C , the device is activated.
- FIGS. 5A-5C a multiple touch enablement protocol is shown.
- a device 50 with a touch sensitive area 51 requires two touches at specific spatial locations to activate the device.
- the spatial location 52 requires a touch of a left thumb 54
- the spatial location 53 requires a swipe of a right thumb 55 along a linear path 56 .
- the screen may be completely blank and the enablement protocol is known only to the authorised user of the device 50 . Once the touch at the location 52 and the swipe at the location 53 are accomplished in line with the protocol, the device 50 is activated.
- the screen may display indicia suggesting where touch input is required to enable a protocol, however the indicia may be specifically configured to falsely indicate where the actual input is required to enable the function, thereby intentionally frustrating the attempts of an unauthorised person to access the electronic device.
- FIGS. 6A-6D and 7 A- 7 D illustrate how a touch is detected and how a characteristic touch signature of an individual may be determined.
- FIG. 6A a finger 60 is approaching the surface 61 of a touch screen.
- FIG. 6B the finger first comes into contact with the surface and is therefore detected.
- the area 62 of the surface covered by the finger is relatively small.
- greater pressure is applied to the surface thereby distorting the finger such that it covers a greater area 63 of the surface, as seen in FIG. 6C .
- 6 D the finger is withdrawn after a certain time that may itself be part of an enablement protocol.
- this change of shape of a finger as it approaches, touches and withdraws from the surface 61 can be detected and may form part of an enablement protocol. This is particularly useful since even if the precise spatial and temporal location of an enablement protocol is known by an unauthorised user, the protocol may only be satisfied if the specific shape and change of shape of the touch matches that of the authorised user.
- FIG. 7 gives a further illustration.
- FIGS. 7A-7D show a finger movement or swipe on a touch-sensitive surface 70 .
- a finger 71 approaches the surface 70
- the finger first contacts the surface.
- the finger then slides across the surface to the point shown in FIG. 7C and is then removed as shown in FIG. 7D .
- the distance 72 covered by the finger may be detected and form part of an enablement protocol.
- the change in shape of the finger as it slides may also be detected. Accordingly it can be seen that virtually any movement, shape or change of shape of a touch object can be detected and form part of an enablement protocol for enabling a particular function on a device.
- FIGS. 8A-8C and 9 A- 9 C disclose a further embodiment of the present disclosure displaying the powerful flexibility and security available with user-defined enablement protocols.
- FIGS. 8A-8C it can be seen that even if spatial or temporal indicators are provided on the device, an enablement protocol will still not be satisfied if it includes detection of additional touch or near touch parameters, e.g. shape or change of shape of a touch object.
- a device 80 with a touch sensitive display 81 shows a spatial indicator 82 forming part of an enablement protocol.
- the enablement protocol for a specific function of this device requires a single touch at the location of the indicator 82 .
- the shape 83 of the touch object also forms part of the enablement protocol as discussed above in relation to FIGS. 6A-6D .
- the resolution is sufficiently high to measure shape and/or change of shape on approach, at touch-down and at withdrawal of touch.
- FIG. 8B a user holds the device 80 and commences the enablement protocol by placing the pad of their thumb 84 on the indicator 82 . This is accomplished in FIG.
- the spatial indicator 82 gives a user some indication of the required enablement protocol
- a touch object does not match the required shape 83
- the device will not be enabled. For example touching the indicator 82 with a stylus or fingertip will not satisfy the enablement protocol.
- a device 90 includes a touch sensitive display 91 showing two spatial and/or temporal indicators 92 and 93 .
- the indicators 92 and 93 may be flashing to indicate the desired temporal sequence between touches at the indicators 92 and 93 .
- the enablement protocol requires a simple touch at the location of the indicator 92 , but not at the location of the indicator 93 . Rather, the protocol requires a swipe along an arc 96 that begins in the lower right hand corner and terminates at the indicator 93 .
- the device is held such that that portion of the enablement protocol requiring a left thumb touch 97 at the location of the indicator 92 is satisfied, and the right thumb 98 proceeds from the lower right hand corner up to the location of the indicator 93 within the time specified by the enablement protocol.
- the enablement protocol enables the function of the device 90 and the device is, in this embodiment, activated.
- an unauthorised person attempting to access the electronic device could attempt to replicate the enablement protocol of an authorised person by carefully inspecting the touch surface for residual finger marks.
- the enablement protocol may be selected to avoid indicating, from an inspection of finger marks alone, the actual enablement protocol. For example, a circular enablement protocol would not indicate the direction which the user traced out the circle. Similarly, the direction of a swipe would be difficult to interpret from only residual finger marks, and enablement protocols with temporal features such as the rhythm of a touch sequence could not be ascertained from residual finger marks.
Abstract
Description
- This application claims priority under the Paris Convention to Australian Provisional Patent Application No. 2007902519 filed on May 11, 2007, and Australian Provisional Patent Application No. 2007902517 filed on May 11, 2007, the contents of both of which are incorporated herein by reference.
- The present disclosure relates to user-defined enablement protocols for electronic devices. The disclosure has been developed primarily for providing an enhanced security means for enabling or executing functions of electronic devices through an interface such as a touch-sensitive display, and will be described hereinafter with reference to this application. However it will be appreciated that the disclosure is not limited to this particular field of use.
- Any discussion of the prior art throughout the specification should in no way be considered as an admission that such prior art is widely known or forms part of the common general knowledge in the field.
- Electronic devices often require some form of enablement protocol to activate them or to enable a particular function. For example to activate a mobile phone or to log on to a computer system a user may need to enter an alphanumeric code e.g. ‘123’. A known drawback here is that if an unauthorised person learns the alphanumeric code, it is simple for them to activate the mobile phone or log in to the computer system. If the device is equipped with a touch screen so that the alphanumeric keys can be shown on a display, an additional level of security is available in that the locations of the keys can be scrambled either before or during entry of the code, making it more difficult for another person to learn the code by observation. However once the code is known, it is once again simple for an unauthorised person to enter it. US Patent Publication No. 2008/00823015 discloses a non-alphanumeric method for unlocking an electronic device whereby direction keys are used to traverse a path through a displayed maze, and the device is unlocked if the path matches a path preset by the user. Once again, there is no security once the unlock code (i.e. the path) is known, and the restriction to using direction keys limits the range and complexity of possible enablement protocols.
- Compared with a keyboard with hard keys, touch screens offer a much greater range of enablement protocols for an electronic device. Apart from the well-known ‘touch to select’ operation mode where a user selects or enables a function by touching a displayed icon, a user can also provide ‘gestural input’ in the form of one or more swipes or taps for example. US Patent Publication No. 2007/0150842 for instance discloses a method for unlocking an electronic device whereby a person touches a displayed image and moves it along an indicated path. Besides unlocking a device, many other functions can be enabled with gestures on a touch screen. US Patent Publication No. 2006/0026535 for instance discloses various mechanisms for interacting with one or more GUI (Graphical User Interface) elements based on a user interface mode and in response to one or more detected touches. The graphical elements that may be activated include a virtual scroll wheel, a virtual keyboard, a toolbar and a control panel. Functions applied include rotating, translating or re-sizing a displayed image, and scrolling through a list. U.S. Pat. Nos. 5,825,352 and 5,943,043 disclose various two finger gestural inputs that emulate mouse functions, while US Patent Publication No. 2007/0177803 discloses the possibility of allowing a user to assign an arbitrary function to a predetermined gesture. However in all these cases the user is either guided by displayed material or the gesture is predetermined, so there is limited opportunity for altering the functionality provided by the device or the mechanisms by which those functions are enabled.
- The present disclosure overcomes or ameliorates at least one of the disadvantages of the prior art, and provides a useful alternative.
- In a first broad aspect, the present disclosure provides a user interface method, comprising the steps of:
- i. defining an enablement protocol for a function; and
ii. recording and retaining said enablement protocol of said function, such that a user enables said function by substantially reproducing said enablement protocol in the absence of spatial or temporal indication of at least a portion of said enablement protocol. - In one embodiment, defining the enablement protocol for a function comprises user-engagement with an interface, which comprises rises one or more touches or near touches of a touch screen. However, it will be appreciated that a user may also define an enablement protocol by engaging other types or configurations of interfaces, for example a rectangular frame surrounding a space wherein adjacent borders of the frame include optical emitters and the opposing borders include photo-detectors, i.e. essentially a touch-sensitive screen apparatus minus the actual touch surface. Alternatively, or additionally, the enablement protocol may comprise voice recognition of pre-defined commands (audible instructions), or the tracking of eye movements, or some other type of gesture of one or more parts of the human body. Whatever form a user's enablement protocol takes, the user will preferably define or record that protocol in the same form. For example if the enablement protocol is a sequence of touches with a particular rhythm at the top left corner of a touch screen, the user will usually define or record that protocol by actually performing the sequence of touches. Alternatively the user could define or record the protocol by other means, such as a spoken or typed description.
- In a second aspect, the present disclosure provides user interface method, comprising the steps of:
- i. defining an enablement protocol for a function of an electronic device, said enablement protocol comprising at least one user-defined touch on a touch-sensitive area operatively associated with said electronic device; and
ii. recording and retaining said enablement protocol of said function, such that a user enables said function by substantially reproducing said enablement protocol in the absence of spatial or temporal indication of at least a portion of said enablement protocol. - It will be appreciated that the enablement protocol which comprises a plurality of spatial and temporal data points (i.e. touches and/or gestures) may be selected from one or more of the following non-exhaustive list of spatial or temporal indications: a simple touch, movement of touch, direction of touch, timing or rhythm between touches, force of touch, number of sequential or simultaneous touches, position of touch or shape or change of shape of touch, etc.
- According to a third aspect the disclosure provides a method for user-defined enablement of a function through an interface, said method comprising the steps of:
- i. recording an enablement protocol comprising at least one user-defined touch of said interface;
ii. assigning said function to said enablement protocol; and
iii. executing or enabling said function when one or more touches received on or near said interface substantially coincide with the recorded enablement protocol, wherein said interface is without spatial or temporal indication of where at least a portion of said enablement protocol is to be received to execute or enable said function. - It will be appreciated that the function can be enabled by one or more touches, and that the user is required to provide the enablement protocol without spatial or temporal indication of where the enablement protocol should be received to execute or enable the function. In one aspect, the enablement protocol may comprise a plurality of user-defined touches of the interface, and at least one spatial or temporal indication of the plurality of the user-defined touches of the enablement protocol is not shown or displayed on the interface. In such embodiments, for example, the enablement protocol may comprise a swipe of a touch screen with a 1 second dwell time at the conclusion of the swipe. The screen may only show where the swipe should be received and not indicate that the user must hold their finger down at the end of the swipe for 1 second. In this case only one of the two spatial/temporal indications is shown to the user.
- In another somewhat related aspect, the function may be enabled when the user input meets a predefined number of spatial or temporal indications. For example, the enablement protocol may comprise a partial swipe of a touch screen, and then a 1 second dwell time, and then a further partial swipe, i.e. swipe+dwell+swipe. The function may be enabled if only two out of these three inputs are received, e.g. swipe+dwell, or swipe+swipe, or dwell+swipe.
- According to a fourth aspect the present disclosure provides a method for enablement of a function through an interface, said method comprising the steps of:
- i. recording an enablement protocol substantially according to a pre-defined enablement protocol, wherein said enablement protocol includes at least one user-defined touch of said interface;
ii. assigning said function to the recorded enablement protocol; and
iii. executing or enabling said function when one or more touches received on or near said interface substantially coincide with said recorded enablement protocol, wherein said interface is without spatial or temporal indication of where at least a portion of said enablement protocol is to be received to execute or enable said function. - In this aspect, the user may be presented with a choice of suggested enablement protocols from which they can choose and substantially mimic, thereby recording their own specific and unique enablement protocol. The user can then assign a function to their unique recorded enablement protocol and execute or enable the function when they input an enablement protocol that substantially coincides with the recorded enablement protocol, wherein the interface is without spatial or temporal indication of where at least a portion of the enablement protocol is to be received to execute or enable the function.
- According to a fifth aspect the present disclosure provides a method for userdefined enablement of a function through an interface, said method comprising the steps of:
- i. recording an enablement protocol comprising one or more user-defined touches of said interface; and
ii. assigning said function to said enablement protocol, wherein in response to a user's request for execution or enablement of said function the user is prompted to perform said enablement protocol, wherein said function is executed or enabled when one or more touches received on or near said interface substantially coincide with the recorded enablement protocol, wherein said interface is without spatial or temporal indication of where at least a portion of said enablement protocol is to be received to execute or enable said function. - In this aspect the enablement protocol effectively acts as a ‘gatekeeper’ or security password control over a particular function. Essentially, the user records an enablement protocol, assigns a function to the enablement protocol and then when wishing to execute or enable the function is prompted to input their unique enablement protocol which, if coinciding with the recorded enablement protocol, allows the function to be executed or enabled.
- Prior art devices typically require touches or near touches within one or more pre-existing or specified areas to enable a particular pre-defined function. For instance, a simple keypad on a touch display provides alphanumeric material at specific predefined locations, and touches or near touches at these locations must be accomplished in a specific sequence to enable the function, e.g. to activate a device containing the touch display. However, the user-defined interface method according to the present disclosure has no such spatial or temporal indication to a user of where touches or near touches should be received to enable the function, thereby allowing a user to have complete freedom in regard to defining how the functionality of an electronic device associated with the interface is enabled by touch.
- To explain, a simple embodiment may involve the activation of a personal electronic device such as PDA or mobile phone. Normally a security code (optionally user-defined) is required to enable the device. This security code is normally a series of alphanumeric keystrokes either on a keypad separate from the display or on the display itself.
- In one embodiment the present disclosure envisages no such spatial or temporal indication to a user. Rather a user can define a security code or ‘enablement protocol’ on the device or indeed on an apparatus separate from but operatively associated with the device by defining one or more touches to enable the aforementioned function. One such ‘touch’ may be for instance the touch of a thumb in the upper right hand corner of the screen or other touch sensitive area, or both. Alternatively, several touches at user-defined locations on the screen may enable the device. In another embodiment, a sequence of touches either at the user-defined locations or indeed anywhere on the screen can enable the functionality required. The timing or protocol (temporal sequence) of these touches must then be recreated to enable said function. In more complex arrangements both the spatial and temporal nature of the touches are examined to ensure that they meet the user-defined enablement protocol. In further arrangements, different types of touching bodies could enable different functions by performing otherwise identical enablement protocols. For example a swipe down one side of a touch screen with a stylus enables function A, while the same swipe with a finger enables function B. This embodiment expands the range of functions that can be performed with a limited range of gestures, or alternatively provides an additional security feature.
- It can be seen then that any ‘touch screen’ input device can have a series of functions which may only be enabled by the user who defined the aforementioned ‘enablement protocol’, since there is no spatial or temporal indication to assist another party who may wish to enable the aforementioned functions.
- In another embodiment, enablement protocols may be linked to the profiles of multiple users such that a single device may be enabled according to the profile of a user who defined a specific enablement protocol.
- As an example, user A defines touches in the upper and lower left hand corners of a display to initiate a device. When touches are provided in those upper and lower corners the device is activated according to user A's profile. User B may have an entirely separate profile and may wish the device to be set up according to their specific needs. They therefore define an enablement protocol different from that of user A; not only will other parties not be able to access the device without knowing the enablement protocol of either party, but once user B initiates their particular enablement protocol the device will be activated according to their specific profile.
- Indeed even if another person knew some aspects of an enablement protocol, such as the locations to be touched or the particular gesture, other aspects such as the rhythm, pattern or sequence of the touches or their touch signature e.g. finger shape, will be different and therefore they will not be able to replicate the enablement protocol.
- The disclosure as described is particularly useful for users with some form of physical handicap that may prevent them from using conventional touch screen devices. By providing complete freedom to an individual user to define enablement protocols for the various functions of a device, they may define the touches in terms of spatial and temporal parameters that they can easily replicate.
- Since there is no spatial or temporal indication of the enablement protocol to a user, even a simple single touch or near touch at a specific location may be sufficient as an enablement protocol with enhanced security. This is to be distinguished from known fingerprint readers, which require a direct touch within a specific area.
- While enablement by a single touch or near touch is possible with the present disclosure it is generally preferred that two or more touches or near touches be used to define the enablement protocol. Touches and near touches may be detected in several ways, including a touch-down, movement of touch, direction of touch, timing or rhythm (protocol) between touches i.e. temporal sequence, force of touch, number of sequential or simultaneous touches, position of touch, and shape or change of shape of touch approaching, contacting or leaving the touch sensitive area.
- One or more of these detection mechanisms may be used to complete the enablement protocol. For instance, if the touch location is correct but the shape of the touch is incorrect the protocol will not be satisfied (i.e. there is not substantial coincidence between the received and the recorded protocol) and the associated function will not be enabled. Similarly if the timing or rhythm between touches is correct but the force applied by those touches is incorrect, once again the protocol will not be satisfied. It will be appreciated that although only selected touch technologies can directly measure the force of an applied touch, certain other touch technologies, especially those with high spatial resolution, can indirectly measure touch force from deformation of the touch object.
- Touches or near touches may be detected by any known touch sensing technology including analogue resistive, surface capacitive, surface acoustic wave, projected capacitive, optical and bending wave technologies, bearing in mind that some touch technologies are better suited than others for distinguishing multiple simultaneous touches, should these be part of an enablement protocol. Enablement protocols that include touch shape or change of shape will generally require high spatial resolution, such as may be offered by optical touch detection technologies disclosed in U.S. Pat. Nos. 5,914,709 and 7,352,940. However any known touch technology, including those mentioned above, can be applied with the present disclosure, with enablement protocols defined within the limitations of that technology. Certain touch technologies, e.g. resistive and bending wave, require a direct touch for activation, whereas other touch technologies, such as optical touch where a sheet of light is established in front of a display, can also sense an object hovering close to the display. The principles of the present disclosure apply whether the touch sensitive device is activated by a touch or a near touch of an interface.
- In a further aspect the present disclosure provides a user interface device comprising: a touch sensitive area adapted to detect touches; a display device to display one or more functions; and a storage medium including a user-defined enablement protocol for enabling functions operatively associated with said display device, said enablement protocol including one or more user-defined touches on or near said touchsensitive area, there being no spatial or temporal indication to said user of at least a portion of said enablement protocol.
- While it is preferred that the enablement protocol be performed on the display device, this is not absolutely necessary. One could envisage the enablement protocol being applied to a touch-sensitive area or ‘protocol apparatus’ separate from the display device. The user would simply follow the defined enablement protocol and the ‘protocol apparatus’ then communicates with the display device e.g. PDA or computer to enable the function held by that device. Indeed the function itself may be held by a third device, such that instigation of the enablement protocol, the display device and the various functions to be applied to the display device are all held separately.
- In another aspect of the present disclosure there is provided a user-defined interface method, comprising the steps of:
- i. defining for an electronic device an enablement protocol for a function of said device whereby said enablement protocol comprises one or more user-defined touches,
ii. retaining said enablement protocol of said function, and
iii. enabling said function by said one or more user-defined touches wherein, if a user is provided with spatial or temporal indication for a portion of said enablement protocol, said enablement protocol additionally requires detection of an additional touch parameter. - Preferably the additional parameter required for the enablement protocol would include one or more of touch movement, direction of touch, timing or rhythm between touches i.e. temporal sequence, force of touch, number of sequential or simultaneous touches, shape and/or change of shape of touch including in different planes e.g. vertical and horizontal, on approach or withdrawal.
- To explain, some devices may have a spatial or temporal indication for a portion of an enablement protocol. For instance, a mobile phone may have locations shown on the display that must be pressed in a certain sequence for partial enablement of the protocol. One mechanism to complete the enablement protocol could be temporal i.e. the locations must be pressed in the precise timing and sequence required. In one alternative, if spatial and temporal data are provided e.g. by flashing of the locations, the size and shape of the touch may be detected to confirm enablement by an authorised user. This detection of touch size and shape may be accomplished not only in the plane of the display or touch sensitive area but also in planes above the display e.g. the shape of the touch approaching or leaving the touch sensitive area, if the touch sensor can detect near touches. This touch signature is virtually impossible to replicate by another user and therefore provides a secure individual enablement protocol. It will be clear to a person skilled in the art that there are various other options available to confirm enablement.
- Accordingly it can be seen that according to several aspects of the present disclosure there is significant flexibility and security as compared with prior art systems. Unless the context clearly requires otherwise, throughout the description and the claims, the words ‘comprise’, ‘comprising’, and the like are to be construed in an inclusive sense as opposed to an exclusive or exhaustive sense; that is to say, in the sense of ‘including, but not limited to’.
- The skilled person will appreciate that the term ‘coincide’, when used herein and when referring to a touch input being compared with a recorded or retained touch input, is meant to define that the inputs should substantially correspond. In one embodiment, as discussed above, the coincidence is substantial overlap in overall touch area, and in another embodiment may also include one or more other parameters, such as dwell time.
- In a further embodiment ‘coincide’ may refer to a predetermined number of matches between the spatial and/or temporal data points of the touch or near touch and those of the enablement protocol.
- It will be appreciated that when referring to ‘touch’ herein the term is also intended to encompass within its scope the term ‘near touch’.
- In one embodiment the touch-sensitive area is larger than the area of the display device and contains the display device within its borders. Generally, conventional touch-sensitive input devices include a display device that is at least as large as the touch-sensitive area. In one embodiment the touch-sensitive area is larger than the display itself and includes the display device within its borders. This is highly unusual but provides advantages over conventional systems, for example the opportunity of producing a smaller display with consequent power reduction. Also, by including the touch-sensitive area outside the display, the display itself is not obscured while input is applied via the touch-sensitive area. The touch portions/locations may be located on a first face of the touch-sensitive area, generally the display side, and/or a second opposite face of the touch-sensitive area. In one particular embodiment the touch portions/locations for enablement of the functions are arranged on the first and second opposite faces of the touch-sensitive area in such a way that they can be simultaneously operated by the thumb and fingers of a user.
- Preferred embodiments of the disclosure will now be described, by way of example only, with reference to the accompanying drawings in which:
-
FIGS. 1 and 2 illustrate conventional user interface devices according to the prior art; -
FIGS. 3A-3D illustrate a first embodiment of the present disclosure; -
FIGS. 4A-4C illustrate a second embodiment of the present disclosure; -
FIGS. 5A-5C illustrate a third embodiment of the present disclosure.FIGS. 6 and 7 illustrate touch signature recognition according to a fourth embodiment of the present disclosure; -
FIG. 8A-8C illustrate a fifth embodiment of the present disclosure; and -
FIGS. 9A-9C illustrate a sixth embodiment of the present disclosure. -
FIGS. 1 and 2 provide illustrations of prior art functionality in conventional devices. -
FIG. 1 shows amobile phone 10 with analphanumeric keypad 11 and adisplay 12. In many instances the mobile phone can only be enabled when a user-defined security code is entered via thealphanumeric keypad 11. Alternatively the alphanumeric keypad may be provided on a touch-sensitive display (i.e. a touch screen) but once again the conventional device comprises a pre-defined arrangement of the enablement protocol such that if an unauthorised user were aware of the code e.g. 123, they could enable the device simply by engaging the predefined locations fornumerals -
FIG. 2 is an illustration of anautomatic teller machine 20 with a touchsensitive display 21. This technology is similar to the mobile phone inFIG. 1 , and in this instance various functions and spatial locations for enabling those functions are clearly defined on the display. The display includesindicia 22 describing the functions e.g. withdrawal, deposit and account balance, and the spatial locations for those functions are defined by thetouch portions 23. Accordingly any user can engage the functions defined on thedisplay 21 by simply following the spatial arrangement already predefined by the device. - The present disclosure seeks to overcome some of the difficulties associated with the prior art and provides greater flexibility and security as compared the with prior art systems. In certain embodiments the enablement protocol of the present disclosure acts to restrict the performing of a user-instigated function. That is, a user attempts to perform (whether by touch or otherwise) a function on an electronic device, and the function is only carried out if the user performs their defined enablement protocol required for that function. Generally this enablement protocol will be performed in response to an instruction, an indication or a request for user input from the electronic device, similar to a conventional ‘password’ dialog box. Each user of the device may have defined a single enablement protocol that enables all functions of the device, or they may have defined different enablement protocols for different functions. It will be appreciated that the particular function can be virtually anything, such as activating a device from an ‘off’ state or a sleep mode, logging into a computer system, or accessing a file. For example a computer system may provide different levels of access to different classes of user, with one class of users having full access to a given file, a second class of users having read-only access, and other users barred or restricted access.
- In other embodiments that are applicable to functions instigated by touch input, the enablement protocol is embedded in the particular touch input, such as a gesture, that instigates the function. This instigating touch input may be predetermined by the electronic device or by software running on it (e.g. a ‘double tap’ gesture that opens a file), or it may be user-defined (e.g. tracing a person's initials to instigate a phone call to that person). In one example a user may attempt to open a file by double tapping on an icon, but the file will only be opened if the ‘details’ of the double tap gesture, such as the touch object size, the duration of each tap and the delay between the taps, substantially coincide with the pre-recorded gesture (user-defined protocol). In another example a user may assign one particular function to a swipe gesture when the gesture is performed with a finger, and a different function when the swipe is performed with a stylus.
- Some specific embodiments will now be described, involving certain enablement protocols that activate an electronic device. However it will be appreciated that these particular enablement protocols could be associated with other functions of the electronic device.
-
FIGS. 3A-3D illustrate a first embodiment according to the present disclosure.FIG. 3A shows aparticular device 30 with a touch screen or touch-sensitive area 31. A user has defined the location of a touch or near touch at alocation 32 in accordance with an enablement protocol for activation of the device. However, it will be clear from the foregoing description that any function of adevice 30 may similarly include an enablement protocol comprising one or more user defined touches or near touches 32. Importantly, the requiredtouch location 32, i.e. the enablement protocol, is not shown on the device at all. The dottedline 33 indicates the location for illustrative purposes only, and there is no spatial or temporal indication whatsoever on thedevice 30 of where or how the touches should be applied to satisfy the user-defined enablement protocol. In fact there may be no display associated with the touch-sensitive area. - This is clearly a significant advantage over the prior art since it will be virtually impossible for an unauthorised user to determine the enablement protocol without any form of spatial or temporal indication.
- In
FIG. 3B a user commences application of the enablement protocol, which requires a thumb to be placed at thelocation 32. InFIG. 3C it can be seen that the user places theirthumb 34 at the requiredlocation 32, thereby satisfying the enablement protocol that results in activation of the device. - In the embodiment shown in
FIGS. 3A-3D the enablement protocol comprises a single touch at the user-definedlocation 32 on the touch-sensitive area 31 of theparticular device 30. It will be clear, however, that the enablement protocol may comprise other forms of touch. - For instance, the embodiment shown in
FIGS. 4A-4C displays a swipe by the user of their thumb to satisfy the enablement protocol. This embodiment again comprises adevice 40 with a touch-sensitive area 41. The spatial location of the enablement protocol for a certain function is shown by dottedlines 42 again for illustrative purposes only. In this instance, the enablement protocol requires movement of the thumb along thearc 43. The enablement protocol may also include the timing or speed of the swipe from the lower right hand corner to the upper right hand corner, the contact dwell time at the start or end of the swipe, and the shape of the thumb on the surface of the touch-sensitive area. It will be appreciated that this offers an enhanced degree of security in that even if another person knew the correct gesture, say by observation, it would be extremely difficult for them to replicate all other aspects of the enablement protocol. - As shown in
FIG. 4B the user holds the device, swipes theirthumb 44 across thearea 42 in the predefined fashion required by the enablement protocol i.e. speed, timing, area covered etc such that, as shown inFIG. 4C , the device is activated. InFIGS. 5A-5C a multiple touch enablement protocol is shown. In this embodiment adevice 50 with a touchsensitive area 51 requires two touches at specific spatial locations to activate the device. Thespatial location 52 requires a touch of aleft thumb 54, and thespatial location 53 requires a swipe of aright thumb 55 along alinear path 56, Once again, similar toFIGS. 3A-3D andFIGS. 4A-4C , there is no outward indication either spatially or temporally of the enablement protocol. The screen may be completely blank and the enablement protocol is known only to the authorised user of thedevice 50. Once the touch at thelocation 52 and the swipe at thelocation 53 are accomplished in line with the protocol, thedevice 50 is activated. In an alternative embodiment that provides an additional security measure, the screen may display indicia suggesting where touch input is required to enable a protocol, however the indicia may be specifically configured to falsely indicate where the actual input is required to enable the function, thereby intentionally frustrating the attempts of an unauthorised person to access the electronic device. -
FIGS. 6A-6D and 7A-7D illustrate how a touch is detected and how a characteristic touch signature of an individual may be determined. - In
FIG. 6A afinger 60 is approaching thesurface 61 of a touch screen. InFIG. 6B the finger first comes into contact with the surface and is therefore detected. At this point it can be seen that thearea 62 of the surface covered by the finger is relatively small. As the user proceeds with the touch, greater pressure is applied to the surface thereby distorting the finger such that it covers agreater area 63 of the surface, as seen inFIG. 6C . In 6D the finger is withdrawn after a certain time that may itself be part of an enablement protocol. - If the resolution and sensitivity of the touch detection apparatus is sufficient, this change of shape of a finger as it approaches, touches and withdraws from the
surface 61 can be detected and may form part of an enablement protocol. This is particularly useful since even if the precise spatial and temporal location of an enablement protocol is known by an unauthorised user, the protocol may only be satisfied if the specific shape and change of shape of the touch matches that of the authorised user. -
FIG. 7 gives a further illustration.FIGS. 7A-7D show a finger movement or swipe on a touch-sensitive surface 70. InFIG. 7A afinger 71 approaches thesurface 70, and inFIG. 7B the finger first contacts the surface. The finger then slides across the surface to the point shown inFIG. 7C and is then removed as shown inFIG. 7D . Thedistance 72 covered by the finger may be detected and form part of an enablement protocol. Similarly, the change in shape of the finger as it slides may also be detected. Accordingly it can be seen that virtually any movement, shape or change of shape of a touch object can be detected and form part of an enablement protocol for enabling a particular function on a device. Unless an unauthorised user can match with sufficient precision the touch signature of the user who defined the enablement protocol, it is virtually impossible for an unauthorised user to match the enablement protocol and thereby enable the functions of the particular device. Clearly touching with an implement such as a pen will not satisfy an enablement protocol defined by a finger or thumb. -
FIGS. 8A-8C and 9A-9C disclose a further embodiment of the present disclosure displaying the powerful flexibility and security available with user-defined enablement protocols. - In
FIGS. 8A-8C it can be seen that even if spatial or temporal indicators are provided on the device, an enablement protocol will still not be satisfied if it includes detection of additional touch or near touch parameters, e.g. shape or change of shape of a touch object. - In
FIG. 8A adevice 80 with a touchsensitive display 81 shows aspatial indicator 82 forming part of an enablement protocol. In other words, the enablement protocol for a specific function of this device requires a single touch at the location of theindicator 82. However in this embodiment theshape 83 of the touch object, of which there is no indication given on the display, also forms part of the enablement protocol as discussed above in relation toFIGS. 6A-6D . With certain touch detection systems e.g. optical systems, the resolution is sufficiently high to measure shape and/or change of shape on approach, at touch-down and at withdrawal of touch. InFIG. 8B a user holds thedevice 80 and commences the enablement protocol by placing the pad of theirthumb 84 on theindicator 82. This is accomplished inFIG. 8C and the function is enabled. Importantly, even though thespatial indicator 82 gives a user some indication of the required enablement protocol, if a touch object does not match the requiredshape 83, the device will not be enabled. For example touching theindicator 82 with a stylus or fingertip will not satisfy the enablement protocol. - A further embodiment can be seen in
FIGS. 9A-9C . In this embodiment adevice 90 includes a touchsensitive display 91 showing two spatial and/ortemporal indicators indicators indicators lines 94 and 95 (illustrative only, not shown on the display) the enablement protocol requires a simple touch at the location of theindicator 92, but not at the location of theindicator 93. Rather, the protocol requires a swipe along anarc 96 that begins in the lower right hand corner and terminates at theindicator 93. Accordingly, for an authorised user to enable a specific function, the device is held such that that portion of the enablement protocol requiring aleft thumb touch 97 at the location of theindicator 92 is satisfied, and theright thumb 98 proceeds from the lower right hand corner up to the location of theindicator 93 within the time specified by the enablement protocol. When both the spatial and temporal sequence as well as shape, change of shape or any other touch or near touch parameters are satisfied, the enablement protocol enables the function of thedevice 90 and the device is, in this embodiment, activated. There is virtually no prospect, even when the spatial and temporal indicators are provided, that an unauthorised user could satisfy all aspects of the user-defined protocol. - In another embodiment, an unauthorised person attempting to access the electronic device could attempt to replicate the enablement protocol of an authorised person by carefully inspecting the touch surface for residual finger marks. In cases such as these the enablement protocol may be selected to avoid indicating, from an inspection of finger marks alone, the actual enablement protocol. For example, a circular enablement protocol would not indicate the direction which the user traced out the circle. Similarly, the direction of a swipe would be difficult to interpret from only residual finger marks, and enablement protocols with temporal features such as the rhythm of a touch sequence could not be ascertained from residual finger marks.
- It will be clear to persons skilled in the art that the proposed method and device has significant flexibility and security as compared with the prior art. In addition, since all functions of a device may be linked to user-defined enablement protocols, operation of the device is completely user-dependent. This is also particularly helpful for users of such interface devices who cannot use conventional touch techniques, e.g. handicapped persons.
- Although the disclosure has been described with reference to specific embodiments, it will be appreciated by those skilled in the art that the disclosure may be embodied in many other forms.
Claims (15)
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/473,961 US9280279B2 (en) | 2007-05-11 | 2012-05-17 | User-defined gesture enablement protocols for touch input devices |
US15/056,015 US9696847B2 (en) | 2007-05-11 | 2016-02-29 | User-defined gesture enablement protocols for touch input device |
US15/626,585 US20170285853A1 (en) | 2007-05-11 | 2017-06-19 | User-defined gesture enablement protocols for touch input device |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AU2007/902,517 | 2007-05-11 | ||
AU2007/902,519 | 2007-05-11 | ||
AU2007902519A AU2007902519A0 (en) | 2007-05-11 | Gestural input | |
AU2007902517A AU2007902517A0 (en) | 2007-05-11 | Input device |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/473,961 Continuation US9280279B2 (en) | 2007-05-11 | 2012-05-17 | User-defined gesture enablement protocols for touch input devices |
Publications (1)
Publication Number | Publication Date |
---|---|
US20080278455A1 true US20080278455A1 (en) | 2008-11-13 |
Family
ID=39969088
Family Applications (4)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/118,047 Abandoned US20080278455A1 (en) | 2007-05-11 | 2008-05-09 | User-Defined Enablement Protocol |
US13/473,961 Active US9280279B2 (en) | 2007-05-11 | 2012-05-17 | User-defined gesture enablement protocols for touch input devices |
US15/056,015 Active - Reinstated US9696847B2 (en) | 2007-05-11 | 2016-02-29 | User-defined gesture enablement protocols for touch input device |
US15/626,585 Abandoned US20170285853A1 (en) | 2007-05-11 | 2017-06-19 | User-defined gesture enablement protocols for touch input device |
Family Applications After (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/473,961 Active US9280279B2 (en) | 2007-05-11 | 2012-05-17 | User-defined gesture enablement protocols for touch input devices |
US15/056,015 Active - Reinstated US9696847B2 (en) | 2007-05-11 | 2016-02-29 | User-defined gesture enablement protocols for touch input device |
US15/626,585 Abandoned US20170285853A1 (en) | 2007-05-11 | 2017-06-19 | User-defined gesture enablement protocols for touch input device |
Country Status (3)
Country | Link |
---|---|
US (4) | US20080278455A1 (en) |
TW (1) | TW200907742A (en) |
WO (1) | WO2008138030A1 (en) |
Cited By (122)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7593000B1 (en) * | 2008-05-17 | 2009-09-22 | David H. Chin | Touch-based authentication of a mobile device through user generated pattern creation |
US20090249476A1 (en) * | 2008-03-26 | 2009-10-01 | Lisa Anne Seacat | Using Multi-Touch Gestures to Protect Sensitive Content Using a Completely Automated Public Turing Test to Tell Computers and Humans Apart (CAPTCHA) |
US20090289916A1 (en) * | 2008-05-23 | 2009-11-26 | Hon Hai Precision Industry Co., Ltd. | Electronic device and method for switching between locked state and unlocked state |
US20100013780A1 (en) * | 2008-07-17 | 2010-01-21 | Sony Corporation | Information processing device, information processing method, and information processing program |
US20100083108A1 (en) * | 2008-09-26 | 2010-04-01 | Research In Motion Limited | Touch-screen device having soft escape key |
US20100162182A1 (en) * | 2008-12-23 | 2010-06-24 | Samsung Electronics Co., Ltd. | Method and apparatus for unlocking electronic appliance |
US20100201641A1 (en) * | 2007-08-13 | 2010-08-12 | Hideaki Tetsuhashi | Contact type input device, contact type input method, and program |
US20100222049A1 (en) * | 2009-03-02 | 2010-09-02 | Lg Electronics Inc. | Method for releasing a locking in mobile terminal and mobile terminal using the same |
US20100255885A1 (en) * | 2009-04-07 | 2010-10-07 | Samsung Electronics Co., Ltd. | Input device and method for mobile terminal |
US20110006996A1 (en) * | 2009-07-08 | 2011-01-13 | Smith Nathan J | Private data entry |
US20110019105A1 (en) * | 2009-07-27 | 2011-01-27 | Echostar Technologies L.L.C. | Verification of symbols received through a touchpad of a remote control device in an electronic system to allow access to system functions |
US20110025620A1 (en) * | 2008-01-11 | 2011-02-03 | Opdi Technologies A/S | Touch-sensitive device |
US20110032377A1 (en) * | 2009-08-07 | 2011-02-10 | Samsung Electronics Co., Ltd. | Digital photographing apparatus, method of controlling the same, and recording medium storing program to execute the method |
US20110041102A1 (en) * | 2009-08-11 | 2011-02-17 | Jong Hwan Kim | Mobile terminal and method for controlling the same |
US20110099473A1 (en) * | 2009-10-23 | 2011-04-28 | Samsung Electronics Co., Ltd. | Input signal processing device for portable device and method of the same |
US20110106446A1 (en) * | 2008-05-26 | 2011-05-05 | Volkswagen Ag | Display method for a display system, display system and operating method for a navigation system of a vehicle |
US20110161888A1 (en) * | 2009-12-28 | 2011-06-30 | Sony Corporation | Operation direction determination apparatus, remote operating system, operation direction determination method and program |
US20110162066A1 (en) * | 2009-12-29 | 2011-06-30 | Bizmodeline Co., Ltd. | Password processing method and apparatus |
US20110163967A1 (en) * | 2010-01-06 | 2011-07-07 | Imran Chaudhri | Device, Method, and Graphical User Interface for Changing Pages in an Electronic Document |
US20110185321A1 (en) * | 2010-01-26 | 2011-07-28 | Jay Christopher Capela | Device, Method, and Graphical User Interface for Precise Positioning of Objects |
US20110185300A1 (en) * | 2010-01-28 | 2011-07-28 | Microsoft Corporation | Brush, carbon-copy, and fill gestures |
US20110187497A1 (en) * | 2008-05-17 | 2011-08-04 | David H Chin | Comparison of an applied gesture on a touch screen of a mobile device with a remotely stored security gesture |
WO2011094045A2 (en) * | 2010-01-28 | 2011-08-04 | Microsoft Corporation | Copy and staple gestures |
US20110191719A1 (en) * | 2010-02-04 | 2011-08-04 | Microsoft Corporation | Cut, Punch-Out, and Rip Gestures |
US20110205176A1 (en) * | 2008-11-05 | 2011-08-25 | Takashi Okada | Portable electronic device, and power saving method and power saving program for the same |
US20110209103A1 (en) * | 2010-02-25 | 2011-08-25 | Hinckley Kenneth P | Multi-screen hold and drag gesture |
US20110265045A1 (en) * | 2010-04-26 | 2011-10-27 | Via Technologies, Inc. | Electronic system and method for operating touch screen thereof |
US20110282785A1 (en) * | 2008-05-17 | 2011-11-17 | Chin David H | Gesture based authentication for wireless payment by a mobile electronic device |
US20110283241A1 (en) * | 2010-05-14 | 2011-11-17 | Google Inc. | Touch Gesture Actions From A Device's Lock Screen |
US20110289423A1 (en) * | 2010-05-24 | 2011-11-24 | Samsung Electronics Co., Ltd. | Method and apparatus for controlling objects of a user interface |
US20120096376A1 (en) * | 2010-10-14 | 2012-04-19 | Canon Kabushiki Kaisha | Display control apparatus, display control method, and storage medium |
US20120133484A1 (en) * | 2010-11-29 | 2012-05-31 | Research In Motion Limited | Multiple-input device lock and unlock |
US20120196573A1 (en) * | 2009-10-09 | 2012-08-02 | Nec Corporation | Mobile device, and security lock control method and program thereof |
US8239785B2 (en) | 2010-01-27 | 2012-08-07 | Microsoft Corporation | Edge gestures |
US20120252410A1 (en) * | 2011-03-28 | 2012-10-04 | Htc Corporation | Systems and Methods for Gesture Lock Obfuscation |
US20120274583A1 (en) * | 2011-02-08 | 2012-11-01 | Ammon Haggerty | Multimodal Touchscreen Interaction Apparatuses, Methods and Systems |
US20130063009A1 (en) * | 2011-09-13 | 2013-03-14 | Nunoerin, Llc | Interactive furniture device |
US20130086673A1 (en) * | 2011-09-29 | 2013-04-04 | Cisco Technology, Inc. | Techniques for securely unlocking a touch screen user device |
US20130174083A1 (en) * | 2011-12-28 | 2013-07-04 | Acer Incorporated | Electronic device and method of controlling the same |
US8539386B2 (en) | 2010-01-26 | 2013-09-17 | Apple Inc. | Device, method, and graphical user interface for selecting and moving objects |
US8539384B2 (en) | 2010-02-25 | 2013-09-17 | Microsoft Corporation | Multi-screen pinch and expand gestures |
US20130278540A1 (en) * | 2012-04-20 | 2013-10-24 | Esat Yilmaz | Inter Touch Sensor Communications |
US20130298079A1 (en) * | 2012-05-02 | 2013-11-07 | Pantech Co., Ltd. | Apparatus and method for unlocking an electronic device |
US8587422B2 (en) | 2010-03-31 | 2013-11-19 | Tk Holdings, Inc. | Occupant sensing system |
US8612884B2 (en) | 2010-01-26 | 2013-12-17 | Apple Inc. | Device, method, and graphical user interface for resizing objects |
US20140002397A1 (en) * | 2012-06-29 | 2014-01-02 | Samsung Electronics Co., Ltd. | Method and device for handling event invocation using a stylus pen |
US20140006965A1 (en) * | 2012-07-02 | 2014-01-02 | Beijing Xiaomi Technology Co., Ltd. | Method of entering a user interface in a device and the device thereof |
US20140026105A1 (en) * | 2012-07-18 | 2014-01-23 | Research In Motion Limited | Method and Apparatus Pertaining to a Gesture-Controlled Snooze Instruction |
US20140028598A1 (en) * | 2012-07-30 | 2014-01-30 | Samsung Electronics Co., Ltd | Apparatus and method for controlling data transmission in terminal |
US20140033140A1 (en) * | 2012-07-11 | 2014-01-30 | Guang Dong Oppo Mobile Telecommunications Corp., Ltd. | Quick access function setting method for a touch control device |
US8707174B2 (en) | 2010-02-25 | 2014-04-22 | Microsoft Corporation | Multi-screen hold and page-flip gesture |
US8725230B2 (en) | 2010-04-02 | 2014-05-13 | Tk Holdings Inc. | Steering wheel with hand sensors |
EP2731002A1 (en) * | 2012-11-09 | 2014-05-14 | Thales | Method for securing a command on a viewing device with tactile surface and related system |
US20140137234A1 (en) * | 2008-05-17 | 2014-05-15 | David H. Chin | Mobile device authentication through touch-based gestures |
US20140145990A1 (en) * | 2012-11-29 | 2014-05-29 | Egalax_Empia Technology Inc. | Method for unlocking touch screen, electronic device thereof, and recording medium thereof |
US8751970B2 (en) | 2010-02-25 | 2014-06-10 | Microsoft Corporation | Multi-screen synchronous slide gesture |
US8766928B2 (en) | 2009-09-25 | 2014-07-01 | Apple Inc. | Device, method, and graphical user interface for manipulating user interface objects |
US8780069B2 (en) | 2009-09-25 | 2014-07-15 | Apple Inc. | Device, method, and graphical user interface for manipulating user interface objects |
US8799827B2 (en) | 2010-02-19 | 2014-08-05 | Microsoft Corporation | Page manipulations using on and off-screen gestures |
US8799826B2 (en) | 2009-09-25 | 2014-08-05 | Apple Inc. | Device, method, and graphical user interface for moving a calendar entry in a calendar application |
US8836648B2 (en) | 2009-05-27 | 2014-09-16 | Microsoft Corporation | Touch pull-in gesture |
US20140267064A1 (en) * | 2013-03-13 | 2014-09-18 | Htc Corporation | Unlock Method and Mobile Device Using the Same |
US20140298672A1 (en) * | 2012-09-27 | 2014-10-09 | Analog Devices Technology | Locking and unlocking of contacless gesture-based user interface of device having contactless gesture detection system |
US8863016B2 (en) | 2009-09-22 | 2014-10-14 | Apple Inc. | Device, method, and graphical user interface for manipulating user interface objects |
US20140347302A1 (en) * | 2013-05-24 | 2014-11-27 | Boe Technology Group Co., Ltd. | Unlocking method and device for touch screen terminal and touch screen terminal |
US8972879B2 (en) | 2010-07-30 | 2015-03-03 | Apple Inc. | Device, method, and graphical user interface for reordering the front-to-back positions of objects |
US20150066762A1 (en) * | 2013-08-28 | 2015-03-05 | Geoffrey W. Chatterton | Authentication system |
US9007190B2 (en) | 2010-03-31 | 2015-04-14 | Tk Holdings Inc. | Steering wheel sensors |
EP2835732A3 (en) * | 2013-08-06 | 2015-04-22 | Samsung Electronics Co., Ltd | Electronic device provided with touch screen and operating method thereof |
US9052820B2 (en) | 2011-05-27 | 2015-06-09 | Microsoft Technology Licensing, Llc | Multi-application environment |
US9075522B2 (en) | 2010-02-25 | 2015-07-07 | Microsoft Technology Licensing, Llc | Multi-screen bookmark hold gesture |
US9081494B2 (en) | 2010-07-30 | 2015-07-14 | Apple Inc. | Device, method, and graphical user interface for copying formatting attributes |
US20150199553A1 (en) * | 2014-01-13 | 2015-07-16 | Samsung Electronics Co., Ltd. | Method for recognizing fingerprint and mobile terminal supporting the same |
US9098182B2 (en) | 2010-07-30 | 2015-08-04 | Apple Inc. | Device, method, and graphical user interface for copying user interface objects between content regions |
US9104440B2 (en) | 2011-05-27 | 2015-08-11 | Microsoft Technology Licensing, Llc | Multi-application environment |
EP2919108A1 (en) * | 2014-03-13 | 2015-09-16 | LG Electronics, Inc. | Mobile terminal and method of controlling the same |
US9158445B2 (en) | 2011-05-27 | 2015-10-13 | Microsoft Technology Licensing, Llc | Managing an immersive interface in a multi-application immersive environment |
US20150317076A1 (en) * | 2012-12-11 | 2015-11-05 | University Of Washington Through Its Center For Commercialization | Use of hand posture to improve text entry |
WO2015149101A3 (en) * | 2014-04-01 | 2015-11-26 | Linkilike Gmbh | Method for comparing user identities of databases |
US9229918B2 (en) | 2010-12-23 | 2016-01-05 | Microsoft Technology Licensing, Llc | Presenting an application change through a tile |
EP2977884A1 (en) * | 2014-07-22 | 2016-01-27 | LG Electronics Inc. | Mobile terminal and method for controlling the same |
US20160026380A1 (en) * | 2014-07-25 | 2016-01-28 | Hannstar Display (Nanjing) Corporation | Hand-held electronic device, computer-executed method and touch-sensing cover |
US20160034172A1 (en) * | 2014-07-30 | 2016-02-04 | Wistron Corporation | Touch device and control method and method for determining unlocking thereof |
US9261964B2 (en) | 2005-12-30 | 2016-02-16 | Microsoft Technology Licensing, Llc | Unintentional touch rejection |
US20160054827A1 (en) * | 2014-08-21 | 2016-02-25 | Echostar Technologies L.L.C. | Determining handedness on multi-element capacitive devices |
US9274682B2 (en) | 2010-02-19 | 2016-03-01 | Microsoft Technology Licensing, Llc | Off-screen gestures to create on-screen input |
US20160070465A1 (en) * | 2014-09-08 | 2016-03-10 | Lenovo (Singapore) Pte, Ltd. | Managing an on-screen keyboard |
US9310994B2 (en) | 2010-02-19 | 2016-04-12 | Microsoft Technology Licensing, Llc | Use of bezel as an input mechanism |
US9354803B2 (en) | 2005-12-23 | 2016-05-31 | Apple Inc. | Scrolling list with floating adjacent index symbols |
US9367205B2 (en) | 2010-02-19 | 2016-06-14 | Microsoft Technolgoy Licensing, Llc | Radial menus with bezel gestures |
US9436374B2 (en) | 2009-09-25 | 2016-09-06 | Apple Inc. | Device, method, and graphical user interface for scrolling a multi-section document |
US9454304B2 (en) | 2010-02-25 | 2016-09-27 | Microsoft Technology Licensing, Llc | Multi-screen dual tap gesture |
US9477337B2 (en) | 2014-03-14 | 2016-10-25 | Microsoft Technology Licensing, Llc | Conductive trace routing for display and bezel sensors |
US9519356B2 (en) | 2010-02-04 | 2016-12-13 | Microsoft Technology Licensing, Llc | Link gestures |
US9582122B2 (en) | 2012-11-12 | 2017-02-28 | Microsoft Technology Licensing, Llc | Touch-sensitive bezel techniques |
US9658766B2 (en) | 2011-05-27 | 2017-05-23 | Microsoft Technology Licensing, Llc | Edge gesture |
US9696223B2 (en) | 2012-09-17 | 2017-07-04 | Tk Holdings Inc. | Single layer force sensor |
US9696888B2 (en) | 2010-12-20 | 2017-07-04 | Microsoft Technology Licensing, Llc | Application-launching interface for multiple modes |
US9727031B2 (en) | 2012-04-13 | 2017-08-08 | Tk Holdings Inc. | Pressure sensor including a pressure sensitive material for use with control systems and methods of using the same |
JP2017140167A (en) * | 2016-02-09 | 2017-08-17 | 株式会社コナミデジタルエンタテインメント | Game system, game control device, and program |
US20170285921A1 (en) * | 2016-03-31 | 2017-10-05 | Brother Kogyo Kabushiki Kaisha | Information processing apparatus,non-transitory computer-readable medium storing instructions therefor, and information processing method |
WO2017167698A1 (en) * | 2016-04-01 | 2017-10-05 | Thales | Secure method for control of a function by means of a touchpad |
US9965165B2 (en) | 2010-02-19 | 2018-05-08 | Microsoft Technology Licensing, Llc | Multi-finger gestures |
US10156934B2 (en) * | 2016-08-12 | 2018-12-18 | Canon Kabushiki Kaisha | Electronic device and control method thereof |
US10254927B2 (en) | 2009-09-25 | 2019-04-09 | Apple Inc. | Device, method, and graphical user interface for manipulating workspace views |
US10254955B2 (en) | 2011-09-10 | 2019-04-09 | Microsoft Technology Licensing, Llc | Progressively indicating new content in an application-selectable user interface |
US10269156B2 (en) | 2015-06-05 | 2019-04-23 | Manufacturing Resources International, Inc. | System and method for blending order confirmation over menu board background |
US10313037B2 (en) | 2016-05-31 | 2019-06-04 | Manufacturing Resources International, Inc. | Electronic display remote image verification system and method |
US10319271B2 (en) | 2016-03-22 | 2019-06-11 | Manufacturing Resources International, Inc. | Cyclic redundancy check for electronic displays |
US10319408B2 (en) | 2015-03-30 | 2019-06-11 | Manufacturing Resources International, Inc. | Monolithic display with separately controllable sections |
US10510304B2 (en) | 2016-08-10 | 2019-12-17 | Manufacturing Resources International, Inc. | Dynamic dimming LED backlight for LCD array |
US10579250B2 (en) | 2011-09-01 | 2020-03-03 | Microsoft Technology Licensing, Llc | Arranging tiles |
US10705701B2 (en) | 2009-03-16 | 2020-07-07 | Apple Inc. | Device, method, and graphical user interface for moving a current position in content at a variable scrubbing rate |
US10725624B2 (en) | 2015-06-05 | 2020-07-28 | Apple Inc. | Movement between multiple views |
US10922736B2 (en) | 2015-05-15 | 2021-02-16 | Manufacturing Resources International, Inc. | Smart electronic display for restaurants |
US10969944B2 (en) | 2010-12-23 | 2021-04-06 | Microsoft Technology Licensing, Llc | Application reporting in an application-selectable user interface |
US10990251B1 (en) * | 2019-11-08 | 2021-04-27 | Sap Se | Smart augmented reality selector |
US11164508B2 (en) * | 2019-05-16 | 2021-11-02 | Asustek Computer Inc. | Electronic device |
US11272017B2 (en) | 2011-05-27 | 2022-03-08 | Microsoft Technology Licensing, Llc | Application notifications manifest |
US11294484B1 (en) * | 2012-12-31 | 2022-04-05 | Allscripts Software, Llc | Method for ensuring use intentions of a touch screen device |
US11301128B2 (en) * | 2019-05-01 | 2022-04-12 | Google Llc | Intended input to a user interface from detected gesture positions |
US11895362B2 (en) | 2021-10-29 | 2024-02-06 | Manufacturing Resources International, Inc. | Proof of play for images displayed at electronic displays |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TWI403932B (en) * | 2009-09-14 | 2013-08-01 | Acer Inc | Method for operating a touch screen, method for defining a touch gesture on the touch screen, and electronic device thereof |
US9213482B2 (en) * | 2011-11-11 | 2015-12-15 | Elan Microelectronics Corporation | Touch control device and method |
US9880653B2 (en) | 2012-04-30 | 2018-01-30 | Corning Incorporated | Pressure-sensing touch system utilizing total-internal reflection |
CN106873874B (en) * | 2017-01-20 | 2021-01-08 | 维沃移动通信有限公司 | Application program starting method and mobile terminal |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5583946A (en) * | 1993-09-30 | 1996-12-10 | Apple Computer, Inc. | Method and apparatus for recognizing gestures on a computer system |
US6163616A (en) * | 1995-12-29 | 2000-12-19 | Feldman; Stephen E. | System and method for verifying the identity of a person |
US20050122315A1 (en) * | 2001-11-16 | 2005-06-09 | Martin Chalk | Communications device and supporting network |
US20050210417A1 (en) * | 2004-03-23 | 2005-09-22 | Marvit David L | User definable gestures for motion controlled handheld devices |
US7286063B2 (en) * | 2002-11-26 | 2007-10-23 | Asulab S.A. | Method of input of a security code by means of a touch screen for access to a function, an apparatus or a given location, and device for implementing the same |
US20080012835A1 (en) * | 2006-07-12 | 2008-01-17 | N-Trig Ltd. | Hover and touch detection for digitizer |
US20080092245A1 (en) * | 2006-09-15 | 2008-04-17 | Agent Science Technologies, Inc. | Multi-touch device behaviormetric user authentication and dynamic usability system |
US20080098456A1 (en) * | 2006-09-15 | 2008-04-24 | Agent Science Technologies, Inc. | Continuous user identification and situation analysis with identification of anonymous users through behaviormetrics |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7170496B2 (en) * | 2003-01-24 | 2007-01-30 | Bruce Peter Middleton | Zero-front-footprint compact input system |
US9063647B2 (en) * | 2006-05-12 | 2015-06-23 | Microsoft Technology Licensing, Llc | Multi-touch uses, gestures, and implementation |
US8842074B2 (en) * | 2006-09-06 | 2014-09-23 | Apple Inc. | Portable electronic device performing similar operations for different gestures |
US7877707B2 (en) * | 2007-01-06 | 2011-01-25 | Apple Inc. | Detecting and interpreting real-world and security gestures on touch and hover sensitive devices |
US8311530B2 (en) * | 2007-01-26 | 2012-11-13 | Research In Motion Limited | Touch entry of password on a mobile device |
US20080204417A1 (en) * | 2007-02-27 | 2008-08-28 | Pierce Paul M | Multimodal Adaptive User Interface for a Portable Electronic Device |
-
2008
- 2008-05-09 WO PCT/AU2008/000456 patent/WO2008138030A1/en active Application Filing
- 2008-05-09 US US12/118,047 patent/US20080278455A1/en not_active Abandoned
- 2008-05-12 TW TW097117405A patent/TW200907742A/en unknown
-
2012
- 2012-05-17 US US13/473,961 patent/US9280279B2/en active Active
-
2016
- 2016-02-29 US US15/056,015 patent/US9696847B2/en active Active - Reinstated
-
2017
- 2017-06-19 US US15/626,585 patent/US20170285853A1/en not_active Abandoned
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5583946A (en) * | 1993-09-30 | 1996-12-10 | Apple Computer, Inc. | Method and apparatus for recognizing gestures on a computer system |
US6163616A (en) * | 1995-12-29 | 2000-12-19 | Feldman; Stephen E. | System and method for verifying the identity of a person |
US20050122315A1 (en) * | 2001-11-16 | 2005-06-09 | Martin Chalk | Communications device and supporting network |
US7286063B2 (en) * | 2002-11-26 | 2007-10-23 | Asulab S.A. | Method of input of a security code by means of a touch screen for access to a function, an apparatus or a given location, and device for implementing the same |
US20050210417A1 (en) * | 2004-03-23 | 2005-09-22 | Marvit David L | User definable gestures for motion controlled handheld devices |
US20080012835A1 (en) * | 2006-07-12 | 2008-01-17 | N-Trig Ltd. | Hover and touch detection for digitizer |
US20080092245A1 (en) * | 2006-09-15 | 2008-04-17 | Agent Science Technologies, Inc. | Multi-touch device behaviormetric user authentication and dynamic usability system |
US20080098456A1 (en) * | 2006-09-15 | 2008-04-24 | Agent Science Technologies, Inc. | Continuous user identification and situation analysis with identification of anonymous users through behaviormetrics |
Cited By (212)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9354803B2 (en) | 2005-12-23 | 2016-05-31 | Apple Inc. | Scrolling list with floating adjacent index symbols |
US10732814B2 (en) | 2005-12-23 | 2020-08-04 | Apple Inc. | Scrolling list with floating adjacent index symbols |
US9594457B2 (en) | 2005-12-30 | 2017-03-14 | Microsoft Technology Licensing, Llc | Unintentional touch rejection |
US9946370B2 (en) | 2005-12-30 | 2018-04-17 | Microsoft Technology Licensing, Llc | Unintentional touch rejection |
US9952718B2 (en) | 2005-12-30 | 2018-04-24 | Microsoft Technology Licensing, Llc | Unintentional touch rejection |
US10019080B2 (en) | 2005-12-30 | 2018-07-10 | Microsoft Technology Licensing, Llc | Unintentional touch rejection |
US9261964B2 (en) | 2005-12-30 | 2016-02-16 | Microsoft Technology Licensing, Llc | Unintentional touch rejection |
US20100201641A1 (en) * | 2007-08-13 | 2010-08-12 | Hideaki Tetsuhashi | Contact type input device, contact type input method, and program |
US20110025620A1 (en) * | 2008-01-11 | 2011-02-03 | Opdi Technologies A/S | Touch-sensitive device |
US9740336B2 (en) | 2008-01-11 | 2017-08-22 | O-Net Wavetouch Limited | Touch-sensitive device |
US9342187B2 (en) | 2008-01-11 | 2016-05-17 | O-Net Wavetouch Limited | Touch-sensitive device |
US8214891B2 (en) * | 2008-03-26 | 2012-07-03 | International Business Machines Corporation | Using multi-touch gestures to protect sensitive content using a completely automated public turing test to tell computers and humans apart (CAPTCHA) |
US20090249476A1 (en) * | 2008-03-26 | 2009-10-01 | Lisa Anne Seacat | Using Multi-Touch Gestures to Protect Sensitive Content Using a Completely Automated Public Turing Test to Tell Computers and Humans Apart (CAPTCHA) |
US9082117B2 (en) * | 2008-05-17 | 2015-07-14 | David H. Chin | Gesture based authentication for wireless payment by a mobile electronic device |
US8174503B2 (en) | 2008-05-17 | 2012-05-08 | David H. Cain | Touch-based authentication of a mobile device through user generated pattern creation |
US20140137234A1 (en) * | 2008-05-17 | 2014-05-15 | David H. Chin | Mobile device authentication through touch-based gestures |
US8913028B2 (en) * | 2008-05-17 | 2014-12-16 | David H. Chin | Mobile device authentication through touch-based gestures |
US7593000B1 (en) * | 2008-05-17 | 2009-09-22 | David H. Chin | Touch-based authentication of a mobile device through user generated pattern creation |
US9024890B2 (en) | 2008-05-17 | 2015-05-05 | David H. Chin | Comparison of an applied gesture on a touch screen of a mobile device with a remotely stored security gesture |
US20110187497A1 (en) * | 2008-05-17 | 2011-08-04 | David H Chin | Comparison of an applied gesture on a touch screen of a mobile device with a remotely stored security gesture |
US20110282785A1 (en) * | 2008-05-17 | 2011-11-17 | Chin David H | Gesture based authentication for wireless payment by a mobile electronic device |
US20090289916A1 (en) * | 2008-05-23 | 2009-11-26 | Hon Hai Precision Industry Co., Ltd. | Electronic device and method for switching between locked state and unlocked state |
US20110106446A1 (en) * | 2008-05-26 | 2011-05-05 | Volkswagen Ag | Display method for a display system, display system and operating method for a navigation system of a vehicle |
US10330488B2 (en) * | 2008-05-26 | 2019-06-25 | Volkswagen Ag | Display method for a display system, display system and operating method for a navigation system of a vehicle |
US9459116B2 (en) * | 2008-05-26 | 2016-10-04 | Volkswagen Ag | Display method for a display system, display system and operating method for a navigation system of a vehicle |
US9411503B2 (en) * | 2008-07-17 | 2016-08-09 | Sony Corporation | Information processing device, information processing method, and information processing program |
US20100013780A1 (en) * | 2008-07-17 | 2010-01-21 | Sony Corporation | Information processing device, information processing method, and information processing program |
US20100083108A1 (en) * | 2008-09-26 | 2010-04-01 | Research In Motion Limited | Touch-screen device having soft escape key |
US20110205176A1 (en) * | 2008-11-05 | 2011-08-25 | Takashi Okada | Portable electronic device, and power saving method and power saving program for the same |
US9032337B2 (en) * | 2008-12-23 | 2015-05-12 | Samsung Electronics Co., Ltd. | Method and apparatus for unlocking electronic appliance |
US10175875B2 (en) | 2008-12-23 | 2019-01-08 | Samsung Electronics Co., Ltd. | Method and apparatus for unlocking electronic appliance |
US11137895B2 (en) | 2008-12-23 | 2021-10-05 | Samsung Electronics Co., Ltd. | Method and apparatus for unlocking electronic appliance |
US20100162182A1 (en) * | 2008-12-23 | 2010-06-24 | Samsung Electronics Co., Ltd. | Method and apparatus for unlocking electronic appliance |
KR20100098804A (en) * | 2009-03-02 | 2010-09-10 | 엘지전자 주식회사 | Method for releasing locking in mobile terminal and mobile terminal using the same |
US20100222049A1 (en) * | 2009-03-02 | 2010-09-02 | Lg Electronics Inc. | Method for releasing a locking in mobile terminal and mobile terminal using the same |
US8712392B2 (en) * | 2009-03-02 | 2014-04-29 | Lg Electronics Inc. | Method for releasing a locking in mobile terminal and mobile terminal using the same |
KR101582687B1 (en) * | 2009-03-02 | 2016-01-05 | 엘지전자 주식회사 | Method for releasing locking in mobile terminal and mobile terminal using the same |
US9298349B2 (en) | 2009-03-02 | 2016-03-29 | Lg Electronics Inc. | Method for releasing a locking in mobile terminal and mobile terminal using the same |
US10705701B2 (en) | 2009-03-16 | 2020-07-07 | Apple Inc. | Device, method, and graphical user interface for moving a current position in content at a variable scrubbing rate |
US11567648B2 (en) | 2009-03-16 | 2023-01-31 | Apple Inc. | Device, method, and graphical user interface for moving a current position in content at a variable scrubbing rate |
US11907519B2 (en) | 2009-03-16 | 2024-02-20 | Apple Inc. | Device, method, and graphical user interface for moving a current position in content at a variable scrubbing rate |
US20100255885A1 (en) * | 2009-04-07 | 2010-10-07 | Samsung Electronics Co., Ltd. | Input device and method for mobile terminal |
US8836648B2 (en) | 2009-05-27 | 2014-09-16 | Microsoft Corporation | Touch pull-in gesture |
US20110006996A1 (en) * | 2009-07-08 | 2011-01-13 | Smith Nathan J | Private data entry |
US20110019105A1 (en) * | 2009-07-27 | 2011-01-27 | Echostar Technologies L.L.C. | Verification of symbols received through a touchpad of a remote control device in an electronic system to allow access to system functions |
US8520113B2 (en) * | 2009-08-07 | 2013-08-27 | Samsung Electronics Co., Ltd. | Digital photographing apparatus, method of controlling the same, and recording medium storing program to execute the method |
US20110032377A1 (en) * | 2009-08-07 | 2011-02-10 | Samsung Electronics Co., Ltd. | Digital photographing apparatus, method of controlling the same, and recording medium storing program to execute the method |
US9563350B2 (en) * | 2009-08-11 | 2017-02-07 | Lg Electronics Inc. | Mobile terminal and method for controlling the same |
US20110041102A1 (en) * | 2009-08-11 | 2011-02-17 | Jong Hwan Kim | Mobile terminal and method for controlling the same |
US10282070B2 (en) | 2009-09-22 | 2019-05-07 | Apple Inc. | Device, method, and graphical user interface for manipulating user interface objects |
US8863016B2 (en) | 2009-09-22 | 2014-10-14 | Apple Inc. | Device, method, and graphical user interface for manipulating user interface objects |
US11334229B2 (en) | 2009-09-22 | 2022-05-17 | Apple Inc. | Device, method, and graphical user interface for manipulating user interface objects |
US10788965B2 (en) | 2009-09-22 | 2020-09-29 | Apple Inc. | Device, method, and graphical user interface for manipulating user interface objects |
US10564826B2 (en) | 2009-09-22 | 2020-02-18 | Apple Inc. | Device, method, and graphical user interface for manipulating user interface objects |
US11947782B2 (en) | 2009-09-25 | 2024-04-02 | Apple Inc. | Device, method, and graphical user interface for manipulating workspace views |
US8766928B2 (en) | 2009-09-25 | 2014-07-01 | Apple Inc. | Device, method, and graphical user interface for manipulating user interface objects |
US8780069B2 (en) | 2009-09-25 | 2014-07-15 | Apple Inc. | Device, method, and graphical user interface for manipulating user interface objects |
US10254927B2 (en) | 2009-09-25 | 2019-04-09 | Apple Inc. | Device, method, and graphical user interface for manipulating workspace views |
US9310907B2 (en) | 2009-09-25 | 2016-04-12 | Apple Inc. | Device, method, and graphical user interface for manipulating user interface objects |
US10928993B2 (en) | 2009-09-25 | 2021-02-23 | Apple Inc. | Device, method, and graphical user interface for manipulating workspace views |
US8799826B2 (en) | 2009-09-25 | 2014-08-05 | Apple Inc. | Device, method, and graphical user interface for moving a calendar entry in a calendar application |
US11366576B2 (en) | 2009-09-25 | 2022-06-21 | Apple Inc. | Device, method, and graphical user interface for manipulating workspace views |
US9436374B2 (en) | 2009-09-25 | 2016-09-06 | Apple Inc. | Device, method, and graphical user interface for scrolling a multi-section document |
JP2015130182A (en) * | 2009-10-09 | 2015-07-16 | レノボ・イノベーションズ・リミテッド(香港) | Mobile device, security lock control method thereof, and program |
EP2487622A1 (en) * | 2009-10-09 | 2012-08-15 | Nec Corporation | Mobile device, and security lock control method and program thereof |
US20120196573A1 (en) * | 2009-10-09 | 2012-08-02 | Nec Corporation | Mobile device, and security lock control method and program thereof |
EP2487622A4 (en) * | 2009-10-09 | 2014-05-28 | Nec Corp | Mobile device, and security lock control method and program thereof |
US20110099473A1 (en) * | 2009-10-23 | 2011-04-28 | Samsung Electronics Co., Ltd. | Input signal processing device for portable device and method of the same |
US20110161888A1 (en) * | 2009-12-28 | 2011-06-30 | Sony Corporation | Operation direction determination apparatus, remote operating system, operation direction determination method and program |
US20150365236A1 (en) * | 2009-12-29 | 2015-12-17 | Bizmodeline Co., Ltd. | Password processing device |
US20110162066A1 (en) * | 2009-12-29 | 2011-06-30 | Bizmodeline Co., Ltd. | Password processing method and apparatus |
US9830444B2 (en) * | 2009-12-29 | 2017-11-28 | Bizmodeline Co., Ltd. | Password processing device |
US9146669B2 (en) * | 2009-12-29 | 2015-09-29 | Bizmodeline Co., Ltd. | Password processing method and apparatus |
US9444626B2 (en) * | 2009-12-29 | 2016-09-13 | Bizmodeline Co., Ltd. | Password processing device |
US20160342785A1 (en) * | 2009-12-29 | 2016-11-24 | Bizmodeline Co., Ltd. | Password processing device |
US20110163967A1 (en) * | 2010-01-06 | 2011-07-07 | Imran Chaudhri | Device, Method, and Graphical User Interface for Changing Pages in an Electronic Document |
US8539385B2 (en) * | 2010-01-26 | 2013-09-17 | Apple Inc. | Device, method, and graphical user interface for precise positioning of objects |
US8539386B2 (en) | 2010-01-26 | 2013-09-17 | Apple Inc. | Device, method, and graphical user interface for selecting and moving objects |
US20110185321A1 (en) * | 2010-01-26 | 2011-07-28 | Jay Christopher Capela | Device, Method, and Graphical User Interface for Precise Positioning of Objects |
US8612884B2 (en) | 2010-01-26 | 2013-12-17 | Apple Inc. | Device, method, and graphical user interface for resizing objects |
US8677268B2 (en) | 2010-01-26 | 2014-03-18 | Apple Inc. | Device, method, and graphical user interface for resizing objects |
US8239785B2 (en) | 2010-01-27 | 2012-08-07 | Microsoft Corporation | Edge gestures |
WO2011094045A2 (en) * | 2010-01-28 | 2011-08-04 | Microsoft Corporation | Copy and staple gestures |
US8261213B2 (en) | 2010-01-28 | 2012-09-04 | Microsoft Corporation | Brush, carbon-copy, and fill gestures |
US9857970B2 (en) | 2010-01-28 | 2018-01-02 | Microsoft Technology Licensing, Llc | Copy and staple gestures |
US9411504B2 (en) | 2010-01-28 | 2016-08-09 | Microsoft Technology Licensing, Llc | Copy and staple gestures |
US9411498B2 (en) | 2010-01-28 | 2016-08-09 | Microsoft Technology Licensing, Llc | Brush, carbon-copy, and fill gestures |
WO2011094045A3 (en) * | 2010-01-28 | 2011-10-20 | Microsoft Corporation | Copy and staple gestures |
US10282086B2 (en) | 2010-01-28 | 2019-05-07 | Microsoft Technology Licensing, Llc | Brush, carbon-copy, and fill gestures |
US20110185300A1 (en) * | 2010-01-28 | 2011-07-28 | Microsoft Corporation | Brush, carbon-copy, and fill gestures |
US20110191719A1 (en) * | 2010-02-04 | 2011-08-04 | Microsoft Corporation | Cut, Punch-Out, and Rip Gestures |
US9519356B2 (en) | 2010-02-04 | 2016-12-13 | Microsoft Technology Licensing, Llc | Link gestures |
US8799827B2 (en) | 2010-02-19 | 2014-08-05 | Microsoft Corporation | Page manipulations using on and off-screen gestures |
US9310994B2 (en) | 2010-02-19 | 2016-04-12 | Microsoft Technology Licensing, Llc | Use of bezel as an input mechanism |
US10268367B2 (en) | 2010-02-19 | 2019-04-23 | Microsoft Technology Licensing, Llc | Radial menus with bezel gestures |
US9274682B2 (en) | 2010-02-19 | 2016-03-01 | Microsoft Technology Licensing, Llc | Off-screen gestures to create on-screen input |
US9367205B2 (en) | 2010-02-19 | 2016-06-14 | Microsoft Technolgoy Licensing, Llc | Radial menus with bezel gestures |
US9965165B2 (en) | 2010-02-19 | 2018-05-08 | Microsoft Technology Licensing, Llc | Multi-finger gestures |
US9454304B2 (en) | 2010-02-25 | 2016-09-27 | Microsoft Technology Licensing, Llc | Multi-screen dual tap gesture |
US20110209103A1 (en) * | 2010-02-25 | 2011-08-25 | Hinckley Kenneth P | Multi-screen hold and drag gesture |
US8539384B2 (en) | 2010-02-25 | 2013-09-17 | Microsoft Corporation | Multi-screen pinch and expand gestures |
US11055050B2 (en) | 2010-02-25 | 2021-07-06 | Microsoft Technology Licensing, Llc | Multi-device pairing and combined display |
US8751970B2 (en) | 2010-02-25 | 2014-06-10 | Microsoft Corporation | Multi-screen synchronous slide gesture |
US9075522B2 (en) | 2010-02-25 | 2015-07-07 | Microsoft Technology Licensing, Llc | Multi-screen bookmark hold gesture |
US8707174B2 (en) | 2010-02-25 | 2014-04-22 | Microsoft Corporation | Multi-screen hold and page-flip gesture |
US8473870B2 (en) | 2010-02-25 | 2013-06-25 | Microsoft Corporation | Multi-screen hold and drag gesture |
US8587422B2 (en) | 2010-03-31 | 2013-11-19 | Tk Holdings, Inc. | Occupant sensing system |
US9007190B2 (en) | 2010-03-31 | 2015-04-14 | Tk Holdings Inc. | Steering wheel sensors |
US8725230B2 (en) | 2010-04-02 | 2014-05-13 | Tk Holdings Inc. | Steering wheel with hand sensors |
US9134897B2 (en) * | 2010-04-26 | 2015-09-15 | Via Technologies, Inc. | Electronic system and method for operating touch screen thereof |
US20110265045A1 (en) * | 2010-04-26 | 2011-10-27 | Via Technologies, Inc. | Electronic system and method for operating touch screen thereof |
US8136053B1 (en) * | 2010-05-14 | 2012-03-13 | Google Inc. | Direct, gesture-based actions from device's lock screen |
US20110283241A1 (en) * | 2010-05-14 | 2011-11-17 | Google Inc. | Touch Gesture Actions From A Device's Lock Screen |
US20110289423A1 (en) * | 2010-05-24 | 2011-11-24 | Samsung Electronics Co., Ltd. | Method and apparatus for controlling objects of a user interface |
EP2390812B1 (en) * | 2010-05-24 | 2016-07-20 | Samsung Electronics Co., Ltd. | Method and apparatus for controlling objects of a user interface |
US9098182B2 (en) | 2010-07-30 | 2015-08-04 | Apple Inc. | Device, method, and graphical user interface for copying user interface objects between content regions |
US8972879B2 (en) | 2010-07-30 | 2015-03-03 | Apple Inc. | Device, method, and graphical user interface for reordering the front-to-back positions of objects |
US9626098B2 (en) | 2010-07-30 | 2017-04-18 | Apple Inc. | Device, method, and graphical user interface for copying formatting attributes |
US9081494B2 (en) | 2010-07-30 | 2015-07-14 | Apple Inc. | Device, method, and graphical user interface for copying formatting attributes |
US20120096376A1 (en) * | 2010-10-14 | 2012-04-19 | Canon Kabushiki Kaisha | Display control apparatus, display control method, and storage medium |
US8782544B2 (en) * | 2010-10-14 | 2014-07-15 | Canon Kabushiki Kaisha | Display control apparatus, display control method, and storage medium |
US20120133484A1 (en) * | 2010-11-29 | 2012-05-31 | Research In Motion Limited | Multiple-input device lock and unlock |
US9696888B2 (en) | 2010-12-20 | 2017-07-04 | Microsoft Technology Licensing, Llc | Application-launching interface for multiple modes |
US10969944B2 (en) | 2010-12-23 | 2021-04-06 | Microsoft Technology Licensing, Llc | Application reporting in an application-selectable user interface |
US11126333B2 (en) | 2010-12-23 | 2021-09-21 | Microsoft Technology Licensing, Llc | Application reporting in an application-selectable user interface |
US9229918B2 (en) | 2010-12-23 | 2016-01-05 | Microsoft Technology Licensing, Llc | Presenting an application change through a tile |
US20120274583A1 (en) * | 2011-02-08 | 2012-11-01 | Ammon Haggerty | Multimodal Touchscreen Interaction Apparatuses, Methods and Systems |
CN103534674A (en) * | 2011-02-08 | 2014-01-22 | 海沃氏公司 | Multimodal touchscreen interaction apparatuses, methods and systems |
US20120252410A1 (en) * | 2011-03-28 | 2012-10-04 | Htc Corporation | Systems and Methods for Gesture Lock Obfuscation |
US9514297B2 (en) * | 2011-03-28 | 2016-12-06 | Htc Corporation | Systems and methods for gesture lock obfuscation |
US9535597B2 (en) | 2011-05-27 | 2017-01-03 | Microsoft Technology Licensing, Llc | Managing an immersive interface in a multi-application immersive environment |
US9052820B2 (en) | 2011-05-27 | 2015-06-09 | Microsoft Technology Licensing, Llc | Multi-application environment |
US9158445B2 (en) | 2011-05-27 | 2015-10-13 | Microsoft Technology Licensing, Llc | Managing an immersive interface in a multi-application immersive environment |
US9658766B2 (en) | 2011-05-27 | 2017-05-23 | Microsoft Technology Licensing, Llc | Edge gesture |
US9104440B2 (en) | 2011-05-27 | 2015-08-11 | Microsoft Technology Licensing, Llc | Multi-application environment |
US11698721B2 (en) | 2011-05-27 | 2023-07-11 | Microsoft Technology Licensing, Llc | Managing an immersive interface in a multi-application immersive environment |
US11272017B2 (en) | 2011-05-27 | 2022-03-08 | Microsoft Technology Licensing, Llc | Application notifications manifest |
US10303325B2 (en) | 2011-05-27 | 2019-05-28 | Microsoft Technology Licensing, Llc | Multi-application environment |
US9104307B2 (en) | 2011-05-27 | 2015-08-11 | Microsoft Technology Licensing, Llc | Multi-application environment |
US10579250B2 (en) | 2011-09-01 | 2020-03-03 | Microsoft Technology Licensing, Llc | Arranging tiles |
US10254955B2 (en) | 2011-09-10 | 2019-04-09 | Microsoft Technology Licensing, Llc | Progressively indicating new content in an application-selectable user interface |
US20130063009A1 (en) * | 2011-09-13 | 2013-03-14 | Nunoerin, Llc | Interactive furniture device |
US20130086673A1 (en) * | 2011-09-29 | 2013-04-04 | Cisco Technology, Inc. | Techniques for securely unlocking a touch screen user device |
US20130174083A1 (en) * | 2011-12-28 | 2013-07-04 | Acer Incorporated | Electronic device and method of controlling the same |
US9727031B2 (en) | 2012-04-13 | 2017-08-08 | Tk Holdings Inc. | Pressure sensor including a pressure sensitive material for use with control systems and methods of using the same |
US20130278540A1 (en) * | 2012-04-20 | 2013-10-24 | Esat Yilmaz | Inter Touch Sensor Communications |
US20130298079A1 (en) * | 2012-05-02 | 2013-11-07 | Pantech Co., Ltd. | Apparatus and method for unlocking an electronic device |
US11204653B2 (en) * | 2012-06-29 | 2021-12-21 | Samsung Electronics Co., Ltd | Method and device for handling event invocation using a stylus pen |
US10831289B2 (en) * | 2012-06-29 | 2020-11-10 | Samsung Electronics Co., Ltd | Method and device for handling event invocation using a stylus pen |
US9857889B2 (en) * | 2012-06-29 | 2018-01-02 | Samsung Electronic Co., Ltd | Method and device for handling event invocation using a stylus pen |
US20140002397A1 (en) * | 2012-06-29 | 2014-01-02 | Samsung Electronics Co., Ltd. | Method and device for handling event invocation using a stylus pen |
US20140006965A1 (en) * | 2012-07-02 | 2014-01-02 | Beijing Xiaomi Technology Co., Ltd. | Method of entering a user interface in a device and the device thereof |
US9823834B2 (en) * | 2012-07-11 | 2017-11-21 | Guang Dong Oppo Mobile Telecommunications., Ltd. | Quick access gesture setting and accessing method for a touch control device |
US20140033140A1 (en) * | 2012-07-11 | 2014-01-30 | Guang Dong Oppo Mobile Telecommunications Corp., Ltd. | Quick access function setting method for a touch control device |
US20140026105A1 (en) * | 2012-07-18 | 2014-01-23 | Research In Motion Limited | Method and Apparatus Pertaining to a Gesture-Controlled Snooze Instruction |
US20140028598A1 (en) * | 2012-07-30 | 2014-01-30 | Samsung Electronics Co., Ltd | Apparatus and method for controlling data transmission in terminal |
US9696223B2 (en) | 2012-09-17 | 2017-07-04 | Tk Holdings Inc. | Single layer force sensor |
US20140298672A1 (en) * | 2012-09-27 | 2014-10-09 | Analog Devices Technology | Locking and unlocking of contacless gesture-based user interface of device having contactless gesture detection system |
FR2998071A1 (en) * | 2012-11-09 | 2014-05-16 | Thales Sa | METHOD FOR SECURING A CONTROL ON A TOUCH-SURFACE VISUALIZATION DEVICE AND SYSTEM THEREOF |
EP2731002A1 (en) * | 2012-11-09 | 2014-05-14 | Thales | Method for securing a command on a viewing device with tactile surface and related system |
US9582122B2 (en) | 2012-11-12 | 2017-02-28 | Microsoft Technology Licensing, Llc | Touch-sensitive bezel techniques |
US10656750B2 (en) | 2012-11-12 | 2020-05-19 | Microsoft Technology Licensing, Llc | Touch-sensitive bezel techniques |
US9389762B2 (en) * | 2012-11-29 | 2016-07-12 | Egalax_Empia Technology Inc. | Method for unlocking touch screen, electronic device thereof, and recording medium thereof |
US20140145990A1 (en) * | 2012-11-29 | 2014-05-29 | Egalax_Empia Technology Inc. | Method for unlocking touch screen, electronic device thereof, and recording medium thereof |
US20150317076A1 (en) * | 2012-12-11 | 2015-11-05 | University Of Washington Through Its Center For Commercialization | Use of hand posture to improve text entry |
US10540083B2 (en) * | 2012-12-11 | 2020-01-21 | University Of Washington | Use of hand posture to improve text entry |
US11294484B1 (en) * | 2012-12-31 | 2022-04-05 | Allscripts Software, Llc | Method for ensuring use intentions of a touch screen device |
US9158399B2 (en) * | 2013-03-13 | 2015-10-13 | Htc Corporation | Unlock method and mobile device using the same |
DE102013211011B4 (en) * | 2013-03-13 | 2017-05-24 | Htc Corporation | Unlock method and mobile device using same |
US20140267064A1 (en) * | 2013-03-13 | 2014-09-18 | Htc Corporation | Unlock Method and Mobile Device Using the Same |
US20140347302A1 (en) * | 2013-05-24 | 2014-11-27 | Boe Technology Group Co., Ltd. | Unlocking method and device for touch screen terminal and touch screen terminal |
US9400595B2 (en) * | 2013-05-24 | 2016-07-26 | Boe Technology Group Co., Ltd. | Unlocking method and device for touch screen terminal and touch screen terminal |
EP2835732A3 (en) * | 2013-08-06 | 2015-04-22 | Samsung Electronics Co., Ltd | Electronic device provided with touch screen and operating method thereof |
US10776479B2 (en) * | 2013-08-28 | 2020-09-15 | Paypal, Inc. | Authentication system |
US20150066762A1 (en) * | 2013-08-28 | 2015-03-05 | Geoffrey W. Chatterton | Authentication system |
US20150199553A1 (en) * | 2014-01-13 | 2015-07-16 | Samsung Electronics Co., Ltd. | Method for recognizing fingerprint and mobile terminal supporting the same |
US10101843B2 (en) * | 2014-01-13 | 2018-10-16 | Samsung Electronics Co., Ltd. | Method for recognizing fingerprint and mobile terminal supporting the same |
US9521241B2 (en) | 2014-03-13 | 2016-12-13 | Lg Electronics Inc. | Mobile terminal and method of controlling the same |
EP2919108A1 (en) * | 2014-03-13 | 2015-09-16 | LG Electronics, Inc. | Mobile terminal and method of controlling the same |
CN104915133A (en) * | 2014-03-13 | 2015-09-16 | Lg电子株式会社 | Mobile terminal and method of controlling the same |
US9477337B2 (en) | 2014-03-14 | 2016-10-25 | Microsoft Technology Licensing, Llc | Conductive trace routing for display and bezel sensors |
US9946383B2 (en) | 2014-03-14 | 2018-04-17 | Microsoft Technology Licensing, Llc | Conductive trace routing for display and bezel sensors |
WO2015149101A3 (en) * | 2014-04-01 | 2015-11-26 | Linkilike Gmbh | Method for comparing user identities of databases |
US9645668B2 (en) | 2014-07-22 | 2017-05-09 | Lg Electronics Inc. | Mobile terminal and method for controlling the same |
EP2977884A1 (en) * | 2014-07-22 | 2016-01-27 | LG Electronics Inc. | Mobile terminal and method for controlling the same |
US20160026380A1 (en) * | 2014-07-25 | 2016-01-28 | Hannstar Display (Nanjing) Corporation | Hand-held electronic device, computer-executed method and touch-sensing cover |
US20160034172A1 (en) * | 2014-07-30 | 2016-02-04 | Wistron Corporation | Touch device and control method and method for determining unlocking thereof |
US9727233B2 (en) * | 2014-07-30 | 2017-08-08 | Wistron Corporation | Touch device and control method and method for determining unlocking thereof |
US10678381B2 (en) * | 2014-08-21 | 2020-06-09 | DISH Technologies L.L.C. | Determining handedness on multi-element capacitive devices |
US20160054827A1 (en) * | 2014-08-21 | 2016-02-25 | Echostar Technologies L.L.C. | Determining handedness on multi-element capacitive devices |
US10048862B2 (en) * | 2014-09-08 | 2018-08-14 | Lenovo (Singapore) Pte. Ltd. | Managing an on-screen keyboard |
US20160070465A1 (en) * | 2014-09-08 | 2016-03-10 | Lenovo (Singapore) Pte, Ltd. | Managing an on-screen keyboard |
US10319408B2 (en) | 2015-03-30 | 2019-06-11 | Manufacturing Resources International, Inc. | Monolithic display with separately controllable sections |
US10922736B2 (en) | 2015-05-15 | 2021-02-16 | Manufacturing Resources International, Inc. | Smart electronic display for restaurants |
US10725624B2 (en) | 2015-06-05 | 2020-07-28 | Apple Inc. | Movement between multiple views |
US10467610B2 (en) | 2015-06-05 | 2019-11-05 | Manufacturing Resources International, Inc. | System and method for a redundant multi-panel electronic display |
US10269156B2 (en) | 2015-06-05 | 2019-04-23 | Manufacturing Resources International, Inc. | System and method for blending order confirmation over menu board background |
JP2017140167A (en) * | 2016-02-09 | 2017-08-17 | 株式会社コナミデジタルエンタテインメント | Game system, game control device, and program |
US10319271B2 (en) | 2016-03-22 | 2019-06-11 | Manufacturing Resources International, Inc. | Cyclic redundancy check for electronic displays |
US10705697B2 (en) * | 2016-03-31 | 2020-07-07 | Brother Kogyo Kabushiki Kaisha | Information processing apparatus configured to edit images, non-transitory computer-readable medium storing instructions therefor, and information processing method for editing images |
US20170285921A1 (en) * | 2016-03-31 | 2017-10-05 | Brother Kogyo Kabushiki Kaisha | Information processing apparatus,non-transitory computer-readable medium storing instructions therefor, and information processing method |
FR3049733A1 (en) * | 2016-04-01 | 2017-10-06 | Thales Sa | METHOD FOR SECURELY CONTROLLING A FUNCTION USING A TOUCH SLAB |
US10649649B2 (en) | 2016-04-01 | 2020-05-12 | Thales | Secure method for control of a function by means of a touchpad |
WO2017167698A1 (en) * | 2016-04-01 | 2017-10-05 | Thales | Secure method for control of a function by means of a touchpad |
US10313037B2 (en) | 2016-05-31 | 2019-06-04 | Manufacturing Resources International, Inc. | Electronic display remote image verification system and method |
US10756836B2 (en) | 2016-05-31 | 2020-08-25 | Manufacturing Resources International, Inc. | Electronic display remote image verification system and method |
US10510304B2 (en) | 2016-08-10 | 2019-12-17 | Manufacturing Resources International, Inc. | Dynamic dimming LED backlight for LCD array |
US10156934B2 (en) * | 2016-08-12 | 2018-12-18 | Canon Kabushiki Kaisha | Electronic device and control method thereof |
US11301128B2 (en) * | 2019-05-01 | 2022-04-12 | Google Llc | Intended input to a user interface from detected gesture positions |
US11164508B2 (en) * | 2019-05-16 | 2021-11-02 | Asustek Computer Inc. | Electronic device |
US10990251B1 (en) * | 2019-11-08 | 2021-04-27 | Sap Se | Smart augmented reality selector |
US11895362B2 (en) | 2021-10-29 | 2024-02-06 | Manufacturing Resources International, Inc. | Proof of play for images displayed at electronic displays |
Also Published As
Publication number | Publication date |
---|---|
TW200907742A (en) | 2009-02-16 |
WO2008138030A1 (en) | 2008-11-20 |
US20170285853A1 (en) | 2017-10-05 |
US9696847B2 (en) | 2017-07-04 |
US20160253032A1 (en) | 2016-09-01 |
US20120233690A1 (en) | 2012-09-13 |
US9280279B2 (en) | 2016-03-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9696847B2 (en) | User-defined gesture enablement protocols for touch input device | |
US10228833B2 (en) | Input device user interface enhancements | |
JP3589381B2 (en) | Virtual pointing device generation method, apparatus, and computer system | |
EP3066605B1 (en) | Directional touch unlocking for electronic devices | |
JP4384734B2 (en) | Virtual pointing device generation instruction method, computer system, and apparatus | |
JP3504462B2 (en) | Pointing device generation instruction method and computer system | |
CN103765368B (en) | Mobile terminal for performing screen unlock based on motion and method thereof | |
EP2506178B1 (en) | User interface, touch-controlled device and method for authenticating a user of a touch-controlled device | |
KR20120003441A (en) | Bimodal touch sensitive digital notebook | |
JPH1063425A (en) | Method for instracting generation of virtual pointing device, and computer system | |
JPH1063422A (en) | Method for generating at least two virtual pointing device, and computer system | |
CN107407998B (en) | Method for interacting with an electronic and/or computer device implementing a capacitive control surface and a peripheral surface, interface and device implementing the method | |
KR101349526B1 (en) | Automatic teller machine and display method | |
KR20120081422A (en) | Terminal having touch screen and method for inputting letter according to touch event thereof | |
US10237736B2 (en) | Unlocking of a mobile device by a code received via a stencil on a touchscreen | |
KR102647593B1 (en) | Method of unlocking device by performing fingerprint verification and moving gesture and the device | |
CN109634417A (en) | A kind of processing method and electronic equipment | |
KR101706909B1 (en) | Finger Input Devices | |
KR20010102638A (en) | Keyboard system having easy key operation function and method thereof | |
TW202009687A (en) | Method for unlocking a display of a multi-display device | |
WO2015016894A1 (en) | Gesture detection |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: RPO PTY LIMITED, AUSTRALIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ATKINS, GRAHAM ROY;MAXWELL, IAN ANDREW;REEL/FRAME:021279/0291 Effective date: 20080606 |
|
AS | Assignment |
Owner name: BRIDGE BANK, NATIONAL ASSOCIATION, CALIFORNIA Free format text: SECURITY AGREEMENT;ASSIGNOR:RPO PTY LTD;REEL/FRAME:024838/0948 Effective date: 20100813 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |
|
AS | Assignment |
Owner name: RPO PTY LTD, CALIFORNIA Free format text: REASSIGNMENT AND RELEASE OF IP SECURITY INTEREST;ASSIGNOR:BRIDGE BANK, NATIONAL ASSOCIATION;REEL/FRAME:028737/0963 Effective date: 20120802 |
|
AS | Assignment |
Owner name: TRINITY CAPITAL INVESTMENT LLC, ARIZONA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:RPO PTY LTD;REEL/FRAME:029770/0739 Effective date: 20120628 Owner name: ZETTA RESEARCH AND DEVELOPMENT LLC - RPO SERIES, D Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TRINITY CAPITAL INVESTMENT LLC;REEL/FRAME:029770/0778 Effective date: 20120629 |