US20100164897A1 - Virtual keypad systems and methods - Google Patents

Virtual keypad systems and methods Download PDF

Info

Publication number
US20100164897A1
US20100164897A1 US12/666,916 US66691608A US2010164897A1 US 20100164897 A1 US20100164897 A1 US 20100164897A1 US 66691608 A US66691608 A US 66691608A US 2010164897 A1 US2010164897 A1 US 2010164897A1
Authority
US
United States
Prior art keywords
keys
touchpad
keypad
key
zone
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/666,916
Inventor
Philippe Morin
David Kryze
Luca Rigazio
Peter Veprek
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Panasonic Corp
Original Assignee
Panasonic Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Panasonic Corp filed Critical Panasonic Corp
Priority to US12/666,916 priority Critical patent/US20100164897A1/en
Publication of US20100164897A1 publication Critical patent/US20100164897A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/02Input arrangements using manually operated switches, e.g. using keyboards or dials
    • G06F3/023Arrangements for converting discrete items of information into a coded form, e.g. arrangements for interpreting keyboard generated codes as alphanumeric codes, operand codes or instruction codes
    • G06F3/0233Character input methods
    • G06F3/0237Character input methods using prediction or retrieval techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/16Constructional details or arrangements
    • G06F1/1613Constructional details or arrangements for portable computers
    • G06F1/1633Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
    • G06F1/1684Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
    • G06F1/1698Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being a sending/receiving arrangement to establish a cordless communication link, e.g. radio or infrared link, integrated cellular phone
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0346Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04886Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42204User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42204User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
    • H04N21/42206User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor characterized by hardware details
    • H04N21/42224Touch pad or touch panel provided on the remote control
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42204User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
    • H04N21/42226Reprogrammable remote control devices
    • H04N21/42227Reprogrammable remote control devices the keys being reprogrammable, e.g. soft keys
    • H04N21/42228Reprogrammable remote control devices the keys being reprogrammable, e.g. soft keys the reprogrammable keys being displayed on a display screen in order to reduce the number of keys on the remote control device itself
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42204User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
    • H04N21/42206User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor characterized by hardware details
    • H04N21/42222Additional components integrated in the remote control device, e.g. timer, speaker, sensors for detecting position, direction or movement of the remote control, microphone or battery charging device

Definitions

  • the present invention relates to methods and systems for recognizing text input from a remote controller.
  • remote control devices can communicate a variety of commands to the electronic product.
  • the remote control devices have become complex to operate.
  • modern television remote controls can include selection buttons for volume, channel selection, menu selection, and picture viewing.
  • selection buttons for volume, channel selection, menu selection, and picture viewing.
  • the user To operate the remote control, the user must take time away from the program he or she is watching to focus in on the buttons. This can be very distracting to a viewer.
  • a virtual keypad system for inputting text includes a remote controller having at least one touchpad incorporated therein and divided into a plurality of touch zones.
  • a display device is in data communication with the remote controller and is operable to display a user interface including a keypad, where each key of the keypad is mapped to a touch zone of the touchpad.
  • a prediction module in response to an operator pressing a given touch zone to select a particular character, performs one or more key prediction methods to predict one or more next plausible keys.
  • a key mapping module remaps the touch zones of the touchpad to the keys of the keypad based on the one or more next plausible keys.
  • FIG. 1 is an illustration of a text input system according to various aspects of the present disclosure.
  • FIG. 2A is an illustration of a remote controller of the text input system of FIG. 1 that includes a touchpad according to various aspects of the present disclosure.
  • FIG. 2B is a block diagram illustrating internal components of the remote controller of FIG. 2A according to various aspects of the present disclosure.
  • FIG. 3 is a dataflow diagram illustrating a virtual keypad module of the text input system of FIG. 1 according to various aspects of the present disclosure.
  • FIG. 4A is an illustration of the remote controller being held in a portrait position according to various aspects of the present disclosure.
  • FIG. 4B is an illustration of the remote controller being held in a landscape position according to various aspects of the present disclosure.
  • FIG. 5 is a table illustrating a mapping between XY coordinates of the touchpad and keys of a keypad.
  • FIG. 6 is a table illustrating a remapping between the coordinates of the touchpad and the keys of a keypad.
  • FIG. 7 is an illustration of a first embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
  • FIG. 8A is an illustration of a second embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
  • FIG. 8B is an illustration of a third embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
  • FIG. 9 is an illustration of a fourth embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
  • FIG. 10 is an illustration of a fifth embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
  • module or sub-module can refer to a processor (shared, dedicated, or group) and memory that executes one or more software or firmware to programs, and/or other suitable components that can provide the described functionality and/or combinations thereof.
  • FIG. 1 depicts an exemplary text input system 10 implemented according to various aspects of the present disclosure.
  • the exemplary text input system 10 includes a virtual keypad module 12 that facilitates the input of alphanumeric characters by a user for interacting with various services delivered through a display device 14 .
  • the display device 14 can be, but is not limited to, a television (as shown), a projector and screen, or a computer.
  • the services can be, for example, internet based services such as, online shopping and movie subscriptions.
  • the virtual keypad module 12 provides feedback to the user via a graphical user interface (GUI) 18 .
  • the GUI 18 includes a virtual keypad 20 , as will be discussed in more detail below.
  • the virtual keypad module 12 can be implemented within the display device 14 . In various other embodiments, the virtual keypad module 12 can be implemented separate from the display device 14 (such as, for example, on a set top box (not shown)) and can be in data communication with the display device 14 . For ease of the discussion, the remainder of the disclosure will be discussed in the context of the virtual keypad module 12 being implemented within the display device 14 .
  • the text input system 10 further includes a remote controller 16 that generates one or more signals to the display device 14 in response to user input.
  • the virtual keypad module 12 receives and processes the signals. Based on the signals, the virtual keypad module 12 determines an orientation and a holding position of the remote controller 16 , recognizes text input, and/or provides visual feedback to the user via a graphical user interface (GUI) 18 .
  • GUI graphical user interface
  • the virtual keypad module 12 implements selection auto-correction methods that compensate for human typing (i.e., clicking) error. For example, when attempting to input text quickly, users can typically undershoot or overshoot the location and click on a nearby unintended key.
  • the virtual keypad module 12 employs a combination of prediction and auto-correction methods to determine which character(s) is/are most likely to be entered by the user.
  • a prediction method is used to compensate for the possible overshoot and undershoot.
  • the predictions can be used to enlarge an activation area of possible keys while reducing (or zero-ing) activation areas of keys that are not in the next-character prediction list.
  • the virtual keypad module 12 disables the selection auto-correction methods and reverts to a default mode (i.e., without enlarging or reducing the activation area).
  • the virtual keypad module 12 can also disable the selection auto-correction method when the interaction becomes slow because it is assumed that clicking errors do not generally occur during slow interaction.
  • FIGS. 2A and 2B illustrate an exemplary remote controller 16 according to various aspects of the present disclosure.
  • the exterior of the remote controller 16 includes a touchpad 22 and one or more soft keys 24 a - 24 d.
  • touch zones defined by one or more coordinates of the touchpad 22 can be mapped to a particular key of the virtual keypad 20 ( FIG. 1 ).
  • a user can select a particular key of the virtual keypad 20 by gently placing his finger or thumb on the touchpad 22 at or near the associated touch zone (FingerDown event), by dragging a finger or thumb along the touchpad 22 to the associated touch zone (FingerDrag event), and/or by lifting the finger or thumb away from the touchpad 22 (FingerUp event). While the user has a finger or thumb on the touchpad 22 (i.e., between FingerDown and FingerUp events), the user can click on the touchpad 22 by applying greater force (FingerPress event) followed by releasing the force (FingerRelease event) to select a key.
  • a relative access method can be used as an alternative or as a secondary method for selecting keys.
  • the relative access method assumes a position of the user's finger or thumb to be a current coordinate or touch zone of the touchpad 22 . Subsequent gestures by the user are then interpreted relative to that coordinate or touch zone. This allows for an adjustable precision in selection.
  • Important functions of the remote controller 16 can be associated with specific selection buttons 26 a - 26 d of the touchpad 22 .
  • the selection buttons 26 a - 26 d can be designated by a specific touchpad button that is painted or illuminated on the touchpad 22 (as shown) or by a button displayed on an overlay to the virtual keypad 20 ( FIG. 1 ) of the GUI 18 ( FIG. 1 ). This allows the user to use the remote controller 16 in complete darkness without having to look away from the content being displayed on the display device 14 ( FIG. 1 ).
  • the functions can be controlled by simply touching the buttons or be controlled by performing a specific gesture.
  • sliding a finger or thumb up or down on the right side of the touchpad 22 can trigger a volume up or volume down action.
  • sliding a finger or thumb right or left on the top side of the touchpad 22 can trigger a channel up or channel down action.
  • the body of the remote controller 16 can be made of a soft material, allowing the remote controller 16 to be squeezed.
  • the squeezing of the remote controller 16 can be performed by the user to trigger certain actions, particularly in contexts where the GUI 18 ( FIG. 1 ) is just waiting for an acknowledgement without proposing a choice (such as a “next” button in a slideshow).
  • the internal components of the remote controller 16 can include, but are not limited to, input sensors 30 , output actuators 32 , an input controller 34 , an output controller 36 , a processing handler 38 , a wireless transmitter (e.g., RF, Bluetooth, etc.) 40 , and/or combinations thereof.
  • the inputs sensors 30 can include touchpad sensors 42 .
  • the touchpad sensors 42 can be single-position registering touchpad sensors mounted side-by-side that allow for the selection of at least two contact points on the touchpad 22 ( FIG. 2A ) simultaneously.
  • the touchpad sensors 42 can be a single multi-touch capable touchpad sensor that can register, with equal precision, two points of contact at the same time.
  • the touchpad sensors 42 can register pressure information to allow the touchpad 22 ( FIG. 2A ) to be clickable.
  • the input sensors 30 can also include one or more selection button sensors 44 , one or more touchpad button sensors 46 , one or more accelerometers 48 , and one or more holding sensors 50 .
  • the holding sensors 50 can be, for example, capacitive sensors that are located around the border of the remote controller 16 , and/or behind the remote controller 16 .
  • the holding sensors 50 indicate whether the user is touching an area of the remote controller 16 in a proximity of the holding sensor 50 .
  • the accelerometer 48 can be a three-axis accelerometer that indicates a positioning of the remote controller 16 .
  • the input controller 34 reads the real-time data from all active sensors. In various embodiments, some sensors may not be active at all times to reduce power consumption.
  • the processing handler 38 gathers and forms into packets the data to be transmitted and/or processes the real-time data from one or more active sensors to perform local actions.
  • the RF transmitter 40 (RF driver 40 ) generates the signals in packet form to the display device 14 ( FIG. 1 ).
  • the output actuators 32 can include one or more LED panels 52 for displaying the touchpad buttons 26 a - 26 d, depending on the specific state of interaction with the GUI 18 present on-screen.
  • the output actuators 32 can additionally or alternatively include actuators for providing sufficient haptic feedback to the user (such as, for example, vibration actuators 54 , light actuators 55 , and/or speaker actuators 56 ).
  • the above has described one example of the operations performed by the output actuator 32 .
  • the output controller 36 updates the state of all the active actuators.
  • FIG. 3 a dataflow diagram illustrates a more detailed exemplary virtual keypad module 12 .
  • the virtual keypad module 12 may include any number of sub-modules. As can be appreciated, the sub-modules shown in FIG. 3 may be combined and/or further partitioned to similarly perform text input.
  • the data inputs 70 , 72 , 74 , and 76 to the virtual keypad module 12 are received from the remote controller 16 ( FIG. 1 ) and/or received from other modules (not shown) within the display device 14 ( FIG. 1 ).
  • the virtual keypad module 12 includes an orientation recognition module 60 , a hand position recognition module 62 , a prediction module 64 , a key input module 66 , a key mapping module 67 , and a GUI manager module 68 .
  • the orientation recognition module 60 determines an orientation of the remote controller 16 based on data received from the holding sensors 50 ( FIG. 2B ) and the accelerometer 48 ( FIG. 2B ). For example, the user can be holding the remote controller 16 ( FIG. 2A ) in a portrait position, as shown in FIG. 4A , or in a landscape position, as shown in FIG. 4B .
  • the orientation recognition module 60 determines the orientation by way of an Artificial Neural Network (ANN).
  • the ANN can be trained by data indicating both landscape position conditions and portrait position conditions.
  • the orientation is determined by training an ANN with sensory data.
  • the sensory data can comprise three-dimensional acceleration (accx, accy, accz) and an activation state of the n capacitive holding position sensors, which can signal that human skin is ever in proximity (1) or is not in proximity (0). These n+3 values are fed into a single perceptron or linear classifier to determine if the remote controller 16 ( FIG. 2A ) is horizontal or vertical. Perceptron coefficients can be trained on a database and hard-coded by a manufacturer.
  • the hand position recognition module 62 determines a holding style of the remote controller 16 ( FIG. 2A ) based on data received from the holding sensors 50 ( FIG. 2B ) and the accelerometer 48 ( FIG. 2B ).
  • the sensory data 70 , 72 can be used to determine whether the remote controller 16 ( FIG. 2A ) is held with one or two hands; and if it is held with one hand, whether it is held with the left or right hand.
  • the hand position recognition module 62 determines the holding style by way of an ANN.
  • the ANN can be trained by data indicating right-hand conditions, left-hand conditions, and two hands conditions.
  • the hand position is determined similarly as discussed above.
  • Multiple perceptrons can be implemented for the more than one binary decisions (e.g., left hand, right hand, two handed).
  • the determination of the orientation and the holding style gives the virtual keypad module 12 the ability to accommodate the user by automatically adapting the text input methods and the look and feel of the GUI 18 ( FIG. 1 ).
  • the determination of the orientation and holding position allows the user to hold the remote controller 16 ( FIG. 2A ) in the most convenient way based on their personal preference and the actual conditions of use (e.g., standing, sitting, lying down).
  • the user can pick up and operate the remote controller 16 ( FIG. 2A ) without worrying about how they are holding it.
  • the hand position recognition module 62 can further perform user verification based on a combination of holding sensor data 70 , accelerometer data 72 , additional sensor information (such as an image of the palm of the user's hand), and/or bio-sensors.
  • the data can be used to fully determine the identity of the user or, more broadly, infer the category to which the user belongs (e.g., left-handed, right-handed, kid, adult, elderly).
  • User identification can be used, for example, for parental control, personalization, and profile switching.
  • User categorization can be used to adapt the GUI 18 ( FIG. 1 ).
  • the key mapping module 67 generates a map indicating an association between the coordinates or touch zones of the touchpad 22 ( FIG. 2A ) and the keys of the keypad and/or touchpad selection buttons. In various embodiments, the key mapping module 67 generates the map based on the orientation and hand position information determined from the orientation recognition module 60 and the hand position recognition module 62 , respectively. The key mapping module 67 maps the touch zones of the touchpad 22 to the keys of the keypad, by associating a key of the keypad to each XY coordinate of the touch zones. Here, the touch zones are defined by one or more XY coordinates of the touchpad 22 . The key mapping module 67 generates a map from the mapping.
  • the map is a two-dimensional lookup table defined by the coordinates of the touchpad 22 .
  • the map can be a two dimensional (XY) table 80 that is used to assign a key of the keypad and/or a touchpad selection button to each coordinate of the touchpad 22 ( FIG. 2A ).
  • the map can then be referenced by the key input module 66 to determine an action to be taken and can be referenced by the GUI manager module 68 to generate the GUI 18 ( FIG. 1 ).
  • the key input module 66 processes touchpad sensor data 74 and/or the accelerometer data 72 .
  • the key input module 66 interprets the touchpad sensor data 74 to be a coordinate or coordinates of the touchpad 22 ( FIG. 2A ) and determines what action to be taken based on the coordinate or coordinates.
  • the key input module 66 can receive the touchpad sensor data 74 , determine a particular coordinate from the data 74 , and reference the map generated by the key mapping module 67 . Based on the entries in the map, the key input module 66 , for example, can project that the user is hovering over a particular key of the keypad and, thus, entering a particular text.
  • the key input module 66 interprets the accelerometer data 72 as an action to be taken.
  • the accelerometer data 72 can indicate if a user has lifted the remote controller 16 ( FIG. 2A ) quickly to select, for example, an uppercase mode.
  • the accelerometer data 72 can indicate when a user has lowered the remote controller 16 ( FIG. 2A ) quickly to select, for example, a lowercase mode.
  • the prediction module 64 To enhance the precision and speed at which the text is entered, the prediction module 64 generates a prediction of which key and/or word the user is trying to select. The prediction module 64 generates the prediction based on the touchpad sensor data 74 and/or based on a determination of previous text entered. In various embodiments, the prediction module 64 performs one or more next key prediction methods, such as, for example, a language model method, a trajectory analysis method, a hand movement analysis method, a timing analysis method, and/or combinations thereof. In short, the prediction module 64 , in response to an operator pressing a given touch zone to select a particular character, performs one or more key prediction methods to predict one or more next plausible keys. Examples of the operator are a finger of the user, a touch-pen, and the like, that operates the touchpad 22 .
  • the prediction module 64 employs one or more language models known in the art to predict the next key based on previous text entered.
  • the language models for instance, predict the one or more next plausible keys based on previous characters selected by the operator. For example, if the partial word ‘pr’ has been entered, the language model can predict that a vowel is likely to follow and that the letter ‘r’ will not be a possibility.
  • the prediction module 64 employs one or more language models to provide a list of reliable candidates of full words from partial word inputs.
  • the language model generates a list of plausible words based on the previous characters selected by the operator.
  • the full words can be selected by the user for auto-completion. For example, if the partial word ‘Pan’ has been entered, a list can be generated that includes ‘Panasonic’ and ‘Pan-American.’ Instead of typing the remaining characters, the user can simply select one of the full words.
  • the language model can generate the word predictions based on words previously entered. For example, once selected, the words can be remembered and the language model can be adapted to favor the remembered words.
  • the trajectory analysis method can be performed to predict possible next keys based on future path estimation including directions and/or velocities of user's finger or thumb movement on the touchpad 22 ( FIG. 2A ) as indicated by the touchpad sensor data 74 .
  • the trajectory analysis method predicts next plausible keys based on a direction of operator movement on the touchpad 22 . For example, if the user first selects the ‘k’ key and the language model predicts that the next key can be one of ‘I,’ ‘e,’ or ‘a,’ the touchpad sensor data 74 can be evaluated to determine a direction the user is heading and velocity of the movement and, thus, eliminate one or more of the choices.
  • the trajectory analysis method determines a coordinate of the key ‘k’ and subsequent finger movements. From that history of XY coordinates, the future path is determined. The path includes a tolerance to account for short-term prediction (more accurate) and longer-term prediction (less accurate). If the future path estimation is heading away from the coordinates of the predicted key, the choice is eliminated. For example, if the path is heading on an Up/Left diagonal line, then the keys ‘e’ and ‘a’ are eliminated and the key ‘I’ is selected as the predicted next key.
  • the hand movement analysis method can be performed using the holding style information provided by the hand position recognition module 62 and the predictions provided by the prediction module 64 .
  • the hand movement analysis method predicts next plausible keys based on a detection of which hand is moving on the touchpad 22 .
  • the hand movement analysis method can evaluate whether the remote controller 16 ( FIG. 2A ) is held with two hands or one hand. If the remote controller 16 ( FIG. 2A ) is held by two hands and movement by a right hand or left hand is detected, then the choices that are associated with the non-moving hand would be eliminated.
  • the prediction module 64 predicts that the next key can be one of ‘e,’ or ‘a,’ and movement is detected by the right hand, then the keys ‘e’ and ‘a’ are eliminated and the key ‘I’ is selected as the predicted next key.
  • the timing analysis method can be performed when the prediction module 64 is unable to predict the appropriate next key or word via any of the next key prediction methods. Such may be the case when the user is entering proper nouns, such as a last name or a password.
  • the timing analysis method evaluates the time the user takes to move from one key to the next. That is, the timing analysis method predicts next plausible keys based on a timing of operator movement on the touchpad 22 . In more detail, the timing analysis method predicts no next plausible keys when the timing exceeds a predetermined limit.
  • the key mapping module 67 remaps the touch zones of the touchpad 22 to the keys of the keypad based on the no next plausible keys. If the user moves more slowly, it is more likely that a proper noun is being entered. The predictions are then ignored.
  • the key mapping module 67 can remap the coordinates of the touchpad 22 ( FIG. 2A ) to the keys of the virtual keypad 20 ( FIG. 1 ). That is, the key mapping module 67 remaps the touch zones of the touchpad 22 to the keys of the keypad based on the one or more next plausible keys predicted by the prediction module 64 . In various embodiments, the coordinates that are associated with the predicted next key can be expanded to make the key more accessible. For example, as shown in FIG.
  • the map can be adjusted such that the coordinates that were previously mapped to the keys ‘o’ or ‘I’ are now mapped to the key ‘p.’
  • the ‘p’ key will be selected and entered if the user clicks on the touchpad 22 ( FIG. 2A ).
  • the key mapping module 67 can remap the coordinates to the keys based on a relative speed of the user movement. For example, the key mapping module 67 can remap the coordinates such that the predicted key is mapped to a larger touch zone area when a faster movement is detected. The key mapping module 67 can maintain the original coordinates or remap to the original coordinates when slower movements are detected. The key mapping module 67 can scale the area between the larger area and the original area when a movement that is not fast or slow is detected.
  • the GUI manager module 68 generates GUI data 76 for displaying the GUI 18 .
  • the GUI 18 provides visual feedback to the user indicating the actions they have performed with the remote controller 16 ( FIG. 2A ).
  • the GUI 18 can include the virtual keypad 20 , including multiple alphanumeric keys 90 , one or more selection buttons 92 , a selection list 94 , a text display box 96 , a title bar 98 , and/or combinations thereof.
  • the touchpad 22 is divided into two operating zones (a first operating zone and a second operating zone).
  • the key mapping module 67 maps a first subset of keys of the keypad to the first operating zone of the touchpad 22 and maps a second subset of keys of the keypad to the second operating zone of the touchpad 22 .
  • the first operating zone corresponds to a top zone or a right zone of the touchpad 22
  • the second operating zone corresponds to a bottom zone or a left zone of the touchpad 22
  • each of the first subset and the second subset is a group of keys of the keypad.
  • the GUI manager module 68 displays the virtual keypad 20 based on the holding position and the orientation determined by the hand position recognition module 62 and the orientation recognition module 60 , respectively. For example, as shown in FIG. 7 , if two hands are used to hold the remote controller 16 ( FIG. 2A ) in the landscape position, the virtual keypad 20 and the touchpad 22 can be divided into two zones 100 , 102 (e.g., left and right). The keys in the right zone 102 can be associated with a first zone 104 of the touchpad 22 , and the keys in the left zone 100 can be associated with a second zone 106 of the touchpad 22 .
  • the keys in the right zone 102 can be associated with a first zone 104 of the touchpad 22
  • the keys in the left zone 100 can be associated with a second zone 106 of the touchpad 22 .
  • the user can select a key 90 in the right zone 102 with a right thumb or finger, and the user can select a key 90 in the left zone 100 with a left thumb or finger.
  • the keys 90 in each zone 100 , 102 can be distinguished by a particular color or shading of the keys 90 so that the user can determine which keys 90 can be selected with which thumb or finger. This will allow for a natural text input, similar to the experience when entering text on classic keyboards.
  • selection of a predicted word can be made through the display of the two most probable words (e.g., ‘Panasonic,’ ‘Pan-America’).
  • the selection buttons 92 are auto-completion selection buttons. When the user selects one of the selection buttons 92 , a probable word displayed in the selected selection button 92 is displayed as a complete word on the display box 96 .
  • the two words can be displayed on selection buttons 92 .
  • the user can select the selection buttons 92 by pushing soft keys 24 b, 24 d located on the top side of the remote controller 16 with the index fingers, or by dragging the finger or thumb to a dedicated zone located at a designated location of the touchpad 22 .
  • the mapping of the coordinates can provide for an overlap between the two areas.
  • the key mapping module 67 maps a third subset of keys of the first subset of keys to the second operating zone and maps a fourth subset of keys of the second subset of keys to the first operating zone.
  • the third subset of keys include the keys designated by the letters ‘t,’ ‘g,’ and ‘b’ of the keys of the first subset (keys included in the zone 100 of FIG. 7 ).
  • the fourth subset of keys include the keys designated by the letters ‘y,’ ‘h,’ and ‘n’ of the keys of the second subset (keys included in the zone 102 of FIG. 7 ).
  • the keys along a boundary between the divided zones can operate for any zones of the touchpad 22 .
  • the letter ‘g’ in left keyboard area can be selected by the touchpad 22 in the first zone, as well as be selected by the touchpad 22 in the second zone 106 .
  • the overlap keys can be identified on the GUI 18 by shading or color.
  • the touchpad 22 can be divided into two zones 108 , 110 or 112 , 114 .
  • the virtual keypad 20 is not divided.
  • a first zone 110 or 114 of the touchpad 22 can be associated with the entire virtual keypad 20 , and can be referred to as the “Key-Entry Zone.”
  • the second zone 108 or 112 of the touchpad 22 can be associated with the auto-completion buttons 92 ( FIG.
  • selection lists 94 auto-completion selection list
  • the auto-completion buttons 92 or selection lists 94 can be displayed when the user touches the touchpad 22 in the “Auto-Completion Zone.”
  • the user interface can include a small visual notification to signal the availability and status of auto-completion words. This way the user will not be bothered with auto-completion unless he decides to use it.
  • the zones 108 , 110 or 112 , 114 of the touchpad 22 can be configured based on an identification of the user. For example, in the case of a right-handed user, the right zone 110 can be associated with the “Key-Entry Zone” used most often and the left zone 108 can be associated with the “Auto-Completion Zone.” Similarly, for a left-handed user, the left zone 108 can be associated with the “Key-Entry Zone” used most often and the right zone 110 can be associated with the “Auto-Completion Zone.”
  • the touchpad 22 is divided into zones.
  • the virtual keypad 20 is not divided into zones. Only one zone of the touchpad 22 is used by the user at all times. The user will be able to address the entire virtual keypad 20 from the entire touchpad 22 .
  • Such touchpad 22 will principally function as the “Key-Entry Zone.” If auto-completion is needed, the user can switch the touchpad 22 to the “Auto-completion Zone” by using a simple gesture.
  • Such gesture can include, but is not limited to, moving the thumb or finger to a specific area of the touchpad 22 (for instance lower right), or sliding the finger or thumb along the touchpad 22 quickly from right to left.
  • the user can select a key 90 by dragging the thumb or finger on the touchpad 22 .
  • one or more on-screen pointers (such as, for example, a cursor, or an illustration of a thumb or finger) slides to a target key 90 .
  • the on-screen pointers can be displayed according to the hand position (e.g., left hand only, right hand only, or both hands).
  • the key 90 can be selected by clicking the clickable touchpad 22 and/or upon release.
  • a thumb or finger posture can be used to indicate a thumb or finger press as opposed to a thumb or finger that is dragged on the touchpad surface.
  • the selected character associated with the key will be displayed in the text display box 96 .
  • the GUI manager module 68 displays the keys 90 of the virtual keypad 20 based on the predicted next key and the remapping of the coordinates performed by the key mapping module 67 ( FIG. 3 ). For example, by knowing the next likely key, the GUI manager module 68 ( FIG. 3 ) can highlight and/or enlarge the most likely key 116 based on the mapping of the coordinates. That is, an arrangement of the keypad on the user interface is modified based on the one or more next plausible keys predicted by the prediction module 64 . In more detail, one or more keys of the keypad on the user interface is highlighted or enlarged based on the one or more next plausible keys predicted by the prediction module 64 .

Abstract

Accordingly, a virtual keypad system for inputting text is provided. A virtual keypad system includes a remote controller having at least one touchpad incorporated therein and divided into a plurality of touch zones. A display device is in data communication with the remote controller and is operable to display a user interface including a keypad, where each key of the keypad is mapped to a touch zone of the touchpad. A prediction module, in response to an operator pressing a given touch zone to select a particular character, performs one or more key prediction methods to predict one or more next plausible keys. A key mapping module remaps the touch zones of the touchpad to the keys of the keypad based on the one or more next plausible keys.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application claims the benefit of U.S. patent applications Nos. 60/946858 and 11/977346, filed on Jun. 28, 2007 and Oct. 24, 2007, respectively. The disclosures of the above applications are incorporated herein by reference.
  • FIELD
  • The present invention relates to methods and systems for recognizing text input from a remote controller.
  • BACKGROUND ART
  • Many electronic consumer products come with remote control devices. These remote control devices can communicate a variety of commands to the electronic product. With the rise in technological advancements to the electronic products, the remote control devices have become complex to operate. For example, modern television remote controls can include selection buttons for volume, channel selection, menu selection, and picture viewing. To operate the remote control, the user must take time away from the program he or she is watching to focus in on the buttons. This can be very distracting to a viewer.
  • In addition, many Internet based services such as online shopping are now being provided through the television. Additional selection buttons such as keypad buttons must be included on the remote control device to accommodate these new services. The new selection buttons serve to increase the complexity as well as the cost of the remote control devices. Various solutions have been proposed to address such problems. One solution is disclosed in U.S. Pat. No. 6,765,557 to use a touchpad for controlling a home entertainment device such as an Interactive television. However, even this solution cannot completely solve the problems of inconveniency to the user inconvenience.
  • The statements in this section merely provide background information related to the present disclosure and may not constitute prior art.
  • SUMMARY
  • A virtual keypad system for inputting text is provided. A virtual keypad system includes a remote controller having at least one touchpad incorporated therein and divided into a plurality of touch zones. A display device is in data communication with the remote controller and is operable to display a user interface including a keypad, where each key of the keypad is mapped to a touch zone of the touchpad. A prediction module, in response to an operator pressing a given touch zone to select a particular character, performs one or more key prediction methods to predict one or more next plausible keys. A key mapping module remaps the touch zones of the touchpad to the keys of the keypad based on the one or more next plausible keys.
  • Further areas of applicability will become apparent from the description provided herein. It should be understood that the description and specific examples are intended for purposes of illustration only and are not intended to limit the scope of the present disclosure.
  • BRIEF DESCRIPTION OF DRAWINGS
  • The drawings described herein are for illustration purposes only and are not intended to limit the scope of the present teachings in any way.
  • FIG. 1 is an illustration of a text input system according to various aspects of the present disclosure.
  • FIG. 2A is an illustration of a remote controller of the text input system of FIG. 1 that includes a touchpad according to various aspects of the present disclosure.
  • FIG. 2B is a block diagram illustrating internal components of the remote controller of FIG. 2A according to various aspects of the present disclosure.
  • FIG. 3 is a dataflow diagram illustrating a virtual keypad module of the text input system of FIG. 1 according to various aspects of the present disclosure.
  • FIG. 4A is an illustration of the remote controller being held in a portrait position according to various aspects of the present disclosure.
  • FIG. 4B is an illustration of the remote controller being held in a landscape position according to various aspects of the present disclosure.
  • FIG. 5 is a table illustrating a mapping between XY coordinates of the touchpad and keys of a keypad.
  • FIG. 6 is a table illustrating a remapping between the coordinates of the touchpad and the keys of a keypad.
  • FIG. 7 is an illustration of a first embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
  • FIG. 8A is an illustration of a second embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
  • FIG. 8B is an illustration of a third embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
  • FIG. 9 is an illustration of a fourth embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
  • FIG. 10 is an illustration of a fifth embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
  • DETAILED DESCRIPTION
  • The following description is merely exemplary in nature and is not intended to limit the present teachings, their application, or uses. It should be understood that throughout the drawings, corresponding reference numerals indicate like or corresponding parts and features. As used herein, the term module or sub-module can refer to a processor (shared, dedicated, or group) and memory that executes one or more software or firmware to programs, and/or other suitable components that can provide the described functionality and/or combinations thereof.
  • Referring now to FIG. 1, FIG. 1 depicts an exemplary text input system 10 implemented according to various aspects of the present disclosure. The exemplary text input system 10 includes a virtual keypad module 12 that facilitates the input of alphanumeric characters by a user for interacting with various services delivered through a display device 14. The display device 14 can be, but is not limited to, a television (as shown), a projector and screen, or a computer. The services can be, for example, internet based services such as, online shopping and movie subscriptions. The virtual keypad module 12 provides feedback to the user via a graphical user interface (GUI) 18. The GUI 18 includes a virtual keypad 20, as will be discussed in more detail below.
  • In various embodiments, the virtual keypad module 12 can be implemented within the display device 14. In various other embodiments, the virtual keypad module 12 can be implemented separate from the display device 14 (such as, for example, on a set top box (not shown)) and can be in data communication with the display device 14. For ease of the discussion, the remainder of the disclosure will be discussed in the context of the virtual keypad module 12 being implemented within the display device 14.
  • The text input system 10 further includes a remote controller 16 that generates one or more signals to the display device 14 in response to user input. The virtual keypad module 12 receives and processes the signals. Based on the signals, the virtual keypad module 12 determines an orientation and a holding position of the remote controller 16, recognizes text input, and/or provides visual feedback to the user via a graphical user interface (GUI) 18. In particular, the virtual keypad module 12 implements selection auto-correction methods that compensate for human typing (i.e., clicking) error. For example, when attempting to input text quickly, users can typically undershoot or overshoot the location and click on a nearby unintended key. The virtual keypad module 12 employs a combination of prediction and auto-correction methods to determine which character(s) is/are most likely to be entered by the user.
  • In one example, provided fast input speeds, a prediction method is used to compensate for the possible overshoot and undershoot. As will be discussed in more detail below, the predictions can be used to enlarge an activation area of possible keys while reducing (or zero-ing) activation areas of keys that are not in the next-character prediction list. However, if the prediction methods are unable to generate a prediction, even at fast input speeds, the virtual keypad module 12 disables the selection auto-correction methods and reverts to a default mode (i.e., without enlarging or reducing the activation area). The virtual keypad module 12 can also disable the selection auto-correction method when the interaction becomes slow because it is assumed that clicking errors do not generally occur during slow interaction.
  • FIGS. 2A and 2B illustrate an exemplary remote controller 16 according to various aspects of the present disclosure. As shown in FIG. 2A, the exterior of the remote controller 16 includes a touchpad 22 and one or more soft keys 24 a-24 d. In various embodiments, touch zones defined by one or more coordinates of the touchpad 22 can be mapped to a particular key of the virtual keypad 20 (FIG. 1).
  • A user can select a particular key of the virtual keypad 20 by gently placing his finger or thumb on the touchpad 22 at or near the associated touch zone (FingerDown event), by dragging a finger or thumb along the touchpad 22 to the associated touch zone (FingerDrag event), and/or by lifting the finger or thumb away from the touchpad 22 (FingerUp event). While the user has a finger or thumb on the touchpad 22 (i.e., between FingerDown and FingerUp events), the user can click on the touchpad 22 by applying greater force (FingerPress event) followed by releasing the force (FingerRelease event) to select a key.
  • In various other embodiments, a relative access method can be used as an alternative or as a secondary method for selecting keys. The relative access method assumes a position of the user's finger or thumb to be a current coordinate or touch zone of the touchpad 22. Subsequent gestures by the user are then interpreted relative to that coordinate or touch zone. This allows for an adjustable precision in selection.
  • Important functions of the remote controller 16 (such as, for example, volume, channel, and mute) can be associated with specific selection buttons 26 a-26 d of the touchpad 22. The selection buttons 26 a-26 d can be designated by a specific touchpad button that is painted or illuminated on the touchpad 22 (as shown) or by a button displayed on an overlay to the virtual keypad 20 (FIG. 1) of the GUI 18 (FIG. 1). This allows the user to use the remote controller 16 in complete darkness without having to look away from the content being displayed on the display device 14 (FIG. 1).
  • The functions can be controlled by simply touching the buttons or be controlled by performing a specific gesture. In one example, sliding a finger or thumb up or down on the right side of the touchpad 22 can trigger a volume up or volume down action. In another example, sliding a finger or thumb right or left on the top side of the touchpad 22 can trigger a channel up or channel down action.
  • In various embodiments, the body of the remote controller 16 can be made of a soft material, allowing the remote controller 16 to be squeezed. The squeezing of the remote controller 16 can be performed by the user to trigger certain actions, particularly in contexts where the GUI 18 (FIG. 1) is just waiting for an acknowledgement without proposing a choice (such as a “next” button in a slideshow).
  • As shown in FIG. 2B, the internal components of the remote controller 16 can include, but are not limited to, input sensors 30, output actuators 32, an input controller 34, an output controller 36, a processing handler 38, a wireless transmitter (e.g., RF, Bluetooth, etc.) 40, and/or combinations thereof. The following describes operations performed by the sensors including the input sensors 30. The inputs sensors 30 can include touchpad sensors 42. The touchpad sensors 42 can be single-position registering touchpad sensors mounted side-by-side that allow for the selection of at least two contact points on the touchpad 22 (FIG. 2A) simultaneously. Alternatively, the touchpad sensors 42 can be a single multi-touch capable touchpad sensor that can register, with equal precision, two points of contact at the same time. In various embodiments, the touchpad sensors 42 can register pressure information to allow the touchpad 22 (FIG. 2A) to be clickable.
  • The input sensors 30 can also include one or more selection button sensors 44, one or more touchpad button sensors 46, one or more accelerometers 48, and one or more holding sensors 50. The holding sensors 50 can be, for example, capacitive sensors that are located around the border of the remote controller 16, and/or behind the remote controller 16. The holding sensors 50 indicate whether the user is touching an area of the remote controller 16 in a proximity of the holding sensor 50. The accelerometer 48 can be a three-axis accelerometer that indicates a positioning of the remote controller 16. The input controller 34 reads the real-time data from all active sensors. In various embodiments, some sensors may not be active at all times to reduce power consumption. The processing handler 38 gathers and forms into packets the data to be transmitted and/or processes the real-time data from one or more active sensors to perform local actions. The RF transmitter 40 (RF driver 40) generates the signals in packet form to the display device 14 (FIG. 1).
  • The output actuators 32 can include one or more LED panels 52 for displaying the touchpad buttons 26 a-26 d, depending on the specific state of interaction with the GUI 18 present on-screen. The output actuators 32 can additionally or alternatively include actuators for providing sufficient haptic feedback to the user (such as, for example, vibration actuators 54, light actuators 55, and/or speaker actuators 56). The above has described one example of the operations performed by the output actuator 32. The output controller 36 updates the state of all the active actuators.
  • Referring now to FIG. 3, a dataflow diagram illustrates a more detailed exemplary virtual keypad module 12. Various embodiments of the virtual keypad module 12 according to the present disclosure may include any number of sub-modules. As can be appreciated, the sub-modules shown in FIG. 3 may be combined and/or further partitioned to similarly perform text input. The data inputs 70, 72, 74, and 76 to the virtual keypad module 12 are received from the remote controller 16 (FIG. 1) and/or received from other modules (not shown) within the display device 14 (FIG. 1). In various embodiments, the virtual keypad module 12 includes an orientation recognition module 60, a hand position recognition module 62, a prediction module 64, a key input module 66, a key mapping module 67, and a GUI manager module 68. The following describes one example of operations performed by each module. The orientation recognition module 60 determines an orientation of the remote controller 16 based on data received from the holding sensors 50 (FIG. 2B) and the accelerometer 48 (FIG. 2B). For example, the user can be holding the remote controller 16 (FIG. 2A) in a portrait position, as shown in FIG. 4A, or in a landscape position, as shown in FIG. 4B. In various embodiments, the orientation recognition module 60 determines the orientation by way of an Artificial Neural Network (ANN). The ANN can be trained by data indicating both landscape position conditions and portrait position conditions.
  • In one example, the orientation is determined by training an ANN with sensory data. The sensory data can comprise three-dimensional acceleration (accx, accy, accz) and an activation state of the n capacitive holding position sensors, which can signal that human skin is ever in proximity (1) or is not in proximity (0). These n+3 values are fed into a single perceptron or linear classifier to determine if the remote controller 16 (FIG. 2A) is horizontal or vertical. Perceptron coefficients can be trained on a database and hard-coded by a manufacturer. The hand position recognition module 62 determines a holding style of the remote controller 16 (FIG. 2A) based on data received from the holding sensors 50 (FIG. 2B) and the accelerometer 48 (FIG. 2B). For example, the sensory data 70, 72 can be used to determine whether the remote controller 16 (FIG. 2A) is held with one or two hands; and if it is held with one hand, whether it is held with the left or right hand. In various embodiments, the hand position recognition module 62 determines the holding style by way of an ANN. The ANN can be trained by data indicating right-hand conditions, left-hand conditions, and two hands conditions.
  • In one example, the hand position is determined similarly as discussed above. Multiple perceptrons can be implemented for the more than one binary decisions (e.g., left hand, right hand, two handed).
  • As will be discussed in more detail below, the determination of the orientation and the holding style gives the virtual keypad module 12 the ability to accommodate the user by automatically adapting the text input methods and the look and feel of the GUI 18 (FIG. 1). Thus, the determination of the orientation and holding position allows the user to hold the remote controller 16 (FIG. 2A) in the most convenient way based on their personal preference and the actual conditions of use (e.g., standing, sitting, lying down). In the case of operating the remote controller 16 (FIG. 2A) in a dark room, the user can pick up and operate the remote controller 16 (FIG. 2A) without worrying about how they are holding it.
  • The hand position recognition module 62 can further perform user verification based on a combination of holding sensor data 70, accelerometer data 72, additional sensor information (such as an image of the palm of the user's hand), and/or bio-sensors. The data can be used to fully determine the identity of the user or, more broadly, infer the category to which the user belongs (e.g., left-handed, right-handed, kid, adult, elderly). User identification can be used, for example, for parental control, personalization, and profile switching. User categorization can be used to adapt the GUI 18 (FIG. 1).
  • The key mapping module 67 generates a map indicating an association between the coordinates or touch zones of the touchpad 22 (FIG. 2A) and the keys of the keypad and/or touchpad selection buttons. In various embodiments, the key mapping module 67 generates the map based on the orientation and hand position information determined from the orientation recognition module 60 and the hand position recognition module 62, respectively. The key mapping module 67 maps the touch zones of the touchpad 22 to the keys of the keypad, by associating a key of the keypad to each XY coordinate of the touch zones. Here, the touch zones are defined by one or more XY coordinates of the touchpad 22. The key mapping module 67 generates a map from the mapping. Here, the map is a two-dimensional lookup table defined by the coordinates of the touchpad 22. For example, as shown in FIG. 5, the map can be a two dimensional (XY) table 80 that is used to assign a key of the keypad and/or a touchpad selection button to each coordinate of the touchpad 22 (FIG. 2A). As will be discussed in more detail below, the map can then be referenced by the key input module 66 to determine an action to be taken and can be referenced by the GUI manager module 68 to generate the GUI 18 (FIG. 1).
  • Referring back to FIG. 3, the key input module 66 processes touchpad sensor data 74 and/or the accelerometer data 72. In various embodiments, the key input module 66 interprets the touchpad sensor data 74 to be a coordinate or coordinates of the touchpad 22 (FIG. 2A) and determines what action to be taken based on the coordinate or coordinates. For example, the key input module 66 can receive the touchpad sensor data 74, determine a particular coordinate from the data 74, and reference the map generated by the key mapping module 67. Based on the entries in the map, the key input module 66, for example, can project that the user is hovering over a particular key of the keypad and, thus, entering a particular text.
  • In various embodiments, the key input module 66 interprets the accelerometer data 72 as an action to be taken. For example, the accelerometer data 72 can indicate if a user has lifted the remote controller 16 (FIG. 2A) quickly to select, for example, an uppercase mode. The accelerometer data 72 can indicate when a user has lowered the remote controller 16 (FIG. 2A) quickly to select, for example, a lowercase mode.
  • To enhance the precision and speed at which the text is entered, the prediction module 64 generates a prediction of which key and/or word the user is trying to select. The prediction module 64 generates the prediction based on the touchpad sensor data 74 and/or based on a determination of previous text entered. In various embodiments, the prediction module 64 performs one or more next key prediction methods, such as, for example, a language model method, a trajectory analysis method, a hand movement analysis method, a timing analysis method, and/or combinations thereof. In short, the prediction module 64, in response to an operator pressing a given touch zone to select a particular character, performs one or more key prediction methods to predict one or more next plausible keys. Examples of the operator are a finger of the user, a touch-pen, and the like, that operates the touchpad 22.
  • In one example, the prediction module 64 employs one or more language models known in the art to predict the next key based on previous text entered. The language models, for instance, predict the one or more next plausible keys based on previous characters selected by the operator. For example, if the partial word ‘pr’ has been entered, the language model can predict that a vowel is likely to follow and that the letter ‘r’ will not be a possibility.
  • In another example, the prediction module 64 employs one or more language models to provide a list of reliable candidates of full words from partial word inputs. In this case, the language model generates a list of plausible words based on the previous characters selected by the operator. The full words can be selected by the user for auto-completion. For example, if the partial word ‘Pan’ has been entered, a list can be generated that includes ‘Panasonic’ and ‘Pan-American.’ Instead of typing the remaining characters, the user can simply select one of the full words.
  • In various embodiments, the language model can generate the word predictions based on words previously entered. For example, once selected, the words can be remembered and the language model can be adapted to favor the remembered words.
  • In yet another example, the trajectory analysis method can be performed to predict possible next keys based on future path estimation including directions and/or velocities of user's finger or thumb movement on the touchpad 22 (FIG. 2A) as indicated by the touchpad sensor data 74. The trajectory analysis method predicts next plausible keys based on a direction of operator movement on the touchpad 22. For example, if the user first selects the ‘k’ key and the language model predicts that the next key can be one of ‘I,’ ‘e,’ or ‘a,’ the touchpad sensor data 74 can be evaluated to determine a direction the user is heading and velocity of the movement and, thus, eliminate one or more of the choices.
  • For example, the trajectory analysis method determines a coordinate of the key ‘k’ and subsequent finger movements. From that history of XY coordinates, the future path is determined. The path includes a tolerance to account for short-term prediction (more accurate) and longer-term prediction (less accurate). If the future path estimation is heading away from the coordinates of the predicted key, the choice is eliminated. For example, if the path is heading on an Up/Left diagonal line, then the keys ‘e’ and ‘a’ are eliminated and the key ‘I’ is selected as the predicted next key.
  • In another example, the hand movement analysis method can be performed using the holding style information provided by the hand position recognition module 62 and the predictions provided by the prediction module 64. The hand movement analysis method predicts next plausible keys based on a detection of which hand is moving on the touchpad 22. The hand movement analysis method can evaluate whether the remote controller 16 (FIG. 2A) is held with two hands or one hand. If the remote controller 16 (FIG. 2A) is held by two hands and movement by a right hand or left hand is detected, then the choices that are associated with the non-moving hand would be eliminated.
  • For example, if the user first selects the ‘k’ key, the prediction module 64 predicts that the next key can be one of ‘e,’ or ‘a,’ and movement is detected by the right hand, then the keys ‘e’ and ‘a’ are eliminated and the key ‘I’ is selected as the predicted next key.
  • In yet another example, the timing analysis method can be performed when the prediction module 64 is unable to predict the appropriate next key or word via any of the next key prediction methods. Such may be the case when the user is entering proper nouns, such as a last name or a password. The timing analysis method evaluates the time the user takes to move from one key to the next. That is, the timing analysis method predicts next plausible keys based on a timing of operator movement on the touchpad 22. In more detail, the timing analysis method predicts no next plausible keys when the timing exceeds a predetermined limit. Here, the key mapping module 67 remaps the touch zones of the touchpad 22 to the keys of the keypad based on the no next plausible keys. If the user moves more slowly, it is more likely that a proper noun is being entered. The predictions are then ignored.
  • Based on the predictions provided by the prediction module 64, the key mapping module 67 can remap the coordinates of the touchpad 22 (FIG. 2A) to the keys of the virtual keypad 20 (FIG. 1). That is, the key mapping module 67 remaps the touch zones of the touchpad 22 to the keys of the keypad based on the one or more next plausible keys predicted by the prediction module 64. In various embodiments, the coordinates that are associated with the predicted next key can be expanded to make the key more accessible. For example, as shown in FIG. 6, if the predicted next key is ‘p,’ the map can be adjusted such that the coordinates that were previously mapped to the keys ‘o’ or ‘I’ are now mapped to the key ‘p.’ Thus, if the user is actually hovering over the ‘I’ key, the ‘p’ key will be selected and entered if the user clicks on the touchpad 22 (FIG. 2A).
  • Referring back to FIG. 3, in various embodiments, the key mapping module 67 can remap the coordinates to the keys based on a relative speed of the user movement. For example, the key mapping module 67 can remap the coordinates such that the predicted key is mapped to a larger touch zone area when a faster movement is detected. The key mapping module 67 can maintain the original coordinates or remap to the original coordinates when slower movements are detected. The key mapping module 67 can scale the area between the larger area and the original area when a movement that is not fast or slow is detected.
  • The GUI manager module 68 generates GUI data 76 for displaying the GUI 18. The GUI 18 provides visual feedback to the user indicating the actions they have performed with the remote controller 16 (FIG. 2A). As shown in the exemplary GUIs 18 of FIGS. 7 through 10, the GUI 18 can include the virtual keypad 20, including multiple alphanumeric keys 90, one or more selection buttons 92, a selection list 94, a text display box 96, a title bar 98, and/or combinations thereof.
  • As discussed further below, the touchpad 22 is divided into two operating zones (a first operating zone and a second operating zone). The key mapping module 67 maps a first subset of keys of the keypad to the first operating zone of the touchpad 22 and maps a second subset of keys of the keypad to the second operating zone of the touchpad 22. For example, the first operating zone corresponds to a top zone or a right zone of the touchpad 22, and the second operating zone corresponds to a bottom zone or a left zone of the touchpad 22. Here, each of the first subset and the second subset is a group of keys of the keypad. In various embodiments, the GUI manager module 68 displays the virtual keypad 20 based on the holding position and the orientation determined by the hand position recognition module 62 and the orientation recognition module 60, respectively. For example, as shown in FIG. 7, if two hands are used to hold the remote controller 16 (FIG. 2A) in the landscape position, the virtual keypad 20 and the touchpad 22 can be divided into two zones 100, 102 (e.g., left and right). The keys in the right zone 102 can be associated with a first zone 104 of the touchpad 22, and the keys in the left zone 100 can be associated with a second zone 106 of the touchpad 22. The user can select a key 90 in the right zone 102 with a right thumb or finger, and the user can select a key 90 in the left zone 100 with a left thumb or finger. The keys 90 in each zone 100, 102 can be distinguished by a particular color or shading of the keys 90 so that the user can determine which keys 90 can be selected with which thumb or finger. This will allow for a natural text input, similar to the experience when entering text on classic keyboards.
  • In this example, selection of a predicted word (auto-completion) can be made through the display of the two most probable words (e.g., ‘Panasonic,’ ‘Pan-America’). The selection buttons 92 are auto-completion selection buttons. When the user selects one of the selection buttons 92, a probable word displayed in the selected selection button 92 is displayed as a complete word on the display box 96. For example, the two words can be displayed on selection buttons 92. The user can select the selection buttons 92 by pushing soft keys 24 b, 24 d located on the top side of the remote controller 16 with the index fingers, or by dragging the finger or thumb to a dedicated zone located at a designated location of the touchpad 22.
  • In various embodiments, when the touchpad 22 and the virtual keypad 20 are divided into two zones, the mapping of the coordinates can provide for an overlap between the two areas. The key mapping module 67 maps a third subset of keys of the first subset of keys to the second operating zone and maps a fourth subset of keys of the second subset of keys to the first operating zone. For instance, the third subset of keys include the keys designated by the letters ‘t,’ ‘g,’ and ‘b’ of the keys of the first subset (keys included in the zone 100 of FIG. 7). Likewise, the fourth subset of keys include the keys designated by the letters ‘y,’ ‘h,’ and ‘n’ of the keys of the second subset (keys included in the zone 102 of FIG. 7). That enables the keys along a boundary between the divided zones to operate for any zones of the touchpad 22. For example, the letter ‘g’ in left keyboard area can be selected by the touchpad 22 in the first zone, as well as be selected by the touchpad 22 in the second zone 106. The overlap keys can be identified on the GUI 18 by shading or color.
  • Referring now to FIGS. 8A and 8B, in another example, if two hands are used to hold the remote controller 16 in the landscape position, the touchpad 22 can be divided into two zones 108, 110 or 112, 114. The virtual keypad 20, however, is not divided. A first zone 110 or 114 of the touchpad 22 can be associated with the entire virtual keypad 20, and can be referred to as the “Key-Entry Zone.” The second zone 108 or 112 of the touchpad 22 can be associated with the auto-completion buttons 92 (FIG. 7) or selection lists 94 (auto-completion selection list), and can be referred to as the “Auto-Completion Zone.” The auto-completion buttons 92 or selection lists 94 can be displayed when the user touches the touchpad 22 in the “Auto-Completion Zone.” The user interface can include a small visual notification to signal the availability and status of auto-completion words. This way the user will not be bothered with auto-completion unless he decides to use it.
  • In various embodiments, the zones 108, 110 or 112, 114 of the touchpad 22 can be configured based on an identification of the user. For example, in the case of a right-handed user, the right zone 110 can be associated with the “Key-Entry Zone” used most often and the left zone 108 can be associated with the “Auto-Completion Zone.” Similarly, for a left-handed user, the left zone 108 can be associated with the “Key-Entry Zone” used most often and the right zone 110 can be associated with the “Auto-Completion Zone.”
  • Referring now to FIG. 9, in yet another example, if one hand is used to hold the remote controller 16 in the portrait position, the touchpad 22 is divided into zones. The virtual keypad 20, however, is not divided into zones. Only one zone of the touchpad 22 is used by the user at all times. The user will be able to address the entire virtual keypad 20 from the entire touchpad 22. Such touchpad 22 will principally function as the “Key-Entry Zone.” If auto-completion is needed, the user can switch the touchpad 22 to the “Auto-completion Zone” by using a simple gesture. Such gesture can include, but is not limited to, moving the thumb or finger to a specific area of the touchpad 22 (for instance lower right), or sliding the finger or thumb along the touchpad 22 quickly from right to left.
  • In any of the examples shown in FIGS. 7 through 9, the user can select a key 90 by dragging the thumb or finger on the touchpad 22. In response to the movement, one or more on-screen pointers (such as, for example, a cursor, or an illustration of a thumb or finger) slides to a target key 90. The on-screen pointers can be displayed according to the hand position (e.g., left hand only, right hand only, or both hands). The key 90 can be selected by clicking the clickable touchpad 22 and/or upon release. When displaying a thumb or finger as the pointer, a different thumb or finger posture can be used to indicate a thumb or finger press as opposed to a thumb or finger that is dragged on the touchpad surface. The selected character associated with the key will be displayed in the text display box 96.
  • Referring now to FIG. 10, in various embodiments, the GUI manager module 68 displays the keys 90 of the virtual keypad 20 based on the predicted next key and the remapping of the coordinates performed by the key mapping module 67 (FIG. 3). For example, by knowing the next likely key, the GUI manager module 68 (FIG. 3) can highlight and/or enlarge the most likely key 116 based on the mapping of the coordinates. That is, an arrangement of the keypad on the user interface is modified based on the one or more next plausible keys predicted by the prediction module 64. In more detail, one or more keys of the keypad on the user interface is highlighted or enlarged based on the one or more next plausible keys predicted by the prediction module 64. These operations are performed by the GUI manager module 68. However, when the user is not moving the cursor quickly or the user hovers over a given coordinate, the highlighted and/or enlarged key 116 is remapped to the original coordinates of the touchpad 22 and the highlighted and/or enlarged key can be resized back to the original size.
  • Those skilled in the art can now appreciate from the foregoing description that the broad teachings of the present disclosure can be implemented in a variety of forms. Therefore, while this disclosure has been described in connection with particular examples thereof, the true scope of the disclosure should not be so limited since other modifications will become apparent to the skilled practitioner upon a study of the drawings, specification, and the following claims.

Claims (24)

1. A virtual keypad system for inputting text, comprising:
a remote controller having at least one touchpad incorporated therein and divided into a plurality of touch zones;
a display device in data communication with the remote controller and operable to display a user interface including a keypad, where each key of the keypad is mapped to a touch zone of the touchpad;
a prediction module, in response to an operator pressing a given touch zone to select a particular character, performs one or more key prediction methods to predict one or more next plausible keys; and
a key mapping module remaps the touch zones of the touchpad to the keys of the keypad based on the one or more next plausible keys.
2. The system of claim 1,
wherein an arrangement of the keypad on the user interface is modified based on the one or more next plausible keys.
3. The system of claim 2,
wherein one or more keys of the keypad on the user interface is enlarged based on the one or more next plausible keys.
4. The system of claim 2,
wherein one or more keys of the keypad on the user interface is highlighted based on the one or more next plausible keys.
5. The system of claim 1 further comprising
a language model operable to predict the one or more next plausible keys based on previous characters selected by the operator.
6. The system of claim 5,
wherein the language model is further operable to generate a list of plausible words based on the previous characters selected by the operator.
7. The system of claim 1,
wherein the one or more key prediction methods includes a trajectory analysis method, where the next plausible key is predicted based on a direction of operator movement on the touchpad.
8. The system of claim 1,
wherein the one or more key prediction methods includes a hand movement analysis method, where the next plausible key is predicted based on a detection of which hand is moving on the touchpad.
9. The system of claim 1,
wherein the touchpad is divided into two operating zones and wherein the key mapping module maps a first subset of keys of the keypad to a first operating zone of the touchpad and maps a second subset of keys of the keypad to a second operating zone of the touchpad.
10. The system of claim 9,
wherein the first operating zone corresponds to a top zone of the touchpad and wherein the second operating zone corresponds to a bottom zone of the touchpad.
11. The system of claim 9,
wherein the first operating zone corresponds to a right zone of the touchpad and wherein the second operating zone corresponds to a left zone of the touch pad.
12. The system of claim 1,
wherein the user interface includes at least one of auto-completion selection buttons and auto-completion selection list.
13. The system of claim 12,
wherein the touchpad is divided into two operating zones and wherein the key mapping module maps a first operating zone to the keys of the keypad and maps a second operating zone to the at least one of the of auto-completion selection buttons and auto-completion selection.
14. The system of claim 9,
wherein the key mapping module maps a third subset of keys of the first subset of keys to the second operating zone and maps a fourth subset of keys of the second subset of keys to the first operating zone.
15. The system of claim 14,
wherein the third subset of keys include the keys designated. by the letters ‘t,’ ‘g,’ and ‘b,’ and wherein the fourth subset of keys include the keys designated by the letters ‘y,’ ‘h,’ and ‘n.’
16. The system of claim 1,
wherein the key prediction methods include a timing analysis method, where the next plausible keys are predicted based on a timing of operator movement on the touchpad.
17. The system of claim 16,
wherein the timing analysis method predicts no next plausible keys when the timing exceeds a predetermined limit.
18. The system of claim 17,
wherein the key mapping module remaps the touch zones of the touchpad to the keys of the keypad based on the no next plausible keys.
19. The system of claim 1,
wherein the touch zones are defined by one or more XY coordinates of the touchpad and wherein the key mapping module maps the touch zones of the touchpad to the keys of the keypad by associating a key of the keypad to each XY coordinate of the touch zones.
20. The system of claim 19,
wherein the key mapping modules generates a map from the mapping and wherein the map is a two-dimensional lookup table defined by the coordinates of the touchpad.
21. A virtual keypad system for inputting text, comprising:
a remote controller having at least one touchpad incorporated therein and divided into a plurality of touch zones;
a display device in data communication with the remote controller and operable to display a keypad and an area for displaying input from the keypad, where each key on the keypad is associated with a touch zone on the touchpad and, in response to an operator pressing a given touch zone, a character on the key correlating to the given touch zone is displayed in an input area of the display;
a language model adapted to receive characters displayed in the input area and operable to predict next most plausible characters in string of received characters; and
a module that arranges the each key on the keypad for facilitating entry of a desired key based on output from the language model.
22. The system of claim 21,
wherein the arranging comprises enlarging keys on the keypad based on the output from the language model.
23. A virtual keypad system for inputting text, comprising:
a remote controller having at least one touchpad incorporated therein and divided into a plurality of touch zones;
a display device in data communication with the remote controller and operable to display a user interface including a keypad, where each key of the keypad is mapped to a touch zone of the touchpad;
a prediction module performs one or more key prediction methods to predict one or more next plausible keys;
a key mapping module remaps the touch zones of the touchpad to the keys of the keypad based on the one or more next plausible keys;
a user interface manager module modifies an arrangement of the keypad based on the remapping of the touch zones of the touchpad to the keys of the keypad; and
a text input module, in response to an operator pressing a given touch zone, selects a key based on the remapping of the touch zones of the touchpad to the keys of the keypad.
24. The system of claim 23,
wherein the user interface manager module enlarges one or more keys of the keypad based on the remapping of the touch zones of the touchpad to the keys of the keypad.
US12/666,916 2007-06-28 2008-06-26 Virtual keypad systems and methods Abandoned US20100164897A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/666,916 US20100164897A1 (en) 2007-06-28 2008-06-26 Virtual keypad systems and methods

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US94685807P 2007-06-28 2007-06-28
US11/977,346 US8065624B2 (en) 2007-06-28 2007-10-24 Virtual keypad systems and methods
PCT/US2008/068384 WO2009006221A1 (en) 2007-06-28 2008-06-26 Virtual keypad systems and methods
US12/666,916 US20100164897A1 (en) 2007-06-28 2008-06-26 Virtual keypad systems and methods

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US11/977,346 Continuation US8065624B2 (en) 2007-06-28 2007-10-24 Virtual keypad systems and methods

Publications (1)

Publication Number Publication Date
US20100164897A1 true US20100164897A1 (en) 2010-07-01

Family

ID=40162292

Family Applications (2)

Application Number Title Priority Date Filing Date
US11/977,346 Active 2030-01-08 US8065624B2 (en) 2007-06-28 2007-10-24 Virtual keypad systems and methods
US12/666,916 Abandoned US20100164897A1 (en) 2007-06-28 2008-06-26 Virtual keypad systems and methods

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US11/977,346 Active 2030-01-08 US8065624B2 (en) 2007-06-28 2007-10-24 Virtual keypad systems and methods

Country Status (4)

Country Link
US (2) US8065624B2 (en)
EP (1) EP2160672A4 (en)
JP (1) JP5323070B2 (en)
WO (1) WO2009006221A1 (en)

Cited By (276)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080055269A1 (en) * 2006-09-06 2008-03-06 Lemay Stephen O Portable Electronic Device for Instant Messaging
US20090005011A1 (en) * 2007-06-28 2009-01-01 Greg Christie Portable Electronic Device with Conversation Management for Incoming Instant Messages
US20090124293A1 (en) * 2007-11-08 2009-05-14 Tsui Yuan-Mao Method for displaying dialing information and mobile communication device using the method
US20090177981A1 (en) * 2008-01-06 2009-07-09 Greg Christie Portable Electronic Device for Instant Messaging Multiple Recipients
US20100185960A1 (en) * 2003-05-02 2010-07-22 Apple Inc. Method and Apparatus for Displaying Information During an Instant Messaging Session
US20100228539A1 (en) * 2009-03-06 2010-09-09 Motorola, Inc. Method and apparatus for psychomotor and psycholinguistic prediction on touch based device
US20100315266A1 (en) * 2009-06-15 2010-12-16 Microsoft Corporation Predictive interfaces with usability constraints
US20110012843A1 (en) * 2009-07-14 2011-01-20 Chih-Hung Li Touch-controlled electronic apparatus and related control method
US20110148762A1 (en) * 2009-12-22 2011-06-23 Universal Electronics Inc. System and method for multi-mode command input
US20110187647A1 (en) * 2010-02-04 2011-08-04 Charles Howard Woloszynski Method and apparatus for virtual keyboard interactions from secondary surfaces
US20110201387A1 (en) * 2010-02-12 2011-08-18 Microsoft Corporation Real-time typing assistance
US20110212876A1 (en) * 2010-02-10 2011-09-01 Michelle Meek Cleaning composition comprising amylase variants with high stability in the presence of a chelating agent
US20110279397A1 (en) * 2009-01-26 2011-11-17 Zrro Technologies (2009) Ltd. Device and method for monitoring the object's behavior
US20110310019A1 (en) * 2010-06-16 2011-12-22 International Business Machines Corporation Reconfiguration of virtual keyboard
US20120078614A1 (en) * 2010-09-27 2012-03-29 Primesense Ltd. Virtual keyboard for a non-tactile three dimensional user interface
US20120098743A1 (en) * 2010-10-26 2012-04-26 Pei-Ling Lai Input method, input device, and computer system
US20120137244A1 (en) * 2010-11-30 2012-05-31 Inventec Corporation Touch device input device and operation method of the same
CN102937876A (en) * 2011-11-23 2013-02-20 微软公司 Dynamic scaling of a touch sensor
US20130181902A1 (en) * 2012-01-17 2013-07-18 Microsoft Corporation Skinnable touch device grip patterns
US20130229363A1 (en) * 2012-03-02 2013-09-05 Christopher A. Whitman Sensing User Input At Display Area Edge
US20130290410A1 (en) * 2012-04-28 2013-10-31 Alibaba Group Holding Limited Performing autocomplete of content
WO2013157232A3 (en) * 2012-04-19 2014-01-09 Sony Corporation Information processing apparatus, information processing method, program, and information processing system
US8719724B2 (en) 2011-03-16 2014-05-06 Honeywell International Inc. Method for enlarging characters displayed on an adaptive touch screen key pad
WO2014107064A1 (en) * 2013-01-07 2014-07-10 Samsung Electronics Co., Ltd. Display apparatus, method for ui display thereof and computer-readable recording medium
US8850241B2 (en) 2012-03-02 2014-09-30 Microsoft Corporation Multi-stage power adapter configured to provide low power upon initial connection of the power adapter to the host device and high power thereafter upon notification from the host device to the power adapter
US8854323B2 (en) 2010-05-20 2014-10-07 Panasonic Intellectual Property Corporation Of America Operating apparatus, operating method, program, recording medium, and integrated circuit
US8872762B2 (en) 2010-12-08 2014-10-28 Primesense Ltd. Three dimensional user interface cursor control
US8873227B2 (en) 2012-03-02 2014-10-28 Microsoft Corporation Flexible hinge support layer
US8881051B2 (en) 2011-07-05 2014-11-04 Primesense Ltd Zoom-based gesture user interface
WO2014178813A1 (en) * 2013-04-29 2014-11-06 Hewlett-Packard Development Company, L.P. Device for displaying a received user interface
US8892446B2 (en) 2010-01-18 2014-11-18 Apple Inc. Service orchestration for intelligent automated assistant
US8933876B2 (en) 2010-12-13 2015-01-13 Apple Inc. Three dimensional user interface session control
US20150029402A1 (en) * 2013-07-26 2015-01-29 Tianjin Funayuanchuang Technology Co.,Ltd. Remote controller, system, and method for controlling remote controller
US8949477B2 (en) 2012-05-14 2015-02-03 Microsoft Technology Licensing, Llc Accessory device architecture
WO2015052724A1 (en) * 2013-10-07 2015-04-16 Deshmukh Rakesh An indian language keypad
US9030498B2 (en) 2011-08-15 2015-05-12 Apple Inc. Combining explicit select gestures and timeclick in a non-tactile three dimensional user interface
US9035876B2 (en) 2008-01-14 2015-05-19 Apple Inc. Three-dimensional user interface session control
US9064654B2 (en) 2012-03-02 2015-06-23 Microsoft Technology Licensing, Llc Method of manufacturing an input device
US9075566B2 (en) 2012-03-02 2015-07-07 Microsoft Technoogy Licensing, LLC Flexible hinge spine
US9122311B2 (en) 2011-08-24 2015-09-01 Apple Inc. Visual feedback for tactile and non-tactile user interfaces
US9158375B2 (en) 2010-07-20 2015-10-13 Apple Inc. Interactive reality augmentation for natural interaction
US9201501B2 (en) 2010-07-20 2015-12-01 Apple Inc. Adaptive projector
US9218063B2 (en) 2011-08-24 2015-12-22 Apple Inc. Sessionless pointing user interface
US9229534B2 (en) 2012-02-28 2016-01-05 Apple Inc. Asymmetric mapping for tactile and non-tactile user interfaces
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US9261964B2 (en) 2005-12-30 2016-02-16 Microsoft Technology Licensing, Llc Unintentional touch rejection
US9285874B2 (en) 2011-02-09 2016-03-15 Apple Inc. Gaze detection in a 3D mapping environment
US9300784B2 (en) 2013-06-13 2016-03-29 Apple Inc. System and method for emergency calls initiated by voice command
US9304549B2 (en) 2013-03-28 2016-04-05 Microsoft Technology Licensing, Llc Hinge mechanism for rotatable component attachment
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US9330381B2 (en) 2008-01-06 2016-05-03 Apple Inc. Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US9354748B2 (en) 2012-02-13 2016-05-31 Microsoft Technology Licensing, Llc Optical stylus interaction
US9360893B2 (en) 2012-03-02 2016-06-07 Microsoft Technology Licensing, Llc Input device writing surface
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
US9377865B2 (en) 2011-07-05 2016-06-28 Apple Inc. Zoom-based gesture user interface
US9377863B2 (en) 2012-03-26 2016-06-28 Apple Inc. Gaze-enhanced virtual touchscreen
US9426905B2 (en) 2012-03-02 2016-08-23 Microsoft Technology Licensing, Llc Connection device for computing devices
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US9447620B2 (en) 2014-09-30 2016-09-20 Microsoft Technology Licensing, Llc Hinge mechanism with multiple preset positions
US9459758B2 (en) 2011-07-05 2016-10-04 Apple Inc. Gesture-based interface with enhanced features
US9477337B2 (en) 2014-03-14 2016-10-25 Microsoft Technology Licensing, Llc Conductive trace routing for display and bezel sensors
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9501161B2 (en) 2010-10-22 2016-11-22 Hewlett-Packard Development Company, L.P. User interface for facilitating character input
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US20160342294A1 (en) * 2015-05-19 2016-11-24 Google Inc. Multi-switch option scanning
US9535906B2 (en) 2008-07-31 2017-01-03 Apple Inc. Mobile device having human language translation capability with positional feedback
US9544504B2 (en) 2012-11-02 2017-01-10 Microsoft Technology Licensing, Llc Rapid synchronized lighting and shuttering
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
US9582122B2 (en) 2012-11-12 2017-02-28 Microsoft Technology Licensing, Llc Touch-sensitive bezel techniques
US9606986B2 (en) 2014-09-29 2017-03-28 Apple Inc. Integrated word N-gram and class M-gram language models
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US9620104B2 (en) 2013-06-07 2017-04-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9626955B2 (en) 2008-04-05 2017-04-18 Apple Inc. Intelligent text-to-speech conversion
US9633674B2 (en) 2013-06-07 2017-04-25 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9633660B2 (en) 2010-02-25 2017-04-25 Apple Inc. User profiling for voice input processing
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9646614B2 (en) 2000-03-16 2017-05-09 Apple Inc. Fast, language-independent method for user authentication by voice
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US9697822B1 (en) 2013-03-15 2017-07-04 Apple Inc. System and method for updating an adaptive speech recognition model
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US9752361B2 (en) 2015-06-18 2017-09-05 Microsoft Technology Licensing, Llc Multistage hinge
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9779225B2 (en) 2015-04-08 2017-10-03 Google Inc. Method and system to provide access to secure features of a device
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9798393B2 (en) 2011-08-29 2017-10-24 Apple Inc. Text correction processing
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US9824808B2 (en) 2012-08-20 2017-11-21 Microsoft Technology Licensing, Llc Switchable magnetic lock
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US9864415B2 (en) 2015-06-30 2018-01-09 Microsoft Technology Licensing, Llc Multistage friction hinge
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US9870066B2 (en) 2012-03-02 2018-01-16 Microsoft Technology Licensing, Llc Method of manufacturing an input device
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9922642B2 (en) 2013-03-15 2018-03-20 Apple Inc. Training an at least partial voice command system
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9953088B2 (en) 2012-05-14 2018-04-24 Apple Inc. Crowd sourcing information to fulfill user requests
US9959870B2 (en) 2008-12-11 2018-05-01 Apple Inc. Speech recognition involving a mobile device
US9966068B2 (en) 2013-06-08 2018-05-08 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US9971774B2 (en) 2012-09-19 2018-05-15 Apple Inc. Voice-based media searching
US10001857B2 (en) 2015-05-13 2018-06-19 Seiko Epson Corporation Display apparatus and method of controlling display apparatus
US10031556B2 (en) 2012-06-08 2018-07-24 Microsoft Technology Licensing, Llc User experience adaptation
US10037057B2 (en) 2016-09-22 2018-07-31 Microsoft Technology Licensing, Llc Friction hinge
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
US10079014B2 (en) 2012-06-08 2018-09-18 Apple Inc. Name recognition system
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10089072B2 (en) 2016-06-11 2018-10-02 Apple Inc. Intelligent device arbitration and control
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US10107994B2 (en) 2012-06-12 2018-10-23 Microsoft Technology Licensing, Llc Wide field-of-view virtual image projector
US10120420B2 (en) 2014-03-21 2018-11-06 Microsoft Technology Licensing, Llc Lockable display and techniques enabling use of lockable displays
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US10156889B2 (en) 2014-09-15 2018-12-18 Microsoft Technology Licensing, Llc Inductive peripheral retention device
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
US10185542B2 (en) 2013-06-09 2019-01-22 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10199051B2 (en) 2013-02-07 2019-02-05 Apple Inc. Voice trigger for a digital assistant
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US10262324B2 (en) 2010-11-29 2019-04-16 Biocatch Ltd. System, device, and method of differentiating among users based on user-specific page navigation sequence
US10269345B2 (en) 2016-06-11 2019-04-23 Apple Inc. Intelligent task discovery
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US10283110B2 (en) 2009-07-02 2019-05-07 Apple Inc. Methods and apparatuses for automatic speech recognition
US10289433B2 (en) 2014-05-30 2019-05-14 Apple Inc. Domain specific language for encoding assistant dialog
US10298614B2 (en) * 2010-11-29 2019-05-21 Biocatch Ltd. System, device, and method of generating and managing behavioral biometric cookies
US10297253B2 (en) 2016-06-11 2019-05-21 Apple Inc. Application integration with a digital assistant
US10303715B2 (en) 2017-05-16 2019-05-28 Apple Inc. Intelligent automated assistant for media exploration
US10311144B2 (en) 2017-05-16 2019-06-04 Apple Inc. Emoji word sense disambiguation
US10318871B2 (en) 2005-09-08 2019-06-11 Apple Inc. Method and apparatus for building an intelligent automated assistant
US10324733B2 (en) 2014-07-30 2019-06-18 Microsoft Technology Licensing, Llc Shutdown notifications
US10332518B2 (en) 2017-05-09 2019-06-25 Apple Inc. User interface for correcting recognition errors
US10344797B2 (en) 2016-04-05 2019-07-09 Microsoft Technology Licensing, Llc Hinge with multiple preset positions
US10354011B2 (en) 2016-06-09 2019-07-16 Apple Inc. Intelligent automated assistant in a home environment
US10356243B2 (en) 2015-06-05 2019-07-16 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US10395654B2 (en) 2017-05-11 2019-08-27 Apple Inc. Text normalization based on a data-driven learning network
US10397262B2 (en) 2017-07-20 2019-08-27 Biocatch Ltd. Device, system, and method of detecting overlay malware
US10404729B2 (en) 2010-11-29 2019-09-03 Biocatch Ltd. Device, method, and system of generating fraud-alerts for cyber-attacks
US10403278B2 (en) 2017-05-16 2019-09-03 Apple Inc. Methods and systems for phonetic matching in digital assistant services
US10403283B1 (en) 2018-06-01 2019-09-03 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
US10410637B2 (en) 2017-05-12 2019-09-10 Apple Inc. User-specific acoustic models
US10417266B2 (en) 2017-05-09 2019-09-17 Apple Inc. Context-aware ranking of intelligent response suggestions
US10417339B2 (en) 2017-08-18 2019-09-17 Kyocera Document Solutions Inc. Suggestion of alternate user input using different user interface
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US10446141B2 (en) 2014-08-28 2019-10-15 Apple Inc. Automatic speech recognition based on user feedback
US10445429B2 (en) 2017-09-21 2019-10-15 Apple Inc. Natural language understanding using vocabularies with compressed serialized tries
US10474815B2 (en) 2010-11-29 2019-11-12 Biocatch Ltd. System, device, and method of detecting malicious automatic script and code injection
US10474753B2 (en) 2016-09-07 2019-11-12 Apple Inc. Language identification using recurrent neural networks
US10482874B2 (en) 2017-05-15 2019-11-19 Apple Inc. Hierarchical belief states for digital assistants
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10496705B1 (en) 2018-06-03 2019-12-03 Apple Inc. Accelerated task performance
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10521466B2 (en) 2016-06-11 2019-12-31 Apple Inc. Data driven natural language event detection and classification
US10523680B2 (en) * 2015-07-09 2019-12-31 Biocatch Ltd. System, device, and method for detecting a proxy server
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US10568032B2 (en) 2007-04-03 2020-02-18 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US10579784B2 (en) 2016-11-02 2020-03-03 Biocatch Ltd. System, device, and method of secure utilization of fingerprints for user authentication
US10586036B2 (en) 2010-11-29 2020-03-10 Biocatch Ltd. System, device, and method of recovery and resetting of user authentication factor
US10592095B2 (en) 2014-05-23 2020-03-17 Apple Inc. Instantaneous speaking of content on touch devices
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US10592604B2 (en) 2018-03-12 2020-03-17 Apple Inc. Inverse text normalization for automatic speech recognition
US10621585B2 (en) 2010-11-29 2020-04-14 Biocatch Ltd. Contextual mapping of web-pages, and generation of fraud-relatedness score-values
US10636424B2 (en) 2017-11-30 2020-04-28 Apple Inc. Multi-turn canned dialog
US10643611B2 (en) 2008-10-02 2020-05-05 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US10657328B2 (en) 2017-06-02 2020-05-19 Apple Inc. Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
US10684703B2 (en) 2018-06-01 2020-06-16 Apple Inc. Attention aware virtual assistant dismissal
US10685355B2 (en) * 2016-12-04 2020-06-16 Biocatch Ltd. Method, device, and system of detecting mule accounts and accounts used for money laundering
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10706373B2 (en) 2011-06-03 2020-07-07 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10719765B2 (en) 2015-06-25 2020-07-21 Biocatch Ltd. Conditional behavioral biometrics
US10726832B2 (en) 2017-05-11 2020-07-28 Apple Inc. Maintaining privacy of personal information
US10728761B2 (en) 2010-11-29 2020-07-28 Biocatch Ltd. Method, device, and system of detecting a lie of a user who inputs data
US10733993B2 (en) 2016-06-10 2020-08-04 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10733375B2 (en) 2018-01-31 2020-08-04 Apple Inc. Knowledge-based framework for improving natural language understanding
US10733982B2 (en) 2018-01-08 2020-08-04 Apple Inc. Multi-directional dialog
US10747305B2 (en) 2010-11-29 2020-08-18 Biocatch Ltd. Method, system, and device of authenticating identity of a user of an electronic device
US10748546B2 (en) 2017-05-16 2020-08-18 Apple Inc. Digital assistant services based on device capabilities
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US10755703B2 (en) 2017-05-11 2020-08-25 Apple Inc. Offline personal assistant
US10755051B2 (en) 2017-09-29 2020-08-25 Apple Inc. Rule-based natural language processing
US10762293B2 (en) 2010-12-22 2020-09-01 Apple Inc. Using parts-of-speech tagging and named entity recognition for spelling correction
US10776476B2 (en) 2010-11-29 2020-09-15 Biocatch Ltd. System, device, and method of visual login
US10789945B2 (en) 2017-05-12 2020-09-29 Apple Inc. Low-latency intelligent automated assistant
US10789959B2 (en) 2018-03-02 2020-09-29 Apple Inc. Training speaker recognition models for digital assistants
US10791216B2 (en) 2013-08-06 2020-09-29 Apple Inc. Auto-activating smart responses based on activities from remote devices
US10791176B2 (en) 2017-05-12 2020-09-29 Apple Inc. Synchronization and task delegation of a digital assistant
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US10810274B2 (en) 2017-05-15 2020-10-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
US10818288B2 (en) 2018-03-26 2020-10-27 Apple Inc. Natural assistant interaction
US10834590B2 (en) 2010-11-29 2020-11-10 Biocatch Ltd. Method, device, and system of differentiating between a cyber-attacker and a legitimate user
US10839159B2 (en) 2018-09-28 2020-11-17 Apple Inc. Named entity normalization in a spoken dialog system
US10892996B2 (en) 2018-06-01 2021-01-12 Apple Inc. Variable latency device coordination
US10897482B2 (en) 2010-11-29 2021-01-19 Biocatch Ltd. Method, device, and system of back-coloring, forward-coloring, and fraud detection
US10909331B2 (en) 2018-03-30 2021-02-02 Apple Inc. Implicit identification of translation payload with neural machine translation
US10917431B2 (en) 2010-11-29 2021-02-09 Biocatch Ltd. System, method, and device of authenticating a user based on selfie image or selfie video
US10928918B2 (en) 2018-05-07 2021-02-23 Apple Inc. Raise to speak
US10949757B2 (en) 2010-11-29 2021-03-16 Biocatch Ltd. System, device, and method of detecting user identity based on motor-control loop model
US10949514B2 (en) 2010-11-29 2021-03-16 Biocatch Ltd. Device, system, and method of differentiating among users based on detection of hardware components
US10970394B2 (en) 2017-11-21 2021-04-06 Biocatch Ltd. System, device, and method of detecting vishing attacks
US10984780B2 (en) 2018-05-21 2021-04-20 Apple Inc. Global semantic word embeddings using bi-directional recurrent neural networks
US11010127B2 (en) 2015-06-29 2021-05-18 Apple Inc. Virtual assistant for media playback
US11010561B2 (en) 2018-09-27 2021-05-18 Apple Inc. Sentiment prediction from textual data
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US11023513B2 (en) 2007-12-20 2021-06-01 Apple Inc. Method and apparatus for searching using an active ontology
US11029838B2 (en) 2006-09-06 2021-06-08 Apple Inc. Touch screen device, method, and graphical user interface for customizing display of content category icons
US11055395B2 (en) 2016-07-08 2021-07-06 Biocatch Ltd. Step-up authentication
US11140099B2 (en) 2019-05-21 2021-10-05 Apple Inc. Providing message response suggestions
US11145294B2 (en) 2018-05-07 2021-10-12 Apple Inc. Intelligent automated assistant for delivering content from user experiences
US20210329030A1 (en) * 2010-11-29 2021-10-21 Biocatch Ltd. Device, System, and Method of Detecting Vishing Attacks
US11170166B2 (en) 2018-09-28 2021-11-09 Apple Inc. Neural typographical error modeling via generative adversarial networks
US11181988B1 (en) 2020-08-31 2021-11-23 Apple Inc. Incorporating user feedback into text prediction models via joint reward planning
US11204787B2 (en) 2017-01-09 2021-12-21 Apple Inc. Application integration with a digital assistant
US11210674B2 (en) 2010-11-29 2021-12-28 Biocatch Ltd. Method, device, and system of detecting mule accounts and accounts used for money laundering
US11217251B2 (en) 2019-05-06 2022-01-04 Apple Inc. Spoken notifications
US11223619B2 (en) 2010-11-29 2022-01-11 Biocatch Ltd. Device, system, and method of user authentication based on user-specific characteristics of task performance
US11227589B2 (en) 2016-06-06 2022-01-18 Apple Inc. Intelligent list reading
US11231904B2 (en) 2015-03-06 2022-01-25 Apple Inc. Reducing response latency of intelligent automated assistants
US11237797B2 (en) 2019-05-31 2022-02-01 Apple Inc. User activity shortcut suggestions
US11269977B2 (en) 2010-11-29 2022-03-08 Biocatch Ltd. System, apparatus, and method of collecting and processing data in electronic devices
US11269678B2 (en) 2012-05-15 2022-03-08 Apple Inc. Systems and methods for integrating third party services with a digital assistant
USRE48963E1 (en) 2012-03-02 2022-03-08 Microsoft Technology Licensing, Llc Connection device for computing devices
US11281993B2 (en) 2016-12-05 2022-03-22 Apple Inc. Model and ensemble compression for metric learning
US11289073B2 (en) 2019-05-31 2022-03-29 Apple Inc. Device text to speech
US11301477B2 (en) 2017-05-12 2022-04-12 Apple Inc. Feedback analysis of a digital assistant
US11307752B2 (en) 2019-05-06 2022-04-19 Apple Inc. User configurable task triggers
US11314370B2 (en) 2013-12-06 2022-04-26 Apple Inc. Method for extracting salient dialog usage from live data
US11348573B2 (en) 2019-03-18 2022-05-31 Apple Inc. Multimodality in digital assistant systems
US11360641B2 (en) 2019-06-01 2022-06-14 Apple Inc. Increasing the relevance of new available information
US11386266B2 (en) 2018-06-01 2022-07-12 Apple Inc. Text correction
US11388291B2 (en) 2013-03-14 2022-07-12 Apple Inc. System and method for processing voicemail
US11423908B2 (en) 2019-05-06 2022-08-23 Apple Inc. Interpreting spoken requests
US11462215B2 (en) 2018-09-28 2022-10-04 Apple Inc. Multi-modal inputs for voice commands
US11468282B2 (en) 2015-05-15 2022-10-11 Apple Inc. Virtual assistant in a communication session
US11475884B2 (en) 2019-05-06 2022-10-18 Apple Inc. Reducing digital assistant latency when a language is incorrectly determined
US11475898B2 (en) 2018-10-26 2022-10-18 Apple Inc. Low-latency multi-speaker speech recognition
US11488406B2 (en) 2019-09-25 2022-11-01 Apple Inc. Text detection using global geometry estimators
US11495218B2 (en) 2018-06-01 2022-11-08 Apple Inc. Virtual assistant operation in multi-device environments
US11496600B2 (en) 2019-05-31 2022-11-08 Apple Inc. Remote execution of machine-learned models
US11532306B2 (en) 2017-05-16 2022-12-20 Apple Inc. Detecting a trigger of a digital assistant
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
US11606353B2 (en) 2021-07-22 2023-03-14 Biocatch Ltd. System, device, and method of generating and utilizing one-time passwords
US11638059B2 (en) 2019-01-04 2023-04-25 Apple Inc. Content playback on multiple devices
US11657813B2 (en) 2019-05-31 2023-05-23 Apple Inc. Voice identification in digital assistant systems
US11798547B2 (en) 2013-03-15 2023-10-24 Apple Inc. Voice activated device for use with a voice-based digital assistant

Families Citing this family (156)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7286115B2 (en) 2000-05-26 2007-10-23 Tegic Communications, Inc. Directional input system with automatic correction
US7821503B2 (en) 2003-04-09 2010-10-26 Tegic Communications, Inc. Touch screen and graphical user interface
US7030863B2 (en) 2000-05-26 2006-04-18 America Online, Incorporated Virtual keyboard system with automatic correction
US7750891B2 (en) 2003-04-09 2010-07-06 Tegic Communications, Inc. Selective input system based on tracking of motion parameters of an input device
US7779364B2 (en) * 2007-01-26 2010-08-17 Kabushiki Kaisha Toshiba System and method for generating a graphical user input interface via an associated display
US8225203B2 (en) * 2007-02-01 2012-07-17 Nuance Communications, Inc. Spell-check for a keyboard system with automatic correction
US8201087B2 (en) 2007-02-01 2012-06-12 Tegic Communications, Inc. Spell-check for a keyboard system with automatic correction
US8065624B2 (en) * 2007-06-28 2011-11-22 Panasonic Corporation Virtual keypad systems and methods
US20100299710A1 (en) * 2007-09-20 2010-11-25 Samsung Electronics Co. Ltd. Method for inputting user command and video apparatus and input apparatus employing the same
KR101470413B1 (en) * 2007-09-20 2014-12-10 삼성전자주식회사 The method of inputting user command and the image apparatus and input apparatus thereof
US8631358B2 (en) * 2007-10-10 2014-01-14 Apple Inc. Variable device graphical user interface
US20090102818A1 (en) * 2007-10-22 2009-04-23 Motorola, Inc. Method and device for error-free keypad input
US20090185080A1 (en) * 2008-01-18 2009-07-23 Imu Solutions, Inc. Controlling an electronic device by changing an angular orientation of a remote wireless-controller
US20090193361A1 (en) * 2008-01-30 2009-07-30 Research In Motion Limited Electronic device and method of controlling same
US9520743B2 (en) * 2008-03-27 2016-12-13 Echostar Technologies L.L.C. Reduction of power consumption in remote control electronics
US8289688B2 (en) 2008-04-01 2012-10-16 Litl, Llc Portable computer with multiple display configurations
US8612888B2 (en) 2008-04-01 2013-12-17 Litl, Llc Method and apparatus for managing digital media content
US9003315B2 (en) 2008-04-01 2015-04-07 Litl Llc System and method for streamlining user interaction with electronic content
US7907060B2 (en) * 2008-05-08 2011-03-15 Echostar Technologies L.L.C. Systems, methods and apparatus for detecting replacement of a battery in a remote control
CN101598966A (en) * 2008-06-06 2009-12-09 深圳富泰宏精密工业有限公司 The contactor control device and the control method thereof that show two pictures
US8305249B2 (en) 2008-07-18 2012-11-06 EchoStar Technologies, L.L.C. Systems and methods for controlling power consumption in electronic devices
FR2934067B1 (en) * 2008-07-21 2013-01-25 Dav HAPTICALLY RETURN CONTROL DEVICE AND CORRESPONDING CONTROL METHOD
US20100020036A1 (en) * 2008-07-23 2010-01-28 Edward Hui Portable electronic device and method of controlling same
TWI360762B (en) * 2008-09-05 2012-03-21 Mitake Information Corp On-screen virtual keyboard system
US8750938B2 (en) 2008-09-29 2014-06-10 Microsoft Corporation Glow touch feedback for virtual input devices
US8786551B2 (en) * 2008-12-11 2014-07-22 Verizon Patent And Licensing Inc. Predictive keypad/keyboard lighting to guide input
US20100164756A1 (en) * 2008-12-30 2010-07-01 Nokia Corporation Electronic device user input
JP5541166B2 (en) * 2009-01-20 2014-07-09 日本電気株式会社 INPUT DEVICE, INFORMATION PROCESSING DEVICE, INPUT METHOD, AND PROGRAM
FR2941546B1 (en) * 2009-01-27 2011-02-04 Wyplay OPTIMIZED METHOD FOR ENTERING ALPHANUMERIC CHARACTERS FROM A REMOTE CONTROL
US8326358B2 (en) 2009-01-30 2012-12-04 Research In Motion Limited System and method for access control in a portable electronic device
CA2749916A1 (en) * 2009-02-04 2010-08-12 Benjamin Firooz Ghassabian Data entry system
US8134475B2 (en) 2009-03-16 2012-03-13 Echostar Technologies L.L.C. Backlighting remote controls
US20100245268A1 (en) * 2009-03-30 2010-09-30 Stg Interactive S.A. User-friendly process for interacting with informational content on touchscreen devices
GB201108200D0 (en) * 2011-05-16 2011-06-29 Touchtype Ltd User input prediction
GB201016385D0 (en) * 2010-09-29 2010-11-10 Touchtype Ltd System and method for inputting text into electronic devices
KR20100118366A (en) * 2009-04-28 2010-11-05 삼성전자주식회사 Operating method of touch screen and portable device including the same
US9043718B2 (en) * 2009-06-05 2015-05-26 Blackberry Limited System and method for applying a text prediction algorithm to a virtual keyboard
US20100310136A1 (en) * 2009-06-09 2010-12-09 Sony Ericsson Mobile Communications Ab Distinguishing right-hand input and left-hand input based on finger recognition
EP2264563A1 (en) * 2009-06-19 2010-12-22 Tegic Communications, Inc. Virtual keyboard system with automatic correction
US20110041056A1 (en) * 2009-08-14 2011-02-17 Research In Motion Limited Electronic device with touch-sensitive display and method of facilitating input at the electronic device
EP2284653A1 (en) * 2009-08-14 2011-02-16 Research In Motion Limited Electronic device with touch-sensitive display and method of facilitating input at the electronic device
US20110077470A1 (en) * 2009-09-30 2011-03-31 Nellcor Puritan Bennett Llc Patient Monitor Symmetry Control
US8261211B2 (en) * 2009-10-01 2012-09-04 Microsoft Corporation Monitoring pointer trajectory and modifying display interface
US8347221B2 (en) * 2009-10-07 2013-01-01 Research In Motion Limited Touch-sensitive display and method of control
KR101058525B1 (en) 2009-10-09 2011-08-23 삼성전자주식회사 Text input method and display device using the same
KR101595029B1 (en) 2009-11-18 2016-02-17 엘지전자 주식회사 Mobile terminal and method for controlling the same
US20110154260A1 (en) * 2009-12-17 2011-06-23 Motorola Inc Method and apparatus for displaying information in an electronic device
JP5676113B2 (en) * 2010-01-27 2015-02-25 京セラ株式会社 Portable electronic devices
JP5635274B2 (en) 2010-01-27 2014-12-03 京セラ株式会社 Tactile sensation presentation apparatus and tactile sensation presentation method
US8456435B2 (en) 2010-02-26 2013-06-04 Research In Motion Limited Electronic device with touch-sensitive display and method of facilitating input at the electronic device
EP2369445B1 (en) * 2010-02-26 2016-08-17 BlackBerry Limited Electronic device with touch-sensitive display and method of facilitating input at the electronic device
EP2407859A1 (en) * 2010-07-16 2012-01-18 Gigaset Communications GmbH Dynamic adjustment of a user interface on a sensor monitor
KR101701932B1 (en) * 2010-07-22 2017-02-13 삼성전자 주식회사 Input device and control method of thereof
TWI427507B (en) * 2010-07-29 2014-02-21 Hon Hai Prec Ind Co Ltd Handheld device and method for inputting characters thereof
US9639265B2 (en) * 2010-09-03 2017-05-02 Microsoft Technology Licensing, Llc Distance-time based hit-testing for displayed target graphical elements
GB201200643D0 (en) 2012-01-16 2012-02-29 Touchtype Ltd System and method for inputting text
JP5782699B2 (en) 2010-10-15 2015-09-24 ソニー株式会社 Information processing apparatus, input control method for information processing apparatus, and program
TW201222396A (en) * 2010-11-22 2012-06-01 Acer Inc Touch-controlled apparatus and method for displaying a virtual keyboard on the touch-controlled apparatus thereof
US20120131490A1 (en) * 2010-11-22 2012-05-24 Shao-Chieh Lin Touch-controlled device and method for displaying a virtual keyboard on the touch-controlled device thereof
KR101821160B1 (en) * 2010-12-10 2018-03-08 삼성전자 주식회사 Method and apparatus for providing user keypad in a portable terminal
US9354804B2 (en) 2010-12-29 2016-05-31 Microsoft Technology Licensing, Llc Touch event anticipation in a computing device
US9201185B2 (en) 2011-02-04 2015-12-01 Microsoft Technology Licensing, Llc Directional backlighting for display panels
US11087659B2 (en) * 2011-03-11 2021-08-10 Cox Communications, Inc. System, method and device for presenting different functional displays when orientation of the device changes
WO2013024317A1 (en) 2011-08-15 2013-02-21 Telefonaktiebolaget Lm Ericsson (Publ) Resizing selection zones on a touch sensitive display responsive to likelihood of selection
US9262076B2 (en) * 2011-09-12 2016-02-16 Microsoft Technology Licensing, Llc Soft keyboard interface
KR20130034765A (en) * 2011-09-29 2013-04-08 삼성전자주식회사 Method and device for inputting of mobile terminal using a pen
US8490008B2 (en) 2011-11-10 2013-07-16 Research In Motion Limited Touchscreen keyboard predictive display and generation of a set of characters
US9715489B2 (en) 2011-11-10 2017-07-25 Blackberry Limited Displaying a prediction candidate after a typing mistake
US9122672B2 (en) 2011-11-10 2015-09-01 Blackberry Limited In-letter word prediction for virtual keyboard
US9652448B2 (en) 2011-11-10 2017-05-16 Blackberry Limited Methods and systems for removing or replacing on-keyboard prediction candidates
US9310889B2 (en) 2011-11-10 2016-04-12 Blackberry Limited Touchscreen keyboard predictive display and generation of a set of characters
KR20130054579A (en) * 2011-11-17 2013-05-27 삼성전자주식회사 Display apparatus and control method thereof
JP5413448B2 (en) * 2011-12-23 2014-02-12 株式会社デンソー Display system, display device, and operation device
WO2013099362A1 (en) * 2011-12-28 2013-07-04 Ikeda Hiroyuki Portable terminal
US9152323B2 (en) 2012-01-19 2015-10-06 Blackberry Limited Virtual keyboard providing an indication of received input
US9557913B2 (en) 2012-01-19 2017-01-31 Blackberry Limited Virtual keyboard display having a ticker proximate to the virtual keyboard
US9052414B2 (en) 2012-02-07 2015-06-09 Microsoft Technology Licensing, Llc Virtual image device
US9244612B1 (en) * 2012-02-16 2016-01-26 Google Inc. Key selection of a graphical keyboard based on user input posture
US20130222255A1 (en) 2012-02-24 2013-08-29 Research In Motion Limited Portable electronic device including touch-sensitive display and method of controlling same
WO2013123572A1 (en) * 2012-02-24 2013-08-29 Research In Motion Limited Touchscreen keyboard providing word predictions in partitions of the touchscreen keyboard in proximate association with candidate letters
US8749529B2 (en) 2012-03-01 2014-06-10 Microsoft Corporation Sensor-in-pixel display system with near infrared filter
EP2823385B1 (en) * 2012-03-04 2020-03-18 Microsoft Technology Licensing, LLC Portable device and control method thereof
US9201510B2 (en) 2012-04-16 2015-12-01 Blackberry Limited Method and device having touchscreen keyboard with visual cues
US9292192B2 (en) 2012-04-30 2016-03-22 Blackberry Limited Method and apparatus for text selection
US9354805B2 (en) 2012-04-30 2016-05-31 Blackberry Limited Method and apparatus for text selection
US10025487B2 (en) 2012-04-30 2018-07-17 Blackberry Limited Method and apparatus for text selection
US20130285927A1 (en) * 2012-04-30 2013-10-31 Research In Motion Limited Touchscreen keyboard with correction of previously input text
KR101901611B1 (en) * 2012-05-09 2018-09-27 엘지전자 주식회사 Mobile terminal and control method for the mobile terminal
US8884881B2 (en) 2012-05-11 2014-11-11 Blackberry Limited Portable electronic device and method of controlling same
EP2662752B1 (en) * 2012-05-11 2017-09-13 BlackBerry Limited Apparatus and method for character entry in a portable electronic device
US9207860B2 (en) 2012-05-25 2015-12-08 Blackberry Limited Method and apparatus for detecting a gesture
US8947353B2 (en) 2012-06-12 2015-02-03 Microsoft Corporation Photosensor array gesture detection
US9073123B2 (en) 2012-06-13 2015-07-07 Microsoft Technology Licensing, Llc Housing vents
US9684382B2 (en) 2012-06-13 2017-06-20 Microsoft Technology Licensing, Llc Input device configuration having capacitive and pressure sensors
US9459160B2 (en) 2012-06-13 2016-10-04 Microsoft Technology Licensing, Llc Input device sensor configuration
US9063693B2 (en) 2012-06-13 2015-06-23 Microsoft Technology Licensing, Llc Peripheral device storage
JP6071107B2 (en) 2012-06-14 2017-02-01 裕行 池田 Mobile device
US9256089B2 (en) 2012-06-15 2016-02-09 Microsoft Technology Licensing, Llc Object-detecting backlight unit
US20130346904A1 (en) * 2012-06-26 2013-12-26 International Business Machines Corporation Targeted key press zones on an interactive display
US9116552B2 (en) 2012-06-27 2015-08-25 Blackberry Limited Touchscreen keyboard providing selection of word predictions in partitions of the touchscreen keyboard
US9355345B2 (en) 2012-07-23 2016-05-31 Microsoft Technology Licensing, Llc Transparent tags with encoded data
US9298295B2 (en) * 2012-07-25 2016-03-29 Facebook, Inc. Gestures for auto-correct
US9836213B2 (en) * 2012-07-27 2017-12-05 Symbol Technologies, Llc Enhanced user interface for pressure sensitive touch screen
US9524290B2 (en) 2012-08-31 2016-12-20 Blackberry Limited Scoring predictions based on prediction length and typing speed
US20140063067A1 (en) * 2012-08-31 2014-03-06 Research In Motion Limited Method to select word by swiping capacitive keyboard
US9063653B2 (en) 2012-08-31 2015-06-23 Blackberry Limited Ranking predictions based on typing speed and typing confidence
WO2014047161A2 (en) * 2012-09-18 2014-03-27 Google Inc. Posture-adaptive selection
US9152173B2 (en) 2012-10-09 2015-10-06 Microsoft Technology Licensing, Llc Transparent display device
US8654030B1 (en) 2012-10-16 2014-02-18 Microsoft Corporation Antenna placement
WO2014059618A1 (en) 2012-10-17 2014-04-24 Microsoft Corporation Graphic formation via material ablation
WO2014059624A1 (en) 2012-10-17 2014-04-24 Microsoft Corporation Metal alloy injection molding protrusions
CN104903026B (en) 2012-10-17 2017-10-24 微软技术许可有限责任公司 Metal alloy injection is molded overfall
US9035884B2 (en) * 2012-10-17 2015-05-19 Nuance Communications, Inc. Subscription updates in multiple device language models
US9304595B2 (en) * 2012-10-19 2016-04-05 Google Inc. Gesture-keyboard decoding using gesture path deviation
US8704792B1 (en) 2012-10-19 2014-04-22 Google Inc. Density-based filtering of gesture events associated with a user interface of a computing device
US8952892B2 (en) 2012-11-01 2015-02-10 Microsoft Corporation Input location correction tables for input panels
CN103809898A (en) * 2012-11-14 2014-05-21 宇瞻科技股份有限公司 Intelligent input method
US9411510B2 (en) * 2012-12-07 2016-08-09 Apple Inc. Techniques for preventing typographical errors on soft keyboards
US9513748B2 (en) 2012-12-13 2016-12-06 Microsoft Technology Licensing, Llc Combined display panel circuit
US20140208258A1 (en) * 2013-01-22 2014-07-24 Jenny Yuen Predictive Input Using Custom Dictionaries
US9176538B2 (en) 2013-02-05 2015-11-03 Microsoft Technology Licensing, Llc Input device configurations
US10578499B2 (en) 2013-02-17 2020-03-03 Microsoft Technology Licensing, Llc Piezo-actuated virtual buttons for touch surfaces
US9638835B2 (en) 2013-03-05 2017-05-02 Microsoft Technology Licensing, Llc Asymmetric aberration correcting lens
US10120540B2 (en) * 2013-03-14 2018-11-06 Samsung Electronics Co., Ltd. Visual feedback for user interface navigation on television system
US9348429B2 (en) 2013-03-15 2016-05-24 Blackberry Limited Method and apparatus for word prediction using the position of a non-typing digit
EP2778859B1 (en) 2013-03-15 2015-07-08 BlackBerry Limited Method and apparatus for word prediction using the position of a non-typing digit
US9552777B2 (en) 2013-05-10 2017-01-24 Microsoft Technology Licensing, Llc Phase control backlight
US9454251B1 (en) 2013-06-26 2016-09-27 Google Inc. Methods, systems, and media for controlling a remote device using a touch screen of a mobile device in a display inhibited state
FR3014572B1 (en) * 2013-12-05 2016-01-01 Op3Ft METHOD FOR CONTROLLING INTERACTION WITH A TOUCH SCREEN AND EQUIPMENT USING THE SAME
US9448631B2 (en) 2013-12-31 2016-09-20 Microsoft Technology Licensing, Llc Input device haptics and pressure sensing
CN114895838A (en) * 2014-01-06 2022-08-12 华为终端有限公司 Application program display method and terminal
US9317072B2 (en) 2014-01-28 2016-04-19 Microsoft Technology Licensing, Llc Hinge mechanism with preset positions
US9759854B2 (en) 2014-02-17 2017-09-12 Microsoft Technology Licensing, Llc Input device outer layer and backlighting
US10440001B2 (en) * 2014-06-18 2019-10-08 Dell Products, Lp Method to securely authenticate management server over un-encrypted remote console connection
US9477653B2 (en) 2014-06-26 2016-10-25 Blackberry Limited Character entry for an electronic device using a position sensing keyboard
US9930167B2 (en) * 2014-07-07 2018-03-27 Verizon Patent And Licensing Inc. Messaging application with in-application search functionality
US20160034131A1 (en) * 2014-07-31 2016-02-04 Sony Corporation Methods and systems of a graphical user interface shift
US9513671B2 (en) 2014-08-01 2016-12-06 Microsoft Technology Licensing, Llc Peripheral retention device
US10534532B2 (en) 2014-08-08 2020-01-14 Samsung Electronics Co., Ltd. Electronic device and method for processing letter input in electronic device
US10191986B2 (en) 2014-08-11 2019-01-29 Microsoft Technology Licensing, Llc Web resource compatibility with web applications
US9705637B2 (en) 2014-08-19 2017-07-11 Microsoft Technology Licensing, Llc Guard band utilization for wireless data communication
US9397723B2 (en) 2014-08-26 2016-07-19 Microsoft Technology Licensing, Llc Spread spectrum wireless over non-contiguous channels
US9727231B2 (en) 2014-11-19 2017-08-08 Honda Motor Co., Ltd. System and method for providing absolute coordinate and zone mapping between a touchpad and a display screen
US20170371515A1 (en) 2014-11-19 2017-12-28 Honda Motor Co., Ltd. System and method for providing absolute and zone coordinate mapping with graphic animations
CN107408013B (en) * 2015-03-13 2020-07-10 京瓷办公信息系统株式会社 Information processing apparatus, screen display method, and computer-readable medium
CN104850239B (en) * 2015-04-23 2018-06-19 百度在线网络技术(北京)有限公司 A kind of method and apparatus for being used to determine keyboard display state
US10222889B2 (en) 2015-06-03 2019-03-05 Microsoft Technology Licensing, Llc Force inputs and cursor control
US10416799B2 (en) 2015-06-03 2019-09-17 Microsoft Technology Licensing, Llc Force sensing and inadvertent input control of an input device
US10061385B2 (en) 2016-01-22 2018-08-28 Microsoft Technology Licensing, Llc Haptic feedback for a touch input device
CN114579774A (en) * 2017-03-03 2022-06-03 谷歌有限责任公司 System and method for detecting that an application improperly implements presentation of a content item
KR20190065746A (en) * 2017-12-04 2019-06-12 삼성전자주식회사 Electronic apparatus, method for controlling thereof and the computer readable recording medium
US11554322B2 (en) * 2019-04-26 2023-01-17 Sony Interactive Entertainment LLC Game controller with touchpad input
CN111866565B (en) 2019-04-26 2022-09-02 三星电子株式会社 Display device and control method of display device
USD951286S1 (en) * 2019-11-19 2022-05-10 Johnson Systems Inc. Display screen with graphical user interface
EP3832434A1 (en) * 2019-12-03 2021-06-09 XRSpace CO., LTD. Behavior-based configuration method and behavior-based configuration system
CN113448465A (en) * 2021-06-29 2021-09-28 腾讯科技(深圳)有限公司 Virtual cursor control method and device, storage medium and electronic equipment

Citations (77)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5038401A (en) * 1989-04-05 1991-08-06 Pioneer Electronic Corporation Transmitter for remote control with operation switches having changeably displayed forms
US5128672A (en) * 1990-10-30 1992-07-07 Apple Computer, Inc. Dynamic predictive keyboard
US5652630A (en) * 1995-05-31 1997-07-29 International Business Machines Corporation Video receiver display, three axis remote control, and microcontroller for executing programs
US5724106A (en) * 1995-07-17 1998-03-03 Gateway 2000, Inc. Hand held remote control device with trigger button
US5936614A (en) * 1991-04-30 1999-08-10 International Business Machines Corporation User defined keyboard entry system
US5963671A (en) * 1991-11-27 1999-10-05 International Business Machines Corporation Enhancement of soft keyboard operations using trigram prediction
US6222528B1 (en) * 1997-03-07 2001-04-24 Cirque Corporation Method and apparatus for data input
US6359270B1 (en) * 1998-09-04 2002-03-19 Ncr Corporation Communications module mounting for domestic appliance
US6396523B1 (en) * 1999-07-29 2002-05-28 Interlink Electronics, Inc. Home entertainment device remote control
US20020093535A1 (en) * 2001-01-17 2002-07-18 Murphy Michael William User interface for character entry using a minimum number of selection keys
US20020114654A1 (en) * 2001-02-16 2002-08-22 Toshiyasu Abe Improved Keyboard
US20020122197A1 (en) * 2001-02-21 2002-09-05 Eli Abir Method and apparatus for text input
US20020149569A1 (en) * 2001-04-12 2002-10-17 International Business Machines Corporation Touchscreen user interface
US20030080945A1 (en) * 2001-10-29 2003-05-01 Betts-Lacroix Jonathan Keyboard with variable-sized keys
US20030193478A1 (en) * 2002-04-04 2003-10-16 Edwin Ng Reduced keyboard system that emulates QWERTY-type mapping and typing
US20030197736A1 (en) * 2002-01-16 2003-10-23 Murphy Michael W. User interface for character entry using a minimum number of selection keys
US20040041723A1 (en) * 2002-05-27 2004-03-04 Canon Kabushiki Kaisha, Tokyo, Japan Remote control device
US20040064351A1 (en) * 1999-11-22 2004-04-01 Mikurak Michael G. Increased visibility during order management in a network-based supply chain environment
US20040080487A1 (en) * 2002-10-29 2004-04-29 Griffin Jason T. Electronic device having keyboard for thumb typing
US6765557B1 (en) * 2000-04-10 2004-07-20 Interlink Electronics, Inc. Remote control having touch pad to screen mapping
US20040140956A1 (en) * 2003-01-16 2004-07-22 Kushler Clifford A. System and method for continuous stroke word-based text input
US20040169635A1 (en) * 2001-07-12 2004-09-02 Ghassabian Benjamin Firooz Features to enhance data entry through a small data entry unit
US20040239624A1 (en) * 2003-04-02 2004-12-02 Artoun Ramian Freehand symbolic input apparatus and method
US20050052406A1 (en) * 2003-04-09 2005-03-10 James Stephanick Selective input system based on tracking of motion parameters of an input device
US20050099400A1 (en) * 2003-11-06 2005-05-12 Samsung Electronics Co., Ltd. Apparatus and method for providing vitrtual graffiti and recording medium for the same
US20050141770A1 (en) * 2003-12-30 2005-06-30 Nokia Corporation Split on-screen keyboard
US20050162402A1 (en) * 2004-01-27 2005-07-28 Watanachote Susornpol J. Methods of interacting with a computer using a finger(s) touch sensing input device with visual feedback
US20050195159A1 (en) * 2004-02-23 2005-09-08 Hunleth Frank A. Keyboardless text entry
US20050253814A1 (en) * 1999-10-27 2005-11-17 Firooz Ghassabian Integrated keypad system
US7002553B2 (en) * 2001-12-27 2006-02-21 Mark Shkolnikov Active keyboard system for handheld electronic devices
US20060119588A1 (en) * 2004-12-03 2006-06-08 Sung-Min Yoon Apparatus and method of processing information input using a touchpad
US20060152496A1 (en) * 2005-01-13 2006-07-13 602531 British Columbia Ltd. Method, system, apparatus and computer-readable media for directing input associated with keyboard-type device
US20060197750A1 (en) * 2005-03-04 2006-09-07 Apple Computer, Inc. Hand held electronic device with multiple touch sensing devices
US20060227032A1 (en) * 2005-04-11 2006-10-12 Alberto Vidal Apparatus and method to facilitate universal remote control
US20060238510A1 (en) * 2005-04-25 2006-10-26 Georgios Panotopoulos User interface incorporating emulated hard keys
US7139983B2 (en) * 2000-04-10 2006-11-21 Hillcrest Laboratories, Inc. Interactive content guide for television programming
US20070040813A1 (en) * 2003-01-16 2007-02-22 Forword Input, Inc. System and method for continuous stroke word-based text input
US20070046641A1 (en) * 2005-09-01 2007-03-01 Swee Ho Lim Entering a character into an electronic device
US20070060176A1 (en) * 2005-09-13 2007-03-15 Microsoft Corporation Type-ahead keypad input
US20070061754A1 (en) * 2005-08-26 2007-03-15 Veveo, Inc. User interface for visual cooperation between text input and display device
US20070079239A1 (en) * 2000-10-27 2007-04-05 Firooz Ghassabian Data entry system
US20070080954A1 (en) * 2005-10-07 2007-04-12 Research In Motion Limited System and method for using navigational and other commands on a mobile communication device
US20070082708A1 (en) * 2005-10-07 2007-04-12 Research In Motion Limited Device, system, and method for informing users of functions and characters associated with telephone keys
US20070132887A1 (en) * 2005-12-09 2007-06-14 David Boyden On screen display for alpha-numeric input
US20070182595A1 (en) * 2004-06-04 2007-08-09 Firooz Ghasabian Systems to enhance data entry in mobile and fixed environment
US20070188472A1 (en) * 2003-04-18 2007-08-16 Ghassabian Benjamin F Systems to enhance data entry in mobile and fixed environment
US20070195353A1 (en) * 2006-02-17 2007-08-23 Canon Kabushiki Kaisha Image forming apparatus
US20080007434A1 (en) * 2006-07-10 2008-01-10 Luben Hristov Priority and Combination Suppression Techniques (PST/CST) for a Capacitive Keyboard
US20080168364A1 (en) * 2007-01-05 2008-07-10 Apple Computer, Inc. Adaptive acceleration of mouse cursor
US7453439B1 (en) * 2003-01-16 2008-11-18 Forward Input Inc. System and method for continuous stroke word-based text input
US20080316212A1 (en) * 2005-09-20 2008-12-25 Cliff Kushler System and method for a user interface for text editing and menu selection
US20090007001A1 (en) * 2007-06-28 2009-01-01 Matsushita Electric Industrial Co., Ltd. Virtual keypad systems and methods
US20090135162A1 (en) * 2005-03-10 2009-05-28 Koninklijke Philips Electronics, N.V. System and Method For Detecting the Location, Size and Shape of Multiple Objects That Interact With a Touch Screen Display
US7649478B1 (en) * 2005-11-03 2010-01-19 Hyoungsoo Yoon Data entry using sequential keystrokes
US20100103127A1 (en) * 2007-02-23 2010-04-29 Taeun Park Virtual Keyboard Input System Using Pointing Apparatus In Digital Device
US20100245395A1 (en) * 2009-03-31 2010-09-30 Dell Products L.P. Information Handling System with Touch Pad and Point Stick
US7817145B2 (en) * 2008-04-18 2010-10-19 Cirque Corporation Method and system for performing scrolling by movement of a pointing object in a curvilinear path on a touchpad
US20100265200A1 (en) * 2009-04-21 2010-10-21 Samsung Electronics Co., Ltd. Method for processing multi-touch input in mobile device
US7957955B2 (en) * 2007-01-05 2011-06-07 Apple Inc. Method and system for providing word recommendations for text input
US7991784B2 (en) * 2005-01-21 2011-08-02 Prashant Parikh Automatic dynamic contextual data entry completion system
US8036878B2 (en) * 2005-05-18 2011-10-11 Never Wall Treuhand GmbH Device incorporating improved text input mechanism
US8060839B2 (en) * 2006-06-30 2011-11-15 Samsung Electronics Co., Ltd Character input method and mobile communication terminal using the same
US20110279397A1 (en) * 2009-01-26 2011-11-17 Zrro Technologies (2009) Ltd. Device and method for monitoring the object's behavior
US20110279376A1 (en) * 2010-05-12 2011-11-17 Aaron Tang Remote control to operate computer system
US20110279354A1 (en) * 2010-05-12 2011-11-17 Aaron Tang Computer with tv mode
US8065453B2 (en) * 2006-04-06 2011-11-22 Research In Motion Limited Handheld electronic device and associated method employing a multiple-axis input device and learning a context of a text input for use by a disambiguation routine
US8078978B2 (en) * 2007-10-19 2011-12-13 Google Inc. Method and system for predicting text
US8078884B2 (en) * 2006-11-13 2011-12-13 Veveo, Inc. Method of and system for selecting and presenting content based on user identification
US8086602B2 (en) * 2006-04-20 2011-12-27 Veveo Inc. User interface methods and systems for selecting and presenting content based on user navigation and selection actions associated with the content
US8112454B2 (en) * 2006-03-06 2012-02-07 Veveo, Inc. Methods and systems for ordering content items according to learned user preferences
US8117540B2 (en) * 2005-05-18 2012-02-14 Neuer Wall Treuhand Gmbh Method and device incorporating improved text input mechanism
US20120117101A1 (en) * 2010-11-10 2012-05-10 Erland Unruh Text entry with word prediction, completion, or correction supplemented by search of shared corpus
US8296294B2 (en) * 2007-05-25 2012-10-23 Veveo, Inc. Method and system for unified searching across and within multiple documents
US8296484B2 (en) * 2006-03-30 2012-10-23 Harris Corporation Alphanumeric data entry apparatus and method using multicharacter keys of a keypad
US8311829B2 (en) * 2004-06-02 2012-11-13 Tegic Communications, Inc. Multimodal disambiguation of speech recognition
US8316035B2 (en) * 2008-01-16 2012-11-20 International Business Machines Corporation Systems and arrangements of text type-ahead
US8330625B2 (en) * 2007-03-30 2012-12-11 Honeywell International Inc. Aircraft systems with flight management systems that display reports

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH08221169A (en) 1995-02-14 1996-08-30 Matsushita Electric Ind Co Ltd Method and device for displaying software keyboard
JPH09160910A (en) * 1995-12-04 1997-06-20 Matsushita Electric Ind Co Ltd Software keyboard display system
JPH09251347A (en) 1996-03-15 1997-09-22 Matsushita Electric Ind Co Ltd Coordinate input device
JP4459103B2 (en) * 2005-04-13 2010-04-28 シャープ株式会社 Information terminal device, information input method, information input program, and recording medium
DE602006019893D1 (en) * 2006-12-29 2011-03-10 Research In Motion Ltd Portable electronic device with input confirmation and corresponding method

Patent Citations (97)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5038401A (en) * 1989-04-05 1991-08-06 Pioneer Electronic Corporation Transmitter for remote control with operation switches having changeably displayed forms
US5128672A (en) * 1990-10-30 1992-07-07 Apple Computer, Inc. Dynamic predictive keyboard
US5936614A (en) * 1991-04-30 1999-08-10 International Business Machines Corporation User defined keyboard entry system
US5963671A (en) * 1991-11-27 1999-10-05 International Business Machines Corporation Enhancement of soft keyboard operations using trigram prediction
US5652630A (en) * 1995-05-31 1997-07-29 International Business Machines Corporation Video receiver display, three axis remote control, and microcontroller for executing programs
US5724106A (en) * 1995-07-17 1998-03-03 Gateway 2000, Inc. Hand held remote control device with trigger button
US6222528B1 (en) * 1997-03-07 2001-04-24 Cirque Corporation Method and apparatus for data input
US6359270B1 (en) * 1998-09-04 2002-03-19 Ncr Corporation Communications module mounting for domestic appliance
US6396523B1 (en) * 1999-07-29 2002-05-28 Interlink Electronics, Inc. Home entertainment device remote control
US20090037623A1 (en) * 1999-10-27 2009-02-05 Firooz Ghassabian Integrated keypad system
US7020270B1 (en) * 1999-10-27 2006-03-28 Firooz Ghassabian Integrated keypad system
US20050253814A1 (en) * 1999-10-27 2005-11-17 Firooz Ghassabian Integrated keypad system
US8271336B2 (en) * 1999-11-22 2012-09-18 Accenture Global Services Gmbh Increased visibility during order management in a network-based supply chain environment
US20040064351A1 (en) * 1999-11-22 2004-04-01 Mikurak Michael G. Increased visibility during order management in a network-based supply chain environment
US6765557B1 (en) * 2000-04-10 2004-07-20 Interlink Electronics, Inc. Remote control having touch pad to screen mapping
US7139983B2 (en) * 2000-04-10 2006-11-21 Hillcrest Laboratories, Inc. Interactive content guide for television programming
US20070079239A1 (en) * 2000-10-27 2007-04-05 Firooz Ghassabian Data entry system
US20020093535A1 (en) * 2001-01-17 2002-07-18 Murphy Michael William User interface for character entry using a minimum number of selection keys
US20020114654A1 (en) * 2001-02-16 2002-08-22 Toshiyasu Abe Improved Keyboard
US20020122197A1 (en) * 2001-02-21 2002-09-05 Eli Abir Method and apparatus for text input
US6692170B2 (en) * 2001-02-21 2004-02-17 Eli Abir Method and apparatus for text input
US20020149569A1 (en) * 2001-04-12 2002-10-17 International Business Machines Corporation Touchscreen user interface
US20040169635A1 (en) * 2001-07-12 2004-09-02 Ghassabian Benjamin Firooz Features to enhance data entry through a small data entry unit
US7372454B2 (en) * 2001-10-29 2008-05-13 Oqo Incorporated Keyboard with variable-sized keys
US20030080945A1 (en) * 2001-10-29 2003-05-01 Betts-Lacroix Jonathan Keyboard with variable-sized keys
US7002553B2 (en) * 2001-12-27 2006-02-21 Mark Shkolnikov Active keyboard system for handheld electronic devices
US20030197736A1 (en) * 2002-01-16 2003-10-23 Murphy Michael W. User interface for character entry using a minimum number of selection keys
US20030193478A1 (en) * 2002-04-04 2003-10-16 Edwin Ng Reduced keyboard system that emulates QWERTY-type mapping and typing
US7202853B2 (en) * 2002-04-04 2007-04-10 Xrgomics Pte, Ltd. Reduced keyboard system that emulates QWERTY-type mapping and typing
US20040041723A1 (en) * 2002-05-27 2004-03-04 Canon Kabushiki Kaisha, Tokyo, Japan Remote control device
US20040080487A1 (en) * 2002-10-29 2004-04-29 Griffin Jason T. Electronic device having keyboard for thumb typing
US7382358B2 (en) * 2003-01-16 2008-06-03 Forword Input, Inc. System and method for continuous stroke word-based text input
US7098896B2 (en) * 2003-01-16 2006-08-29 Forword Input Inc. System and method for continuous stroke word-based text input
US7453439B1 (en) * 2003-01-16 2008-11-18 Forward Input Inc. System and method for continuous stroke word-based text input
US20040140956A1 (en) * 2003-01-16 2004-07-22 Kushler Clifford A. System and method for continuous stroke word-based text input
US20070040813A1 (en) * 2003-01-16 2007-02-22 Forword Input, Inc. System and method for continuous stroke word-based text input
US20040239624A1 (en) * 2003-04-02 2004-12-02 Artoun Ramian Freehand symbolic input apparatus and method
US20050052406A1 (en) * 2003-04-09 2005-03-10 James Stephanick Selective input system based on tracking of motion parameters of an input device
US7750891B2 (en) * 2003-04-09 2010-07-06 Tegic Communications, Inc. Selective input system based on tracking of motion parameters of an input device
US20070188472A1 (en) * 2003-04-18 2007-08-16 Ghassabian Benjamin F Systems to enhance data entry in mobile and fixed environment
US20050099400A1 (en) * 2003-11-06 2005-05-12 Samsung Electronics Co., Ltd. Apparatus and method for providing vitrtual graffiti and recording medium for the same
US7956845B2 (en) * 2003-11-06 2011-06-07 Samsung Electronics Co., Ltd Apparatus and method for providing virtual graffiti and recording medium for the same
US20050141770A1 (en) * 2003-12-30 2005-06-30 Nokia Corporation Split on-screen keyboard
US20050162402A1 (en) * 2004-01-27 2005-07-28 Watanachote Susornpol J. Methods of interacting with a computer using a finger(s) touch sensing input device with visual feedback
US20050195159A1 (en) * 2004-02-23 2005-09-08 Hunleth Frank A. Keyboardless text entry
US8311829B2 (en) * 2004-06-02 2012-11-13 Tegic Communications, Inc. Multimodal disambiguation of speech recognition
US20070182595A1 (en) * 2004-06-04 2007-08-09 Firooz Ghasabian Systems to enhance data entry in mobile and fixed environment
US20090146848A1 (en) * 2004-06-04 2009-06-11 Ghassabian Firooz Benjamin Systems to enhance data entry in mobile and fixed environment
US20060119588A1 (en) * 2004-12-03 2006-06-08 Sung-Min Yoon Apparatus and method of processing information input using a touchpad
US20060152496A1 (en) * 2005-01-13 2006-07-13 602531 British Columbia Ltd. Method, system, apparatus and computer-readable media for directing input associated with keyboard-type device
US7991784B2 (en) * 2005-01-21 2011-08-02 Prashant Parikh Automatic dynamic contextual data entry completion system
US8311805B2 (en) * 2005-01-21 2012-11-13 Prashant Parikh Automatic dynamic contextual data entry completion system
US20060197750A1 (en) * 2005-03-04 2006-09-07 Apple Computer, Inc. Hand held electronic device with multiple touch sensing devices
US20090135162A1 (en) * 2005-03-10 2009-05-28 Koninklijke Philips Electronics, N.V. System and Method For Detecting the Location, Size and Shape of Multiple Objects That Interact With a Touch Screen Display
US20060227032A1 (en) * 2005-04-11 2006-10-12 Alberto Vidal Apparatus and method to facilitate universal remote control
US20060238510A1 (en) * 2005-04-25 2006-10-26 Georgios Panotopoulos User interface incorporating emulated hard keys
US8117540B2 (en) * 2005-05-18 2012-02-14 Neuer Wall Treuhand Gmbh Method and device incorporating improved text input mechanism
US8036878B2 (en) * 2005-05-18 2011-10-11 Never Wall Treuhand GmbH Device incorporating improved text input mechanism
US20090199092A1 (en) * 2005-06-16 2009-08-06 Firooz Ghassabian Data entry system
US7737999B2 (en) * 2005-08-26 2010-06-15 Veveo, Inc. User interface for visual cooperation between text input and display device
US20070061754A1 (en) * 2005-08-26 2007-03-15 Veveo, Inc. User interface for visual cooperation between text input and display device
US20070046641A1 (en) * 2005-09-01 2007-03-01 Swee Ho Lim Entering a character into an electronic device
US7443316B2 (en) * 2005-09-01 2008-10-28 Motorola, Inc. Entering a character into an electronic device
US7539472B2 (en) * 2005-09-13 2009-05-26 Microsoft Corporation Type-ahead keypad input for an input device
US20070060176A1 (en) * 2005-09-13 2007-03-15 Microsoft Corporation Type-ahead keypad input
US7542029B2 (en) * 2005-09-20 2009-06-02 Cliff Kushler System and method for a user interface for text editing and menu selection
US20080316212A1 (en) * 2005-09-20 2008-12-25 Cliff Kushler System and method for a user interface for text editing and menu selection
US20070082708A1 (en) * 2005-10-07 2007-04-12 Research In Motion Limited Device, system, and method for informing users of functions and characters associated with telephone keys
US7869832B2 (en) * 2005-10-07 2011-01-11 Research In Motion Limited Device, system, and method for informing users of functions and characters associated with telephone keys
US20070080954A1 (en) * 2005-10-07 2007-04-12 Research In Motion Limited System and method for using navigational and other commands on a mobile communication device
US7649478B1 (en) * 2005-11-03 2010-01-19 Hyoungsoo Yoon Data entry using sequential keystrokes
US20070132887A1 (en) * 2005-12-09 2007-06-14 David Boyden On screen display for alpha-numeric input
US8284445B2 (en) * 2006-02-17 2012-10-09 Canon Kabushiki Kaisha Image forming apparatus for forming images from objects
US20070195353A1 (en) * 2006-02-17 2007-08-23 Canon Kabushiki Kaisha Image forming apparatus
US8112454B2 (en) * 2006-03-06 2012-02-07 Veveo, Inc. Methods and systems for ordering content items according to learned user preferences
US8296484B2 (en) * 2006-03-30 2012-10-23 Harris Corporation Alphanumeric data entry apparatus and method using multicharacter keys of a keypad
US8065453B2 (en) * 2006-04-06 2011-11-22 Research In Motion Limited Handheld electronic device and associated method employing a multiple-axis input device and learning a context of a text input for use by a disambiguation routine
US8086602B2 (en) * 2006-04-20 2011-12-27 Veveo Inc. User interface methods and systems for selecting and presenting content based on user navigation and selection actions associated with the content
US8060839B2 (en) * 2006-06-30 2011-11-15 Samsung Electronics Co., Ltd Character input method and mobile communication terminal using the same
US20080007434A1 (en) * 2006-07-10 2008-01-10 Luben Hristov Priority and Combination Suppression Techniques (PST/CST) for a Capacitive Keyboard
US8078884B2 (en) * 2006-11-13 2011-12-13 Veveo, Inc. Method of and system for selecting and presenting content based on user identification
US20080168364A1 (en) * 2007-01-05 2008-07-10 Apple Computer, Inc. Adaptive acceleration of mouse cursor
US7957955B2 (en) * 2007-01-05 2011-06-07 Apple Inc. Method and system for providing word recommendations for text input
US20100103127A1 (en) * 2007-02-23 2010-04-29 Taeun Park Virtual Keyboard Input System Using Pointing Apparatus In Digital Device
US8330625B2 (en) * 2007-03-30 2012-12-11 Honeywell International Inc. Aircraft systems with flight management systems that display reports
US8296294B2 (en) * 2007-05-25 2012-10-23 Veveo, Inc. Method and system for unified searching across and within multiple documents
US20090007001A1 (en) * 2007-06-28 2009-01-01 Matsushita Electric Industrial Co., Ltd. Virtual keypad systems and methods
US8065624B2 (en) * 2007-06-28 2011-11-22 Panasonic Corporation Virtual keypad systems and methods
US8078978B2 (en) * 2007-10-19 2011-12-13 Google Inc. Method and system for predicting text
US8316035B2 (en) * 2008-01-16 2012-11-20 International Business Machines Corporation Systems and arrangements of text type-ahead
US7817145B2 (en) * 2008-04-18 2010-10-19 Cirque Corporation Method and system for performing scrolling by movement of a pointing object in a curvilinear path on a touchpad
US20110279397A1 (en) * 2009-01-26 2011-11-17 Zrro Technologies (2009) Ltd. Device and method for monitoring the object's behavior
US20100245395A1 (en) * 2009-03-31 2010-09-30 Dell Products L.P. Information Handling System with Touch Pad and Point Stick
US20100265200A1 (en) * 2009-04-21 2010-10-21 Samsung Electronics Co., Ltd. Method for processing multi-touch input in mobile device
US20110279376A1 (en) * 2010-05-12 2011-11-17 Aaron Tang Remote control to operate computer system
US20110279354A1 (en) * 2010-05-12 2011-11-17 Aaron Tang Computer with tv mode
US20120117101A1 (en) * 2010-11-10 2012-05-10 Erland Unruh Text entry with word prediction, completion, or correction supplemented by search of shared corpus

Cited By (469)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9646614B2 (en) 2000-03-16 2017-05-09 Apple Inc. Fast, language-independent method for user authentication by voice
US10348654B2 (en) 2003-05-02 2019-07-09 Apple Inc. Method and apparatus for displaying information during an instant messaging session
US8458278B2 (en) 2003-05-02 2013-06-04 Apple Inc. Method and apparatus for displaying information during an instant messaging session
US20100185960A1 (en) * 2003-05-02 2010-07-22 Apple Inc. Method and Apparatus for Displaying Information During an Instant Messaging Session
US10623347B2 (en) 2003-05-02 2020-04-14 Apple Inc. Method and apparatus for displaying information during an instant messaging session
US8554861B2 (en) 2003-05-02 2013-10-08 Apple Inc. Method and apparatus for displaying information during an instant messaging session
US10318871B2 (en) 2005-09-08 2019-06-11 Apple Inc. Method and apparatus for building an intelligent automated assistant
US11928604B2 (en) 2005-09-08 2024-03-12 Apple Inc. Method and apparatus for building an intelligent automated assistant
US9261964B2 (en) 2005-12-30 2016-02-16 Microsoft Technology Licensing, Llc Unintentional touch rejection
US10019080B2 (en) 2005-12-30 2018-07-10 Microsoft Technology Licensing, Llc Unintentional touch rejection
US9594457B2 (en) 2005-12-30 2017-03-14 Microsoft Technology Licensing, Llc Unintentional touch rejection
US9952718B2 (en) 2005-12-30 2018-04-24 Microsoft Technology Licensing, Llc Unintentional touch rejection
US9946370B2 (en) 2005-12-30 2018-04-17 Microsoft Technology Licensing, Llc Unintentional touch rejection
US11762547B2 (en) 2006-09-06 2023-09-19 Apple Inc. Portable electronic device for instant messaging
US10572142B2 (en) 2006-09-06 2020-02-25 Apple Inc. Portable electronic device for instant messaging
US11029838B2 (en) 2006-09-06 2021-06-08 Apple Inc. Touch screen device, method, and graphical user interface for customizing display of content category icons
US11169690B2 (en) 2006-09-06 2021-11-09 Apple Inc. Portable electronic device for instant messaging
US9600174B2 (en) 2006-09-06 2017-03-21 Apple Inc. Portable electronic device for instant messaging
US20080055269A1 (en) * 2006-09-06 2008-03-06 Lemay Stephen O Portable Electronic Device for Instant Messaging
US9304675B2 (en) * 2006-09-06 2016-04-05 Apple Inc. Portable electronic device for instant messaging
US8930191B2 (en) 2006-09-08 2015-01-06 Apple Inc. Paraphrasing of user requests and results by automated digital assistant
US8942986B2 (en) 2006-09-08 2015-01-27 Apple Inc. Determining user intent based on ontologies of domains
US9117447B2 (en) 2006-09-08 2015-08-25 Apple Inc. Using event alert text as input to an automated assistant
US11012942B2 (en) 2007-04-03 2021-05-18 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US10568032B2 (en) 2007-04-03 2020-02-18 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US9954996B2 (en) 2007-06-28 2018-04-24 Apple Inc. Portable electronic device with conversation management for incoming instant messages
US11743375B2 (en) 2007-06-28 2023-08-29 Apple Inc. Portable electronic device with conversation management for incoming instant messages
US20090005011A1 (en) * 2007-06-28 2009-01-01 Greg Christie Portable Electronic Device with Conversation Management for Incoming Instant Messages
US11122158B2 (en) 2007-06-28 2021-09-14 Apple Inc. Portable electronic device with conversation management for incoming instant messages
US8010163B2 (en) * 2007-11-08 2011-08-30 Htc Corporation Method for displaying dialing information and mobile communication device using the method
US20090124293A1 (en) * 2007-11-08 2009-05-14 Tsui Yuan-Mao Method for displaying dialing information and mobile communication device using the method
US11023513B2 (en) 2007-12-20 2021-06-01 Apple Inc. Method and apparatus for searching using an active ontology
US10381016B2 (en) 2008-01-03 2019-08-13 Apple Inc. Methods and apparatus for altering audio output signals
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US9330381B2 (en) 2008-01-06 2016-05-03 Apple Inc. Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars
US9792001B2 (en) 2008-01-06 2017-10-17 Apple Inc. Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars
US20090177981A1 (en) * 2008-01-06 2009-07-09 Greg Christie Portable Electronic Device for Instant Messaging Multiple Recipients
US8407603B2 (en) 2008-01-06 2013-03-26 Apple Inc. Portable electronic device for instant messaging multiple recipients
US10503366B2 (en) 2008-01-06 2019-12-10 Apple Inc. Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars
US10521084B2 (en) 2008-01-06 2019-12-31 Apple Inc. Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars
US11126326B2 (en) 2008-01-06 2021-09-21 Apple Inc. Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars
US9035876B2 (en) 2008-01-14 2015-05-19 Apple Inc. Three-dimensional user interface session control
US9865248B2 (en) 2008-04-05 2018-01-09 Apple Inc. Intelligent text-to-speech conversion
US9626955B2 (en) 2008-04-05 2017-04-18 Apple Inc. Intelligent text-to-speech conversion
US9535906B2 (en) 2008-07-31 2017-01-03 Apple Inc. Mobile device having human language translation capability with positional feedback
US10108612B2 (en) 2008-07-31 2018-10-23 Apple Inc. Mobile device having human language translation capability with positional feedback
US11348582B2 (en) 2008-10-02 2022-05-31 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US10643611B2 (en) 2008-10-02 2020-05-05 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US9959870B2 (en) 2008-12-11 2018-05-01 Apple Inc. Speech recognition involving a mobile device
JP2012515966A (en) * 2009-01-26 2012-07-12 ズッロ・テクノロジーズ・(2009)・リミテッド Device and method for monitoring the behavior of an object
US8830189B2 (en) * 2009-01-26 2014-09-09 Zrro Technologies (2009) Ltd. Device and method for monitoring the object's behavior
US20110279397A1 (en) * 2009-01-26 2011-11-17 Zrro Technologies (2009) Ltd. Device and method for monitoring the object's behavior
US8583421B2 (en) * 2009-03-06 2013-11-12 Motorola Mobility Llc Method and apparatus for psychomotor and psycholinguistic prediction on touch based device
US20100228539A1 (en) * 2009-03-06 2010-09-09 Motorola, Inc. Method and apparatus for psychomotor and psycholinguistic prediction on touch based device
US11080012B2 (en) 2009-06-05 2021-08-03 Apple Inc. Interface for a virtual digital assistant
US10795541B2 (en) 2009-06-05 2020-10-06 Apple Inc. Intelligent organization of tasks items
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US10475446B2 (en) 2009-06-05 2019-11-12 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US20100315266A1 (en) * 2009-06-15 2010-12-16 Microsoft Corporation Predictive interfaces with usability constraints
US10283110B2 (en) 2009-07-02 2019-05-07 Apple Inc. Methods and apparatuses for automatic speech recognition
US20120139873A1 (en) * 2009-07-14 2012-06-07 Chih-Hung Li Touch-controlled electronic apparatus and related control method
US20110012843A1 (en) * 2009-07-14 2011-01-20 Chih-Hung Li Touch-controlled electronic apparatus and related control method
US20110148762A1 (en) * 2009-12-22 2011-06-23 Universal Electronics Inc. System and method for multi-mode command input
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
US10706841B2 (en) 2010-01-18 2020-07-07 Apple Inc. Task flow identification based on user intent
US9548050B2 (en) 2010-01-18 2017-01-17 Apple Inc. Intelligent automated assistant
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US8903716B2 (en) 2010-01-18 2014-12-02 Apple Inc. Personalized vocabulary for digital assistant
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US8892446B2 (en) 2010-01-18 2014-11-18 Apple Inc. Service orchestration for intelligent automated assistant
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US10741185B2 (en) 2010-01-18 2020-08-11 Apple Inc. Intelligent automated assistant
US11423886B2 (en) 2010-01-18 2022-08-23 Apple Inc. Task flow identification based on user intent
US20110187647A1 (en) * 2010-02-04 2011-08-04 Charles Howard Woloszynski Method and apparatus for virtual keyboard interactions from secondary surfaces
US20110212876A1 (en) * 2010-02-10 2011-09-01 Michelle Meek Cleaning composition comprising amylase variants with high stability in the presence of a chelating agent
US10126936B2 (en) 2010-02-12 2018-11-13 Microsoft Technology Licensing, Llc Typing assistance for editing
US20110202836A1 (en) * 2010-02-12 2011-08-18 Microsoft Corporation Typing assistance for editing
US20110202876A1 (en) * 2010-02-12 2011-08-18 Microsoft Corporation User-centric soft keyboard predictive technologies
US9165257B2 (en) 2010-02-12 2015-10-20 Microsoft Technology Licensing, Llc Typing assistance for editing
US20110201387A1 (en) * 2010-02-12 2011-08-18 Microsoft Corporation Real-time typing assistance
US9613015B2 (en) 2010-02-12 2017-04-04 Microsoft Technology Licensing, Llc User-centric soft keyboard predictive technologies
US10156981B2 (en) 2010-02-12 2018-12-18 Microsoft Technology Licensing, Llc User-centric soft keyboard predictive technologies
US8782556B2 (en) 2010-02-12 2014-07-15 Microsoft Corporation User-centric soft keyboard predictive technologies
US9633660B2 (en) 2010-02-25 2017-04-25 Apple Inc. User profiling for voice input processing
US10692504B2 (en) 2010-02-25 2020-06-23 Apple Inc. User profiling for voice input processing
US10049675B2 (en) 2010-02-25 2018-08-14 Apple Inc. User profiling for voice input processing
US8854323B2 (en) 2010-05-20 2014-10-07 Panasonic Intellectual Property Corporation Of America Operating apparatus, operating method, program, recording medium, and integrated circuit
US20110310019A1 (en) * 2010-06-16 2011-12-22 International Business Machines Corporation Reconfiguration of virtual keyboard
US8648809B2 (en) * 2010-06-16 2014-02-11 International Business Machines Corporation Reconfiguration of virtual keyboard
US9201501B2 (en) 2010-07-20 2015-12-01 Apple Inc. Adaptive projector
US9158375B2 (en) 2010-07-20 2015-10-13 Apple Inc. Interactive reality augmentation for natural interaction
US8959013B2 (en) * 2010-09-27 2015-02-17 Apple Inc. Virtual keyboard for a non-tactile three dimensional user interface
US20120078614A1 (en) * 2010-09-27 2012-03-29 Primesense Ltd. Virtual keyboard for a non-tactile three dimensional user interface
US9501161B2 (en) 2010-10-22 2016-11-22 Hewlett-Packard Development Company, L.P. User interface for facilitating character input
US20120098743A1 (en) * 2010-10-26 2012-04-26 Pei-Ling Lai Input method, input device, and computer system
US10776476B2 (en) 2010-11-29 2020-09-15 Biocatch Ltd. System, device, and method of visual login
US11314849B2 (en) 2010-11-29 2022-04-26 Biocatch Ltd. Method, device, and system of detecting a lie of a user who inputs data
US11269977B2 (en) 2010-11-29 2022-03-08 Biocatch Ltd. System, apparatus, and method of collecting and processing data in electronic devices
US11425563B2 (en) 2010-11-29 2022-08-23 Biocatch Ltd. Method, device, and system of differentiating between a cyber-attacker and a legitimate user
US11210674B2 (en) 2010-11-29 2021-12-28 Biocatch Ltd. Method, device, and system of detecting mule accounts and accounts used for money laundering
US10262324B2 (en) 2010-11-29 2019-04-16 Biocatch Ltd. System, device, and method of differentiating among users based on user-specific page navigation sequence
US10298614B2 (en) * 2010-11-29 2019-05-21 Biocatch Ltd. System, device, and method of generating and managing behavioral biometric cookies
US10404729B2 (en) 2010-11-29 2019-09-03 Biocatch Ltd. Device, method, and system of generating fraud-alerts for cyber-attacks
US11580553B2 (en) 2010-11-29 2023-02-14 Biocatch Ltd. Method, device, and system of detecting mule accounts and accounts used for money laundering
US10474815B2 (en) 2010-11-29 2019-11-12 Biocatch Ltd. System, device, and method of detecting malicious automatic script and code injection
US11223619B2 (en) 2010-11-29 2022-01-11 Biocatch Ltd. Device, system, and method of user authentication based on user-specific characteristics of task performance
US10586036B2 (en) 2010-11-29 2020-03-10 Biocatch Ltd. System, device, and method of recovery and resetting of user authentication factor
US10621585B2 (en) 2010-11-29 2020-04-14 Biocatch Ltd. Contextual mapping of web-pages, and generation of fraud-relatedness score-values
US10949757B2 (en) 2010-11-29 2021-03-16 Biocatch Ltd. System, device, and method of detecting user identity based on motor-control loop model
US11330012B2 (en) 2010-11-29 2022-05-10 Biocatch Ltd. System, method, and device of authenticating a user based on selfie image or selfie video
US20210329030A1 (en) * 2010-11-29 2021-10-21 Biocatch Ltd. Device, System, and Method of Detecting Vishing Attacks
US11838118B2 (en) * 2010-11-29 2023-12-05 Biocatch Ltd. Device, system, and method of detecting vishing attacks
US10917431B2 (en) 2010-11-29 2021-02-09 Biocatch Ltd. System, method, and device of authenticating a user based on selfie image or selfie video
US10728761B2 (en) 2010-11-29 2020-07-28 Biocatch Ltd. Method, device, and system of detecting a lie of a user who inputs data
US10949514B2 (en) 2010-11-29 2021-03-16 Biocatch Ltd. Device, system, and method of differentiating among users based on detection of hardware components
US10747305B2 (en) 2010-11-29 2020-08-18 Biocatch Ltd. Method, system, and device of authenticating identity of a user of an electronic device
US10897482B2 (en) 2010-11-29 2021-01-19 Biocatch Ltd. Method, device, and system of back-coloring, forward-coloring, and fraud detection
US11250435B2 (en) 2010-11-29 2022-02-15 Biocatch Ltd. Contextual mapping of web-pages, and generation of fraud-relatedness score-values
US10834590B2 (en) 2010-11-29 2020-11-10 Biocatch Ltd. Method, device, and system of differentiating between a cyber-attacker and a legitimate user
US20120137244A1 (en) * 2010-11-30 2012-05-31 Inventec Corporation Touch device input device and operation method of the same
US8872762B2 (en) 2010-12-08 2014-10-28 Primesense Ltd. Three dimensional user interface cursor control
US8933876B2 (en) 2010-12-13 2015-01-13 Apple Inc. Three dimensional user interface session control
US10762293B2 (en) 2010-12-22 2020-09-01 Apple Inc. Using parts-of-speech tagging and named entity recognition for spelling correction
US9454225B2 (en) 2011-02-09 2016-09-27 Apple Inc. Gaze-based display control
US9285874B2 (en) 2011-02-09 2016-03-15 Apple Inc. Gaze detection in a 3D mapping environment
US9342146B2 (en) 2011-02-09 2016-05-17 Apple Inc. Pointing-based display interaction
US8719724B2 (en) 2011-03-16 2014-05-06 Honeywell International Inc. Method for enlarging characters displayed on an adaptive touch screen key pad
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US10102359B2 (en) 2011-03-21 2018-10-16 Apple Inc. Device access using voice authentication
US10417405B2 (en) 2011-03-21 2019-09-17 Apple Inc. Device access using voice authentication
US11120372B2 (en) 2011-06-03 2021-09-14 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10706373B2 (en) 2011-06-03 2020-07-07 Apple Inc. Performing actions associated with task items that represent tasks to perform
US11350253B2 (en) 2011-06-03 2022-05-31 Apple Inc. Active transport based notifications
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US9377865B2 (en) 2011-07-05 2016-06-28 Apple Inc. Zoom-based gesture user interface
US9459758B2 (en) 2011-07-05 2016-10-04 Apple Inc. Gesture-based interface with enhanced features
US8881051B2 (en) 2011-07-05 2014-11-04 Primesense Ltd Zoom-based gesture user interface
US9030498B2 (en) 2011-08-15 2015-05-12 Apple Inc. Combining explicit select gestures and timeclick in a non-tactile three dimensional user interface
US9218063B2 (en) 2011-08-24 2015-12-22 Apple Inc. Sessionless pointing user interface
US9122311B2 (en) 2011-08-24 2015-09-01 Apple Inc. Visual feedback for tactile and non-tactile user interfaces
US9798393B2 (en) 2011-08-29 2017-10-24 Apple Inc. Text correction processing
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
WO2013078171A1 (en) * 2011-11-23 2013-05-30 Microsoft Corporation Dynamic scaling of touch sensor
US20130127738A1 (en) * 2011-11-23 2013-05-23 Microsoft Corporation Dynamic scaling of touch sensor
CN102937876A (en) * 2011-11-23 2013-02-20 微软公司 Dynamic scaling of a touch sensor
EP2597548A3 (en) * 2011-11-23 2013-12-11 Microsoft Corporation Dynamic scaling of touch sensor
US20130181902A1 (en) * 2012-01-17 2013-07-18 Microsoft Corporation Skinnable touch device grip patterns
CN104054043A (en) * 2012-01-17 2014-09-17 微软公司 Skinnable touch device grip patterns
US9519419B2 (en) * 2012-01-17 2016-12-13 Microsoft Technology Licensing, Llc Skinnable touch device grip patterns
US9354748B2 (en) 2012-02-13 2016-05-31 Microsoft Technology Licensing, Llc Optical stylus interaction
US9229534B2 (en) 2012-02-28 2016-01-05 Apple Inc. Asymmetric mapping for tactile and non-tactile user interfaces
US9460029B2 (en) 2012-03-02 2016-10-04 Microsoft Technology Licensing, Llc Pressure sensitive keys
US8854799B2 (en) 2012-03-02 2014-10-07 Microsoft Corporation Flux fountain
US9275809B2 (en) 2012-03-02 2016-03-01 Microsoft Technology Licensing, Llc Device camera angle
US9766663B2 (en) 2012-03-02 2017-09-19 Microsoft Technology Licensing, Llc Hinge for component attachment
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US9176901B2 (en) 2012-03-02 2015-11-03 Microsoft Technology Licensing, Llc Flux fountain
USRE48963E1 (en) 2012-03-02 2022-03-08 Microsoft Technology Licensing, Llc Connection device for computing devices
US9793073B2 (en) 2012-03-02 2017-10-17 Microsoft Technology Licensing, Llc Backlighting a fabric enclosure of a flexible cover
US9098117B2 (en) 2012-03-02 2015-08-04 Microsoft Technology Licensing, Llc Classifying the intent of user input
US9268373B2 (en) 2012-03-02 2016-02-23 Microsoft Technology Licensing, Llc Flexible hinge spine
US9075566B2 (en) 2012-03-02 2015-07-07 Microsoft Technoogy Licensing, LLC Flexible hinge spine
US9064654B2 (en) 2012-03-02 2015-06-23 Microsoft Technology Licensing, Llc Method of manufacturing an input device
US9298236B2 (en) 2012-03-02 2016-03-29 Microsoft Technology Licensing, Llc Multi-stage power adapter configured to provide a first power level upon initial connection of the power adapter to the host device and a second power level thereafter upon notification from the host device to the power adapter
US9304948B2 (en) * 2012-03-02 2016-04-05 Microsoft Technology Licensing, Llc Sensing user input at display area edge
US9852855B2 (en) 2012-03-02 2017-12-26 Microsoft Technology Licensing, Llc Pressure sensitive key normalization
US9047207B2 (en) 2012-03-02 2015-06-02 Microsoft Technology Licensing, Llc Mobile device power state
US10963087B2 (en) 2012-03-02 2021-03-30 Microsoft Technology Licensing, Llc Pressure sensitive keys
US9465412B2 (en) 2012-03-02 2016-10-11 Microsoft Technology Licensing, Llc Input device layers and nesting
US8947864B2 (en) 2012-03-02 2015-02-03 Microsoft Corporation Flexible hinge and removable attachment
US9870066B2 (en) 2012-03-02 2018-01-16 Microsoft Technology Licensing, Llc Method of manufacturing an input device
US9619071B2 (en) 2012-03-02 2017-04-11 Microsoft Technology Licensing, Llc Computing device and an apparatus having sensors configured for measuring spatial information indicative of a position of the computing devices
US9111703B2 (en) 2012-03-02 2015-08-18 Microsoft Technology Licensing, Llc Sensor stack venting
US9304949B2 (en) 2012-03-02 2016-04-05 Microsoft Technology Licensing, Llc Sensing user input at display area edge
US9904327B2 (en) 2012-03-02 2018-02-27 Microsoft Technology Licensing, Llc Flexible hinge and removable attachment
US8935774B2 (en) 2012-03-02 2015-01-13 Microsoft Corporation Accessory device authentication
US8903517B2 (en) 2012-03-02 2014-12-02 Microsoft Corporation Computer device and an apparatus having sensors configured for measuring spatial information indicative of a position of the computing devices
US8896993B2 (en) 2012-03-02 2014-11-25 Microsoft Corporation Input device layers and nesting
US9946307B2 (en) 2012-03-02 2018-04-17 Microsoft Technology Licensing, Llc Classifying the intent of user input
US9710093B2 (en) 2012-03-02 2017-07-18 Microsoft Technology Licensing, Llc Pressure sensitive key normalization
US9618977B2 (en) 2012-03-02 2017-04-11 Microsoft Technology Licensing, Llc Input device securing techniques
US9116550B2 (en) 2012-03-02 2015-08-25 Microsoft Technology Licensing, Llc Device kickstand
US8873227B2 (en) 2012-03-02 2014-10-28 Microsoft Corporation Flexible hinge support layer
US9678542B2 (en) 2012-03-02 2017-06-13 Microsoft Technology Licensing, Llc Multiple position input device cover
US9176900B2 (en) 2012-03-02 2015-11-03 Microsoft Technology Licensing, Llc Flexible hinge and removable attachment
US9426905B2 (en) 2012-03-02 2016-08-23 Microsoft Technology Licensing, Llc Connection device for computing devices
US9158384B2 (en) 2012-03-02 2015-10-13 Microsoft Technology Licensing, Llc Flexible hinge protrusion attachment
US9411751B2 (en) 2012-03-02 2016-08-09 Microsoft Technology Licensing, Llc Key formation
US8850241B2 (en) 2012-03-02 2014-09-30 Microsoft Corporation Multi-stage power adapter configured to provide low power upon initial connection of the power adapter to the host device and high power thereafter upon notification from the host device to the power adapter
US9158383B2 (en) 2012-03-02 2015-10-13 Microsoft Technology Licensing, Llc Force concentrator
US8830668B2 (en) 2012-03-02 2014-09-09 Microsoft Corporation Flexible hinge and removable attachment
US8791382B2 (en) 2012-03-02 2014-07-29 Microsoft Corporation Input device securing techniques
US8780541B2 (en) 2012-03-02 2014-07-15 Microsoft Corporation Flexible hinge and removable attachment
US10013030B2 (en) 2012-03-02 2018-07-03 Microsoft Technology Licensing, Llc Multiple position input device cover
US8780540B2 (en) 2012-03-02 2014-07-15 Microsoft Corporation Flexible hinge and removable attachment
US20130229363A1 (en) * 2012-03-02 2013-09-05 Christopher A. Whitman Sensing User Input At Display Area Edge
US9360893B2 (en) 2012-03-02 2016-06-07 Microsoft Technology Licensing, Llc Input device writing surface
US9146620B2 (en) 2012-03-02 2015-09-29 Microsoft Technology Licensing, Llc Input device assembly
US11069336B2 (en) 2012-03-02 2021-07-20 Apple Inc. Systems and methods for name pronunciation
US9134808B2 (en) 2012-03-02 2015-09-15 Microsoft Technology Licensing, Llc Device kickstand
US9134807B2 (en) 2012-03-02 2015-09-15 Microsoft Technology Licensing, Llc Pressure sensitive key normalization
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US9377863B2 (en) 2012-03-26 2016-06-28 Apple Inc. Gaze-enhanced virtual touchscreen
US11169611B2 (en) 2012-03-26 2021-11-09 Apple Inc. Enhanced virtual touchpad
US9772744B2 (en) 2012-04-19 2017-09-26 Sony Corporation Information processing apparatus, information processing method, program, and information processing system
US10162480B2 (en) 2012-04-19 2018-12-25 Sony Corporation Information processing apparatus, information processing method, program, and information processing system
US10942620B2 (en) 2012-04-19 2021-03-09 Sony Corporation Information processing apparatus, information processing method, program, and information processing system
WO2013157232A3 (en) * 2012-04-19 2014-01-09 Sony Corporation Information processing apparatus, information processing method, program, and information processing system
US10110708B2 (en) * 2012-04-28 2018-10-23 Alibaba Group Holding Limited Performing autocomplete of content
US20130290410A1 (en) * 2012-04-28 2013-10-31 Alibaba Group Holding Limited Performing autocomplete of content
US20170163770A1 (en) * 2012-04-28 2017-06-08 Alibaba Group Holding Limited Performing autocomplete of content
US9584626B2 (en) * 2012-04-28 2017-02-28 Alibaba Group Holding Limited Performing autocomplete of content
US9953088B2 (en) 2012-05-14 2018-04-24 Apple Inc. Crowd sourcing information to fulfill user requests
US9348605B2 (en) 2012-05-14 2016-05-24 Microsoft Technology Licensing, Llc System and method for accessory device architecture that passes human interface device (HID) data via intermediate processor
US8949477B2 (en) 2012-05-14 2015-02-03 Microsoft Technology Licensing, Llc Accessory device architecture
US9959241B2 (en) 2012-05-14 2018-05-01 Microsoft Technology Licensing, Llc System and method for accessory device architecture that passes via intermediate processor a descriptor when processing in a low power state
US11269678B2 (en) 2012-05-15 2022-03-08 Apple Inc. Systems and methods for integrating third party services with a digital assistant
US10079014B2 (en) 2012-06-08 2018-09-18 Apple Inc. Name recognition system
US10031556B2 (en) 2012-06-08 2018-07-24 Microsoft Technology Licensing, Llc User experience adaptation
US10107994B2 (en) 2012-06-12 2018-10-23 Microsoft Technology Licensing, Llc Wide field-of-view virtual image projector
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9824808B2 (en) 2012-08-20 2017-11-21 Microsoft Technology Licensing, Llc Switchable magnetic lock
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9971774B2 (en) 2012-09-19 2018-05-15 Apple Inc. Voice-based media searching
US9544504B2 (en) 2012-11-02 2017-01-10 Microsoft Technology Licensing, Llc Rapid synchronized lighting and shuttering
US9582122B2 (en) 2012-11-12 2017-02-28 Microsoft Technology Licensing, Llc Touch-sensitive bezel techniques
US10656750B2 (en) 2012-11-12 2020-05-19 Microsoft Technology Licensing, Llc Touch-sensitive bezel techniques
US10986393B2 (en) 2013-01-07 2021-04-20 Samsung Electronics Co., Ltd. Display apparatus, method for UI display thereof and computer-readable recording medium
US11930237B2 (en) 2013-01-07 2024-03-12 Samsung Electronics Co., Ltd. Display apparatus, method for UI display thereof and computer-readable recording medium
US11363324B2 (en) 2013-01-07 2022-06-14 Samsung Electronics Co., Ltd. Display apparatus, method for UI display thereof and computer-readable recording medium
WO2014107064A1 (en) * 2013-01-07 2014-07-10 Samsung Electronics Co., Ltd. Display apparatus, method for ui display thereof and computer-readable recording medium
US11665387B2 (en) 2013-01-07 2023-05-30 Samsung Electronics Co., Ltd. Display apparatus, method for UI display thereof and computer-readable recording medium
US10199051B2 (en) 2013-02-07 2019-02-05 Apple Inc. Voice trigger for a digital assistant
US10978090B2 (en) 2013-02-07 2021-04-13 Apple Inc. Voice trigger for a digital assistant
US10714117B2 (en) 2013-02-07 2020-07-14 Apple Inc. Voice trigger for a digital assistant
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
US11388291B2 (en) 2013-03-14 2022-07-12 Apple Inc. System and method for processing voicemail
US11798547B2 (en) 2013-03-15 2023-10-24 Apple Inc. Voice activated device for use with a voice-based digital assistant
US9697822B1 (en) 2013-03-15 2017-07-04 Apple Inc. System and method for updating an adaptive speech recognition model
US9922642B2 (en) 2013-03-15 2018-03-20 Apple Inc. Training an at least partial voice command system
US9304549B2 (en) 2013-03-28 2016-04-05 Microsoft Technology Licensing, Llc Hinge mechanism for rotatable component attachment
WO2014178813A1 (en) * 2013-04-29 2014-11-06 Hewlett-Packard Development Company, L.P. Device for displaying a received user interface
US9633674B2 (en) 2013-06-07 2017-04-25 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9966060B2 (en) 2013-06-07 2018-05-08 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
US9620104B2 (en) 2013-06-07 2017-04-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9966068B2 (en) 2013-06-08 2018-05-08 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US10657961B2 (en) 2013-06-08 2020-05-19 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US11048473B2 (en) 2013-06-09 2021-06-29 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US11727219B2 (en) 2013-06-09 2023-08-15 Apple Inc. System and method for inferring user intent from speech inputs
US10769385B2 (en) 2013-06-09 2020-09-08 Apple Inc. System and method for inferring user intent from speech inputs
US10185542B2 (en) 2013-06-09 2019-01-22 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
US9300784B2 (en) 2013-06-13 2016-03-29 Apple Inc. System and method for emergency calls initiated by voice command
US20150029402A1 (en) * 2013-07-26 2015-01-29 Tianjin Funayuanchuang Technology Co.,Ltd. Remote controller, system, and method for controlling remote controller
US10791216B2 (en) 2013-08-06 2020-09-29 Apple Inc. Auto-activating smart responses based on activities from remote devices
WO2015052724A1 (en) * 2013-10-07 2015-04-16 Deshmukh Rakesh An indian language keypad
US11314370B2 (en) 2013-12-06 2022-04-26 Apple Inc. Method for extracting salient dialog usage from live data
US9477337B2 (en) 2014-03-14 2016-10-25 Microsoft Technology Licensing, Llc Conductive trace routing for display and bezel sensors
US9946383B2 (en) 2014-03-14 2018-04-17 Microsoft Technology Licensing, Llc Conductive trace routing for display and bezel sensors
US10120420B2 (en) 2014-03-21 2018-11-06 Microsoft Technology Licensing, Llc Lockable display and techniques enabling use of lockable displays
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US10592095B2 (en) 2014-05-23 2020-03-17 Apple Inc. Instantaneous speaking of content on touch devices
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
US10497365B2 (en) 2014-05-30 2019-12-03 Apple Inc. Multi-command single utterance input method
US10714095B2 (en) 2014-05-30 2020-07-14 Apple Inc. Intelligent assistant for home automation
US10169329B2 (en) 2014-05-30 2019-01-01 Apple Inc. Exemplar-based natural language processing
US11133008B2 (en) 2014-05-30 2021-09-28 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US10699717B2 (en) 2014-05-30 2020-06-30 Apple Inc. Intelligent assistant for home automation
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US10083690B2 (en) 2014-05-30 2018-09-25 Apple Inc. Better resolution when referencing to concepts
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US10289433B2 (en) 2014-05-30 2019-05-14 Apple Inc. Domain specific language for encoding assistant dialog
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US10657966B2 (en) 2014-05-30 2020-05-19 Apple Inc. Better resolution when referencing to concepts
US11257504B2 (en) 2014-05-30 2022-02-22 Apple Inc. Intelligent assistant for home automation
US10417344B2 (en) 2014-05-30 2019-09-17 Apple Inc. Exemplar-based natural language processing
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US10878809B2 (en) 2014-05-30 2020-12-29 Apple Inc. Multi-command single utterance input method
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US10904611B2 (en) 2014-06-30 2021-01-26 Apple Inc. Intelligent automated assistant for TV user interactions
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US9668024B2 (en) 2014-06-30 2017-05-30 Apple Inc. Intelligent automated assistant for TV user interactions
US10324733B2 (en) 2014-07-30 2019-06-18 Microsoft Technology Licensing, Llc Shutdown notifications
US10446141B2 (en) 2014-08-28 2019-10-15 Apple Inc. Automatic speech recognition based on user feedback
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US10431204B2 (en) 2014-09-11 2019-10-01 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US10156889B2 (en) 2014-09-15 2018-12-18 Microsoft Technology Licensing, Llc Inductive peripheral retention device
US9606986B2 (en) 2014-09-29 2017-03-28 Apple Inc. Integrated word N-gram and class M-gram language models
US10453443B2 (en) 2014-09-30 2019-10-22 Apple Inc. Providing an indication of the suitability of speech recognition
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US10438595B2 (en) 2014-09-30 2019-10-08 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US9447620B2 (en) 2014-09-30 2016-09-20 Microsoft Technology Licensing, Llc Hinge mechanism with multiple preset positions
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US9964998B2 (en) 2014-09-30 2018-05-08 Microsoft Technology Licensing, Llc Hinge mechanism with multiple preset positions
US9986419B2 (en) 2014-09-30 2018-05-29 Apple Inc. Social reminders
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US10390213B2 (en) 2014-09-30 2019-08-20 Apple Inc. Social reminders
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
US11556230B2 (en) 2014-12-02 2023-01-17 Apple Inc. Data detection
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US11231904B2 (en) 2015-03-06 2022-01-25 Apple Inc. Reducing response latency of intelligent automated assistants
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US11087759B2 (en) 2015-03-08 2021-08-10 Apple Inc. Virtual assistant activation
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US10311871B2 (en) 2015-03-08 2019-06-04 Apple Inc. Competing devices responding to voice triggers
US10529332B2 (en) 2015-03-08 2020-01-07 Apple Inc. Virtual assistant activation
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US10930282B2 (en) 2015-03-08 2021-02-23 Apple Inc. Competing devices responding to voice triggers
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9779225B2 (en) 2015-04-08 2017-10-03 Google Inc. Method and system to provide access to secure features of a device
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US10001857B2 (en) 2015-05-13 2018-06-19 Seiko Epson Corporation Display apparatus and method of controlling display apparatus
US11468282B2 (en) 2015-05-15 2022-10-11 Apple Inc. Virtual assistant in a communication session
US20160342294A1 (en) * 2015-05-19 2016-11-24 Google Inc. Multi-switch option scanning
US10067670B2 (en) * 2015-05-19 2018-09-04 Google Llc Multi-switch option scanning
US11127397B2 (en) 2015-05-27 2021-09-21 Apple Inc. Device voice control
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US10356243B2 (en) 2015-06-05 2019-07-16 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US10681212B2 (en) 2015-06-05 2020-06-09 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US9752361B2 (en) 2015-06-18 2017-09-05 Microsoft Technology Licensing, Llc Multistage hinge
US11238349B2 (en) 2015-06-25 2022-02-01 Biocatch Ltd. Conditional behavioural biometrics
US10719765B2 (en) 2015-06-25 2020-07-21 Biocatch Ltd. Conditional behavioral biometrics
US11010127B2 (en) 2015-06-29 2021-05-18 Apple Inc. Virtual assistant for media playback
US9864415B2 (en) 2015-06-30 2018-01-09 Microsoft Technology Licensing, Llc Multistage friction hinge
US10606322B2 (en) 2015-06-30 2020-03-31 Microsoft Technology Licensing, Llc Multistage friction hinge
US11323451B2 (en) 2015-07-09 2022-05-03 Biocatch Ltd. System, device, and method for detection of proxy server
US10523680B2 (en) * 2015-07-09 2019-12-31 Biocatch Ltd. System, device, and method for detecting a proxy server
US10834090B2 (en) * 2015-07-09 2020-11-10 Biocatch Ltd. System, device, and method for detection of proxy server
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US11500672B2 (en) 2015-09-08 2022-11-15 Apple Inc. Distributed personal assistant
US11126400B2 (en) 2015-09-08 2021-09-21 Apple Inc. Zero latency digital assistant
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US11526368B2 (en) 2015-11-06 2022-12-13 Apple Inc. Intelligent automated assistant in a messaging environment
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10354652B2 (en) 2015-12-02 2019-07-16 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10942703B2 (en) 2015-12-23 2021-03-09 Apple Inc. Proactive assistance based on dialog communication between devices
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US10344797B2 (en) 2016-04-05 2019-07-09 Microsoft Technology Licensing, Llc Hinge with multiple preset positions
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US11227589B2 (en) 2016-06-06 2022-01-18 Apple Inc. Intelligent list reading
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US11069347B2 (en) 2016-06-08 2021-07-20 Apple Inc. Intelligent automated assistant for media exploration
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US10354011B2 (en) 2016-06-09 2019-07-16 Apple Inc. Intelligent automated assistant in a home environment
US11037565B2 (en) 2016-06-10 2021-06-15 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10733993B2 (en) 2016-06-10 2020-08-04 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10521466B2 (en) 2016-06-11 2019-12-31 Apple Inc. Data driven natural language event detection and classification
US10269345B2 (en) 2016-06-11 2019-04-23 Apple Inc. Intelligent task discovery
US10089072B2 (en) 2016-06-11 2018-10-02 Apple Inc. Intelligent device arbitration and control
US10942702B2 (en) 2016-06-11 2021-03-09 Apple Inc. Intelligent device arbitration and control
US11152002B2 (en) 2016-06-11 2021-10-19 Apple Inc. Application integration with a digital assistant
US10297253B2 (en) 2016-06-11 2019-05-21 Apple Inc. Application integration with a digital assistant
US10580409B2 (en) 2016-06-11 2020-03-03 Apple Inc. Application integration with a digital assistant
US11055395B2 (en) 2016-07-08 2021-07-06 Biocatch Ltd. Step-up authentication
US10474753B2 (en) 2016-09-07 2019-11-12 Apple Inc. Language identification using recurrent neural networks
US10037057B2 (en) 2016-09-22 2018-07-31 Microsoft Technology Licensing, Llc Friction hinge
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US10553215B2 (en) 2016-09-23 2020-02-04 Apple Inc. Intelligent automated assistant
US10579784B2 (en) 2016-11-02 2020-03-03 Biocatch Ltd. System, device, and method of secure utilization of fingerprints for user authentication
US10685355B2 (en) * 2016-12-04 2020-06-16 Biocatch Ltd. Method, device, and system of detecting mule accounts and accounts used for money laundering
US11281993B2 (en) 2016-12-05 2022-03-22 Apple Inc. Model and ensemble compression for metric learning
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US11204787B2 (en) 2017-01-09 2021-12-21 Apple Inc. Application integration with a digital assistant
US11656884B2 (en) 2017-01-09 2023-05-23 Apple Inc. Application integration with a digital assistant
US10741181B2 (en) 2017-05-09 2020-08-11 Apple Inc. User interface for correcting recognition errors
US10332518B2 (en) 2017-05-09 2019-06-25 Apple Inc. User interface for correcting recognition errors
US10417266B2 (en) 2017-05-09 2019-09-17 Apple Inc. Context-aware ranking of intelligent response suggestions
US10726832B2 (en) 2017-05-11 2020-07-28 Apple Inc. Maintaining privacy of personal information
US10755703B2 (en) 2017-05-11 2020-08-25 Apple Inc. Offline personal assistant
US11599331B2 (en) 2017-05-11 2023-03-07 Apple Inc. Maintaining privacy of personal information
US10847142B2 (en) 2017-05-11 2020-11-24 Apple Inc. Maintaining privacy of personal information
US10395654B2 (en) 2017-05-11 2019-08-27 Apple Inc. Text normalization based on a data-driven learning network
US11405466B2 (en) 2017-05-12 2022-08-02 Apple Inc. Synchronization and task delegation of a digital assistant
US11301477B2 (en) 2017-05-12 2022-04-12 Apple Inc. Feedback analysis of a digital assistant
US10789945B2 (en) 2017-05-12 2020-09-29 Apple Inc. Low-latency intelligent automated assistant
US11380310B2 (en) 2017-05-12 2022-07-05 Apple Inc. Low-latency intelligent automated assistant
US10410637B2 (en) 2017-05-12 2019-09-10 Apple Inc. User-specific acoustic models
US10791176B2 (en) 2017-05-12 2020-09-29 Apple Inc. Synchronization and task delegation of a digital assistant
US10810274B2 (en) 2017-05-15 2020-10-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
US10482874B2 (en) 2017-05-15 2019-11-19 Apple Inc. Hierarchical belief states for digital assistants
US10303715B2 (en) 2017-05-16 2019-05-28 Apple Inc. Intelligent automated assistant for media exploration
US11532306B2 (en) 2017-05-16 2022-12-20 Apple Inc. Detecting a trigger of a digital assistant
US10909171B2 (en) 2017-05-16 2021-02-02 Apple Inc. Intelligent automated assistant for media exploration
US10748546B2 (en) 2017-05-16 2020-08-18 Apple Inc. Digital assistant services based on device capabilities
US10403278B2 (en) 2017-05-16 2019-09-03 Apple Inc. Methods and systems for phonetic matching in digital assistant services
US10311144B2 (en) 2017-05-16 2019-06-04 Apple Inc. Emoji word sense disambiguation
US11217255B2 (en) 2017-05-16 2022-01-04 Apple Inc. Far-field extension for digital assistant services
US10657328B2 (en) 2017-06-02 2020-05-19 Apple Inc. Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling
US10397262B2 (en) 2017-07-20 2019-08-27 Biocatch Ltd. Device, system, and method of detecting overlay malware
US10417339B2 (en) 2017-08-18 2019-09-17 Kyocera Document Solutions Inc. Suggestion of alternate user input using different user interface
US10445429B2 (en) 2017-09-21 2019-10-15 Apple Inc. Natural language understanding using vocabularies with compressed serialized tries
US10755051B2 (en) 2017-09-29 2020-08-25 Apple Inc. Rule-based natural language processing
US10970394B2 (en) 2017-11-21 2021-04-06 Biocatch Ltd. System, device, and method of detecting vishing attacks
US10636424B2 (en) 2017-11-30 2020-04-28 Apple Inc. Multi-turn canned dialog
US10733982B2 (en) 2018-01-08 2020-08-04 Apple Inc. Multi-directional dialog
US10733375B2 (en) 2018-01-31 2020-08-04 Apple Inc. Knowledge-based framework for improving natural language understanding
US10789959B2 (en) 2018-03-02 2020-09-29 Apple Inc. Training speaker recognition models for digital assistants
US10592604B2 (en) 2018-03-12 2020-03-17 Apple Inc. Inverse text normalization for automatic speech recognition
US11710482B2 (en) 2018-03-26 2023-07-25 Apple Inc. Natural assistant interaction
US10818288B2 (en) 2018-03-26 2020-10-27 Apple Inc. Natural assistant interaction
US10909331B2 (en) 2018-03-30 2021-02-02 Apple Inc. Implicit identification of translation payload with neural machine translation
US11145294B2 (en) 2018-05-07 2021-10-12 Apple Inc. Intelligent automated assistant for delivering content from user experiences
US10928918B2 (en) 2018-05-07 2021-02-23 Apple Inc. Raise to speak
US11169616B2 (en) 2018-05-07 2021-11-09 Apple Inc. Raise to speak
US11854539B2 (en) 2018-05-07 2023-12-26 Apple Inc. Intelligent automated assistant for delivering content from user experiences
US10984780B2 (en) 2018-05-21 2021-04-20 Apple Inc. Global semantic word embeddings using bi-directional recurrent neural networks
US11431642B2 (en) 2018-06-01 2022-08-30 Apple Inc. Variable latency device coordination
US10892996B2 (en) 2018-06-01 2021-01-12 Apple Inc. Variable latency device coordination
US11386266B2 (en) 2018-06-01 2022-07-12 Apple Inc. Text correction
US10684703B2 (en) 2018-06-01 2020-06-16 Apple Inc. Attention aware virtual assistant dismissal
US10984798B2 (en) 2018-06-01 2021-04-20 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
US10403283B1 (en) 2018-06-01 2019-09-03 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
US11009970B2 (en) 2018-06-01 2021-05-18 Apple Inc. Attention aware virtual assistant dismissal
US11495218B2 (en) 2018-06-01 2022-11-08 Apple Inc. Virtual assistant operation in multi-device environments
US10720160B2 (en) 2018-06-01 2020-07-21 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
US10944859B2 (en) 2018-06-03 2021-03-09 Apple Inc. Accelerated task performance
US10504518B1 (en) 2018-06-03 2019-12-10 Apple Inc. Accelerated task performance
US10496705B1 (en) 2018-06-03 2019-12-03 Apple Inc. Accelerated task performance
US11010561B2 (en) 2018-09-27 2021-05-18 Apple Inc. Sentiment prediction from textual data
US11170166B2 (en) 2018-09-28 2021-11-09 Apple Inc. Neural typographical error modeling via generative adversarial networks
US10839159B2 (en) 2018-09-28 2020-11-17 Apple Inc. Named entity normalization in a spoken dialog system
US11462215B2 (en) 2018-09-28 2022-10-04 Apple Inc. Multi-modal inputs for voice commands
US11475898B2 (en) 2018-10-26 2022-10-18 Apple Inc. Low-latency multi-speaker speech recognition
US11638059B2 (en) 2019-01-04 2023-04-25 Apple Inc. Content playback on multiple devices
US11348573B2 (en) 2019-03-18 2022-05-31 Apple Inc. Multimodality in digital assistant systems
US11475884B2 (en) 2019-05-06 2022-10-18 Apple Inc. Reducing digital assistant latency when a language is incorrectly determined
US11217251B2 (en) 2019-05-06 2022-01-04 Apple Inc. Spoken notifications
US11307752B2 (en) 2019-05-06 2022-04-19 Apple Inc. User configurable task triggers
US11423908B2 (en) 2019-05-06 2022-08-23 Apple Inc. Interpreting spoken requests
US11140099B2 (en) 2019-05-21 2021-10-05 Apple Inc. Providing message response suggestions
US11289073B2 (en) 2019-05-31 2022-03-29 Apple Inc. Device text to speech
US11360739B2 (en) 2019-05-31 2022-06-14 Apple Inc. User activity shortcut suggestions
US11657813B2 (en) 2019-05-31 2023-05-23 Apple Inc. Voice identification in digital assistant systems
US11237797B2 (en) 2019-05-31 2022-02-01 Apple Inc. User activity shortcut suggestions
US11496600B2 (en) 2019-05-31 2022-11-08 Apple Inc. Remote execution of machine-learned models
US11360641B2 (en) 2019-06-01 2022-06-14 Apple Inc. Increasing the relevance of new available information
US11488406B2 (en) 2019-09-25 2022-11-01 Apple Inc. Text detection using global geometry estimators
US11181988B1 (en) 2020-08-31 2021-11-23 Apple Inc. Incorporating user feedback into text prediction models via joint reward planning
US11606353B2 (en) 2021-07-22 2023-03-14 Biocatch Ltd. System, device, and method of generating and utilizing one-time passwords

Also Published As

Publication number Publication date
WO2009006221A1 (en) 2009-01-08
US20090007001A1 (en) 2009-01-01
JP5323070B2 (en) 2013-10-23
EP2160672A4 (en) 2011-07-27
EP2160672A1 (en) 2010-03-10
JP2010536076A (en) 2010-11-25
US8065624B2 (en) 2011-11-22

Similar Documents

Publication Publication Date Title
US8065624B2 (en) Virtual keypad systems and methods
US20220261112A1 (en) Systems, devices, and methods for touch-free typing
US10359932B2 (en) Method and apparatus for providing character input interface
US10191573B2 (en) Pointer display device, pointer display/detection method, pointer display/detection program and information apparatus
US20160364138A1 (en) Front touchscreen and back touchpad operated user interface employing semi-persistent button groups
US20100259482A1 (en) Keyboard gesturing
US20170017393A1 (en) Method for controlling interactive objects from a touchpad of a computerized device
US20050162402A1 (en) Methods of interacting with a computer using a finger(s) touch sensing input device with visual feedback
US20110209087A1 (en) Method and device for controlling an inputting data
KR20130088752A (en) Multidirectional button, key, and keyboard
CN104360813A (en) Display equipment and information processing method thereof
US20110126100A1 (en) Method of providing gui for guiding start position of user operation and digital device using the same
KR20160097410A (en) Method of providing touchless input interface based on gesture recognition and the apparatus applied thereto
CN111007977A (en) Intelligent virtual interaction method and device
Benko et al. Imprecision, inaccuracy, and frustration: The tale of touch input
WO2016115976A1 (en) Smart wearable input apparatus
JP2016076232A (en) Display device and control method for the same
CN105700806A (en) Directional gesture input method and device for intelligent televisions or intelligent glasses
WO2021178255A1 (en) Electronic input system
CN114040857A (en) Method for operating an operating system in a vehicle and operating system in a vehicle
Ni A framework of freehand gesture interaction: techniques, guidelines, and applications
US20150106764A1 (en) Enhanced Input Selection
WO2015013662A1 (en) Method for controlling a virtual keyboard from a touchpad of a computerized device
Lee et al. Embodied interaction on constrained interfaces for augmented reality
WO2018035353A1 (en) Front touchscreen and back touchpad operated user interface employing semi-persistent button groups

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION