US20100164897A1 - Virtual keypad systems and methods - Google Patents
Virtual keypad systems and methods Download PDFInfo
- Publication number
- US20100164897A1 US20100164897A1 US12/666,916 US66691608A US2010164897A1 US 20100164897 A1 US20100164897 A1 US 20100164897A1 US 66691608 A US66691608 A US 66691608A US 2010164897 A1 US2010164897 A1 US 2010164897A1
- Authority
- US
- United States
- Prior art keywords
- keys
- touchpad
- keypad
- key
- zone
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/02—Input arrangements using manually operated switches, e.g. using keyboards or dials
- G06F3/023—Arrangements for converting discrete items of information into a coded form, e.g. arrangements for interpreting keyboard generated codes as alphanumeric codes, operand codes or instruction codes
- G06F3/0233—Character input methods
- G06F3/0237—Character input methods using prediction or retrieval techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/1698—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being a sending/receiving arrangement to establish a cordless communication link, e.g. radio or infrared link, integrated cellular phone
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0346—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04886—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42204—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42204—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
- H04N21/42206—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor characterized by hardware details
- H04N21/42224—Touch pad or touch panel provided on the remote control
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42204—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
- H04N21/42226—Reprogrammable remote control devices
- H04N21/42227—Reprogrammable remote control devices the keys being reprogrammable, e.g. soft keys
- H04N21/42228—Reprogrammable remote control devices the keys being reprogrammable, e.g. soft keys the reprogrammable keys being displayed on a display screen in order to reduce the number of keys on the remote control device itself
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42204—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
- H04N21/42206—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor characterized by hardware details
- H04N21/42222—Additional components integrated in the remote control device, e.g. timer, speaker, sensors for detecting position, direction or movement of the remote control, microphone or battery charging device
Definitions
- the present invention relates to methods and systems for recognizing text input from a remote controller.
- remote control devices can communicate a variety of commands to the electronic product.
- the remote control devices have become complex to operate.
- modern television remote controls can include selection buttons for volume, channel selection, menu selection, and picture viewing.
- selection buttons for volume, channel selection, menu selection, and picture viewing.
- the user To operate the remote control, the user must take time away from the program he or she is watching to focus in on the buttons. This can be very distracting to a viewer.
- a virtual keypad system for inputting text includes a remote controller having at least one touchpad incorporated therein and divided into a plurality of touch zones.
- a display device is in data communication with the remote controller and is operable to display a user interface including a keypad, where each key of the keypad is mapped to a touch zone of the touchpad.
- a prediction module in response to an operator pressing a given touch zone to select a particular character, performs one or more key prediction methods to predict one or more next plausible keys.
- a key mapping module remaps the touch zones of the touchpad to the keys of the keypad based on the one or more next plausible keys.
- FIG. 1 is an illustration of a text input system according to various aspects of the present disclosure.
- FIG. 2A is an illustration of a remote controller of the text input system of FIG. 1 that includes a touchpad according to various aspects of the present disclosure.
- FIG. 2B is a block diagram illustrating internal components of the remote controller of FIG. 2A according to various aspects of the present disclosure.
- FIG. 3 is a dataflow diagram illustrating a virtual keypad module of the text input system of FIG. 1 according to various aspects of the present disclosure.
- FIG. 4A is an illustration of the remote controller being held in a portrait position according to various aspects of the present disclosure.
- FIG. 4B is an illustration of the remote controller being held in a landscape position according to various aspects of the present disclosure.
- FIG. 5 is a table illustrating a mapping between XY coordinates of the touchpad and keys of a keypad.
- FIG. 6 is a table illustrating a remapping between the coordinates of the touchpad and the keys of a keypad.
- FIG. 7 is an illustration of a first embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
- FIG. 8A is an illustration of a second embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
- FIG. 8B is an illustration of a third embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
- FIG. 9 is an illustration of a fourth embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
- FIG. 10 is an illustration of a fifth embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure.
- module or sub-module can refer to a processor (shared, dedicated, or group) and memory that executes one or more software or firmware to programs, and/or other suitable components that can provide the described functionality and/or combinations thereof.
- FIG. 1 depicts an exemplary text input system 10 implemented according to various aspects of the present disclosure.
- the exemplary text input system 10 includes a virtual keypad module 12 that facilitates the input of alphanumeric characters by a user for interacting with various services delivered through a display device 14 .
- the display device 14 can be, but is not limited to, a television (as shown), a projector and screen, or a computer.
- the services can be, for example, internet based services such as, online shopping and movie subscriptions.
- the virtual keypad module 12 provides feedback to the user via a graphical user interface (GUI) 18 .
- the GUI 18 includes a virtual keypad 20 , as will be discussed in more detail below.
- the virtual keypad module 12 can be implemented within the display device 14 . In various other embodiments, the virtual keypad module 12 can be implemented separate from the display device 14 (such as, for example, on a set top box (not shown)) and can be in data communication with the display device 14 . For ease of the discussion, the remainder of the disclosure will be discussed in the context of the virtual keypad module 12 being implemented within the display device 14 .
- the text input system 10 further includes a remote controller 16 that generates one or more signals to the display device 14 in response to user input.
- the virtual keypad module 12 receives and processes the signals. Based on the signals, the virtual keypad module 12 determines an orientation and a holding position of the remote controller 16 , recognizes text input, and/or provides visual feedback to the user via a graphical user interface (GUI) 18 .
- GUI graphical user interface
- the virtual keypad module 12 implements selection auto-correction methods that compensate for human typing (i.e., clicking) error. For example, when attempting to input text quickly, users can typically undershoot or overshoot the location and click on a nearby unintended key.
- the virtual keypad module 12 employs a combination of prediction and auto-correction methods to determine which character(s) is/are most likely to be entered by the user.
- a prediction method is used to compensate for the possible overshoot and undershoot.
- the predictions can be used to enlarge an activation area of possible keys while reducing (or zero-ing) activation areas of keys that are not in the next-character prediction list.
- the virtual keypad module 12 disables the selection auto-correction methods and reverts to a default mode (i.e., without enlarging or reducing the activation area).
- the virtual keypad module 12 can also disable the selection auto-correction method when the interaction becomes slow because it is assumed that clicking errors do not generally occur during slow interaction.
- FIGS. 2A and 2B illustrate an exemplary remote controller 16 according to various aspects of the present disclosure.
- the exterior of the remote controller 16 includes a touchpad 22 and one or more soft keys 24 a - 24 d.
- touch zones defined by one or more coordinates of the touchpad 22 can be mapped to a particular key of the virtual keypad 20 ( FIG. 1 ).
- a user can select a particular key of the virtual keypad 20 by gently placing his finger or thumb on the touchpad 22 at or near the associated touch zone (FingerDown event), by dragging a finger or thumb along the touchpad 22 to the associated touch zone (FingerDrag event), and/or by lifting the finger or thumb away from the touchpad 22 (FingerUp event). While the user has a finger or thumb on the touchpad 22 (i.e., between FingerDown and FingerUp events), the user can click on the touchpad 22 by applying greater force (FingerPress event) followed by releasing the force (FingerRelease event) to select a key.
- a relative access method can be used as an alternative or as a secondary method for selecting keys.
- the relative access method assumes a position of the user's finger or thumb to be a current coordinate or touch zone of the touchpad 22 . Subsequent gestures by the user are then interpreted relative to that coordinate or touch zone. This allows for an adjustable precision in selection.
- Important functions of the remote controller 16 can be associated with specific selection buttons 26 a - 26 d of the touchpad 22 .
- the selection buttons 26 a - 26 d can be designated by a specific touchpad button that is painted or illuminated on the touchpad 22 (as shown) or by a button displayed on an overlay to the virtual keypad 20 ( FIG. 1 ) of the GUI 18 ( FIG. 1 ). This allows the user to use the remote controller 16 in complete darkness without having to look away from the content being displayed on the display device 14 ( FIG. 1 ).
- the functions can be controlled by simply touching the buttons or be controlled by performing a specific gesture.
- sliding a finger or thumb up or down on the right side of the touchpad 22 can trigger a volume up or volume down action.
- sliding a finger or thumb right or left on the top side of the touchpad 22 can trigger a channel up or channel down action.
- the body of the remote controller 16 can be made of a soft material, allowing the remote controller 16 to be squeezed.
- the squeezing of the remote controller 16 can be performed by the user to trigger certain actions, particularly in contexts where the GUI 18 ( FIG. 1 ) is just waiting for an acknowledgement without proposing a choice (such as a “next” button in a slideshow).
- the internal components of the remote controller 16 can include, but are not limited to, input sensors 30 , output actuators 32 , an input controller 34 , an output controller 36 , a processing handler 38 , a wireless transmitter (e.g., RF, Bluetooth, etc.) 40 , and/or combinations thereof.
- the inputs sensors 30 can include touchpad sensors 42 .
- the touchpad sensors 42 can be single-position registering touchpad sensors mounted side-by-side that allow for the selection of at least two contact points on the touchpad 22 ( FIG. 2A ) simultaneously.
- the touchpad sensors 42 can be a single multi-touch capable touchpad sensor that can register, with equal precision, two points of contact at the same time.
- the touchpad sensors 42 can register pressure information to allow the touchpad 22 ( FIG. 2A ) to be clickable.
- the input sensors 30 can also include one or more selection button sensors 44 , one or more touchpad button sensors 46 , one or more accelerometers 48 , and one or more holding sensors 50 .
- the holding sensors 50 can be, for example, capacitive sensors that are located around the border of the remote controller 16 , and/or behind the remote controller 16 .
- the holding sensors 50 indicate whether the user is touching an area of the remote controller 16 in a proximity of the holding sensor 50 .
- the accelerometer 48 can be a three-axis accelerometer that indicates a positioning of the remote controller 16 .
- the input controller 34 reads the real-time data from all active sensors. In various embodiments, some sensors may not be active at all times to reduce power consumption.
- the processing handler 38 gathers and forms into packets the data to be transmitted and/or processes the real-time data from one or more active sensors to perform local actions.
- the RF transmitter 40 (RF driver 40 ) generates the signals in packet form to the display device 14 ( FIG. 1 ).
- the output actuators 32 can include one or more LED panels 52 for displaying the touchpad buttons 26 a - 26 d, depending on the specific state of interaction with the GUI 18 present on-screen.
- the output actuators 32 can additionally or alternatively include actuators for providing sufficient haptic feedback to the user (such as, for example, vibration actuators 54 , light actuators 55 , and/or speaker actuators 56 ).
- the above has described one example of the operations performed by the output actuator 32 .
- the output controller 36 updates the state of all the active actuators.
- FIG. 3 a dataflow diagram illustrates a more detailed exemplary virtual keypad module 12 .
- the virtual keypad module 12 may include any number of sub-modules. As can be appreciated, the sub-modules shown in FIG. 3 may be combined and/or further partitioned to similarly perform text input.
- the data inputs 70 , 72 , 74 , and 76 to the virtual keypad module 12 are received from the remote controller 16 ( FIG. 1 ) and/or received from other modules (not shown) within the display device 14 ( FIG. 1 ).
- the virtual keypad module 12 includes an orientation recognition module 60 , a hand position recognition module 62 , a prediction module 64 , a key input module 66 , a key mapping module 67 , and a GUI manager module 68 .
- the orientation recognition module 60 determines an orientation of the remote controller 16 based on data received from the holding sensors 50 ( FIG. 2B ) and the accelerometer 48 ( FIG. 2B ). For example, the user can be holding the remote controller 16 ( FIG. 2A ) in a portrait position, as shown in FIG. 4A , or in a landscape position, as shown in FIG. 4B .
- the orientation recognition module 60 determines the orientation by way of an Artificial Neural Network (ANN).
- the ANN can be trained by data indicating both landscape position conditions and portrait position conditions.
- the orientation is determined by training an ANN with sensory data.
- the sensory data can comprise three-dimensional acceleration (accx, accy, accz) and an activation state of the n capacitive holding position sensors, which can signal that human skin is ever in proximity (1) or is not in proximity (0). These n+3 values are fed into a single perceptron or linear classifier to determine if the remote controller 16 ( FIG. 2A ) is horizontal or vertical. Perceptron coefficients can be trained on a database and hard-coded by a manufacturer.
- the hand position recognition module 62 determines a holding style of the remote controller 16 ( FIG. 2A ) based on data received from the holding sensors 50 ( FIG. 2B ) and the accelerometer 48 ( FIG. 2B ).
- the sensory data 70 , 72 can be used to determine whether the remote controller 16 ( FIG. 2A ) is held with one or two hands; and if it is held with one hand, whether it is held with the left or right hand.
- the hand position recognition module 62 determines the holding style by way of an ANN.
- the ANN can be trained by data indicating right-hand conditions, left-hand conditions, and two hands conditions.
- the hand position is determined similarly as discussed above.
- Multiple perceptrons can be implemented for the more than one binary decisions (e.g., left hand, right hand, two handed).
- the determination of the orientation and the holding style gives the virtual keypad module 12 the ability to accommodate the user by automatically adapting the text input methods and the look and feel of the GUI 18 ( FIG. 1 ).
- the determination of the orientation and holding position allows the user to hold the remote controller 16 ( FIG. 2A ) in the most convenient way based on their personal preference and the actual conditions of use (e.g., standing, sitting, lying down).
- the user can pick up and operate the remote controller 16 ( FIG. 2A ) without worrying about how they are holding it.
- the hand position recognition module 62 can further perform user verification based on a combination of holding sensor data 70 , accelerometer data 72 , additional sensor information (such as an image of the palm of the user's hand), and/or bio-sensors.
- the data can be used to fully determine the identity of the user or, more broadly, infer the category to which the user belongs (e.g., left-handed, right-handed, kid, adult, elderly).
- User identification can be used, for example, for parental control, personalization, and profile switching.
- User categorization can be used to adapt the GUI 18 ( FIG. 1 ).
- the key mapping module 67 generates a map indicating an association between the coordinates or touch zones of the touchpad 22 ( FIG. 2A ) and the keys of the keypad and/or touchpad selection buttons. In various embodiments, the key mapping module 67 generates the map based on the orientation and hand position information determined from the orientation recognition module 60 and the hand position recognition module 62 , respectively. The key mapping module 67 maps the touch zones of the touchpad 22 to the keys of the keypad, by associating a key of the keypad to each XY coordinate of the touch zones. Here, the touch zones are defined by one or more XY coordinates of the touchpad 22 . The key mapping module 67 generates a map from the mapping.
- the map is a two-dimensional lookup table defined by the coordinates of the touchpad 22 .
- the map can be a two dimensional (XY) table 80 that is used to assign a key of the keypad and/or a touchpad selection button to each coordinate of the touchpad 22 ( FIG. 2A ).
- the map can then be referenced by the key input module 66 to determine an action to be taken and can be referenced by the GUI manager module 68 to generate the GUI 18 ( FIG. 1 ).
- the key input module 66 processes touchpad sensor data 74 and/or the accelerometer data 72 .
- the key input module 66 interprets the touchpad sensor data 74 to be a coordinate or coordinates of the touchpad 22 ( FIG. 2A ) and determines what action to be taken based on the coordinate or coordinates.
- the key input module 66 can receive the touchpad sensor data 74 , determine a particular coordinate from the data 74 , and reference the map generated by the key mapping module 67 . Based on the entries in the map, the key input module 66 , for example, can project that the user is hovering over a particular key of the keypad and, thus, entering a particular text.
- the key input module 66 interprets the accelerometer data 72 as an action to be taken.
- the accelerometer data 72 can indicate if a user has lifted the remote controller 16 ( FIG. 2A ) quickly to select, for example, an uppercase mode.
- the accelerometer data 72 can indicate when a user has lowered the remote controller 16 ( FIG. 2A ) quickly to select, for example, a lowercase mode.
- the prediction module 64 To enhance the precision and speed at which the text is entered, the prediction module 64 generates a prediction of which key and/or word the user is trying to select. The prediction module 64 generates the prediction based on the touchpad sensor data 74 and/or based on a determination of previous text entered. In various embodiments, the prediction module 64 performs one or more next key prediction methods, such as, for example, a language model method, a trajectory analysis method, a hand movement analysis method, a timing analysis method, and/or combinations thereof. In short, the prediction module 64 , in response to an operator pressing a given touch zone to select a particular character, performs one or more key prediction methods to predict one or more next plausible keys. Examples of the operator are a finger of the user, a touch-pen, and the like, that operates the touchpad 22 .
- the prediction module 64 employs one or more language models known in the art to predict the next key based on previous text entered.
- the language models for instance, predict the one or more next plausible keys based on previous characters selected by the operator. For example, if the partial word ‘pr’ has been entered, the language model can predict that a vowel is likely to follow and that the letter ‘r’ will not be a possibility.
- the prediction module 64 employs one or more language models to provide a list of reliable candidates of full words from partial word inputs.
- the language model generates a list of plausible words based on the previous characters selected by the operator.
- the full words can be selected by the user for auto-completion. For example, if the partial word ‘Pan’ has been entered, a list can be generated that includes ‘Panasonic’ and ‘Pan-American.’ Instead of typing the remaining characters, the user can simply select one of the full words.
- the language model can generate the word predictions based on words previously entered. For example, once selected, the words can be remembered and the language model can be adapted to favor the remembered words.
- the trajectory analysis method can be performed to predict possible next keys based on future path estimation including directions and/or velocities of user's finger or thumb movement on the touchpad 22 ( FIG. 2A ) as indicated by the touchpad sensor data 74 .
- the trajectory analysis method predicts next plausible keys based on a direction of operator movement on the touchpad 22 . For example, if the user first selects the ‘k’ key and the language model predicts that the next key can be one of ‘I,’ ‘e,’ or ‘a,’ the touchpad sensor data 74 can be evaluated to determine a direction the user is heading and velocity of the movement and, thus, eliminate one or more of the choices.
- the trajectory analysis method determines a coordinate of the key ‘k’ and subsequent finger movements. From that history of XY coordinates, the future path is determined. The path includes a tolerance to account for short-term prediction (more accurate) and longer-term prediction (less accurate). If the future path estimation is heading away from the coordinates of the predicted key, the choice is eliminated. For example, if the path is heading on an Up/Left diagonal line, then the keys ‘e’ and ‘a’ are eliminated and the key ‘I’ is selected as the predicted next key.
- the hand movement analysis method can be performed using the holding style information provided by the hand position recognition module 62 and the predictions provided by the prediction module 64 .
- the hand movement analysis method predicts next plausible keys based on a detection of which hand is moving on the touchpad 22 .
- the hand movement analysis method can evaluate whether the remote controller 16 ( FIG. 2A ) is held with two hands or one hand. If the remote controller 16 ( FIG. 2A ) is held by two hands and movement by a right hand or left hand is detected, then the choices that are associated with the non-moving hand would be eliminated.
- the prediction module 64 predicts that the next key can be one of ‘e,’ or ‘a,’ and movement is detected by the right hand, then the keys ‘e’ and ‘a’ are eliminated and the key ‘I’ is selected as the predicted next key.
- the timing analysis method can be performed when the prediction module 64 is unable to predict the appropriate next key or word via any of the next key prediction methods. Such may be the case when the user is entering proper nouns, such as a last name or a password.
- the timing analysis method evaluates the time the user takes to move from one key to the next. That is, the timing analysis method predicts next plausible keys based on a timing of operator movement on the touchpad 22 . In more detail, the timing analysis method predicts no next plausible keys when the timing exceeds a predetermined limit.
- the key mapping module 67 remaps the touch zones of the touchpad 22 to the keys of the keypad based on the no next plausible keys. If the user moves more slowly, it is more likely that a proper noun is being entered. The predictions are then ignored.
- the key mapping module 67 can remap the coordinates of the touchpad 22 ( FIG. 2A ) to the keys of the virtual keypad 20 ( FIG. 1 ). That is, the key mapping module 67 remaps the touch zones of the touchpad 22 to the keys of the keypad based on the one or more next plausible keys predicted by the prediction module 64 . In various embodiments, the coordinates that are associated with the predicted next key can be expanded to make the key more accessible. For example, as shown in FIG.
- the map can be adjusted such that the coordinates that were previously mapped to the keys ‘o’ or ‘I’ are now mapped to the key ‘p.’
- the ‘p’ key will be selected and entered if the user clicks on the touchpad 22 ( FIG. 2A ).
- the key mapping module 67 can remap the coordinates to the keys based on a relative speed of the user movement. For example, the key mapping module 67 can remap the coordinates such that the predicted key is mapped to a larger touch zone area when a faster movement is detected. The key mapping module 67 can maintain the original coordinates or remap to the original coordinates when slower movements are detected. The key mapping module 67 can scale the area between the larger area and the original area when a movement that is not fast or slow is detected.
- the GUI manager module 68 generates GUI data 76 for displaying the GUI 18 .
- the GUI 18 provides visual feedback to the user indicating the actions they have performed with the remote controller 16 ( FIG. 2A ).
- the GUI 18 can include the virtual keypad 20 , including multiple alphanumeric keys 90 , one or more selection buttons 92 , a selection list 94 , a text display box 96 , a title bar 98 , and/or combinations thereof.
- the touchpad 22 is divided into two operating zones (a first operating zone and a second operating zone).
- the key mapping module 67 maps a first subset of keys of the keypad to the first operating zone of the touchpad 22 and maps a second subset of keys of the keypad to the second operating zone of the touchpad 22 .
- the first operating zone corresponds to a top zone or a right zone of the touchpad 22
- the second operating zone corresponds to a bottom zone or a left zone of the touchpad 22
- each of the first subset and the second subset is a group of keys of the keypad.
- the GUI manager module 68 displays the virtual keypad 20 based on the holding position and the orientation determined by the hand position recognition module 62 and the orientation recognition module 60 , respectively. For example, as shown in FIG. 7 , if two hands are used to hold the remote controller 16 ( FIG. 2A ) in the landscape position, the virtual keypad 20 and the touchpad 22 can be divided into two zones 100 , 102 (e.g., left and right). The keys in the right zone 102 can be associated with a first zone 104 of the touchpad 22 , and the keys in the left zone 100 can be associated with a second zone 106 of the touchpad 22 .
- the keys in the right zone 102 can be associated with a first zone 104 of the touchpad 22
- the keys in the left zone 100 can be associated with a second zone 106 of the touchpad 22 .
- the user can select a key 90 in the right zone 102 with a right thumb or finger, and the user can select a key 90 in the left zone 100 with a left thumb or finger.
- the keys 90 in each zone 100 , 102 can be distinguished by a particular color or shading of the keys 90 so that the user can determine which keys 90 can be selected with which thumb or finger. This will allow for a natural text input, similar to the experience when entering text on classic keyboards.
- selection of a predicted word can be made through the display of the two most probable words (e.g., ‘Panasonic,’ ‘Pan-America’).
- the selection buttons 92 are auto-completion selection buttons. When the user selects one of the selection buttons 92 , a probable word displayed in the selected selection button 92 is displayed as a complete word on the display box 96 .
- the two words can be displayed on selection buttons 92 .
- the user can select the selection buttons 92 by pushing soft keys 24 b, 24 d located on the top side of the remote controller 16 with the index fingers, or by dragging the finger or thumb to a dedicated zone located at a designated location of the touchpad 22 .
- the mapping of the coordinates can provide for an overlap between the two areas.
- the key mapping module 67 maps a third subset of keys of the first subset of keys to the second operating zone and maps a fourth subset of keys of the second subset of keys to the first operating zone.
- the third subset of keys include the keys designated by the letters ‘t,’ ‘g,’ and ‘b’ of the keys of the first subset (keys included in the zone 100 of FIG. 7 ).
- the fourth subset of keys include the keys designated by the letters ‘y,’ ‘h,’ and ‘n’ of the keys of the second subset (keys included in the zone 102 of FIG. 7 ).
- the keys along a boundary between the divided zones can operate for any zones of the touchpad 22 .
- the letter ‘g’ in left keyboard area can be selected by the touchpad 22 in the first zone, as well as be selected by the touchpad 22 in the second zone 106 .
- the overlap keys can be identified on the GUI 18 by shading or color.
- the touchpad 22 can be divided into two zones 108 , 110 or 112 , 114 .
- the virtual keypad 20 is not divided.
- a first zone 110 or 114 of the touchpad 22 can be associated with the entire virtual keypad 20 , and can be referred to as the “Key-Entry Zone.”
- the second zone 108 or 112 of the touchpad 22 can be associated with the auto-completion buttons 92 ( FIG.
- selection lists 94 auto-completion selection list
- the auto-completion buttons 92 or selection lists 94 can be displayed when the user touches the touchpad 22 in the “Auto-Completion Zone.”
- the user interface can include a small visual notification to signal the availability and status of auto-completion words. This way the user will not be bothered with auto-completion unless he decides to use it.
- the zones 108 , 110 or 112 , 114 of the touchpad 22 can be configured based on an identification of the user. For example, in the case of a right-handed user, the right zone 110 can be associated with the “Key-Entry Zone” used most often and the left zone 108 can be associated with the “Auto-Completion Zone.” Similarly, for a left-handed user, the left zone 108 can be associated with the “Key-Entry Zone” used most often and the right zone 110 can be associated with the “Auto-Completion Zone.”
- the touchpad 22 is divided into zones.
- the virtual keypad 20 is not divided into zones. Only one zone of the touchpad 22 is used by the user at all times. The user will be able to address the entire virtual keypad 20 from the entire touchpad 22 .
- Such touchpad 22 will principally function as the “Key-Entry Zone.” If auto-completion is needed, the user can switch the touchpad 22 to the “Auto-completion Zone” by using a simple gesture.
- Such gesture can include, but is not limited to, moving the thumb or finger to a specific area of the touchpad 22 (for instance lower right), or sliding the finger or thumb along the touchpad 22 quickly from right to left.
- the user can select a key 90 by dragging the thumb or finger on the touchpad 22 .
- one or more on-screen pointers (such as, for example, a cursor, or an illustration of a thumb or finger) slides to a target key 90 .
- the on-screen pointers can be displayed according to the hand position (e.g., left hand only, right hand only, or both hands).
- the key 90 can be selected by clicking the clickable touchpad 22 and/or upon release.
- a thumb or finger posture can be used to indicate a thumb or finger press as opposed to a thumb or finger that is dragged on the touchpad surface.
- the selected character associated with the key will be displayed in the text display box 96 .
- the GUI manager module 68 displays the keys 90 of the virtual keypad 20 based on the predicted next key and the remapping of the coordinates performed by the key mapping module 67 ( FIG. 3 ). For example, by knowing the next likely key, the GUI manager module 68 ( FIG. 3 ) can highlight and/or enlarge the most likely key 116 based on the mapping of the coordinates. That is, an arrangement of the keypad on the user interface is modified based on the one or more next plausible keys predicted by the prediction module 64 . In more detail, one or more keys of the keypad on the user interface is highlighted or enlarged based on the one or more next plausible keys predicted by the prediction module 64 .
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Hardware Design (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Input From Keyboards Or The Like (AREA)
- Position Input By Displaying (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Accordingly, a virtual keypad system for inputting text is provided. A virtual keypad system includes a remote controller having at least one touchpad incorporated therein and divided into a plurality of touch zones. A display device is in data communication with the remote controller and is operable to display a user interface including a keypad, where each key of the keypad is mapped to a touch zone of the touchpad. A prediction module, in response to an operator pressing a given touch zone to select a particular character, performs one or more key prediction methods to predict one or more next plausible keys. A key mapping module remaps the touch zones of the touchpad to the keys of the keypad based on the one or more next plausible keys.
Description
- This application claims the benefit of U.S. patent applications Nos. 60/946858 and 11/977346, filed on Jun. 28, 2007 and Oct. 24, 2007, respectively. The disclosures of the above applications are incorporated herein by reference.
- The present invention relates to methods and systems for recognizing text input from a remote controller.
- Many electronic consumer products come with remote control devices. These remote control devices can communicate a variety of commands to the electronic product. With the rise in technological advancements to the electronic products, the remote control devices have become complex to operate. For example, modern television remote controls can include selection buttons for volume, channel selection, menu selection, and picture viewing. To operate the remote control, the user must take time away from the program he or she is watching to focus in on the buttons. This can be very distracting to a viewer.
- In addition, many Internet based services such as online shopping are now being provided through the television. Additional selection buttons such as keypad buttons must be included on the remote control device to accommodate these new services. The new selection buttons serve to increase the complexity as well as the cost of the remote control devices. Various solutions have been proposed to address such problems. One solution is disclosed in U.S. Pat. No. 6,765,557 to use a touchpad for controlling a home entertainment device such as an Interactive television. However, even this solution cannot completely solve the problems of inconveniency to the user inconvenience.
- The statements in this section merely provide background information related to the present disclosure and may not constitute prior art.
- A virtual keypad system for inputting text is provided. A virtual keypad system includes a remote controller having at least one touchpad incorporated therein and divided into a plurality of touch zones. A display device is in data communication with the remote controller and is operable to display a user interface including a keypad, where each key of the keypad is mapped to a touch zone of the touchpad. A prediction module, in response to an operator pressing a given touch zone to select a particular character, performs one or more key prediction methods to predict one or more next plausible keys. A key mapping module remaps the touch zones of the touchpad to the keys of the keypad based on the one or more next plausible keys.
- Further areas of applicability will become apparent from the description provided herein. It should be understood that the description and specific examples are intended for purposes of illustration only and are not intended to limit the scope of the present disclosure.
- The drawings described herein are for illustration purposes only and are not intended to limit the scope of the present teachings in any way.
-
FIG. 1 is an illustration of a text input system according to various aspects of the present disclosure. -
FIG. 2A is an illustration of a remote controller of the text input system ofFIG. 1 that includes a touchpad according to various aspects of the present disclosure. -
FIG. 2B is a block diagram illustrating internal components of the remote controller ofFIG. 2A according to various aspects of the present disclosure. -
FIG. 3 is a dataflow diagram illustrating a virtual keypad module of the text input system ofFIG. 1 according to various aspects of the present disclosure. -
FIG. 4A is an illustration of the remote controller being held in a portrait position according to various aspects of the present disclosure. -
FIG. 4B is an illustration of the remote controller being held in a landscape position according to various aspects of the present disclosure. -
FIG. 5 is a table illustrating a mapping between XY coordinates of the touchpad and keys of a keypad. -
FIG. 6 is a table illustrating a remapping between the coordinates of the touchpad and the keys of a keypad. -
FIG. 7 is an illustration of a first embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure. -
FIG. 8A is an illustration of a second embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure. -
FIG. 8B is an illustration of a third embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure. -
FIG. 9 is an illustration of a fourth embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure. -
FIG. 10 is an illustration of a fifth embodiment of a virtual keypad graphical user interface according to various aspects of the present disclosure. - The following description is merely exemplary in nature and is not intended to limit the present teachings, their application, or uses. It should be understood that throughout the drawings, corresponding reference numerals indicate like or corresponding parts and features. As used herein, the term module or sub-module can refer to a processor (shared, dedicated, or group) and memory that executes one or more software or firmware to programs, and/or other suitable components that can provide the described functionality and/or combinations thereof.
- Referring now to
FIG. 1 ,FIG. 1 depicts an exemplarytext input system 10 implemented according to various aspects of the present disclosure. The exemplarytext input system 10 includes avirtual keypad module 12 that facilitates the input of alphanumeric characters by a user for interacting with various services delivered through adisplay device 14. Thedisplay device 14 can be, but is not limited to, a television (as shown), a projector and screen, or a computer. The services can be, for example, internet based services such as, online shopping and movie subscriptions. Thevirtual keypad module 12 provides feedback to the user via a graphical user interface (GUI) 18. TheGUI 18 includes avirtual keypad 20, as will be discussed in more detail below. - In various embodiments, the
virtual keypad module 12 can be implemented within thedisplay device 14. In various other embodiments, thevirtual keypad module 12 can be implemented separate from the display device 14 (such as, for example, on a set top box (not shown)) and can be in data communication with thedisplay device 14. For ease of the discussion, the remainder of the disclosure will be discussed in the context of thevirtual keypad module 12 being implemented within thedisplay device 14. - The
text input system 10 further includes aremote controller 16 that generates one or more signals to thedisplay device 14 in response to user input. Thevirtual keypad module 12 receives and processes the signals. Based on the signals, thevirtual keypad module 12 determines an orientation and a holding position of theremote controller 16, recognizes text input, and/or provides visual feedback to the user via a graphical user interface (GUI) 18. In particular, thevirtual keypad module 12 implements selection auto-correction methods that compensate for human typing (i.e., clicking) error. For example, when attempting to input text quickly, users can typically undershoot or overshoot the location and click on a nearby unintended key. Thevirtual keypad module 12 employs a combination of prediction and auto-correction methods to determine which character(s) is/are most likely to be entered by the user. - In one example, provided fast input speeds, a prediction method is used to compensate for the possible overshoot and undershoot. As will be discussed in more detail below, the predictions can be used to enlarge an activation area of possible keys while reducing (or zero-ing) activation areas of keys that are not in the next-character prediction list. However, if the prediction methods are unable to generate a prediction, even at fast input speeds, the
virtual keypad module 12 disables the selection auto-correction methods and reverts to a default mode (i.e., without enlarging or reducing the activation area). Thevirtual keypad module 12 can also disable the selection auto-correction method when the interaction becomes slow because it is assumed that clicking errors do not generally occur during slow interaction. -
FIGS. 2A and 2B illustrate an exemplaryremote controller 16 according to various aspects of the present disclosure. As shown inFIG. 2A , the exterior of theremote controller 16 includes atouchpad 22 and one or more soft keys 24 a-24 d. In various embodiments, touch zones defined by one or more coordinates of thetouchpad 22 can be mapped to a particular key of the virtual keypad 20 (FIG. 1 ). - A user can select a particular key of the
virtual keypad 20 by gently placing his finger or thumb on thetouchpad 22 at or near the associated touch zone (FingerDown event), by dragging a finger or thumb along thetouchpad 22 to the associated touch zone (FingerDrag event), and/or by lifting the finger or thumb away from the touchpad 22 (FingerUp event). While the user has a finger or thumb on the touchpad 22 (i.e., between FingerDown and FingerUp events), the user can click on thetouchpad 22 by applying greater force (FingerPress event) followed by releasing the force (FingerRelease event) to select a key. - In various other embodiments, a relative access method can be used as an alternative or as a secondary method for selecting keys. The relative access method assumes a position of the user's finger or thumb to be a current coordinate or touch zone of the
touchpad 22. Subsequent gestures by the user are then interpreted relative to that coordinate or touch zone. This allows for an adjustable precision in selection. - Important functions of the remote controller 16 (such as, for example, volume, channel, and mute) can be associated with specific selection buttons 26 a-26 d of the
touchpad 22. The selection buttons 26 a-26 d can be designated by a specific touchpad button that is painted or illuminated on the touchpad 22 (as shown) or by a button displayed on an overlay to the virtual keypad 20 (FIG. 1 ) of the GUI 18 (FIG. 1 ). This allows the user to use theremote controller 16 in complete darkness without having to look away from the content being displayed on the display device 14 (FIG. 1 ). - The functions can be controlled by simply touching the buttons or be controlled by performing a specific gesture. In one example, sliding a finger or thumb up or down on the right side of the
touchpad 22 can trigger a volume up or volume down action. In another example, sliding a finger or thumb right or left on the top side of thetouchpad 22 can trigger a channel up or channel down action. - In various embodiments, the body of the
remote controller 16 can be made of a soft material, allowing theremote controller 16 to be squeezed. The squeezing of theremote controller 16 can be performed by the user to trigger certain actions, particularly in contexts where the GUI 18 (FIG. 1 ) is just waiting for an acknowledgement without proposing a choice (such as a “next” button in a slideshow). - As shown in
FIG. 2B , the internal components of theremote controller 16 can include, but are not limited to,input sensors 30,output actuators 32, aninput controller 34, anoutput controller 36, aprocessing handler 38, a wireless transmitter (e.g., RF, Bluetooth, etc.) 40, and/or combinations thereof. The following describes operations performed by the sensors including theinput sensors 30. Theinputs sensors 30 can includetouchpad sensors 42. Thetouchpad sensors 42 can be single-position registering touchpad sensors mounted side-by-side that allow for the selection of at least two contact points on the touchpad 22 (FIG. 2A ) simultaneously. Alternatively, thetouchpad sensors 42 can be a single multi-touch capable touchpad sensor that can register, with equal precision, two points of contact at the same time. In various embodiments, thetouchpad sensors 42 can register pressure information to allow the touchpad 22 (FIG. 2A ) to be clickable. - The
input sensors 30 can also include one or more selection button sensors 44, one or moretouchpad button sensors 46, one ormore accelerometers 48, and one ormore holding sensors 50. The holdingsensors 50 can be, for example, capacitive sensors that are located around the border of theremote controller 16, and/or behind theremote controller 16. The holdingsensors 50 indicate whether the user is touching an area of theremote controller 16 in a proximity of the holdingsensor 50. Theaccelerometer 48 can be a three-axis accelerometer that indicates a positioning of theremote controller 16. Theinput controller 34 reads the real-time data from all active sensors. In various embodiments, some sensors may not be active at all times to reduce power consumption. Theprocessing handler 38 gathers and forms into packets the data to be transmitted and/or processes the real-time data from one or more active sensors to perform local actions. The RF transmitter 40 (RF driver 40) generates the signals in packet form to the display device 14 (FIG. 1 ). - The
output actuators 32 can include one ormore LED panels 52 for displaying the touchpad buttons 26 a-26 d, depending on the specific state of interaction with theGUI 18 present on-screen. Theoutput actuators 32 can additionally or alternatively include actuators for providing sufficient haptic feedback to the user (such as, for example,vibration actuators 54,light actuators 55, and/or speaker actuators 56). The above has described one example of the operations performed by theoutput actuator 32. Theoutput controller 36 updates the state of all the active actuators. - Referring now to
FIG. 3 , a dataflow diagram illustrates a more detailed exemplaryvirtual keypad module 12. Various embodiments of thevirtual keypad module 12 according to the present disclosure may include any number of sub-modules. As can be appreciated, the sub-modules shown inFIG. 3 may be combined and/or further partitioned to similarly perform text input. Thedata inputs virtual keypad module 12 are received from the remote controller 16 (FIG. 1 ) and/or received from other modules (not shown) within the display device 14 (FIG. 1 ). In various embodiments, thevirtual keypad module 12 includes anorientation recognition module 60, a handposition recognition module 62, aprediction module 64, akey input module 66, akey mapping module 67, and aGUI manager module 68. The following describes one example of operations performed by each module. Theorientation recognition module 60 determines an orientation of theremote controller 16 based on data received from the holding sensors 50 (FIG. 2B ) and the accelerometer 48 (FIG. 2B ). For example, the user can be holding the remote controller 16 (FIG. 2A ) in a portrait position, as shown inFIG. 4A , or in a landscape position, as shown inFIG. 4B . In various embodiments, theorientation recognition module 60 determines the orientation by way of an Artificial Neural Network (ANN). The ANN can be trained by data indicating both landscape position conditions and portrait position conditions. - In one example, the orientation is determined by training an ANN with sensory data. The sensory data can comprise three-dimensional acceleration (accx, accy, accz) and an activation state of the n capacitive holding position sensors, which can signal that human skin is ever in proximity (1) or is not in proximity (0). These n+3 values are fed into a single perceptron or linear classifier to determine if the remote controller 16 (
FIG. 2A ) is horizontal or vertical. Perceptron coefficients can be trained on a database and hard-coded by a manufacturer. The handposition recognition module 62 determines a holding style of the remote controller 16 (FIG. 2A ) based on data received from the holding sensors 50 (FIG. 2B ) and the accelerometer 48 (FIG. 2B ). For example, thesensory data FIG. 2A ) is held with one or two hands; and if it is held with one hand, whether it is held with the left or right hand. In various embodiments, the handposition recognition module 62 determines the holding style by way of an ANN. The ANN can be trained by data indicating right-hand conditions, left-hand conditions, and two hands conditions. - In one example, the hand position is determined similarly as discussed above. Multiple perceptrons can be implemented for the more than one binary decisions (e.g., left hand, right hand, two handed).
- As will be discussed in more detail below, the determination of the orientation and the holding style gives the
virtual keypad module 12 the ability to accommodate the user by automatically adapting the text input methods and the look and feel of the GUI 18 (FIG. 1 ). Thus, the determination of the orientation and holding position allows the user to hold the remote controller 16 (FIG. 2A ) in the most convenient way based on their personal preference and the actual conditions of use (e.g., standing, sitting, lying down). In the case of operating the remote controller 16 (FIG. 2A ) in a dark room, the user can pick up and operate the remote controller 16 (FIG. 2A ) without worrying about how they are holding it. - The hand
position recognition module 62 can further perform user verification based on a combination of holdingsensor data 70,accelerometer data 72, additional sensor information (such as an image of the palm of the user's hand), and/or bio-sensors. The data can be used to fully determine the identity of the user or, more broadly, infer the category to which the user belongs (e.g., left-handed, right-handed, kid, adult, elderly). User identification can be used, for example, for parental control, personalization, and profile switching. User categorization can be used to adapt the GUI 18 (FIG. 1 ). - The
key mapping module 67 generates a map indicating an association between the coordinates or touch zones of the touchpad 22 (FIG. 2A ) and the keys of the keypad and/or touchpad selection buttons. In various embodiments, thekey mapping module 67 generates the map based on the orientation and hand position information determined from theorientation recognition module 60 and the handposition recognition module 62, respectively. Thekey mapping module 67 maps the touch zones of thetouchpad 22 to the keys of the keypad, by associating a key of the keypad to each XY coordinate of the touch zones. Here, the touch zones are defined by one or more XY coordinates of thetouchpad 22. Thekey mapping module 67 generates a map from the mapping. Here, the map is a two-dimensional lookup table defined by the coordinates of thetouchpad 22. For example, as shown inFIG. 5 , the map can be a two dimensional (XY) table 80 that is used to assign a key of the keypad and/or a touchpad selection button to each coordinate of the touchpad 22 (FIG. 2A ). As will be discussed in more detail below, the map can then be referenced by thekey input module 66 to determine an action to be taken and can be referenced by theGUI manager module 68 to generate the GUI 18 (FIG. 1 ). - Referring back to
FIG. 3 , thekey input module 66 processestouchpad sensor data 74 and/or theaccelerometer data 72. In various embodiments, thekey input module 66 interprets thetouchpad sensor data 74 to be a coordinate or coordinates of the touchpad 22 (FIG. 2A ) and determines what action to be taken based on the coordinate or coordinates. For example, thekey input module 66 can receive thetouchpad sensor data 74, determine a particular coordinate from thedata 74, and reference the map generated by thekey mapping module 67. Based on the entries in the map, thekey input module 66, for example, can project that the user is hovering over a particular key of the keypad and, thus, entering a particular text. - In various embodiments, the
key input module 66 interprets theaccelerometer data 72 as an action to be taken. For example, theaccelerometer data 72 can indicate if a user has lifted the remote controller 16 (FIG. 2A ) quickly to select, for example, an uppercase mode. Theaccelerometer data 72 can indicate when a user has lowered the remote controller 16 (FIG. 2A ) quickly to select, for example, a lowercase mode. - To enhance the precision and speed at which the text is entered, the
prediction module 64 generates a prediction of which key and/or word the user is trying to select. Theprediction module 64 generates the prediction based on thetouchpad sensor data 74 and/or based on a determination of previous text entered. In various embodiments, theprediction module 64 performs one or more next key prediction methods, such as, for example, a language model method, a trajectory analysis method, a hand movement analysis method, a timing analysis method, and/or combinations thereof. In short, theprediction module 64, in response to an operator pressing a given touch zone to select a particular character, performs one or more key prediction methods to predict one or more next plausible keys. Examples of the operator are a finger of the user, a touch-pen, and the like, that operates thetouchpad 22. - In one example, the
prediction module 64 employs one or more language models known in the art to predict the next key based on previous text entered. The language models, for instance, predict the one or more next plausible keys based on previous characters selected by the operator. For example, if the partial word ‘pr’ has been entered, the language model can predict that a vowel is likely to follow and that the letter ‘r’ will not be a possibility. - In another example, the
prediction module 64 employs one or more language models to provide a list of reliable candidates of full words from partial word inputs. In this case, the language model generates a list of plausible words based on the previous characters selected by the operator. The full words can be selected by the user for auto-completion. For example, if the partial word ‘Pan’ has been entered, a list can be generated that includes ‘Panasonic’ and ‘Pan-American.’ Instead of typing the remaining characters, the user can simply select one of the full words. - In various embodiments, the language model can generate the word predictions based on words previously entered. For example, once selected, the words can be remembered and the language model can be adapted to favor the remembered words.
- In yet another example, the trajectory analysis method can be performed to predict possible next keys based on future path estimation including directions and/or velocities of user's finger or thumb movement on the touchpad 22 (
FIG. 2A ) as indicated by thetouchpad sensor data 74. The trajectory analysis method predicts next plausible keys based on a direction of operator movement on thetouchpad 22. For example, if the user first selects the ‘k’ key and the language model predicts that the next key can be one of ‘I,’ ‘e,’ or ‘a,’ thetouchpad sensor data 74 can be evaluated to determine a direction the user is heading and velocity of the movement and, thus, eliminate one or more of the choices. - For example, the trajectory analysis method determines a coordinate of the key ‘k’ and subsequent finger movements. From that history of XY coordinates, the future path is determined. The path includes a tolerance to account for short-term prediction (more accurate) and longer-term prediction (less accurate). If the future path estimation is heading away from the coordinates of the predicted key, the choice is eliminated. For example, if the path is heading on an Up/Left diagonal line, then the keys ‘e’ and ‘a’ are eliminated and the key ‘I’ is selected as the predicted next key.
- In another example, the hand movement analysis method can be performed using the holding style information provided by the hand
position recognition module 62 and the predictions provided by theprediction module 64. The hand movement analysis method predicts next plausible keys based on a detection of which hand is moving on thetouchpad 22. The hand movement analysis method can evaluate whether the remote controller 16 (FIG. 2A ) is held with two hands or one hand. If the remote controller 16 (FIG. 2A ) is held by two hands and movement by a right hand or left hand is detected, then the choices that are associated with the non-moving hand would be eliminated. - For example, if the user first selects the ‘k’ key, the
prediction module 64 predicts that the next key can be one of ‘e,’ or ‘a,’ and movement is detected by the right hand, then the keys ‘e’ and ‘a’ are eliminated and the key ‘I’ is selected as the predicted next key. - In yet another example, the timing analysis method can be performed when the
prediction module 64 is unable to predict the appropriate next key or word via any of the next key prediction methods. Such may be the case when the user is entering proper nouns, such as a last name or a password. The timing analysis method evaluates the time the user takes to move from one key to the next. That is, the timing analysis method predicts next plausible keys based on a timing of operator movement on thetouchpad 22. In more detail, the timing analysis method predicts no next plausible keys when the timing exceeds a predetermined limit. Here, thekey mapping module 67 remaps the touch zones of thetouchpad 22 to the keys of the keypad based on the no next plausible keys. If the user moves more slowly, it is more likely that a proper noun is being entered. The predictions are then ignored. - Based on the predictions provided by the
prediction module 64, thekey mapping module 67 can remap the coordinates of the touchpad 22 (FIG. 2A ) to the keys of the virtual keypad 20 (FIG. 1 ). That is, thekey mapping module 67 remaps the touch zones of thetouchpad 22 to the keys of the keypad based on the one or more next plausible keys predicted by theprediction module 64. In various embodiments, the coordinates that are associated with the predicted next key can be expanded to make the key more accessible. For example, as shown inFIG. 6 , if the predicted next key is ‘p,’ the map can be adjusted such that the coordinates that were previously mapped to the keys ‘o’ or ‘I’ are now mapped to the key ‘p.’ Thus, if the user is actually hovering over the ‘I’ key, the ‘p’ key will be selected and entered if the user clicks on the touchpad 22 (FIG. 2A ). - Referring back to
FIG. 3 , in various embodiments, thekey mapping module 67 can remap the coordinates to the keys based on a relative speed of the user movement. For example, thekey mapping module 67 can remap the coordinates such that the predicted key is mapped to a larger touch zone area when a faster movement is detected. Thekey mapping module 67 can maintain the original coordinates or remap to the original coordinates when slower movements are detected. Thekey mapping module 67 can scale the area between the larger area and the original area when a movement that is not fast or slow is detected. - The
GUI manager module 68 generatesGUI data 76 for displaying theGUI 18. TheGUI 18 provides visual feedback to the user indicating the actions they have performed with the remote controller 16 (FIG. 2A ). As shown in theexemplary GUIs 18 ofFIGS. 7 through 10 , theGUI 18 can include thevirtual keypad 20, including multiplealphanumeric keys 90, one ormore selection buttons 92, aselection list 94, atext display box 96, atitle bar 98, and/or combinations thereof. - As discussed further below, the
touchpad 22 is divided into two operating zones (a first operating zone and a second operating zone). Thekey mapping module 67 maps a first subset of keys of the keypad to the first operating zone of thetouchpad 22 and maps a second subset of keys of the keypad to the second operating zone of thetouchpad 22. For example, the first operating zone corresponds to a top zone or a right zone of thetouchpad 22, and the second operating zone corresponds to a bottom zone or a left zone of thetouchpad 22. Here, each of the first subset and the second subset is a group of keys of the keypad. In various embodiments, theGUI manager module 68 displays thevirtual keypad 20 based on the holding position and the orientation determined by the handposition recognition module 62 and theorientation recognition module 60, respectively. For example, as shown inFIG. 7 , if two hands are used to hold the remote controller 16 (FIG. 2A ) in the landscape position, thevirtual keypad 20 and thetouchpad 22 can be divided into twozones 100, 102 (e.g., left and right). The keys in theright zone 102 can be associated with afirst zone 104 of thetouchpad 22, and the keys in theleft zone 100 can be associated with asecond zone 106 of thetouchpad 22. The user can select a key 90 in theright zone 102 with a right thumb or finger, and the user can select a key 90 in theleft zone 100 with a left thumb or finger. Thekeys 90 in eachzone keys 90 so that the user can determine whichkeys 90 can be selected with which thumb or finger. This will allow for a natural text input, similar to the experience when entering text on classic keyboards. - In this example, selection of a predicted word (auto-completion) can be made through the display of the two most probable words (e.g., ‘Panasonic,’ ‘Pan-America’). The
selection buttons 92 are auto-completion selection buttons. When the user selects one of theselection buttons 92, a probable word displayed in the selectedselection button 92 is displayed as a complete word on thedisplay box 96. For example, the two words can be displayed onselection buttons 92. The user can select theselection buttons 92 by pushingsoft keys remote controller 16 with the index fingers, or by dragging the finger or thumb to a dedicated zone located at a designated location of thetouchpad 22. - In various embodiments, when the
touchpad 22 and thevirtual keypad 20 are divided into two zones, the mapping of the coordinates can provide for an overlap between the two areas. Thekey mapping module 67 maps a third subset of keys of the first subset of keys to the second operating zone and maps a fourth subset of keys of the second subset of keys to the first operating zone. For instance, the third subset of keys include the keys designated by the letters ‘t,’ ‘g,’ and ‘b’ of the keys of the first subset (keys included in thezone 100 ofFIG. 7 ). Likewise, the fourth subset of keys include the keys designated by the letters ‘y,’ ‘h,’ and ‘n’ of the keys of the second subset (keys included in thezone 102 ofFIG. 7 ). That enables the keys along a boundary between the divided zones to operate for any zones of thetouchpad 22. For example, the letter ‘g’ in left keyboard area can be selected by thetouchpad 22 in the first zone, as well as be selected by thetouchpad 22 in thesecond zone 106. The overlap keys can be identified on theGUI 18 by shading or color. - Referring now to
FIGS. 8A and 8B , in another example, if two hands are used to hold theremote controller 16 in the landscape position, thetouchpad 22 can be divided into twozones virtual keypad 20, however, is not divided. Afirst zone touchpad 22 can be associated with the entirevirtual keypad 20, and can be referred to as the “Key-Entry Zone.” Thesecond zone touchpad 22 can be associated with the auto-completion buttons 92 (FIG. 7 ) or selection lists 94 (auto-completion selection list), and can be referred to as the “Auto-Completion Zone.” The auto-completion buttons 92 or selection lists 94 can be displayed when the user touches thetouchpad 22 in the “Auto-Completion Zone.” The user interface can include a small visual notification to signal the availability and status of auto-completion words. This way the user will not be bothered with auto-completion unless he decides to use it. - In various embodiments, the
zones touchpad 22 can be configured based on an identification of the user. For example, in the case of a right-handed user, theright zone 110 can be associated with the “Key-Entry Zone” used most often and theleft zone 108 can be associated with the “Auto-Completion Zone.” Similarly, for a left-handed user, theleft zone 108 can be associated with the “Key-Entry Zone” used most often and theright zone 110 can be associated with the “Auto-Completion Zone.” - Referring now to
FIG. 9 , in yet another example, if one hand is used to hold theremote controller 16 in the portrait position, thetouchpad 22 is divided into zones. Thevirtual keypad 20, however, is not divided into zones. Only one zone of thetouchpad 22 is used by the user at all times. The user will be able to address the entirevirtual keypad 20 from theentire touchpad 22.Such touchpad 22 will principally function as the “Key-Entry Zone.” If auto-completion is needed, the user can switch thetouchpad 22 to the “Auto-completion Zone” by using a simple gesture. Such gesture can include, but is not limited to, moving the thumb or finger to a specific area of the touchpad 22 (for instance lower right), or sliding the finger or thumb along thetouchpad 22 quickly from right to left. - In any of the examples shown in
FIGS. 7 through 9 , the user can select a key 90 by dragging the thumb or finger on thetouchpad 22. In response to the movement, one or more on-screen pointers (such as, for example, a cursor, or an illustration of a thumb or finger) slides to atarget key 90. The on-screen pointers can be displayed according to the hand position (e.g., left hand only, right hand only, or both hands). The key 90 can be selected by clicking theclickable touchpad 22 and/or upon release. When displaying a thumb or finger as the pointer, a different thumb or finger posture can be used to indicate a thumb or finger press as opposed to a thumb or finger that is dragged on the touchpad surface. The selected character associated with the key will be displayed in thetext display box 96. - Referring now to
FIG. 10 , in various embodiments, theGUI manager module 68 displays thekeys 90 of thevirtual keypad 20 based on the predicted next key and the remapping of the coordinates performed by the key mapping module 67 (FIG. 3 ). For example, by knowing the next likely key, the GUI manager module 68 (FIG. 3 ) can highlight and/or enlarge the most likely key 116 based on the mapping of the coordinates. That is, an arrangement of the keypad on the user interface is modified based on the one or more next plausible keys predicted by theprediction module 64. In more detail, one or more keys of the keypad on the user interface is highlighted or enlarged based on the one or more next plausible keys predicted by theprediction module 64. These operations are performed by theGUI manager module 68. However, when the user is not moving the cursor quickly or the user hovers over a given coordinate, the highlighted and/orenlarged key 116 is remapped to the original coordinates of thetouchpad 22 and the highlighted and/or enlarged key can be resized back to the original size. - Those skilled in the art can now appreciate from the foregoing description that the broad teachings of the present disclosure can be implemented in a variety of forms. Therefore, while this disclosure has been described in connection with particular examples thereof, the true scope of the disclosure should not be so limited since other modifications will become apparent to the skilled practitioner upon a study of the drawings, specification, and the following claims.
Claims (24)
1. A virtual keypad system for inputting text, comprising:
a remote controller having at least one touchpad incorporated therein and divided into a plurality of touch zones;
a display device in data communication with the remote controller and operable to display a user interface including a keypad, where each key of the keypad is mapped to a touch zone of the touchpad;
a prediction module, in response to an operator pressing a given touch zone to select a particular character, performs one or more key prediction methods to predict one or more next plausible keys; and
a key mapping module remaps the touch zones of the touchpad to the keys of the keypad based on the one or more next plausible keys.
2. The system of claim 1 ,
wherein an arrangement of the keypad on the user interface is modified based on the one or more next plausible keys.
3. The system of claim 2 ,
wherein one or more keys of the keypad on the user interface is enlarged based on the one or more next plausible keys.
4. The system of claim 2 ,
wherein one or more keys of the keypad on the user interface is highlighted based on the one or more next plausible keys.
5. The system of claim 1 further comprising
a language model operable to predict the one or more next plausible keys based on previous characters selected by the operator.
6. The system of claim 5 ,
wherein the language model is further operable to generate a list of plausible words based on the previous characters selected by the operator.
7. The system of claim 1 ,
wherein the one or more key prediction methods includes a trajectory analysis method, where the next plausible key is predicted based on a direction of operator movement on the touchpad.
8. The system of claim 1 ,
wherein the one or more key prediction methods includes a hand movement analysis method, where the next plausible key is predicted based on a detection of which hand is moving on the touchpad.
9. The system of claim 1 ,
wherein the touchpad is divided into two operating zones and wherein the key mapping module maps a first subset of keys of the keypad to a first operating zone of the touchpad and maps a second subset of keys of the keypad to a second operating zone of the touchpad.
10. The system of claim 9 ,
wherein the first operating zone corresponds to a top zone of the touchpad and wherein the second operating zone corresponds to a bottom zone of the touchpad.
11. The system of claim 9 ,
wherein the first operating zone corresponds to a right zone of the touchpad and wherein the second operating zone corresponds to a left zone of the touch pad.
12. The system of claim 1 ,
wherein the user interface includes at least one of auto-completion selection buttons and auto-completion selection list.
13. The system of claim 12 ,
wherein the touchpad is divided into two operating zones and wherein the key mapping module maps a first operating zone to the keys of the keypad and maps a second operating zone to the at least one of the of auto-completion selection buttons and auto-completion selection.
14. The system of claim 9 ,
wherein the key mapping module maps a third subset of keys of the first subset of keys to the second operating zone and maps a fourth subset of keys of the second subset of keys to the first operating zone.
15. The system of claim 14 ,
wherein the third subset of keys include the keys designated. by the letters ‘t,’ ‘g,’ and ‘b,’ and wherein the fourth subset of keys include the keys designated by the letters ‘y,’ ‘h,’ and ‘n.’
16. The system of claim 1 ,
wherein the key prediction methods include a timing analysis method, where the next plausible keys are predicted based on a timing of operator movement on the touchpad.
17. The system of claim 16 ,
wherein the timing analysis method predicts no next plausible keys when the timing exceeds a predetermined limit.
18. The system of claim 17 ,
wherein the key mapping module remaps the touch zones of the touchpad to the keys of the keypad based on the no next plausible keys.
19. The system of claim 1 ,
wherein the touch zones are defined by one or more XY coordinates of the touchpad and wherein the key mapping module maps the touch zones of the touchpad to the keys of the keypad by associating a key of the keypad to each XY coordinate of the touch zones.
20. The system of claim 19 ,
wherein the key mapping modules generates a map from the mapping and wherein the map is a two-dimensional lookup table defined by the coordinates of the touchpad.
21. A virtual keypad system for inputting text, comprising:
a remote controller having at least one touchpad incorporated therein and divided into a plurality of touch zones;
a display device in data communication with the remote controller and operable to display a keypad and an area for displaying input from the keypad, where each key on the keypad is associated with a touch zone on the touchpad and, in response to an operator pressing a given touch zone, a character on the key correlating to the given touch zone is displayed in an input area of the display;
a language model adapted to receive characters displayed in the input area and operable to predict next most plausible characters in string of received characters; and
a module that arranges the each key on the keypad for facilitating entry of a desired key based on output from the language model.
22. The system of claim 21 ,
wherein the arranging comprises enlarging keys on the keypad based on the output from the language model.
23. A virtual keypad system for inputting text, comprising:
a remote controller having at least one touchpad incorporated therein and divided into a plurality of touch zones;
a display device in data communication with the remote controller and operable to display a user interface including a keypad, where each key of the keypad is mapped to a touch zone of the touchpad;
a prediction module performs one or more key prediction methods to predict one or more next plausible keys;
a key mapping module remaps the touch zones of the touchpad to the keys of the keypad based on the one or more next plausible keys;
a user interface manager module modifies an arrangement of the keypad based on the remapping of the touch zones of the touchpad to the keys of the keypad; and
a text input module, in response to an operator pressing a given touch zone, selects a key based on the remapping of the touch zones of the touchpad to the keys of the keypad.
24. The system of claim 23 ,
wherein the user interface manager module enlarges one or more keys of the keypad based on the remapping of the touch zones of the touchpad to the keys of the keypad.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/666,916 US20100164897A1 (en) | 2007-06-28 | 2008-06-26 | Virtual keypad systems and methods |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US94685807P | 2007-06-28 | 2007-06-28 | |
US11/977,346 US8065624B2 (en) | 2007-06-28 | 2007-10-24 | Virtual keypad systems and methods |
PCT/US2008/068384 WO2009006221A1 (en) | 2007-06-28 | 2008-06-26 | Virtual keypad systems and methods |
US12/666,916 US20100164897A1 (en) | 2007-06-28 | 2008-06-26 | Virtual keypad systems and methods |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/977,346 Continuation US8065624B2 (en) | 2007-06-28 | 2007-10-24 | Virtual keypad systems and methods |
Publications (1)
Publication Number | Publication Date |
---|---|
US20100164897A1 true US20100164897A1 (en) | 2010-07-01 |
Family
ID=40162292
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/977,346 Active 2030-01-08 US8065624B2 (en) | 2007-06-28 | 2007-10-24 | Virtual keypad systems and methods |
US12/666,916 Abandoned US20100164897A1 (en) | 2007-06-28 | 2008-06-26 | Virtual keypad systems and methods |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/977,346 Active 2030-01-08 US8065624B2 (en) | 2007-06-28 | 2007-10-24 | Virtual keypad systems and methods |
Country Status (4)
Country | Link |
---|---|
US (2) | US8065624B2 (en) |
EP (1) | EP2160672A4 (en) |
JP (1) | JP5323070B2 (en) |
WO (1) | WO2009006221A1 (en) |
Cited By (277)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080055269A1 (en) * | 2006-09-06 | 2008-03-06 | Lemay Stephen O | Portable Electronic Device for Instant Messaging |
US20090005011A1 (en) * | 2007-06-28 | 2009-01-01 | Greg Christie | Portable Electronic Device with Conversation Management for Incoming Instant Messages |
US20090124293A1 (en) * | 2007-11-08 | 2009-05-14 | Tsui Yuan-Mao | Method for displaying dialing information and mobile communication device using the method |
US20090177981A1 (en) * | 2008-01-06 | 2009-07-09 | Greg Christie | Portable Electronic Device for Instant Messaging Multiple Recipients |
US20100185960A1 (en) * | 2003-05-02 | 2010-07-22 | Apple Inc. | Method and Apparatus for Displaying Information During an Instant Messaging Session |
US20100228539A1 (en) * | 2009-03-06 | 2010-09-09 | Motorola, Inc. | Method and apparatus for psychomotor and psycholinguistic prediction on touch based device |
US20100315266A1 (en) * | 2009-06-15 | 2010-12-16 | Microsoft Corporation | Predictive interfaces with usability constraints |
US20110012843A1 (en) * | 2009-07-14 | 2011-01-20 | Chih-Hung Li | Touch-controlled electronic apparatus and related control method |
US20110148762A1 (en) * | 2009-12-22 | 2011-06-23 | Universal Electronics Inc. | System and method for multi-mode command input |
US20110187647A1 (en) * | 2010-02-04 | 2011-08-04 | Charles Howard Woloszynski | Method and apparatus for virtual keyboard interactions from secondary surfaces |
US20110201387A1 (en) * | 2010-02-12 | 2011-08-18 | Microsoft Corporation | Real-time typing assistance |
US20110212876A1 (en) * | 2010-02-10 | 2011-09-01 | Michelle Meek | Cleaning composition comprising amylase variants with high stability in the presence of a chelating agent |
US20110279397A1 (en) * | 2009-01-26 | 2011-11-17 | Zrro Technologies (2009) Ltd. | Device and method for monitoring the object's behavior |
US20110310019A1 (en) * | 2010-06-16 | 2011-12-22 | International Business Machines Corporation | Reconfiguration of virtual keyboard |
US20120078614A1 (en) * | 2010-09-27 | 2012-03-29 | Primesense Ltd. | Virtual keyboard for a non-tactile three dimensional user interface |
US20120098743A1 (en) * | 2010-10-26 | 2012-04-26 | Pei-Ling Lai | Input method, input device, and computer system |
US20120137244A1 (en) * | 2010-11-30 | 2012-05-31 | Inventec Corporation | Touch device input device and operation method of the same |
CN102937876A (en) * | 2011-11-23 | 2013-02-20 | 微软公司 | Dynamic scaling of a touch sensor |
US20130181902A1 (en) * | 2012-01-17 | 2013-07-18 | Microsoft Corporation | Skinnable touch device grip patterns |
US20130229363A1 (en) * | 2012-03-02 | 2013-09-05 | Christopher A. Whitman | Sensing User Input At Display Area Edge |
US20130290410A1 (en) * | 2012-04-28 | 2013-10-31 | Alibaba Group Holding Limited | Performing autocomplete of content |
WO2013157232A3 (en) * | 2012-04-19 | 2014-01-09 | Sony Corporation | Information processing apparatus, information processing method, program, and information processing system |
US8719724B2 (en) | 2011-03-16 | 2014-05-06 | Honeywell International Inc. | Method for enlarging characters displayed on an adaptive touch screen key pad |
WO2014107064A1 (en) * | 2013-01-07 | 2014-07-10 | Samsung Electronics Co., Ltd. | Display apparatus, method for ui display thereof and computer-readable recording medium |
US8850241B2 (en) | 2012-03-02 | 2014-09-30 | Microsoft Corporation | Multi-stage power adapter configured to provide low power upon initial connection of the power adapter to the host device and high power thereafter upon notification from the host device to the power adapter |
US8854323B2 (en) | 2010-05-20 | 2014-10-07 | Panasonic Intellectual Property Corporation Of America | Operating apparatus, operating method, program, recording medium, and integrated circuit |
US8872762B2 (en) | 2010-12-08 | 2014-10-28 | Primesense Ltd. | Three dimensional user interface cursor control |
US8873227B2 (en) | 2012-03-02 | 2014-10-28 | Microsoft Corporation | Flexible hinge support layer |
US8881051B2 (en) | 2011-07-05 | 2014-11-04 | Primesense Ltd | Zoom-based gesture user interface |
WO2014178813A1 (en) * | 2013-04-29 | 2014-11-06 | Hewlett-Packard Development Company, L.P. | Device for displaying a received user interface |
US8892446B2 (en) | 2010-01-18 | 2014-11-18 | Apple Inc. | Service orchestration for intelligent automated assistant |
US8933876B2 (en) | 2010-12-13 | 2015-01-13 | Apple Inc. | Three dimensional user interface session control |
US20150029402A1 (en) * | 2013-07-26 | 2015-01-29 | Tianjin Funayuanchuang Technology Co.,Ltd. | Remote controller, system, and method for controlling remote controller |
US8949477B2 (en) | 2012-05-14 | 2015-02-03 | Microsoft Technology Licensing, Llc | Accessory device architecture |
WO2015052724A1 (en) * | 2013-10-07 | 2015-04-16 | Deshmukh Rakesh | An indian language keypad |
US9030498B2 (en) | 2011-08-15 | 2015-05-12 | Apple Inc. | Combining explicit select gestures and timeclick in a non-tactile three dimensional user interface |
US9035876B2 (en) | 2008-01-14 | 2015-05-19 | Apple Inc. | Three-dimensional user interface session control |
US9064654B2 (en) | 2012-03-02 | 2015-06-23 | Microsoft Technology Licensing, Llc | Method of manufacturing an input device |
US9075566B2 (en) | 2012-03-02 | 2015-07-07 | Microsoft Technoogy Licensing, LLC | Flexible hinge spine |
US9122311B2 (en) | 2011-08-24 | 2015-09-01 | Apple Inc. | Visual feedback for tactile and non-tactile user interfaces |
US9158375B2 (en) | 2010-07-20 | 2015-10-13 | Apple Inc. | Interactive reality augmentation for natural interaction |
US9201501B2 (en) | 2010-07-20 | 2015-12-01 | Apple Inc. | Adaptive projector |
US9218063B2 (en) | 2011-08-24 | 2015-12-22 | Apple Inc. | Sessionless pointing user interface |
US9229534B2 (en) | 2012-02-28 | 2016-01-05 | Apple Inc. | Asymmetric mapping for tactile and non-tactile user interfaces |
US9261964B2 (en) | 2005-12-30 | 2016-02-16 | Microsoft Technology Licensing, Llc | Unintentional touch rejection |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US9285874B2 (en) | 2011-02-09 | 2016-03-15 | Apple Inc. | Gaze detection in a 3D mapping environment |
US9300784B2 (en) | 2013-06-13 | 2016-03-29 | Apple Inc. | System and method for emergency calls initiated by voice command |
US9304549B2 (en) | 2013-03-28 | 2016-04-05 | Microsoft Technology Licensing, Llc | Hinge mechanism for rotatable component attachment |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US9330381B2 (en) | 2008-01-06 | 2016-05-03 | Apple Inc. | Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US9354748B2 (en) | 2012-02-13 | 2016-05-31 | Microsoft Technology Licensing, Llc | Optical stylus interaction |
US9360893B2 (en) | 2012-03-02 | 2016-06-07 | Microsoft Technology Licensing, Llc | Input device writing surface |
US9368114B2 (en) | 2013-03-14 | 2016-06-14 | Apple Inc. | Context-sensitive handling of interruptions |
US9377863B2 (en) | 2012-03-26 | 2016-06-28 | Apple Inc. | Gaze-enhanced virtual touchscreen |
US9377865B2 (en) | 2011-07-05 | 2016-06-28 | Apple Inc. | Zoom-based gesture user interface |
US9426905B2 (en) | 2012-03-02 | 2016-08-23 | Microsoft Technology Licensing, Llc | Connection device for computing devices |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US9447620B2 (en) | 2014-09-30 | 2016-09-20 | Microsoft Technology Licensing, Llc | Hinge mechanism with multiple preset positions |
US9459758B2 (en) | 2011-07-05 | 2016-10-04 | Apple Inc. | Gesture-based interface with enhanced features |
US9477337B2 (en) | 2014-03-14 | 2016-10-25 | Microsoft Technology Licensing, Llc | Conductive trace routing for display and bezel sensors |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9502031B2 (en) | 2014-05-27 | 2016-11-22 | Apple Inc. | Method for supporting dynamic grammars in WFST-based ASR |
US9501161B2 (en) | 2010-10-22 | 2016-11-22 | Hewlett-Packard Development Company, L.P. | User interface for facilitating character input |
US20160342294A1 (en) * | 2015-05-19 | 2016-11-24 | Google Inc. | Multi-switch option scanning |
US9535906B2 (en) | 2008-07-31 | 2017-01-03 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US9544504B2 (en) | 2012-11-02 | 2017-01-10 | Microsoft Technology Licensing, Llc | Rapid synchronized lighting and shuttering |
US9576574B2 (en) | 2012-09-10 | 2017-02-21 | Apple Inc. | Context-sensitive handling of interruptions by intelligent digital assistant |
US9582122B2 (en) | 2012-11-12 | 2017-02-28 | Microsoft Technology Licensing, Llc | Touch-sensitive bezel techniques |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
US9606986B2 (en) | 2014-09-29 | 2017-03-28 | Apple Inc. | Integrated word N-gram and class M-gram language models |
US9620104B2 (en) | 2013-06-07 | 2017-04-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9620105B2 (en) | 2014-05-15 | 2017-04-11 | Apple Inc. | Analyzing audio input for efficient speech and music recognition |
US9626955B2 (en) | 2008-04-05 | 2017-04-18 | Apple Inc. | Intelligent text-to-speech conversion |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US9633674B2 (en) | 2013-06-07 | 2017-04-25 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
US9633660B2 (en) | 2010-02-25 | 2017-04-25 | Apple Inc. | User profiling for voice input processing |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US9646614B2 (en) | 2000-03-16 | 2017-05-09 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US9697822B1 (en) | 2013-03-15 | 2017-07-04 | Apple Inc. | System and method for updating an adaptive speech recognition model |
US9711141B2 (en) | 2014-12-09 | 2017-07-18 | Apple Inc. | Disambiguating heteronyms in speech synthesis |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US9734193B2 (en) | 2014-05-30 | 2017-08-15 | Apple Inc. | Determining domain salience ranking from ambiguous words in natural speech |
US9752361B2 (en) | 2015-06-18 | 2017-09-05 | Microsoft Technology Licensing, Llc | Multistage hinge |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US9779225B2 (en) | 2015-04-08 | 2017-10-03 | Google Inc. | Method and system to provide access to secure features of a device |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US9798393B2 (en) | 2011-08-29 | 2017-10-24 | Apple Inc. | Text correction processing |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US9824808B2 (en) | 2012-08-20 | 2017-11-21 | Microsoft Technology Licensing, Llc | Switchable magnetic lock |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US9864415B2 (en) | 2015-06-30 | 2018-01-09 | Microsoft Technology Licensing, Llc | Multistage friction hinge |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US9870066B2 (en) | 2012-03-02 | 2018-01-16 | Microsoft Technology Licensing, Llc | Method of manufacturing an input device |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9922642B2 (en) | 2013-03-15 | 2018-03-20 | Apple Inc. | Training an at least partial voice command system |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9953088B2 (en) | 2012-05-14 | 2018-04-24 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US9959870B2 (en) | 2008-12-11 | 2018-05-01 | Apple Inc. | Speech recognition involving a mobile device |
US9966068B2 (en) | 2013-06-08 | 2018-05-08 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US9966065B2 (en) | 2014-05-30 | 2018-05-08 | Apple Inc. | Multi-command single utterance input method |
US9971774B2 (en) | 2012-09-19 | 2018-05-15 | Apple Inc. | Voice-based media searching |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10001857B2 (en) | 2015-05-13 | 2018-06-19 | Seiko Epson Corporation | Display apparatus and method of controlling display apparatus |
US10031556B2 (en) | 2012-06-08 | 2018-07-24 | Microsoft Technology Licensing, Llc | User experience adaptation |
US10037057B2 (en) | 2016-09-22 | 2018-07-31 | Microsoft Technology Licensing, Llc | Friction hinge |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10079014B2 (en) | 2012-06-08 | 2018-09-18 | Apple Inc. | Name recognition system |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10089072B2 (en) | 2016-06-11 | 2018-10-02 | Apple Inc. | Intelligent device arbitration and control |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US10107994B2 (en) | 2012-06-12 | 2018-10-23 | Microsoft Technology Licensing, Llc | Wide field-of-view virtual image projector |
US10120420B2 (en) | 2014-03-21 | 2018-11-06 | Microsoft Technology Licensing, Llc | Lockable display and techniques enabling use of lockable displays |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US10156889B2 (en) | 2014-09-15 | 2018-12-18 | Microsoft Technology Licensing, Llc | Inductive peripheral retention device |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
US10185542B2 (en) | 2013-06-09 | 2019-01-22 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10199051B2 (en) | 2013-02-07 | 2019-02-05 | Apple Inc. | Voice trigger for a digital assistant |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US10262324B2 (en) | 2010-11-29 | 2019-04-16 | Biocatch Ltd. | System, device, and method of differentiating among users based on user-specific page navigation sequence |
US10269345B2 (en) | 2016-06-11 | 2019-04-23 | Apple Inc. | Intelligent task discovery |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US10283110B2 (en) | 2009-07-02 | 2019-05-07 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US10289433B2 (en) | 2014-05-30 | 2019-05-14 | Apple Inc. | Domain specific language for encoding assistant dialog |
US10298614B2 (en) * | 2010-11-29 | 2019-05-21 | Biocatch Ltd. | System, device, and method of generating and managing behavioral biometric cookies |
US10297253B2 (en) | 2016-06-11 | 2019-05-21 | Apple Inc. | Application integration with a digital assistant |
US10303715B2 (en) | 2017-05-16 | 2019-05-28 | Apple Inc. | Intelligent automated assistant for media exploration |
US10311144B2 (en) | 2017-05-16 | 2019-06-04 | Apple Inc. | Emoji word sense disambiguation |
US10318871B2 (en) | 2005-09-08 | 2019-06-11 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US10324733B2 (en) | 2014-07-30 | 2019-06-18 | Microsoft Technology Licensing, Llc | Shutdown notifications |
US10332518B2 (en) | 2017-05-09 | 2019-06-25 | Apple Inc. | User interface for correcting recognition errors |
US10344797B2 (en) | 2016-04-05 | 2019-07-09 | Microsoft Technology Licensing, Llc | Hinge with multiple preset positions |
US10356243B2 (en) | 2015-06-05 | 2019-07-16 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10354011B2 (en) | 2016-06-09 | 2019-07-16 | Apple Inc. | Intelligent automated assistant in a home environment |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US10395654B2 (en) | 2017-05-11 | 2019-08-27 | Apple Inc. | Text normalization based on a data-driven learning network |
US10397262B2 (en) | 2017-07-20 | 2019-08-27 | Biocatch Ltd. | Device, system, and method of detecting overlay malware |
US10403283B1 (en) | 2018-06-01 | 2019-09-03 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US10403278B2 (en) | 2017-05-16 | 2019-09-03 | Apple Inc. | Methods and systems for phonetic matching in digital assistant services |
US10404729B2 (en) | 2010-11-29 | 2019-09-03 | Biocatch Ltd. | Device, method, and system of generating fraud-alerts for cyber-attacks |
US10410637B2 (en) | 2017-05-12 | 2019-09-10 | Apple Inc. | User-specific acoustic models |
US10417339B2 (en) | 2017-08-18 | 2019-09-17 | Kyocera Document Solutions Inc. | Suggestion of alternate user input using different user interface |
US10417266B2 (en) | 2017-05-09 | 2019-09-17 | Apple Inc. | Context-aware ranking of intelligent response suggestions |
US10445429B2 (en) | 2017-09-21 | 2019-10-15 | Apple Inc. | Natural language understanding using vocabularies with compressed serialized tries |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US10474815B2 (en) | 2010-11-29 | 2019-11-12 | Biocatch Ltd. | System, device, and method of detecting malicious automatic script and code injection |
US10474753B2 (en) | 2016-09-07 | 2019-11-12 | Apple Inc. | Language identification using recurrent neural networks |
US10482874B2 (en) | 2017-05-15 | 2019-11-19 | Apple Inc. | Hierarchical belief states for digital assistants |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10496705B1 (en) | 2018-06-03 | 2019-12-03 | Apple Inc. | Accelerated task performance |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US10521466B2 (en) | 2016-06-11 | 2019-12-31 | Apple Inc. | Data driven natural language event detection and classification |
US10523680B2 (en) * | 2015-07-09 | 2019-12-31 | Biocatch Ltd. | System, device, and method for detecting a proxy server |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US10568032B2 (en) | 2007-04-03 | 2020-02-18 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US10579784B2 (en) | 2016-11-02 | 2020-03-03 | Biocatch Ltd. | System, device, and method of secure utilization of fingerprints for user authentication |
US10586036B2 (en) | 2010-11-29 | 2020-03-10 | Biocatch Ltd. | System, device, and method of recovery and resetting of user authentication factor |
US10592604B2 (en) | 2018-03-12 | 2020-03-17 | Apple Inc. | Inverse text normalization for automatic speech recognition |
US10592095B2 (en) | 2014-05-23 | 2020-03-17 | Apple Inc. | Instantaneous speaking of content on touch devices |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US10621585B2 (en) | 2010-11-29 | 2020-04-14 | Biocatch Ltd. | Contextual mapping of web-pages, and generation of fraud-relatedness score-values |
US10636424B2 (en) | 2017-11-30 | 2020-04-28 | Apple Inc. | Multi-turn canned dialog |
US10643611B2 (en) | 2008-10-02 | 2020-05-05 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US10657328B2 (en) | 2017-06-02 | 2020-05-19 | Apple Inc. | Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US10684703B2 (en) | 2018-06-01 | 2020-06-16 | Apple Inc. | Attention aware virtual assistant dismissal |
US10685355B2 (en) * | 2016-12-04 | 2020-06-16 | Biocatch Ltd. | Method, device, and system of detecting mule accounts and accounts used for money laundering |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10706373B2 (en) | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US10719765B2 (en) | 2015-06-25 | 2020-07-21 | Biocatch Ltd. | Conditional behavioral biometrics |
US10728761B2 (en) | 2010-11-29 | 2020-07-28 | Biocatch Ltd. | Method, device, and system of detecting a lie of a user who inputs data |
US10726832B2 (en) | 2017-05-11 | 2020-07-28 | Apple Inc. | Maintaining privacy of personal information |
US10733982B2 (en) | 2018-01-08 | 2020-08-04 | Apple Inc. | Multi-directional dialog |
US10733993B2 (en) | 2016-06-10 | 2020-08-04 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10733375B2 (en) | 2018-01-31 | 2020-08-04 | Apple Inc. | Knowledge-based framework for improving natural language understanding |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10747305B2 (en) | 2010-11-29 | 2020-08-18 | Biocatch Ltd. | Method, system, and device of authenticating identity of a user of an electronic device |
US10748546B2 (en) | 2017-05-16 | 2020-08-18 | Apple Inc. | Digital assistant services based on device capabilities |
US10755703B2 (en) | 2017-05-11 | 2020-08-25 | Apple Inc. | Offline personal assistant |
US10755051B2 (en) | 2017-09-29 | 2020-08-25 | Apple Inc. | Rule-based natural language processing |
US10762293B2 (en) | 2010-12-22 | 2020-09-01 | Apple Inc. | Using parts-of-speech tagging and named entity recognition for spelling correction |
US10776476B2 (en) | 2010-11-29 | 2020-09-15 | Biocatch Ltd. | System, device, and method of visual login |
US10791176B2 (en) | 2017-05-12 | 2020-09-29 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US10791216B2 (en) | 2013-08-06 | 2020-09-29 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
US10789959B2 (en) | 2018-03-02 | 2020-09-29 | Apple Inc. | Training speaker recognition models for digital assistants |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US10789945B2 (en) | 2017-05-12 | 2020-09-29 | Apple Inc. | Low-latency intelligent automated assistant |
US10810274B2 (en) | 2017-05-15 | 2020-10-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
US10818288B2 (en) | 2018-03-26 | 2020-10-27 | Apple Inc. | Natural assistant interaction |
US10834590B2 (en) | 2010-11-29 | 2020-11-10 | Biocatch Ltd. | Method, device, and system of differentiating between a cyber-attacker and a legitimate user |
US10839159B2 (en) | 2018-09-28 | 2020-11-17 | Apple Inc. | Named entity normalization in a spoken dialog system |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
US10897482B2 (en) | 2010-11-29 | 2021-01-19 | Biocatch Ltd. | Method, device, and system of back-coloring, forward-coloring, and fraud detection |
US10909331B2 (en) | 2018-03-30 | 2021-02-02 | Apple Inc. | Implicit identification of translation payload with neural machine translation |
US10917431B2 (en) | 2010-11-29 | 2021-02-09 | Biocatch Ltd. | System, method, and device of authenticating a user based on selfie image or selfie video |
US10928918B2 (en) | 2018-05-07 | 2021-02-23 | Apple Inc. | Raise to speak |
US10949757B2 (en) | 2010-11-29 | 2021-03-16 | Biocatch Ltd. | System, device, and method of detecting user identity based on motor-control loop model |
US10949514B2 (en) | 2010-11-29 | 2021-03-16 | Biocatch Ltd. | Device, system, and method of differentiating among users based on detection of hardware components |
US10970394B2 (en) | 2017-11-21 | 2021-04-06 | Biocatch Ltd. | System, device, and method of detecting vishing attacks |
US10984780B2 (en) | 2018-05-21 | 2021-04-20 | Apple Inc. | Global semantic word embeddings using bi-directional recurrent neural networks |
US11010127B2 (en) | 2015-06-29 | 2021-05-18 | Apple Inc. | Virtual assistant for media playback |
US11010561B2 (en) | 2018-09-27 | 2021-05-18 | Apple Inc. | Sentiment prediction from textual data |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US11023513B2 (en) | 2007-12-20 | 2021-06-01 | Apple Inc. | Method and apparatus for searching using an active ontology |
US11029838B2 (en) | 2006-09-06 | 2021-06-08 | Apple Inc. | Touch screen device, method, and graphical user interface for customizing display of content category icons |
US11055395B2 (en) | 2016-07-08 | 2021-07-06 | Biocatch Ltd. | Step-up authentication |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
US11145294B2 (en) | 2018-05-07 | 2021-10-12 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US20210329030A1 (en) * | 2010-11-29 | 2021-10-21 | Biocatch Ltd. | Device, System, and Method of Detecting Vishing Attacks |
US11170166B2 (en) | 2018-09-28 | 2021-11-09 | Apple Inc. | Neural typographical error modeling via generative adversarial networks |
US11181988B1 (en) | 2020-08-31 | 2021-11-23 | Apple Inc. | Incorporating user feedback into text prediction models via joint reward planning |
US11204787B2 (en) | 2017-01-09 | 2021-12-21 | Apple Inc. | Application integration with a digital assistant |
US11210674B2 (en) | 2010-11-29 | 2021-12-28 | Biocatch Ltd. | Method, device, and system of detecting mule accounts and accounts used for money laundering |
US11217251B2 (en) | 2019-05-06 | 2022-01-04 | Apple Inc. | Spoken notifications |
US11223619B2 (en) | 2010-11-29 | 2022-01-11 | Biocatch Ltd. | Device, system, and method of user authentication based on user-specific characteristics of task performance |
US11227589B2 (en) | 2016-06-06 | 2022-01-18 | Apple Inc. | Intelligent list reading |
US11231904B2 (en) | 2015-03-06 | 2022-01-25 | Apple Inc. | Reducing response latency of intelligent automated assistants |
US11237797B2 (en) | 2019-05-31 | 2022-02-01 | Apple Inc. | User activity shortcut suggestions |
US11269678B2 (en) | 2012-05-15 | 2022-03-08 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
US11269977B2 (en) | 2010-11-29 | 2022-03-08 | Biocatch Ltd. | System, apparatus, and method of collecting and processing data in electronic devices |
USRE48963E1 (en) | 2012-03-02 | 2022-03-08 | Microsoft Technology Licensing, Llc | Connection device for computing devices |
US11281993B2 (en) | 2016-12-05 | 2022-03-22 | Apple Inc. | Model and ensemble compression for metric learning |
US11289073B2 (en) | 2019-05-31 | 2022-03-29 | Apple Inc. | Device text to speech |
US11301477B2 (en) | 2017-05-12 | 2022-04-12 | Apple Inc. | Feedback analysis of a digital assistant |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
US11314370B2 (en) | 2013-12-06 | 2022-04-26 | Apple Inc. | Method for extracting salient dialog usage from live data |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
US11360641B2 (en) | 2019-06-01 | 2022-06-14 | Apple Inc. | Increasing the relevance of new available information |
US11386266B2 (en) | 2018-06-01 | 2022-07-12 | Apple Inc. | Text correction |
US11388291B2 (en) | 2013-03-14 | 2022-07-12 | Apple Inc. | System and method for processing voicemail |
US11423908B2 (en) | 2019-05-06 | 2022-08-23 | Apple Inc. | Interpreting spoken requests |
US11462215B2 (en) | 2018-09-28 | 2022-10-04 | Apple Inc. | Multi-modal inputs for voice commands |
US11468282B2 (en) | 2015-05-15 | 2022-10-11 | Apple Inc. | Virtual assistant in a communication session |
US11475884B2 (en) | 2019-05-06 | 2022-10-18 | Apple Inc. | Reducing digital assistant latency when a language is incorrectly determined |
US11475898B2 (en) | 2018-10-26 | 2022-10-18 | Apple Inc. | Low-latency multi-speaker speech recognition |
US11488406B2 (en) | 2019-09-25 | 2022-11-01 | Apple Inc. | Text detection using global geometry estimators |
US11495218B2 (en) | 2018-06-01 | 2022-11-08 | Apple Inc. | Virtual assistant operation in multi-device environments |
US11496600B2 (en) | 2019-05-31 | 2022-11-08 | Apple Inc. | Remote execution of machine-learned models |
US11532306B2 (en) | 2017-05-16 | 2022-12-20 | Apple Inc. | Detecting a trigger of a digital assistant |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US11606353B2 (en) | 2021-07-22 | 2023-03-14 | Biocatch Ltd. | System, device, and method of generating and utilizing one-time passwords |
US11638059B2 (en) | 2019-01-04 | 2023-04-25 | Apple Inc. | Content playback on multiple devices |
US11657813B2 (en) | 2019-05-31 | 2023-05-23 | Apple Inc. | Voice identification in digital assistant systems |
US11798547B2 (en) | 2013-03-15 | 2023-10-24 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
US20240080339A1 (en) * | 2010-11-29 | 2024-03-07 | Biocatch Ltd. | Device, System, and Method of Detecting Vishing Attacks |
Families Citing this family (156)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7750891B2 (en) | 2003-04-09 | 2010-07-06 | Tegic Communications, Inc. | Selective input system based on tracking of motion parameters of an input device |
US7286115B2 (en) | 2000-05-26 | 2007-10-23 | Tegic Communications, Inc. | Directional input system with automatic correction |
US7030863B2 (en) | 2000-05-26 | 2006-04-18 | America Online, Incorporated | Virtual keyboard system with automatic correction |
US7821503B2 (en) | 2003-04-09 | 2010-10-26 | Tegic Communications, Inc. | Touch screen and graphical user interface |
US7779364B2 (en) * | 2007-01-26 | 2010-08-17 | Kabushiki Kaisha Toshiba | System and method for generating a graphical user input interface via an associated display |
US8225203B2 (en) * | 2007-02-01 | 2012-07-17 | Nuance Communications, Inc. | Spell-check for a keyboard system with automatic correction |
US8201087B2 (en) | 2007-02-01 | 2012-06-12 | Tegic Communications, Inc. | Spell-check for a keyboard system with automatic correction |
US8065624B2 (en) * | 2007-06-28 | 2011-11-22 | Panasonic Corporation | Virtual keypad systems and methods |
US20100299710A1 (en) * | 2007-09-20 | 2010-11-25 | Samsung Electronics Co. Ltd. | Method for inputting user command and video apparatus and input apparatus employing the same |
KR101470413B1 (en) * | 2007-09-20 | 2014-12-10 | 삼성전자주식회사 | The method of inputting user command and the image apparatus and input apparatus thereof |
US8631358B2 (en) * | 2007-10-10 | 2014-01-14 | Apple Inc. | Variable device graphical user interface |
US20090102818A1 (en) * | 2007-10-22 | 2009-04-23 | Motorola, Inc. | Method and device for error-free keypad input |
US20090185080A1 (en) * | 2008-01-18 | 2009-07-23 | Imu Solutions, Inc. | Controlling an electronic device by changing an angular orientation of a remote wireless-controller |
US20090193361A1 (en) * | 2008-01-30 | 2009-07-30 | Research In Motion Limited | Electronic device and method of controlling same |
US9520743B2 (en) * | 2008-03-27 | 2016-12-13 | Echostar Technologies L.L.C. | Reduction of power consumption in remote control electronics |
US8612888B2 (en) | 2008-04-01 | 2013-12-17 | Litl, Llc | Method and apparatus for managing digital media content |
US8624844B2 (en) | 2008-04-01 | 2014-01-07 | Litl Llc | Portable computer with multiple display configurations |
US9003315B2 (en) | 2008-04-01 | 2015-04-07 | Litl Llc | System and method for streamlining user interaction with electronic content |
US7907060B2 (en) * | 2008-05-08 | 2011-03-15 | Echostar Technologies L.L.C. | Systems, methods and apparatus for detecting replacement of a battery in a remote control |
CN101598966A (en) * | 2008-06-06 | 2009-12-09 | 深圳富泰宏精密工业有限公司 | The contactor control device and the control method thereof that show two pictures |
US8305249B2 (en) | 2008-07-18 | 2012-11-06 | EchoStar Technologies, L.L.C. | Systems and methods for controlling power consumption in electronic devices |
FR2934067B1 (en) * | 2008-07-21 | 2013-01-25 | Dav | HAPTICALLY RETURN CONTROL DEVICE AND CORRESPONDING CONTROL METHOD |
US20100020036A1 (en) * | 2008-07-23 | 2010-01-28 | Edward Hui | Portable electronic device and method of controlling same |
TWI360762B (en) * | 2008-09-05 | 2012-03-21 | Mitake Information Corp | On-screen virtual keyboard system |
US8750938B2 (en) | 2008-09-29 | 2014-06-10 | Microsoft Corporation | Glow touch feedback for virtual input devices |
US8786551B2 (en) * | 2008-12-11 | 2014-07-22 | Verizon Patent And Licensing Inc. | Predictive keypad/keyboard lighting to guide input |
US20100164756A1 (en) * | 2008-12-30 | 2010-07-01 | Nokia Corporation | Electronic device user input |
US20110276528A1 (en) * | 2009-01-20 | 2011-11-10 | Kenta Shioiri | Input device, information processing device, input method, and program |
FR2941546B1 (en) * | 2009-01-27 | 2011-02-04 | Wyplay | OPTIMIZED METHOD FOR ENTERING ALPHANUMERIC CHARACTERS FROM A REMOTE CONTROL |
US8326358B2 (en) | 2009-01-30 | 2012-12-04 | Research In Motion Limited | System and method for access control in a portable electronic device |
SG177239A1 (en) * | 2009-02-04 | 2012-02-28 | Keyless Systems Ltd | Data entry system |
US8134475B2 (en) | 2009-03-16 | 2012-03-13 | Echostar Technologies L.L.C. | Backlighting remote controls |
US20100245268A1 (en) * | 2009-03-30 | 2010-09-30 | Stg Interactive S.A. | User-friendly process for interacting with informational content on touchscreen devices |
GB201108200D0 (en) | 2011-05-16 | 2011-06-29 | Touchtype Ltd | User input prediction |
GB201016385D0 (en) * | 2010-09-29 | 2010-11-10 | Touchtype Ltd | System and method for inputting text into electronic devices |
KR20100118366A (en) * | 2009-04-28 | 2010-11-05 | 삼성전자주식회사 | Operating method of touch screen and portable device including the same |
US9043718B2 (en) * | 2009-06-05 | 2015-05-26 | Blackberry Limited | System and method for applying a text prediction algorithm to a virtual keyboard |
US20100310136A1 (en) * | 2009-06-09 | 2010-12-09 | Sony Ericsson Mobile Communications Ab | Distinguishing right-hand input and left-hand input based on finger recognition |
EP2264563A1 (en) * | 2009-06-19 | 2010-12-22 | Tegic Communications, Inc. | Virtual keyboard system with automatic correction |
EP2284653A1 (en) * | 2009-08-14 | 2011-02-16 | Research In Motion Limited | Electronic device with touch-sensitive display and method of facilitating input at the electronic device |
US20110041056A1 (en) * | 2009-08-14 | 2011-02-17 | Research In Motion Limited | Electronic device with touch-sensitive display and method of facilitating input at the electronic device |
US20110077470A1 (en) * | 2009-09-30 | 2011-03-31 | Nellcor Puritan Bennett Llc | Patient Monitor Symmetry Control |
US8261211B2 (en) * | 2009-10-01 | 2012-09-04 | Microsoft Corporation | Monitoring pointer trajectory and modifying display interface |
US8347221B2 (en) * | 2009-10-07 | 2013-01-01 | Research In Motion Limited | Touch-sensitive display and method of control |
KR101058525B1 (en) | 2009-10-09 | 2011-08-23 | 삼성전자주식회사 | Text input method and display device using the same |
KR101595029B1 (en) * | 2009-11-18 | 2016-02-17 | 엘지전자 주식회사 | Mobile terminal and method for controlling the same |
US20110154260A1 (en) * | 2009-12-17 | 2011-06-23 | Motorola Inc | Method and apparatus for displaying information in an electronic device |
JP5676113B2 (en) * | 2010-01-27 | 2015-02-25 | 京セラ株式会社 | Portable electronic devices |
JP5635274B2 (en) | 2010-01-27 | 2014-12-03 | 京セラ株式会社 | Tactile sensation presentation apparatus and tactile sensation presentation method |
EP2369445B1 (en) * | 2010-02-26 | 2016-08-17 | BlackBerry Limited | Electronic device with touch-sensitive display and method of facilitating input at the electronic device |
US8456435B2 (en) | 2010-02-26 | 2013-06-04 | Research In Motion Limited | Electronic device with touch-sensitive display and method of facilitating input at the electronic device |
EP2407859A1 (en) * | 2010-07-16 | 2012-01-18 | Gigaset Communications GmbH | Dynamic adjustment of a user interface on a sensor monitor |
KR101701932B1 (en) * | 2010-07-22 | 2017-02-13 | 삼성전자 주식회사 | Input device and control method of thereof |
TWI427507B (en) * | 2010-07-29 | 2014-02-21 | Hon Hai Prec Ind Co Ltd | Handheld device and method for inputting characters thereof |
US9639265B2 (en) * | 2010-09-03 | 2017-05-02 | Microsoft Technology Licensing, Llc | Distance-time based hit-testing for displayed target graphical elements |
GB201200643D0 (en) | 2012-01-16 | 2012-02-29 | Touchtype Ltd | System and method for inputting text |
JP5782699B2 (en) * | 2010-10-15 | 2015-09-24 | ソニー株式会社 | Information processing apparatus, input control method for information processing apparatus, and program |
US20120131490A1 (en) * | 2010-11-22 | 2012-05-24 | Shao-Chieh Lin | Touch-controlled device and method for displaying a virtual keyboard on the touch-controlled device thereof |
TW201222396A (en) * | 2010-11-22 | 2012-06-01 | Acer Inc | Touch-controlled apparatus and method for displaying a virtual keyboard on the touch-controlled apparatus thereof |
KR101821160B1 (en) * | 2010-12-10 | 2018-03-08 | 삼성전자 주식회사 | Method and apparatus for providing user keypad in a portable terminal |
US9354804B2 (en) | 2010-12-29 | 2016-05-31 | Microsoft Technology Licensing, Llc | Touch event anticipation in a computing device |
US9201185B2 (en) | 2011-02-04 | 2015-12-01 | Microsoft Technology Licensing, Llc | Directional backlighting for display panels |
US11087659B2 (en) | 2011-03-11 | 2021-08-10 | Cox Communications, Inc. | System, method and device for presenting different functional displays when orientation of the device changes |
WO2013024317A1 (en) | 2011-08-15 | 2013-02-21 | Telefonaktiebolaget Lm Ericsson (Publ) | Resizing selection zones on a touch sensitive display responsive to likelihood of selection |
US9262076B2 (en) * | 2011-09-12 | 2016-02-16 | Microsoft Technology Licensing, Llc | Soft keyboard interface |
KR20130034765A (en) * | 2011-09-29 | 2013-04-08 | 삼성전자주식회사 | Method and device for inputting of mobile terminal using a pen |
US8490008B2 (en) | 2011-11-10 | 2013-07-16 | Research In Motion Limited | Touchscreen keyboard predictive display and generation of a set of characters |
US9310889B2 (en) | 2011-11-10 | 2016-04-12 | Blackberry Limited | Touchscreen keyboard predictive display and generation of a set of characters |
US9122672B2 (en) | 2011-11-10 | 2015-09-01 | Blackberry Limited | In-letter word prediction for virtual keyboard |
US9652448B2 (en) | 2011-11-10 | 2017-05-16 | Blackberry Limited | Methods and systems for removing or replacing on-keyboard prediction candidates |
US9715489B2 (en) | 2011-11-10 | 2017-07-25 | Blackberry Limited | Displaying a prediction candidate after a typing mistake |
KR20130054579A (en) * | 2011-11-17 | 2013-05-27 | 삼성전자주식회사 | Display apparatus and control method thereof |
JP5413448B2 (en) * | 2011-12-23 | 2014-02-12 | 株式会社デンソー | Display system, display device, and operation device |
US10423328B2 (en) * | 2011-12-28 | 2019-09-24 | Hiroyuki Ikeda | Portable terminal for controlling two cursors within a virtual keyboard according to setting of movement by a single key at a time or a plurality of keys at a time |
US9557913B2 (en) | 2012-01-19 | 2017-01-31 | Blackberry Limited | Virtual keyboard display having a ticker proximate to the virtual keyboard |
US9152323B2 (en) | 2012-01-19 | 2015-10-06 | Blackberry Limited | Virtual keyboard providing an indication of received input |
US9052414B2 (en) | 2012-02-07 | 2015-06-09 | Microsoft Technology Licensing, Llc | Virtual image device |
US9244612B1 (en) * | 2012-02-16 | 2016-01-26 | Google Inc. | Key selection of a graphical keyboard based on user input posture |
CA2865272C (en) | 2012-02-24 | 2019-11-05 | Blackberry Limited | Virtual keyboard with dynamically reconfigurable layout |
DE112012000189B4 (en) * | 2012-02-24 | 2023-06-15 | Blackberry Limited | Touch screen keyboard for providing word predictions in partitions of the touch screen keyboard in close association with candidate letters |
US8749529B2 (en) | 2012-03-01 | 2014-06-10 | Microsoft Corporation | Sensor-in-pixel display system with near infrared filter |
WO2013133478A1 (en) | 2012-03-04 | 2013-09-12 | Lg Electronics Inc. | Portable device and control method thereof |
US9201510B2 (en) | 2012-04-16 | 2015-12-01 | Blackberry Limited | Method and device having touchscreen keyboard with visual cues |
US9354805B2 (en) | 2012-04-30 | 2016-05-31 | Blackberry Limited | Method and apparatus for text selection |
US10025487B2 (en) | 2012-04-30 | 2018-07-17 | Blackberry Limited | Method and apparatus for text selection |
US9292192B2 (en) | 2012-04-30 | 2016-03-22 | Blackberry Limited | Method and apparatus for text selection |
US20130285927A1 (en) * | 2012-04-30 | 2013-10-31 | Research In Motion Limited | Touchscreen keyboard with correction of previously input text |
KR101901611B1 (en) * | 2012-05-09 | 2018-09-27 | 엘지전자 주식회사 | Mobile terminal and control method for the mobile terminal |
US8884881B2 (en) | 2012-05-11 | 2014-11-11 | Blackberry Limited | Portable electronic device and method of controlling same |
EP2662752B1 (en) * | 2012-05-11 | 2017-09-13 | BlackBerry Limited | Apparatus and method for character entry in a portable electronic device |
US9207860B2 (en) | 2012-05-25 | 2015-12-08 | Blackberry Limited | Method and apparatus for detecting a gesture |
US8947353B2 (en) | 2012-06-12 | 2015-02-03 | Microsoft Corporation | Photosensor array gesture detection |
US9684382B2 (en) | 2012-06-13 | 2017-06-20 | Microsoft Technology Licensing, Llc | Input device configuration having capacitive and pressure sensors |
US9459160B2 (en) | 2012-06-13 | 2016-10-04 | Microsoft Technology Licensing, Llc | Input device sensor configuration |
US9073123B2 (en) | 2012-06-13 | 2015-07-07 | Microsoft Technology Licensing, Llc | Housing vents |
US9063693B2 (en) | 2012-06-13 | 2015-06-23 | Microsoft Technology Licensing, Llc | Peripheral device storage |
JP6071107B2 (en) | 2012-06-14 | 2017-02-01 | 裕行 池田 | Mobile device |
US9256089B2 (en) | 2012-06-15 | 2016-02-09 | Microsoft Technology Licensing, Llc | Object-detecting backlight unit |
US20130346904A1 (en) * | 2012-06-26 | 2013-12-26 | International Business Machines Corporation | Targeted key press zones on an interactive display |
US9116552B2 (en) | 2012-06-27 | 2015-08-25 | Blackberry Limited | Touchscreen keyboard providing selection of word predictions in partitions of the touchscreen keyboard |
US9355345B2 (en) | 2012-07-23 | 2016-05-31 | Microsoft Technology Licensing, Llc | Transparent tags with encoded data |
US9298295B2 (en) * | 2012-07-25 | 2016-03-29 | Facebook, Inc. | Gestures for auto-correct |
US9836213B2 (en) * | 2012-07-27 | 2017-12-05 | Symbol Technologies, Llc | Enhanced user interface for pressure sensitive touch screen |
US9524290B2 (en) | 2012-08-31 | 2016-12-20 | Blackberry Limited | Scoring predictions based on prediction length and typing speed |
US9063653B2 (en) | 2012-08-31 | 2015-06-23 | Blackberry Limited | Ranking predictions based on typing speed and typing confidence |
US20140063067A1 (en) * | 2012-08-31 | 2014-03-06 | Research In Motion Limited | Method to select word by swiping capacitive keyboard |
US9471220B2 (en) | 2012-09-18 | 2016-10-18 | Google Inc. | Posture-adaptive selection |
US9152173B2 (en) | 2012-10-09 | 2015-10-06 | Microsoft Technology Licensing, Llc | Transparent display device |
US8654030B1 (en) | 2012-10-16 | 2014-02-18 | Microsoft Corporation | Antenna placement |
US8983849B2 (en) | 2012-10-17 | 2015-03-17 | Nuance Communications, Inc. | Multiple device intelligent language model synchronization |
CN104903026B (en) | 2012-10-17 | 2017-10-24 | 微软技术许可有限责任公司 | Metal alloy injection is molded overfall |
WO2014059624A1 (en) | 2012-10-17 | 2014-04-24 | Microsoft Corporation | Metal alloy injection molding protrusions |
WO2014059618A1 (en) | 2012-10-17 | 2014-04-24 | Microsoft Corporation | Graphic formation via material ablation |
US8704792B1 (en) | 2012-10-19 | 2014-04-22 | Google Inc. | Density-based filtering of gesture events associated with a user interface of a computing device |
US9304595B2 (en) * | 2012-10-19 | 2016-04-05 | Google Inc. | Gesture-keyboard decoding using gesture path deviation |
US8952892B2 (en) | 2012-11-01 | 2015-02-10 | Microsoft Corporation | Input location correction tables for input panels |
CN103809898A (en) * | 2012-11-14 | 2014-05-21 | 宇瞻科技股份有限公司 | Intelligent input method |
US9411510B2 (en) * | 2012-12-07 | 2016-08-09 | Apple Inc. | Techniques for preventing typographical errors on soft keyboards |
US9513748B2 (en) | 2012-12-13 | 2016-12-06 | Microsoft Technology Licensing, Llc | Combined display panel circuit |
US20140208258A1 (en) * | 2013-01-22 | 2014-07-24 | Jenny Yuen | Predictive Input Using Custom Dictionaries |
US9176538B2 (en) | 2013-02-05 | 2015-11-03 | Microsoft Technology Licensing, Llc | Input device configurations |
US10578499B2 (en) | 2013-02-17 | 2020-03-03 | Microsoft Technology Licensing, Llc | Piezo-actuated virtual buttons for touch surfaces |
US9638835B2 (en) | 2013-03-05 | 2017-05-02 | Microsoft Technology Licensing, Llc | Asymmetric aberration correcting lens |
US10120540B2 (en) * | 2013-03-14 | 2018-11-06 | Samsung Electronics Co., Ltd. | Visual feedback for user interface navigation on television system |
US9348429B2 (en) | 2013-03-15 | 2016-05-24 | Blackberry Limited | Method and apparatus for word prediction using the position of a non-typing digit |
EP2778859B1 (en) | 2013-03-15 | 2015-07-08 | BlackBerry Limited | Method and apparatus for word prediction using the position of a non-typing digit |
US9552777B2 (en) | 2013-05-10 | 2017-01-24 | Microsoft Technology Licensing, Llc | Phase control backlight |
US9454251B1 (en) * | 2013-06-26 | 2016-09-27 | Google Inc. | Methods, systems, and media for controlling a remote device using a touch screen of a mobile device in a display inhibited state |
FR3014572B1 (en) * | 2013-12-05 | 2016-01-01 | Op3Ft | METHOD FOR CONTROLLING INTERACTION WITH A TOUCH SCREEN AND EQUIPMENT USING THE SAME |
US9448631B2 (en) | 2013-12-31 | 2016-09-20 | Microsoft Technology Licensing, Llc | Input device haptics and pressure sensing |
CN114895838A (en) * | 2014-01-06 | 2022-08-12 | 华为终端有限公司 | Application program display method and terminal |
US9317072B2 (en) | 2014-01-28 | 2016-04-19 | Microsoft Technology Licensing, Llc | Hinge mechanism with preset positions |
US9759854B2 (en) | 2014-02-17 | 2017-09-12 | Microsoft Technology Licensing, Llc | Input device outer layer and backlighting |
US10440001B2 (en) * | 2014-06-18 | 2019-10-08 | Dell Products, Lp | Method to securely authenticate management server over un-encrypted remote console connection |
US9477653B2 (en) | 2014-06-26 | 2016-10-25 | Blackberry Limited | Character entry for an electronic device using a position sensing keyboard |
US9930167B2 (en) * | 2014-07-07 | 2018-03-27 | Verizon Patent And Licensing Inc. | Messaging application with in-application search functionality |
US20160034131A1 (en) * | 2014-07-31 | 2016-02-04 | Sony Corporation | Methods and systems of a graphical user interface shift |
US9513671B2 (en) | 2014-08-01 | 2016-12-06 | Microsoft Technology Licensing, Llc | Peripheral retention device |
US10534532B2 (en) | 2014-08-08 | 2020-01-14 | Samsung Electronics Co., Ltd. | Electronic device and method for processing letter input in electronic device |
US10191986B2 (en) | 2014-08-11 | 2019-01-29 | Microsoft Technology Licensing, Llc | Web resource compatibility with web applications |
US9705637B2 (en) | 2014-08-19 | 2017-07-11 | Microsoft Technology Licensing, Llc | Guard band utilization for wireless data communication |
US9397723B2 (en) | 2014-08-26 | 2016-07-19 | Microsoft Technology Licensing, Llc | Spread spectrum wireless over non-contiguous channels |
US9727231B2 (en) | 2014-11-19 | 2017-08-08 | Honda Motor Co., Ltd. | System and method for providing absolute coordinate and zone mapping between a touchpad and a display screen |
US20170371515A1 (en) | 2014-11-19 | 2017-12-28 | Honda Motor Co., Ltd. | System and method for providing absolute and zone coordinate mapping with graphic animations |
US20180024707A1 (en) * | 2015-03-13 | 2018-01-25 | Kyocera Document Solutions Inc. | Information processing device and screen display method |
CN104850239B (en) * | 2015-04-23 | 2018-06-19 | 百度在线网络技术(北京)有限公司 | A kind of method and apparatus for being used to determine keyboard display state |
US10222889B2 (en) | 2015-06-03 | 2019-03-05 | Microsoft Technology Licensing, Llc | Force inputs and cursor control |
US10416799B2 (en) | 2015-06-03 | 2019-09-17 | Microsoft Technology Licensing, Llc | Force sensing and inadvertent input control of an input device |
US10061385B2 (en) | 2016-01-22 | 2018-08-28 | Microsoft Technology Licensing, Llc | Haptic feedback for a touch input device |
CN108811514B (en) * | 2017-03-03 | 2022-03-01 | 谷歌有限责任公司 | System and method for detecting that an application executing on a client device improperly implements presentation of a content item |
KR20190065746A (en) * | 2017-12-04 | 2019-06-12 | 삼성전자주식회사 | Electronic apparatus, method for controlling thereof and the computer readable recording medium |
US11554322B2 (en) * | 2019-04-26 | 2023-01-17 | Sony Interactive Entertainment LLC | Game controller with touchpad input |
CN111866565B (en) | 2019-04-26 | 2022-09-02 | 三星电子株式会社 | Display device and control method of display device |
USD951286S1 (en) * | 2019-11-19 | 2022-05-10 | Johnson Systems Inc. | Display screen with graphical user interface |
EP3832434A1 (en) * | 2019-12-03 | 2021-06-09 | XRSpace CO., LTD. | Behavior-based configuration method and behavior-based configuration system |
CN113448465A (en) * | 2021-06-29 | 2021-09-28 | 腾讯科技(深圳)有限公司 | Virtual cursor control method and device, storage medium and electronic equipment |
Citations (77)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5038401A (en) * | 1989-04-05 | 1991-08-06 | Pioneer Electronic Corporation | Transmitter for remote control with operation switches having changeably displayed forms |
US5128672A (en) * | 1990-10-30 | 1992-07-07 | Apple Computer, Inc. | Dynamic predictive keyboard |
US5652630A (en) * | 1995-05-31 | 1997-07-29 | International Business Machines Corporation | Video receiver display, three axis remote control, and microcontroller for executing programs |
US5724106A (en) * | 1995-07-17 | 1998-03-03 | Gateway 2000, Inc. | Hand held remote control device with trigger button |
US5936614A (en) * | 1991-04-30 | 1999-08-10 | International Business Machines Corporation | User defined keyboard entry system |
US5963671A (en) * | 1991-11-27 | 1999-10-05 | International Business Machines Corporation | Enhancement of soft keyboard operations using trigram prediction |
US6222528B1 (en) * | 1997-03-07 | 2001-04-24 | Cirque Corporation | Method and apparatus for data input |
US6359270B1 (en) * | 1998-09-04 | 2002-03-19 | Ncr Corporation | Communications module mounting for domestic appliance |
US6396523B1 (en) * | 1999-07-29 | 2002-05-28 | Interlink Electronics, Inc. | Home entertainment device remote control |
US20020093535A1 (en) * | 2001-01-17 | 2002-07-18 | Murphy Michael William | User interface for character entry using a minimum number of selection keys |
US20020114654A1 (en) * | 2001-02-16 | 2002-08-22 | Toshiyasu Abe | Improved Keyboard |
US20020122197A1 (en) * | 2001-02-21 | 2002-09-05 | Eli Abir | Method and apparatus for text input |
US20020149569A1 (en) * | 2001-04-12 | 2002-10-17 | International Business Machines Corporation | Touchscreen user interface |
US20030080945A1 (en) * | 2001-10-29 | 2003-05-01 | Betts-Lacroix Jonathan | Keyboard with variable-sized keys |
US20030193478A1 (en) * | 2002-04-04 | 2003-10-16 | Edwin Ng | Reduced keyboard system that emulates QWERTY-type mapping and typing |
US20030197736A1 (en) * | 2002-01-16 | 2003-10-23 | Murphy Michael W. | User interface for character entry using a minimum number of selection keys |
US20040041723A1 (en) * | 2002-05-27 | 2004-03-04 | Canon Kabushiki Kaisha, Tokyo, Japan | Remote control device |
US20040064351A1 (en) * | 1999-11-22 | 2004-04-01 | Mikurak Michael G. | Increased visibility during order management in a network-based supply chain environment |
US20040080487A1 (en) * | 2002-10-29 | 2004-04-29 | Griffin Jason T. | Electronic device having keyboard for thumb typing |
US6765557B1 (en) * | 2000-04-10 | 2004-07-20 | Interlink Electronics, Inc. | Remote control having touch pad to screen mapping |
US20040140956A1 (en) * | 2003-01-16 | 2004-07-22 | Kushler Clifford A. | System and method for continuous stroke word-based text input |
US20040169635A1 (en) * | 2001-07-12 | 2004-09-02 | Ghassabian Benjamin Firooz | Features to enhance data entry through a small data entry unit |
US20040239624A1 (en) * | 2003-04-02 | 2004-12-02 | Artoun Ramian | Freehand symbolic input apparatus and method |
US20050052406A1 (en) * | 2003-04-09 | 2005-03-10 | James Stephanick | Selective input system based on tracking of motion parameters of an input device |
US20050099400A1 (en) * | 2003-11-06 | 2005-05-12 | Samsung Electronics Co., Ltd. | Apparatus and method for providing vitrtual graffiti and recording medium for the same |
US20050141770A1 (en) * | 2003-12-30 | 2005-06-30 | Nokia Corporation | Split on-screen keyboard |
US20050162402A1 (en) * | 2004-01-27 | 2005-07-28 | Watanachote Susornpol J. | Methods of interacting with a computer using a finger(s) touch sensing input device with visual feedback |
US20050195159A1 (en) * | 2004-02-23 | 2005-09-08 | Hunleth Frank A. | Keyboardless text entry |
US20050253814A1 (en) * | 1999-10-27 | 2005-11-17 | Firooz Ghassabian | Integrated keypad system |
US7002553B2 (en) * | 2001-12-27 | 2006-02-21 | Mark Shkolnikov | Active keyboard system for handheld electronic devices |
US20060119588A1 (en) * | 2004-12-03 | 2006-06-08 | Sung-Min Yoon | Apparatus and method of processing information input using a touchpad |
US20060152496A1 (en) * | 2005-01-13 | 2006-07-13 | 602531 British Columbia Ltd. | Method, system, apparatus and computer-readable media for directing input associated with keyboard-type device |
US20060197750A1 (en) * | 2005-03-04 | 2006-09-07 | Apple Computer, Inc. | Hand held electronic device with multiple touch sensing devices |
US20060227032A1 (en) * | 2005-04-11 | 2006-10-12 | Alberto Vidal | Apparatus and method to facilitate universal remote control |
US20060238510A1 (en) * | 2005-04-25 | 2006-10-26 | Georgios Panotopoulos | User interface incorporating emulated hard keys |
US7139983B2 (en) * | 2000-04-10 | 2006-11-21 | Hillcrest Laboratories, Inc. | Interactive content guide for television programming |
US20070040813A1 (en) * | 2003-01-16 | 2007-02-22 | Forword Input, Inc. | System and method for continuous stroke word-based text input |
US20070046641A1 (en) * | 2005-09-01 | 2007-03-01 | Swee Ho Lim | Entering a character into an electronic device |
US20070060176A1 (en) * | 2005-09-13 | 2007-03-15 | Microsoft Corporation | Type-ahead keypad input |
US20070061754A1 (en) * | 2005-08-26 | 2007-03-15 | Veveo, Inc. | User interface for visual cooperation between text input and display device |
US20070079239A1 (en) * | 2000-10-27 | 2007-04-05 | Firooz Ghassabian | Data entry system |
US20070080954A1 (en) * | 2005-10-07 | 2007-04-12 | Research In Motion Limited | System and method for using navigational and other commands on a mobile communication device |
US20070082708A1 (en) * | 2005-10-07 | 2007-04-12 | Research In Motion Limited | Device, system, and method for informing users of functions and characters associated with telephone keys |
US20070132887A1 (en) * | 2005-12-09 | 2007-06-14 | David Boyden | On screen display for alpha-numeric input |
US20070182595A1 (en) * | 2004-06-04 | 2007-08-09 | Firooz Ghasabian | Systems to enhance data entry in mobile and fixed environment |
US20070188472A1 (en) * | 2003-04-18 | 2007-08-16 | Ghassabian Benjamin F | Systems to enhance data entry in mobile and fixed environment |
US20070195353A1 (en) * | 2006-02-17 | 2007-08-23 | Canon Kabushiki Kaisha | Image forming apparatus |
US20080007434A1 (en) * | 2006-07-10 | 2008-01-10 | Luben Hristov | Priority and Combination Suppression Techniques (PST/CST) for a Capacitive Keyboard |
US20080168364A1 (en) * | 2007-01-05 | 2008-07-10 | Apple Computer, Inc. | Adaptive acceleration of mouse cursor |
US7453439B1 (en) * | 2003-01-16 | 2008-11-18 | Forward Input Inc. | System and method for continuous stroke word-based text input |
US20080316212A1 (en) * | 2005-09-20 | 2008-12-25 | Cliff Kushler | System and method for a user interface for text editing and menu selection |
US20090007001A1 (en) * | 2007-06-28 | 2009-01-01 | Matsushita Electric Industrial Co., Ltd. | Virtual keypad systems and methods |
US20090135162A1 (en) * | 2005-03-10 | 2009-05-28 | Koninklijke Philips Electronics, N.V. | System and Method For Detecting the Location, Size and Shape of Multiple Objects That Interact With a Touch Screen Display |
US7649478B1 (en) * | 2005-11-03 | 2010-01-19 | Hyoungsoo Yoon | Data entry using sequential keystrokes |
US20100103127A1 (en) * | 2007-02-23 | 2010-04-29 | Taeun Park | Virtual Keyboard Input System Using Pointing Apparatus In Digital Device |
US20100245395A1 (en) * | 2009-03-31 | 2010-09-30 | Dell Products L.P. | Information Handling System with Touch Pad and Point Stick |
US7817145B2 (en) * | 2008-04-18 | 2010-10-19 | Cirque Corporation | Method and system for performing scrolling by movement of a pointing object in a curvilinear path on a touchpad |
US20100265200A1 (en) * | 2009-04-21 | 2010-10-21 | Samsung Electronics Co., Ltd. | Method for processing multi-touch input in mobile device |
US7957955B2 (en) * | 2007-01-05 | 2011-06-07 | Apple Inc. | Method and system for providing word recommendations for text input |
US7991784B2 (en) * | 2005-01-21 | 2011-08-02 | Prashant Parikh | Automatic dynamic contextual data entry completion system |
US8036878B2 (en) * | 2005-05-18 | 2011-10-11 | Never Wall Treuhand GmbH | Device incorporating improved text input mechanism |
US8060839B2 (en) * | 2006-06-30 | 2011-11-15 | Samsung Electronics Co., Ltd | Character input method and mobile communication terminal using the same |
US20110279354A1 (en) * | 2010-05-12 | 2011-11-17 | Aaron Tang | Computer with tv mode |
US20110279397A1 (en) * | 2009-01-26 | 2011-11-17 | Zrro Technologies (2009) Ltd. | Device and method for monitoring the object's behavior |
US20110279376A1 (en) * | 2010-05-12 | 2011-11-17 | Aaron Tang | Remote control to operate computer system |
US8065453B2 (en) * | 2006-04-06 | 2011-11-22 | Research In Motion Limited | Handheld electronic device and associated method employing a multiple-axis input device and learning a context of a text input for use by a disambiguation routine |
US8078978B2 (en) * | 2007-10-19 | 2011-12-13 | Google Inc. | Method and system for predicting text |
US8078884B2 (en) * | 2006-11-13 | 2011-12-13 | Veveo, Inc. | Method of and system for selecting and presenting content based on user identification |
US8086602B2 (en) * | 2006-04-20 | 2011-12-27 | Veveo Inc. | User interface methods and systems for selecting and presenting content based on user navigation and selection actions associated with the content |
US8112454B2 (en) * | 2006-03-06 | 2012-02-07 | Veveo, Inc. | Methods and systems for ordering content items according to learned user preferences |
US8117540B2 (en) * | 2005-05-18 | 2012-02-14 | Neuer Wall Treuhand Gmbh | Method and device incorporating improved text input mechanism |
US20120117101A1 (en) * | 2010-11-10 | 2012-05-10 | Erland Unruh | Text entry with word prediction, completion, or correction supplemented by search of shared corpus |
US8296484B2 (en) * | 2006-03-30 | 2012-10-23 | Harris Corporation | Alphanumeric data entry apparatus and method using multicharacter keys of a keypad |
US8296294B2 (en) * | 2007-05-25 | 2012-10-23 | Veveo, Inc. | Method and system for unified searching across and within multiple documents |
US8311829B2 (en) * | 2004-06-02 | 2012-11-13 | Tegic Communications, Inc. | Multimodal disambiguation of speech recognition |
US8316035B2 (en) * | 2008-01-16 | 2012-11-20 | International Business Machines Corporation | Systems and arrangements of text type-ahead |
US8330625B2 (en) * | 2007-03-30 | 2012-12-11 | Honeywell International Inc. | Aircraft systems with flight management systems that display reports |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH08221169A (en) | 1995-02-14 | 1996-08-30 | Matsushita Electric Ind Co Ltd | Method and device for displaying software keyboard |
JPH09160910A (en) * | 1995-12-04 | 1997-06-20 | Matsushita Electric Ind Co Ltd | Software keyboard display system |
JPH09251347A (en) | 1996-03-15 | 1997-09-22 | Matsushita Electric Ind Co Ltd | Coordinate input device |
JP4459103B2 (en) * | 2005-04-13 | 2010-04-28 | シャープ株式会社 | Information terminal device, information input method, information input program, and recording medium |
DE602006019893D1 (en) * | 2006-12-29 | 2011-03-10 | Research In Motion Ltd | Portable electronic device with input confirmation and corresponding method |
-
2007
- 2007-10-24 US US11/977,346 patent/US8065624B2/en active Active
-
2008
- 2008-06-26 JP JP2010515128A patent/JP5323070B2/en active Active
- 2008-06-26 US US12/666,916 patent/US20100164897A1/en not_active Abandoned
- 2008-06-26 EP EP08772060A patent/EP2160672A4/en not_active Withdrawn
- 2008-06-26 WO PCT/US2008/068384 patent/WO2009006221A1/en active Search and Examination
Patent Citations (97)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5038401A (en) * | 1989-04-05 | 1991-08-06 | Pioneer Electronic Corporation | Transmitter for remote control with operation switches having changeably displayed forms |
US5128672A (en) * | 1990-10-30 | 1992-07-07 | Apple Computer, Inc. | Dynamic predictive keyboard |
US5936614A (en) * | 1991-04-30 | 1999-08-10 | International Business Machines Corporation | User defined keyboard entry system |
US5963671A (en) * | 1991-11-27 | 1999-10-05 | International Business Machines Corporation | Enhancement of soft keyboard operations using trigram prediction |
US5652630A (en) * | 1995-05-31 | 1997-07-29 | International Business Machines Corporation | Video receiver display, three axis remote control, and microcontroller for executing programs |
US5724106A (en) * | 1995-07-17 | 1998-03-03 | Gateway 2000, Inc. | Hand held remote control device with trigger button |
US6222528B1 (en) * | 1997-03-07 | 2001-04-24 | Cirque Corporation | Method and apparatus for data input |
US6359270B1 (en) * | 1998-09-04 | 2002-03-19 | Ncr Corporation | Communications module mounting for domestic appliance |
US6396523B1 (en) * | 1999-07-29 | 2002-05-28 | Interlink Electronics, Inc. | Home entertainment device remote control |
US20090037623A1 (en) * | 1999-10-27 | 2009-02-05 | Firooz Ghassabian | Integrated keypad system |
US7020270B1 (en) * | 1999-10-27 | 2006-03-28 | Firooz Ghassabian | Integrated keypad system |
US20050253814A1 (en) * | 1999-10-27 | 2005-11-17 | Firooz Ghassabian | Integrated keypad system |
US8271336B2 (en) * | 1999-11-22 | 2012-09-18 | Accenture Global Services Gmbh | Increased visibility during order management in a network-based supply chain environment |
US20040064351A1 (en) * | 1999-11-22 | 2004-04-01 | Mikurak Michael G. | Increased visibility during order management in a network-based supply chain environment |
US6765557B1 (en) * | 2000-04-10 | 2004-07-20 | Interlink Electronics, Inc. | Remote control having touch pad to screen mapping |
US7139983B2 (en) * | 2000-04-10 | 2006-11-21 | Hillcrest Laboratories, Inc. | Interactive content guide for television programming |
US20070079239A1 (en) * | 2000-10-27 | 2007-04-05 | Firooz Ghassabian | Data entry system |
US20020093535A1 (en) * | 2001-01-17 | 2002-07-18 | Murphy Michael William | User interface for character entry using a minimum number of selection keys |
US20020114654A1 (en) * | 2001-02-16 | 2002-08-22 | Toshiyasu Abe | Improved Keyboard |
US20020122197A1 (en) * | 2001-02-21 | 2002-09-05 | Eli Abir | Method and apparatus for text input |
US6692170B2 (en) * | 2001-02-21 | 2004-02-17 | Eli Abir | Method and apparatus for text input |
US20020149569A1 (en) * | 2001-04-12 | 2002-10-17 | International Business Machines Corporation | Touchscreen user interface |
US20040169635A1 (en) * | 2001-07-12 | 2004-09-02 | Ghassabian Benjamin Firooz | Features to enhance data entry through a small data entry unit |
US7372454B2 (en) * | 2001-10-29 | 2008-05-13 | Oqo Incorporated | Keyboard with variable-sized keys |
US20030080945A1 (en) * | 2001-10-29 | 2003-05-01 | Betts-Lacroix Jonathan | Keyboard with variable-sized keys |
US7002553B2 (en) * | 2001-12-27 | 2006-02-21 | Mark Shkolnikov | Active keyboard system for handheld electronic devices |
US20030197736A1 (en) * | 2002-01-16 | 2003-10-23 | Murphy Michael W. | User interface for character entry using a minimum number of selection keys |
US20030193478A1 (en) * | 2002-04-04 | 2003-10-16 | Edwin Ng | Reduced keyboard system that emulates QWERTY-type mapping and typing |
US7202853B2 (en) * | 2002-04-04 | 2007-04-10 | Xrgomics Pte, Ltd. | Reduced keyboard system that emulates QWERTY-type mapping and typing |
US20040041723A1 (en) * | 2002-05-27 | 2004-03-04 | Canon Kabushiki Kaisha, Tokyo, Japan | Remote control device |
US20040080487A1 (en) * | 2002-10-29 | 2004-04-29 | Griffin Jason T. | Electronic device having keyboard for thumb typing |
US7382358B2 (en) * | 2003-01-16 | 2008-06-03 | Forword Input, Inc. | System and method for continuous stroke word-based text input |
US7098896B2 (en) * | 2003-01-16 | 2006-08-29 | Forword Input Inc. | System and method for continuous stroke word-based text input |
US7453439B1 (en) * | 2003-01-16 | 2008-11-18 | Forward Input Inc. | System and method for continuous stroke word-based text input |
US20040140956A1 (en) * | 2003-01-16 | 2004-07-22 | Kushler Clifford A. | System and method for continuous stroke word-based text input |
US20070040813A1 (en) * | 2003-01-16 | 2007-02-22 | Forword Input, Inc. | System and method for continuous stroke word-based text input |
US20040239624A1 (en) * | 2003-04-02 | 2004-12-02 | Artoun Ramian | Freehand symbolic input apparatus and method |
US20050052406A1 (en) * | 2003-04-09 | 2005-03-10 | James Stephanick | Selective input system based on tracking of motion parameters of an input device |
US7750891B2 (en) * | 2003-04-09 | 2010-07-06 | Tegic Communications, Inc. | Selective input system based on tracking of motion parameters of an input device |
US20070188472A1 (en) * | 2003-04-18 | 2007-08-16 | Ghassabian Benjamin F | Systems to enhance data entry in mobile and fixed environment |
US20050099400A1 (en) * | 2003-11-06 | 2005-05-12 | Samsung Electronics Co., Ltd. | Apparatus and method for providing vitrtual graffiti and recording medium for the same |
US7956845B2 (en) * | 2003-11-06 | 2011-06-07 | Samsung Electronics Co., Ltd | Apparatus and method for providing virtual graffiti and recording medium for the same |
US20050141770A1 (en) * | 2003-12-30 | 2005-06-30 | Nokia Corporation | Split on-screen keyboard |
US20050162402A1 (en) * | 2004-01-27 | 2005-07-28 | Watanachote Susornpol J. | Methods of interacting with a computer using a finger(s) touch sensing input device with visual feedback |
US20050195159A1 (en) * | 2004-02-23 | 2005-09-08 | Hunleth Frank A. | Keyboardless text entry |
US8311829B2 (en) * | 2004-06-02 | 2012-11-13 | Tegic Communications, Inc. | Multimodal disambiguation of speech recognition |
US20070182595A1 (en) * | 2004-06-04 | 2007-08-09 | Firooz Ghasabian | Systems to enhance data entry in mobile and fixed environment |
US20090146848A1 (en) * | 2004-06-04 | 2009-06-11 | Ghassabian Firooz Benjamin | Systems to enhance data entry in mobile and fixed environment |
US20060119588A1 (en) * | 2004-12-03 | 2006-06-08 | Sung-Min Yoon | Apparatus and method of processing information input using a touchpad |
US20060152496A1 (en) * | 2005-01-13 | 2006-07-13 | 602531 British Columbia Ltd. | Method, system, apparatus and computer-readable media for directing input associated with keyboard-type device |
US8311805B2 (en) * | 2005-01-21 | 2012-11-13 | Prashant Parikh | Automatic dynamic contextual data entry completion system |
US7991784B2 (en) * | 2005-01-21 | 2011-08-02 | Prashant Parikh | Automatic dynamic contextual data entry completion system |
US20060197750A1 (en) * | 2005-03-04 | 2006-09-07 | Apple Computer, Inc. | Hand held electronic device with multiple touch sensing devices |
US20090135162A1 (en) * | 2005-03-10 | 2009-05-28 | Koninklijke Philips Electronics, N.V. | System and Method For Detecting the Location, Size and Shape of Multiple Objects That Interact With a Touch Screen Display |
US20060227032A1 (en) * | 2005-04-11 | 2006-10-12 | Alberto Vidal | Apparatus and method to facilitate universal remote control |
US20060238510A1 (en) * | 2005-04-25 | 2006-10-26 | Georgios Panotopoulos | User interface incorporating emulated hard keys |
US8117540B2 (en) * | 2005-05-18 | 2012-02-14 | Neuer Wall Treuhand Gmbh | Method and device incorporating improved text input mechanism |
US8036878B2 (en) * | 2005-05-18 | 2011-10-11 | Never Wall Treuhand GmbH | Device incorporating improved text input mechanism |
US20090199092A1 (en) * | 2005-06-16 | 2009-08-06 | Firooz Ghassabian | Data entry system |
US7737999B2 (en) * | 2005-08-26 | 2010-06-15 | Veveo, Inc. | User interface for visual cooperation between text input and display device |
US20070061754A1 (en) * | 2005-08-26 | 2007-03-15 | Veveo, Inc. | User interface for visual cooperation between text input and display device |
US20070046641A1 (en) * | 2005-09-01 | 2007-03-01 | Swee Ho Lim | Entering a character into an electronic device |
US7443316B2 (en) * | 2005-09-01 | 2008-10-28 | Motorola, Inc. | Entering a character into an electronic device |
US7539472B2 (en) * | 2005-09-13 | 2009-05-26 | Microsoft Corporation | Type-ahead keypad input for an input device |
US20070060176A1 (en) * | 2005-09-13 | 2007-03-15 | Microsoft Corporation | Type-ahead keypad input |
US7542029B2 (en) * | 2005-09-20 | 2009-06-02 | Cliff Kushler | System and method for a user interface for text editing and menu selection |
US20080316212A1 (en) * | 2005-09-20 | 2008-12-25 | Cliff Kushler | System and method for a user interface for text editing and menu selection |
US20070082708A1 (en) * | 2005-10-07 | 2007-04-12 | Research In Motion Limited | Device, system, and method for informing users of functions and characters associated with telephone keys |
US7869832B2 (en) * | 2005-10-07 | 2011-01-11 | Research In Motion Limited | Device, system, and method for informing users of functions and characters associated with telephone keys |
US20070080954A1 (en) * | 2005-10-07 | 2007-04-12 | Research In Motion Limited | System and method for using navigational and other commands on a mobile communication device |
US7649478B1 (en) * | 2005-11-03 | 2010-01-19 | Hyoungsoo Yoon | Data entry using sequential keystrokes |
US20070132887A1 (en) * | 2005-12-09 | 2007-06-14 | David Boyden | On screen display for alpha-numeric input |
US8284445B2 (en) * | 2006-02-17 | 2012-10-09 | Canon Kabushiki Kaisha | Image forming apparatus for forming images from objects |
US20070195353A1 (en) * | 2006-02-17 | 2007-08-23 | Canon Kabushiki Kaisha | Image forming apparatus |
US8112454B2 (en) * | 2006-03-06 | 2012-02-07 | Veveo, Inc. | Methods and systems for ordering content items according to learned user preferences |
US8296484B2 (en) * | 2006-03-30 | 2012-10-23 | Harris Corporation | Alphanumeric data entry apparatus and method using multicharacter keys of a keypad |
US8065453B2 (en) * | 2006-04-06 | 2011-11-22 | Research In Motion Limited | Handheld electronic device and associated method employing a multiple-axis input device and learning a context of a text input for use by a disambiguation routine |
US8086602B2 (en) * | 2006-04-20 | 2011-12-27 | Veveo Inc. | User interface methods and systems for selecting and presenting content based on user navigation and selection actions associated with the content |
US8060839B2 (en) * | 2006-06-30 | 2011-11-15 | Samsung Electronics Co., Ltd | Character input method and mobile communication terminal using the same |
US20080007434A1 (en) * | 2006-07-10 | 2008-01-10 | Luben Hristov | Priority and Combination Suppression Techniques (PST/CST) for a Capacitive Keyboard |
US8078884B2 (en) * | 2006-11-13 | 2011-12-13 | Veveo, Inc. | Method of and system for selecting and presenting content based on user identification |
US20080168364A1 (en) * | 2007-01-05 | 2008-07-10 | Apple Computer, Inc. | Adaptive acceleration of mouse cursor |
US7957955B2 (en) * | 2007-01-05 | 2011-06-07 | Apple Inc. | Method and system for providing word recommendations for text input |
US20100103127A1 (en) * | 2007-02-23 | 2010-04-29 | Taeun Park | Virtual Keyboard Input System Using Pointing Apparatus In Digital Device |
US8330625B2 (en) * | 2007-03-30 | 2012-12-11 | Honeywell International Inc. | Aircraft systems with flight management systems that display reports |
US8296294B2 (en) * | 2007-05-25 | 2012-10-23 | Veveo, Inc. | Method and system for unified searching across and within multiple documents |
US20090007001A1 (en) * | 2007-06-28 | 2009-01-01 | Matsushita Electric Industrial Co., Ltd. | Virtual keypad systems and methods |
US8065624B2 (en) * | 2007-06-28 | 2011-11-22 | Panasonic Corporation | Virtual keypad systems and methods |
US8078978B2 (en) * | 2007-10-19 | 2011-12-13 | Google Inc. | Method and system for predicting text |
US8316035B2 (en) * | 2008-01-16 | 2012-11-20 | International Business Machines Corporation | Systems and arrangements of text type-ahead |
US7817145B2 (en) * | 2008-04-18 | 2010-10-19 | Cirque Corporation | Method and system for performing scrolling by movement of a pointing object in a curvilinear path on a touchpad |
US20110279397A1 (en) * | 2009-01-26 | 2011-11-17 | Zrro Technologies (2009) Ltd. | Device and method for monitoring the object's behavior |
US20100245395A1 (en) * | 2009-03-31 | 2010-09-30 | Dell Products L.P. | Information Handling System with Touch Pad and Point Stick |
US20100265200A1 (en) * | 2009-04-21 | 2010-10-21 | Samsung Electronics Co., Ltd. | Method for processing multi-touch input in mobile device |
US20110279354A1 (en) * | 2010-05-12 | 2011-11-17 | Aaron Tang | Computer with tv mode |
US20110279376A1 (en) * | 2010-05-12 | 2011-11-17 | Aaron Tang | Remote control to operate computer system |
US20120117101A1 (en) * | 2010-11-10 | 2012-05-10 | Erland Unruh | Text entry with word prediction, completion, or correction supplemented by search of shared corpus |
Cited By (473)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9646614B2 (en) | 2000-03-16 | 2017-05-09 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US10623347B2 (en) | 2003-05-02 | 2020-04-14 | Apple Inc. | Method and apparatus for displaying information during an instant messaging session |
US8458278B2 (en) | 2003-05-02 | 2013-06-04 | Apple Inc. | Method and apparatus for displaying information during an instant messaging session |
US10348654B2 (en) | 2003-05-02 | 2019-07-09 | Apple Inc. | Method and apparatus for displaying information during an instant messaging session |
US20100185960A1 (en) * | 2003-05-02 | 2010-07-22 | Apple Inc. | Method and Apparatus for Displaying Information During an Instant Messaging Session |
US8554861B2 (en) | 2003-05-02 | 2013-10-08 | Apple Inc. | Method and apparatus for displaying information during an instant messaging session |
US10318871B2 (en) | 2005-09-08 | 2019-06-11 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US11928604B2 (en) | 2005-09-08 | 2024-03-12 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US9594457B2 (en) | 2005-12-30 | 2017-03-14 | Microsoft Technology Licensing, Llc | Unintentional touch rejection |
US9261964B2 (en) | 2005-12-30 | 2016-02-16 | Microsoft Technology Licensing, Llc | Unintentional touch rejection |
US9952718B2 (en) | 2005-12-30 | 2018-04-24 | Microsoft Technology Licensing, Llc | Unintentional touch rejection |
US10019080B2 (en) | 2005-12-30 | 2018-07-10 | Microsoft Technology Licensing, Llc | Unintentional touch rejection |
US9946370B2 (en) | 2005-12-30 | 2018-04-17 | Microsoft Technology Licensing, Llc | Unintentional touch rejection |
US11762547B2 (en) | 2006-09-06 | 2023-09-19 | Apple Inc. | Portable electronic device for instant messaging |
US11169690B2 (en) | 2006-09-06 | 2021-11-09 | Apple Inc. | Portable electronic device for instant messaging |
US9600174B2 (en) | 2006-09-06 | 2017-03-21 | Apple Inc. | Portable electronic device for instant messaging |
US9304675B2 (en) * | 2006-09-06 | 2016-04-05 | Apple Inc. | Portable electronic device for instant messaging |
US20080055269A1 (en) * | 2006-09-06 | 2008-03-06 | Lemay Stephen O | Portable Electronic Device for Instant Messaging |
US10572142B2 (en) | 2006-09-06 | 2020-02-25 | Apple Inc. | Portable electronic device for instant messaging |
US11029838B2 (en) | 2006-09-06 | 2021-06-08 | Apple Inc. | Touch screen device, method, and graphical user interface for customizing display of content category icons |
US9117447B2 (en) | 2006-09-08 | 2015-08-25 | Apple Inc. | Using event alert text as input to an automated assistant |
US8942986B2 (en) | 2006-09-08 | 2015-01-27 | Apple Inc. | Determining user intent based on ontologies of domains |
US8930191B2 (en) | 2006-09-08 | 2015-01-06 | Apple Inc. | Paraphrasing of user requests and results by automated digital assistant |
US11012942B2 (en) | 2007-04-03 | 2021-05-18 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US10568032B2 (en) | 2007-04-03 | 2020-02-18 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US11122158B2 (en) | 2007-06-28 | 2021-09-14 | Apple Inc. | Portable electronic device with conversation management for incoming instant messages |
US11743375B2 (en) | 2007-06-28 | 2023-08-29 | Apple Inc. | Portable electronic device with conversation management for incoming instant messages |
US9954996B2 (en) | 2007-06-28 | 2018-04-24 | Apple Inc. | Portable electronic device with conversation management for incoming instant messages |
US20090005011A1 (en) * | 2007-06-28 | 2009-01-01 | Greg Christie | Portable Electronic Device with Conversation Management for Incoming Instant Messages |
US8010163B2 (en) * | 2007-11-08 | 2011-08-30 | Htc Corporation | Method for displaying dialing information and mobile communication device using the method |
US20090124293A1 (en) * | 2007-11-08 | 2009-05-14 | Tsui Yuan-Mao | Method for displaying dialing information and mobile communication device using the method |
US11023513B2 (en) | 2007-12-20 | 2021-06-01 | Apple Inc. | Method and apparatus for searching using an active ontology |
US10381016B2 (en) | 2008-01-03 | 2019-08-13 | Apple Inc. | Methods and apparatus for altering audio output signals |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US11126326B2 (en) | 2008-01-06 | 2021-09-21 | Apple Inc. | Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars |
US10521084B2 (en) | 2008-01-06 | 2019-12-31 | Apple Inc. | Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars |
US10503366B2 (en) | 2008-01-06 | 2019-12-10 | Apple Inc. | Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars |
US20090177981A1 (en) * | 2008-01-06 | 2009-07-09 | Greg Christie | Portable Electronic Device for Instant Messaging Multiple Recipients |
US8407603B2 (en) | 2008-01-06 | 2013-03-26 | Apple Inc. | Portable electronic device for instant messaging multiple recipients |
US9330381B2 (en) | 2008-01-06 | 2016-05-03 | Apple Inc. | Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars |
US9792001B2 (en) | 2008-01-06 | 2017-10-17 | Apple Inc. | Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars |
US9035876B2 (en) | 2008-01-14 | 2015-05-19 | Apple Inc. | Three-dimensional user interface session control |
US9626955B2 (en) | 2008-04-05 | 2017-04-18 | Apple Inc. | Intelligent text-to-speech conversion |
US9865248B2 (en) | 2008-04-05 | 2018-01-09 | Apple Inc. | Intelligent text-to-speech conversion |
US9535906B2 (en) | 2008-07-31 | 2017-01-03 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US10108612B2 (en) | 2008-07-31 | 2018-10-23 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US11348582B2 (en) | 2008-10-02 | 2022-05-31 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US10643611B2 (en) | 2008-10-02 | 2020-05-05 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US9959870B2 (en) | 2008-12-11 | 2018-05-01 | Apple Inc. | Speech recognition involving a mobile device |
US20110279397A1 (en) * | 2009-01-26 | 2011-11-17 | Zrro Technologies (2009) Ltd. | Device and method for monitoring the object's behavior |
JP2012515966A (en) * | 2009-01-26 | 2012-07-12 | ズッロ・テクノロジーズ・(2009)・リミテッド | Device and method for monitoring the behavior of an object |
US8830189B2 (en) * | 2009-01-26 | 2014-09-09 | Zrro Technologies (2009) Ltd. | Device and method for monitoring the object's behavior |
US8583421B2 (en) * | 2009-03-06 | 2013-11-12 | Motorola Mobility Llc | Method and apparatus for psychomotor and psycholinguistic prediction on touch based device |
US20100228539A1 (en) * | 2009-03-06 | 2010-09-09 | Motorola, Inc. | Method and apparatus for psychomotor and psycholinguistic prediction on touch based device |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US10475446B2 (en) | 2009-06-05 | 2019-11-12 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US10795541B2 (en) | 2009-06-05 | 2020-10-06 | Apple Inc. | Intelligent organization of tasks items |
US11080012B2 (en) | 2009-06-05 | 2021-08-03 | Apple Inc. | Interface for a virtual digital assistant |
US20100315266A1 (en) * | 2009-06-15 | 2010-12-16 | Microsoft Corporation | Predictive interfaces with usability constraints |
US10283110B2 (en) | 2009-07-02 | 2019-05-07 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US20120139873A1 (en) * | 2009-07-14 | 2012-06-07 | Chih-Hung Li | Touch-controlled electronic apparatus and related control method |
US20110012843A1 (en) * | 2009-07-14 | 2011-01-20 | Chih-Hung Li | Touch-controlled electronic apparatus and related control method |
US20110148762A1 (en) * | 2009-12-22 | 2011-06-23 | Universal Electronics Inc. | System and method for multi-mode command input |
US10741185B2 (en) | 2010-01-18 | 2020-08-11 | Apple Inc. | Intelligent automated assistant |
US12087308B2 (en) | 2010-01-18 | 2024-09-10 | Apple Inc. | Intelligent automated assistant |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US10706841B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Task flow identification based on user intent |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US8892446B2 (en) | 2010-01-18 | 2014-11-18 | Apple Inc. | Service orchestration for intelligent automated assistant |
US8903716B2 (en) | 2010-01-18 | 2014-12-02 | Apple Inc. | Personalized vocabulary for digital assistant |
US9548050B2 (en) | 2010-01-18 | 2017-01-17 | Apple Inc. | Intelligent automated assistant |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US11423886B2 (en) | 2010-01-18 | 2022-08-23 | Apple Inc. | Task flow identification based on user intent |
US20110187647A1 (en) * | 2010-02-04 | 2011-08-04 | Charles Howard Woloszynski | Method and apparatus for virtual keyboard interactions from secondary surfaces |
US20110212876A1 (en) * | 2010-02-10 | 2011-09-01 | Michelle Meek | Cleaning composition comprising amylase variants with high stability in the presence of a chelating agent |
US20110202836A1 (en) * | 2010-02-12 | 2011-08-18 | Microsoft Corporation | Typing assistance for editing |
US10156981B2 (en) | 2010-02-12 | 2018-12-18 | Microsoft Technology Licensing, Llc | User-centric soft keyboard predictive technologies |
US9165257B2 (en) | 2010-02-12 | 2015-10-20 | Microsoft Technology Licensing, Llc | Typing assistance for editing |
US20110201387A1 (en) * | 2010-02-12 | 2011-08-18 | Microsoft Corporation | Real-time typing assistance |
US20110202876A1 (en) * | 2010-02-12 | 2011-08-18 | Microsoft Corporation | User-centric soft keyboard predictive technologies |
US8782556B2 (en) | 2010-02-12 | 2014-07-15 | Microsoft Corporation | User-centric soft keyboard predictive technologies |
US10126936B2 (en) | 2010-02-12 | 2018-11-13 | Microsoft Technology Licensing, Llc | Typing assistance for editing |
US9613015B2 (en) | 2010-02-12 | 2017-04-04 | Microsoft Technology Licensing, Llc | User-centric soft keyboard predictive technologies |
US10049675B2 (en) | 2010-02-25 | 2018-08-14 | Apple Inc. | User profiling for voice input processing |
US9633660B2 (en) | 2010-02-25 | 2017-04-25 | Apple Inc. | User profiling for voice input processing |
US10692504B2 (en) | 2010-02-25 | 2020-06-23 | Apple Inc. | User profiling for voice input processing |
US8854323B2 (en) | 2010-05-20 | 2014-10-07 | Panasonic Intellectual Property Corporation Of America | Operating apparatus, operating method, program, recording medium, and integrated circuit |
US20110310019A1 (en) * | 2010-06-16 | 2011-12-22 | International Business Machines Corporation | Reconfiguration of virtual keyboard |
US8648809B2 (en) * | 2010-06-16 | 2014-02-11 | International Business Machines Corporation | Reconfiguration of virtual keyboard |
US9201501B2 (en) | 2010-07-20 | 2015-12-01 | Apple Inc. | Adaptive projector |
US9158375B2 (en) | 2010-07-20 | 2015-10-13 | Apple Inc. | Interactive reality augmentation for natural interaction |
US20120078614A1 (en) * | 2010-09-27 | 2012-03-29 | Primesense Ltd. | Virtual keyboard for a non-tactile three dimensional user interface |
US8959013B2 (en) * | 2010-09-27 | 2015-02-17 | Apple Inc. | Virtual keyboard for a non-tactile three dimensional user interface |
US9501161B2 (en) | 2010-10-22 | 2016-11-22 | Hewlett-Packard Development Company, L.P. | User interface for facilitating character input |
US20120098743A1 (en) * | 2010-10-26 | 2012-04-26 | Pei-Ling Lai | Input method, input device, and computer system |
US10262324B2 (en) | 2010-11-29 | 2019-04-16 | Biocatch Ltd. | System, device, and method of differentiating among users based on user-specific page navigation sequence |
US11250435B2 (en) | 2010-11-29 | 2022-02-15 | Biocatch Ltd. | Contextual mapping of web-pages, and generation of fraud-relatedness score-values |
US10621585B2 (en) | 2010-11-29 | 2020-04-14 | Biocatch Ltd. | Contextual mapping of web-pages, and generation of fraud-relatedness score-values |
US10949514B2 (en) | 2010-11-29 | 2021-03-16 | Biocatch Ltd. | Device, system, and method of differentiating among users based on detection of hardware components |
US11314849B2 (en) | 2010-11-29 | 2022-04-26 | Biocatch Ltd. | Method, device, and system of detecting a lie of a user who inputs data |
US11425563B2 (en) | 2010-11-29 | 2022-08-23 | Biocatch Ltd. | Method, device, and system of differentiating between a cyber-attacker and a legitimate user |
US10298614B2 (en) * | 2010-11-29 | 2019-05-21 | Biocatch Ltd. | System, device, and method of generating and managing behavioral biometric cookies |
US10586036B2 (en) | 2010-11-29 | 2020-03-10 | Biocatch Ltd. | System, device, and method of recovery and resetting of user authentication factor |
US11269977B2 (en) | 2010-11-29 | 2022-03-08 | Biocatch Ltd. | System, apparatus, and method of collecting and processing data in electronic devices |
US10776476B2 (en) | 2010-11-29 | 2020-09-15 | Biocatch Ltd. | System, device, and method of visual login |
US10404729B2 (en) | 2010-11-29 | 2019-09-03 | Biocatch Ltd. | Device, method, and system of generating fraud-alerts for cyber-attacks |
US12101354B2 (en) * | 2010-11-29 | 2024-09-24 | Biocatch Ltd. | Device, system, and method of detecting vishing attacks |
US11330012B2 (en) | 2010-11-29 | 2022-05-10 | Biocatch Ltd. | System, method, and device of authenticating a user based on selfie image or selfie video |
US10747305B2 (en) | 2010-11-29 | 2020-08-18 | Biocatch Ltd. | Method, system, and device of authenticating identity of a user of an electronic device |
US20240080339A1 (en) * | 2010-11-29 | 2024-03-07 | Biocatch Ltd. | Device, System, and Method of Detecting Vishing Attacks |
US10834590B2 (en) | 2010-11-29 | 2020-11-10 | Biocatch Ltd. | Method, device, and system of differentiating between a cyber-attacker and a legitimate user |
US11580553B2 (en) | 2010-11-29 | 2023-02-14 | Biocatch Ltd. | Method, device, and system of detecting mule accounts and accounts used for money laundering |
US11838118B2 (en) * | 2010-11-29 | 2023-12-05 | Biocatch Ltd. | Device, system, and method of detecting vishing attacks |
US10949757B2 (en) | 2010-11-29 | 2021-03-16 | Biocatch Ltd. | System, device, and method of detecting user identity based on motor-control loop model |
US10897482B2 (en) | 2010-11-29 | 2021-01-19 | Biocatch Ltd. | Method, device, and system of back-coloring, forward-coloring, and fraud detection |
US10728761B2 (en) | 2010-11-29 | 2020-07-28 | Biocatch Ltd. | Method, device, and system of detecting a lie of a user who inputs data |
US11223619B2 (en) | 2010-11-29 | 2022-01-11 | Biocatch Ltd. | Device, system, and method of user authentication based on user-specific characteristics of task performance |
US11210674B2 (en) | 2010-11-29 | 2021-12-28 | Biocatch Ltd. | Method, device, and system of detecting mule accounts and accounts used for money laundering |
US10917431B2 (en) | 2010-11-29 | 2021-02-09 | Biocatch Ltd. | System, method, and device of authenticating a user based on selfie image or selfie video |
US10474815B2 (en) | 2010-11-29 | 2019-11-12 | Biocatch Ltd. | System, device, and method of detecting malicious automatic script and code injection |
US20210329030A1 (en) * | 2010-11-29 | 2021-10-21 | Biocatch Ltd. | Device, System, and Method of Detecting Vishing Attacks |
US20120137244A1 (en) * | 2010-11-30 | 2012-05-31 | Inventec Corporation | Touch device input device and operation method of the same |
US8872762B2 (en) | 2010-12-08 | 2014-10-28 | Primesense Ltd. | Three dimensional user interface cursor control |
US8933876B2 (en) | 2010-12-13 | 2015-01-13 | Apple Inc. | Three dimensional user interface session control |
US10762293B2 (en) | 2010-12-22 | 2020-09-01 | Apple Inc. | Using parts-of-speech tagging and named entity recognition for spelling correction |
US9454225B2 (en) | 2011-02-09 | 2016-09-27 | Apple Inc. | Gaze-based display control |
US9285874B2 (en) | 2011-02-09 | 2016-03-15 | Apple Inc. | Gaze detection in a 3D mapping environment |
US9342146B2 (en) | 2011-02-09 | 2016-05-17 | Apple Inc. | Pointing-based display interaction |
US8719724B2 (en) | 2011-03-16 | 2014-05-06 | Honeywell International Inc. | Method for enlarging characters displayed on an adaptive touch screen key pad |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US10102359B2 (en) | 2011-03-21 | 2018-10-16 | Apple Inc. | Device access using voice authentication |
US10417405B2 (en) | 2011-03-21 | 2019-09-17 | Apple Inc. | Device access using voice authentication |
US10706373B2 (en) | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US11120372B2 (en) | 2011-06-03 | 2021-09-14 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US11350253B2 (en) | 2011-06-03 | 2022-05-31 | Apple Inc. | Active transport based notifications |
US8881051B2 (en) | 2011-07-05 | 2014-11-04 | Primesense Ltd | Zoom-based gesture user interface |
US9459758B2 (en) | 2011-07-05 | 2016-10-04 | Apple Inc. | Gesture-based interface with enhanced features |
US9377865B2 (en) | 2011-07-05 | 2016-06-28 | Apple Inc. | Zoom-based gesture user interface |
US9030498B2 (en) | 2011-08-15 | 2015-05-12 | Apple Inc. | Combining explicit select gestures and timeclick in a non-tactile three dimensional user interface |
US9218063B2 (en) | 2011-08-24 | 2015-12-22 | Apple Inc. | Sessionless pointing user interface |
US9122311B2 (en) | 2011-08-24 | 2015-09-01 | Apple Inc. | Visual feedback for tactile and non-tactile user interfaces |
US9798393B2 (en) | 2011-08-29 | 2017-10-24 | Apple Inc. | Text correction processing |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
CN102937876A (en) * | 2011-11-23 | 2013-02-20 | 微软公司 | Dynamic scaling of a touch sensor |
US20130127738A1 (en) * | 2011-11-23 | 2013-05-23 | Microsoft Corporation | Dynamic scaling of touch sensor |
WO2013078171A1 (en) * | 2011-11-23 | 2013-05-30 | Microsoft Corporation | Dynamic scaling of touch sensor |
EP2597548A3 (en) * | 2011-11-23 | 2013-12-11 | Microsoft Corporation | Dynamic scaling of touch sensor |
US9519419B2 (en) * | 2012-01-17 | 2016-12-13 | Microsoft Technology Licensing, Llc | Skinnable touch device grip patterns |
US20130181902A1 (en) * | 2012-01-17 | 2013-07-18 | Microsoft Corporation | Skinnable touch device grip patterns |
CN104054043A (en) * | 2012-01-17 | 2014-09-17 | 微软公司 | Skinnable touch device grip patterns |
US9354748B2 (en) | 2012-02-13 | 2016-05-31 | Microsoft Technology Licensing, Llc | Optical stylus interaction |
US9229534B2 (en) | 2012-02-28 | 2016-01-05 | Apple Inc. | Asymmetric mapping for tactile and non-tactile user interfaces |
USRE48963E1 (en) | 2012-03-02 | 2022-03-08 | Microsoft Technology Licensing, Llc | Connection device for computing devices |
US9946307B2 (en) | 2012-03-02 | 2018-04-17 | Microsoft Technology Licensing, Llc | Classifying the intent of user input |
US9304948B2 (en) * | 2012-03-02 | 2016-04-05 | Microsoft Technology Licensing, Llc | Sensing user input at display area edge |
US9460029B2 (en) | 2012-03-02 | 2016-10-04 | Microsoft Technology Licensing, Llc | Pressure sensitive keys |
US8780540B2 (en) | 2012-03-02 | 2014-07-15 | Microsoft Corporation | Flexible hinge and removable attachment |
US9793073B2 (en) | 2012-03-02 | 2017-10-17 | Microsoft Technology Licensing, Llc | Backlighting a fabric enclosure of a flexible cover |
US9426905B2 (en) | 2012-03-02 | 2016-08-23 | Microsoft Technology Licensing, Llc | Connection device for computing devices |
US8780541B2 (en) | 2012-03-02 | 2014-07-15 | Microsoft Corporation | Flexible hinge and removable attachment |
US9304949B2 (en) | 2012-03-02 | 2016-04-05 | Microsoft Technology Licensing, Llc | Sensing user input at display area edge |
US9298236B2 (en) | 2012-03-02 | 2016-03-29 | Microsoft Technology Licensing, Llc | Multi-stage power adapter configured to provide a first power level upon initial connection of the power adapter to the host device and a second power level thereafter upon notification from the host device to the power adapter |
US8791382B2 (en) | 2012-03-02 | 2014-07-29 | Microsoft Corporation | Input device securing techniques |
US9852855B2 (en) | 2012-03-02 | 2017-12-26 | Microsoft Technology Licensing, Llc | Pressure sensitive key normalization |
US8830668B2 (en) | 2012-03-02 | 2014-09-09 | Microsoft Corporation | Flexible hinge and removable attachment |
US9619071B2 (en) | 2012-03-02 | 2017-04-11 | Microsoft Technology Licensing, Llc | Computing device and an apparatus having sensors configured for measuring spatial information indicative of a position of the computing devices |
US20130229363A1 (en) * | 2012-03-02 | 2013-09-05 | Christopher A. Whitman | Sensing User Input At Display Area Edge |
US8850241B2 (en) | 2012-03-02 | 2014-09-30 | Microsoft Corporation | Multi-stage power adapter configured to provide low power upon initial connection of the power adapter to the host device and high power thereafter upon notification from the host device to the power adapter |
US9870066B2 (en) | 2012-03-02 | 2018-01-16 | Microsoft Technology Licensing, Llc | Method of manufacturing an input device |
US10963087B2 (en) | 2012-03-02 | 2021-03-30 | Microsoft Technology Licensing, Llc | Pressure sensitive keys |
US9134808B2 (en) | 2012-03-02 | 2015-09-15 | Microsoft Technology Licensing, Llc | Device kickstand |
US9275809B2 (en) | 2012-03-02 | 2016-03-01 | Microsoft Technology Licensing, Llc | Device camera angle |
US9904327B2 (en) | 2012-03-02 | 2018-02-27 | Microsoft Technology Licensing, Llc | Flexible hinge and removable attachment |
US8854799B2 (en) | 2012-03-02 | 2014-10-07 | Microsoft Corporation | Flux fountain |
US9465412B2 (en) | 2012-03-02 | 2016-10-11 | Microsoft Technology Licensing, Llc | Input device layers and nesting |
US8873227B2 (en) | 2012-03-02 | 2014-10-28 | Microsoft Corporation | Flexible hinge support layer |
US8896993B2 (en) | 2012-03-02 | 2014-11-25 | Microsoft Corporation | Input device layers and nesting |
US9766663B2 (en) | 2012-03-02 | 2017-09-19 | Microsoft Technology Licensing, Llc | Hinge for component attachment |
US9618977B2 (en) | 2012-03-02 | 2017-04-11 | Microsoft Technology Licensing, Llc | Input device securing techniques |
US8903517B2 (en) | 2012-03-02 | 2014-12-02 | Microsoft Corporation | Computer device and an apparatus having sensors configured for measuring spatial information indicative of a position of the computing devices |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US9268373B2 (en) | 2012-03-02 | 2016-02-23 | Microsoft Technology Licensing, Llc | Flexible hinge spine |
US8935774B2 (en) | 2012-03-02 | 2015-01-13 | Microsoft Corporation | Accessory device authentication |
US11069336B2 (en) | 2012-03-02 | 2021-07-20 | Apple Inc. | Systems and methods for name pronunciation |
US8947864B2 (en) | 2012-03-02 | 2015-02-03 | Microsoft Corporation | Flexible hinge and removable attachment |
US9710093B2 (en) | 2012-03-02 | 2017-07-18 | Microsoft Technology Licensing, Llc | Pressure sensitive key normalization |
US9047207B2 (en) | 2012-03-02 | 2015-06-02 | Microsoft Technology Licensing, Llc | Mobile device power state |
US9176900B2 (en) | 2012-03-02 | 2015-11-03 | Microsoft Technology Licensing, Llc | Flexible hinge and removable attachment |
US9176901B2 (en) | 2012-03-02 | 2015-11-03 | Microsoft Technology Licensing, Llc | Flux fountain |
US9064654B2 (en) | 2012-03-02 | 2015-06-23 | Microsoft Technology Licensing, Llc | Method of manufacturing an input device |
US9075566B2 (en) | 2012-03-02 | 2015-07-07 | Microsoft Technoogy Licensing, LLC | Flexible hinge spine |
US10013030B2 (en) | 2012-03-02 | 2018-07-03 | Microsoft Technology Licensing, Llc | Multiple position input device cover |
US9360893B2 (en) | 2012-03-02 | 2016-06-07 | Microsoft Technology Licensing, Llc | Input device writing surface |
US9158384B2 (en) | 2012-03-02 | 2015-10-13 | Microsoft Technology Licensing, Llc | Flexible hinge protrusion attachment |
US9158383B2 (en) | 2012-03-02 | 2015-10-13 | Microsoft Technology Licensing, Llc | Force concentrator |
US9146620B2 (en) | 2012-03-02 | 2015-09-29 | Microsoft Technology Licensing, Llc | Input device assembly |
US9098117B2 (en) | 2012-03-02 | 2015-08-04 | Microsoft Technology Licensing, Llc | Classifying the intent of user input |
US9134807B2 (en) | 2012-03-02 | 2015-09-15 | Microsoft Technology Licensing, Llc | Pressure sensitive key normalization |
US9678542B2 (en) | 2012-03-02 | 2017-06-13 | Microsoft Technology Licensing, Llc | Multiple position input device cover |
US9111703B2 (en) | 2012-03-02 | 2015-08-18 | Microsoft Technology Licensing, Llc | Sensor stack venting |
US9116550B2 (en) | 2012-03-02 | 2015-08-25 | Microsoft Technology Licensing, Llc | Device kickstand |
US9411751B2 (en) | 2012-03-02 | 2016-08-09 | Microsoft Technology Licensing, Llc | Key formation |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
US11169611B2 (en) | 2012-03-26 | 2021-11-09 | Apple Inc. | Enhanced virtual touchpad |
US9377863B2 (en) | 2012-03-26 | 2016-06-28 | Apple Inc. | Gaze-enhanced virtual touchscreen |
US9772744B2 (en) | 2012-04-19 | 2017-09-26 | Sony Corporation | Information processing apparatus, information processing method, program, and information processing system |
US10162480B2 (en) | 2012-04-19 | 2018-12-25 | Sony Corporation | Information processing apparatus, information processing method, program, and information processing system |
US10942620B2 (en) | 2012-04-19 | 2021-03-09 | Sony Corporation | Information processing apparatus, information processing method, program, and information processing system |
WO2013157232A3 (en) * | 2012-04-19 | 2014-01-09 | Sony Corporation | Information processing apparatus, information processing method, program, and information processing system |
US20170163770A1 (en) * | 2012-04-28 | 2017-06-08 | Alibaba Group Holding Limited | Performing autocomplete of content |
US10110708B2 (en) * | 2012-04-28 | 2018-10-23 | Alibaba Group Holding Limited | Performing autocomplete of content |
US9584626B2 (en) * | 2012-04-28 | 2017-02-28 | Alibaba Group Holding Limited | Performing autocomplete of content |
US20130290410A1 (en) * | 2012-04-28 | 2013-10-31 | Alibaba Group Holding Limited | Performing autocomplete of content |
US9348605B2 (en) | 2012-05-14 | 2016-05-24 | Microsoft Technology Licensing, Llc | System and method for accessory device architecture that passes human interface device (HID) data via intermediate processor |
US9953088B2 (en) | 2012-05-14 | 2018-04-24 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US8949477B2 (en) | 2012-05-14 | 2015-02-03 | Microsoft Technology Licensing, Llc | Accessory device architecture |
US9959241B2 (en) | 2012-05-14 | 2018-05-01 | Microsoft Technology Licensing, Llc | System and method for accessory device architecture that passes via intermediate processor a descriptor when processing in a low power state |
US11269678B2 (en) | 2012-05-15 | 2022-03-08 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
US10079014B2 (en) | 2012-06-08 | 2018-09-18 | Apple Inc. | Name recognition system |
US10031556B2 (en) | 2012-06-08 | 2018-07-24 | Microsoft Technology Licensing, Llc | User experience adaptation |
US10107994B2 (en) | 2012-06-12 | 2018-10-23 | Microsoft Technology Licensing, Llc | Wide field-of-view virtual image projector |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9824808B2 (en) | 2012-08-20 | 2017-11-21 | Microsoft Technology Licensing, Llc | Switchable magnetic lock |
US9576574B2 (en) | 2012-09-10 | 2017-02-21 | Apple Inc. | Context-sensitive handling of interruptions by intelligent digital assistant |
US9971774B2 (en) | 2012-09-19 | 2018-05-15 | Apple Inc. | Voice-based media searching |
US9544504B2 (en) | 2012-11-02 | 2017-01-10 | Microsoft Technology Licensing, Llc | Rapid synchronized lighting and shuttering |
US10656750B2 (en) | 2012-11-12 | 2020-05-19 | Microsoft Technology Licensing, Llc | Touch-sensitive bezel techniques |
US9582122B2 (en) | 2012-11-12 | 2017-02-28 | Microsoft Technology Licensing, Llc | Touch-sensitive bezel techniques |
WO2014107064A1 (en) * | 2013-01-07 | 2014-07-10 | Samsung Electronics Co., Ltd. | Display apparatus, method for ui display thereof and computer-readable recording medium |
US11665387B2 (en) | 2013-01-07 | 2023-05-30 | Samsung Electronics Co., Ltd. | Display apparatus, method for UI display thereof and computer-readable recording medium |
US11363324B2 (en) | 2013-01-07 | 2022-06-14 | Samsung Electronics Co., Ltd. | Display apparatus, method for UI display thereof and computer-readable recording medium |
US11930237B2 (en) | 2013-01-07 | 2024-03-12 | Samsung Electronics Co., Ltd. | Display apparatus, method for UI display thereof and computer-readable recording medium |
US10986393B2 (en) | 2013-01-07 | 2021-04-20 | Samsung Electronics Co., Ltd. | Display apparatus, method for UI display thereof and computer-readable recording medium |
US10199051B2 (en) | 2013-02-07 | 2019-02-05 | Apple Inc. | Voice trigger for a digital assistant |
US10714117B2 (en) | 2013-02-07 | 2020-07-14 | Apple Inc. | Voice trigger for a digital assistant |
US10978090B2 (en) | 2013-02-07 | 2021-04-13 | Apple Inc. | Voice trigger for a digital assistant |
US9368114B2 (en) | 2013-03-14 | 2016-06-14 | Apple Inc. | Context-sensitive handling of interruptions |
US11388291B2 (en) | 2013-03-14 | 2022-07-12 | Apple Inc. | System and method for processing voicemail |
US9697822B1 (en) | 2013-03-15 | 2017-07-04 | Apple Inc. | System and method for updating an adaptive speech recognition model |
US9922642B2 (en) | 2013-03-15 | 2018-03-20 | Apple Inc. | Training an at least partial voice command system |
US11798547B2 (en) | 2013-03-15 | 2023-10-24 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
US9304549B2 (en) | 2013-03-28 | 2016-04-05 | Microsoft Technology Licensing, Llc | Hinge mechanism for rotatable component attachment |
WO2014178813A1 (en) * | 2013-04-29 | 2014-11-06 | Hewlett-Packard Development Company, L.P. | Device for displaying a received user interface |
US9966060B2 (en) | 2013-06-07 | 2018-05-08 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9633674B2 (en) | 2013-06-07 | 2017-04-25 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
US9620104B2 (en) | 2013-06-07 | 2017-04-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US10657961B2 (en) | 2013-06-08 | 2020-05-19 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US9966068B2 (en) | 2013-06-08 | 2018-05-08 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US10185542B2 (en) | 2013-06-09 | 2019-01-22 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
US10769385B2 (en) | 2013-06-09 | 2020-09-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
US11048473B2 (en) | 2013-06-09 | 2021-06-29 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US11727219B2 (en) | 2013-06-09 | 2023-08-15 | Apple Inc. | System and method for inferring user intent from speech inputs |
US9300784B2 (en) | 2013-06-13 | 2016-03-29 | Apple Inc. | System and method for emergency calls initiated by voice command |
US20150029402A1 (en) * | 2013-07-26 | 2015-01-29 | Tianjin Funayuanchuang Technology Co.,Ltd. | Remote controller, system, and method for controlling remote controller |
US10791216B2 (en) | 2013-08-06 | 2020-09-29 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
US12010262B2 (en) | 2013-08-06 | 2024-06-11 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
WO2015052724A1 (en) * | 2013-10-07 | 2015-04-16 | Deshmukh Rakesh | An indian language keypad |
US11314370B2 (en) | 2013-12-06 | 2022-04-26 | Apple Inc. | Method for extracting salient dialog usage from live data |
US9477337B2 (en) | 2014-03-14 | 2016-10-25 | Microsoft Technology Licensing, Llc | Conductive trace routing for display and bezel sensors |
US9946383B2 (en) | 2014-03-14 | 2018-04-17 | Microsoft Technology Licensing, Llc | Conductive trace routing for display and bezel sensors |
US10120420B2 (en) | 2014-03-21 | 2018-11-06 | Microsoft Technology Licensing, Llc | Lockable display and techniques enabling use of lockable displays |
US9620105B2 (en) | 2014-05-15 | 2017-04-11 | Apple Inc. | Analyzing audio input for efficient speech and music recognition |
US10592095B2 (en) | 2014-05-23 | 2020-03-17 | Apple Inc. | Instantaneous speaking of content on touch devices |
US9502031B2 (en) | 2014-05-27 | 2016-11-22 | Apple Inc. | Method for supporting dynamic grammars in WFST-based ASR |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US10169329B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Exemplar-based natural language processing |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US10699717B2 (en) | 2014-05-30 | 2020-06-30 | Apple Inc. | Intelligent assistant for home automation |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US10083690B2 (en) | 2014-05-30 | 2018-09-25 | Apple Inc. | Better resolution when referencing to concepts |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US10289433B2 (en) | 2014-05-30 | 2019-05-14 | Apple Inc. | Domain specific language for encoding assistant dialog |
US11133008B2 (en) | 2014-05-30 | 2021-09-28 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US10497365B2 (en) | 2014-05-30 | 2019-12-03 | Apple Inc. | Multi-command single utterance input method |
US10878809B2 (en) | 2014-05-30 | 2020-12-29 | Apple Inc. | Multi-command single utterance input method |
US9734193B2 (en) | 2014-05-30 | 2017-08-15 | Apple Inc. | Determining domain salience ranking from ambiguous words in natural speech |
US10657966B2 (en) | 2014-05-30 | 2020-05-19 | Apple Inc. | Better resolution when referencing to concepts |
US10714095B2 (en) | 2014-05-30 | 2020-07-14 | Apple Inc. | Intelligent assistant for home automation |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US10417344B2 (en) | 2014-05-30 | 2019-09-17 | Apple Inc. | Exemplar-based natural language processing |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US11257504B2 (en) | 2014-05-30 | 2022-02-22 | Apple Inc. | Intelligent assistant for home automation |
US9966065B2 (en) | 2014-05-30 | 2018-05-08 | Apple Inc. | Multi-command single utterance input method |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10904611B2 (en) | 2014-06-30 | 2021-01-26 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US9668024B2 (en) | 2014-06-30 | 2017-05-30 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10324733B2 (en) | 2014-07-30 | 2019-06-18 | Microsoft Technology Licensing, Llc | Shutdown notifications |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10431204B2 (en) | 2014-09-11 | 2019-10-01 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US10156889B2 (en) | 2014-09-15 | 2018-12-18 | Microsoft Technology Licensing, Llc | Inductive peripheral retention device |
US9606986B2 (en) | 2014-09-29 | 2017-03-28 | Apple Inc. | Integrated word N-gram and class M-gram language models |
US9986419B2 (en) | 2014-09-30 | 2018-05-29 | Apple Inc. | Social reminders |
US10438595B2 (en) | 2014-09-30 | 2019-10-08 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US9964998B2 (en) | 2014-09-30 | 2018-05-08 | Microsoft Technology Licensing, Llc | Hinge mechanism with multiple preset positions |
US9447620B2 (en) | 2014-09-30 | 2016-09-20 | Microsoft Technology Licensing, Llc | Hinge mechanism with multiple preset positions |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US10390213B2 (en) | 2014-09-30 | 2019-08-20 | Apple Inc. | Social reminders |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US10453443B2 (en) | 2014-09-30 | 2019-10-22 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US11556230B2 (en) | 2014-12-02 | 2023-01-17 | Apple Inc. | Data detection |
US9711141B2 (en) | 2014-12-09 | 2017-07-18 | Apple Inc. | Disambiguating heteronyms in speech synthesis |
US11231904B2 (en) | 2015-03-06 | 2022-01-25 | Apple Inc. | Reducing response latency of intelligent automated assistants |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US10311871B2 (en) | 2015-03-08 | 2019-06-04 | Apple Inc. | Competing devices responding to voice triggers |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US10930282B2 (en) | 2015-03-08 | 2021-02-23 | Apple Inc. | Competing devices responding to voice triggers |
US10529332B2 (en) | 2015-03-08 | 2020-01-07 | Apple Inc. | Virtual assistant activation |
US11087759B2 (en) | 2015-03-08 | 2021-08-10 | Apple Inc. | Virtual assistant activation |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9779225B2 (en) | 2015-04-08 | 2017-10-03 | Google Inc. | Method and system to provide access to secure features of a device |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US10001857B2 (en) | 2015-05-13 | 2018-06-19 | Seiko Epson Corporation | Display apparatus and method of controlling display apparatus |
US11468282B2 (en) | 2015-05-15 | 2022-10-11 | Apple Inc. | Virtual assistant in a communication session |
US20160342294A1 (en) * | 2015-05-19 | 2016-11-24 | Google Inc. | Multi-switch option scanning |
US10067670B2 (en) * | 2015-05-19 | 2018-09-04 | Google Llc | Multi-switch option scanning |
US11127397B2 (en) | 2015-05-27 | 2021-09-21 | Apple Inc. | Device voice control |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US10356243B2 (en) | 2015-06-05 | 2019-07-16 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10681212B2 (en) | 2015-06-05 | 2020-06-09 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
US9752361B2 (en) | 2015-06-18 | 2017-09-05 | Microsoft Technology Licensing, Llc | Multistage hinge |
US11238349B2 (en) | 2015-06-25 | 2022-02-01 | Biocatch Ltd. | Conditional behavioural biometrics |
US10719765B2 (en) | 2015-06-25 | 2020-07-21 | Biocatch Ltd. | Conditional behavioral biometrics |
US11010127B2 (en) | 2015-06-29 | 2021-05-18 | Apple Inc. | Virtual assistant for media playback |
US10606322B2 (en) | 2015-06-30 | 2020-03-31 | Microsoft Technology Licensing, Llc | Multistage friction hinge |
US9864415B2 (en) | 2015-06-30 | 2018-01-09 | Microsoft Technology Licensing, Llc | Multistage friction hinge |
US11323451B2 (en) | 2015-07-09 | 2022-05-03 | Biocatch Ltd. | System, device, and method for detection of proxy server |
US10523680B2 (en) * | 2015-07-09 | 2019-12-31 | Biocatch Ltd. | System, device, and method for detecting a proxy server |
US10834090B2 (en) * | 2015-07-09 | 2020-11-10 | Biocatch Ltd. | System, device, and method for detection of proxy server |
US11500672B2 (en) | 2015-09-08 | 2022-11-15 | Apple Inc. | Distributed personal assistant |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US11126400B2 (en) | 2015-09-08 | 2021-09-21 | Apple Inc. | Zero latency digital assistant |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US11526368B2 (en) | 2015-11-06 | 2022-12-13 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10354652B2 (en) | 2015-12-02 | 2019-07-16 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10942703B2 (en) | 2015-12-23 | 2021-03-09 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US10344797B2 (en) | 2016-04-05 | 2019-07-09 | Microsoft Technology Licensing, Llc | Hinge with multiple preset positions |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US11227589B2 (en) | 2016-06-06 | 2022-01-18 | Apple Inc. | Intelligent list reading |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
US11069347B2 (en) | 2016-06-08 | 2021-07-20 | Apple Inc. | Intelligent automated assistant for media exploration |
US10354011B2 (en) | 2016-06-09 | 2019-07-16 | Apple Inc. | Intelligent automated assistant in a home environment |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US11037565B2 (en) | 2016-06-10 | 2021-06-15 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10733993B2 (en) | 2016-06-10 | 2020-08-04 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10297253B2 (en) | 2016-06-11 | 2019-05-21 | Apple Inc. | Application integration with a digital assistant |
US11152002B2 (en) | 2016-06-11 | 2021-10-19 | Apple Inc. | Application integration with a digital assistant |
US10269345B2 (en) | 2016-06-11 | 2019-04-23 | Apple Inc. | Intelligent task discovery |
US10089072B2 (en) | 2016-06-11 | 2018-10-02 | Apple Inc. | Intelligent device arbitration and control |
US10521466B2 (en) | 2016-06-11 | 2019-12-31 | Apple Inc. | Data driven natural language event detection and classification |
US10580409B2 (en) | 2016-06-11 | 2020-03-03 | Apple Inc. | Application integration with a digital assistant |
US10942702B2 (en) | 2016-06-11 | 2021-03-09 | Apple Inc. | Intelligent device arbitration and control |
US11055395B2 (en) | 2016-07-08 | 2021-07-06 | Biocatch Ltd. | Step-up authentication |
US10474753B2 (en) | 2016-09-07 | 2019-11-12 | Apple Inc. | Language identification using recurrent neural networks |
US10037057B2 (en) | 2016-09-22 | 2018-07-31 | Microsoft Technology Licensing, Llc | Friction hinge |
US10553215B2 (en) | 2016-09-23 | 2020-02-04 | Apple Inc. | Intelligent automated assistant |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US10579784B2 (en) | 2016-11-02 | 2020-03-03 | Biocatch Ltd. | System, device, and method of secure utilization of fingerprints for user authentication |
US10685355B2 (en) * | 2016-12-04 | 2020-06-16 | Biocatch Ltd. | Method, device, and system of detecting mule accounts and accounts used for money laundering |
US11281993B2 (en) | 2016-12-05 | 2022-03-22 | Apple Inc. | Model and ensemble compression for metric learning |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US11656884B2 (en) | 2017-01-09 | 2023-05-23 | Apple Inc. | Application integration with a digital assistant |
US11204787B2 (en) | 2017-01-09 | 2021-12-21 | Apple Inc. | Application integration with a digital assistant |
US10741181B2 (en) | 2017-05-09 | 2020-08-11 | Apple Inc. | User interface for correcting recognition errors |
US10332518B2 (en) | 2017-05-09 | 2019-06-25 | Apple Inc. | User interface for correcting recognition errors |
US10417266B2 (en) | 2017-05-09 | 2019-09-17 | Apple Inc. | Context-aware ranking of intelligent response suggestions |
US10395654B2 (en) | 2017-05-11 | 2019-08-27 | Apple Inc. | Text normalization based on a data-driven learning network |
US10847142B2 (en) | 2017-05-11 | 2020-11-24 | Apple Inc. | Maintaining privacy of personal information |
US11599331B2 (en) | 2017-05-11 | 2023-03-07 | Apple Inc. | Maintaining privacy of personal information |
US10726832B2 (en) | 2017-05-11 | 2020-07-28 | Apple Inc. | Maintaining privacy of personal information |
US10755703B2 (en) | 2017-05-11 | 2020-08-25 | Apple Inc. | Offline personal assistant |
US11380310B2 (en) | 2017-05-12 | 2022-07-05 | Apple Inc. | Low-latency intelligent automated assistant |
US10789945B2 (en) | 2017-05-12 | 2020-09-29 | Apple Inc. | Low-latency intelligent automated assistant |
US11301477B2 (en) | 2017-05-12 | 2022-04-12 | Apple Inc. | Feedback analysis of a digital assistant |
US10791176B2 (en) | 2017-05-12 | 2020-09-29 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US10410637B2 (en) | 2017-05-12 | 2019-09-10 | Apple Inc. | User-specific acoustic models |
US11405466B2 (en) | 2017-05-12 | 2022-08-02 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US10810274B2 (en) | 2017-05-15 | 2020-10-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
US10482874B2 (en) | 2017-05-15 | 2019-11-19 | Apple Inc. | Hierarchical belief states for digital assistants |
US11217255B2 (en) | 2017-05-16 | 2022-01-04 | Apple Inc. | Far-field extension for digital assistant services |
US10909171B2 (en) | 2017-05-16 | 2021-02-02 | Apple Inc. | Intelligent automated assistant for media exploration |
US10403278B2 (en) | 2017-05-16 | 2019-09-03 | Apple Inc. | Methods and systems for phonetic matching in digital assistant services |
US10311144B2 (en) | 2017-05-16 | 2019-06-04 | Apple Inc. | Emoji word sense disambiguation |
US10748546B2 (en) | 2017-05-16 | 2020-08-18 | Apple Inc. | Digital assistant services based on device capabilities |
US11532306B2 (en) | 2017-05-16 | 2022-12-20 | Apple Inc. | Detecting a trigger of a digital assistant |
US10303715B2 (en) | 2017-05-16 | 2019-05-28 | Apple Inc. | Intelligent automated assistant for media exploration |
US10657328B2 (en) | 2017-06-02 | 2020-05-19 | Apple Inc. | Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling |
US10397262B2 (en) | 2017-07-20 | 2019-08-27 | Biocatch Ltd. | Device, system, and method of detecting overlay malware |
US10417339B2 (en) | 2017-08-18 | 2019-09-17 | Kyocera Document Solutions Inc. | Suggestion of alternate user input using different user interface |
US10445429B2 (en) | 2017-09-21 | 2019-10-15 | Apple Inc. | Natural language understanding using vocabularies with compressed serialized tries |
US10755051B2 (en) | 2017-09-29 | 2020-08-25 | Apple Inc. | Rule-based natural language processing |
US10970394B2 (en) | 2017-11-21 | 2021-04-06 | Biocatch Ltd. | System, device, and method of detecting vishing attacks |
US10636424B2 (en) | 2017-11-30 | 2020-04-28 | Apple Inc. | Multi-turn canned dialog |
US10733982B2 (en) | 2018-01-08 | 2020-08-04 | Apple Inc. | Multi-directional dialog |
US10733375B2 (en) | 2018-01-31 | 2020-08-04 | Apple Inc. | Knowledge-based framework for improving natural language understanding |
US10789959B2 (en) | 2018-03-02 | 2020-09-29 | Apple Inc. | Training speaker recognition models for digital assistants |
US10592604B2 (en) | 2018-03-12 | 2020-03-17 | Apple Inc. | Inverse text normalization for automatic speech recognition |
US11710482B2 (en) | 2018-03-26 | 2023-07-25 | Apple Inc. | Natural assistant interaction |
US10818288B2 (en) | 2018-03-26 | 2020-10-27 | Apple Inc. | Natural assistant interaction |
US10909331B2 (en) | 2018-03-30 | 2021-02-02 | Apple Inc. | Implicit identification of translation payload with neural machine translation |
US10928918B2 (en) | 2018-05-07 | 2021-02-23 | Apple Inc. | Raise to speak |
US11169616B2 (en) | 2018-05-07 | 2021-11-09 | Apple Inc. | Raise to speak |
US11145294B2 (en) | 2018-05-07 | 2021-10-12 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US11854539B2 (en) | 2018-05-07 | 2023-12-26 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US10984780B2 (en) | 2018-05-21 | 2021-04-20 | Apple Inc. | Global semantic word embeddings using bi-directional recurrent neural networks |
US10984798B2 (en) | 2018-06-01 | 2021-04-20 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US11431642B2 (en) | 2018-06-01 | 2022-08-30 | Apple Inc. | Variable latency device coordination |
US10403283B1 (en) | 2018-06-01 | 2019-09-03 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US10720160B2 (en) | 2018-06-01 | 2020-07-21 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US11386266B2 (en) | 2018-06-01 | 2022-07-12 | Apple Inc. | Text correction |
US10684703B2 (en) | 2018-06-01 | 2020-06-16 | Apple Inc. | Attention aware virtual assistant dismissal |
US11495218B2 (en) | 2018-06-01 | 2022-11-08 | Apple Inc. | Virtual assistant operation in multi-device environments |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
US11009970B2 (en) | 2018-06-01 | 2021-05-18 | Apple Inc. | Attention aware virtual assistant dismissal |
US10504518B1 (en) | 2018-06-03 | 2019-12-10 | Apple Inc. | Accelerated task performance |
US10496705B1 (en) | 2018-06-03 | 2019-12-03 | Apple Inc. | Accelerated task performance |
US10944859B2 (en) | 2018-06-03 | 2021-03-09 | Apple Inc. | Accelerated task performance |
US11010561B2 (en) | 2018-09-27 | 2021-05-18 | Apple Inc. | Sentiment prediction from textual data |
US11462215B2 (en) | 2018-09-28 | 2022-10-04 | Apple Inc. | Multi-modal inputs for voice commands |
US11170166B2 (en) | 2018-09-28 | 2021-11-09 | Apple Inc. | Neural typographical error modeling via generative adversarial networks |
US10839159B2 (en) | 2018-09-28 | 2020-11-17 | Apple Inc. | Named entity normalization in a spoken dialog system |
US11475898B2 (en) | 2018-10-26 | 2022-10-18 | Apple Inc. | Low-latency multi-speaker speech recognition |
US11638059B2 (en) | 2019-01-04 | 2023-04-25 | Apple Inc. | Content playback on multiple devices |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
US11423908B2 (en) | 2019-05-06 | 2022-08-23 | Apple Inc. | Interpreting spoken requests |
US11217251B2 (en) | 2019-05-06 | 2022-01-04 | Apple Inc. | Spoken notifications |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
US11475884B2 (en) | 2019-05-06 | 2022-10-18 | Apple Inc. | Reducing digital assistant latency when a language is incorrectly determined |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
US11657813B2 (en) | 2019-05-31 | 2023-05-23 | Apple Inc. | Voice identification in digital assistant systems |
US11360739B2 (en) | 2019-05-31 | 2022-06-14 | Apple Inc. | User activity shortcut suggestions |
US11496600B2 (en) | 2019-05-31 | 2022-11-08 | Apple Inc. | Remote execution of machine-learned models |
US11289073B2 (en) | 2019-05-31 | 2022-03-29 | Apple Inc. | Device text to speech |
US11237797B2 (en) | 2019-05-31 | 2022-02-01 | Apple Inc. | User activity shortcut suggestions |
US11360641B2 (en) | 2019-06-01 | 2022-06-14 | Apple Inc. | Increasing the relevance of new available information |
US11488406B2 (en) | 2019-09-25 | 2022-11-01 | Apple Inc. | Text detection using global geometry estimators |
US11181988B1 (en) | 2020-08-31 | 2021-11-23 | Apple Inc. | Incorporating user feedback into text prediction models via joint reward planning |
US11606353B2 (en) | 2021-07-22 | 2023-03-14 | Biocatch Ltd. | System, device, and method of generating and utilizing one-time passwords |
Also Published As
Publication number | Publication date |
---|---|
US8065624B2 (en) | 2011-11-22 |
WO2009006221A1 (en) | 2009-01-08 |
EP2160672A4 (en) | 2011-07-27 |
US20090007001A1 (en) | 2009-01-01 |
EP2160672A1 (en) | 2010-03-10 |
JP2010536076A (en) | 2010-11-25 |
JP5323070B2 (en) | 2013-10-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8065624B2 (en) | Virtual keypad systems and methods | |
US20220261112A1 (en) | Systems, devices, and methods for touch-free typing | |
US10359932B2 (en) | Method and apparatus for providing character input interface | |
US10191573B2 (en) | Pointer display device, pointer display/detection method, pointer display/detection program and information apparatus | |
US20100259482A1 (en) | Keyboard gesturing | |
US20160364138A1 (en) | Front touchscreen and back touchpad operated user interface employing semi-persistent button groups | |
US20170017393A1 (en) | Method for controlling interactive objects from a touchpad of a computerized device | |
US20050162402A1 (en) | Methods of interacting with a computer using a finger(s) touch sensing input device with visual feedback | |
KR20130088752A (en) | Multidirectional button, key, and keyboard | |
CN104360813A (en) | Display equipment and information processing method thereof | |
KR20160097410A (en) | Method of providing touchless input interface based on gesture recognition and the apparatus applied thereto | |
Benko et al. | Imprecision, inaccuracy, and frustration: The tale of touch input | |
CN111007977A (en) | Intelligent virtual interaction method and device | |
CN105700806A (en) | Directional gesture input method and device for intelligent televisions or intelligent glasses | |
JP2016076232A (en) | Display device and control method for the same | |
WO2015013662A1 (en) | Method for controlling a virtual keyboard from a touchpad of a computerized device | |
Ni | A framework of freehand gesture interaction: techniques, guidelines, and applications | |
WO2021178255A1 (en) | Electronic input system | |
CN114040857A (en) | Method for operating an operating system in a vehicle and operating system in a vehicle | |
US20150106764A1 (en) | Enhanced Input Selection | |
EP4400959A1 (en) | Touchless user interface operating method based on an electronic device | |
Lee et al. | Embodied interaction on constrained interfaces for augmented reality | |
KR20150049661A (en) | Apparatus and method for processing input information of touchpad | |
WO2018035353A1 (en) | Front touchscreen and back touchpad operated user interface employing semi-persistent button groups | |
Gaur | AUGMENTED TOUCH INTERACTIONS WITH FINGER CONTACT SHAPE AND ORIENTATION |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |