Patent application number | Description | Published |
20100207904 | TARGETING IN A STYLUS-BASED USER INTERFACE - Aspects of the invention provide virtual hover zones. When a user lowers a hovering stylus while remaining within a hover zone, cursor control is modified to be more easily controllable by the user. If the user pauses the stylus in mid-air before lowering the stylus, and if the stylus remains within the hover zone, then upon touchdown the cursor may be moved to the projection of the location where the stylus was paused. Any action that may be taken in response to the touch down may be sent to the projection location as well. Also provided are cursor control zones. A dampening zone may be used to provide dampened cursor movement feedback in response to movement input provided by a pointing device. Also, a dead zone may be used to prohibit cursor movement in response to movement input provided by the pointing device. | 08-19-2010 |
20100318930 | ASSISTING USER INTERFACE ELEMENT USE - Methods of controlling the display and use of a UI element are disclosed. In an embodiment, the UI element may configured so that it initially maintains a topmost position but eventually allows other applications to assume the topmost position. In an embodiment, the display of the element may be adjusted in response to an input so that the UI element is not visible on the display. In an embodiment, the use of the UI element may allow for seamless dragging of the UI element even if the user inadvertently fails to make consistent contact with the touch-sensitive display while dragging the UI element. | 12-16-2010 |
20110157062 | Touch Input Data Handling - A system for enabling a tablet input object is described. A tablet input object can take various inputs from touch, a mouse, and a pen and sends their information to an application or operating system. Also, a pen message pathway may also be used to handle touch messages, thereby reusing an existing pen message pathway for messages created by something other than a pen. | 06-30-2011 |
20110216028 | Methods For Allowing Applications To Filter Out Or Opt Into Tablet Input - Methods and systems for enabling a tablet input object is described. A tablet input object can take various inputs from touch, a mouse, and a pen and send their information to an application. | 09-08-2011 |
20130044070 | Unintentional Touch Rejection - A method for rejecting an unintentional palm touch is disclosed. In at least some embodiments, a touch is detected by a touch-sensitive surface associated with a display. Characteristics of the touch may be used to generate a set of parameters related to the touch. In an embodiment, firmware is used to determine a reliability value for the touch. The reliability value and the location of the touch is provided to a software module. The software module uses the reliability value and an activity context to determine a confidence level of the touch. In an embodiment, the confidence level may include an evaluation of changes in the reliability value over time. If the confidence level for the touch is too low, it may be rejected. | 02-21-2013 |
20130314316 | TARGETING IN A STYLUS-BASED USER INTERFACE - Aspects of the invention provide virtual hover zones. When a user lowers a hovering stylus while remaining within a hover zone, cursor control is modified to be more easily controllable by the user. If the user pauses the stylus in mid-air before lowering the stylus, and if the stylus remains within the hover zone, then upon touchdown the cursor may be moved to the projection of the location where the stylus was paused. Any action that may be taken in response to the touch down may be sent to the projection location as well. Also provided are cursor control zones. A dampening zone may be used to provide dampened cursor movement feedback in response to movement input provided by a pointing device. Also, a dead zone may be used to prohibit cursor movement in response to movement input provided by the pointing device. | 11-28-2013 |
20140111462 | Unintentional Touch Rejection - A method for rejecting an unintentional palm touch is disclosed. In at least some embodiments, a touch is detected by a touch-sensitive surface associated with a display. Characteristics of the touch may be used to generate a set of parameters related to the touch. In an embodiment, firmware is used to determine a reliability value for the touch. The reliability value and the location of the touch is provided to a software module. The software module uses the reliability value and an activity context to determine a confidence level of the touch. In an embodiment, the confidence level may include an evaluation of changes in the reliability value over time. If the confidence level for the touch is too low, it may be rejected. | 04-24-2014 |
20140285457 | Touch Input Data Handling - A system for enabling a tablet input object is described. A tablet input object can take various inputs from touch, a mouse, and a pen and sends their information to an application or operating system. Also, a pen message pathway may also be used to handle touch messages, thereby reusing an existing pen message pathway for messages created by something other than a pen. | 09-25-2014 |
Patent application number | Description | Published |
20090195506 | Dynamic Soft Keyboard - In accordance with one or more aspects of a dynamic soft keyboard, a user input is received via a soft keyboard having multiple keys. Information describing a current input environment for the soft keyboard is obtained, and a determination is made as to which one or more keys of the multiple keys was intended to be selected by the user input. This determination is made based at least in part on the current input environment. | 08-06-2009 |
20090284478 | Multi-Contact and Single-Contact Input - This document describes tools capable of initiating a function based on one or more tactile contacts received through a contact detection device, such as a touch pad. In some embodiments, the tools identify tactile contacts in accordance with the tool's input mode. The tools may use the input mode to determine what gestures may be identified for the tactile contacts. In some embodiments, these tools switch input modes based on a number or characteristic of tactile contacts electronically represented in contact data. By so doing, the tools may more accurately determine appropriate gestures or provide a broader range of functions based on tactile contacts received through a contact detection device. | 11-19-2009 |
20090284479 | Multi-Touch Input Platform - This document describes tools having or interacting with a touch-sensitive device with one or more contact detectors that detect a tactile input from a user making contact with the contact detectors and an input/output module that persistently identifies contact data representing the detected tactile input. This identification is unique, thereby differentiating one or more tactile inputs from other current tactile inputs represented in the contact data. Using this unique identification, the input/output module can initiate an application to provide a function, such as data entry or a mapped function, associated with the detected tactile input. These data or mapped functions may, in many instances, provide a greater depth or breadth of functions by which users may interact with applications and computer systems. | 11-19-2009 |
20100103117 | MULTI-TOUCH MANIPULATION OF APPLICATION OBJECTS - The manipulation system described herein provides a common platform and application-programming interface (API) for applications to communicate with various multi-touch hardware devices, and facilitates the interpretation of multi-touch input as one or more manipulations. Manipulations map more directly to user intentions than do individual touch inputs and add support for basic transformation of objects using multiple touch contacts. An application can use manipulations to support rotating, resizing, and translating multiple objects at the same time. The manipulation system outputs two-dimensional (2D) affine transforms that contain rotation, scale, and translation information. Thus, using the manipulation system the application author can focus more on building touch-capable applications and let the manipulation system handle the underlying transformations and communication with the multi-touch hardware. | 04-29-2010 |
20100103118 | MULTI-TOUCH OBJECT INERTIA SIMULATION - The inertia system provides a common platform and application-programming interface (API) for applications to extend the input received from various multi-touch hardware devices to simulate real-world behavior of application objects. To move naturally, application objects should exhibit physical characteristics such as elasticity and deceleration. When a user lifts all contacts from an object, the inertia system provides additional manipulation events to the application so that the application can handle the events as if the user was still moving the object with touch. The inertia system generates the events based on a simulation of the behavior of the objects. If the user moves an object into another object, the inertia system simulates the boundary characteristics of the objects. Thus, the inertia system provides more realistic movement for application objects manipulated using multi-touch hardware and the API provides a consistent feel to manipulations across applications. | 04-29-2010 |
20130067382 | SOFT KEYBOARD INTERFACE - User experience of the performance of a computing device is improved through an operating system that processes inputs from a soft keyboard to provide information that can be used to accurately determine keys a user intended to strike while typing. For each detected tap, the operating system provides a probability that one or more keys were the intended target for the user. These probabilities may be computed from probability distribution functions that are dynamically determined based on user and/or system factors, such as typing rate and keyboard style or layout. Other components may use the probabilities to select a key corresponding to a detected keyboard tap as representing the intended user input. The selection may be made based on the probabilities alone or in combination with contextual factors that yield an overall probability that a detected tap arose from a user targeting a specific key. | 03-14-2013 |
20130278529 | MULTI-TOUCH MANIPULATION OF APPLICATION OBJECTS - The manipulation system described herein provides a common platform and application-programming interface (API) for applications to communicate with various multi-touch hardware devices, and facilitates the interpretation of multi-touch input as one or more manipulations. Manipulations map more directly to user intentions than do individual touch inputs and add support for basic transformation of objects using multiple touch contacts. An application can use manipulations to support rotating, resizing, and translating multiple objects at the same time. The manipulation system outputs two-dimensional (2D) affine transforms that contain rotation, scale, and translation information. Thus, using the manipulation system the application author can focus more on building touch-capable applications and let the manipulation system handle the underlying transformations and communication with the multi-touch hardware. | 10-24-2013 |
20130285961 | MULTI-TOUCH OBJECT INERTIA SIMULATION - The inertia system provides a common platform and application-programming interface (API) for applications to extend the input received from various multi-touch hardware devices to simulate real-world behavior of application objects. To move naturally, application objects should exhibit physical characteristics such as elasticity and deceleration. When a user lifts all contacts from an object, the inertia system provides additional manipulation events to the application so that the application can handle the events as if the user was still moving the object with touch. The inertia system generates the events based on a simulation of the behavior of the objects. If the user moves an object into another object, the inertia system simulates the boundary characteristics of the objects. Thus, the inertia system provides more realistic movement for application objects manipulated using multi-touch hardware and the API provides a consistent feel to manipulations across applications. | 10-31-2013 |
20150022478 | MULTI-TOUCH MANIPULATION OF APPLICATION OBJECTS - The manipulation system described herein provides a common platform and application-programming interface (API) for applications to communicate with various multi-touch hardware devices, and facilitates the interpretation of multi-touch input as one or more manipulations. Manipulations map more directly to user intentions than do individual touch inputs and add support for basic transformation of objects using multiple touch contacts. An application can use manipulations to support rotating, resizing, and translating multiple objects at the same time. The manipulation system outputs two-dimensional (2D) affine transforms that contain rotation, scale, and translation information. Thus, using the manipulation system the application author can focus more on building touch-capable applications and let the manipulation system handle the underlying transformations and communication with the multi-touch hardware. | 01-22-2015 |
Patent application number | Description | Published |
20090292989 | PANNING CONTENT UTILIZING A DRAG OPERATION - Computer-readable media, computerized methods, and computer systems for intuitively invoking a panning action (e.g., moving content within a content region of a display area) by applying a user-initiated input at the content region rendered at a touchscreen interface are provided. Initially, aspects of the user-initiated input include a location of actuation (e.g., touch point on the touchscreen interface) and a gesture. Upon ascertaining that the actuation location occurred within the content region and that the gesture is a drag operation, based on a distance of uninterrupted tactile contact with the touchscreen interface, a panning mode may be initiated. When in the panning mode, and if the application rendering the content at the display area supports scrolling functionality, the gesture will control movement of the content within the content region. In particular, the drag operation of the gesture will pan the content within the display area when surfaced at the touchscreen interface. | 11-26-2009 |
20120001861 | Unintentional Touch Rejection - A method for rejecting an unintentional palm touch is disclosed. In at least some embodiments, a touch is detected by a touch-sensitive surface associated with a display. Characteristics of the touch may be used to generate a set of parameters related to the touch. In an embodiment, firmware is used to determine a reliability value for the touch. The reliability value and the location of the touch is provided to a software module. The software module uses the reliability value and an activity context to determine a confidence level of the touch. In an embodiment, the confidence level may include an evaluation of changes in the reliability value over time. If the confidence level for the touch is too low, it may be rejected. | 01-05-2012 |
20130147749 | PANNING CONTENT UTILIZING A DRAG OPERATION - Computer-readable media, computerized methods, and computer systems for intuitively invoking a panning action (e.g., moving content within a content region of a display area) by applying a user-initiated input at the content region rendered at a touchscreen interface are provided. Initially, aspects of the user-initiated input include a location of actuation (e.g., touch point on the touchscreen interface) and a gesture. Upon ascertaining that the actuation location occurred within the content region and that the gesture is a drag operation, based on a distance of uninterrupted tactile contact with the touchscreen interface, a panning mode may be initiated. When in the panning mode, and if the application rendering the content at the display area supports scrolling functionality, the gesture will control movement of the content within the content region. In particular, the drag operation of the gesture will pan the content within the display area when surfaced at the touchscreen interface. | 06-13-2013 |