Patent application number | Description | Published |
20090055739 | CONTEXT-AWARE ADAPTIVE USER INTERFACE - Technologies, systems, and methods for context-aware adaptation of user interface where monitored context includes ambient environmental and temporal conditions, user state, and the like. For example, when a user has been using an application for a long time, ambient lighting conditions are becoming darker, and the user is inferred to be experiencing increased eye strain and fatigue, the user interface may be adapted by increasing the contrast. Such adaptation may be based on rules, pre-defined or otherwise. The processing of sensor data typically results in context codes and detection of context patterns that may be used to adapt user interface for an optimized user experience. | 02-26-2009 |
20100318293 | RETRACING STEPS - Techniques for creating breadcrumbs for a trail of activity are described. The trail of activity may be created by recording movement information based on inferred actions of walking, not walking, or changing floor levels. The movement information may be recorded with an accelerometer and a pressure sensor. A representation of a list of breadcrumbs may be visually displayed on a user interface of a mobile device, in a reverse order to retrace steps. In some implementations, a compass may additionally or alternatively be used to collect directional information relative to the earth's magnetic poles. | 12-16-2010 |
20110083089 | MONITORING POINTER TRAJECTORY AND MODIFYING DISPLAY INTERFACE - Apparatus and methods for improving touch-screen interface usability and accuracy by determining the trajectory of a pointer as it approaches the touch-screen and modifying the touch-screen display accordingly. The system may predict an object on the display the user is likely to select next. The system may designate this object as a Designated Target Object, or DTO. The system may modify the appearance of the DTO by, for example, changing the size of the DTO, or by changing its shape, style, coloring, perspective, positioning, etc. | 04-07-2011 |
20110173204 | ASSIGNING GESTURE DICTIONARIES - Techniques for assigning a gesture dictionary in a gesture-based system to a user comprise capturing data representative of a user in a physical space. In a gesture-based system, gestures may control aspects of a computing environment or application, where the gestures may be derived from a user's position or movement in a physical space. In an example embodiment, the system may monitor a user's gestures and select a particular gesture dictionary in response to the manner in which the user performs the gestures. The gesture dictionary may be assigned in real time with respect to the capture of the data representative of a user's gesture. The system may generate calibration tests for assigning a gesture dictionary. The system may track the user during a set of short gesture calibration tests and assign the gesture dictionary based on a compilation of the data captured that represents the user's gestures. | 07-14-2011 |
20120105257 | Multimodal Input System - The subject disclosure relates to user input into a computer system, and a technology by which one or more users interact with a computer system via a combination of input modalities. When the input data of two or more input modalities are related, they are combined to interpret an intended meaning of the input. For example, speech when combined with one input gesture has one intended meaning, e.g., convert the speech to verbatim text for consumption by a program, while the exact speech when combined with a different input gesture has a different meaning, e.g., convert the speech to a command that controls the operation of that same program. | 05-03-2012 |
20120109868 | Real-Time Adaptive Output - The subject disclosure relates to a technology by which output data in the form of audio, visual, haptic, and/or other output is automatically selected and tailored by a system, including adapting in real time, to address one or more users' specific needs, context and implicit/explicit intent. State data and preference data are input into a real time adaptive output system that uses the data to select among output modalities, e.g., to change output mechanisms, add/remove output mechanisms, and/or change rendering characteristics. The output may be rendered on one or more output mechanisms to a single user or multiple users, including via a remote output mechanism. | 05-03-2012 |
20120124126 | CONTEXTUAL AND TASK FOCUSED COMPUTING - Concepts and technologies are described herein for contextual and task-focused computing. In accordance with the concepts and technologies disclosed herein, a discovery engine analyzes application data describing applications, recognizes tasks associated with the applications, and stores task data identifying and describing the tasks in a data storage location. The task data is searchable by search engines, indexing and search services, and task engines configured to provide tasks to one or more client devices operating alone or in a synchronized manner, the tasks being provided on demand or based upon activity associated with the one or more client devices. A task engine receives or obtains contextual data describing context associate with the client devices and/or social networking data associated with one or more users of the client devices. Based upon the contextual data and/or the social networking data, the task engine identifies one or more relevant tasks and provides to the client devices information for accessing the relevant tasks, or packaged data corresponding to the relevant tasks. | 05-17-2012 |
20120143681 | ROOM-BASED COMPUTING ENVIRONMENTS - Concepts and technologies for creating and accessing room-based computing environments are disclosed. Resources are categorized and/or bundled into categories or bundles of resources. Resources are associated with the room-based computing environment and various data relating to the resources is stored, including data relating to permissions for accessing the resources. Upon detecting access of the room-based computing environment, a room engine can authenticate an entity associated with the access and determine what contents of the room-based computing environment are to be presented based upon the permissions information and/or other considerations. The environment is generated and presented to the entity via one or more user interfaces. | 06-07-2012 |
20120150784 | Immersive Planning of Events Including Vacations - The subject disclosure is directed towards a web service or the like that assists users in generating a plan, such as a vacation plan. In one aspect, a user chooses a model that generates a plan, including by selecting content objects (e.g., found by searching) corresponding to plan objects. Selection is based upon user input, along with rules, constraints and/or equations associated with the model. A presentation mechanism produces a presentation (e.g., an audiovisual experience) from the content/plan objects, such as a linear narrative, a timeline, a schedule, a calendar, a gallery, a list, and/or a map. The plan may be annotated with annotation data. The plan may be interacted with to re-plan it, and may be saved and/or provided to another user for viewing and/or re-planning. Plan versions may be compared to see the changes made. | 06-14-2012 |
20120150787 | Addition of Plan-Generation Models and Expertise by Crowd Contributorst - The subject disclosure is directed towards a web service that maintains a set of models used to generate plans, such as vacation plans, in which the set of models includes models that are authored by crowd contributors via the service. The models include rules, constraints and/or equations, and may be text based and declarative such that any author can edit an existing model or combination of existing models into a new model. Users can access the models to generate a plan according to user parameters, view a presentation of that plan, and interact to provide new parameters to the model and/or with objects in the plan to modify the plan and view a presentation of the modified plan. | 06-14-2012 |
20120151348 | Using Cinematographic Techniques for Conveying and Interacting with Plan Sagas - The subject disclosure is directed towards obtaining a linear narrative synthesized from a set of objects, such as objects corresponding to a plan, and using cinematographic and other effects to convey additional information with that linear narrative when presented to a user. A user interacts with data from which the linear narrative is synthesized, such as to add transition effects between objects, change the lighting, focus, size (zoom), pan and so forth to emphasize or de-emphasize an object, and/or to highlight a relationship between objects. A user instruction may correspond to a theme (e.g., style or mood), with the effects, possibly including audio, selected based upon that theme. | 06-14-2012 |
20120151350 | Synthesis of a Linear Narrative from Search Content - The subject disclosure is directed towards automatically synthesizing content found via one or more searches into a linear narrative such as a slideshow and/or other audiovisual presentation, for playback to a user. A model in conjunction with user input parameters may assist in obtaining the search content, comprising content objects. The model applies rules, constraints and/or equations to generate a plan comprising plan objects, and a content synthesizer processes the plan objects into the linear narrative. The user may interact to change the input parameters and/or the set of plan objects, resulting in a modified narrative being re-synthesized for playback. | 06-14-2012 |
20120159326 | RICH INTERACTIVE SAGA CREATION - One or more techniques and/or systems are disclosed for creating a saga from signal-rich digital memories. User-related content, such as media elements and/or other signals, are captured and used to generate a digital memory graph, comprising the captured user-related content and associated metadata. An interactive saga of digital media elements is created using the digital memory graph by combining at least a portion of a plurality of digital media elements, from the captured user-related content, based on one or more user interactions. | 06-21-2012 |
20120159341 | INTERACTIONS WITH CONTEXTUAL AND TASK-BASED COMPUTING ENVIRONMENTS - Concepts and technologies are described herein for interacting with contextual and task-focused computing environments. Tasks associated with applications are described by task data. Tasks and/or batches of tasks relevant to activities occurring at a client are identified, and a UI for presenting the tasks is generated. The UIs can include tasks and workflows corresponding to batches of tasks. Workflows can be executed, interrupted, and resumed on demand. Interrupted workflows are stored with data indicating progress, contextual information, UI information, and other information. The workflow is stored and/or shared. When execution of the workflow is resumed, the same or a different UI can be provided, based upon the device used to resume execution of the workflow. Thus, multiple devices and users can access workflows in parallel to provide collaborative task execution. | 06-21-2012 |
20120166411 | DISCOVERY OF REMOTELY EXECUTED APPLICATIONS - A search engine discovers and indexes applications in a search index and receives queries from devices. The search engine is configured to obtain contextual data describing context associated with the devices and/or social networking data associated with one or more users of the devices. Based upon the contextual data and/or the social networking data, the search engine modifies the query and executes the query to identify applications. The search engine generates search results corresponding to the identified applications. The search engine also is configured to generate advertising relevant to the modified query, and to rank the search results in accordance with the query, the contextual data, and/or the social networking data. The ranked search results and the advertising are presented to the client as search results and/or in a web store format. Activity of the client and the search engine can be tracked and reported to authorized entities. | 06-28-2012 |
20120166522 | SUPPORTING INTELLIGENT USER INTERFACE INTERACTIONS - Concepts and technologies are described herein for supporting intelligent user interface interactions. Commands accepted by applications can be published or determined. Before or during access of the application, the commands can be presented at clients to indicate commands available for interfacing with the application. The commands can be presented with information indicating how the user interface and/or input device of the client may be used to execute the available commands. Input received from the client can be compared to the available commands to determine if the input matches an available command. Contextual data relating to the client, preferences, and/or other data also can be retrieved and analyzed to determine the intent of the client. The intent can be used to identify an intended command and to modify the input to match the intended command. The modified input can be transmitted to the application. | 06-28-2012 |
20120293439 | MONITORING POINTER TRAJECTORY AND MODIFYING DISPLAY INTERFACE - Apparatus and methods for improving touch-screen interface usability and accuracy by determining the trajectory of a pointer as it approaches the touch-screen and modifying the touch-screen display accordingly. The system may predict an object on the display the user is likely to select next. The system may designate this object as a Designated Target Object, or DTO. The system may modify the appearance of the DTO by, for example, changing the size of the DTO, or by changing its shape, style, coloring, perspective, positioning, etc. | 11-22-2012 |
20140109023 | ASSIGNING GESTURE DICTIONARIES - Techniques for assigning a gesture dictionary in a gesture-based system to a user comprise capturing data representative of a user in a physical space. In a gesture-based system, gestures may control aspects of a computing environment or application, where the gestures may be derived from a user's position or movement in a physical space. In an example embodiment, the system may monitor a user's gestures and select a particular gesture dictionary in response to the manner in which the user performs the gestures. The gesture dictionary may be assigned in real time with respect to the capture of the data representative of a user's gesture. The system may generate calibration tests for assigning a gesture dictionary. The system may track the user during a set of short gesture calibration tests and assign the gesture dictionary based on a compilation of the data captured that represents the user's gestures. | 04-17-2014 |