Patent application number | Description | Published |
20090167700 | INSERTION MARKER PLACEMENT ON TOUCH SENSITIVE DISPLAY - In accordance with some embodiments, a computer-implemented method is performed at a portable electronic device with a touch screen display. The method can include: displaying graphics on the touch screen display, detecting a finger contact on the touch screen display, and, in response to the detected finger contact, inserting an insertion marker in the graphics at a first location. The method can further include detecting a finger movement on the touch screen display and, irrespective of initial distance from finger to insertion marker on the touch screen display, moving the insertion marker in accordance with the detected finger movement from the first location to a second location in the graphics. | 07-02-2009 |
20100171712 | Device, Method, and Graphical User Interface for Manipulating a User Interface Object - In some embodiments, an electronic device with a display and a touch-sensitive surface displays a user interface object. The device detects a first contact and a second contact concurrently on the touch-sensitive surface. The device determines which contact of the first contact and the second contact is a topmost contact, a bottommost contact, a leftmost contact, and a rightmost contact on the touch-sensitive surface. While continuing to detect the first contact and the second contact, the device detects movement of the first contact across the touch-sensitive surface, and concurrently moves two edges of the user interface object that correspond to the first contact in accordance with the detected movement of the first contact, including horizontally moving one of the two edges and vertically moving the other of the two edges. | 07-08-2010 |
20110069016 | Device, Method, and Graphical User Interface for Manipulating User Interface Objects - A computing device with a touch screen display simultaneously displays on the touch screen display a plurality of user interface objects and at least one destination object. The computing device detects a first input by a user on a destination object displayed on the touch screen display. While continuing to detect the first input by the user on the destination object, the computing device detects a second input by the user on a first user interface object displayed on the touch screen display. In response to detecting the second input by the user on the first user interface object, the computing device performs an action on the first user interface object. The action is associated with the destination object. | 03-24-2011 |
20110069017 | Device, Method, and Graphical User Interface for Manipulating User Interface Objects - A computing device with a touch screen display simultaneously displays on the touch screen display a plurality of user interface objects and at least one destination object. The computing device detects a first input by a user on a destination object displayed on the touch screen display. While continuing to detect the first input by the user on the destination object, the computing device detects a second input by the user on a first user interface object displayed on the touch screen display. In response to detecting the second input by the user on the first user interface object, the computing device performs an action on the first user interface object. The action is associated with the destination object. | 03-24-2011 |
20110072375 | Device, Method, and Graphical User Interface for Manipulating User Interface Objects - A computing device with a touch screen display simultaneously displays on the touch screen display a plurality of user interface objects and at least one destination object. The computing device detects a first input by a user on a destination object displayed on the touch screen display. While continuing to detect the first input by the user on the destination object, the computing device detects a second input by the user on a first user interface object displayed on the touch screen display. In response to detecting the second input by the user on the first user interface object, the computing device performs an action on the first user interface object. The action is associated with the destination object. | 03-24-2011 |
20110072394 | Device, Method, and Graphical User Interface for Manipulating User Interface Objects - A computing device with a touch screen display simultaneously displays on the touch screen display a plurality of user interface objects and at least one destination object. The computing device detects a first input by a user on a destination object displayed on the touch screen display. While continuing to detect the first input by the user on the destination object, the computing device detects a second input by the user on a first user interface object displayed on the touch screen display. In response to detecting the second input by the user on the first user interface object, the computing device performs an action on the first user interface object. The action is associated with the destination object. | 03-24-2011 |
20110078622 | Device, Method, and Graphical User Interface for Moving a Calendar Entry in a Calendar Application - In some embodiments, a multifunction device with a display and a touch-sensitive surface displays a multi-week view in a calendar application on the display and detects a first input by a user. In response to detecting the first input by the user, the device selects a first calendar entry in the multi-week view in the calendar application. While continuing to detect selection of the first calendar entry by the user, the device detects a first multifinger gesture on the touch-sensitive surface, and in response to detecting the first multifinger gesture on the touch-sensitive surface, the device expands display of a single week in the multi-week view; and maintains display of the first calendar entry on the display. In some embodiments, the device moves the first calendar entry to a date and time in the calendar application in accordance with a second input by the user. | 03-31-2011 |
20110078624 | Device, Method, and Graphical User Interface for Manipulating Workspace Views - In some embodiments, a multifunction device with a display and a touch-sensitive surface creates a plurality of workspace views. A respective workspace view is configured to contain content assigned by a user to the respective workspace view. The content includes application windows. The device displays a first workspace view in the plurality of workspace views on the display without displaying other workspace views in the plurality of workspace views and detects a first multifinger gesture on the touch-sensitive surface. In response to detecting the first multifinger gesture on the touch-sensitive surface, the device replaces display of the first workspace view with concurrent display of the plurality of workspace views. | 03-31-2011 |
20110109538 | ENVIRONMENT SENSITIVE DISPLAY TAGS - This is directed to dynamic tags or screen savers for display on an electronic device. The tags can include several dynamic elements that move across the display. The particular characteristics of the elements can be controlled in part by the output of one or more sensors detecting the environment of the device. For example, the color scheme used for a tag can be selected based on the colors of an image captured by a camera, and the orientation of the movement can be selected from the output of a motion sensing component. The tag can adjust automatically based on the sensor outputs to provide an aesthetically pleasing display that a user can use as an fashion accessory. | 05-12-2011 |
20120030566 | SYSTEM WITH TOUCH-BASED SELECTION OF DATA ITEMS - Computing equipment may display data items in a list on a touch screen display. The computing equipment may use the touch screen display to detect touch gestures. A user may select a data item using a touch gesture such as a tap gesture. In response, the computing equipment may display a selectable option. When the option is displayed, movable markers may be placed in the list. The markers can be dragged to new locations to adjust how many of the data items are selected and highlighted in the list. Ranges of selected items may be merged by moving the markers to unify separate groups of selected items. A region that contains multiple selectable options may be displayed adjacent to a selected item. The selectable options may correspond to different ways to select and deselect items. Multifinger swipe gestures may be used to select and deselect data items. | 02-02-2012 |
20120030567 | SYSTEM WITH CONTEXTUAL DASHBOARD AND DROPBOARD FEATURES - A user may select content that has been displayed. The selected content may be provided to multiple applications as input in response to detection of a user command such as a touch gesture. The applications may be widgets that are displayed in respective application regions surrounding a focus region. The selected text may be presented in the focus region. Each widget may produce output in its application region that is based on the selected input. A user can launch a desired widget using a swipe gesture towards the desired widget. A user may transfer the selected content using a swipe from the focus region to an application region. A user can select which widgets are included in the application regions. Displayed data items may be related to selected content. A data item may be dragged onto a widget icon to transfer the data item to an associated widget. | 02-02-2012 |
20120036460 | Device, Method, and Graphical User Interface for Creating a New Folder - An electronic device concurrently displays a plurality of user interface objects and a list of folder icons. The list of folder icons includes a first new folder icon. The device detects a first input by a user using a first user input device, selects a folder icon in the list of folder icons in accordance with the first input, and indicates selection of the folder icon in the list of folder icons. The device also detects a second input by the user using a second user input device on one or more of the displayed user interface objects, moves the one or more user interface objects into a folder that corresponds to the selected folder icon, and when the selected folder icon is the first new folder icon, displays a second new folder icon in the list of folders icons. | 02-09-2012 |
20120084689 | Managing Items in a User Interface - User interface changes related to moving items in a user interface are disclosed. An operation (e.g., a drag operation) can be initiated on selected items by moving a cursor or pointing device in the user interface, and an animation can be presented illustrating representations of the selected items moving from their respective original locations toward a current location of the cursor or pointing device and forming a cluster in proximity to the current location of the cursor or pointing device. As the cluster of items is moved over a container object in the user interface, the representations of the items can adopt the appearance style defined by that container object. The representations of the items can also be shown to depart from the cluster and move toward anticipated locations of the items in the container object as a preview of a drop operation into the container object. | 04-05-2012 |
20120254795 | Methods and Graphical User Interfaces for Editing on a Portable Multifunction Device - Methods and graphical user interfaces for editing on a portable multifunction device with a touch screen display are disclosed. While displaying an application interface of an application, the device detects a multitouch edit initiation gesture on the touch screen display. In response to detection of the multitouch edit initiation gesture, the device displays a plurality of user-selectable edit option icons in an area of the touch screen display that is independent of a location of the multitouch edit initiation gesture. The device also displays a start point object and an end point object to select content displayed by the application in the application interface. | 10-04-2012 |