Patent application number | Description | Published |
20130321400 | 3D Map Views for 3D Maps - Some embodiments provide a non-transitory machine-readable medium that stores a program which when executed on a device by at least one processing unit provides different viewing modes for viewing a three-dimensional (3D) map. The program renders a first view of the 3D map for display in a first viewing mode based on a first set of map data. The program receives input to adjust the view of the 3D map. In response to the input, the program renders a second view of the 3D map for display in a second viewing mode based on a second set of map data different from the first set of map data. | 12-05-2013 |
20130321401 | Virtual Camera for 3D Maps - Some embodiments provide a non-transitory machine-readable medium that stores a mapping application which when executed on a device by at least one processing unit provides automated animation of a three-dimensional (3D) map along a navigation route. The mapping application identifies a first set of attributes for determining a first position of a virtual camera in the 3D map at a first instance in time. Based on the identified first set of attributes, the mapping application determines the position of the virtual camera in the 3D map at the first instance in time. The mapping application identifies a second set of attributes for determining a second position of the virtual camera in the 3D map at a second instance in time. Based on the identified second set of attributes, the mapping application determines the position of the virtual camera in the 3D map at the second instance in time. The mapping application renders an animated 3D map view of the 3D map from the first instance in time to the second instance in time based on the first and second positions of the virtual camera in the 3D map. | 12-05-2013 |
20130322702 | Rendering Maps - Some embodiments provide a mapping application for rendering map portions. The mapping application includes a map receiver for receiving map tiles from a mapping service in response to a request for the map tiles needed for a particular map view. Each map tile includes vector data describing a map region. The mapping application includes a set of mesh building modules. Each mesh building module is for using the vector data in at least one map tile to build a mesh for a particular layer of the particular map view. The mapping application includes a mesh aggregation module for combining layers from several mesh builders into a renderable tile for the particular map view. The mapping application includes a rendering engine for rendering the particular map view. | 12-05-2013 |
20130328924 | Constructing Road Geometry - Some embodiments provide a method for a mapping service. The method generates an initial set of geometries for a road graph defined for a map region. The road graph includes several road segments and junctions aggregated into roads. The method identifies an overlap between a first geometry of a first road segment and a second geometry of a second road segment. The first road segment and the second road segment are parts of different roads and do not meet at a junction. The method automatically modifies at least one of the first and second geometries in order to prevent the first geometry and second geometry from overlapping. In some embodiments each geometry is defined by a set of vertices that specify its boundaries. The method of some embodiments automatically modifies the vertices of at least one of the first and second geometries. | 12-12-2013 |
20130332057 | Representing Traffic Along a Route - Some embodiments provide a mapping application that has a novel way of displaying traffic congestion along roads in the map. The mapping application in some embodiments defines a traffic congestion representation to run parallel to its corresponding road portion when the map is viewed at a particular zoom level, and defines a traffic congestion representation to be placed over its corresponding road portion when the map is viewed at another zoom level. The mapping application in some embodiments differentiates the appearance of the traffic congestion representation that signifies heavy traffic congestion from the appearance of the traffic congestion representation that signifies moderate traffic congestion. In some of these embodiments, the mapping application does not generate a traffic congestion representation for areas along a road that are not congested. | 12-12-2013 |
20140365113 | Navigation Application with Several Navigation Modes - A method for providing navigation instructions on a device is described. As the device traverses a navigated route according to a first mode of transportation, the method displays a first turn-by-turn navigation presentation defined for the first mode. Based on data gathered by the device, the method determines that the device is navigating the route according to a second mode of transportation. The method automatically displays a second, different turn-by-turn navigation presentation defined for the second mode. | 12-11-2014 |
20140365122 | Navigation Peek Ahead and Behind in a Navigation Application - A method of providing a sequence of turn-by-turn navigation instructions on a device traversing a route is provided. Each turn-by-turn navigation instruction is associated with a location on the route. As the device traverses along the route, the method displays a turn-by-turn navigation instruction associated with a current location of the device. The method receives a touch input through a touch input interface of the device while displaying a first turn-by-turn navigation instruction and a first map region that displays the current location and a first location associated with the first turn-by-turn navigation instruction. In response to receiving the touch input, the method displays a second turn-by-turn navigation instruction and a second map region that displays a second location associated with the second turn-by-turn navigation instruction. Without receiving additional input, the method automatically returns to the display of the first turn-by-turn navigation instruction and the first map region. | 12-11-2014 |
20140365154 | COMPASS CALIBRATION - A method that performs a series of interactive operations to calibrate a compass in a mobile device. The method requires a user to move the device to a variety of different orientations. In order to ensure that the device moves to a sufficient number and variety of orientations, the method instructs the user to rotate the device in a series of interactive operations. The interactive operations provide feedback to inform the user how well the user is performing the interactive operations. In some embodiments, the feedback is tactile (e.g., a vibration). In some embodiments the feedback is audible (e.g., a beep or buzz). In some embodiments, the feedback is visual (e.g., an image or images on a video display of the device). The feedback in some embodiments is continuous (e.g., a changing visual display) and in some embodiments is discrete (e.g., the device beeps after taking a good reading). | 12-11-2014 |
20140365965 | NIGHT MODE - A device that provides a map and/or navigation application that displays items on the map and/or navigation instructions differently in different modes. The applications of some embodiments provide a day mode and a night mode. In some embodiments the application uses the day mode as a default and activates the night mode when the time is after sunset at the location of the device. Some embodiments activate night mode when multiple conditions are satisfied (for example, when (1) the time is after sunset at the location of the device and (2) the ambient light level is below a threshold brightness). | 12-11-2014 |
20150046867 | CONTEXT SENSITIVE ACTIONS - Techniques for performing context-sensitive actions in response to touch input are provided. A user interface of an application can be displayed. Touch input can be received in a region of the displayed user interface, and a context can be determined. A first action may be performed if the context is a first context and a second action may instead be performed if the context is a second context different from the first context. In some embodiments, an action may be performed if the context is a first context and the touch input is a first touch input, and may also be performed if the context is a second context and the touch input is a second touch input | 02-12-2015 |
20150046884 | CONTEXT SENSITIVE ACTIONS IN RESPONSE TO TOUCH INPUT - Techniques for performing context-sensitive actions in response to touch input are provided. A user interface of an application can be displayed. Touch input can be received in a region of the displayed user interface, and a context can be determined. A first action may be performed if the context is a first context and a second action may instead be performed if the context is a second context different from the first context. In some embodiments, an action may be performed if the context is a first context and the touch input is a first touch input, and may also be performed if the context is a second context and the touch input is a second touch input. | 02-12-2015 |