US8762041B2 - Method, device and system for presenting navigational information - Google Patents
Method, device and system for presenting navigational information Download PDFInfo
- Publication number
- US8762041B2 US8762041B2 US12/819,521 US81952110A US8762041B2 US 8762041 B2 US8762041 B2 US 8762041B2 US 81952110 A US81952110 A US 81952110A US 8762041 B2 US8762041 B2 US 8762041B2
- Authority
- US
- United States
- Prior art keywords
- objects
- image
- route
- mobile device
- input images
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
- 238000000034 method Methods 0.000 title claims abstract description 62
- 238000012545 processing Methods 0.000 claims abstract description 65
- 238000009877 rendering Methods 0.000 claims description 16
- 230000033001 locomotion Effects 0.000 claims description 15
- 238000004590 computer program Methods 0.000 claims description 11
- 238000005516 engineering process Methods 0.000 description 35
- 238000004891 communication Methods 0.000 description 18
- 230000001413 cellular effect Effects 0.000 description 8
- 238000001514 detection method Methods 0.000 description 5
- 230000000694 effects Effects 0.000 description 5
- 230000008569 process Effects 0.000 description 5
- 238000010586 diagram Methods 0.000 description 4
- 238000000605 extraction Methods 0.000 description 3
- 230000000007 visual effect Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 229920001690 polydopamine Polymers 0.000 description 2
- 238000007781 pre-processing Methods 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 230000006641 stabilisation Effects 0.000 description 2
- 238000011105 stabilization Methods 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 238000012952 Resampling Methods 0.000 description 1
- 206010038743 Restlessness Diseases 0.000 description 1
- 241000278713 Theora Species 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 238000013475 authorization Methods 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000000903 blocking effect Effects 0.000 description 1
- 238000005266 casting Methods 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 230000001934 delay Effects 0.000 description 1
- 230000001815 facial effect Effects 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000002156 mixing Methods 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/34—Route searching; Route guidance
- G01C21/36—Input/output arrangements for on-board computers
- G01C21/3605—Destination input or retrieval
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/34—Route searching; Route guidance
- G01C21/36—Input/output arrangements for on-board computers
- G01C21/3626—Details of the output of route guidance instructions
- G01C21/3632—Guidance using simplified or iconic instructions, e.g. using arrows
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/34—Route searching; Route guidance
- G01C21/36—Input/output arrangements for on-board computers
- G01C21/3626—Details of the output of route guidance instructions
- G01C21/3647—Guidance involving output of stored or live camera images or video streams
Definitions
- the present disclosure relates generally to mobile electronic devices and, in particular, to techniques for presenting navigational information on a mobile device.
- Mobile devices such as hand-held GPS units and in-vehicle GPS units have been developed for providing navigation guidance.
- Wireless communication-enabled mobile devices such as GPS-enabled cellular telephones, smart phones, handheld PDAs and the like have also been equipped with navigation functionality for providing navigation guidance. These devices are configured to determine the location of the mobile device, to allow a user to specify a destination and to provide navigational information for traversing a route from the current location to the specified destination.
- Navigational information such as turn-by-turn route guidance instructions
- the navigational information is presented as directional arrows on maps showing the user's current location, text displayed on the mobile device screen and as audio alerts such as “turn left now”.
- the user is required to correlate the information provided with the observed environment, potentially resulting in delays and errors.
- FIG. 1 is a flowchart of a method for presenting navigational information for a mobile device in accordance with embodiments of the present technology
- FIG. 2 is a flowchart of a method of processing an input image in accordance with embodiments of the present technology
- FIG. 3 is an exemplary illustration, in accordance with embodiments of the present technology, of modifying the input image
- FIG. 4 is an exemplary illustration, in accordance with embodiments of the present technology, of modifying the input image
- FIG. 5 illustrates an exemplary method of creating an output image by combining two layers in accordance with embodiments of the present technology
- FIG. 6 illustrates an exemplary method of creating an output image by combining three layers in accordance with embodiments of the present technology
- FIG. 7( a ) illustrates a block diagram of an exemplary system configured for presenting navigational information in accordance with embodiments of the present technology
- FIG. 7( b ) illustrates a block diagram of an exemplary image generation module for generating an output image from an input image in accordance with embodiments of the present technology
- FIG. 8 illustrates a block diagram of an exemplary mobile device in accordance with embodiments of the present technology
- the present technology generally provides a method, mobile device and computer program product for presenting navigational information.
- an aspect of the present technology is a method comprising: obtaining one or more input images; generating one or more output images by obtaining navigational information for the mobile device and processing the one or more input images to integrate the obtained navigational information, wherein processing the one or more input images includes recognizing one or more objects in the one or more input images; and providing the one or more output images for presentation.
- Another aspect of the present technology is a computer program product comprising code adapted to perform acts associated with the foregoing method when the code is loaded into memory and executed on a processor of a mobile device.
- Yet another aspect of the present technology is a mobile device configured to obtain one or more input images, generate one or more output images by obtaining navigational information and processing the one or more input images to integrate the navigational information, wherein processing the one or more input images comprises recognizing one or more objects in the one or more input images, and to provide the one or more output images for presentation.
- FIG. 1 illustrates a method 100 for presenting navigational information for a mobile device in accordance with embodiments of the present technology.
- the method 100 comprises obtaining one or more input images 130 , generating one or more output images 140 by processing of the input images to integrate navigational information therein, and presenting the one or more output images 150 .
- an input image of a user's current location may be obtained from the camera on the user's mobile phone.
- the obtained image may be processed to recognize the streets shown in the input image and an output image may be presented highlighting the street comprising the navigation route in the input image.
- Embodiments of the present technology may be used in a variety of augmented reality applications such as automobile and pedestrian navigation, electronic tourist guides, gaming applications and activities such as geocaching and the like.
- An image is captured by an image-capturing device such as a still camera, video camera, infra-red camera or the like.
- the captured image is indicative of a current location of the mobile device.
- the video camera of a mobile device may be used to continuously obtain an image at the user's current location.
- the captured image is indicative of a destination for the mobile device.
- a tourist wishing to find a route to a particular landmark may capture an image including the landmark, for example in the distance, using the mobile device camera. The user may then specify the landmark as a desired destination by indicating it on the captured image.
- the image may be captured in a suitable image format as would be readily understood by a worker skilled in the art.
- Formats for capturing images may include raw image formats, Joint Photographic Experts Group File (JPEG) and the like for still and video images and interlaced and progressive scanning formats for digital video images.
- JPEG Joint Photographic Experts Group File
- the captured image hereinafter referred to as the input image, is converted to a digitized representation if it is not in a digital format.
- the digitized representation comprises pixels with image parameters such as resolution, pixel density, pixel bit depth, image size and magnification based in part on the image-capturing device.
- Standard digital formats for still images include raster formats such as Joint Photographic Experts Group File Interchange Format (JPEG/JFIF), Tagged Image File Format (TIFF), raw image formats, Graphics Interchange Format (GIF), Tagged Image File Format (TIFF), Bitmap and the like, and vector formats, such as Computer Graphics Metafile (CGM), Scalable Vector Graphics (SVG), Encapsulated Postscript (EPS), Windows Metafile (WMF), Enhanced Metafile (EMF) and the like.
- Standard digital formats for moving images include H-261, Theora and Moving Picture Experts Group (MPEG) formats such as MPEG-1, MPEG-2, MPEG-4 and the like.
- processing of the input image comprises converting the image from one format to another, for example to decompress the input image, and to convert it to a format amenable to selectable modification of pixels in the input image.
- processing of the image comprises reconstructing the still image from the compressed video file, as would be readily understood by a worker skilled in the art.
- the input image may also be associated with image parameters such as the location, orientation, tilt and zoom that may be used for processing of the input image.
- image parameters may be obtained from sensors operatively associated with the image-capturing device.
- the input image may be a frame from a series of images captured periodically at a given frame rate or it may be a still image captured substantially immediately prior to processing and subsequent output of a processed image.
- the input image may be subjected to various pre-processing operations such as resampling, noise reduction, image stabilization and the like.
- the image may be resampled in order to change the size of the image to fit the dimensions of the display.
- Noise reduction and image stabilization may be necessary to improve the quality of the image before it can be subjected to more detailed processing.
- the pre-processing step may also include adjustments to image attributes such as color, balance, brightness, luminance, lighting, saturation and contrast.
- An output image is formed by processing the input image to integrate navigational information therein.
- navigational information may be integrated into the input image as a visual representation appearing in the output image.
- Navigational information as used herein may comprise information that may be relevant to a user at a current location or for a user traversing a route.
- navigational information may comprise map information, a visual representation of route information, places of interest along the route or at the current location, weather and traffic information, route directions, travel speed, or the like, or a combination thereof.
- FIG. 2 is a flowchart illustrating the steps involved in a method 200 for generating an output image by processing of the input image in accordance with embodiments of the present technology.
- the method 200 comprises obtaining an input image and navigational information 205 , recognizing one or more objects 210 in the input image, processing the input image 220 , at least in part based on the one or more objects recognized, to integrate navigational information with the input image, and generating the output image 230 .
- an object such as a street along which navigation route proceeds may be recognized 210 in the input image and the input image may be modified 220 to highlight the navigation route along the recognized street.
- processing the input image to integrate navigational information may comprise recognizing one or more objects in the input image.
- processing the input image to integrate navigational information further comprises modifying the input image at least in part based on the one or more objects recognized.
- Modifying the input image may comprise adding, modifying, removing or replacing one or more objects from the input image.
- processing of the input image is performed by the mobile device.
- the processing is performed at a remote server, exemplarily operated by a service and application provider.
- the processing is performed by both.
- the distribution of processing may be based on factors such as the computational capabilities of the respective processing modules, communications capabilities between the mobile device and the remote server and the availability of databases for maps and images.
- One or more load balancing activities may be performed for the distribution of the processing burden, as would be readily understood by a worker skilled in the art.
- the input image is processed to recognize objects. Recognizing objects in the input image, as used herein, may comprise identifying pixels in the image that correspond to particular objects such as a particular street or building.
- the objects recognized in the input image include objects related to navigational information, for example, objects that are indicative of navigational instructions wherein such objects are called navigational objects.
- Navigational objects may include objects such as streets, intersections, buildings, landmarks, traffic lights and signs, turns, roundabouts, dead ends and the like.
- the objects recognized in the input image also include other objects such as cars, trees, people and the like. Accordingly, the step of recognizing may comprise distinguishing or differentiating navigational objects from other objects.
- recognizing objects in the input image comprises distinguishing or detecting objects such as streets, buildings, vehicles and the like. In some embodiments, recognizing objects further comprises identifying particular instances of more general objects. For example, the image is first processed to detect the portions of the image that represent the “street”. Then, the one or more objects classified as “street” may be identified as a specific street by name, for example, “First Ave”, “Second Ave.”, etc.”
- the image is processed to distinguish objects such as the street from the rest of the image which might contain other objects such as lawn, sidewalks, buildings, and the like.
- Detection of particular objects in an image may be performed using various object recognition and pattern detection techniques that will be readily known to a worker skilled in the art. Detection of particular instances of objects may be aided by information regarding location of the image, for example using GPS coordinates acquired when capturing the image.
- Object and pattern recognition techniques and algorithms are well known in various fields related to computer vision such as facial detection, baggage scanning, etc.
- Typical object recognition techniques comprise segmenting the image into segments or sets of pixels that share some common visual characteristic using techniques such as contrast enhancement.
- the segmented regions or objects may be subjected to feature extraction.
- Typical features detected by feature extraction algorithms include edges, corners, blobs and ridges. Other properties such as colour, texture, shape and size of the objects may also be analyzed.
- Various rough and detailed classification steps may be successively applied to the objects to compare their feature sets with a set of standard patterns such as patterns for tree, car, street, etc. stored in a database and to determine the object classes.
- the objects detected in the input image may be identified by comparing the input image to maps or other images stored in a database.
- the input image may be tagged with information related to location in which the image was captured in addition to parameters such as the tilt, orientation, scale and field of view of the image-capturing device used to capture the input image.
- location information may be obtained from a GPS receiver, orientation information from a magnometer, tilt information from tilt sensors and information related to the scale of the input image and its field of view may be derived from the zoom parameter of the image capturing device.
- Databases containing extensive automobile and pedestrian navigation maps are well-known and widely used in various navigation devices and applications. Such databases may be stored locally on a mobile device or on a remotely accessible server. Some applications also provide images of locations and routes in addition to map and terrain information. Exemplarily, the images may be stored in the database and the database entry may be linked to or tagged with reference to the location in the captured image. For example, a reference may indicate New York City, Times Square, north-west corner, latitude and longitude. Images in a database may also be stored with reference to image-capturing device parameters such as the location, orientation, tilt and magnification of the image-capturing device. For example, a reference may indicate: New York City, Times Square, south-west corner, latitude and longitude, facing N15W.
- an input image is processed to distinguish and identify objects in a predetermined portion of the input image, for example in the entire input image or a particular quadrant of the input image. For example, all the buildings, streets and landmarks appearing in the input image may be identified.
- target objects are specified and the input image may be processed to recognize the specified target objects which may or may not be navigational objects.
- the image may be processed to recognize one or more navigational objects including streets representing the navigation route and one or more landmarks appearing along the navigation route.
- processing an input image to detect target objects therein comprises as a first step, determining an approximate region where a particular target object, for example the navigation route, is expected to be detected.
- An approximate region of the input image may be identified for each target object by comparing the input image to maps and images in a database. This approximate region of the image may then be subjected to further detailed processing to distinguish the pixels in the input image representing the target objects.
- the approximate region of the input image where the navigation route is expected to be detected may be further processed to distinguish pixels representing the navigation route from pixels representing other objects such as sidewalks, neighbouring buildings etc.
- the series of input images are compared using information related to the motion of the image-capturing device.
- a particular object of interest for example, a building
- the position of this particular object for example, the building
- the pixels corresponding to this object of interest may be predicted in succeeding input images using information known about the motion of the image-capturing device.
- Motion of the mobile device may comprise changes in position as well as orientation and tilt of the image capturing device. Accordingly, various sensors for detecting position and motion such as GPS receivers, magnometers, tilt sensors, gyros, and accelerometers may be provided to detect various aspects of motion. A worker skilled in the art will be familiar with appropriate sensors for detection of motion and position. Motion information may be obtained for either or both of the image-capturing device and the mobile device.
- processing an input image to recognize target objects comprises recognition of a reference object.
- the reference object may be a particular landmark building, a particular corner of a building, a traffic light, a tree or other prominent object in an input image.
- an input image may be quickly scanned to determine the location of this reference object.
- the location of other objects in relation to the location of the reference object may be known, either based on a previous input image or based on maps or images obtained from the database. Accordingly, once the reference object has been recognized in an input image, other objects in the input image may be detected based on the location of the reference object and its relationship to the other objects.
- processing the input image further comprises modifying the input image to integrate navigational information.
- Navigational information may be obtained based at least in part on a determination of the location of the mobile device, and optionally based on one or more objects recognized in the input image.
- the one or more objects may be navigational objects or other objects.
- the input image is modified at least in part based on one or more objects recognized in the input image.
- the one or more recognized objects may be navigational objects or other objects.
- the input image may be modified to add, remove or modify objects in the input image.
- the objects are indicative of navigational information. For example, graphical, textual or symbolic objects may be added to the image.
- objects in the input image such as particular buildings may be outlined or highlighted.
- an object obstructing the entrance to a building may be removed.
- the input image is modified to outline particular objects.
- the boundary between the street and the sidewalk may be outlined with a white line.
- the method comprises distinguishing streets from sidewalks using techniques such as contrast enhancement, edge extraction and the like, for example as described herein.
- an object is outlined by changing the colour of the pixels corresponding to the extracted edges of the object.
- the outline may be a continuous solid line or a line pattern such as a dashed or dotted line.
- the colour may be changed to a predetermined value, changed from its initial colour by a predetermined amount, or a combination thereof, or the like, so as to make the outline readily identifiable.
- the input image is modified to highlight particular objects in the input image. For example, a particular restaurant or street in the image may be highlighted in yellow.
- the object may be identified using object recognition techniques known in the art.
- an object is highlighted by modifying the colour of the pixels corresponding to the object in the input image.
- Objects in the input image may be highlighted with a uniform translucent or opaque color or with a pattern such as a hatched pattern.
- FIG. 3 is an exemplary illustration of modifying an input image to highlight and outline particular objects in accordance with embodiments of the present technology.
- FIG. 3( a ) shows an image of a street as captured by a camera. Accordingly, FIG. 3( a ) may be considered to be the input image 310 for the purposes of this example.
- FIG. 3( b ) is a line drawing 305 of the input image 310 indicating the main features of the input image as well as the modifications to highlight 320 , outline 330 and add objects 340 , 350 , 360 .
- FIG. 3( c ) shows the resulting output image created by modifying the obtained input image 310 to highlight a navigation route 320 and outline a destination 330 .
- text objects 340 , 350 and 360 respectively indicating the street name, end of the route and the destination address have been added to the input image.
- the input image is modified to add rendered objects on the input image.
- the objects may be navigational objects or other objects.
- the rendered object comprises a rendered outline of the object (for example, a wireframe outline).
- the outline may be a line rendered with a predetermined point size or dash pattern.
- a rendered object comprises a rendered body.
- the object may be rendered by filling its interior with a pattern, color or color gradient.
- the object may be rendered with a uniform color or a hatch pattern.
- the object may be rendered as opaque or translucent.
- both the outline and body of an object are rendered. Accordingly, the outline may be rendered to be distinct from the body of the object.
- the object may be rendered as a planar and two-dimensional object or as a three-dimensional object.
- objects are rendered to match the depth and perspective of the input image.
- the rendered object corresponds to an existing object in the input image.
- the outline of a building in the input image may be rendered.
- the outline may be rendered, at least in part, by utilizing the recognized edges of the building.
- an object in the input image may be highlighted by filling the interior of the rendered object with the desired highlight colour, or modifying all or a portion of the rendered object colour so as to make it stand out as an object of interest.
- the rendered objects are indicative of a part of an object in the input image.
- the input image may be modified to render a particular lane on a multi-lane highway.
- the rendered object is indicative of an imaginary object.
- an outline of the navigation route through the field or square i.e. a “route” shaped object
- the route object may be scaled and shaped at least in part based on the recognizing of the square, streets and buildings around the square, and optionally other objects such as trees etc. on the square.
- FIG. 4( a ) shows an input image as obtained from a camera
- FIG. 4( b ) is a line drawing of the obtained input image indicating the square 410 and the rendered pedestrian route 420 through the square.
- FIG. 4( c ) shows the output image created by rendering a navigation route 420 on the square 410 of the input image.
- the input image is modified to render objects on the input image that may be at least partially missing from the input image as captured by the image-capturing device.
- the rendered object is indicative of an object interior to or otherwise obstructed by an object in the input image.
- the image may be modified to render an outline of a path obstructed by a vehicle or to render an object indicating a café inside a building.
- the rendered object is indicative of an object partially or completely obscured in the image due to poor visibility conditions such as fog or poor light.
- the input image may be modified to add objects to aid navigation in poor visibility conditions. For example, the outline of a path not clearly visible due to foggy conditions may be added to the input image.
- the rendered object overlaps, at least partially, with one or more objects in the input image.
- the rendered object may be, at least partially, in front of or behind other objects in the input image.
- the rendered object may be, at least partially, in front of some objects as well as at least partially behind other objects in the input image.
- the street to be highlighted may be partially blocked by a car on the street. Accordingly, in order to highlight the street, the rendered street object can be placed behind the car.
- the input image may be modified to add textual objects.
- the text may be related to navigation instructions, speed of travel, time or distance remaining to destination, information related to objects in the field of view, weather information, or other information.
- the text may be added to label streets and landmarks in the input image.
- the input image is modified to add symbolic objects.
- a left turn icon may be added to indicate an upcoming turn.
- objects added to the input image are based on a user specification or a user preference. For example, a user may select a mode where the street names will be displayed but not other points of interest.
- the user can select a “parking” mode when attempting to park a car.
- the input image, in this mode may be modified to render an outline of a car, scaled to have the size of the user's car and placed in various possible parking spots in the vicinity to indicate to the user whether or not the car will fit in that spot.
- a user may select an object in an input image which will be highlighted whenever it appears in the field of view. For example, the user may specify a destination by selecting a landmark in the input image. Accordingly, the landmark may then be highlighted whenever it appears in the field of view. As another example, a tourist wandering around in a neighbourhood can keep themselves oriented with respect to a particular building or a tree in a park by specifying that it always be outlined when it appears in the field of view.
- an indication of the landmark is included in the output image even when it is obstructed.
- an arrow may appear at an edge of the output image indicating the location of the landmark.
- the input image is modified to render the navigation route.
- the input image may be modified to render one or more alternate or additional navigation routes.
- the navigation routes or paths may be ranked or prioritized based on different parameters such as condition of route, length of route, difficulty of route, parameters related to terrain such as elevation changes, congestion and traffic along route, time of day, and the like.
- the user may specify one or more preferred parameters or preferred routes.
- the route may be rendered in a colour indicative of a ranking or priority level assigned to the route. For example, a preferred safe navigation route may be rendered in green while an alternate navigation route may be rendered in a red colour.
- the route is rendered in a colour, colour gradient or range of colours indicative of a particular aspect of the route.
- the colour of the route may be indicative of aspects such as elevation, changes in elevation, condition of the route, and the like.
- the input image is modified to improve user perceptibility.
- image properties such as the colour, saturation, brightness, contrast, size of image or field of view may be configured to facilitate perception of objects in the output image.
- the input image is modified to change the colour of one or more objects.
- the colour of substantially the entire image is modified.
- the input image may be modified to be in monochrome while one or more objects (such as the navigation route or the building comprising the destination) may be rendered in colour.
- the objects added to the input image are rendered based, at least in part, on one or more objects recognized in the image.
- the objects added to the input image are placed or aligned based, at least in part, on one or more objects recognized in the image. For example, a text label for a street in the input image may be placed on the object recognized as “First Street” in the image. As another example, a path through a park may be placed in the middle of the object recognized as the park in the image.
- objects are rendered at least in part based on the location, size and shape of one or more objects in the image.
- the rendered route object 420 in FIG. 4 is scaled for the size of the object recognized as square in the input image.
- the path is shaped based on detecting in the input image, the user's current location and the street to which the user is headed.
- the objects added to the input image are placed or aligned based, at least in part, on the context. For example, in order to indicate an upcoming left turn to a user traversing a route along “First Street”, a left turn arrow may be placed on the object recognized as “First Street”.
- the colour in which an object is rendered is selected based on the colour of one or more objects recognized in the input image.
- the text may be in a dark colour if placed on a light background of sky but light if placed on a dark background of trees.
- the rendering of objects may be indicative of the objects being navigational objects or other objects.
- the colour, texture or outline pattern of the rendered navigation route may be selected to be different from that used for other objects.
- objects are rendered at least in part using maps and other images stored in a database. For example, when the view of the street ahead is blocked by a vehicle, maps and images in the database may be retrieved and utilized to determine how the street continues beyond the blocking vehicle. For example, the street may continue straight then turns right after 10 m. Accordingly, an outline of the route beyond the vehicle may be rendered, based on the information retrieved from the database, and placed on top of the vehicle in the image, or otherwise blended into the image to convey continuity of the route past the obstruction.
- an object is rendered by modifying a previously rendered object in accordance with information obtained about the motion of the image capturing device.
- a previously rendered object such as a building or street
- a previously rendered object may be moved, rotated, scaled based on knowledge of the translational, rotational or tilt movement of the image-capturing device.
- a worker skilled in the art will be familiar with various techniques for transforming objects.
- an outline of a building may be rendered on a first input image. If the image-capturing device is known to have moved 20 m towards the building, the rendered outline of the building may be scaled accordingly for the subsequent input image. Such scaling of a building may further depend on the initial and final absolute distance to the building, along with other factors.
- Objects may be rendered using rendering techniques familiar to a worker skilled in the art (for example, pixel-by-pixel rendering, rasterization, ray casting, etc.).
- rendering techniques familiar to a worker skilled in the art (for example, pixel-by-pixel rendering, rasterization, ray casting, etc.).
- the selection of rendering technique used may depend on processing capacity, processing time available, frame rate for capturing images or updating images on the display.
- the output image is created by modifying pixels in the input image.
- the pixels modified correspond to the one or more objects to be modified or added to the input image. For example, when an object (for example a building) is to be highlighted, the object representing the building is first identified in the image using object recognition techniques. Then, the colour of the pixels corresponding to the building is modified to show the building as highlighted in that colour. Similarly, an object may be outlined by modifying the colour of the pixels that correspond to the edges of the object in the input image.
- the output image is created by modifying a subset of the pixels in the input image corresponding to an object to be modified or added. For example, in order to highlight the navigation route with a translucent effect, a subset of the pixels (in an alternating hatch pattern) corresponding to the navigation route may be changed in colour. Similarly, a navigation route may be outlined by a dotted or dashed line by modifying the colour of a subset of pixels (in the appropriate pattern for producing a dashed or dotted effect) representing the edges of the route.
- objects to be added to the input image are placed on one or more layers distinct from the layer containing the input image.
- the output image is created by combining the layers with the input image using techniques for combining layers as would be readily known to a worker skilled in the art. For example, the output image may be created by combining a first layer containing the input image with a second layer containing the rendered route.
- Objects added to the input image may be translucent or opaque.
- a worker skilled in the art will be familiar with various methods of indicating the degree of transparency of a pixel and for combining or blending layers with overlapping objects of different transparency.
- the degree of transparency may be indicated by an alpha value defined to range between 0 for completely opaque to 1 for completely transparent.
- the transparency of each pixel may be stored in its alpha channel.
- the layers may be combined using alpha compositing rules which would be readily known to a worker skilled in the art.
- FIG. 5 illustrates an exemplary method of creating an output image to highlight a street in the input image in accordance with embodiments of the present technology.
- FIG. 5( a ) shows the input image 510 on a first (bottom) layer 501 .
- the input image 510 is depicted as a line drawing.
- a second (top) layer 502 is created and an object corresponding to the street 515 is rendered on this layer.
- the street is rendered by filling the pixels on the second (top) layer 502 corresponding to the street 515 with the desired highlight colour (for example, grey in FIG. 5( a )).
- the two layers may be combined to produce an output image 520 as shown in FIG. 5( b ) or 5 ( c ).
- the layers 501 and 502 will be combined such that the pixels 525 of the output image 520 in the position of the street 515 will be substantially entirely those of the top layer 502 as shown in FIG. 5( b ). As shown in FIG. 5( c ), if a translucent effect is desired for the highlighted street in the output image 520 , the pixels 530 of the output image 520 in the position of the street, will alternately be those of the input image on the bottom layer and the rendered street on the top layer.
- a plurality of objects are added to the input image to create the output image.
- the objects may be placed on the same layer or on different layers. Accordingly, each layer may comprise one or more objects.
- An object on a layer may be individually positioned and altered without affecting other objects on that layer or on other layers.
- a rendered objects overlaps with other rendered objects or objects in the input image.
- the rendered objects may be placed in front of some objects and behind other objects. Placing a first object in front of other objects may be accomplished by placing the first object in a top layer and placing objects behind the first object in lower layers and so on.
- the overlapping objects placed on different layers may be combined sequentially using information about the transparency of the objects on each layer.
- FIG. 6 illustrates an exemplary method of combining layers containing overlapping objects in accordance with embodiments of the present technology.
- FIG. 6( a ) shows an exemplary input image 610 with a street 612 and a car 615 on the street 612 .
- the input image is shown as a line drawing.
- the “rendered route” object 618 has to be placed above the “street” 612 in the image but below the “car” 615 of the image.
- this may be achieved by processing the input image to recognize the pixels corresponding to the car 615 and creating a third layer with pixels corresponding to the car.
- the first (bottom) layer 601 comprises the input image 610
- the second (middle) layer 602 comprises the rendered route 618
- the third (top) layer 603 comprises the car 615 of the input image 610
- the pixels corresponding to the car on the third layer may be defined as opaque and the pixels on the second layer corresponding to the rendered route may be defined to be “translucent”.
- the output image 620 is created by combining the layers 603 , 602 , and 601 , from top to bottom such that in the position of the car 615 , the pixels of the output image contain pixels of the third layer 603 .
- the pixels of the output image in the position of the rendered route 618 contain pixels alternately taken from the second layer 602 and the first layer 601 . In the remaining parts of the output image, the pixels correspond to the pixels of the input image taken from the first layer 601 .
- objects added to the input image for the formation of the output image are updated at different rates.
- objects on a layer may be updated at the same rate while objects on different layers may be updated at different rates.
- one layer may contain objects that are relatively static (for example text indicating the temperature) while another layer may contain more dynamic objects (for example the rendered route). Accordingly, some layers integrated with the input image may be updated more frequently than others.
- an output image formed by processing the input image to integrate navigational information, is presented on the mobile device.
- the input image may need to be re-sampled if the pixel dimension of the image capturing device are different from those of the display module.
- the input image is re-sampled before processing of the input image.
- the input image is re-sampled after it has been processed to recognize objects and before objects have been rendered.
- the re-sampling is performed after the input image has been modified to integrate navigation information and before the resulting output image is displayed on the display module.
- the delay time may be fixed and predetermined.
- the delay time may be based on the parameters such as amount of processing required, the processing capacity, the frame rate at which the input image is obtained, delay to compensate for vibrations, or other sources of delay.
- aspects of processing of the input image is based, at least in part, on the delay time.
- the delay time is variable and a maximum delay time is defined.
- the maximum delay time is defined based on the motion of the mobile device. For example, the output image may be updated more frequently when the mobile device is moving more rapidly so that the image on the display corresponds closely to the user's view. The motion may be either related to changes in location or orientation. If a delay time is expected to exceed a maximum delay time, the delay time may be reduced, for example by sacrificing image quality.
- an intermediate output image is presented for display while the processing continues.
- the intermediate output image is the input image as captured.
- the intermediate output image is the input image at an intermediate stage of processing, for example, when some of the layers containing navigational information have been updated and integrated with the input image but other layers have not yet been updated or integrated.
- the output image is presented on one or more displays associated with the mobile device.
- the presentation of the output image is accompanied by audio alerts and notifications presented on an audio output device such as a speaker.
- the presentation of the output image is also be accompanied by tactile alerts. For example, a tactile or audio alert may be issued to indicate that the image capturing device is facing away from the navigation route.
- Acts associated with the method described herein can be implemented as coded instructions in a computer program product.
- the computer program product is a computer-readable medium upon which software code is recorded to execute the method when the computer program product is loaded into memory and executed on the microprocessor of the mobile device.
- Acts associated with the method described herein can be implemented as coded instructions in plural computer program products. For example, a first portion of the method may be performed using one mobile device, and a second portion of the method may be performed on another mobile device or a device such as a server, or the like.
- each computer program product is a computer-readable medium upon which software code is recorded to execute appropriate portions of the method when a computer program product is loaded into memory and executed on the microprocessor of a mobile device.
- Embodiments of the present technology may be implemented on a mobile device.
- the mobile device may be a cellular telephone, satellite telephone, smartphone, PDA, laptop, net book, a tablet PC, a portable navigation system such as a handheld or vehicular GPS unit, or another device that can be configured to allow a determination of its location. Additional types of mobile devices would be readily known to a worker skilled in the art.
- the mobile device comprises a display control module.
- the display control module may be configured to obtain one or more input images and one or more output images and to control presentation of the obtained output images which are generated by obtaining navigational information and processing the input image at least in part based on the obtained navigational information, for example to integrate said navigational information.
- FIG. 7( a ) is a schematic of an exemplary configuration utilizing a display control module in accordance with embodiments of the present technology.
- the display control module 710 obtains one or more input images 720 from an operatively linked image-capturing module 725 .
- the display control module 710 obtains one or more output images 730 from an operatively linked image generation module 735 .
- additional parameters (tilt, motion, etc.) 740 may also be obtained by the display control module 710 .
- the display control module then provides the one or more output images for presentation by one or more operatively linked image display modules 750 .
- a processing module may be operatively coupled with the display control module, image capturing module, the image generation module and the image display modules.
- the processing module may be configured to coordinate the functions of the various modules and to process different types of data using one or more processors.
- the processing module is integrated with the display control module.
- the processing module is the display control module.
- the processing module comprises or is operatively associated with a memory module configured for storage of different types of data. Appropriate processing and memory modules would be readily understood by a worker skilled in the art.
- the mobile device comprises a location identification module 715 .
- the location identification module is configured to allow a determination of location of the mobile device.
- the location identification module comprises a GPS receiver.
- GPS Global Positioning System
- the present disclosure refers expressly to the “Global Positioning System”, it should be understood that this term and its abbreviation “GPS” are being used expansively to include any satellite-based navigation-signal broadcast system, and would therefore include other systems used around the world including the Beidou (COMPASS) system being developed by China, the multi-national Galileo system being developed by the European Union, in collaboration with China, India, Morocco, Saudi Arabia and South Korea, Russia's GLONASS system, India's proposed Navigational Satellite System (IRNSS), and Japan's proposed QZSS regional system.
- COMPASS Beidou
- GLONASS GLONASS
- IRNSS India's proposed Navigational Satellite System
- QZSS Japan's proposed QZSS regional system.
- the mobile device comprises an image-capture module 725 configured to capture one or more images.
- the image-capture module may be integrated or operatively coupled with one or both of the display control module and the image generation module.
- the image-capture module is integrated or operatively associated with sensors 745 such as tilt sensors, orientation sensors, accelerometers, light sensors and the like.
- sensor modules are further configured to process signals detected by the sensors.
- the mobile device comprises one or more image parameter modules configured to determine parameters indicative of the captured images including but not limited to tilt parameter, orientation parameter, zoom parameter, motion parameter, and light parameter.
- the image parameter modules are integrated or operatively connected with the respective sensor modules.
- the image parameter modules may be integrated or operatively connected with one or both of the display control and image generation modules.
- the mobile device comprises one or more image display modules 750 .
- the image display modules may be communicatively coupled or integrated with one or more of the display control module, the image generation module and the image-capture module.
- the image generation module 735 is configured to obtain one or more input images from an image capture module and generate one or more output images at least in part based on the obtained input image.
- the image generation module generates the one or more output images by processing the obtained one or more input images to integrate navigational information obtained from a communicatively linked navigational information module 760 wherein processing the one or more input images includes recognizing one or more objects in the one or more input images.
- the mobile device is communicatively linked to an image generation module 735 .
- the image generation module 735 is located remotely at one or more servers accessed via a network.
- the mobile device comprises an image generation module 735 .
- the image generation module may be integrated or operatively linked with the display control module 715 .
- the image generation module may be integrated or operatively linked with the image-capture module.
- an image generation module is communicatively linked to one or more map and image databases 780 storing map and image information.
- one or more map and image databases are stored locally in the image generation module.
- one or more map and image databases are communicatively linked to the image generation module.
- the map and image databases are located remotely at one or more map and image servers accessed via a network.
- the image generation module 735 comprises an object recognition module 736 configured to process the input image to recognize one or more objects.
- the image generation module 735 also comprises an object rendering module 737 configured to render objects indicative of navigational information and at least in part based one or more objects recognized in the input image.
- the image generation module 735 also comprises an integration module 738 , configured to integrate the navigational information with the input image.
- the processing required for the generation of the output image may be performed by the mobile device, by a server managed exemplarily by a service and application provider, or by a combination of both.
- Information used in said processing may at least partly be stored on the mobile device, on a server of a service and application provider, or on a server of a data and content provider communicatively linked with the appropriate processing entities, such as described above, or a combination thereof.
- the distribution of the processing burden is based on the amount of data available for processing and the available communication bandwidth for data transfer between the various processing entities.
- the mobile device comprises a navigational information module 760 configured to obtain location data from the location identification module and generate navigational information, such as navigational instructions, at least in part based on the obtained location data.
- a navigational information module 760 configured to obtain location data from the location identification module and generate navigational information, such as navigational instructions, at least in part based on the obtained location data.
- one or more map databases with map information are stored locally in the navigational information module.
- one or more map databases are communicatively linked to the navigational information module.
- the map databases are located remotely at one or more map servers accessed via a network.
- the navigational information module may be integrated or operatively coupled with one or both of the location identification module and the image generation module.
- one or more remote servers monitor the current location of a user, for the purpose of obtaining real-time information, such as traffic updates and other location based services. Accordingly, the mobile device may be configured to regularly update its location information with the server. In some embodiments, the server also serves to provide navigation information based, at least partly, on the current location of a user. Accordingly, the mobile device may be configured to provide other route related requests to the server.
- the mobile device comprises one or more sensing modules including but not limited to light sensors and accelerometers.
- the sensor modules are further configured to process signals detected by the sensors.
- the mobile device is communicatively coupled to one or more external devices.
- the mobile device may be coupled to a windshield mounted camera or to car stereo speakers and display.
- the mobile device may be communicatively coupled to the external devices by means of a wireless or a hard-wired connection.
- Wireless communication can be achieved using any short-range or long-range communication protocols that may be readily known to a worker skilled in the art.
- the mobile device may be linked to some external devices using BluetoothTM and to others using radio frequency transmission.
- the mobile device may comprise additional short-range and long-range communication modules. Examples of short-range communications modules include a BluetoothTM module, an infrared module with associated circuits and components, and a TEL MUTE line for communication with similarly linked devices.
- FIG. 8 shows a block diagram depicting certain main components of an exemplary mobile device 810 with wireless communications capability in accordance with embodiments of the present technology. It should be understood that this figure is intentionally simplified to show only certain components; the mobile device 810 may include other components beyond those shown in FIG. 8 .
- the mobile device 810 is operatively associated with a GPS system 820 allowing for determining its location.
- the mobile device 810 is linked to a cellular network 830 through a base-station 832 , and thereby to one or more servers of a service and application provider, and to one or more servers of a data and content provider.
- Additional mobile devices 810 a may be linked to the cellular network 830 using the same base station 832 or a separate base station 832 a .
- the cellular network 830 thus serves for data transfer between peer mobile devices 810 , 810 a.
- the mobile device 810 further comprises a location identification module 818 , which is configured to determine the location of the mobile device.
- the location identification module 818 includes a GPS receiver chipset for receiving GPS radio signals transmitted from the one or more orbiting GPS satellites 820 .
- the GPS receiver chipset can be embedded within the device or externally connected, such as, for example, a BluetoothTM GPS puck or dongle. Other systems for determining location may be used in place of GPS, as would be readily understood by a worker skilled in the art.
- Navigational notifications are generated at least in part based on location data from the location identification module 818 .
- the device 810 comprises a processing module 812 , which includes a microprocessor 814 (or simply a “processor”) and operatively associated memory 816 (in the form of RAM or flash memory or both), to enable a variety of device functions and to execute an operating system for running software applications loaded on the device.
- the processing module 812 is configured to obtain one or more input images exemplarily from a video camera 890 , to obtain navigational information and instructions exemplarily from a remote server, to generate an output image and to control the display of the output image exemplarily by one or more displays 824 .
- the output image is generated by processing the input images to integrate the obtained navigational information wherein processing the input images includes recognizing one or more objects in the input images.
- the mobile device 810 includes a radiofrequency (RF) transceiver 822 for communicating wirelessly with the base station 832 of a wireless network 830 .
- the base station 832 may be a cellular base station, wireless access point, or the like.
- the base station 832 may vary as the mobile device travels, for example, using well-known handoff processes in cellular networks.
- the RF transceiver 822 may optionally be alternatively or additionally used for communicating directly with a peer device such as a third party mobile device, for example as may occur in some ad-hoc networks.
- the RF transceiver enables access to a wireless communication channel for transmitting and receiving data.
- the RF transceiver 822 may further allow for a wireless voice channel for transmitting and receiving voice communications, for example concurrently with transmission and reception of data over the same or a separate logical or physical channel.
- the mobile device 810 sends and receives communication signals via the RF transceiver 822 .
- the mobile device 810 may communicate in accordance with one or more appropriate technologies such as: Global Systems for Mobile communications (GSM), General Packet Radio Service (GPRS), Code Division Multiple Access (CDMA) technologies, Wideband CDMA (WCDMA), whether 2G, 3G, High speed packet access (HSPA), Universal Mobile Telecommunication System (UMTS) based technologies, Long Term Evolution (LTE) technologies, Orthogonal Frequency Division Multiplexing (OFDM) technologies, Ultra-Wideband (UWB) technologies, WiFiTM or WiMAXTM technologies, or other communication technologies and protocols as would readily be understood by a worker skilled in the art.
- GSM Global Systems for Mobile communications
- GPRS General Packet Radio Service
- CDMA Code Division Multiple Access
- WCDMA Wideband CDMA
- HSPA High speed packet access
- UMTS Universal Mobile Telecommunication System
- OFDM Orthogonal Frequency Division Multiplexing
- UWB Ultra
- the mobile device 810 may be capable of operation using multiple protocols.
- the base station 832 may be part of a wireless network, such as a cellular network, local-area network, wide-area network, wireless hotspot network, or the like.
- the mobile device, base station, network components, and the like may be configured for data communication, voice communication, or a combination thereof, possibly using additional components, configurations and procedures where appropriate, such as SIM cards, authorization and authentication procedures, handoff procedures, and the like, as would be readily understood by a worker skilled in the art.
- the mobile device 810 is linked to route information servers 860 which may include public or private route information servers.
- the route information servers are typically map servers that provide map data to networked computing devices or mobile devices upon request, for example, in vector format or alternatively as bitmaps.
- the route information servers e.g. the map servers
- These route information servers are communicatively connected to the for transmitting route information to each mobile device 810 in response to location data received from each mobile device.
- the “location data” comprises (1) the current location of the respective mobile device, e.g.
- the location data may also include a default location that has been specified, set or preconfigured by the sender or by the recipient for use in generating the route information in the event that the current location cannot be determined.
- the mobile device comprises one or more input/output devices or user interfaces (UI's) 824 .
- the mobile device 810 may include one or more of the following: a display (e.g. a small LCD screen), a thumbwheel and/or trackball, a keyboard, a touch screen, a keypad, a button, a speaker, a still camera and a video camera.
- the mobile device comprises one or more sensor modules 826 including but not limited to a microphone, an orientation sensor, an accelerometer, a light sensor etc.
- the mobile device comprises short-range communication modules 880 for communication between the mobile device and other similarly enabled external devices 882 .
- short-range communication modules 880 may include a Bluetooth communication module for communicating with a windshield mounted camera or a car stereo display. These devices 882 may also be used exemplarily to obtain information about the current location or to display navigational information.
Landscapes
- Engineering & Computer Science (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Automation & Control Theory (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Navigation (AREA)
Abstract
Description
Claims (21)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/819,521 US8762041B2 (en) | 2010-06-21 | 2010-06-21 | Method, device and system for presenting navigational information |
US14/288,848 US20140297185A1 (en) | 2010-06-21 | 2014-05-28 | Method, Device and System for Presenting Navigational Information |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/819,521 US8762041B2 (en) | 2010-06-21 | 2010-06-21 | Method, device and system for presenting navigational information |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/288,848 Continuation US20140297185A1 (en) | 2010-06-21 | 2014-05-28 | Method, Device and System for Presenting Navigational Information |
Publications (2)
Publication Number | Publication Date |
---|---|
US20110313653A1 US20110313653A1 (en) | 2011-12-22 |
US8762041B2 true US8762041B2 (en) | 2014-06-24 |
Family
ID=45329390
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/819,521 Active 2031-05-24 US8762041B2 (en) | 2010-06-21 | 2010-06-21 | Method, device and system for presenting navigational information |
US14/288,848 Abandoned US20140297185A1 (en) | 2010-06-21 | 2014-05-28 | Method, Device and System for Presenting Navigational Information |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/288,848 Abandoned US20140297185A1 (en) | 2010-06-21 | 2014-05-28 | Method, Device and System for Presenting Navigational Information |
Country Status (1)
Country | Link |
---|---|
US (2) | US8762041B2 (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140244160A1 (en) * | 2013-01-04 | 2014-08-28 | MoneyDesktop, Inc. a Delaware Corporation | Augmented reality financial institution branch locator |
US20160019786A1 (en) * | 2014-07-17 | 2016-01-21 | Thinkware Corporation | System and method for providing augmented reality notification |
US9544491B2 (en) * | 2014-06-17 | 2017-01-10 | Furuno Electric Co., Ltd. | Maritime camera and control system |
WO2018081851A1 (en) * | 2016-11-03 | 2018-05-11 | Buy Somewhere Pty Ltd | Visualisation system and software architecture therefor |
US10122925B2 (en) | 2010-08-17 | 2018-11-06 | Nokia Technologies Oy | Method, apparatus, and computer program product for capturing image data |
US10267641B2 (en) * | 2015-09-29 | 2019-04-23 | Xiaomi Inc. | Navigation method and device |
US20210180980A1 (en) * | 2018-08-30 | 2021-06-17 | Continental Automotive Gmbh | Roadway mapping device |
Families Citing this family (114)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101019915B1 (en) * | 2010-03-05 | 2011-03-08 | 팅크웨어(주) | Server, navigation system, navigation for vehicle and video providing method of navigation for vehicel |
JP5057184B2 (en) * | 2010-03-31 | 2012-10-24 | アイシン・エィ・ダブリュ株式会社 | Image processing system and vehicle control system |
JP2011247860A (en) * | 2010-05-31 | 2011-12-08 | Denso Corp | Navigation apparatus |
WO2011161487A1 (en) | 2010-06-21 | 2011-12-29 | Nokia Corporation | Apparatus, method and computer program for adjustable noise cancellation |
KR101357262B1 (en) * | 2010-08-13 | 2014-01-29 | 주식회사 팬택 | Apparatus and Method for Recognizing Object using filter information |
JP2012065263A (en) * | 2010-09-17 | 2012-03-29 | Olympus Imaging Corp | Imaging apparatus |
KR101229795B1 (en) * | 2010-12-20 | 2013-02-05 | 주식회사 케이티 | Method of providing user interface of route notifying terminal and apparatus for the same |
TW201227381A (en) * | 2010-12-20 | 2012-07-01 | Ind Tech Res Inst | Real-time traffic situation awareness system and method |
US8612149B2 (en) * | 2011-02-10 | 2013-12-17 | Blackberry Limited | System and method of relative location detection using image perspective analysis |
US9398210B2 (en) * | 2011-02-24 | 2016-07-19 | Digimarc Corporation | Methods and systems for dealing with perspective distortion in connection with smartphone cameras |
US9664527B2 (en) * | 2011-02-25 | 2017-05-30 | Nokia Technologies Oy | Method and apparatus for providing route information in image media |
US9342610B2 (en) * | 2011-08-25 | 2016-05-17 | Microsoft Technology Licensing, Llc | Portals: registered objects as virtualized, personalized displays |
US20130061147A1 (en) * | 2011-09-07 | 2013-03-07 | Nokia Corporation | Method and apparatus for determining directions and navigating to geo-referenced places within images and videos |
JP2013080326A (en) * | 2011-10-03 | 2013-05-02 | Sony Corp | Image processing device, image processing method, and program |
US9047688B2 (en) | 2011-10-21 | 2015-06-02 | Here Global B.V. | Depth cursor and depth measurement in images |
US8553942B2 (en) | 2011-10-21 | 2013-10-08 | Navteq B.V. | Reimaging based on depthmap information |
US9116011B2 (en) | 2011-10-21 | 2015-08-25 | Here Global B.V. | Three dimensional routing |
US9404764B2 (en) | 2011-12-30 | 2016-08-02 | Here Global B.V. | Path side imagery |
US9024970B2 (en) | 2011-12-30 | 2015-05-05 | Here Global B.V. | Path side image on map overlay |
US8930141B2 (en) * | 2011-12-30 | 2015-01-06 | Nokia Corporation | Apparatus, method and computer program for displaying points of interest |
US8749634B2 (en) * | 2012-03-01 | 2014-06-10 | H4 Engineering, Inc. | Apparatus and method for automatic video recording |
US9082239B2 (en) | 2012-03-14 | 2015-07-14 | Flextronics Ap, Llc | Intelligent vehicle for assisting vehicle occupants |
US9384609B2 (en) | 2012-03-14 | 2016-07-05 | Autoconnect Holdings Llc | Vehicle to vehicle safety and traffic communications |
CN102607579A (en) * | 2012-03-14 | 2012-07-25 | 深圳市赛格导航科技股份有限公司 | Vehicle-mounted navigation terminal and system |
WO2014172327A1 (en) | 2013-04-15 | 2014-10-23 | Flextronics Ap, Llc | Synchronization between vehicle and user device calendar |
US20140310379A1 (en) | 2013-04-15 | 2014-10-16 | Flextronics Ap, Llc | Vehicle initiated communications with third parties via virtual personality |
US9378601B2 (en) | 2012-03-14 | 2016-06-28 | Autoconnect Holdings Llc | Providing home automation information via communication with a vehicle |
US9412273B2 (en) | 2012-03-14 | 2016-08-09 | Autoconnect Holdings Llc | Radar sensing and emergency response vehicle detection |
US9147298B2 (en) * | 2012-03-14 | 2015-09-29 | Flextronics Ap, Llc | Behavior modification via altered map routes based on user profile information |
KR101923929B1 (en) * | 2012-06-06 | 2018-11-30 | 삼성전자주식회사 | Mobile communication terminal device for providing augmented reality service and method for changing to display for augmented reality service |
US20130328867A1 (en) * | 2012-06-06 | 2013-12-12 | Samsung Electronics Co. Ltd. | Apparatus and method for providing augmented reality information using three dimension map |
KR102021050B1 (en) * | 2012-06-06 | 2019-09-11 | 삼성전자주식회사 | Method for providing navigation information, machine-readable storage medium, mobile terminal and server |
US9142003B2 (en) * | 2012-06-10 | 2015-09-22 | Apple Inc. | Adaptive frame rate control |
US9488492B2 (en) * | 2014-03-18 | 2016-11-08 | Sri International | Real-time system for multi-modal 3D geospatial mapping, object recognition, scene annotation and analytics |
JP6064544B2 (en) * | 2012-11-27 | 2017-01-25 | ソニー株式会社 | Image processing apparatus, image processing method, program, and terminal device |
WO2014101742A1 (en) * | 2012-12-27 | 2014-07-03 | Harman International Industries, Incorporated | Vehicle navigation |
CN103968824B (en) * | 2013-01-28 | 2018-04-10 | 华为终端(东莞)有限公司 | One kind finds augmented reality mesh calibration method and terminal |
US9251627B2 (en) * | 2013-03-05 | 2016-02-02 | Sears Brands, L.L.C. | Removable dashboard instrument system |
KR20230173231A (en) | 2013-03-11 | 2023-12-26 | 매직 립, 인코포레이티드 | System and method for augmented and virtual reality |
US9514650B2 (en) | 2013-03-13 | 2016-12-06 | Honda Motor Co., Ltd. | System and method for warning a driver of pedestrians and other obstacles when turning |
KR102560629B1 (en) | 2013-03-15 | 2023-07-26 | 매직 립, 인코포레이티드 | Display system and method |
EP2976694B1 (en) | 2013-03-20 | 2019-10-09 | Nokia Technologies Oy | A touch display device with tactile feedback |
CN104123313B (en) * | 2013-04-28 | 2017-08-01 | 腾讯科技(深圳)有限公司 | Obtain the method and server of guide to visitors information |
US9316503B2 (en) | 2013-04-28 | 2016-04-19 | Tencent Technology (Shenzhen) Company Limited | Method, server, and system for obtaining navigation information |
CN103383259B (en) * | 2013-07-11 | 2016-05-04 | 北京奇虎科技有限公司 | Realize the method and system of route guiding by streetscape service |
US20150019125A1 (en) * | 2013-07-15 | 2015-01-15 | Eduard Mazanec | Gps navigation directions display device, system, and process with points of interest |
KR102222336B1 (en) * | 2013-08-19 | 2021-03-04 | 삼성전자주식회사 | User terminal device for displaying map and method thereof |
US9292764B2 (en) * | 2013-09-17 | 2016-03-22 | Qualcomm Incorporated | Method and apparatus for selectively providing information on objects in a captured image |
KR20150033162A (en) * | 2013-09-23 | 2015-04-01 | 삼성전자주식회사 | Compositor and system-on-chip having the same, and driving method thereof |
KR20150034997A (en) * | 2013-09-27 | 2015-04-06 | 네이버 주식회사 | Method and system for notifying destination by route guide |
US20150149081A1 (en) * | 2013-11-22 | 2015-05-28 | Oscar L. Kramer, JR. | Methods for Enhancing the Display of Electronic Nautical Charts with Depths Corrected for Tide |
US9485416B2 (en) * | 2014-03-11 | 2016-11-01 | Wipro Limited | Method and a guided imaging unit for guiding a user to capture an image |
US9638538B2 (en) | 2014-10-14 | 2017-05-02 | Uber Technologies, Inc. | Street-level guidance via route path |
US9569692B2 (en) * | 2014-10-31 | 2017-02-14 | The Nielsen Company (Us), Llc | Context-based image recognition for consumer market research |
US9443425B2 (en) * | 2014-11-06 | 2016-09-13 | Myine Electronics, Inc. | Methods and systems for destination congestion avoidance |
US10104023B2 (en) * | 2015-04-16 | 2018-10-16 | Oath Inc. | Location sharing |
US10068147B2 (en) * | 2015-04-30 | 2018-09-04 | Samsung Electronics Co., Ltd. | System and method for insertion of photograph taker into a photograph |
US9927246B2 (en) * | 2015-05-27 | 2018-03-27 | Here Global B.V. | Method, apparatus and computer program product for providing navigation information in relation to augmented reality guidance |
US10147211B2 (en) | 2015-07-15 | 2018-12-04 | Fyusion, Inc. | Artificially rendering images using viewpoint interpolation and extrapolation |
US11095869B2 (en) | 2015-09-22 | 2021-08-17 | Fyusion, Inc. | System and method for generating combined embedded multi-view interactive digital media representations |
US10222932B2 (en) | 2015-07-15 | 2019-03-05 | Fyusion, Inc. | Virtual reality environment based manipulation of multilayered multi-view interactive digital media representations |
US11006095B2 (en) | 2015-07-15 | 2021-05-11 | Fyusion, Inc. | Drone based capture of a multi-view interactive digital media |
US10242474B2 (en) | 2015-07-15 | 2019-03-26 | Fyusion, Inc. | Artificially rendering images using viewpoint interpolation and extrapolation |
CN105100501B (en) * | 2015-08-27 | 2018-05-18 | 黑龙江科技大学 | A kind of mobile phone computing system based on Internet of Things |
US11783864B2 (en) | 2015-09-22 | 2023-10-10 | Fyusion, Inc. | Integration of audio into a multi-view interactive digital media representation |
US10692126B2 (en) | 2015-11-17 | 2020-06-23 | Nio Usa, Inc. | Network-based system for selling and servicing cars |
KR101866728B1 (en) * | 2016-04-25 | 2018-06-15 | 현대자동차주식회사 | Navigation apparatus, vehicle and method for controlling vehicle |
US20180012197A1 (en) | 2016-07-07 | 2018-01-11 | NextEv USA, Inc. | Battery exchange licensing program based on state of charge of battery pack |
US9928734B2 (en) | 2016-08-02 | 2018-03-27 | Nio Usa, Inc. | Vehicle-to-pedestrian communication systems |
US11202017B2 (en) | 2016-10-06 | 2021-12-14 | Fyusion, Inc. | Live style transfer on a mobile device |
CN106504395A (en) * | 2016-10-25 | 2017-03-15 | 合肥东上多媒体科技有限公司 | A kind of based on the e-platform for carrying out digitization displaying to museum |
US10168173B2 (en) * | 2016-10-26 | 2019-01-01 | Google Llc | Systems and methods for using visual landmarks in initial navigation |
US10110871B2 (en) * | 2016-10-31 | 2018-10-23 | Disney Enterprises, Inc. | Recording high fidelity digital immersive experiences through off-device computation |
US10031523B2 (en) | 2016-11-07 | 2018-07-24 | Nio Usa, Inc. | Method and system for behavioral sharing in autonomous vehicles |
US10708547B2 (en) | 2016-11-11 | 2020-07-07 | Nio Usa, Inc. | Using vehicle sensor data to monitor environmental and geologic conditions |
US10694357B2 (en) | 2016-11-11 | 2020-06-23 | Nio Usa, Inc. | Using vehicle sensor data to monitor pedestrian health |
US10410064B2 (en) | 2016-11-11 | 2019-09-10 | Nio Usa, Inc. | System for tracking and identifying vehicles and pedestrians |
US10515390B2 (en) | 2016-11-21 | 2019-12-24 | Nio Usa, Inc. | Method and system for data optimization |
US11892311B2 (en) | 2016-11-26 | 2024-02-06 | Thinkware Corporation | Image processing apparatus, image processing method, computer program and computer readable recording medium |
CN108121764B (en) | 2016-11-26 | 2022-03-11 | 星克跃尔株式会社 | Image processing device, image processing method, computer program, and computer-readable recording medium |
US10249104B2 (en) | 2016-12-06 | 2019-04-02 | Nio Usa, Inc. | Lease observation and event recording |
US10074223B2 (en) | 2017-01-13 | 2018-09-11 | Nio Usa, Inc. | Secured vehicle for user use only |
US10031521B1 (en) | 2017-01-16 | 2018-07-24 | Nio Usa, Inc. | Method and system for using weather information in operation of autonomous vehicles |
US9984572B1 (en) | 2017-01-16 | 2018-05-29 | Nio Usa, Inc. | Method and system for sharing parking space availability among autonomous vehicles |
US10471829B2 (en) | 2017-01-16 | 2019-11-12 | Nio Usa, Inc. | Self-destruct zone and autonomous vehicle navigation |
US10286915B2 (en) | 2017-01-17 | 2019-05-14 | Nio Usa, Inc. | Machine learning for personalized driving |
US10464530B2 (en) | 2017-01-17 | 2019-11-05 | Nio Usa, Inc. | Voice biometric pre-purchase enrollment for autonomous vehicles |
US10437879B2 (en) | 2017-01-18 | 2019-10-08 | Fyusion, Inc. | Visual search using multi-view interactive digital media representations |
US10897469B2 (en) | 2017-02-02 | 2021-01-19 | Nio Usa, Inc. | System and method for firewalls between vehicle networks |
US10684136B2 (en) | 2017-02-28 | 2020-06-16 | International Business Machines Corporation | User-friendly navigation system |
US10313651B2 (en) | 2017-05-22 | 2019-06-04 | Fyusion, Inc. | Snapshots at predefined intervals or angles |
US11069147B2 (en) | 2017-06-26 | 2021-07-20 | Fyusion, Inc. | Modification of multi-view interactive digital media representation |
US10234302B2 (en) | 2017-06-27 | 2019-03-19 | Nio Usa, Inc. | Adaptive route and motion planning based on learned external and internal vehicle environment |
US10369974B2 (en) | 2017-07-14 | 2019-08-06 | Nio Usa, Inc. | Control and coordination of driverless fuel replenishment for autonomous vehicles |
US10710633B2 (en) | 2017-07-14 | 2020-07-14 | Nio Usa, Inc. | Control of complex parking maneuvers and autonomous fuel replenishment of driverless vehicles |
US10837790B2 (en) | 2017-08-01 | 2020-11-17 | Nio Usa, Inc. | Productive and accident-free driving modes for a vehicle |
US10393528B2 (en) * | 2017-08-02 | 2019-08-27 | Wing Aviation Llc | Systems and methods for navigation path determination for unmanned vehicles |
US10635109B2 (en) | 2017-10-17 | 2020-04-28 | Nio Usa, Inc. | Vehicle path-planner monitor and controller |
US10935978B2 (en) | 2017-10-30 | 2021-03-02 | Nio Usa, Inc. | Vehicle self-localization using particle filters and visual odometry |
US10606274B2 (en) | 2017-10-30 | 2020-03-31 | Nio Usa, Inc. | Visual place recognition based self-localization for autonomous vehicles |
US10717412B2 (en) | 2017-11-13 | 2020-07-21 | Nio Usa, Inc. | System and method for controlling a vehicle using secondary access methods |
KR102014261B1 (en) | 2017-12-12 | 2019-08-26 | 엘지전자 주식회사 | Vehicle control device mounted on vehicle and method for controlling the vehicle |
US10592747B2 (en) | 2018-04-26 | 2020-03-17 | Fyusion, Inc. | Method and apparatus for 3-D auto tagging |
US10369966B1 (en) | 2018-05-23 | 2019-08-06 | Nio Usa, Inc. | Controlling access to a vehicle using wireless access devices |
US20200019782A1 (en) * | 2018-07-13 | 2020-01-16 | International Business Machines Corporation | Accommodating object occlusion in point-of-view displays |
CN112639579B (en) | 2018-08-31 | 2023-09-15 | 奇跃公司 | Spatially resolved dynamic dimming for augmented reality devices |
US11118931B2 (en) * | 2018-11-28 | 2021-09-14 | International Business Machines Corporation | Enhanced endpoint directions in a mapping program |
JP2022517207A (en) | 2019-01-11 | 2022-03-07 | マジック リープ, インコーポレイテッド | Time-multiplexed display of virtual content at various depths |
JP7261090B2 (en) * | 2019-05-30 | 2023-04-19 | フォルシアクラリオン・エレクトロニクス株式会社 | In-vehicle device, driving assistance method, and driving assistance system |
CN114222942A (en) | 2019-08-12 | 2022-03-22 | 奇跃公司 | System and method for virtual and augmented reality |
KR20210129974A (en) | 2020-04-21 | 2021-10-29 | 현대자동차주식회사 | Apparatus for displaying of a vehicle and method thereof |
KR20210129976A (en) * | 2020-04-21 | 2021-10-29 | 현대자동차주식회사 | Apparatus for displaying of a vehicle and method thereof |
US11302077B2 (en) | 2020-05-29 | 2022-04-12 | Snap Inc. | Augmented reality guidance that generates guidance markers |
DE102021128251A1 (en) | 2021-10-29 | 2023-05-04 | Joynext Gmbh | Viewing a route using a mobile device |
Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5815411A (en) | 1993-09-10 | 1998-09-29 | Criticom Corporation | Electro-optic vision system which exploits position and attitude |
US6181302B1 (en) | 1996-04-24 | 2001-01-30 | C. Macgill Lynde | Marine navigation binoculars with virtual display superimposing real world image |
US6208933B1 (en) | 1998-12-04 | 2001-03-27 | Northrop Grumman Corporation | Cartographic overlay on sensor video |
US6903707B2 (en) | 2000-08-09 | 2005-06-07 | Information Decision Technologies, Llc | Method for using a motorized camera mount for tracking in augmented reality |
US6917370B2 (en) | 2002-05-13 | 2005-07-12 | Charles Benton | Interacting augmented reality and virtual reality |
US20060038833A1 (en) | 2004-08-19 | 2006-02-23 | Mallinson Dominic S | Portable augmented reality device and method |
US7089110B2 (en) * | 2002-04-30 | 2006-08-08 | Telmap Ltd. | Dynamic navigation system |
US20060177103A1 (en) * | 2005-01-07 | 2006-08-10 | Evan Hildreth | Optical flow based tilt sensor |
US20060195858A1 (en) | 2004-04-15 | 2006-08-31 | Yusuke Takahashi | Video object recognition device and recognition method, video annotation giving device and giving method, and program |
US20070088497A1 (en) | 2005-06-14 | 2007-04-19 | Jung Mun H | Matching camera-photographed image with map data in portable terminal and travel route guidance method |
US20080039120A1 (en) * | 2006-02-24 | 2008-02-14 | Telmap Ltd. | Visual inputs for navigation |
US7460953B2 (en) | 2004-06-30 | 2008-12-02 | Navteq North America, Llc | Method of operating a navigation system using images |
US7511736B2 (en) | 2003-03-10 | 2009-03-31 | Charles Benton | Augmented reality navigation system |
WO2010012310A1 (en) | 2008-07-31 | 2010-02-04 | Tele Atlas B.V. | Method of displaying navigation data in 3d |
US20110098918A1 (en) * | 2009-10-28 | 2011-04-28 | Google Inc. | Navigation Images |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6192314B1 (en) * | 1998-03-25 | 2001-02-20 | Navigation Technologies Corp. | Method and system for route calculation in a navigation application |
US20070070069A1 (en) * | 2005-09-26 | 2007-03-29 | Supun Samarasekera | System and method for enhanced situation awareness and visualization of environments |
US20110178702A1 (en) * | 2010-01-20 | 2011-07-21 | Sony Ericsson Mobile Communications Ab | Optimum travel times |
CN103562680B (en) * | 2011-06-03 | 2016-06-29 | 苹果公司 | For comparing and select the apparatus and method for of alternative navigation route |
-
2010
- 2010-06-21 US US12/819,521 patent/US8762041B2/en active Active
-
2014
- 2014-05-28 US US14/288,848 patent/US20140297185A1/en not_active Abandoned
Patent Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5815411A (en) | 1993-09-10 | 1998-09-29 | Criticom Corporation | Electro-optic vision system which exploits position and attitude |
US6181302B1 (en) | 1996-04-24 | 2001-01-30 | C. Macgill Lynde | Marine navigation binoculars with virtual display superimposing real world image |
US6208933B1 (en) | 1998-12-04 | 2001-03-27 | Northrop Grumman Corporation | Cartographic overlay on sensor video |
US6903707B2 (en) | 2000-08-09 | 2005-06-07 | Information Decision Technologies, Llc | Method for using a motorized camera mount for tracking in augmented reality |
US7089110B2 (en) * | 2002-04-30 | 2006-08-08 | Telmap Ltd. | Dynamic navigation system |
US6917370B2 (en) | 2002-05-13 | 2005-07-12 | Charles Benton | Interacting augmented reality and virtual reality |
US7511736B2 (en) | 2003-03-10 | 2009-03-31 | Charles Benton | Augmented reality navigation system |
US20060195858A1 (en) | 2004-04-15 | 2006-08-31 | Yusuke Takahashi | Video object recognition device and recognition method, video annotation giving device and giving method, and program |
US7460953B2 (en) | 2004-06-30 | 2008-12-02 | Navteq North America, Llc | Method of operating a navigation system using images |
US20060038833A1 (en) | 2004-08-19 | 2006-02-23 | Mallinson Dominic S | Portable augmented reality device and method |
US20060177103A1 (en) * | 2005-01-07 | 2006-08-10 | Evan Hildreth | Optical flow based tilt sensor |
US20070088497A1 (en) | 2005-06-14 | 2007-04-19 | Jung Mun H | Matching camera-photographed image with map data in portable terminal and travel route guidance method |
US20080039120A1 (en) * | 2006-02-24 | 2008-02-14 | Telmap Ltd. | Visual inputs for navigation |
WO2010012310A1 (en) | 2008-07-31 | 2010-02-04 | Tele Atlas B.V. | Method of displaying navigation data in 3d |
US20110098918A1 (en) * | 2009-10-28 | 2011-04-28 | Google Inc. | Navigation Images |
Non-Patent Citations (20)
Title |
---|
"Augmented Driving" (video) http://www.dailymotion.com/video/xcyr2b-augmented-driving-auto?start=57. |
"Augmented Reality on Android-DailyMobile.se" (video) http://www.youtube.com/watch?v=PdmSVCDmJGA. |
"AugSatNav-Augmented Reality Navigation System by Phyora" (video) http://www.youtube.com/watch?v=nFJRGab4oKQ. |
"Instar-Car Navigation (2002)-Ars Electronica Futurelab" (video) http://www.youtube.com/watch?v=sPCk2vGk8g0. |
"IPhone Augmented Reality Navigation" (video) http://www.youtube.com/watch?v=rgXzdUb-fug. |
"Augmented Driving" (video) http://www.dailymotion.com/video/xcyr2b—augmented-driving—auto?start=57. |
"Augmented Reality on Android—DailyMobile.se" (video) http://www.youtube.com/watch?v=PdmSVCDmJGA. |
"AugSatNav—Augmented Reality Navigation System by Phyora" (video) http://www.youtube.com/watch?v=nFJRGab4oKQ. |
"Instar—Car Navigation (2002)—Ars Electronica Futurelab" (video) http://www.youtube.com/watch?v=sPCk2vGk8g0. |
"IPhone Augmented Reality Navigation" (video) http://www.youtube.com/watch?v=rgXzdUb—fug. |
Azuma, R., A Survey of Augmented Reality, Presence: Teleoperators and Virtual Environments 6, Aug. 4, 1997, p. 355-385. |
Hu, Z. et al., Towards a New Generation of Car Navigation System-Data Fusion Technology in Solving On-Board Camera Registration Problem, Computer Science Department, Kumamoto University, Japan, 12 pages. |
Hu, Z. et al., Towards a New Generation of Car Navigation System—Data Fusion Technology in Solving On-Board Camera Registration Problem, Computer Science Department, Kumamoto University, Japan, 12 pages. |
Hu, Z. et al., Vision-based Car Navigation System (webpage), Jan. 29, 2010, http://navi.cs.kumamoto-u.ac.jp/english/projects/08.html. |
Ito, K. et al., Car Navigation System with Image Recognition, Abstract, Big. Techn. Papers-Intern. Conf. on Consumer Elect., Pioneer, 2009, 2 pages. |
Ito, K. et al., Car Navigation System with Image Recognition, Abstract, Big. Techn. Papers—Intern. Conf. on Consumer Elect., Pioneer, 2009, 2 pages. |
Making Virtual Solid, LLC, "Virtual CableTM Car Navigation-Safe, Simple and Intuitive" (webpage), Jan. 29, 2010, http://www.mvs.net/. |
Making Virtual Solid, LLC, "Virtual CableTM Car Navigation—Safe, Simple and Intuitive" (webpage), Jan. 29, 2010, http://www.mvs.net/. |
Nakatsuru, T. et al., Image Overlay on Optical See-through Displays for Vehicle Navigation, Proceedings of the Second IEEE and ACM International Symposium on Mixed and Augmented Reality, 2003. |
Tae-Hyun, H. et al., Detection of Traffic Lights for Vision-Based Car Navigation System, L.W. Chang, W.N. Lie and R. Chiang (Eds.): PSIVT 2006, LNCS 4319, pp. 682-691, 2006. |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10122925B2 (en) | 2010-08-17 | 2018-11-06 | Nokia Technologies Oy | Method, apparatus, and computer program product for capturing image data |
US20140244160A1 (en) * | 2013-01-04 | 2014-08-28 | MoneyDesktop, Inc. a Delaware Corporation | Augmented reality financial institution branch locator |
US9383218B2 (en) * | 2013-01-04 | 2016-07-05 | Mx Technologies, Inc. | Augmented reality financial institution branch locator |
US9544491B2 (en) * | 2014-06-17 | 2017-01-10 | Furuno Electric Co., Ltd. | Maritime camera and control system |
US20160019786A1 (en) * | 2014-07-17 | 2016-01-21 | Thinkware Corporation | System and method for providing augmented reality notification |
US9773412B2 (en) * | 2014-07-17 | 2017-09-26 | Thinkware Corporation | System and method for providing augmented reality notification |
US9905128B2 (en) * | 2014-07-17 | 2018-02-27 | Thinkware Corporation | System and method for providing augmented reality notification |
US10267641B2 (en) * | 2015-09-29 | 2019-04-23 | Xiaomi Inc. | Navigation method and device |
WO2018081851A1 (en) * | 2016-11-03 | 2018-05-11 | Buy Somewhere Pty Ltd | Visualisation system and software architecture therefor |
US20210180980A1 (en) * | 2018-08-30 | 2021-06-17 | Continental Automotive Gmbh | Roadway mapping device |
US12078505B2 (en) * | 2018-08-30 | 2024-09-03 | Continental Automotive Gmbh | Roadway mapping device |
Also Published As
Publication number | Publication date |
---|---|
US20110313653A1 (en) | 2011-12-22 |
US20140297185A1 (en) | 2014-10-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8762041B2 (en) | Method, device and system for presenting navigational information | |
CA2743941C (en) | Method, device and system for presenting navigational information | |
US11543256B2 (en) | Electronic apparatus and control method thereof | |
US9429438B2 (en) | Updating map data from camera images | |
CN108680173B (en) | Electronic device, control method of electronic device, and computer-readable recording medium | |
US9454847B2 (en) | System and method of indicating transition between street level images | |
KR101147748B1 (en) | A mobile telecommunication device having a geographic information providing function and the method thereof | |
US9351109B2 (en) | Method, device and system for enhancing location information | |
US8682579B2 (en) | System and method for displaying address information on a map | |
KR101303948B1 (en) | Apparatus and Method for Providing Augmented Reality Information of invisible Reality Object | |
KR102255432B1 (en) | Electronic apparatus and control method thereof | |
CA2762743C (en) | Updating map data from camera images | |
JP5886688B2 (en) | Information processing apparatus, information processing method, and program | |
US20080228393A1 (en) | Navigation device and method | |
CN111710189B (en) | Control method for electronic device, and recording medium | |
JP7082588B2 (en) | Programs, navigation systems, navigation devices | |
JP7046555B2 (en) | In-vehicle device, server, display method, transmission method | |
KR102506263B1 (en) | Apparatus, method, and computer readable recording medium for aerial image process, navigation apparatus and server for path guide using aerial map | |
US20240104925A1 (en) | Satellite-enabled off-road obstacle detection | |
EP2437556B1 (en) | Method, system and computer program product for enhancing location information | |
JP2016146186A (en) | Information processing device, information processing method, and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: RESEARCH IN MOTION LIMITED, CANADA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:LINDNER, JEFFERY ERHARD;REEL/FRAME:024566/0739 Effective date: 20100616 |
|
AS | Assignment |
Owner name: BLACKBERRY LIMITED, ONTARIO Free format text: CHANGE OF NAME;ASSIGNOR:RESEARCH IN MOTION LIMITED;REEL/FRAME:032683/0403 Effective date: 20130709 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551) Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
AS | Assignment |
Owner name: MALIKIE INNOVATIONS LIMITED, IRELAND Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BLACKBERRY LIMITED;REEL/FRAME:064104/0103 Effective date: 20230511 |
|
AS | Assignment |
Owner name: MALIKIE INNOVATIONS LIMITED, IRELAND Free format text: NUNC PRO TUNC ASSIGNMENT;ASSIGNOR:BLACKBERRY LIMITED;REEL/FRAME:064270/0001 Effective date: 20230511 |