US20110188760A1 - System and Method for Creating and Displaying Map Projections related to Real-Time Images - Google Patents
System and Method for Creating and Displaying Map Projections related to Real-Time Images Download PDFInfo
- Publication number
- US20110188760A1 US20110188760A1 US12/699,545 US69954510A US2011188760A1 US 20110188760 A1 US20110188760 A1 US 20110188760A1 US 69954510 A US69954510 A US 69954510A US 2011188760 A1 US2011188760 A1 US 2011188760A1
- Authority
- US
- United States
- Prior art keywords
- real
- time
- map projection
- image
- objects
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 claims abstract 20
- 239000003550 marker Substances 0.000 claims abstract 5
- 238000000605 extraction Methods 0.000 claims 3
- 238000012800 visualization Methods 0.000 claims 3
- 230000009466 transformation Effects 0.000 claims 2
- 239000013598 vector Substances 0.000 claims 2
- 230000003044 adaptive effect Effects 0.000 claims 1
- 238000005553 drilling Methods 0.000 claims 1
- 230000004048 modification Effects 0.000 claims 1
- 238000012986 modification Methods 0.000 claims 1
- 238000004091 panning Methods 0.000 claims 1
- 230000001131 transforming effect Effects 0.000 claims 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/05—Geographic models
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B29/00—Maps; Plans; Charts; Diagrams, e.g. route diagram
- G09B29/10—Map spot or coordinate position indicators; Map reading aids
- G09B29/106—Map spot or coordinate position indicators; Map reading aids using electronic means
Definitions
- This application relates to an interactive visual presentation of a map projection on a user interface. Specifically, the application relates to a system and method for displaying a map projection on a user interface for depicting a reflected view of real-time objects displayed within a user/device's viewing area.
- GPS navigation systems that provide information about a user's surroundings and geographical information.
- Some GPS navigation systems appear as an animated 2D image on a specific device such as a BlackBerry®, iPhone®, cell phone or a specific GPS device such as Garmin®.
- These 2D systems are designed to provide navigation information on an animated street map view.
- the GPS systems can provide a street view, a satellite view and a hybrid view of a map.
- it can be distracting for a user to view their surroundings while looking at a secondary device. It can also be disorienting for a user to try to relate the animated map view to the real world images seen by the user and attempt navigation at the same time.
- these GPS devices/applications provide a limited amount of information about the user's surroundings. For example, they may provide an animated street map view, a compass, and directional information to navigate the user to a desired location.
- 3D GPS systems that provide a limited amount of navigation information on windshields of vehicles or airplanes.
- Such navigation systems referred to as heads up displays, project basic directional information for guiding a user regarding their current location and/or destination location. For example, in one case a virtual cable line is projected on a windshield to show the direction which the user should navigate on the highway.
- a 3D compass is projected onto the window of a fighter plane to show the current position, altitude and bearing of the plane to allow a pilot to know which direction they are facing.
- AR augmented reality
- AR is a term for a live view of a physical real-world environment whose elements are merged with virtual computer-generated imagery—creating a mixed reality.
- the computer-generated images are displayed as a layer over a user's view of the physical world.
- AR systems are also now in everyday use on mobile devices, such as iPhone® and BlackBerry® devices, where the device's screen is used to show merged live video from its camera with virtual elements placed on the display.
- Navigation using Augmented Reality methods typically is done using annotations and text descriptions on the live scene to provide information about the physical world. However, this approach occludes the live scene and provides a limited amount of information about the physical world. Further, it is difficult for a user to relate the virtual imagery to the physical world.
- the existing GPS systems present limited amount of information about a user's surroundings and present difficulties for a user to navigate to a desired location while referring to and trying correlate an animated map view on a GPS device screen to the real-world.
- a navigation system that provides correlation between real-world images and virtual map images and allows improved navigation.
- the virtual map images include for example, virtual aerial and/or satellite imagery, raster or vector map images.
- the navigation system displays a map projection comprising a reflected virtual view of real-time objects seen within a user/device's viewing area to aid in navigation and improve understanding of the current surroundings.
- a method for creating and displaying a map projection of a device's real-time viewing area to depict virtual objects providing a reflected view of real-time objects within the device's viewing area comprising: displaying a real-time image of the device's viewing area taken from a geographical location on a display; retrieving the map projection for revealing the reflected view as an elevated view of a ground surface about the device's current geographical location and in accordance with the device's viewing area; superimposing the map projection on the display in an upper portion of the real-time image; and defining one or more markers configured to show a relationship between the map projection and the real-time image, each marker overlaid on the display and configured to connect between the virtual object in the map projection and the corresponding real-time object on the real-time image.
- the map projection includes aerial and/or satellite imagery, a raster or a vector map projection.
- a method and system for transforming the map projection displays a reflected view of real-time objects in the real-time plane.
- the map projection is transformed to a surface having one of a parabolic curve surface, an arcuate surface, a flat surface, an angled planar surface, a surface of revolution curve, a surface shifted relative to the real-time objects displayed, a surface having at least a portion thereof magnified relative to the map projection, a surface having at least a portion thereof compressed relative to the map projection, a surface having a subset of virtual objects provided in the map projection.
- the selection of the transform being based on pre-defined criteria (i.e. user preferences for providing improved visibility of certain virtual objects or focusing on a specific region of the virtual objects).
- FIG. 1A is a block diagram of a data processing system for a visualization tool and FIG. 1B is a block diagram of further details;
- FIG. 2 is a block diagram of the visualization tool and associated components according to one embodiment
- FIG. 3 is a block diagram of the visualization tool having a transform module according to an alternate embodiment
- FIG. 4 is a representative screen shot of the visualization tool showing the projection image displayed on a real-time image and relationships therebetween in accordance with an embodiment thereof;
- FIGS. 5A-9C are representative views depicting alternate transforms applied by the transform module of the visualization tool to the projection image according to alternate embodiments
- FIG. 10 is a representative screen shot of the visualization tool showing the projection image displayed on a real-time image using one or more of a vector map, vectors, annotations and symbols to represent the projected virtual image in accordance with one embodiment;
- FIGS. 11-14 and 16 are representative views depicting alternate transforms applied by the transform module to the projection image according to alternate embodiments.
- FIG. 15 is a schematic diagram illustrating the process for creating and displaying a map projection according to one embodiment.
- a visualization data processing system 100 includes a visualization tool 12 for retrieving and processing a collection of virtual image information 14 as input data elements to a user interface 202 .
- the virtual image information 14 provides satellite and/or aerial imagery as projection images 22 (also referred to as map projections or virtual images herein).
- the virtual image information 14 further provides geographical information 24 associated with the projection images 22 .
- the geographical information 24 can include, for example, location information associated with the projected images 22 , names of locations within the projection images, or other characterizing information, physical characteristics of objects within the projection images 22 .
- the geographical information 24 can include any information that may be used for example, in graphical information systems (GIS).
- the visualization tool 12 is further configured for processing user/device events 109 .
- the user/device events 109 comprise real-time images 34 that are viewed/captured by a device 101 (see FIG. 1B ) as the location of the device 101 changes.
- the real-time images 34 may also be referred to as images captured within the device's 101 viewing area.
- the user/device events 109 further comprise real-time event information 32 that defines any temporal and spatial information related to the real-time images 34 viewed or captured at a predefined instance.
- the real-time event information 32 may provide information about the device's 101 current location, facing direction, distance to one or more pre-defined real-time objects, landmarks within and outside the device's 101 current viewing area, physical characteristics (i.e. dimensions) of objects within the viewing area, distances between objects in the viewing area, distance between a user/device's current positioning to one or more objects in the real-time viewing area.
- the event information 32 may further include information about locations on the real-time image where there is a lack of objects (i.e. determining an upper portion of the real-time image where there are no buildings, trees, people or other objects). This determination of a lack of objects may be used to define a location (i.e. an upper portion of the real-time image) where the projections image may be overlaid without blocking any objects.
- the virtual image information 14 can be combined with a respective set of associations 16 which define relationships of the projected images 22 and real-time images 34 by the tool 12 to generate an interactive visual representation 18 on the visual interface (VI) 202 .
- the set of associations 16 may be predefined by a user (i.e. analyst) of the tool 12 to define relationships between the virtual image information 14 and real-time images 34 provided by user events 109 .
- the set of associations 16 may be generated by the tool 12 based on projection geographical information 24 (i.e. location of virtual objects) and real-time event information 32 (i.e. estimated location of real-time objects) related to the real-time images 34 captured.
- the associations 16 may be generated by the tool 12 for example to link one or more objects in the device's 101 viewing area (the objects captured in real-time images 34 ) to corresponding virtual objects within the projection images 22 .
- the associations 16 are displayed on the visual representation 18 as markers connected between a virtual object in the projection image 22 and a corresponding real-time object in the real-time image 34 .
- the markers can alternatively be colour coded or otherwise visually emphasized to distinguish associations between real-time objects and virtual objects.
- the associations 16 are made by correlating the estimated geographic location of each real-time object (provided by the real-time event information 32 ) to the geographical location of each virtual object (provided by the projection geographical information 24 ).
- the definition of associations 16 may be user dependent, semi automated (i.e. defined by the tool 12 but modifiable by a user of the device 101 or fully automated (i.e. fully defined by the tool 12 ).
- Management of the virtual image information 14 , tool 12 , and associations 16 are driven by updated user events 109 of a user (not shown) of the tool 12 .
- the user events 109 comprise real-time images 34 captured and updated as the location of the device 101 ( FIG. 1B ) changes.
- the user events 109 further comprise event information 32 defining location and other GIS information related to each real-time image 34 .
- the user events 109 may be updated by interaction of a user (not shown) with the user interface 108 (see FIG. 1B ) during interaction with the visual representation 18 .
- the visual representation 18 shows connectivity between the projection images 22 and the real-time images 34 .
- the data processing system 100 of a device 101 includes the user interface device(s) 108 for interacting with the tool 12 , the user interface device(s) 108 being connected to a memory 102 via a BUS 106 .
- the device 101 comprises a computing device and may include for example a laptop or desktop computer, a mobile phone, a Personal Digital Assistant (PDA), virtual reality goggles, monocle, heads-up display systems, virtual reality device other types of computing devices as will be envisaged by a person skilled in the art.
- the interface device(s) 108 are coupled to a processor 104 via the BUS 106 , to interact with user events 109 to monitor or otherwise instruct the operation of the tool 12 via an operating system 110 .
- the user interface device(s) 108 can include one or more user input devices such as but not limited to a QWERTY keyboard, a keypad, a trackwheel, a stylus, a mouse, a microphone, a digital compass and an accelerometer.
- the visual interface 202 is considered to be a user output device, such as but not limited to a computer screen display, a mobile device display (such as a cell phone screen), goggles having a screen display (such as a virtual reality goggles). If the screen is touch sensitive, then the display can also be used as a user input device as controlled by the processor 104 .
- the data processing system 100 can include a computer readable storage medium 46 coupled to the processor 104 for providing instructions to the processor 104 and/or the tool 12 .
- the computer readable medium 46 can include hardware and/or software such as, by way of example only, magnetic disks, magnetic tape, optically readable medium such as CD/DVD ROMS, and memory cards.
- the computer readable medium 46 may take the form of a small disk, floppy diskette, cassette, hard disk drive, solid-state memory card, or RAM provided in the memory 102 . It should be noted that the above listed example computer readable mediums 46 can be used either alone or in combination.
- System 100 further comprises a network interface 47 comply the system 100 for communication with one or more public or private networks searches a LAN and/or the Internet.
- the tool 12 interacts via link 116 with a VI manager 112 (also known as a visualization renderer) of the system 100 for presenting the visual representation 18 on the visual interface 202 .
- the tool 12 processes virtual image information 14 , associations 16 , and user events 109 from data files or tables 122 of the memory 102 .
- the associations 16 may either be user-defined, or provided by the tool 12 (or a combination thereof). If the tool 12 provides the associations 16 , the associations 16 are determined based on analyzing the location of real-time objects (as provided by event information 32 for example by determining the distance between the real-time objects and the device 101 and determining the location of the real-time object by knowing the current location of the device 101 .
- the tool 12 then processes the information received from the tables 122 for subsequent presentation on the visual representation 18 . It is recognized that the virtual image information 14 , associations 16 and the user events 109 could be stored in the same or separate tables 122 , as desired.
- the tool 12 can receive requests for storing, retrieving, amending, or creating the virtual image information 14 , associations 16 via the tool 12 and/or directly via link 120 from the VI manager 112 , as driven by the user events 109 and/or independent operation of the tool 12 . Accordingly, the tool 12 and manager 112 coordinate the processing of data objects 14 , association set 16 and user events 109 with respect to the content of the screen representation 18 displayed in the visual interface 202 .
- the visualization tool 12 and the visual interface 202 may exist on separate devices (not shown) such that the process of creating the map projection is performed on a first device and the second device is used to render the map projection and the real-time image on the display.
- the task related instructions can comprise code and/or machine readable instructions for implementing predetermined functions/operations including those of an operating system, tool 12 , or other information processing system, for example, in response to command or input provided by a user of the system 100 .
- the processor 104 also referred to as module(s) for specific components of the tool 12 ) as used herein is a configured device and/or set of machine-readable instructions for performing operations as described by example above.
- the processor/modules in general may comprise any one or combination of, hardware, firmware, and/or software.
- the processor/modules acts upon information by manipulating, analyzing, modifying, converting or transmitting information for use by an executable procedure or an information device, and/or by routing the information with respect to an output device.
- the processor/modules may use or comprise the capabilities of a controller or microprocessor, for example. Accordingly, any of the functionality provided by the systems and process of the accompanying figures may be implemented in hardware, software or a combination of both. Accordingly, the use of a processor/modules as a device and/or as a set of machine readable instructions is hereafter referred to generically as a processor/module for sake of simplicity.
- storage means the devices and data connected to the computer through input/output operations such as hard disk and tape systems and other forms of storage not including computer memory and other in-computer storage.
- storage is divided into: (1) primary storage, which holds data in memory (sometimes called random access memory or RAM) and other “built-in” devices such as the processor's L1 cache, and (2) secondary storage, which holds data on hard disks, tapes, and other devices requiring input/output operations.
- primary storage which holds data in memory (sometimes called random access memory or RAM) and other “built-in” devices such as the processor's L1 cache
- secondary storage which holds data on hard disks, tapes, and other devices requiring input/output operations.
- Primary storage can be much faster to access than secondary storage because of the proximity of the storage to the processor or because of the nature of the storage devices.
- secondary storage can hold much more data than primary storage.
- primary storage includes read-only memory (ROM) and L1 and L2 cache memory.
- ROM read-only memory
- L1 and L2 cache memory In addition to hard disks, secondary storage includes a range of device types and technologies, including diskettes, Zip drives, redundant array of independent disks (RAID) systems, and holographic storage. Devices that hold storage are collectively known as storage media.
- a database is a further embodiment of memory 102 as a collection of information that is organized so that it can easily be accessed, managed, and updated.
- databases can be classified according to types of content: bibliographic, full-text, numeric, and images.
- databases are sometimes classified according to their organizational approach.
- a relational database is a tabular database in which data is defined so that it can be reorganized and accessed in a number of different ways.
- a distributed database is one that can be dispersed or replicated among different points in a network.
- An object-oriented programming database is one that is congruent with the data defined in object classes and subclasses.
- Computer databases typically contain aggregations of data records or files, such as sales transactions, product catalogs and inventories, and customer profiles.
- a database manager provides users the capabilities of controlling read/write access, specifying report generation, and analyzing usage.
- Databases and database managers are prevalent in large mainframe systems, but are also present in smaller distributed workstation and mid-range systems such as the AS/400 and on personal computers.
- SQL Structured Query Language
- IBM's DB2, Microsoft's Access and database products from Oracle, Sybase, and Computer Associates.
- Memory is a further embodiment of memory 102 storage as the electronic holding place for instructions and data that the computer's microprocessor can reach quickly.
- its memory When the computer is in normal operation, its memory usually contains the main parts of the operating system and some or all of the application programs and related data that are being used. Memory is often used as a shorter synonym for random access memory (RAM). This kind of memory is located on one or more microchips that are physically close to the microprocessor in the computer.
- RAM random access memory
- the visualization tool 12 comprises an extraction module 302 , an associations module 304 , and a leader module 306 .
- the visualization renderer 112 displays a real-time image 34 based on device 101 current geographical location, facing direction and viewing angle of the device 101 (step 1502 ).
- the extraction module 302 is configured for extracting projection images 22 based on user/device events 109 and real-time images 34 viewed by the device 101 (step 1504 ). As described herein, the projection images 22 provides a reflected view of objects seen in the real-time plane (real-time images 34 ).
- the projection images 22 may refer to satellite and/or aerial imagery and/or annotated vector graphics which provide a reflected view of objects in the real-time plane.
- the reflected view of objects provided by projection images 22 is taken from an elevated position about the device's 101 current geographical location and in accordance with the direction angle of the viewing area (as provided by event information 32 ).
- the reflected view 22 is a mirror image of the real-time objects in the real-time image 22 revealing a hidden view/angle of objects seen in the viewing area.
- the user/device events 109 provide event information 32 which defines the geographical coordinates of the device 101 and other GIS information related to the device's 101 viewing area.
- the event information 32 can provide information defining the device 101 current location and facing direction as it captures real-time images 34 within its viewing area.
- the event information 32 can provide for example, the current location of the device 101 , the direction the device 101 is facing, the estimated elevation of the device 101 relative to the surrounding ground surface, the estimated distance of the device 101 from surrounding objects and/or pre-defined landmarks associated with the real-time images 34 . Accordingly, based on the device 101 current geographical location and viewing area (i.e.
- the extracting module 302 retrieves the corresponding projection images 22 (also referred to as map projections) from the data store 122 .
- the projection images 22 depict an elevated view of a ground surface about the device 101 current geographical location and in accordance with the device 101 viewing area (i.e. direction the device 101 is facing).
- the projection images 22 can comprise aerial and/or satellite imagery and/or annotated vector graphics which provide a reflected top view of the device's 101 viewing area in accordance with the device's current geographical co-ordinates.
- the projection images 22 allow a user of the device 101 to reveal portions of the real-time objects hidden (i.e. a top surface of the real-time objects not seen by the user) from the device 101 viewing area.
- the visualization renderer 112 is configured to display the projection image 22 on the display 18 and overlaid in an upper portion of the real-time image 34 (step 1506 ).
- the associations module 304 is configured to associate real-time objects provided in real-time images 34 to virtual objects (also referred to as reflected objects) in projection images 22 (step 1508 ).
- the event information 32 provides estimated coordinates of one or more pre-defined real-time objects within the real-time image 34 .
- the estimated coordinates of each real-time object may be calculated for example, using the device 101 current location and estimated distance to each real-time object (provided by event information 32 ).
- the projection geographical information 24 contains geographical coordinates of one or more virtual objects provided in projection images 22 .
- the leader module 306 is configured to communicate with the visualization renderer 112 to display and overlay one or more markers for showing a relationship between virtual objects in the projection images 22 and real-time objects in real-time images 34 .
- Each marker is overlaid on the display and configured to connect between a virtual object in the projection image 22 and a corresponding real-time object in real-time images 34 (step 1508 ).
- the relationships between the objects being defined by the associations module 304 .
- the leader module 306 may further be configured to provide textual information characterizing the relationship between the projected images 22 and real-time images 34 .
- the textual information may include for example identifiers showing current/destination locations, names of buildings, names of landmarks within the viewing area, directional information, etc. An example of textual information is shown in FIG. 4 where the identifier “My Location” is shown overlaid on the projected image 22 (shown as step 1510 ).
- the leader module 306 is thus configured to render markers onto the scene (i.e. visualization representation 18 ), in communication with visualization renderer 112 connecting a physical object (i.e. an object in the real-time image 34 ) to its virtual representation (i.e. a virtual object in the projection image 22 ).
- the markers are overlaid such that they are placed on top of the real-time image 34 and the projection image 22 and extending therebetween.
- the leader module 306 is provided information about the dimension and physical characteristics of one or more objects in the real-time image 34 . Accordingly, in the present embodiment the leader module 306 uses the dimension information to determine the placement of the marker. For example, the marker may be drawn such that it extends from the virtual object to the top surface of the corresponding real-time object (i.e. the marker is drawn to a certain predefined height on the real-time image 34 ).
- the visualization renderer 112 is further configured to visualize and display the real-time image 34 , the projection image 22 and the one or more markers on the display 18 .
- the visualization renderer 112 is configured to display the real-time image 34 on the display 18 .
- the visualization renderer 112 is further configured to superimpose the projection image 22 in an upper portion of the real-time image 34 .
- the visualization renderer 112 may further be configured to determine an area on the real-time image 34 that is less filled or less crowded with real-time objects for overlaying the projected image 22 therein.
- the projection image 22 is overlaid in an upper portion of the real-time image 34 where there are less real-time objects (i.e. the projection image 22 is displayed on top of the sky portion of the real-time image 34 on the display 18 .
- the visualization renderer 112 is further configured to communicate with the leader module 306 and overlay the markers connecting objects within the displayed real-time image 34 to virtual objects in the displayed projected image 22 .
- the real-time image 34 depicts the device 101 viewing area taken from a specific geographical location and is displayed on the visual representation 18 .
- the map projection 22 depicts an elevated view of a ground surface about the device's current geographical location and in accordance with the device's 101 viewing area (i.e. which direction and which angle the device is facing).
- This map projection 22 is overlaid on the real-time image 34 in an upper portion 410 thereof.
- the map projection is overlaid in a manner to minimize blocking objects (i.e. 412 ) in the lower portion 414 .
- FIG. 412 blocking objects
- markers 404 connecting virtual objects (i.e. 406 ) with real-time objects (i.e. 408 ).
- the one or more markers may be shown by colour coding the virtual object in the projected image 22 to the real-time object in image 34 .
- a specific building or landmark may have the same colour in the projection image 22 and the real-time image 34 as a way of showing their relationship.
- the projection images 22 referred to herein may include images such as aerial images and/or satellite images and/or annotated vector graphics depicting a top view of the ground surface about the device 101 geographical location.
- the projection images 22 may also contain other symbolic representations of map projections such as vectors, objects, user-defined/pre-defined/generated by tool 12 annotations, location specific annotation and images as may be understood by a person skilled in the art.
- the projection images 22 may also be transparent, semi-transparent or opaque and may be predefined by a user of the visualization tool 12 .
- FIG. 10 illustrates an example screen shot of the visual representation 18 where the projection images 22 uses map symbols, vectors and annotations to indicate terrains, objects, different elevations and surfaces within the real-time image 34 .
- FIG. 10 illustrates an example screen shot of the visual representation 18 where the projection images 22 uses map symbols, vectors and annotations to indicate terrains, objects, different elevations and surfaces within the real-time image 34 .
- one or more markers 1000 provided by the leader module 306 are used to correlate or associate objects within the real-time image 34 to the projection or virtual image 22 .
- the real-time image 34 is displayed underneath the projection image 22 .
- the projection image 22 is displayed in an upper portion 1004 of visual representation 18 (and overlaid on an upper portion of the real-time image 34 ).
- the projection image 22 is semi-transparent such that the real-time image 34 is still visible.
- the visualization tool 12 determines that the location or viewing angle of the device 101 has been changed (i.e. via real-time event information 32 ).
- the visualization tool 12 extracts an updated projection image 22 based on the updated coordinates of the device 101 and determines updated associations 16 .
- the visualization tool 12 further communicates with visualization renderer 112 to display an updated real-time image 34 , an updated projection image 22 and one or more markers showing the relationships therebetween on the display 18 .
- the display 18 is updated in real-time to provide the user of the device 101 with a live map.
- the visual interface 202 allows a user to provide external data sources (via user interface 108 ) to include non-geographical features such as colleagues, enemies, events or other physical entities on the display 18 .
- the user interface 108 is configured to allow communication with the tool 12 for a user to add objects and to manipulate the virtual projection image 22 .
- the user interface 108 is further configured to interact with a user to allow placement and modification of at least one of the one or more markers, annotations, vectors, and symbols.
- each of the one or more markers, annotations, vectors and symbols may be configured for defining a relationship between the map projection and the real-time image such as to still allow visibility of the real-time image and the map projection.
- the visualization tool 12 further comprises a transform module 308 .
- the transform module 308 is configured for receiving one or more projection images 22 and for applying a transform thereto to obtain a transformed map projection image 30 .
- the transform module 308 is configured to transform the projection image 22 such as to allow clearer visibility and usability of the projection image 22 for a user.
- the transform applied may depend upon the use of the image 22 . For example, in some cases it may be desirable to have better viewing of objects located closer to the user while in other cases it may be more important to have emphasis on objects located farther from the user.
- the origin of the marker (i.e. 404 ) is computed by applying the surface deformation transform to the geo-coordinates of the object's physical location (geographical location of device 101 provided by event information 32 ).
- the transformed point is extruded downward with a vector annotation (i.e. as seen in FIG. 4 ), appearing to stop at the plane of the physical surface (i.e. extending from the projection image 22 to the real-time image 34 ).
- This stopping point on the physical surface is a perspective function of the distance from the device 101 to the physical location of the real-time object. If the 3D geometry of the physical real-time object is known, this can be used to extend the marker line from the sky to the top of the object, instead of extending all the way down to the surface plane of the real-time image 34 .
- the surface of the projection image 22 (also referred to as a reflected map surface in the figures) is a flat, planar surface such that the angle between a real-time object 502 and a virtual object 504 located above the real-time object is 90 degrees. It is noted that the term planar is also used to described a flat surface as will be understood by a person skilled in the art.
- the virtual object 504 is positioned directly above the real-time object 502 (i.e. a one-to-one vertical match).
- the present embodiment is useful for viewing a reflection of an object (i.e. a virtual object) that is located near the observer (i.e. user of the device 101 ). Correspondingly, in this case, it can be difficult to view virtual objects in the projection image 22 that are far from the device 101 location.
- the flat surface of the projection image 22 refers to the surface along which the projection image is reflected or the plane where the projection image 22 lies.
- the transform module 308 applies a slope to the plane of the projection image 22 to provide a desired viewing angle.
- a user of the device 101 standing a predefined location 602 can view a virtual object 604 at a pre-defined angle (i.e. the viewing angle being less than 90 degrees) such as to aid in viewing virtual objects that are positioned directly above where a user of the device 101 is standing on the ground surface.
- the angle of the surface of the map projection 22 allows easier reading of distant objects in the virtual plane (i.e. seen as virtual objects in the projected image 22 ). As illustrated in FIG.
- the projected image 22 may also be shifted a pre-defined distance 608 relative to the device 101 current location on the ground surface 606 . Further, the virtual objects displayed in the projection image 22 may be magnified relative to real-time objects in the real-time image 34 .
- the transform module 308 applies a curve or parabolic transform to the projection image 22 such that the plane or base of projection image 22 is curved in at least one portion.
- This curve transform allows a user of the tool 12 to view locations near the device 101 with more clarity as it appears to stretch out at least a first portion 702 of the projected image 22 which is located close to the device 101 (i.e. location of device 101 on ground shown as 706 ).
- a second portion 706 of the projected image 22 remains flat. The second portion 706 being located farthest from the geographical location of the device 706 .
- the height of the projection surface varies as function of the distance to the observer.
- the transform module 308 applies a spherical transform to the projection image 22 .
- a first portion of the projected image 802 located close to the position of the device 101 i.e. 806
- a second portion of the projected image 804 located away from the device 101 is compressed.
- virtual objects located near a user of the device 101 are magnified for emphasis and clarity while virtual objects located away from the user are compressed.
- the reflection of objects near the observer is located at an angle in the map projection 22 plane.
- the transform module 308 applies a multi-axis curve or surface of revolution transform to the projected image 22 .
- increased emphasis is placed on virtual objects located closest to the device's present location. That is, as shown in FIG. 9B , the virtual objects 904 located closest to the device's 101 geographical location 902 are magnified in the visual representation 18 for better viewing.
- the position of the observer (user of the device 101 ) on the real-time image 34 is shown at element 901 .
- the corresponding reflected image of the observer in the projection image 22 is seen as above the observer and along the axis of the surface of revolution curve as element 903 .
- FIG. 9C provides exemplary functions for defining the surface of revolution curve.
- the surface of revolution curve is a multi-axis curve that is manipulated in the X, Y and Z axes.
- the equations in FIG. 9C define how the surface of the projection image 22 is deformed. For example, thinking of the surface as a mesh, each point on the mesh (i.e. each virtual object having coordinates on the projection image 22 ) is translated or transformed from an initial position (X,Y,Z) to its new position (X′, Y′, Z′). The original points prior to transformation are shown as “OP” while the transformed points are shown as “DP”. Relative to a user of the device 101 viewpoint, the surface appears to curve upwards in the (X,Z) plane and also outward like a ripple effect in the (X,Y) plane.
- the transform module 308 applies scaling and shifting to the flat surface of the projected image 22 to obtain a transformed image that is shifted, and scaled relative to the original projected image 22 .
- the transformed image provides magnification of a certain portion of the viewing area of the user (i.e. the transformed image is a reflection of a certain portion of the real-time image such the virtual objects appear magnified relative to their corresponding real-time objects).
- the shifting allows the user to view the projected image 22 at a more desirable angle. In this case, the angle between the user of the device 1102 and the reflection of the location of the user in the transformed projection image 1100 is reduced to less than 45 degrees.
- the orientation and positioning of the map projection image 22 is linked to the orientation of the user so that the projection image 22 is updated as the user changes orientation.
- the transform module 308 applies a variable width scaling to the projected image 22 .
- the projected image 22 is scaled along the x and y axes based on the distance to the user/device 101 location 1202 . That is objects located closer to the device 101 (i.e. shown as area 1204 ) are magnified while objects located farther from the device 101 (i.e. shown as area 1206 ) are compressed.
- This type of transform provides emphasis on local areas near the user of the device 101 and keeps the observer position 1208 on the image projection 22 in front of the user location to allow the user of the device improved readability of the image projection 22 .
- the transform module 308 applies a variable distance emphasis to the projection image 22 . That is, pre-defined distances and regions are magnified for better readability of beyond the horizon geography. That is, by magnifying virtual objects at pre-defined distances (i.e. magnifying virtual objects in the projected image 22 ), the real-time objects located beyond the horizon are brought into view on the projection image 22 .
- the transform module 308 provides a reflected map where the reflection is based on part of the real-time image 34 . That is, the projection image 22 provides a reflection of a certain pre-defined area of the real-time image 34 . In this way, the projection image 22 provides partial coverage of the real-time image 34 , thereby reducing clutter in the map projection image 22 and reducing the burden on the CPU of the device 101 .
- the transform module 308 applies a quadratic function transform to the projection image 22 .
- the quadratic function results in a transformed projection image 22 having of one or more curved portions such as a parabola.
- the quadratic function transform in FIG. 16 may be adaptive depending upon the observer (i.e. user of the device 101 ) viewing angle. That is, as the device 101 viewing area (i.e. dependent upon the viewing angle and direction facing) of the device changes, the quadratic function and thus the shape of the parabola is updated to allow clearer visibility and usability of the projection image 22 .
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- General Physics & Mathematics (AREA)
- Geometry (AREA)
- Computer Graphics (AREA)
- Remote Sensing (AREA)
- Mathematical Physics (AREA)
- Business, Economics & Management (AREA)
- Educational Administration (AREA)
- Educational Technology (AREA)
- Computer Hardware Design (AREA)
- General Engineering & Computer Science (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
- This application relates to an interactive visual presentation of a map projection on a user interface. Specifically, the application relates to a system and method for displaying a map projection on a user interface for depicting a reflected view of real-time objects displayed within a user/device's viewing area.
- Currently, there exists many navigation systems that provide information about a user's surroundings and geographical information. Some GPS navigation systems appear as an animated 2D image on a specific device such as a BlackBerry®, iPhone®, cell phone or a specific GPS device such as Garmin®. These 2D systems are designed to provide navigation information on an animated street map view. The GPS systems can provide a street view, a satellite view and a hybrid view of a map. However, it can be distracting for a user to view their surroundings while looking at a secondary device. It can also be disorienting for a user to try to relate the animated map view to the real world images seen by the user and attempt navigation at the same time. As well, these GPS devices/applications provide a limited amount of information about the user's surroundings. For example, they may provide an animated street map view, a compass, and directional information to navigate the user to a desired location.
- There are also some 3D GPS systems that provide a limited amount of navigation information on windshields of vehicles or airplanes. Such navigation systems referred to as heads up displays, project basic directional information for guiding a user regarding their current location and/or destination location. For example, in one case a virtual cable line is projected on a windshield to show the direction which the user should navigate on the highway. Alternatively, a 3D compass is projected onto the window of a fighter plane to show the current position, altitude and bearing of the plane to allow a pilot to know which direction they are facing.
- Generally, augmented reality (AR) is a term for a live view of a physical real-world environment whose elements are merged with virtual computer-generated imagery—creating a mixed reality. The computer-generated images are displayed as a layer over a user's view of the physical world. With this extra information presented to the user, the physical world can be enhanced or augmented beyond the user's normal experience. AR systems are also now in everyday use on mobile devices, such as iPhone® and BlackBerry® devices, where the device's screen is used to show merged live video from its camera with virtual elements placed on the display. Navigation using Augmented Reality methods typically is done using annotations and text descriptions on the live scene to provide information about the physical world. However, this approach occludes the live scene and provides a limited amount of information about the physical world. Further, it is difficult for a user to relate the virtual imagery to the physical world.
- Accordingly, the existing GPS systems present limited amount of information about a user's surroundings and present difficulties for a user to navigate to a desired location while referring to and trying correlate an animated map view on a GPS device screen to the real-world.
- According to one aspect there is provided a navigation system that provides correlation between real-world images and virtual map images and allows improved navigation. The virtual map images include for example, virtual aerial and/or satellite imagery, raster or vector map images. According to the present aspect, the navigation system displays a map projection comprising a reflected virtual view of real-time objects seen within a user/device's viewing area to aid in navigation and improve understanding of the current surroundings.
- According to one aspect there is provided a method for creating and displaying a map projection of a device's real-time viewing area to depict virtual objects providing a reflected view of real-time objects within the device's viewing area, the method comprising: displaying a real-time image of the device's viewing area taken from a geographical location on a display; retrieving the map projection for revealing the reflected view as an elevated view of a ground surface about the device's current geographical location and in accordance with the device's viewing area; superimposing the map projection on the display in an upper portion of the real-time image; and defining one or more markers configured to show a relationship between the map projection and the real-time image, each marker overlaid on the display and configured to connect between the virtual object in the map projection and the corresponding real-time object on the real-time image. According to one aspect, the map projection includes aerial and/or satellite imagery, a raster or a vector map projection.
- According to a further aspect of the invention, there is provided a method and system for transforming the map projection. The map projection displays a reflected view of real-time objects in the real-time plane. The map projection is transformed to a surface having one of a parabolic curve surface, an arcuate surface, a flat surface, an angled planar surface, a surface of revolution curve, a surface shifted relative to the real-time objects displayed, a surface having at least a portion thereof magnified relative to the map projection, a surface having at least a portion thereof compressed relative to the map projection, a surface having a subset of virtual objects provided in the map projection. The selection of the transform being based on pre-defined criteria (i.e. user preferences for providing improved visibility of certain virtual objects or focusing on a specific region of the virtual objects).
- A better understanding of these and other embodiments of the present invention can be obtained with reference to the following drawings and detailed description of the preferred embodiments, in which:
-
FIG. 1A is a block diagram of a data processing system for a visualization tool andFIG. 1B is a block diagram of further details; -
FIG. 2 is a block diagram of the visualization tool and associated components according to one embodiment; -
FIG. 3 is a block diagram of the visualization tool having a transform module according to an alternate embodiment; -
FIG. 4 is a representative screen shot of the visualization tool showing the projection image displayed on a real-time image and relationships therebetween in accordance with an embodiment thereof; -
FIGS. 5A-9C are representative views depicting alternate transforms applied by the transform module of the visualization tool to the projection image according to alternate embodiments; -
FIG. 10 is a representative screen shot of the visualization tool showing the projection image displayed on a real-time image using one or more of a vector map, vectors, annotations and symbols to represent the projected virtual image in accordance with one embodiment; -
FIGS. 11-14 and 16 are representative views depicting alternate transforms applied by the transform module to the projection image according to alternate embodiments; and -
FIG. 15 is a schematic diagram illustrating the process for creating and displaying a map projection according to one embodiment. - Referring to
FIG. 1A , a visualizationdata processing system 100 includes avisualization tool 12 for retrieving and processing a collection ofvirtual image information 14 as input data elements to auser interface 202. Thevirtual image information 14 provides satellite and/or aerial imagery as projection images 22 (also referred to as map projections or virtual images herein). Thevirtual image information 14 further providesgeographical information 24 associated with theprojection images 22. Thegeographical information 24 can include, for example, location information associated with the projectedimages 22, names of locations within the projection images, or other characterizing information, physical characteristics of objects within theprojection images 22. Thegeographical information 24 can include any information that may be used for example, in graphical information systems (GIS). Thevisualization tool 12 is further configured for processing user/device events 109. The user/device events 109 comprise real-time images 34 that are viewed/captured by a device 101 (seeFIG. 1B ) as the location of thedevice 101 changes. The real-time images 34 may also be referred to as images captured within the device's 101 viewing area. The user/device events 109 further comprise real-time event information 32 that defines any temporal and spatial information related to the real-time images 34 viewed or captured at a predefined instance. For example, the real-time event information 32 may provide information about the device's 101 current location, facing direction, distance to one or more pre-defined real-time objects, landmarks within and outside the device's 101 current viewing area, physical characteristics (i.e. dimensions) of objects within the viewing area, distances between objects in the viewing area, distance between a user/device's current positioning to one or more objects in the real-time viewing area. - The
event information 32 may further include information about locations on the real-time image where there is a lack of objects (i.e. determining an upper portion of the real-time image where there are no buildings, trees, people or other objects). This determination of a lack of objects may be used to define a location (i.e. an upper portion of the real-time image) where the projections image may be overlaid without blocking any objects. - The
virtual image information 14 can be combined with a respective set ofassociations 16 which define relationships of the projectedimages 22 and real-time images 34 by thetool 12 to generate an interactivevisual representation 18 on the visual interface (VI) 202. The set ofassociations 16 may be predefined by a user (i.e. analyst) of thetool 12 to define relationships between thevirtual image information 14 and real-time images 34 provided byuser events 109. Alternatively, the set ofassociations 16 may be generated by thetool 12 based on projection geographical information 24 (i.e. location of virtual objects) and real-time event information 32 (i.e. estimated location of real-time objects) related to the real-time images 34 captured. Theassociations 16 may be generated by thetool 12 for example to link one or more objects in the device's 101 viewing area (the objects captured in real-time images 34) to corresponding virtual objects within theprojection images 22. Theassociations 16 are displayed on thevisual representation 18 as markers connected between a virtual object in theprojection image 22 and a corresponding real-time object in the real-time image 34. As will be described, the markers can alternatively be colour coded or otherwise visually emphasized to distinguish associations between real-time objects and virtual objects. As will be described below, theassociations 16 are made by correlating the estimated geographic location of each real-time object (provided by the real-time event information 32) to the geographical location of each virtual object (provided by the projection geographical information 24). The definition ofassociations 16 may be user dependent, semi automated (i.e. defined by thetool 12 but modifiable by a user of thedevice 101 or fully automated (i.e. fully defined by the tool 12). - Management of the
virtual image information 14,tool 12, andassociations 16 are driven by updateduser events 109 of a user (not shown) of thetool 12. Theuser events 109 comprise real-time images 34 captured and updated as the location of the device 101 (FIG. 1B ) changes. Theuser events 109 furthercomprise event information 32 defining location and other GIS information related to each real-time image 34. Alternatively, theuser events 109 may be updated by interaction of a user (not shown) with the user interface 108 (seeFIG. 1B ) during interaction with thevisual representation 18. As will be described, thevisual representation 18 shows connectivity between theprojection images 22 and the real-time images 34. - Referring to
FIG. 1B , thedata processing system 100 of adevice 101 includes the user interface device(s) 108 for interacting with thetool 12, the user interface device(s) 108 being connected to amemory 102 via a BUS 106. Thedevice 101 comprises a computing device and may include for example a laptop or desktop computer, a mobile phone, a Personal Digital Assistant (PDA), virtual reality goggles, monocle, heads-up display systems, virtual reality device other types of computing devices as will be envisaged by a person skilled in the art. The interface device(s) 108 are coupled to aprocessor 104 via the BUS 106, to interact withuser events 109 to monitor or otherwise instruct the operation of thetool 12 via anoperating system 110. The user interface device(s) 108 can include one or more user input devices such as but not limited to a QWERTY keyboard, a keypad, a trackwheel, a stylus, a mouse, a microphone, a digital compass and an accelerometer. Thevisual interface 202 is considered to be a user output device, such as but not limited to a computer screen display, a mobile device display (such as a cell phone screen), goggles having a screen display (such as a virtual reality goggles). If the screen is touch sensitive, then the display can also be used as a user input device as controlled by theprocessor 104. Further, it is recognized that thedata processing system 100 can include a computerreadable storage medium 46 coupled to theprocessor 104 for providing instructions to theprocessor 104 and/or thetool 12. The computerreadable medium 46 can include hardware and/or software such as, by way of example only, magnetic disks, magnetic tape, optically readable medium such as CD/DVD ROMS, and memory cards. In each case, the computerreadable medium 46 may take the form of a small disk, floppy diskette, cassette, hard disk drive, solid-state memory card, or RAM provided in thememory 102. It should be noted that the above listed example computerreadable mediums 46 can be used either alone or in combination.System 100 further comprises anetwork interface 47 comply thesystem 100 for communication with one or more public or private networks searches a LAN and/or the Internet. - Referring again to
FIG. 1B , thetool 12 interacts vialink 116 with a VI manager 112 (also known as a visualization renderer) of thesystem 100 for presenting thevisual representation 18 on thevisual interface 202. Thetool 12 processesvirtual image information 14,associations 16, anduser events 109 from data files or tables 122 of thememory 102. As described above, theassociations 16 may either be user-defined, or provided by the tool 12 (or a combination thereof). If thetool 12 provides theassociations 16, theassociations 16 are determined based on analyzing the location of real-time objects (as provided byevent information 32 for example by determining the distance between the real-time objects and thedevice 101 and determining the location of the real-time object by knowing the current location of thedevice 101. Thetool 12 then processes the information received from the tables 122 for subsequent presentation on thevisual representation 18. It is recognized that thevirtual image information 14,associations 16 and theuser events 109 could be stored in the same or separate tables 122, as desired. Thetool 12 can receive requests for storing, retrieving, amending, or creating thevirtual image information 14,associations 16 via thetool 12 and/or directly via link 120 from theVI manager 112, as driven by theuser events 109 and/or independent operation of thetool 12. Accordingly, thetool 12 andmanager 112 coordinate the processing of data objects 14, association set 16 anduser events 109 with respect to the content of thescreen representation 18 displayed in thevisual interface 202. - As will be understood by a person skilled in the art, the
visualization tool 12 and thevisual interface 202 may exist on separate devices (not shown) such that the process of creating the map projection is performed on a first device and the second device is used to render the map projection and the real-time image on the display. - The task related instructions can comprise code and/or machine readable instructions for implementing predetermined functions/operations including those of an operating system,
tool 12, or other information processing system, for example, in response to command or input provided by a user of thesystem 100. The processor 104 (also referred to as module(s) for specific components of the tool 12) as used herein is a configured device and/or set of machine-readable instructions for performing operations as described by example above. - As used herein, the processor/modules in general may comprise any one or combination of, hardware, firmware, and/or software. The processor/modules acts upon information by manipulating, analyzing, modifying, converting or transmitting information for use by an executable procedure or an information device, and/or by routing the information with respect to an output device. The processor/modules may use or comprise the capabilities of a controller or microprocessor, for example. Accordingly, any of the functionality provided by the systems and process of the accompanying figures may be implemented in hardware, software or a combination of both. Accordingly, the use of a processor/modules as a device and/or as a set of machine readable instructions is hereafter referred to generically as a processor/module for sake of simplicity.
- It will be understood by a person skilled in the art that the
memory 102 storage described herein is the place where data is held in an electromagnetic or optical form for access by a computer processor. In one embodiment, storage means the devices and data connected to the computer through input/output operations such as hard disk and tape systems and other forms of storage not including computer memory and other in-computer storage. In a second embodiment, in a more formal usage, storage is divided into: (1) primary storage, which holds data in memory (sometimes called random access memory or RAM) and other “built-in” devices such as the processor's L1 cache, and (2) secondary storage, which holds data on hard disks, tapes, and other devices requiring input/output operations. Primary storage can be much faster to access than secondary storage because of the proximity of the storage to the processor or because of the nature of the storage devices. On the other hand, secondary storage can hold much more data than primary storage. In addition to RAM, primary storage includes read-only memory (ROM) and L1 and L2 cache memory. In addition to hard disks, secondary storage includes a range of device types and technologies, including diskettes, Zip drives, redundant array of independent disks (RAID) systems, and holographic storage. Devices that hold storage are collectively known as storage media. - A database is a further embodiment of
memory 102 as a collection of information that is organized so that it can easily be accessed, managed, and updated. In one view, databases can be classified according to types of content: bibliographic, full-text, numeric, and images. In computing, databases are sometimes classified according to their organizational approach. As well, a relational database is a tabular database in which data is defined so that it can be reorganized and accessed in a number of different ways. A distributed database is one that can be dispersed or replicated among different points in a network. An object-oriented programming database is one that is congruent with the data defined in object classes and subclasses. - Computer databases typically contain aggregations of data records or files, such as sales transactions, product catalogs and inventories, and customer profiles. Typically, a database manager provides users the capabilities of controlling read/write access, specifying report generation, and analyzing usage. Databases and database managers are prevalent in large mainframe systems, but are also present in smaller distributed workstation and mid-range systems such as the AS/400 and on personal computers. SQL (Structured Query Language) is a standard language for making interactive queries from and updating a database such as IBM's DB2, Microsoft's Access, and database products from Oracle, Sybase, and Computer Associates.
- Memory is a further embodiment of
memory 102 storage as the electronic holding place for instructions and data that the computer's microprocessor can reach quickly. When the computer is in normal operation, its memory usually contains the main parts of the operating system and some or all of the application programs and related data that are being used. Memory is often used as a shorter synonym for random access memory (RAM). This kind of memory is located on one or more microchips that are physically close to the microprocessor in the computer. - Referring to
FIGS. 2 and 15 , shown is one embodiment of thevisualization tool 12. In the present embodiment, thevisualization tool 12 comprises anextraction module 302, anassociations module 304, and aleader module 306. Thevisualization renderer 112 displays a real-time image 34 based ondevice 101 current geographical location, facing direction and viewing angle of the device 101 (step 1502). Theextraction module 302 is configured for extractingprojection images 22 based on user/device events 109 and real-time images 34 viewed by the device 101 (step 1504). As described herein, theprojection images 22 provides a reflected view of objects seen in the real-time plane (real-time images 34). Theprojection images 22 may refer to satellite and/or aerial imagery and/or annotated vector graphics which provide a reflected view of objects in the real-time plane. The reflected view of objects provided byprojection images 22 is taken from an elevated position about the device's 101 current geographical location and in accordance with the direction angle of the viewing area (as provided by event information 32). In one example, the reflectedview 22 is a mirror image of the real-time objects in the real-time image 22 revealing a hidden view/angle of objects seen in the viewing area. - Thus, the user/
device events 109 provideevent information 32 which defines the geographical coordinates of thedevice 101 and other GIS information related to the device's 101 viewing area. Theevent information 32 can provide information defining thedevice 101 current location and facing direction as it captures real-time images 34 within its viewing area. Thus, theevent information 32 can provide for example, the current location of thedevice 101, the direction thedevice 101 is facing, the estimated elevation of thedevice 101 relative to the surrounding ground surface, the estimated distance of thedevice 101 from surrounding objects and/or pre-defined landmarks associated with the real-time images 34. Accordingly, based on thedevice 101 current geographical location and viewing area (i.e. the facing direction of the device 101) provided byevent information 32, the extractingmodule 302 retrieves the corresponding projection images 22 (also referred to as map projections) from thedata store 122. Theprojection images 22 depict an elevated view of a ground surface about thedevice 101 current geographical location and in accordance with thedevice 101 viewing area (i.e. direction thedevice 101 is facing). In this way, theprojection images 22 can comprise aerial and/or satellite imagery and/or annotated vector graphics which provide a reflected top view of the device's 101 viewing area in accordance with the device's current geographical co-ordinates. As will be described, in one embodiment, theprojection images 22 allow a user of thedevice 101 to reveal portions of the real-time objects hidden (i.e. a top surface of the real-time objects not seen by the user) from thedevice 101 viewing area. Thevisualization renderer 112 is configured to display theprojection image 22 on thedisplay 18 and overlaid in an upper portion of the real-time image 34 (step 1506). - Once the
projection images 22 are extracted, theassociations module 304 is configured to associate real-time objects provided in real-time images 34 to virtual objects (also referred to as reflected objects) in projection images 22 (step 1508). Theevent information 32 provides estimated coordinates of one or more pre-defined real-time objects within the real-time image 34. The estimated coordinates of each real-time object may be calculated for example, using thedevice 101 current location and estimated distance to each real-time object (provided by event information 32). The projectiongeographical information 24 contains geographical coordinates of one or more virtual objects provided inprojection images 22. Accordingly, theleader module 306 is configured to communicate with thevisualization renderer 112 to display and overlay one or more markers for showing a relationship between virtual objects in theprojection images 22 and real-time objects in real-time images 34. Each marker is overlaid on the display and configured to connect between a virtual object in theprojection image 22 and a corresponding real-time object in real-time images 34 (step 1508). The relationships between the objects being defined by theassociations module 304. Theleader module 306 may further be configured to provide textual information characterizing the relationship between the projectedimages 22 and real-time images 34. The textual information may include for example identifiers showing current/destination locations, names of buildings, names of landmarks within the viewing area, directional information, etc. An example of textual information is shown inFIG. 4 where the identifier “My Location” is shown overlaid on the projected image 22 (shown as step 1510). - The
leader module 306 is thus configured to render markers onto the scene (i.e. visualization representation 18), in communication withvisualization renderer 112 connecting a physical object (i.e. an object in the real-time image 34) to its virtual representation (i.e. a virtual object in the projection image 22). The markers are overlaid such that they are placed on top of the real-time image 34 and theprojection image 22 and extending therebetween. In one embodiment, theleader module 306 is provided information about the dimension and physical characteristics of one or more objects in the real-time image 34. Accordingly, in the present embodiment theleader module 306 uses the dimension information to determine the placement of the marker. For example, the marker may be drawn such that it extends from the virtual object to the top surface of the corresponding real-time object (i.e. the marker is drawn to a certain predefined height on the real-time image 34). - The
visualization renderer 112 is further configured to visualize and display the real-time image 34, theprojection image 22 and the one or more markers on thedisplay 18. According to one embodiment, thevisualization renderer 112 is configured to display the real-time image 34 on thedisplay 18. Thevisualization renderer 112 is further configured to superimpose theprojection image 22 in an upper portion of the real-time image 34. Thevisualization renderer 112 may further be configured to determine an area on the real-time image 34 that is less filled or less crowded with real-time objects for overlaying the projectedimage 22 therein. - For example, referring to
FIG. 4 , theprojection image 22 is overlaid in an upper portion of the real-time image 34 where there are less real-time objects (i.e. theprojection image 22 is displayed on top of the sky portion of the real-time image 34 on thedisplay 18. - The
visualization renderer 112 is further configured to communicate with theleader module 306 and overlay the markers connecting objects within the displayed real-time image 34 to virtual objects in the displayed projectedimage 22. - Referring again to
FIG. 4 there is illustrated an example screen shot of thevisual representation 18. As seen inFIG. 4 , the real-time image 34 depicts thedevice 101 viewing area taken from a specific geographical location and is displayed on thevisual representation 18. Themap projection 22 depicts an elevated view of a ground surface about the device's current geographical location and in accordance with the device's 101 viewing area (i.e. which direction and which angle the device is facing). Thismap projection 22 is overlaid on the real-time image 34 in anupper portion 410 thereof. As can be seen the map projection is overlaid in a manner to minimize blocking objects (i.e. 412) in thelower portion 414. As can be seen inFIG. 4 , there are one ormore markers 404 connecting virtual objects (i.e. 406) with real-time objects (i.e. 408). There is also atextual marker 402 showing the user's current location. In one embodiment, the one or more markers may be shown by colour coding the virtual object in the projectedimage 22 to the real-time object inimage 34. For example, a specific building or landmark may have the same colour in theprojection image 22 and the real-time image 34 as a way of showing their relationship. - It is noted that the
projection images 22 referred to herein, may include images such as aerial images and/or satellite images and/or annotated vector graphics depicting a top view of the ground surface about thedevice 101 geographical location. Theprojection images 22 may also contain other symbolic representations of map projections such as vectors, objects, user-defined/pre-defined/generated bytool 12 annotations, location specific annotation and images as may be understood by a person skilled in the art. Theprojection images 22 may also be transparent, semi-transparent or opaque and may be predefined by a user of thevisualization tool 12.FIG. 10 illustrates an example screen shot of thevisual representation 18 where theprojection images 22 uses map symbols, vectors and annotations to indicate terrains, objects, different elevations and surfaces within the real-time image 34. InFIG. 10 , one ormore markers 1000 provided by theleader module 306 are used to correlate or associate objects within the real-time image 34 to the projection orvirtual image 22. Referring toFIG. 10 , the real-time image 34 is displayed underneath theprojection image 22. Further, theprojection image 22 is displayed in an upper portion 1004 of visual representation 18 (and overlaid on an upper portion of the real-time image 34). In the present example, theprojection image 22 is semi-transparent such that the real-time image 34 is still visible. - According to the present embodiment, as the user of the
device 101 moves the device 101 (i.e. the viewing screen) location or positioning, thevisualization tool 12 determines that the location or viewing angle of thedevice 101 has been changed (i.e. via real-time event information 32). Thus, thevisualization tool 12 extracts an updatedprojection image 22 based on the updated coordinates of thedevice 101 and determines updatedassociations 16. Thevisualization tool 12 further communicates withvisualization renderer 112 to display an updated real-time image 34, an updatedprojection image 22 and one or more markers showing the relationships therebetween on thedisplay 18. Thus, thedisplay 18 is updated in real-time to provide the user of thedevice 101 with a live map. - According to one embodiment, the
visual interface 202 allows a user to provide external data sources (via user interface 108) to include non-geographical features such as colleagues, enemies, events or other physical entities on thedisplay 18. In one embodiment, theuser interface 108 is configured to allow communication with thetool 12 for a user to add objects and to manipulate thevirtual projection image 22. In a further embodiment, theuser interface 108 is further configured to interact with a user to allow placement and modification of at least one of the one or more markers, annotations, vectors, and symbols. As described herein each of the one or more markers, annotations, vectors and symbols may be configured for defining a relationship between the map projection and the real-time image such as to still allow visibility of the real-time image and the map projection. - Referring to
FIG. 3 , shown is a block diagram illustrating an embodiment of thevisualization tool 12. In the present embodiment, thevisualization tool 12 further comprises atransform module 308. Thetransform module 308 is configured for receiving one ormore projection images 22 and for applying a transform thereto to obtain a transformedmap projection image 30. Thetransform module 308 is configured to transform theprojection image 22 such as to allow clearer visibility and usability of theprojection image 22 for a user. The transform applied may depend upon the use of theimage 22. For example, in some cases it may be desirable to have better viewing of objects located closer to the user while in other cases it may be more important to have emphasis on objects located farther from the user. - Once a transform is applied to the
projection image 22 to obtain a transform image, the origin of the marker (i.e. 404) is computed by applying the surface deformation transform to the geo-coordinates of the object's physical location (geographical location ofdevice 101 provided by event information 32). In one aspect, the transformed point is extruded downward with a vector annotation (i.e. as seen inFIG. 4 ), appearing to stop at the plane of the physical surface (i.e. extending from theprojection image 22 to the real-time image 34). This stopping point on the physical surface is a perspective function of the distance from thedevice 101 to the physical location of the real-time object. If the 3D geometry of the physical real-time object is known, this can be used to extend the marker line from the sky to the top of the object, instead of extending all the way down to the surface plane of the real-time image 34. - Referring to
FIGS. 5A and 5B shown are examples of aprojection image 22 where no transform is applied to theprojection image 22. According to the present embodiment, the surface of the projection image 22 (also referred to as a reflected map surface in the figures) is a flat, planar surface such that the angle between a real-time object 502 and avirtual object 504 located above the real-time object is 90 degrees. It is noted that the term planar is also used to described a flat surface as will be understood by a person skilled in the art. In this case, thevirtual object 504 is positioned directly above the real-time object 502 (i.e. a one-to-one vertical match). The present embodiment is useful for viewing a reflection of an object (i.e. a virtual object) that is located near the observer (i.e. user of the device 101). Correspondingly, in this case, it can be difficult to view virtual objects in theprojection image 22 that are far from thedevice 101 location. - It is further noted that the flat surface of the
projection image 22 refers to the surface along which the projection image is reflected or the plane where theprojection image 22 lies. - As illustrated in
FIGS. 6A and 6B , in an alternate embodiment, thetransform module 308 applies a slope to the plane of theprojection image 22 to provide a desired viewing angle. In this way, a user of thedevice 101 standing apredefined location 602 can view avirtual object 604 at a pre-defined angle (i.e. the viewing angle being less than 90 degrees) such as to aid in viewing virtual objects that are positioned directly above where a user of thedevice 101 is standing on the ground surface. In this case, the angle of the surface of themap projection 22 allows easier reading of distant objects in the virtual plane (i.e. seen as virtual objects in the projected image 22). As illustrated inFIG. 6B , the projectedimage 22 may also be shifted apre-defined distance 608 relative to thedevice 101 current location on theground surface 606. Further, the virtual objects displayed in theprojection image 22 may be magnified relative to real-time objects in the real-time image 34. - Referring to
FIGS. 7A and 7B , in one embodiment, thetransform module 308 applies a curve or parabolic transform to theprojection image 22 such that the plane or base ofprojection image 22 is curved in at least one portion. This curve transform allows a user of thetool 12 to view locations near thedevice 101 with more clarity as it appears to stretch out at least afirst portion 702 of the projectedimage 22 which is located close to the device 101 (i.e. location ofdevice 101 on ground shown as 706). In the present embodiment, asecond portion 706 of the projectedimage 22 remains flat. Thesecond portion 706 being located farthest from the geographical location of thedevice 706. The height of the projection surface varies as function of the distance to the observer. - Referring to
FIG. 8 , in one embodiment, thetransform module 308 applies a spherical transform to theprojection image 22. In the present embodiment, a first portion of the projectedimage 802 located close to the position of the device 101 (i.e. 806) is stretched out and magnified, while a second portion of the projectedimage 804 located away from thedevice 101 is compressed. In this manner, virtual objects located near a user of thedevice 101 are magnified for emphasis and clarity while virtual objects located away from the user are compressed. Further the reflection of objects near the observer is located at an angle in themap projection 22 plane. - Referring to
FIGS. 9A-9C , according to one embodiment, thetransform module 308 applies a multi-axis curve or surface of revolution transform to the projectedimage 22. In the present embodiment, increased emphasis is placed on virtual objects located closest to the device's present location. That is, as shown inFIG. 9B , thevirtual objects 904 located closest to the device's 101geographical location 902 are magnified in thevisual representation 18 for better viewing. - As shown in
FIG. 9A , the position of the observer (user of the device 101) on the real-time image 34 is shown at element 901. The corresponding reflected image of the observer in theprojection image 22 is seen as above the observer and along the axis of the surface of revolution curve as element 903. -
FIG. 9C provides exemplary functions for defining the surface of revolution curve. As can be seen the surface of revolution curve is a multi-axis curve that is manipulated in the X, Y and Z axes. The equations inFIG. 9C define how the surface of theprojection image 22 is deformed. For example, thinking of the surface as a mesh, each point on the mesh (i.e. each virtual object having coordinates on the projection image 22) is translated or transformed from an initial position (X,Y,Z) to its new position (X′, Y′, Z′). The original points prior to transformation are shown as “OP” while the transformed points are shown as “DP”. Relative to a user of thedevice 101 viewpoint, the surface appears to curve upwards in the (X,Z) plane and also outward like a ripple effect in the (X,Y) plane. - Referring to
FIG. 11 , according to one embodiment, thetransform module 308 applies scaling and shifting to the flat surface of the projectedimage 22 to obtain a transformed image that is shifted, and scaled relative to the original projectedimage 22. In this case, the transformed image provides magnification of a certain portion of the viewing area of the user (i.e. the transformed image is a reflection of a certain portion of the real-time image such the virtual objects appear magnified relative to their corresponding real-time objects). Further, the shifting allows the user to view the projectedimage 22 at a more desirable angle. In this case, the angle between the user of thedevice 1102 and the reflection of the location of the user in the transformedprojection image 1100 is reduced to less than 45 degrees. According to the present embodiment, the orientation and positioning of themap projection image 22 is linked to the orientation of the user so that theprojection image 22 is updated as the user changes orientation. - Reflected Map Projection with Variable
Width Scaling Transform 1200 - Referring to
FIG. 12 , according to one embodiment, thetransform module 308 applies a variable width scaling to the projectedimage 22. In this case, the projectedimage 22 is scaled along the x and y axes based on the distance to the user/device 101location 1202. That is objects located closer to the device 101 (i.e. shown as area 1204) are magnified while objects located farther from the device 101 (i.e. shown as area 1206) are compressed. This type of transform provides emphasis on local areas near the user of thedevice 101 and keeps theobserver position 1208 on theimage projection 22 in front of the user location to allow the user of the device improved readability of theimage projection 22. - Reflected Map Projection with
Variable Distance Emphasis 1300 - Referring to
FIG. 13 , according to one embodiment, thetransform module 308 applies a variable distance emphasis to theprojection image 22. That is, pre-defined distances and regions are magnified for better readability of beyond the horizon geography. That is, by magnifying virtual objects at pre-defined distances (i.e. magnifying virtual objects in the projected image 22), the real-time objects located beyond the horizon are brought into view on theprojection image 22. - Referring to
FIG. 14 , according to one embodiment, thetransform module 308 provides a reflected map where the reflection is based on part of the real-time image 34. That is, theprojection image 22 provides a reflection of a certain pre-defined area of the real-time image 34. In this way, theprojection image 22 provides partial coverage of the real-time image 34, thereby reducing clutter in themap projection image 22 and reducing the burden on the CPU of thedevice 101. - Referring to
FIG. 16 , according to one embodiment, thetransform module 308 applies a quadratic function transform to theprojection image 22. The quadratic function results in a transformedprojection image 22 having of one or more curved portions such as a parabola. The quadratic function transform inFIG. 16 may be adaptive depending upon the observer (i.e. user of the device 101) viewing angle. That is, as thedevice 101 viewing area (i.e. dependent upon the viewing angle and direction facing) of the device changes, the quadratic function and thus the shape of the parabola is updated to allow clearer visibility and usability of theprojection image 22. - Although preferred embodiments of the invention have been described herein, it will be understood by those skilled in the art that variations may be made thereto without departing from the spirit of the invention or the scope of the appended claims.
Claims (21)
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/699,545 US8436872B2 (en) | 2010-02-03 | 2010-02-03 | System and method for creating and displaying map projections related to real-time images |
PCT/CA2011/000043 WO2011094839A1 (en) | 2010-02-03 | 2011-01-17 | System and method for creating and displaying map projections related to real-time images |
CN201180018116.2A CN102870147B (en) | 2010-02-03 | 2011-01-17 | System and method for creating and displaying map projections related to real-time images |
EP11739288.6A EP2531990A4 (en) | 2010-02-03 | 2011-01-17 | System and method for creating and displaying map projections related to real-time images |
US13/826,781 US9047699B2 (en) | 2010-02-03 | 2013-03-14 | System and method for creating and displaying map projections related to real-time images |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/699,545 US8436872B2 (en) | 2010-02-03 | 2010-02-03 | System and method for creating and displaying map projections related to real-time images |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/826,781 Continuation US9047699B2 (en) | 2010-02-03 | 2013-03-14 | System and method for creating and displaying map projections related to real-time images |
Publications (2)
Publication Number | Publication Date |
---|---|
US20110188760A1 true US20110188760A1 (en) | 2011-08-04 |
US8436872B2 US8436872B2 (en) | 2013-05-07 |
Family
ID=44341712
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/699,545 Active 2032-01-01 US8436872B2 (en) | 2010-02-03 | 2010-02-03 | System and method for creating and displaying map projections related to real-time images |
US13/826,781 Active 2030-03-04 US9047699B2 (en) | 2010-02-03 | 2013-03-14 | System and method for creating and displaying map projections related to real-time images |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/826,781 Active 2030-03-04 US9047699B2 (en) | 2010-02-03 | 2013-03-14 | System and method for creating and displaying map projections related to real-time images |
Country Status (4)
Country | Link |
---|---|
US (2) | US8436872B2 (en) |
EP (1) | EP2531990A4 (en) |
CN (1) | CN102870147B (en) |
WO (1) | WO2011094839A1 (en) |
Cited By (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120001939A1 (en) * | 2010-06-30 | 2012-01-05 | Nokia Corporation | Methods, apparatuses and computer program products for automatically generating suggested information layers in augmented reality |
US20120056899A1 (en) * | 2010-09-08 | 2012-03-08 | Matei Stroila | Generating a multi-layered geographic image and the use thereof |
US20130044128A1 (en) * | 2011-08-17 | 2013-02-21 | James C. Liu | Context adaptive user interface for augmented reality display |
JP2013218459A (en) * | 2012-04-06 | 2013-10-24 | Sony Corp | Information processor, information processing method and information processing system |
US20140055401A1 (en) * | 2010-11-05 | 2014-02-27 | Bluespace Corporation | Method and apparatus for controlling contents in realtime fashion |
US20140188877A1 (en) * | 2012-12-21 | 2014-07-03 | Orange | Method for Managing a System of Geographical Information Adapted for Use With at Least One Pointing Device, with Creation of Purely Virtual Digital Objects |
US8823744B2 (en) | 2012-03-23 | 2014-09-02 | Bluebeam Software, Inc. | Method for indicating annotations associated with a particular display view of a three-dimensional model independent of any display view |
US9069440B2 (en) | 2012-06-05 | 2015-06-30 | Apple Inc. | Method, system and apparatus for providing a three-dimensional transition animation for a map view change |
US20150199849A1 (en) * | 2014-01-13 | 2015-07-16 | Lg Electronics Inc. | Head mounted display and method of controlling thereof |
US9153195B2 (en) | 2011-08-17 | 2015-10-06 | Microsoft Technology Licensing, Llc | Providing contextual personal information by a mixed reality device |
US20150332511A1 (en) * | 2014-05-13 | 2015-11-19 | Spaceview Inc. | Method for interactive catalog for 3d objects within the 2d environment |
US9213405B2 (en) | 2010-12-16 | 2015-12-15 | Microsoft Technology Licensing, Llc | Comprehension and intent-based content for augmented reality displays |
US9639968B2 (en) * | 2014-02-18 | 2017-05-02 | Harman International Industries, Inc. | Generating an augmented view of a location of interest |
US9648075B1 (en) * | 2012-12-18 | 2017-05-09 | Google Inc. | Systems and methods for providing an event map |
CN107635131A (en) * | 2017-09-01 | 2018-01-26 | 北京雷石天地电子技术有限公司 | Method and system for realizing virtual reality |
US10001898B1 (en) | 2011-07-12 | 2018-06-19 | Domo, Inc. | Automated provisioning of relational information for a summary data visualization |
US10026226B1 (en) * | 2014-06-10 | 2018-07-17 | Ripple Inc | Rendering an augmented reality object |
US10074266B2 (en) | 2012-12-21 | 2018-09-11 | Orange | Method for managing a system of geographical information adapted for use with at least one pointing device, with creation of associations between digital objects |
US10331864B1 (en) | 2010-12-27 | 2019-06-25 | Mlb Advanced Media, L.P. | Watermarking systems and methods |
CN110075519A (en) * | 2019-05-06 | 2019-08-02 | 网易(杭州)网络有限公司 | Information processing method and device, storage medium and electronic equipment in virtual reality |
US10474352B1 (en) | 2011-07-12 | 2019-11-12 | Domo, Inc. | Dynamic expansion of data visualizations |
US10552978B2 (en) | 2017-06-27 | 2020-02-04 | International Business Machines Corporation | Dynamic image and image marker tracking |
US20200117010A1 (en) * | 2018-10-16 | 2020-04-16 | Hyundai Motor Company | Method for correcting image distortion in a hud system |
US10724871B2 (en) * | 2016-11-03 | 2020-07-28 | The Charles Stark Draper Laboratory, Inc. | Camera-based heading-hold navigation |
US10726624B2 (en) | 2011-07-12 | 2020-07-28 | Domo, Inc. | Automatic creation of drill paths |
CN111625210A (en) * | 2019-02-27 | 2020-09-04 | 杭州海康威视系统技术有限公司 | Large screen control method, device and equipment |
US10896528B2 (en) | 2016-12-20 | 2021-01-19 | Advanced New Technologies Co., Ltd. | Method and apparatus for generating and updating tiles maps in virtual maps |
US10930038B2 (en) | 2014-06-10 | 2021-02-23 | Lab Of Misfits Ar, Inc. | Dynamic location based digital element |
US20210118238A1 (en) * | 2013-03-01 | 2021-04-22 | Apple Inc. | Registration between actual mobile device position and environmental model |
US11127210B2 (en) | 2011-08-24 | 2021-09-21 | Microsoft Technology Licensing, Llc | Touch and social cues as inputs into a computer |
CN113744335A (en) * | 2021-08-24 | 2021-12-03 | 北京体育大学 | Sports guiding method, system and storage medium based on field mark |
US20220028176A1 (en) * | 2017-08-11 | 2022-01-27 | Objectvideo Labs, Llc | Enhancing monitoring system with augmented reality |
US11410394B2 (en) | 2020-11-04 | 2022-08-09 | West Texas Technology Partners, Inc. | Method for interactive catalog for 3D objects within the 2D environment |
US12008697B2 (en) | 2014-06-10 | 2024-06-11 | Ripple, Inc. Of Delaware | Dynamic location based digital element |
Families Citing this family (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8902260B2 (en) * | 2010-09-01 | 2014-12-02 | Google Inc. | Simplified creation of customized maps |
WO2012154262A2 (en) | 2011-02-21 | 2012-11-15 | TransRobotics, Inc. | System and method for sensing distance and/or movement |
GB2489685B (en) * | 2011-03-31 | 2017-01-25 | Geovs Ltd | A Display System |
US20140005886A1 (en) * | 2012-06-29 | 2014-01-02 | Microsoft Corporation | Controlling automotive functionality using internal- and external-facing sensors |
US10206610B2 (en) | 2012-10-05 | 2019-02-19 | TransRobotics, Inc. | Systems and methods for high resolution distance sensing and applications |
CN104102678B (en) * | 2013-04-15 | 2018-06-05 | 腾讯科技(深圳)有限公司 | The implementation method and realization device of augmented reality |
US9996974B2 (en) | 2013-08-30 | 2018-06-12 | Qualcomm Incorporated | Method and apparatus for representing a physical scene |
US10089863B2 (en) * | 2013-08-30 | 2018-10-02 | Komatsu Ltd. | Management system and management method for mining machine |
US10074182B2 (en) * | 2013-11-14 | 2018-09-11 | Microsoft Technology Licensing, Llc | Presenting markup in a scene using depth fading |
US9351118B2 (en) | 2014-04-11 | 2016-05-24 | Keith Crutchfield | Apparatus, systems and methods for visually connecting people |
CN105321144A (en) * | 2015-07-02 | 2016-02-10 | 深圳市雅都软件股份有限公司 | Map dynamic projection display system and method |
JP2018535568A (en) | 2015-09-02 | 2018-11-29 | サムロール エルエルシー | Camera system and method for registering images and presenting a series of aligned images |
US10542060B2 (en) | 2016-09-19 | 2020-01-21 | Ebay Inc. | Interactive real-time visualization system for large-scale streaming data |
US10360723B2 (en) * | 2017-08-08 | 2019-07-23 | Christie Digital Systems Usa, Inc. | Method and apparatus for automated projection mapping previsualization |
CN108088422B (en) * | 2018-01-24 | 2020-12-15 | 成都纵横自动化技术股份有限公司 | Method for determining real overlapping rate |
CA3113406A1 (en) | 2018-09-18 | 2020-03-26 | TransRobotics, Inc. | Technologies for acting based on object detection |
EP3719532B1 (en) | 2019-04-04 | 2022-12-28 | Transrobotics, Inc. | Technologies for acting based on object tracking |
US20220301264A1 (en) * | 2021-03-22 | 2022-09-22 | Apple Inc. | Devices, methods, and graphical user interfaces for maps |
CN113239952B (en) * | 2021-03-30 | 2023-03-24 | 西北工业大学 | Aerial image geographical positioning method based on spatial scale attention mechanism and vector map |
JP7319635B2 (en) * | 2021-11-02 | 2023-08-02 | グリー株式会社 | Computer program, virtual space display device, and virtual space display method |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040107072A1 (en) * | 2002-12-03 | 2004-06-03 | Arne Dietrich | Ins-based user orientation and navigation |
US20040169653A1 (en) * | 1995-04-20 | 2004-09-02 | Yoshinori Endo | Bird's-eye view forming method, map display apparatus and navigation system |
US20050149262A1 (en) * | 2003-05-16 | 2005-07-07 | Wataru Oikubo | Navigation apparatus |
US20060041375A1 (en) * | 2004-08-19 | 2006-02-23 | Geographic Data Technology, Inc. | Automated georeferencing of digitized map images |
US20060174211A1 (en) * | 1999-06-09 | 2006-08-03 | Microsoft Corporation | Methods, apparatus and data structures for providing a user interface which facilitates decision making |
US20080186255A1 (en) * | 2006-12-07 | 2008-08-07 | Cohen Philip R | Systems and methods for data annotation, recordation, and communication |
US20090028384A1 (en) * | 2005-04-18 | 2009-01-29 | Alexander Bovyrin | Three-dimensional road map estimation from video sequences by tracking pedestrians |
US20090171580A1 (en) * | 2007-12-26 | 2009-07-02 | Sony Corporation | Navigation apparatus and navigation apparatus display method |
US20090186629A1 (en) * | 2008-01-17 | 2009-07-23 | At&T Mobility Ii Llc | Caller Identification with Caller Geographical Location |
US20100328344A1 (en) * | 2009-06-25 | 2010-12-30 | Nokia Corporation | Method and apparatus for an augmented reality user interface |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7563748B2 (en) | 2003-06-23 | 2009-07-21 | Cognis Ip Management Gmbh | Alcohol alkoxylate carriers for pesticide active ingredients |
WO2005057133A1 (en) | 2003-11-25 | 2005-06-23 | Honeywell International, Inc. | Perspective vertical situation display system and method |
US7843451B2 (en) * | 2007-05-25 | 2010-11-30 | Google Inc. | Efficient rendering of panoramic images, and applications thereof |
US8237791B2 (en) * | 2008-03-19 | 2012-08-07 | Microsoft Corporation | Visualizing camera feeds on a map |
-
2010
- 2010-02-03 US US12/699,545 patent/US8436872B2/en active Active
-
2011
- 2011-01-17 WO PCT/CA2011/000043 patent/WO2011094839A1/en active Application Filing
- 2011-01-17 CN CN201180018116.2A patent/CN102870147B/en active Active
- 2011-01-17 EP EP11739288.6A patent/EP2531990A4/en not_active Withdrawn
-
2013
- 2013-03-14 US US13/826,781 patent/US9047699B2/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040169653A1 (en) * | 1995-04-20 | 2004-09-02 | Yoshinori Endo | Bird's-eye view forming method, map display apparatus and navigation system |
US20060174211A1 (en) * | 1999-06-09 | 2006-08-03 | Microsoft Corporation | Methods, apparatus and data structures for providing a user interface which facilitates decision making |
US20040107072A1 (en) * | 2002-12-03 | 2004-06-03 | Arne Dietrich | Ins-based user orientation and navigation |
US20050149262A1 (en) * | 2003-05-16 | 2005-07-07 | Wataru Oikubo | Navigation apparatus |
US20060041375A1 (en) * | 2004-08-19 | 2006-02-23 | Geographic Data Technology, Inc. | Automated georeferencing of digitized map images |
US20090028384A1 (en) * | 2005-04-18 | 2009-01-29 | Alexander Bovyrin | Three-dimensional road map estimation from video sequences by tracking pedestrians |
US20080186255A1 (en) * | 2006-12-07 | 2008-08-07 | Cohen Philip R | Systems and methods for data annotation, recordation, and communication |
US20090171580A1 (en) * | 2007-12-26 | 2009-07-02 | Sony Corporation | Navigation apparatus and navigation apparatus display method |
US20090186629A1 (en) * | 2008-01-17 | 2009-07-23 | At&T Mobility Ii Llc | Caller Identification with Caller Geographical Location |
US20100328344A1 (en) * | 2009-06-25 | 2010-12-30 | Nokia Corporation | Method and apparatus for an augmented reality user interface |
Cited By (58)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9910866B2 (en) * | 2010-06-30 | 2018-03-06 | Nokia Technologies Oy | Methods, apparatuses and computer program products for automatically generating suggested information layers in augmented reality |
US20120001939A1 (en) * | 2010-06-30 | 2012-01-05 | Nokia Corporation | Methods, apparatuses and computer program products for automatically generating suggested information layers in augmented reality |
US20120056899A1 (en) * | 2010-09-08 | 2012-03-08 | Matei Stroila | Generating a multi-layered geographic image and the use thereof |
US8723886B2 (en) * | 2010-09-08 | 2014-05-13 | Navteq B.V. | Generating a multi-layered geographic image and the use thereof |
US9508184B2 (en) | 2010-09-08 | 2016-11-29 | Here Global B.V. | Generating a multi-layered geographic image and the use thereof |
US20140055401A1 (en) * | 2010-11-05 | 2014-02-27 | Bluespace Corporation | Method and apparatus for controlling contents in realtime fashion |
US9213405B2 (en) | 2010-12-16 | 2015-12-15 | Microsoft Technology Licensing, Llc | Comprehension and intent-based content for augmented reality displays |
US10331864B1 (en) | 2010-12-27 | 2019-06-25 | Mlb Advanced Media, L.P. | Watermarking systems and methods |
US10001898B1 (en) | 2011-07-12 | 2018-06-19 | Domo, Inc. | Automated provisioning of relational information for a summary data visualization |
US10726624B2 (en) | 2011-07-12 | 2020-07-28 | Domo, Inc. | Automatic creation of drill paths |
US10474352B1 (en) | 2011-07-12 | 2019-11-12 | Domo, Inc. | Dynamic expansion of data visualizations |
US10223832B2 (en) | 2011-08-17 | 2019-03-05 | Microsoft Technology Licensing, Llc | Providing location occupancy analysis via a mixed reality device |
US9153195B2 (en) | 2011-08-17 | 2015-10-06 | Microsoft Technology Licensing, Llc | Providing contextual personal information by a mixed reality device |
US10019962B2 (en) * | 2011-08-17 | 2018-07-10 | Microsoft Technology Licensing, Llc | Context adaptive user interface for augmented reality display |
US20130044128A1 (en) * | 2011-08-17 | 2013-02-21 | James C. Liu | Context adaptive user interface for augmented reality display |
US11127210B2 (en) | 2011-08-24 | 2021-09-21 | Microsoft Technology Licensing, Llc | Touch and social cues as inputs into a computer |
WO2013141947A3 (en) * | 2012-03-23 | 2014-12-18 | Bluebeam Software, Inc. | Method for indicating annotations associated with a particular display view of a three-dimensional model |
US8823744B2 (en) | 2012-03-23 | 2014-09-02 | Bluebeam Software, Inc. | Method for indicating annotations associated with a particular display view of a three-dimensional model independent of any display view |
US10140771B2 (en) | 2012-04-06 | 2018-11-27 | Sony Corporation | Information processing apparatus for displaying an image such that the image is superimposed on real scenery |
US9685002B2 (en) | 2012-04-06 | 2017-06-20 | Sony Corporation | Information processing apparatus and information processing system having a marker detecting unit and an extracting unit, and information processing method by using the same |
JP2013218459A (en) * | 2012-04-06 | 2013-10-24 | Sony Corp | Information processor, information processing method and information processing system |
US9069440B2 (en) | 2012-06-05 | 2015-06-30 | Apple Inc. | Method, system and apparatus for providing a three-dimensional transition animation for a map view change |
US9648075B1 (en) * | 2012-12-18 | 2017-05-09 | Google Inc. | Systems and methods for providing an event map |
US10074266B2 (en) | 2012-12-21 | 2018-09-11 | Orange | Method for managing a system of geographical information adapted for use with at least one pointing device, with creation of associations between digital objects |
US20140188877A1 (en) * | 2012-12-21 | 2014-07-03 | Orange | Method for Managing a System of Geographical Information Adapted for Use With at Least One Pointing Device, with Creation of Purely Virtual Digital Objects |
US11532136B2 (en) * | 2013-03-01 | 2022-12-20 | Apple Inc. | Registration between actual mobile device position and environmental model |
US20210118238A1 (en) * | 2013-03-01 | 2021-04-22 | Apple Inc. | Registration between actual mobile device position and environmental model |
US9639986B2 (en) * | 2014-01-13 | 2017-05-02 | Lg Electronics Inc. | Head mounted display and method of controlling thereof |
US20150199849A1 (en) * | 2014-01-13 | 2015-07-16 | Lg Electronics Inc. | Head mounted display and method of controlling thereof |
US9639968B2 (en) * | 2014-02-18 | 2017-05-02 | Harman International Industries, Inc. | Generating an augmented view of a location of interest |
US20150332511A1 (en) * | 2014-05-13 | 2015-11-19 | Spaceview Inc. | Method for interactive catalog for 3d objects within the 2d environment |
US11544418B2 (en) | 2014-05-13 | 2023-01-03 | West Texas Technology Partners, Llc | Method for replacing 3D objects in 2D environment |
US9971853B2 (en) | 2014-05-13 | 2018-05-15 | Atheer, Inc. | Method for replacing 3D objects in 2D environment |
US9977844B2 (en) | 2014-05-13 | 2018-05-22 | Atheer, Inc. | Method for providing a projection to align 3D objects in 2D environment |
US10002208B2 (en) * | 2014-05-13 | 2018-06-19 | Atheer, Inc. | Method for interactive catalog for 3D objects within the 2D environment |
US10860749B2 (en) | 2014-05-13 | 2020-12-08 | Atheer, Inc. | Method for interactive catalog for 3D objects within the 2D environment |
US10635757B2 (en) | 2014-05-13 | 2020-04-28 | Atheer, Inc. | Method for replacing 3D objects in 2D environment |
US11403797B2 (en) | 2014-06-10 | 2022-08-02 | Ripple, Inc. Of Delaware | Dynamic location based digital element |
US10026226B1 (en) * | 2014-06-10 | 2018-07-17 | Ripple Inc | Rendering an augmented reality object |
US12154233B2 (en) | 2014-06-10 | 2024-11-26 | Ripple, Inc. Of Delaware | Audio content of a digital object associated with a geographical location |
US12008697B2 (en) | 2014-06-10 | 2024-06-11 | Ripple, Inc. Of Delaware | Dynamic location based digital element |
US10930038B2 (en) | 2014-06-10 | 2021-02-23 | Lab Of Misfits Ar, Inc. | Dynamic location based digital element |
US11069138B2 (en) | 2014-06-10 | 2021-07-20 | Ripple, Inc. Of Delaware | Audio content of a digital object associated with a geographical location |
US11532140B2 (en) * | 2014-06-10 | 2022-12-20 | Ripple, Inc. Of Delaware | Audio content of a digital object associated with a geographical location |
US10724871B2 (en) * | 2016-11-03 | 2020-07-28 | The Charles Stark Draper Laboratory, Inc. | Camera-based heading-hold navigation |
US10896528B2 (en) | 2016-12-20 | 2021-01-19 | Advanced New Technologies Co., Ltd. | Method and apparatus for generating and updating tiles maps in virtual maps |
US10552978B2 (en) | 2017-06-27 | 2020-02-04 | International Business Machines Corporation | Dynamic image and image marker tracking |
US11983827B2 (en) * | 2017-08-11 | 2024-05-14 | Objectvideo Labs, Llc | Enhancing monitoring system with augmented reality |
US20220028176A1 (en) * | 2017-08-11 | 2022-01-27 | Objectvideo Labs, Llc | Enhancing monitoring system with augmented reality |
CN107635131A (en) * | 2017-09-01 | 2018-01-26 | 北京雷石天地电子技术有限公司 | Method and system for realizing virtual reality |
US20200117010A1 (en) * | 2018-10-16 | 2020-04-16 | Hyundai Motor Company | Method for correcting image distortion in a hud system |
KR102518661B1 (en) * | 2018-10-16 | 2023-04-06 | 현대자동차주식회사 | Method for correction image distortion of hud system |
US10996478B2 (en) * | 2018-10-16 | 2021-05-04 | Hyundai Motor Company | Method for correcting image distortion in a HUD system |
KR20200042754A (en) * | 2018-10-16 | 2020-04-24 | 현대자동차주식회사 | Method for correction image distortion of hud system |
CN111625210A (en) * | 2019-02-27 | 2020-09-04 | 杭州海康威视系统技术有限公司 | Large screen control method, device and equipment |
CN110075519A (en) * | 2019-05-06 | 2019-08-02 | 网易(杭州)网络有限公司 | Information processing method and device, storage medium and electronic equipment in virtual reality |
US11410394B2 (en) | 2020-11-04 | 2022-08-09 | West Texas Technology Partners, Inc. | Method for interactive catalog for 3D objects within the 2D environment |
CN113744335A (en) * | 2021-08-24 | 2021-12-03 | 北京体育大学 | Sports guiding method, system and storage medium based on field mark |
Also Published As
Publication number | Publication date |
---|---|
WO2011094839A1 (en) | 2011-08-11 |
EP2531990A1 (en) | 2012-12-12 |
EP2531990A4 (en) | 2014-06-18 |
CN102870147B (en) | 2015-02-18 |
US9047699B2 (en) | 2015-06-02 |
US20140104315A1 (en) | 2014-04-17 |
US8436872B2 (en) | 2013-05-07 |
CN102870147A (en) | 2013-01-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9047699B2 (en) | System and method for creating and displaying map projections related to real-time images | |
Zollmann et al. | Visualization techniques in augmented reality: A taxonomy, methods and patterns | |
US20240290049A1 (en) | Displaying Content in an Augmented Reality System | |
US7995078B2 (en) | Compound lenses for multi-source data presentation | |
US7983473B2 (en) | Transparency adjustment of a presentation | |
US7109998B2 (en) | Stationary semantic zooming | |
US11188739B2 (en) | Processing uncertain content in a computer graphics system | |
US9465513B2 (en) | Visual representation of map navigation history | |
US20070064018A1 (en) | Detail-in-context lenses for online maps | |
US8350872B2 (en) | Graphical user interfaces and occlusion prevention for fisheye lenses with line segment foci | |
US10950040B2 (en) | Labeling for three-dimensional occluded shapes | |
US20070165050A1 (en) | Method and system for geographically-based and time-based online advertising | |
Sandor et al. | Egocentric space-distorting visualizations for rapid environment exploration in mobile mixed reality | |
Maass et al. | Dynamic annotation of interactive environments using object-integrated billboards | |
Zhang et al. | Dynamic labeling management in virtual and augmented environments | |
JP4137923B2 (en) | Image display method and apparatus | |
Maass et al. | Use of depth cues for the annotation of 3d geo-virtual environments | |
Mota et al. | Spatially-Embedded Lens Visualization: A Design Space | |
Agrawal | Augmented Reality, an Emerging Technology and its View Management Problem. | |
Wang et al. | A 3-D Terrain Visualization Database for Highway Information Management | |
Bell et al. | Maintaining Visibility Constraints for View Management in 3D User Interfaces | |
POI | Egocentric Space-Distorting Visualizations for Rapid Environment Exploration in Mobile Mixed Reality |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: OCULUS INFO INC., CANADA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WRIGHT, WILLIAM;KAPLER, THOMAS;SENIOR, ROBIN;REEL/FRAME:024034/0131 Effective date: 20100129 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: UNCHARTED SOFTWARE INC., CANADA Free format text: CHANGE OF NAME;ASSIGNOR:OCULUS INFO INC.;REEL/FRAME:035638/0515 Effective date: 20141218 |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |