USRE45264E1 - Methods and apparatus for generating three-dimensional image data models - Google Patents
Methods and apparatus for generating three-dimensional image data models Download PDFInfo
- Publication number
- USRE45264E1 USRE45264E1 US13/861,336 US201313861336A USRE45264E US RE45264 E1 USRE45264 E1 US RE45264E1 US 201313861336 A US201313861336 A US 201313861336A US RE45264 E USRE45264 E US RE45264E
- Authority
- US
- United States
- Prior art keywords
- image data
- dimensional
- continuum
- user
- software
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related, expires
Links
- 238000000034 method Methods 0.000 title claims abstract description 24
- 238000013499 data model Methods 0.000 title 1
- 239000002131 composite material Substances 0.000 claims description 35
- 230000033001 locomotion Effects 0.000 claims description 21
- 230000002452 interceptive effect Effects 0.000 claims description 9
- 230000001413 cellular effect Effects 0.000 claims description 3
- 238000013500 data storage Methods 0.000 claims description 3
- 230000004044 response Effects 0.000 claims description 2
- 238000013316 zoning Methods 0.000 claims description 2
- 238000005507 spraying Methods 0.000 claims 3
- 238000007619 statistical method Methods 0.000 abstract description 2
- 238000007405 data analysis Methods 0.000 abstract 1
- 238000012545 processing Methods 0.000 description 12
- 238000003860 storage Methods 0.000 description 11
- 238000004458 analytical method Methods 0.000 description 7
- 238000004364 calculation method Methods 0.000 description 6
- 230000008901 benefit Effects 0.000 description 5
- 238000004891 communication Methods 0.000 description 5
- 230000002596 correlated effect Effects 0.000 description 5
- 230000000875 corresponding effect Effects 0.000 description 4
- 238000013481 data capture Methods 0.000 description 4
- 238000012805 post-processing Methods 0.000 description 4
- 230000008859 change Effects 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 230000007704 transition Effects 0.000 description 3
- 230000000007 visual effect Effects 0.000 description 3
- 230000004075 alteration Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000004519 manufacturing process Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 101000606504 Drosophila melanogaster Tyrosine-protein kinase-like otk Proteins 0.000 description 1
- 229910000831 Steel Inorganic materials 0.000 description 1
- 230000004913 activation Effects 0.000 description 1
- 239000011449 brick Substances 0.000 description 1
- 230000001276 controlling effect Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000007340 echolocation Effects 0.000 description 1
- 238000013213 extrapolation Methods 0.000 description 1
- 238000005111 flow chemistry technique Methods 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000001788 irregular Effects 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000000737 periodic effect Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000001953 sensory effect Effects 0.000 description 1
- 239000007921 spray Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 239000010959 steel Substances 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/05—Geographic models
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/698—Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/29—Geographical information databases
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/587—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using geographical or spatial information, e.g. location
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/7867—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title and artist information, manually generated time, location and usage information, user ratings
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/787—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using geographical or spatial information, e.g. location
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/953—Querying, e.g. by the use of web search engines
- G06F16/9537—Spatial or temporal dependent retrieval, e.g. spatiotemporal queries
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/10—Office automation; Time management
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0278—Product appraisal
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/06—Buying, selling or leasing transactions
- G06Q30/0601—Electronic shopping [e-shopping]
- G06Q30/0623—Item investigation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/06—Buying, selling or leasing transactions
- G06Q30/0601—Electronic shopping [e-shopping]
- G06Q30/0623—Item investigation
- G06Q30/0625—Directed, with specific intent or strategy
- G06Q30/0629—Directed, with specific intent or strategy for generating comparisons
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
- G06Q50/10—Services
- G06Q50/16—Real estate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/60—Editing figures and text; Combining figures or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/20—Finite element generation, e.g. wire-frame surface description, tesselation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4038—Image mosaicing, e.g. composing plane images from plane sub-images
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/28—Mobile studios
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/181—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
Definitions
- the present invention relates to methods and apparatus of generating image data. More specifically, the present invention relates to generating data descriptive of a continuum of three dimensional images, such as those comprising a geographic landscape.
- Photographs Images of geographic areas created on a local basis have been created in numerous formats. Photographs, movie cameras, video camera recorders, and more recently digital recorders have all been utilized to capture moving images of a geographic parcel. Photographs can be useful because they are easily associated with a particular real estate parcel, however they are limited in the scope of the subject matter that they can include. In addition, photographs must be manually acquired and docketed in order to be associated with a property and subsequently retrieved. Panoramas can increase the scope of the subject matter to a point, but are limited by a point of capture of the panoramic views.
- Movie pictures, video recorders and digital recorders and other “motion pictures” provide for increased scope of image capture. However, it is sometimes difficult to locate and view a particular portion of interest of images captured as motion pictures. In addition, correlation of particular portions of a motion picture with a particular real estate parcel can be difficult if the motion picture includes more than one real estate parcel. For example, it is relatively difficult to locate and view a particular real estate parcel in the context of its neighborhood setting, or particular aspects of its setting. Aerial images, such as satellite pictures of geographic areas have also been captured and specific parcels of land or landmarks can be located on the aerial images.
- the present invention provides methods and apparatus for generating a continuum of three dimensional image data.
- the continuum of three dimensional image data provides a street level representation of geographic locations.
- Two or more sets of image data are captured of a subject area, wherein each of the two or more images are captured from disparate points on a continuum.
- FIG. 1 illustrates a continuum from which image data sets capture a subject.
- FIG. 2 illustrates multiple images captured from disparate points along a continuum.
- FIG. 3 illustrates multiple image data sets that overlap subject matter.
- FIG. 4 illustrates a three-dimensional model generated form the image data sets.
- FIG. 5 illustrates a user interface generated from the image data sets and including multiple modalities of image data.
- FIG. 6 illustrates an exemplary user interface for traversing multiple modalities of image data.
- FIG. 7 illustrates an exemplary composite of image data set portions.
- FIG. 8 illustrates apparatus what may be used to implement the present invention.
- the present invention provides methods and apparatus for presenting three-dimensional model of a subject, such as a geographic area.
- the three-dimensional models can include a polygon based model format generated from a point cloud array and sprayed with actual image data of the subject, captured from disparate points along a continuum.
- a user interface is presented with user interactive controls which allow a user to traverse the three-dimensional model.
- a user location within the model will determine which portions of the three-dimensional model are viewed.
- Some embodiments can include user interactive devices which allow a user to traverse the three-dimensional model in three directions: a) an essentially up and down vertical movement; b) an essentially side-to-side horizontal movement; and c) a depth movement into and out of the three-dimensional model.
- Image data and three-dimensional polygon based models presented to an inventor at a given instance can be dependent upon the relative position selected by the user within the user interface.
- icons, overlays and metadata can be presented to a user upon demand to provide further information relating to a given geographic area within view.
- Still another aspect can include presentation of one or more geographic areas based upon criteria ascertainable from related data.
- image data can be correlated with geospatial data so that a particular street address or Cartesian coordinate, such as a latitude/longitude location, is designated and specific image data is presented which correlates with the geospatial data.
- the specific image data can be presented sprayed over three-dimensional structures generated from two-dimensional image data captured from that specific location.
- some embodiments can also include indications on the user interface of user selectable data related to a particular location.
- User selectable data can include, for example: cellular phone signal strength; wireless network data, including network name, location and signal strength; global position system satellite data and signal strength; radio signal strength; school district data, average yard size, property price range; building features; proximity to amenities; political boundaries: zoning restrictions; and almost any other measurable quantity.
- Datum can all be correlated according to a geographic location and presented in various combinations or upon demand.
- the data can be presented in conjunction with or independent of, an image data continuum or image data set.
- Video DriveByTM refers to street level video data captured in multiple angles encompassing a 360° view.
- Video FlyByTM refers to Aerial/Satellite oblique (angular) view images with polygon line views.
- RibbonViewTM refers to a film strip like view of properties which stands up at approximately 90° from a flat or angled aerial/satellite ortho image map and provides direct-on front images of properties to be displayed.
- Flash Viewer refers to direct streaming of video to an online user via a web browser.
- image data is captured from disparate points along a continuum.
- the continuum 100 includes the path of a vehicle carrying a digital camera, or other image data capturing device.
- Image data sets are captured at disparate points 101 - 106 along the continuum.
- Some preferred embodiments include capturing each image data set at an angle that is generally orthogonal to the subject matter 107 .
- Positional data and orientation of the camera capturing image data sets is recorded for each disparate point 101 - 106 along the continuum from which an image data set is captured.
- Positional data can include any data indicative of where the subject matter of an image data set is located.
- Some preferred embodiments of positional data include Cartesian coordinates that are context sensitive according to the mechanism used to generate coordinate data.
- Positional data can be generated, for example, by an image data recording device, such as a camera, associated with a device for recording a global position (GPS device). Time stamps associated with image data sets and time stamps associated with the GPS data can be utilized to correlate the GPS location data with image data set recorded by the image data recording device.
- an image data recording device such as a camera
- GPS device global position
- an altimeter can be used to record an altitude from which a camera records image data sets.
- the altitude data can be associated with an image data set, for example, metadata correlated with the image data set.
- Such embodiments can therefore include recordation of a latitude, longitude and altitude coordinate for a given image data set.
- geospatial data such as latitude and longitude coordinates
- GPS data can be time stamped and collected once every second.
- GPS reception can be interrupted, depending upon location relative to large object, such as multistory buildings, or cold cover. Therefore, some additional embodiments can include an accelerometer for recording motion associated with a camera and a GPS unit operatively attached to the camera.
- Data from the accelerometer can indicate movement of the camera.
- Some accelerometers such as micro electro-mechanical system (MEMs) accelerometers can easily be incorporated into a camera system assembly attached to a vehicle.
- MEMs micro electro-mechanical system
- Use of multiple MEM accelerometers positioned to measure movement in four or more directions along an x-axis, y-axis, and z-axis in relation to a camera can also be used to calculate direction of movement.
- the accelerometer can therefore be used to extrapolate a current position of the camera, based upon a last set of GPS geospatial data recorded.
- Geospatial data can be used to indicate an initial geographic position.
- a change in geospatial data can be additionally utilized to indicate velocity and direction of image data set capture.
- Accelerometer data may also be used to indicate a velocity and direction of image data set capture.
- Accelerometer data may also be used to indicate a velocity and direction of camera movement. Calculations of time elapsed at the indicated velocity (such as for example, the Kalman Filter) can yield a calculated position at a time of image capture, even if the time of image capture is between GPS readings.
- one standard can include tracking a camera position with a GPS unit that records location at a rate of once per second.
- the camera can record image data at a faster rate than once per second, such as, for example, one of: 12 images per second, 24 images per second or 29.97 images per second.
- An initial camera position can be recorded which correlates with a GPS reading, subsequent image data capture will occur in between the one second GPS reading interval.
- the camera position can be determined with a calculation based upon the velocity of camera movement supplied by the accelerometer and time elapsed since a previous GPS reading.
- Still other embodiments can utilize optical flow methodology and visual odometry to facilitate calculations of a camera position and the velocity of a vehicle or person from which a series of image data sets are captured.
- Visual odometry can be accomplished with a single omni-directional camera or with stereo cameras, and is based upon processing which tracks the position of salient features in a series of feature sets and calculates movement based upon the relative positions of the features.
- camera based simultaneous localization and mapping (SLAM) of visual image data can also be used to facilitate calculations of velocity of a change in position of a camera used to capture image data sets.
- the velocity will be directly tied to the motion of a vehicle to which the camera is mounted, or a person carrying a camera rig.
- Orientation of a camera can include a direction of image capture recorded by the camera. Orientation can be designated, for example, in relation to the cardinal directions, i.e. north, south, east and west. Any means available to record such a designation, such as an electronic compass, is within the scope of the present invention. However, it may be desirable to include a means to record the orientation with a greater degree of accuracy than is typically available through the use of an electronic compass.
- orientation can be determined according to a fixed position of a camera in relation to the direction of travel of a vehicle (or person) used to transport the camera.
- a plurality of cameras can be fixedly attached to a vehicle capturing Video DriveByTM data.
- Each camera therefore maintains a constant direction of image capture in relation to the heading of the vehicle.
- Mechanics of the camera such as, for example, lens parameters and shutter speed, can indicate a depth of field during camera image data capture.
- Some embodiments can also include simultaneously capturing multiple image data sets and correlating two or more of the image data sets. Correlation can be accomplished via a time stamp or other chronological or synchronous.
- the position of a camera can be combined with a direction of image capture and the depth of field of the camera, to determine a location of image data captured by the camera at a particular instance in time.
- the present invention can also include apparatus for utilizing echo location to determine a distance of an object from a camera, capturing an image data set and storing a correlated distance with the image data set.
- radar data can be correlated with a camera image data set to calculate the location of various objects captured by the camera.
- a time stamp can also be combined with data to quantify a location for a particular image formed by the captured image data.
- data used to calculate a location of an image is stored in a metadata file space associated with the image data.
- some embodiments can store metadata in the exchangeable image file format (EXIF), TIFFTAGS or International Press Telecommunication Council (IPTC) formats.
- Image data may be stored, for example in JPEG or TIFF formats.
- other metadata formats can also be used.
- the image data and metafile data are stored on an external data storage device, such as a hard disk drive operatively attached to the camera.
- the data can be stored in the camera.
- the metadata can include data descriptive of some or all of: date and time; camera settings such aperture, shutter speed and focal length; geospatial data from a GPS receiver unit; accelerometer data inertial guidance system data; camera orientation; and camera fixed position related to vehicle travel.
- image data sets 201 - 203 are illustrated which capture images of a subject 200 from different points along a continuum.
- image data sets 304 - 307 overlap the subject matter captured. The overlap allows for features present in one image data set to be present in a second image data set, and preferably in a tertiary image data set.
- FIG. 4 a block diagram illustrates three-dimensional structures 405 - 407 created from image data captured.
- the structures can be created by generating a point cloud array from a plurality of features contained in the image data sets and converting the point cloud array to a polygon based model.
- the polygon based model can be associated with a location relative to the image data set and image data can be sprayed over the polygon based model.
- image data sprayed over the polygon based models includes the composite of image data.
- a user can traverse the presentation of the composite of image data and the three-dimensional structures. Movement through the user interface can be accomplished with any known interaction device, such as, for example one or more of: a keyboard, mouse, video game controller Apple IPhone, digital cellular phone, remote controller, WiiMote and a joystick.
- any known interaction device such as, for example one or more of: a keyboard, mouse, video game controller Apple IPhone, digital cellular phone, remote controller, WiiMote and a joystick.
- the user interface can allow a user to move in three directions through a scene presented.
- the directions can include an x coordinate providing a vertical dimension 401 , a y dimension providing a horizontal dimension 402 and a z coordinate providing a depth dimension 403 .
- Three-dimensional modeling sprayed with actual image data allows the user to proceed in “off-track” paths that do not match a path created by the disparate points from which data is collected.
- a time dimension can also be traversed, such as, for example, image data sets of a geospatial designation across time chronology such as a period of seasons or years.
- Some embodiments can include a “take me there” time designation, wherein a user can indicate a time period and geospatial description.
- a three-dimensional model can be generated with image data sets from the indicated time period for the designated geospatial area.
- planes of image data 405 can change according to the perspective chosen by the user.
- a three-dimensional structure of a house can be sprayed with image data included in the continuum composite of image data.
- the actual image data will correlate with a surface of the three-dimensional structure facing the disparate points from which the image data was captured.
- additional surface of the house 406 can also be sprayed wither texture and color captured by the actual image data. In this way, as a user moves in a z direction beyond front surface of the house, the user will be presented with additional surfaces of the house, with each surface having a color and texture extrapolated from the image data from the front of the house.
- an object such as a tree 405 can also be created in three dimensions with color and texture based upon the image data composite.
- user interactive device controls can be provided which allow a user to remove objects such as the tree 405 and the house 406 . Removal of an object will allow the user to see beyond the removed object into the scene provided. For example, if a tree 405 obscures the view of the front of a house 406 , activation of a user interactive control can remove the tree 405 allowing an unobstructed view of the three-dimensional model of the house 406 . If actual image data is not available for those portions of the house now viewable in the unobstructed view, then the present invention allows for color and texture data matching the known surfaces of the house to be sprayed over those portions now viewable.
- Some embodiments can additionally allow a user to modify color and texture of three-dimensional models presented to them. Such modifications facilitate “what-if” scenarios. For example, a buyer contemplating purchasing a particular property can be presented with a three-dimensional model of the property with three-dimensional representations of objects on the property sprayed with actual image data captured of the property. According to some embodiments, the buyer can then modify the view presented to represent possible changes that may be made to the property. Changes may include, to continue the example, providing a brick façade over a painted surface, or removal of shrubbery, or addition of shrubbery. Addition of items can be provided from a menu of three-dimensional objects made available to the user. Color selections and textures can also be made available via a user selectable menu. In this fashion a unique blend of a composite continuum of image data and user defined image data can be presented, including three-dimensional modeled objects derived from the image data and user modifications to such models.
- a user may be able to select a building and designate from a user interface a type and color of roofing the user would like to see illustrated for the building. Accordingly, a user may designate an architectural shingle in various hues, or a Spanish tile or steel roof and the present invention will spray the pattern and color selected onto a three-dimensional model of the building.
- a user may upload a color or hue or texture to be sprayed upon a three-dimensional model of a structure.
- the color and hue may also be stored and made available for subsequent application to portions of an image data set.
- the present invention can also include generation of a user interface 500 with multiple representations of image data relating to a geographic area.
- an aerial view of a geographic area 504 can be combined with one or more two-dimensional continuums of composite image data 505 - 506 .
- the two-dimensional continuums of composite image data 505 - 506 can be generated according to methods and apparatus described in a related co-pending patent application by the same inventors.
- one or more continuums of disparate points from which image data is captured can also be indicated on the aerial view 504 .
- Three-dimensional models 510 - 503 can also be included in the user interface 500 and located on the aerial view according to geospatial data generated during image data capture and processing of the image data.
- the three-dimensional models can be generated from multiple image data sets captured from disparate points along a continuum sprayed with the image data captured.
- a user may locate a particular geographic area via different vehicles made available to a user.
- a user can enter a geospatial designation, such as, for example, one or more of: a street address; a Cartesian coordinate; a “Favorite” name designating a geospatial area; a popular name for a geospatial area; a coded indication for a geospatial area, such as a multiple listing number or data file number; and metadata.
- a “take me there” command can be executed to bring the user to the requested geospatial area.
- the present invention provides a vehicle “cockpit” interface which allows a user to virtually “fly” over an aerial view and then transition into a virtual “dashboard” and “drive” mode along a street level two-dimensional continuum of actual image data, and in some preferred embodiments, a continuum of composite image data.
- the user may also transition into a virtual three-dimensional modality which includes three-dimensional models sprayed with image data captured from a continuum of disparate points.
- the three-dimensional models can be viewed from a virtual “drive by” mode or a virtual “walkabout.”
- the virtual walkabout can allow a user to go off path and view the geographic area from amongst three-dimensional models sprayed with image data captured from a continuum of disparate points.
- the user may manipulate the appearance of the three-dimensional models.
- a take me there command can be additionally operative to designate a modality, such as the aerial view, two-dimensional ribbon view, or three-dimensional drive by or three-dimensional walk about view into which a user will enter.
- a modality such as the aerial view, two-dimensional ribbon view, or three-dimensional drive by or three-dimensional walk about view into which a user will enter.
- a user can be provided with tools that provide analysis of the image data shown. Analysis can include, for example, designation of a particular portion of a geographic area 508 .
- the portion 508 can be representative of a real estate parcel, as indicated by municipal records, or some other designation.
- a user may draw a polygon 508 onto an aerial image of a geographic area 500 and ask for analysis of the area contained within the polygon 508 .
- Analysis can include, by way of example, a calculation of an area included within the polygon. Other analysis can include more sophisticated calculations. For example, it is within the scope of the invention to designate property boundaries, such as, through county record data and locate an image data of the property.
- the image data can be used to generate a three-dimensional model of the property, including a model of a house within the property boundaries and other features, such as trees and utility demarcations.
- the present invention may also be used to generate a size of a front yard, and, by extrapolation, the size of the side and back yards.
- An approximate size of a house can also be calculated based the dimensions of the three-dimensional model generated.
- Other features, such as how many stories, the house includes and the type of façade the house has may also be ascertained from analyzing the image data and the model.
- Statistical representations can also be made, according to the analysis.
- Statistical representations can provide a summary of characteristics of a single real estate parcel, or a designated area.
- statistical analysis can include one or more of an average, median and high/low value designation for: the size of a property within a given neighborhood, the size of a front yard, how much road frontage each property includes, the average size of the houses within the designated area, the number of stories comprising the buildings within the area, siding types of buildings within the area, the size of side yards between houses, height of trees within the area, and almost any other data that may be calculated from the image data sets and the three-dimensional models.
- the interface 600 can include a portion with an aerial view 601 and a polygon designation of geographic area of interest, such as for example, a particular real estate parcel or neighborhood location.
- the interface may also include a portion with a two-dimensional continuum of image data 603 A.
- the continuum of image data is preferably a composite of image data sets captured from disparate points.
- Another portion 603 B can include image data sprayed over three-dimensional models 604 generated from point clouds of features captured by the image data sets.
- image data sprayed over the three-dimensional polygon formats includes a composite image formed by aligning two or more of the image data sets.
- the present invention creates a composite through alignment of portions of data from more than one data set. Alignment can be accomplished in image data processing.
- image data processing the images 701 - 703 are aligned to form a composite image 700 .
- the composite image 700 is essentially two dimensional image data arranged as a second continuum, or ribbon.
- the second continuum includes ongoing image data 701 - 703 captured from the points defining the first continuum.
- the series of points of image capture in the first continuum includes positions of a vehicle carrying an image capture device, such as a camera, as the vehicle traverses a path proximate to a geographic area.
- the camera is positioned to capture image data of the geographic area.
- Image data 701 - 703 is periodically captured as the vehicle traverses the path. The motion of the vehicle, combined with the periodic capture of image data 701 - 703 , thereby results in image data 701 - 703 being captured from disparate points along the first continuum.
- a preferred embodiment includes capture of image data with a motion vector of the camera in space maintained generally orthogonal to a subject for which image data will be captured. Orthogonal capture of the image data can facilitate consistency for subsequent composite of portions of the image data captured. Therefore, data captured at an angle of between about 75° and 105° can provide data most easily assembled into a continuum of data. However, other angles may be used to create different effects. In addition, in some embodiments physical limitation may dictate the use of other angles. In such instances, post processing can be utilized to compensate for the angle of image data capture.
- a continuous pictorial representation includes RibbonViewTM by Real Data Systems. RibbonViewTM correlates a ribbon of geographic image data with geospatial designations to facilitate identification of a particular geographic area, as discussed more fully below.
- image capture processing can be accomplished in real time or via post image capture processing.
- select portions 704 - 707 of two or more sets of captured image data are aligned to generate the composite image 708 .
- some preferred embodiments include vertical slices of data 704 - 707 aligned in a horizontal plane to form the composite image 708 .
- the length of a horizontal plane defining a composite image 708 is only limited by the length of a continuum along which points are defined and from which image data 704 - 707 is captured.
- the use of only slices of data 704 - 707 from any particular captured image provides for a higher quality image 708 .
- the quality is increased, for example, when a temporary obstruction, such as a passing car, person or animal, captured in one image data set, is only represented in a thin slice of a continuous ribbon 704 - 707 data.
- alignment of multiple thin slices of image data 704 - 707 is facilitated from the perspective of which aberrations typical human sensory is capable of distinguishing. Numerous thin slices 704 - 707 are perceived as a smoother transition across the horizontal plane defining the composite image 708 removing a choppiness that may be experienced with the use of larger data images.
- the present invention can include a uniform width of each slice of data 704 - 707 or a variable width.
- the width of a particular slice may vary, for example, upon one or more of the velocity of a vehicle from which image data sets are captured, the sample rate of a camera used to capture an image data set 701 - 703 , the resolution of a picture comprising an image data set 701 - 703 and the path of a camera.
- a high resolution image generated by a 2.1 mega pixel camera may have a 1600 by 1200 resolution and allow for a thinner slice 704 - 707 that includes a width of between about 5 to 700 pixels of an image data set.
- Embodiments with very high quality can include a slice 704 - 707 of between 1 to 2 pixels.
- some embodiments can include a slice 704 - 707 of an image data set 701 - 703 that includes a percentage of the image data set 701 - 703 . Accordingly, some preferred embodiments can include a slice 704 - 707 of between about 5% to about 12% of an image data set. Other preferred embodiments can include up to about 50% of an image data set. However, it should be understood that some embodiments include a slice 704 - 707 that constitutes any fraction of the complete image data set.
- a database or other data processing mechanism can track each portion or slice 704 - 708 utilized to construct a continuum and associate the slice 704 - 708 with an original image 701 - 703 from which the slice 704 - 708 is derived.
- User interactive devices can execute the retrieval of an entire original image 701 - 703 or a series of original images 701 - 703 upon request. In some instances, one or more original images 701 - 703 may provide detail not included in the composite image 708 .
- a selected portion of an image data set may be determined by the physical attributes of the equipment used to capture an image data set.
- a typical camera lens can impart some distortion to an image data set, as represented in the illustration by an elongated portion and a compressed portion.
- Utilizing only a portion of an image data set, such as, for example, a center portion vertical slice, can minimize the effect of distortion introduced by a lens, or other source, to a composite image. Distortion is minimized when the composite image is made to include those portions of the image data set corresponding with the center of the lens.
- Specific embodiments may likewise account for other aberrations that may be present in a set of image data.
- the length of a composite image generated according to the present invention is limited only by the ability to capture image data from additional points on a continuum and store the captured image data for post processing.
- the post processing allows for the alignment of portions of the image data compiled into a composite two-dimensional view that can continue so long as additional image data is made available to be added to it.
- FIG. 8 illustrates a controller 800 that may be utilized to implement some embodiments of the present invention.
- the controller 800 comprises a processor unit 810 , such as one or more processors, coupled to a communication device 820 configured to communicate via a communication network (not shown in FIG. 8 ).
- the communication device 820 may be used to communicate, for example, with one or more online devices, such as a personal computer, laptop or a handheld device.
- the processor 810 is also in communication with a storage device 830 .
- the storage device 830 may comprise any appropriate information storage device, including combinations of magnetic storage devices (e.g., magnetic tape and hard disk drives), optical storage devices, and/or semiconductor memory devices such as Random Access Memory (RAM) devices and Read Only Memory (ROM) devices.
- RAM Random Access Memory
- ROM Read Only Memory
- the storage device 830 can store a program 840 for controlling the processor 810 .
- the processor 810 performs instructions of the program 840 , and thereby operates in accordance with the present invention.
- the processor 810 may also cause the communication device 820 to transmit information, including, in some instances, control commands to operate apparatus to implement the processes described above.
- the storage device 830 can additionally store related data in a database 830 A and database 830 B, as needed.
- apparatus includes a video and data server farm.
- the video and data server farm includes at least one video storage server that stores video image files containing video drive-by data that corresponds to a geographic location, a database server that processes a data query received from a user over the Internet that corresponds to a geographic location of interest, and an image server.
- the database server identifies video image files stored in the video storage server that correspond to the geographic location of interest contained in the data query, and transfers the video image files over a pre-processing network to the image processing server.
- the image processing server converts the video drive-by data to post processed video data corresponding to a desired image format, and transfers the post processed video data via post-processing network to the Internet response to the query.
- a landing zone server can also be included which receives the video drive-by data from a portable memory device and permits the viewing and analysis of the video drive-by data prior to storage in the video storage server. Still further, a map server is preferably provided to present a static image an overhead view of the geographic location of interest.
- Embodiments can also include one or more servers described above included in one or more physical units. Each server does not need to be a disparate apparatus. Still other embodiments can include one or more or the servers described above included in multiple physical units. Some embodiments can even include a single server, as described which includes multiple physical apparatus units at disparate locations.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Business, Economics & Management (AREA)
- Databases & Information Systems (AREA)
- General Engineering & Computer Science (AREA)
- Strategic Management (AREA)
- Data Mining & Analysis (AREA)
- Accounting & Taxation (AREA)
- Finance (AREA)
- Economics (AREA)
- General Business, Economics & Management (AREA)
- Marketing (AREA)
- Multimedia (AREA)
- Development Economics (AREA)
- Software Systems (AREA)
- Tourism & Hospitality (AREA)
- Library & Information Science (AREA)
- Human Resources & Organizations (AREA)
- Entrepreneurship & Innovation (AREA)
- Geometry (AREA)
- Signal Processing (AREA)
- Computer Graphics (AREA)
- Remote Sensing (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Primary Health Care (AREA)
- Game Theory and Decision Science (AREA)
- Operations Research (AREA)
- Quality & Reliability (AREA)
- Computer Hardware Design (AREA)
- Human Computer Interaction (AREA)
- Artificial Intelligence (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- Instructional Devices (AREA)
- Processing Or Creating Images (AREA)
Abstract
The present invention provides methods and apparatus for generating a three dimensional output which includes a continuum of image data sprayed over three-dimensional models. The three-dimensional models can be representative of features captured by the image data wherein image data can be captured at multiple disparate points along another continuum. The user interface can also include multiple modalities of image data and statistical analysis of the image data.
Description
The present application claims priority to pending patent application Ser. No. 11/216,465, filed, Aug. 31, 2005 and entitled, “Apparatus and Method for Producing Video Drive-By Data Corresponding to a Geographic Location,”; and also pending application Ser. No. 11/702,708 filed Feb. 6, 2007 and entitled Methods and Apparatus for Generating a Continuum of Image Data; and also pending application Ser. No. 11/702,707 filed Feb. 6, 2007 and entitled Methods and Apparatus for Presenting a Continuum of Image Data; the contents of which are each relied upon and incorporated by reference. This application is a continuation-in-part of, and claims the benefit under 35 U.S.C. §120 from, nonprovisional U.S. patent application Ser. No. 11/216,465, entitled “Apparatus And Method For Producing Video Drive-By Data Corresponding To A Geographic Location,” filed on Aug. 31, 2005, now U.S. Pat. No. 7,389,181. U.S. patent application Ser. No. 11/216,465 claims the benefit under 35 U.S.C. §119 from provisional U.S. patent application Ser. No. 60/605,498, entitled “Video and data processing system”, filed on Aug. 31, 2004. This application is also a continuation-in-part of, and claims the benefit under 35 U.S.C. §120 from, nonprovisional U.S. patent application Ser. No. 11/702,707, entitled “Methods And Apparatus For Presenting A Continuum Of Image Data,” filed on Feb. 6, 2007. This application is also a continuation-in-part of, and claims the benefit under 35 U.S.C. §120 from, nonprovisional U.S. patent application Ser. No. 11/702,708, entitled “Methods And Apparatus For Generating A Continuum Of Image Data,” filed on Feb. 6, 2007, now U.S. Pat. No. 7,929,800. This application also claims the benefit under 35 U.S.C. §119 from provisional U.S. patent application Ser. No. 60/902,953, filed on Feb. 22, 2007. This application incorporates by reference U.S. patent application Ser. No. 11/216,465. This application also incorporates by reference U.S. patent application Ser. No. 11/702,707. This application also incorporates by reference U.S. patent application Ser. No. 11/702,708.
The present invention relates to methods and apparatus of generating image data. More specifically, the present invention relates to generating data descriptive of a continuum of three dimensional images, such as those comprising a geographic landscape.
Images of geographic areas created on a local basis have been created in numerous formats. Photographs, movie cameras, video camera recorders, and more recently digital recorders have all been utilized to capture moving images of a geographic parcel. Photographs can be useful because they are easily associated with a particular real estate parcel, however they are limited in the scope of the subject matter that they can include. In addition, photographs must be manually acquired and docketed in order to be associated with a property and subsequently retrieved. Panoramas can increase the scope of the subject matter to a point, but are limited by a point of capture of the panoramic views.
Movie pictures, video recorders and digital recorders and other “motion pictures” provide for increased scope of image capture. However, it is sometimes difficult to locate and view a particular portion of interest of images captured as motion pictures. In addition, correlation of particular portions of a motion picture with a particular real estate parcel can be difficult if the motion picture includes more than one real estate parcel. For example, it is relatively difficult to locate and view a particular real estate parcel in the context of its neighborhood setting, or particular aspects of its setting. Aerial images, such as satellite pictures of geographic areas have also been captured and specific parcels of land or landmarks can be located on the aerial images.
In addition, methods and apparatus have been known for generating point clouds and generating representations of particular objects from processing of the point clouds.
However, prior to the present invention, there has not been a mechanism for generating a continuum of object representations based upon point cloud data.
Accordingly, the present invention provides methods and apparatus for generating a continuum of three dimensional image data. In some embodiments, the continuum of three dimensional image data provides a street level representation of geographic locations. Two or more sets of image data are captured of a subject area, wherein each of the two or more images are captured from disparate points on a continuum.
As presented herein, various embodiments of the present invention will be described, followed by some specific examples of various components that can be utilized to implement the embodiments. The following drawings facilitate the description of some embodiments:
The present invention provides methods and apparatus for presenting three-dimensional model of a subject, such as a geographic area. The three-dimensional models can include a polygon based model format generated from a point cloud array and sprayed with actual image data of the subject, captured from disparate points along a continuum.
A user interface is presented with user interactive controls which allow a user to traverse the three-dimensional model. A user location within the model will determine which portions of the three-dimensional model are viewed. Some embodiments can include user interactive devices which allow a user to traverse the three-dimensional model in three directions: a) an essentially up and down vertical movement; b) an essentially side-to-side horizontal movement; and c) a depth movement into and out of the three-dimensional model. Image data and three-dimensional polygon based models presented to an inventor at a given instance can be dependent upon the relative position selected by the user within the user interface.
In another aspect, icons, overlays and metadata can be presented to a user upon demand to provide further information relating to a given geographic area within view.
Still another aspect can include presentation of one or more geographic areas based upon criteria ascertainable from related data. For example, image data can be correlated with geospatial data so that a particular street address or Cartesian coordinate, such as a latitude/longitude location, is designated and specific image data is presented which correlates with the geospatial data. The specific image data can be presented sprayed over three-dimensional structures generated from two-dimensional image data captured from that specific location.
In additional aspects, some embodiments can also include indications on the user interface of user selectable data related to a particular location. User selectable data can include, for example: cellular phone signal strength; wireless network data, including network name, location and signal strength; global position system satellite data and signal strength; radio signal strength; school district data, average yard size, property price range; building features; proximity to amenities; political boundaries: zoning restrictions; and almost any other measurable quantity. Datum can all be correlated according to a geographic location and presented in various combinations or upon demand. In addition, the data can be presented in conjunction with or independent of, an image data continuum or image data set.
As used herein, Video DriveBy™ refers to street level video data captured in multiple angles encompassing a 360° view.
As used herein, Video FlyBy™ refers to Aerial/Satellite oblique (angular) view images with polygon line views.
As used herein, RibbonView™ refers to a film strip like view of properties which stands up at approximately 90° from a flat or angled aerial/satellite ortho image map and provides direct-on front images of properties to be displayed.
As used herein, Flash Viewer (Streaming Video) refers to direct streaming of video to an online user via a web browser.
Methods
According to the present invention, image data is captured from disparate points along a continuum. Referring now to FIG. 1 , in some embodiments the continuum 100 includes the path of a vehicle carrying a digital camera, or other image data capturing device. Image data sets are captured at disparate points 101-106 along the continuum.
Some preferred embodiments include capturing each image data set at an angle that is generally orthogonal to the subject matter 107.
Positional data and orientation of the camera capturing image data sets is recorded for each disparate point 101-106 along the continuum from which an image data set is captured. Positional data can include any data indicative of where the subject matter of an image data set is located. Some preferred embodiments of positional data include Cartesian coordinates that are context sensitive according to the mechanism used to generate coordinate data.
Positional data can be generated, for example, by an image data recording device, such as a camera, associated with a device for recording a global position (GPS device). Time stamps associated with image data sets and time stamps associated with the GPS data can be utilized to correlate the GPS location data with image data set recorded by the image data recording device.
In still another aspect, in some embodiments, an altimeter can be used to record an altitude from which a camera records image data sets. The altitude data can be associated with an image data set, for example, metadata correlated with the image data set. Such embodiments can therefore include recordation of a latitude, longitude and altitude coordinate for a given image data set. In addition, it is also within the scope of this invention to record a time of generation of an image data set and a depth of focus for an image data set.
According to the present invention, geospatial data, such as latitude and longitude coordinates, can be generated by the GPS and stored with image data generated by the camera. In some embodiments, GPS data can be time stamped and collected once every second. However, in some instances, GPS reception can be interrupted, depending upon location relative to large object, such as multistory buildings, or cold cover. Therefore, some additional embodiments can include an accelerometer for recording motion associated with a camera and a GPS unit operatively attached to the camera.
Data from the accelerometer can indicate movement of the camera. Some accelerometers, such as micro electro-mechanical system (MEMs) accelerometers can easily be incorporated into a camera system assembly attached to a vehicle. Use of multiple MEM accelerometers positioned to measure movement in four or more directions along an x-axis, y-axis, and z-axis in relation to a camera can also be used to calculate direction of movement. The accelerometer can therefore be used to extrapolate a current position of the camera, based upon a last set of GPS geospatial data recorded.
Geospatial data can be used to indicate an initial geographic position. A change in geospatial data can be additionally utilized to indicate velocity and direction of image data set capture. Accelerometer data may also be used to indicate a velocity and direction of image data set capture. Accelerometer data may also be used to indicate a velocity and direction of camera movement. Calculations of time elapsed at the indicated velocity (such as for example, the Kalman Filter) can yield a calculated position at a time of image capture, even if the time of image capture is between GPS readings.
For example, one standard can include tracking a camera position with a GPS unit that records location at a rate of once per second. The camera can record image data at a faster rate than once per second, such as, for example, one of: 12 images per second, 24 images per second or 29.97 images per second. An initial camera position can be recorded which correlates with a GPS reading, subsequent image data capture will occur in between the one second GPS reading interval. The camera position can be determined with a calculation based upon the velocity of camera movement supplied by the accelerometer and time elapsed since a previous GPS reading.
Still other embodiments can utilize optical flow methodology and visual odometry to facilitate calculations of a camera position and the velocity of a vehicle or person from which a series of image data sets are captured. Visual odometry can be accomplished with a single omni-directional camera or with stereo cameras, and is based upon processing which tracks the position of salient features in a series of feature sets and calculates movement based upon the relative positions of the features. In some embodiments, camera based simultaneous localization and mapping (SLAM) of visual image data can also be used to facilitate calculations of velocity of a change in position of a camera used to capture image data sets. Typically, the velocity will be directly tied to the motion of a vehicle to which the camera is mounted, or a person carrying a camera rig.
Orientation of a camera can include a direction of image capture recorded by the camera. Orientation can be designated, for example, in relation to the cardinal directions, i.e. north, south, east and west. Any means available to record such a designation, such as an electronic compass, is within the scope of the present invention. However, it may be desirable to include a means to record the orientation with a greater degree of accuracy than is typically available through the use of an electronic compass.
Therefore, in some embodiments, orientation can be determined according to a fixed position of a camera in relation to the direction of travel of a vehicle (or person) used to transport the camera. For example, a plurality of cameras can be fixedly attached to a vehicle capturing Video DriveBy™ data. Each camera therefore maintains a constant direction of image capture in relation to the heading of the vehicle. Mechanics of the camera, such as, for example, lens parameters and shutter speed, can indicate a depth of field during camera image data capture. Some embodiments can also include simultaneously capturing multiple image data sets and correlating two or more of the image data sets. Correlation can be accomplished via a time stamp or other chronological or synchronous.
The position of a camera can be combined with a direction of image capture and the depth of field of the camera, to determine a location of image data captured by the camera at a particular instance in time. The present invention can also include apparatus for utilizing echo location to determine a distance of an object from a camera, capturing an image data set and storing a correlated distance with the image data set. For example, radar data can be correlated with a camera image data set to calculate the location of various objects captured by the camera. A time stamp can also be combined with data to quantify a location for a particular image formed by the captured image data.
In some embodiments of the present invention data used to calculate a location of an image is stored in a metadata file space associated with the image data. For example, some embodiments can store metadata in the exchangeable image file format (EXIF), TIFFTAGS or International Press Telecommunication Council (IPTC) formats. Image data may be stored, for example in JPEG or TIFF formats. However, other metadata formats can also be used. Typically, due to the size of data files that are generated during capture of Video DriveBy™ data, the image data and metafile data are stored on an external data storage device, such as a hard disk drive operatively attached to the camera. However, in some embodiments, the data can be stored in the camera.
As discussed above, the metadata can include data descriptive of some or all of: date and time; camera settings such aperture, shutter speed and focal length; geospatial data from a GPS receiver unit; accelerometer data inertial guidance system data; camera orientation; and camera fixed position related to vehicle travel.
Referring now to FIG. 2 , exemplary image data sets 201-203 are illustrated which capture images of a subject 200 from different points along a continuum. In some embodiments, as illustrated in FIG. 3 , image data sets 304-307 overlap the subject matter captured. The overlap allows for features present in one image data set to be present in a second image data set, and preferably in a tertiary image data set.
Referring now to FIG. 4 a block diagram illustrates three-dimensional structures 405-407 created from image data captured. In some embodiments, the structures can be created by generating a point cloud array from a plurality of features contained in the image data sets and converting the point cloud array to a polygon based model. The polygon based model can be associated with a location relative to the image data set and image data can be sprayed over the polygon based model. In some preferred embodiments, image data sprayed over the polygon based models includes the composite of image data.
According to the present invention, a user can traverse the presentation of the composite of image data and the three-dimensional structures. Movement through the user interface can be accomplished with any known interaction device, such as, for example one or more of: a keyboard, mouse, video game controller Apple IPhone, digital cellular phone, remote controller, WiiMote and a joystick.
The user interface can allow a user to move in three directions through a scene presented. The directions can include an x coordinate providing a vertical dimension 401, a y dimension providing a horizontal dimension 402 and a z coordinate providing a depth dimension 403. Three-dimensional modeling sprayed with actual image data allows the user to proceed in “off-track” paths that do not match a path created by the disparate points from which data is collected.
In some embodiments, a time dimension can also be traversed, such as, for example, image data sets of a geospatial designation across time chronology such as a period of seasons or years. Some embodiments can include a “take me there” time designation, wherein a user can indicate a time period and geospatial description. According to the present invention, a three-dimensional model can be generated with image data sets from the indicated time period for the designated geospatial area.
As the user moves into a scene presented in the user interface, planes of image data 405 can change according to the perspective chosen by the user. For example, a three-dimensional structure of a house can be sprayed with image data included in the continuum composite of image data. The actual image data will correlate with a surface of the three-dimensional structure facing the disparate points from which the image data was captured. According to the present invention, additional surface of the house 406 can also be sprayed wither texture and color captured by the actual image data. In this way, as a user moves in a z direction beyond front surface of the house, the user will be presented with additional surfaces of the house, with each surface having a color and texture extrapolated from the image data from the front of the house.
In addition, an object such as a tree 405 can also be created in three dimensions with color and texture based upon the image data composite. In some embodiments, user interactive device controls can be provided which allow a user to remove objects such as the tree 405 and the house 406. Removal of an object will allow the user to see beyond the removed object into the scene provided. For example, if a tree 405 obscures the view of the front of a house 406, activation of a user interactive control can remove the tree 405 allowing an unobstructed view of the three-dimensional model of the house 406. If actual image data is not available for those portions of the house now viewable in the unobstructed view, then the present invention allows for color and texture data matching the known surfaces of the house to be sprayed over those portions now viewable.
Some embodiments can additionally allow a user to modify color and texture of three-dimensional models presented to them. Such modifications facilitate “what-if” scenarios. For example, a buyer contemplating purchasing a particular property can be presented with a three-dimensional model of the property with three-dimensional representations of objects on the property sprayed with actual image data captured of the property. According to some embodiments, the buyer can then modify the view presented to represent possible changes that may be made to the property. Changes may include, to continue the example, providing a brick façade over a painted surface, or removal of shrubbery, or addition of shrubbery. Addition of items can be provided from a menu of three-dimensional objects made available to the user. Color selections and textures can also be made available via a user selectable menu. In this fashion a unique blend of a composite continuum of image data and user defined image data can be presented, including three-dimensional modeled objects derived from the image data and user modifications to such models.
By way of non-limiting example, in some embodiments, a user may be able to select a building and designate from a user interface a type and color of roofing the user would like to see illustrated for the building. Accordingly, a user may designate an architectural shingle in various hues, or a Spanish tile or steel roof and the present invention will spray the pattern and color selected onto a three-dimensional model of the building. In still other embodiments, it is within the scope of some embodiments of the present invention to allow a user to upload a texture and color to be sprayed upon a three-dimensional model, or a portion of the three-dimensional model, such as a roof portion. According to such embodiments, a user may upload a color or hue or texture to be sprayed upon a three-dimensional model of a structure. The color and hue may also be stored and made available for subsequent application to portions of an image data set.
Referring now to FIG. 5 , the present invention can also include generation of a user interface 500 with multiple representations of image data relating to a geographic area. For example, an aerial view of a geographic area 504 can be combined with one or more two-dimensional continuums of composite image data 505-506. The two-dimensional continuums of composite image data 505-506 can be generated according to methods and apparatus described in a related co-pending patent application by the same inventors. In addition, one or more continuums of disparate points from which image data is captured can also be indicated on the aerial view 504. Three-dimensional models 510-503 can also be included in the user interface 500 and located on the aerial view according to geospatial data generated during image data capture and processing of the image data.
As described in a related application, the three-dimensional models can be generated from multiple image data sets captured from disparate points along a continuum sprayed with the image data captured.
According to some embodiments of the present invention, a user may locate a particular geographic area via different vehicles made available to a user. For example, a user can enter a geospatial designation, such as, for example, one or more of: a street address; a Cartesian coordinate; a “Favorite” name designating a geospatial area; a popular name for a geospatial area; a coded indication for a geospatial area, such as a multiple listing number or data file number; and metadata. According to the present invention, a “take me there” command can be executed to bring the user to the requested geospatial area. Alternatively, the present invention provides a vehicle “cockpit” interface which allows a user to virtually “fly” over an aerial view and then transition into a virtual “dashboard” and “drive” mode along a street level two-dimensional continuum of actual image data, and in some preferred embodiments, a continuum of composite image data. The user may also transition into a virtual three-dimensional modality which includes three-dimensional models sprayed with image data captured from a continuum of disparate points. The three-dimensional models can be viewed from a virtual “drive by” mode or a virtual “walkabout.” The virtual walkabout can allow a user to go off path and view the geographic area from amongst three-dimensional models sprayed with image data captured from a continuum of disparate points. In addition, as discussed above, the user may manipulate the appearance of the three-dimensional models.
A take me there command can be additionally operative to designate a modality, such as the aerial view, two-dimensional ribbon view, or three-dimensional drive by or three-dimensional walk about view into which a user will enter.
In another aspect of the present invention, a user can be provided with tools that provide analysis of the image data shown. Analysis can include, for example, designation of a particular portion of a geographic area 508. The portion 508 can be representative of a real estate parcel, as indicated by municipal records, or some other designation. For example, a user may draw a polygon 508 onto an aerial image of a geographic area 500 and ask for analysis of the area contained within the polygon 508.
Analysis can include, by way of example, a calculation of an area included within the polygon. Other analysis can include more sophisticated calculations. For example, it is within the scope of the invention to designate property boundaries, such as, through county record data and locate an image data of the property. The image data can be used to generate a three-dimensional model of the property, including a model of a house within the property boundaries and other features, such as trees and utility demarcations. Based upon the location of the house within the property boundaries, the present invention may also be used to generate a size of a front yard, and, by extrapolation, the size of the side and back yards. An approximate size of a house can also be calculated based the dimensions of the three-dimensional model generated. Other features, such as how many stories, the house includes and the type of façade the house has may also be ascertained from analyzing the image data and the model.
Statistical representations can also be made, according to the analysis. Statistical representations can provide a summary of characteristics of a single real estate parcel, or a designated area. By way of example, statistical analysis can include one or more of an average, median and high/low value designation for: the size of a property within a given neighborhood, the size of a front yard, how much road frontage each property includes, the average size of the houses within the designated area, the number of stories comprising the buildings within the area, siding types of buildings within the area, the size of side yards between houses, height of trees within the area, and almost any other data that may be calculated from the image data sets and the three-dimensional models.
Referring now to FIG. 6 , a view of an alternative exemplary user interface 600 according to the present invention is illustrated. The interface 600 can include a portion with an aerial view 601 and a polygon designation of geographic area of interest, such as for example, a particular real estate parcel or neighborhood location. The interface may also include a portion with a two-dimensional continuum of image data 603A. The continuum of image data is preferably a composite of image data sets captured from disparate points. Another portion 603B can include image data sprayed over three-dimensional models 604 generated from point clouds of features captured by the image data sets.
Referring now to FIG. 7 , in some preferred embodiments, image data sprayed over the three-dimensional polygon formats includes a composite image formed by aligning two or more of the image data sets. Unlike stitching processes previously known, the present invention creates a composite through alignment of portions of data from more than one data set. Alignment can be accomplished in image data processing. Using image data processing, the images 701-703 are aligned to form a composite image 700. The composite image 700 is essentially two dimensional image data arranged as a second continuum, or ribbon. The second continuum includes ongoing image data 701-703 captured from the points defining the first continuum.
In some particular embodiments, the series of points of image capture in the first continuum includes positions of a vehicle carrying an image capture device, such as a camera, as the vehicle traverses a path proximate to a geographic area. The camera is positioned to capture image data of the geographic area. Image data 701-703 is periodically captured as the vehicle traverses the path. The motion of the vehicle, combined with the periodic capture of image data 701-703, thereby results in image data 701-703 being captured from disparate points along the first continuum.
A preferred embodiment includes capture of image data with a motion vector of the camera in space maintained generally orthogonal to a subject for which image data will be captured. Orthogonal capture of the image data can facilitate consistency for subsequent composite of portions of the image data captured. Therefore, data captured at an angle of between about 75° and 105° can provide data most easily assembled into a continuum of data. However, other angles may be used to create different effects. In addition, in some embodiments physical limitation may dictate the use of other angles. In such instances, post processing can be utilized to compensate for the angle of image data capture.
During image data processing, some or all of the images are aligned to form a composite image in the form of a continuous pictorial representation of the geographic area. One commercial embodiment of a continuous pictorial representation includes RibbonView™ by Real Data Systems. RibbonView™ correlates a ribbon of geographic image data with geospatial designations to facilitate identification of a particular geographic area, as discussed more fully below. In various embodiments, image capture processing can be accomplished in real time or via post image capture processing.
In some embodiments, select portions 704-707 of two or more sets of captured image data are aligned to generate the composite image 708. As illustrated, some preferred embodiments include vertical slices of data 704-707 aligned in a horizontal plane to form the composite image 708. Unlike a traditional photograph taken with a macro lens, according to the present invention, the length of a horizontal plane defining a composite image 708 is only limited by the length of a continuum along which points are defined and from which image data 704-707 is captured.
The use of only slices of data 704-707 from any particular captured image provides for a higher quality image 708. The quality is increased, for example, when a temporary obstruction, such as a passing car, person or animal, captured in one image data set, is only represented in a thin slice of a continuous ribbon 704-707 data. In addition, alignment of multiple thin slices of image data 704-707 is facilitated from the perspective of which aberrations typical human sensory is capable of distinguishing. Numerous thin slices 704-707 are perceived as a smoother transition across the horizontal plane defining the composite image 708 removing a choppiness that may be experienced with the use of larger data images.
The present invention can include a uniform width of each slice of data 704-707 or a variable width. The width of a particular slice may vary, for example, upon one or more of the velocity of a vehicle from which image data sets are captured, the sample rate of a camera used to capture an image data set 701-703, the resolution of a picture comprising an image data set 701-703 and the path of a camera. A high resolution image generated by a 2.1 mega pixel camera may have a 1600 by 1200 resolution and allow for a thinner slice 704-707 that includes a width of between about 5 to 700 pixels of an image data set. Embodiments with very high quality can include a slice 704-707 of between 1 to 2 pixels.
From a different perspective, some embodiments can include a slice 704-707 of an image data set 701-703 that includes a percentage of the image data set 701-703. Accordingly, some preferred embodiments can include a slice 704-707 of between about 5% to about 12% of an image data set. Other preferred embodiments can include up to about 50% of an image data set. However, it should be understood that some embodiments include a slice 704-707 that constitutes any fraction of the complete image data set.
It should be noted that although preferred embodiments may utilize vertical rectangular slices 704-709, the scope of the present invention is not limited by which portion of an image data set 701-703 is utilized to generate a composite image 270. Therefore, it is within the scope of the invention to use any portion of any image data set 701-703 to generate a composite image. Accordingly, slices of an image 701-703 other than vertical slices 704-709 may be apportioned and combined into a composite image 270. Slices may therefore include a slice angled at 60° or 75°, or other angle conducive to a particular application. In addition, it is also within the scope of the present invention to utilize irregular shaped portions of two or more image data sets 701-703 to generate a composite image 270.
In some embodiments, a database or other data processing mechanism, can track each portion or slice 704-708 utilized to construct a continuum and associate the slice 704-708 with an original image 701-703 from which the slice 704-708 is derived. User interactive devices can execute the retrieval of an entire original image 701-703 or a series of original images 701-703 upon request. In some instances, one or more original images 701-703 may provide detail not included in the composite image 708.
In some embodiments, a selected portion of an image data set may be determined by the physical attributes of the equipment used to capture an image data set. For example, a typical camera lens can impart some distortion to an image data set, as represented in the illustration by an elongated portion and a compressed portion. Utilizing only a portion of an image data set, such as, for example, a center portion vertical slice, can minimize the effect of distortion introduced by a lens, or other source, to a composite image. Distortion is minimized when the composite image is made to include those portions of the image data set corresponding with the center of the lens. Specific embodiments may likewise account for other aberrations that may be present in a set of image data.
It will be apparent to those schooled in the art that the length of a composite image generated according to the present invention is limited only by the ability to capture image data from additional points on a continuum and store the captured image data for post processing. The post processing allows for the alignment of portions of the image data compiled into a composite two-dimensional view that can continue so long as additional image data is made available to be added to it.
Apparatus
The teachings of the present invention may be implemented with any apparatus capable of embodying the innovative concepts described herein. Image capture can be accomplished, for example, via a digital camera capable of capturing 12 or more images per second. In addition, FIG. 8 illustrates a controller 800 that may be utilized to implement some embodiments of the present invention. The controller 800 comprises a processor unit 810, such as one or more processors, coupled to a communication device 820 configured to communicate via a communication network (not shown in FIG. 8 ). The communication device 820 may be used to communicate, for example, with one or more online devices, such as a personal computer, laptop or a handheld device.
The processor 810 is also in communication with a storage device 830. The storage device 830 may comprise any appropriate information storage device, including combinations of magnetic storage devices (e.g., magnetic tape and hard disk drives), optical storage devices, and/or semiconductor memory devices such as Random Access Memory (RAM) devices and Read Only Memory (ROM) devices.
The storage device 830 can store a program 840 for controlling the processor 810. The processor 810 performs instructions of the program 840, and thereby operates in accordance with the present invention. The processor 810 may also cause the communication device 820 to transmit information, including, in some instances, control commands to operate apparatus to implement the processes described above. The storage device 830 can additionally store related data in a database 830A and database 830B, as needed.
In some preferred embodiments, apparatus includes a video and data server farm. The video and data server farm includes at least one video storage server that stores video image files containing video drive-by data that corresponds to a geographic location, a database server that processes a data query received from a user over the Internet that corresponds to a geographic location of interest, and an image server. In operation, the database server identifies video image files stored in the video storage server that correspond to the geographic location of interest contained in the data query, and transfers the video image files over a pre-processing network to the image processing server. The image processing server converts the video drive-by data to post processed video data corresponding to a desired image format, and transfers the post processed video data via post-processing network to the Internet response to the query.
A landing zone server can also be included which receives the video drive-by data from a portable memory device and permits the viewing and analysis of the video drive-by data prior to storage in the video storage server. Still further, a map server is preferably provided to present a static image an overhead view of the geographic location of interest.
Embodiments can also include one or more servers described above included in one or more physical units. Each server does not need to be a disparate apparatus. Still other embodiments can include one or more or the servers described above included in multiple physical units. Some embodiments can even include a single server, as described which includes multiple physical apparatus units at disparate locations.
A number of embodiments of the present invention have been described. Nevertheless, it will be understood that various modifications may be made without departing from the spirit and scope of the invention. For example, various methods or equipment may be used to implement the process steps described herein or to create a device according to the inventive concepts provided above and further described in the claims. In addition, various integration of components, as well as software and firmware can be implemented. Accordingly, other embodiments are within the scope of the following claims.
Claims (25)
1. Apparatus for generating a three-dimensional model based upon two dimensional image data, the apparatus comprising:
a processor operatively attached to a data storage device and executable computer code stored upon the data storage device and executable upon demand to cause the processor to:
receive two or more images of a subject, wherein the two or more images comprise two dimensional image data captured from disparate points on a geospatial continuum and comprise portions aligned in a dimension consistent with the continuum;
generate a two dimensional composite image of the subject comprising the aligned portions of the two or more images;
create a human readable output comprising one or more three-dimensional models of the subject, the one or more three-dimensional models comprising one or more polygon based models generated from a point cloud array based upon structures included in the two dimensional composite image of the subject and sprayed with two dimensional image data of the subject captured from disparate points along the continuum.
2. The apparatus of claim 1 wherein the human readable output additionally comprises user interactive controls operative to traverse the three-dimensional model.
3. The apparatus of claim 1 wherein a selection of one or more three-dimensional models is presented based upon an indication of a geospatial location.
4. The apparatus of claim 3 wherein the geospatial location is indicated with a Cartesian coordinate.
5. The apparatus of claim 3 wherein the geospatial location is indicated with a latitude and longitude designation.
6. The apparatus of claim 1 wherein a selection of one or more three-dimensional models is presented based upon an indication of a street address.
7. The apparatus of claim 1 wherein a selection of one or more three-dimensional models is presented based upon an aerial view of a geographic area comprising polygon outline indicating a selected portion of the geographic area.
8. The apparatus of claim 3 wherein the software is additionally operative to provide a human readable output comprising user interactive devices operable to allow a user to traverse the three-dimensional model in one or more of three directions including: a) an essentially up and down vertical movement; b) an essentially side-to-side horizontal movement; and c) a depth movement into and out of the three-dimensional model.
9. The apparatus of claim 3 wherein the image data and three-dimensional polygon based models presented to a user at a given instance are based upon a relative position selected by the user within the user interface and a response to a command issued via a user interactive device.
10. The apparatus of claim 3 wherein the software is additionally operative to provide at least one entry of metadata relating to a given geographic area within view.
11. The apparatus of claim 3 wherein specific image data captured of a particular geographic area is presented sprayed over three-dimensional structures generated from two-dimensional image data captured from that specific location.
12. The apparatus of claim 3 wherein the software is additionally operative to provide an indication of cellular phone signal strength.
13. The apparatus of claim 3 wherein the software is additionally operative to provide an indication of a wireless network data signal strength.
14. The apparatus of claim 3 wherein the software is additionally operative to provide an indication of a global position system satellite data and a signal strength.
15. The apparatus of claim 3 wherein the software is additionally operative to provide an indication of a radio signal strength.
16. The apparatus of claim 3 wherein the software is additionally operative to provide an indication of a school district data encompassing a designated geospatial area.
17. The apparatus of claim 3 wherein the software is additionally operative to provide an indication of an average yard size included in a designated geospatial area.
18. The apparatus of claim 3 wherein the software is additionally operative to provide price range for real estate parcels included in a designated geographic area.
19. The apparatus of claim 3 wherein the software is additionally operative to provide one or more of: building features; proximity to amenities; political boundaries; and zoning restrictions.
20. A method of generating a three-dimensional model of a geographic subject area, the method comprising:
capturing image data of a subject with an image capture device from disparate points along a continuum; wherein the image data comprise two or more images comprise two dimensional image data captured from disparate points on a geospatial continuum and comprise portions aligned in a dimension consistent with the continuum;
generate a composite image of the subject comprising the aligned portions of the two or more images;
generating with a computerized system a point cloud array based upon a structure withing the composite image;
generating a three-dimensional model of an object included on the geographic area, wherein the three-dimensional model comprises one or more polygon based models generated from the point cloud array; and
spraying at least a portion of the image data captured on the three dimensional model.
21. The method of claim 20 wherein the spraying of the image data captured on the three-dimensional model creates a human readable output.
22. The method of claim 21 additionally comprising the step of creating a composite of image data comprising a portion of multiple disparate image data sets and spraying the composite image data on the subject.
23. The method of claim 22 additionally comprising the step of generating a user interactive control operative to traverse the three-dimensional model.
24. The method of claim 22 additionally comprising the steps of:
generating a selection of one or more three-dimensional models; and
presented selection of one or more three-dimensional models based upon an indication of geospatial location.
25. The method of claim 24 wherein the geospatial location is indicated with a Cartesian coordinate.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/861,336 USRE45264E1 (en) | 2004-08-31 | 2013-04-11 | Methods and apparatus for generating three-dimensional image data models |
Applications Claiming Priority (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US60549804P | 2004-08-31 | 2004-08-31 | |
US11/216,465 US7389181B2 (en) | 2004-08-31 | 2005-08-31 | Apparatus and method for producing video drive-by data corresponding to a geographic location |
US11/702,707 US20080189031A1 (en) | 2007-02-06 | 2007-02-06 | Methods and apparatus for presenting a continuum of image data |
US11/702,708 US7929800B2 (en) | 2007-02-06 | 2007-02-06 | Methods and apparatus for generating a continuum of image data |
US90295307P | 2007-02-22 | 2007-02-22 | |
US12/036,197 US8207964B1 (en) | 2008-02-22 | 2008-02-22 | Methods and apparatus for generating three-dimensional image data models |
US13/861,336 USRE45264E1 (en) | 2004-08-31 | 2013-04-11 | Methods and apparatus for generating three-dimensional image data models |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/036,197 Reissue US8207964B1 (en) | 2004-08-31 | 2008-02-22 | Methods and apparatus for generating three-dimensional image data models |
Publications (1)
Publication Number | Publication Date |
---|---|
USRE45264E1 true USRE45264E1 (en) | 2014-12-02 |
Family
ID=46272907
Family Applications (18)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/036,197 Ceased US8207964B1 (en) | 2004-08-31 | 2008-02-22 | Methods and apparatus for generating three-dimensional image data models |
US13/481,912 Expired - Fee Related US9384277B2 (en) | 2004-08-31 | 2012-05-28 | Three dimensional image data models |
US13/791,957 Abandoned US20130191359A1 (en) | 2004-08-31 | 2013-03-09 | Method and Apparatus of Receiving a List of Links to Street View Data of Real Estate Properties |
US13/791,943 Expired - Fee Related US9311396B2 (en) | 2004-08-31 | 2013-03-09 | Method of providing street view data of a real estate property |
US13/791,938 Abandoned US20130201340A1 (en) | 2004-08-31 | 2013-03-09 | Method of viewing street View Data of a Real Estate Property |
US13/791,951 Abandoned US20130191725A1 (en) | 2004-08-31 | 2013-03-09 | Method and Apparatus of Viewing Street View Data of a Real Estate Property |
US13/791,947 Abandoned US20130191295A1 (en) | 2004-08-31 | 2013-03-09 | Method of Receiving a List of Links to Street View Data of Real Estate Properties |
US13/791,945 Abandoned US20130188052A1 (en) | 2004-08-31 | 2013-03-09 | Method of Providing Street View Data of a Comparable Real Estate Property |
US13/791,955 Expired - Fee Related US8890866B2 (en) | 2004-08-31 | 2013-03-09 | Method and apparatus of providing street view data of a comparable real estate property |
US13/791,961 Expired - Fee Related US8558848B2 (en) | 2004-08-31 | 2013-03-09 | Wireless internet-accessible drive-by street view system and method |
US13/791,953 Expired - Fee Related US9311397B2 (en) | 2004-08-31 | 2013-03-09 | Method and apparatus of providing street view data of a real estate property |
US13/795,959 Abandoned US20130188008A1 (en) | 2004-08-31 | 2013-03-12 | Generating Video Drive-By Data Corresponding to a Geographic Location and Related Three Dimensional Models |
US13/796,544 Expired - Fee Related US8902226B2 (en) | 2004-08-31 | 2013-03-12 | Method for using drive-by image data to generate a valuation report of a selected real estate property |
US13/830,164 Abandoned US20130201341A1 (en) | 2004-08-31 | 2013-03-14 | Generating a Point of Interest Database with Specific Geographic Locations Derived from Street-Level Image Data |
US13/861,336 Expired - Fee Related USRE45264E1 (en) | 2004-08-31 | 2013-04-11 | Methods and apparatus for generating three-dimensional image data models |
US14/023,028 Abandoned US20140019301A1 (en) | 2004-08-31 | 2013-09-10 | Internet-accessible drive-by street view system and method |
US14/243,712 Abandoned US20140211015A1 (en) | 2004-08-31 | 2014-04-02 | Method and apparatus for providing image data of property |
US15/200,784 Abandoned US20170094165A1 (en) | 2004-08-31 | 2016-07-01 | Method and apparatus for providing image data of property |
Family Applications Before (14)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/036,197 Ceased US8207964B1 (en) | 2004-08-31 | 2008-02-22 | Methods and apparatus for generating three-dimensional image data models |
US13/481,912 Expired - Fee Related US9384277B2 (en) | 2004-08-31 | 2012-05-28 | Three dimensional image data models |
US13/791,957 Abandoned US20130191359A1 (en) | 2004-08-31 | 2013-03-09 | Method and Apparatus of Receiving a List of Links to Street View Data of Real Estate Properties |
US13/791,943 Expired - Fee Related US9311396B2 (en) | 2004-08-31 | 2013-03-09 | Method of providing street view data of a real estate property |
US13/791,938 Abandoned US20130201340A1 (en) | 2004-08-31 | 2013-03-09 | Method of viewing street View Data of a Real Estate Property |
US13/791,951 Abandoned US20130191725A1 (en) | 2004-08-31 | 2013-03-09 | Method and Apparatus of Viewing Street View Data of a Real Estate Property |
US13/791,947 Abandoned US20130191295A1 (en) | 2004-08-31 | 2013-03-09 | Method of Receiving a List of Links to Street View Data of Real Estate Properties |
US13/791,945 Abandoned US20130188052A1 (en) | 2004-08-31 | 2013-03-09 | Method of Providing Street View Data of a Comparable Real Estate Property |
US13/791,955 Expired - Fee Related US8890866B2 (en) | 2004-08-31 | 2013-03-09 | Method and apparatus of providing street view data of a comparable real estate property |
US13/791,961 Expired - Fee Related US8558848B2 (en) | 2004-08-31 | 2013-03-09 | Wireless internet-accessible drive-by street view system and method |
US13/791,953 Expired - Fee Related US9311397B2 (en) | 2004-08-31 | 2013-03-09 | Method and apparatus of providing street view data of a real estate property |
US13/795,959 Abandoned US20130188008A1 (en) | 2004-08-31 | 2013-03-12 | Generating Video Drive-By Data Corresponding to a Geographic Location and Related Three Dimensional Models |
US13/796,544 Expired - Fee Related US8902226B2 (en) | 2004-08-31 | 2013-03-12 | Method for using drive-by image data to generate a valuation report of a selected real estate property |
US13/830,164 Abandoned US20130201341A1 (en) | 2004-08-31 | 2013-03-14 | Generating a Point of Interest Database with Specific Geographic Locations Derived from Street-Level Image Data |
Family Applications After (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/023,028 Abandoned US20140019301A1 (en) | 2004-08-31 | 2013-09-10 | Internet-accessible drive-by street view system and method |
US14/243,712 Abandoned US20140211015A1 (en) | 2004-08-31 | 2014-04-02 | Method and apparatus for providing image data of property |
US15/200,784 Abandoned US20170094165A1 (en) | 2004-08-31 | 2016-07-01 | Method and apparatus for providing image data of property |
Country Status (1)
Country | Link |
---|---|
US (18) | US8207964B1 (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9230167B2 (en) | 2011-12-08 | 2016-01-05 | The Nielsen Company (Us), Llc. | Methods, apparatus, and articles of manufacture to measure geographical features using an image of a geographical location |
US9378509B2 (en) | 2012-05-09 | 2016-06-28 | The Nielsen Company (Us), Llc | Methods, apparatus, and articles of manufacture to measure geographical features using an image of a geographical location |
US9547866B2 (en) | 2013-03-14 | 2017-01-17 | The Nielsen Company (Us), Llc | Methods and apparatus to estimate demography based on aerial images |
US10885097B2 (en) | 2015-09-25 | 2021-01-05 | The Nielsen Company (Us), Llc | Methods and apparatus to profile geographic areas of interest |
US11470303B1 (en) | 2010-06-24 | 2022-10-11 | Steven M. Hoffberg | Two dimensional to three dimensional moving image converter |
Families Citing this family (200)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7949191B1 (en) * | 2007-04-04 | 2011-05-24 | A9.Com, Inc. | Method and system for searching for information on a network in response to an image query sent by a user from a mobile communications device |
US9953459B2 (en) | 2008-11-05 | 2018-04-24 | Hover Inc. | Computer vision database platform for a three-dimensional mapping system |
US8606774B1 (en) * | 2009-05-18 | 2013-12-10 | Google Inc. | Methods and systems for 3D shape retrieval |
US8611592B2 (en) * | 2009-08-26 | 2013-12-17 | Apple Inc. | Landmark identification using metadata |
US9286624B2 (en) * | 2009-09-10 | 2016-03-15 | Google Inc. | System and method of displaying annotations on geographic object surfaces |
KR20110052124A (en) * | 2009-11-12 | 2011-05-18 | 삼성전자주식회사 | Method for generating and referencing panorama image and mobile terminal using the same |
US9749597B2 (en) * | 2010-08-10 | 2017-08-29 | International Business Machines Corporation | Precise target positioning in geographical imaging |
US9317133B2 (en) * | 2010-10-08 | 2016-04-19 | Nokia Technologies Oy | Method and apparatus for generating augmented reality content |
US20120150573A1 (en) * | 2010-12-13 | 2012-06-14 | Omar Soubra | Real-time site monitoring design |
US8810598B2 (en) | 2011-04-08 | 2014-08-19 | Nant Holdings Ip, Llc | Interference based augmented reality hosting platforms |
US9552376B2 (en) | 2011-06-09 | 2017-01-24 | MemoryWeb, LLC | Method and apparatus for managing digital files |
US9092674B2 (en) * | 2011-06-23 | 2015-07-28 | International Business Machines Corportion | Method for enhanced location based and context sensitive augmented reality translation |
US9501856B2 (en) * | 2012-02-13 | 2016-11-22 | Nokia Technologies Oy | Method and apparatus for generating panoramic maps with elements of subtle movement |
US9501700B2 (en) | 2012-02-15 | 2016-11-22 | Xactware Solutions, Inc. | System and method for construction estimation using aerial images |
EP2820576A4 (en) | 2012-02-29 | 2015-11-18 | Nokia Technologies Oy | Method and apparatus for rendering items in a user interface |
US10296516B2 (en) * | 2012-05-21 | 2019-05-21 | Here Global B.V. | Method and apparatus for navigation using multiple synchronized mobile devices |
US8923893B2 (en) | 2012-08-07 | 2014-12-30 | Symbol Technologies, Inc. | Real-time planogram generation and maintenance |
US9154805B2 (en) * | 2012-09-12 | 2015-10-06 | Advanced Micro Devices, Inc. | Video and image compression based on position of the image generating device |
US9488489B2 (en) | 2012-09-28 | 2016-11-08 | Google Inc. | Personalized mapping with photo tours |
US10771748B2 (en) * | 2012-11-27 | 2020-09-08 | Metropolitan Life Insurance Co. | System and method for interactive aerial imaging |
US9071756B2 (en) | 2012-12-11 | 2015-06-30 | Facebook, Inc. | Systems and methods for digital video stabilization via constraint-based rotation smoothing |
US10924627B2 (en) * | 2012-12-31 | 2021-02-16 | Virtually Anywhere | Content management for virtual tours |
US10196850B2 (en) | 2013-01-07 | 2019-02-05 | WexEnergy LLC | Frameless supplemental window for fenestration |
US9691163B2 (en) * | 2013-01-07 | 2017-06-27 | Wexenergy Innovations Llc | System and method of measuring distances related to an object utilizing ancillary objects |
US9845636B2 (en) | 2013-01-07 | 2017-12-19 | WexEnergy LLC | Frameless supplemental window for fenestration |
US10883303B2 (en) | 2013-01-07 | 2021-01-05 | WexEnergy LLC | Frameless supplemental window for fenestration |
US20140253540A1 (en) * | 2013-03-07 | 2014-09-11 | Yoav DORI | Method and system of incorporating real world objects into a virtual environment |
US9753950B2 (en) * | 2013-03-15 | 2017-09-05 | Pictometry International Corp. | Virtual property reporting for automatic structure detection |
US8756085B1 (en) * | 2013-03-15 | 2014-06-17 | State Farm Mutual Automobile Insurance Company | Systems and methods for assessing property damage |
US9756138B2 (en) * | 2013-04-08 | 2017-09-05 | Here Global B.V. | Desktop application synchronization to process data captured on a mobile device |
US10134194B2 (en) * | 2013-07-17 | 2018-11-20 | Evernote Corporation | Marking up scenes using a wearable augmented reality device |
US11721066B2 (en) | 2013-07-23 | 2023-08-08 | Hover Inc. | 3D building model materials auto-populator |
US9679227B2 (en) | 2013-08-02 | 2017-06-13 | Xactware Solutions, Inc. | System and method for detecting features in aerial images using disparity mapping and segmentation techniques |
US9903719B2 (en) * | 2013-09-03 | 2018-02-27 | Litel Instruments | System and method for advanced navigation |
US9208171B1 (en) * | 2013-09-05 | 2015-12-08 | Google Inc. | Geographically locating and posing images in a large-scale image repository and processing framework |
US9244940B1 (en) | 2013-09-27 | 2016-01-26 | Google Inc. | Navigation paths for panorama |
US9288368B2 (en) * | 2013-10-08 | 2016-03-15 | Delightfit, Inc. | Video and map data synchronization for simulated athletic training |
US9582516B2 (en) | 2013-10-17 | 2017-02-28 | Nant Holdings Ip, Llc | Wide area augmented reality location-based services |
US9836885B1 (en) | 2013-10-25 | 2017-12-05 | Appliance Computing III, Inc. | Image-based rendering of real spaces |
US11004259B2 (en) | 2013-10-25 | 2021-05-11 | Hover Inc. | Estimating dimensions of geo-referenced ground level imagery using orthogonal imagery |
CN103914521B (en) * | 2014-03-20 | 2015-02-25 | 北京建筑大学 | Street view image storage method and device based on mixed tile pyramids |
US9640223B2 (en) * | 2014-03-27 | 2017-05-02 | Tvu Networks Corporation | Methods, apparatus and systems for time-based and geographic navigation of video content |
US10327034B2 (en) | 2014-03-27 | 2019-06-18 | Tvu Networks Corporation | Methods, apparatus and systems for exchange of video content |
US9992246B2 (en) | 2014-03-27 | 2018-06-05 | Tvu Networks Corporation | Methods, apparatus, and systems for instantly sharing video content on social media |
US10062118B1 (en) | 2014-04-02 | 2018-08-28 | Liberty Mutual Insurance Company | Concepts for providing an insurance quote |
USD781317S1 (en) | 2014-04-22 | 2017-03-14 | Google Inc. | Display screen with graphical user interface or portion thereof |
US9934222B2 (en) | 2014-04-22 | 2018-04-03 | Google Llc | Providing a thumbnail image that follows a main image |
USD781318S1 (en) | 2014-04-22 | 2017-03-14 | Google Inc. | Display screen with graphical user interface or portion thereof |
USD780777S1 (en) | 2014-04-22 | 2017-03-07 | Google Inc. | Display screen with graphical user interface or portion thereof |
US9972121B2 (en) | 2014-04-22 | 2018-05-15 | Google Llc | Selecting time-distributed panoramic images for display |
US9189839B1 (en) | 2014-04-24 | 2015-11-17 | Google Inc. | Automatically generating panorama tours |
US10019761B2 (en) * | 2014-04-25 | 2018-07-10 | State Farm Mutual Automobile Insurance Company | System and method for virtual inspection of a structure |
US10248731B1 (en) * | 2014-05-16 | 2019-04-02 | Corelogic Solutions, Llc | System and method for linking data records for parcels |
US9245183B2 (en) | 2014-06-26 | 2016-01-26 | International Business Machines Corporation | Geographical area condition determination |
US9002647B1 (en) | 2014-06-27 | 2015-04-07 | Google Inc. | Generating turn-by-turn direction previews |
US9418472B2 (en) | 2014-07-17 | 2016-08-16 | Google Inc. | Blending between street view and earth view |
US20160042478A1 (en) * | 2014-08-05 | 2016-02-11 | Mastercard International Incorporated | Methods and Systems for Verifying Images Associated With Offered Properties |
US9724177B2 (en) | 2014-08-19 | 2017-08-08 | Align Technology, Inc. | Viewfinder with real-time tracking for intraoral scanning |
US9870596B2 (en) | 2014-09-18 | 2018-01-16 | International Business Machines Corporation | Predicting community development trends |
CN104361586A (en) * | 2014-10-31 | 2015-02-18 | 国网上海市电力公司 | Electric pole and tower maintenance and pre-warning system based on panoramas |
US9805061B2 (en) * | 2014-11-18 | 2017-10-31 | International Business Machines Corporation | Image search for a location |
US9830724B2 (en) * | 2014-11-21 | 2017-11-28 | Mastercard International Incorporated | Detecting real estate vacancies using satellite imagery |
US9689976B2 (en) | 2014-12-19 | 2017-06-27 | Xidrone Systems, Inc. | Deterent for unmanned aerial systems |
US9715009B1 (en) | 2014-12-19 | 2017-07-25 | Xidrone Systems, Inc. | Deterent for unmanned aerial systems |
US9851870B2 (en) * | 2015-03-17 | 2017-12-26 | Raytheon Company | Multi-dimensional video navigation system and method using interactive map paths |
US9924240B2 (en) | 2015-05-01 | 2018-03-20 | Google Llc | Systems and methods for interactive video generation and rendering |
US20170134815A1 (en) * | 2015-05-06 | 2017-05-11 | Joshua LOMELINO | Systems for creating and distributing videos of real estate |
JP6622990B2 (en) * | 2015-06-29 | 2019-12-18 | オリンパス株式会社 | Information discrimination device, information discrimination method, and information discrimination program |
US10498955B2 (en) | 2015-08-03 | 2019-12-03 | Disney Enterprises, Inc. | Commercial drone detection |
CA2995850A1 (en) * | 2015-08-31 | 2017-03-09 | Ryan Kottenstette | Systems and methods for analyzing remote sensing imagery |
US10220172B2 (en) | 2015-11-25 | 2019-03-05 | Resmed Limited | Methods and systems for providing interface components for respiratory therapy |
US9972125B2 (en) * | 2015-12-16 | 2018-05-15 | Google Inc. | Split tile map rendering |
CN105513127B (en) * | 2015-12-25 | 2018-01-26 | 武汉大学 | Shaft regularization three-dimensional modeling method and system based on density peaks cluster |
US10352689B2 (en) | 2016-01-28 | 2019-07-16 | Symbol Technologies, Llc | Methods and systems for high precision locationing with depth values |
US10839569B2 (en) * | 2016-02-03 | 2020-11-17 | NorthStar Memorial Group LLC | System for geospatial mapping of cemetery properties |
CN105677899B (en) * | 2016-02-04 | 2019-01-29 | 东南大学 | A kind of enhanced vector numerical map production method comprising road direction of travel |
US10204116B2 (en) * | 2016-02-08 | 2019-02-12 | Corelogic Solutions, Llc | Image metadata enrichment generator |
US10147224B2 (en) * | 2016-02-16 | 2018-12-04 | Samsung Electronics Co., Ltd. | Method and apparatus for generating omni media texture mapping metadata |
US10824320B2 (en) * | 2016-03-07 | 2020-11-03 | Facebook, Inc. | Systems and methods for presenting content |
EP3430591A4 (en) * | 2016-03-16 | 2019-11-27 | ADCOR Magnet Systems, LLC | System for georeferenced, geo-oriented real time video streams |
US10341617B2 (en) * | 2016-03-23 | 2019-07-02 | Purdue Research Foundation | Public safety camera identification and monitoring system and method |
US10506201B2 (en) * | 2016-03-23 | 2019-12-10 | Purdue Research Foundation | Public safety camera identification and monitoring system and method |
US10368037B2 (en) * | 2016-03-23 | 2019-07-30 | Purdue Research Foundation | Public safety camera monitoring system and method |
US10410441B2 (en) * | 2016-05-16 | 2019-09-10 | Wi-Tronix, Llc | Real-time data acquisition and recording system viewer |
CN106055610B (en) * | 2016-05-25 | 2020-02-14 | 维沃移动通信有限公司 | Voice information retrieval method and mobile terminal |
US10198955B1 (en) | 2016-09-08 | 2019-02-05 | Amazon Technologies, Inc. | Drone marker and landing zone verification |
US10121117B1 (en) | 2016-09-08 | 2018-11-06 | Amazon Technologies, Inc. | Drone location signature filters |
US10049589B1 (en) | 2016-09-08 | 2018-08-14 | Amazon Technologies, Inc. | Obstacle awareness based guidance to clear landing space |
US10643089B2 (en) | 2016-10-13 | 2020-05-05 | Ricoh Company, Ltd. | Information processing system to obtain and manage images of a property |
US10706839B1 (en) | 2016-10-24 | 2020-07-07 | United Services Automobile Association (Usaa) | Electronic signatures via voice for virtual assistants' interactions |
US11042161B2 (en) | 2016-11-16 | 2021-06-22 | Symbol Technologies, Llc | Navigation control method and apparatus in a mobile automation system |
JP6116746B1 (en) * | 2016-11-17 | 2017-04-19 | 株式会社ネクスト | Information processing apparatus, information processing method, and program |
CN106713847A (en) * | 2016-11-28 | 2017-05-24 | 天津商企生产力促进有限公司 | Electromechanical integrated monitor based on virtual three-dimensional static scene |
US20180150924A1 (en) * | 2016-11-30 | 2018-05-31 | Corelogic Solutions, Llc | Computer-based photo re-use across property reports |
US20180150923A1 (en) * | 2016-11-30 | 2018-05-31 | Corelogic Solutions, Llc | Property study workflow system |
US20180150925A1 (en) * | 2016-11-30 | 2018-05-31 | Corelogic Solutions, Llc | Computer-based control of timing with which property review orders are processed |
CN108462851A (en) * | 2017-02-21 | 2018-08-28 | 北京视联动力国际信息技术有限公司 | The method and mobile terminal of data processing, server, the second mobile terminal |
US10360709B2 (en) * | 2017-04-05 | 2019-07-23 | Microsoft Technology Licensing, Llc | Rendering images on map using orientations |
GB2561882A (en) * | 2017-04-27 | 2018-10-31 | Nokia Technologies Oy | Virtual reality viewport adaption |
US11367092B2 (en) | 2017-05-01 | 2022-06-21 | Symbol Technologies, Llc | Method and apparatus for extracting and processing price text from an image set |
AU2018261257B2 (en) | 2017-05-01 | 2020-10-08 | Symbol Technologies, Llc | Method and apparatus for object status detection |
US10949798B2 (en) | 2017-05-01 | 2021-03-16 | Symbol Technologies, Llc | Multimodal localization and mapping for a mobile automation apparatus |
US10726273B2 (en) | 2017-05-01 | 2020-07-28 | Symbol Technologies, Llc | Method and apparatus for shelf feature and object placement detection from shelf images |
US10663590B2 (en) | 2017-05-01 | 2020-05-26 | Symbol Technologies, Llc | Device and method for merging lidar data |
US11449059B2 (en) | 2017-05-01 | 2022-09-20 | Symbol Technologies, Llc | Obstacle detection for a mobile automation apparatus |
US11093896B2 (en) | 2017-05-01 | 2021-08-17 | Symbol Technologies, Llc | Product status detection system |
US10505057B2 (en) | 2017-05-01 | 2019-12-10 | Symbol Technologies, Llc | Device and method for operating cameras and light sources wherein parasitic reflections from a paired light source are not reflected into the paired camera |
US10591918B2 (en) | 2017-05-01 | 2020-03-17 | Symbol Technologies, Llc | Fixed segmented lattice planning for a mobile automation apparatus |
US10628959B2 (en) | 2017-05-03 | 2020-04-21 | International Business Machines Corporation | Location determination using street view images |
WO2018201423A1 (en) | 2017-05-05 | 2018-11-08 | Symbol Technologies, Llc | Method and apparatus for detecting and interpreting price label text |
CN108882018B (en) | 2017-05-09 | 2020-10-20 | 阿里巴巴(中国)有限公司 | Video playing and data providing method in virtual scene, client and server |
US10984182B2 (en) * | 2017-05-12 | 2021-04-20 | Loveland Innovations, LLC | Systems and methods for context-rich annotation and report generation for UAV microscan data |
US10533364B2 (en) | 2017-05-30 | 2020-01-14 | WexEnergy LLC | Frameless supplemental window for fenestration |
JP7199627B2 (en) | 2017-05-31 | 2023-01-06 | ウーブン プラネット ノース アメリカ,インコーポレイティド | A vehicle system equipped with imaging equipment for high-definition near-real-time map generation |
US11272090B2 (en) * | 2017-06-02 | 2022-03-08 | L-Tron Corporation | Event data and location-linked spherical imaging system |
US10467804B2 (en) | 2017-06-29 | 2019-11-05 | Open Space Labs, Inc. | Automated spatial indexing of images based on floorplan features |
WO2019033090A1 (en) * | 2017-08-11 | 2019-02-14 | Teluss Title Company | Distributed ledger based system and method for the settlement and transfer of title to real estate |
US10460465B2 (en) * | 2017-08-31 | 2019-10-29 | Hover Inc. | Method for generating roof outlines from lateral images |
US10572763B2 (en) | 2017-09-07 | 2020-02-25 | Symbol Technologies, Llc | Method and apparatus for support surface edge detection |
US10521914B2 (en) | 2017-09-07 | 2019-12-31 | Symbol Technologies, Llc | Multi-sensor object recognition system and method |
CN107590854A (en) * | 2017-09-15 | 2018-01-16 | 福建四创软件有限公司 | Reservoir region three-dimensional live methods of exhibiting based on WEBGIS |
US11200513B2 (en) | 2017-10-13 | 2021-12-14 | Carrier Corporation | Real estate image analysis |
US10936176B1 (en) * | 2017-11-17 | 2021-03-02 | United Services Automobile Association (Usaa) | Systems and methods for interactive maps |
RU2670429C1 (en) * | 2017-11-24 | 2018-10-23 | ООО "Ай Ти Ви групп" | Systems and methods of tracking moving objects on video image |
WO2019107536A1 (en) * | 2017-11-30 | 2019-06-06 | 三菱電機株式会社 | Three-dimensional map generating system, three-dimensional map generating method, and three-dimensional map generating program |
CN107958441B (en) * | 2017-12-01 | 2021-02-12 | 深圳市科比特航空科技有限公司 | Image splicing method and device, computer equipment and storage medium |
US10907940B1 (en) | 2017-12-12 | 2021-02-02 | Xidrone Systems, Inc. | Deterrent for unmanned aerial systems using data mining and/or machine learning for improved target detection and classification |
CN108182218B (en) * | 2017-12-21 | 2021-11-19 | 深圳先进技术研究院 | Video character recognition method and system based on geographic information system and electronic equipment |
US10791268B2 (en) | 2018-02-07 | 2020-09-29 | Structionsite Inc. | Construction photograph integration with 3D model images |
US10339384B2 (en) | 2018-02-07 | 2019-07-02 | Structionsite Inc. | Construction photograph integration with 3D model images |
CN108600607A (en) * | 2018-03-13 | 2018-09-28 | 上海网罗电子科技有限公司 | A kind of fire-fighting panoramic information methods of exhibiting based on unmanned plane |
CN108509191A (en) * | 2018-03-16 | 2018-09-07 | 杭州聚秀科技有限公司 | A kind of housing property management WEB graphic methods |
US10740911B2 (en) | 2018-04-05 | 2020-08-11 | Symbol Technologies, Llc | Method, system and apparatus for correcting translucency artifacts in data representing a support structure |
US11327504B2 (en) | 2018-04-05 | 2022-05-10 | Symbol Technologies, Llc | Method, system and apparatus for mobile automation apparatus localization |
US10809078B2 (en) | 2018-04-05 | 2020-10-20 | Symbol Technologies, Llc | Method, system and apparatus for dynamic path generation |
US10823572B2 (en) | 2018-04-05 | 2020-11-03 | Symbol Technologies, Llc | Method, system and apparatus for generating navigational data |
US10832436B2 (en) | 2018-04-05 | 2020-11-10 | Symbol Technologies, Llc | Method, system and apparatus for recovering label positions |
US10467758B1 (en) | 2018-07-13 | 2019-11-05 | Structionsite Inc. | Imagery-based construction progress tracking |
US10864860B2 (en) * | 2018-07-24 | 2020-12-15 | Black Sesame International Holding Limited | Model-based method for 360 degree surround view using cameras and radars mounted around a vehicle |
CN108965718B (en) * | 2018-08-03 | 2021-03-23 | 北京微播视界科技有限公司 | Image generation method and device |
CN109446281B (en) * | 2018-09-05 | 2020-11-24 | 广州市城市规划勘测设计研究院 | ArcGIS-based online intelligent mapping system construction method, system and mapping method |
CN109389053B (en) * | 2018-09-20 | 2021-08-06 | 同济大学 | Method and system for detecting position information of vehicle to be detected around target vehicle |
TWI734037B (en) * | 2018-09-28 | 2021-07-21 | 愛探極溫度行銷有限公司 | Real estate holding and inheritance calculation management system |
US11506483B2 (en) | 2018-10-05 | 2022-11-22 | Zebra Technologies Corporation | Method, system and apparatus for support structure depth determination |
US11010920B2 (en) | 2018-10-05 | 2021-05-18 | Zebra Technologies Corporation | Method, system and apparatus for object detection in point clouds |
EP3881163A4 (en) | 2018-11-12 | 2022-08-03 | Open Space Labs, Inc. | Automated spatial indexing of images to video |
US11090811B2 (en) | 2018-11-13 | 2021-08-17 | Zebra Technologies Corporation | Method and apparatus for labeling of support structures |
US11003188B2 (en) | 2018-11-13 | 2021-05-11 | Zebra Technologies Corporation | Method, system and apparatus for obstacle handling in navigational path generation |
US11210552B2 (en) | 2018-11-14 | 2021-12-28 | Cape Analytics, Inc. | Systems, methods, and computer readable media for predictive analytics and change detection from remotely sensed imagery |
US11079240B2 (en) | 2018-12-07 | 2021-08-03 | Zebra Technologies Corporation | Method, system and apparatus for adaptive particle filter localization |
US11416000B2 (en) | 2018-12-07 | 2022-08-16 | Zebra Technologies Corporation | Method and apparatus for navigational ray tracing |
US11100303B2 (en) | 2018-12-10 | 2021-08-24 | Zebra Technologies Corporation | Method, system and apparatus for auxiliary label detection and association |
US11015938B2 (en) | 2018-12-12 | 2021-05-25 | Zebra Technologies Corporation | Method, system and apparatus for navigational assistance |
US10731970B2 (en) | 2018-12-13 | 2020-08-04 | Zebra Technologies Corporation | Method, system and apparatus for support structure detection |
CN109740867B (en) * | 2018-12-14 | 2021-06-11 | 广州供电局有限公司 | Large-area power failure emergency repair scheme configuration method and device based on situation analysis |
CA3028708A1 (en) | 2018-12-28 | 2020-06-28 | Zih Corp. | Method, system and apparatus for dynamic loop closure in mapping trajectories |
US10936178B2 (en) | 2019-01-07 | 2021-03-02 | MemoryWeb, LLC | Systems and methods for analyzing and organizing digital photos and videos |
CN109474645A (en) * | 2019-01-29 | 2019-03-15 | 江苏省测绘工程院 | A kind of application method of multi-source encryption GPS data |
US10762690B1 (en) | 2019-02-11 | 2020-09-01 | Apple Inc. | Simulated overhead perspective images with removal of obstructions |
CN109918466A (en) * | 2019-03-08 | 2019-06-21 | 江西憶源多媒体科技有限公司 | A kind of real-time map information overall situation rendering method based on video analysis |
USD949159S1 (en) * | 2019-06-02 | 2022-04-19 | Apple Inc. | Display screen or portion thereof with graphical user interface |
US11960286B2 (en) | 2019-06-03 | 2024-04-16 | Zebra Technologies Corporation | Method, system and apparatus for dynamic task sequencing |
US11662739B2 (en) | 2019-06-03 | 2023-05-30 | Zebra Technologies Corporation | Method, system and apparatus for adaptive ceiling-based localization |
US11151743B2 (en) | 2019-06-03 | 2021-10-19 | Zebra Technologies Corporation | Method, system and apparatus for end of aisle detection |
US11080566B2 (en) | 2019-06-03 | 2021-08-03 | Zebra Technologies Corporation | Method, system and apparatus for gap detection in support structures with peg regions |
US11402846B2 (en) | 2019-06-03 | 2022-08-02 | Zebra Technologies Corporation | Method, system and apparatus for mitigating data capture light leakage |
US11341663B2 (en) | 2019-06-03 | 2022-05-24 | Zebra Technologies Corporation | Method, system and apparatus for detecting support structure obstructions |
US11200677B2 (en) | 2019-06-03 | 2021-12-14 | Zebra Technologies Corporation | Method, system and apparatus for shelf edge detection |
CN110599583B (en) * | 2019-07-26 | 2022-03-18 | 深圳眸瞳科技有限公司 | Unmanned aerial vehicle flight trajectory generation method and device, computer equipment and storage medium |
US11971871B1 (en) * | 2019-09-17 | 2024-04-30 | Sketch Align, Inc. | System controlled by data bearing records |
EP4032312A4 (en) * | 2019-09-18 | 2023-09-13 | Lopez, Camilo | System and method for generating a video |
US20210133874A1 (en) | 2019-11-04 | 2021-05-06 | Tellus App, Inc. | Architecture for exchange, publication, and distributed investment of property-backed vehicles |
US11507103B2 (en) | 2019-12-04 | 2022-11-22 | Zebra Technologies Corporation | Method, system and apparatus for localization-based historical obstacle handling |
US11094113B2 (en) | 2019-12-04 | 2021-08-17 | Geomni, Inc. | Systems and methods for modeling structures using point clouds derived from stereoscopic image pairs |
US11107238B2 (en) | 2019-12-13 | 2021-08-31 | Zebra Technologies Corporation | Method, system and apparatus for detecting item facings |
US11776140B2 (en) * | 2019-12-16 | 2023-10-03 | Insurance Services Office, Inc. | Computer vision systems and methods for automatic alignment of parcels with geotagged aerial imagery |
CN111240617B (en) * | 2019-12-31 | 2020-09-25 | 飞燕航空遥感技术有限公司 | Video delivery method and system based on three-dimensional map and environment monitoring method and system |
US11146727B2 (en) * | 2020-03-16 | 2021-10-12 | Ke.Com (Beijing) Technology Co., Ltd. | Method and device for generating a panoramic image |
US11822333B2 (en) | 2020-03-30 | 2023-11-21 | Zebra Technologies Corporation | Method, system and apparatus for data capture illumination control |
EP3893482A1 (en) * | 2020-04-07 | 2021-10-13 | Parkling GmbH | Transportable recording device for recording data for a located street panorama image |
TWI744861B (en) * | 2020-04-15 | 2021-11-01 | 財團法人工業技術研究院 | Application service server, information processing method based on located region, and user equipment |
US20210358043A1 (en) * | 2020-05-12 | 2021-11-18 | Tellus App, Inc. | Systems and methods for interface-driven engagement |
CN111666434B (en) * | 2020-05-26 | 2021-11-02 | 武汉大学 | Streetscape picture retrieval method based on depth global features |
US11436812B2 (en) | 2020-05-29 | 2022-09-06 | Open Space Labs, Inc. | Machine learning based object identification using scaled diagram and three-dimensional model |
US20220383436A1 (en) * | 2020-07-09 | 2022-12-01 | TensorFlight Poland sp. z o. o. | Automated Property Inspections |
US11302034B2 (en) * | 2020-07-09 | 2022-04-12 | Tensorflight, Inc. | Automated property inspections |
US11450024B2 (en) | 2020-07-17 | 2022-09-20 | Zebra Technologies Corporation | Mixed depth object detection |
US12094494B2 (en) | 2020-09-17 | 2024-09-17 | Camilo Lopez | System and method for generating and editing a video |
CN112163412B (en) * | 2020-09-30 | 2024-02-09 | 善恒展创(深圳)发展合伙企业(有限合伙) | Data verification method and device, electronic equipment and storage medium |
US11367265B2 (en) | 2020-10-15 | 2022-06-21 | Cape Analytics, Inc. | Method and system for automated debris detection |
US11593915B2 (en) | 2020-10-21 | 2023-02-28 | Zebra Technologies Corporation | Parallax-tolerant panoramic image generation |
US11392891B2 (en) | 2020-11-03 | 2022-07-19 | Zebra Technologies Corporation | Item placement detection and optimization in material handling systems |
US11847832B2 (en) | 2020-11-11 | 2023-12-19 | Zebra Technologies Corporation | Object classification for autonomous navigation systems |
CN112508784A (en) * | 2020-11-20 | 2021-03-16 | 中南大学 | Panoramic image method of planar object contour model based on image stitching |
CN112730743A (en) * | 2020-12-17 | 2021-04-30 | 中科三清科技有限公司 | Interaction method and device for air quality mode forecast data |
US11954882B2 (en) | 2021-06-17 | 2024-04-09 | Zebra Technologies Corporation | Feature-based georegistration for mobile computing devices |
US11875413B2 (en) | 2021-07-06 | 2024-01-16 | Cape Analytics, Inc. | System and method for property condition analysis |
US11676298B1 (en) | 2021-12-16 | 2023-06-13 | Cape Analytics, Inc. | System and method for change analysis |
US11861843B2 (en) | 2022-01-19 | 2024-01-02 | Cape Analytics, Inc. | System and method for object analysis |
WO2024070532A1 (en) * | 2022-09-26 | 2024-04-04 | 株式会社リコー | Information processing device, information processing method, program, and information processing system |
Citations (69)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5487116A (en) | 1993-05-25 | 1996-01-23 | Matsushita Electric Industrial Co., Ltd. | Vehicle recognition apparatus |
US5633946A (en) | 1994-05-19 | 1997-05-27 | Geospan Corporation | Method and apparatus for collecting and processing visual and spatial position information from a moving platform |
US5684514A (en) | 1991-01-11 | 1997-11-04 | Advanced Interaction, Inc. | Apparatus and method for assembling content addressable video |
US6043837A (en) | 1997-05-08 | 2000-03-28 | Be Here Corporation | Method and apparatus for electronically distributing images from a panoptic camera system |
US6108728A (en) | 1997-11-14 | 2000-08-22 | Fujitsu Limited | Continuous recording system, method, and recording medium recording continuous recording program and capable of being read by computer |
US6111583A (en) * | 1997-09-29 | 2000-08-29 | Skyline Software Systems Ltd. | Apparatus and method for three-dimensional terrain rendering |
US6237006B1 (en) | 1996-10-15 | 2001-05-22 | Mercury Interactive Corporation | Methods for graphically representing web sites and hierarchical node structures |
US6266615B1 (en) | 1999-09-27 | 2001-07-24 | Televigation, Inc. | Method and system for an interactive and real-time distributed navigation system |
US6366892B1 (en) | 2000-04-05 | 2002-04-02 | Titan Film Capitol Llc. | Method and system for automatically producing loan documents |
US20020047895A1 (en) | 2000-10-06 | 2002-04-25 | Bernardo Enrico Di | System and method for creating, storing, and utilizing composite images of a geographic location |
US6421003B1 (en) * | 2000-05-19 | 2002-07-16 | Sri International | Attitude determination using multiple baselines in a navigational positioning system |
US20020101426A1 (en) | 2001-01-31 | 2002-08-01 | Pioneer Corporation | Information display method |
US6504571B1 (en) | 1998-05-18 | 2003-01-07 | International Business Machines Corporation | System and methods for querying digital image archives using recorded parameters |
US6525731B1 (en) | 1999-11-09 | 2003-02-25 | Ibm Corporation | Dynamic view-dependent texture mapping |
US6594600B1 (en) | 1997-10-24 | 2003-07-15 | Commissariat A L'energie Atomique | Method for calibrating the initial position and the orientation of one or several mobile cameras |
US6597818B2 (en) | 1997-05-09 | 2003-07-22 | Sarnoff Corporation | Method and apparatus for performing geo-spatial registration of imagery |
US6625315B2 (en) | 1998-10-23 | 2003-09-23 | Facet Technology Corp. | Method and apparatus for identifying objects depicted in a videostream |
US20030210806A1 (en) | 2002-05-07 | 2003-11-13 | Hitachi, Ltd. | Navigational information service with image capturing and sharing |
US6678394B1 (en) | 1999-11-30 | 2004-01-13 | Cognex Technology And Investment Corporation | Obstacle detection system |
US20040024709A1 (en) | 2002-08-05 | 2004-02-05 | Yu Paul D. | System and method for determining the identity of a party associated with a transaction |
US20040061774A1 (en) | 2002-04-10 | 2004-04-01 | Wachtel Robert A. | Digital imaging system using overlapping images to formulate a seamless composite image and implemented using either a digital imaging sensor array |
US6741790B1 (en) | 1997-05-29 | 2004-05-25 | Red Hen Systems, Inc. | GPS video mapping system |
US6759979B2 (en) * | 2002-01-22 | 2004-07-06 | E-Businesscontrols Corp. | GPS-enhanced system and method for automatically capturing and co-registering virtual models of a site |
US20040143543A1 (en) | 2003-01-17 | 2004-07-22 | Goldman Robert P. | Electronic real estate settlement |
US20040177373A1 (en) | 2003-02-28 | 2004-09-09 | Tsuyoshi Kawabe | Graphic data distribution system and a data converter, and a graphic data distribution method |
US6798924B2 (en) | 1997-09-10 | 2004-09-28 | Ricoh Company, Ltd. | System and method for displaying an image indicating a positional relation between partially overlapping images |
US20040196282A1 (en) | 2003-02-14 | 2004-10-07 | Oh Byong Mok | Modeling and editing image panoramas |
US6904160B2 (en) | 2000-10-18 | 2005-06-07 | Red Hen Systems, Inc. | Method for matching geographic information with recorded images |
US20050128212A1 (en) | 2003-03-06 | 2005-06-16 | Edecker Ada M. | System and method for minimizing the amount of data necessary to create a virtual three-dimensional environment |
US6919892B1 (en) | 2002-08-14 | 2005-07-19 | Avaworks, Incorporated | Photo realistic talking head creation system and method |
US6963657B1 (en) | 1999-09-24 | 2005-11-08 | Honda Giken Kogyo Kabushiki Kaisha | Object recognition system |
US20050270288A1 (en) * | 2003-03-05 | 2005-12-08 | Arcas Blaise A Y | Method for encoding and serving geospatial or other vector data as images |
US6985638B1 (en) | 1999-06-01 | 2006-01-10 | Shin Aoki | Efficient perspective transformation of partial images in composing a single image |
US6985929B1 (en) * | 2000-08-31 | 2006-01-10 | The United States Of America As Represented By The Secretary Of The Navy | Distributed object-oriented geospatial information distribution system and method thereof |
US20060059073A1 (en) | 2004-09-15 | 2006-03-16 | Walzak Rebecca B | System and method for analyzing financial risk |
US7027054B1 (en) | 2002-08-14 | 2006-04-11 | Avaworks, Incorporated | Do-it-yourself photo realistic talking head creation system and method |
US7050102B1 (en) | 1995-01-31 | 2006-05-23 | Vincent Robert S | Spatial referenced photographic system with navigation arrangement |
US20060155594A1 (en) | 2005-01-13 | 2006-07-13 | Jess Almeida | Adaptive step-by-step process with guided conversation logs for improving the quality of transaction data |
US20060155570A1 (en) | 2005-01-13 | 2006-07-13 | Jess Almeida | Aggregation and control of documents in the document repository using meta data and context information and creation of online info binder |
US7139422B2 (en) | 2002-10-18 | 2006-11-21 | Leco Corporation | Indentation hardness test system |
US7171058B2 (en) | 2003-07-31 | 2007-01-30 | Eastman Kodak Company | Method and computer program product for producing an image of a desired aspect ratio |
US7174301B2 (en) | 2000-10-23 | 2007-02-06 | Costar Group, Inc. | System and method for accessing geographic-based data |
US7181081B2 (en) | 2001-05-04 | 2007-02-20 | Legend Films Inc. | Image sequence enhancement system and method |
US7225207B1 (en) * | 2001-10-10 | 2007-05-29 | Google Inc. | Server for geospatially organized flat file data |
US7305365B1 (en) | 2002-06-27 | 2007-12-04 | Microsoft Corporation | System and method for controlling access to location information |
US7317558B2 (en) | 2002-03-28 | 2008-01-08 | Sanyo Electric Co., Ltd. | System and method for image processing of multiple images |
US7324666B2 (en) | 2002-11-15 | 2008-01-29 | Whitegold Solutions, Inc. | Methods for assigning geocodes to street addressable entities |
US20080113672A1 (en) * | 1996-09-09 | 2008-05-15 | Tracbeam Llc | Multiple location estimators for wireless location |
US20080120215A1 (en) | 2006-11-20 | 2008-05-22 | Wagner Gerald C | Systems and methods related to lifetime mortgages |
US7457706B2 (en) * | 2000-08-16 | 2008-11-25 | Harris Corporation | 3-dimensional interactive image modeling system |
US7467356B2 (en) | 2003-07-25 | 2008-12-16 | Three-B International Limited | Graphical user interface for 3d virtual display browser using virtual display windows |
US7573488B2 (en) | 2001-04-13 | 2009-08-11 | Abb Ab | System and method for organizing two and three dimensional image data |
US7599988B2 (en) * | 2002-08-05 | 2009-10-06 | Metacarta, Inc. | Desktop client interaction with a geographical text search system |
US7619626B2 (en) | 2003-03-01 | 2009-11-17 | The Boeing Company | Mapping images from one or more sources into an image for display |
US20100004995A1 (en) | 2008-07-07 | 2010-01-07 | Google Inc. | Claiming Real Estate in Panoramic or 3D Mapping Environments for Advertising |
US7659915B2 (en) | 2004-04-02 | 2010-02-09 | K-Nfb Reading Technology, Inc. | Portable reading device with mode processing |
US7698212B1 (en) | 2003-11-21 | 2010-04-13 | Peirson Chris A | Online settlement statement and funding control system and method |
US7702821B2 (en) | 2005-09-15 | 2010-04-20 | Eye-Fi, Inc. | Content-aware digital media storage device and methods of using the same |
US7747526B1 (en) | 2006-03-27 | 2010-06-29 | Fannie Mae | System and method for transferring mortgage loan servicing rights |
US7746343B1 (en) | 2005-06-27 | 2010-06-29 | Google Inc. | Streaming and interactive visualization of filled polygon data in a geographic information system |
US7777740B2 (en) | 2003-02-20 | 2010-08-17 | Binary Simplex, Inc. | Spatial decomposition methods using bit manipulation |
US7978970B2 (en) | 2006-09-29 | 2011-07-12 | Microscan Systems, Inc. | Systems and/or devices for providing diffuse light |
US7990394B2 (en) | 2007-05-25 | 2011-08-02 | Google Inc. | Viewing and navigating within panoramic images, and applications thereof |
US7999848B2 (en) | 2004-06-11 | 2011-08-16 | Stratech Systems Limited | Method and system for rail track scanning and foreign object detection |
US8042056B2 (en) | 2004-03-16 | 2011-10-18 | Leica Geosystems Ag | Browsers for large geometric data visualization |
US8050521B2 (en) | 2002-07-27 | 2011-11-01 | Archaio, Llc | System and method for simultaneously viewing, coordinating, manipulating and interpreting three-dimensional and two-dimensional digital images of structures for providing true scale measurements and permitting rapid emergency information distribution |
US8111307B2 (en) | 2008-10-25 | 2012-02-07 | Omnivision Technologies, Inc. | Defective color and panchromatic CFA image |
US8180539B2 (en) | 2007-12-06 | 2012-05-15 | Denso Corporation | Location range setting apparatus, control method and controller for apparatus mounted in mobile object, and automotive air conditioner and control method for the same |
US8195386B2 (en) | 2004-09-28 | 2012-06-05 | National University Corporation Kumamoto University | Movable-body navigation information display method and movable-body navigation information display unit |
Family Cites Families (51)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4396942A (en) * | 1979-04-19 | 1983-08-02 | Jackson Gates | Video surveys |
FR2461305B1 (en) * | 1979-07-06 | 1985-12-06 | Thomson Csf | MAP INDICATOR SYSTEM MORE PARTICULARLY FOR AIR NAVIGATION |
US4553206A (en) * | 1983-10-03 | 1985-11-12 | Wang Laboratories, Inc. | Image storage and retrieval |
US4635136A (en) * | 1984-02-06 | 1987-01-06 | Rochester Institute Of Technology | Method and apparatus for storing a massive inventory of labeled images |
NL8503378A (en) * | 1985-12-06 | 1987-07-01 | Theo Jogchum Poelstra | NEW METHOD FOR BUILDING AND TRACKING DATA FILES FOR ROAD TRAFFIC. |
US4870576A (en) * | 1986-03-19 | 1989-09-26 | Realpro, Ltd. | Real estate search and location system and method |
US5032989A (en) * | 1986-03-19 | 1991-07-16 | Realpro, Ltd. | Real estate search and location system and method |
CA1288516C (en) * | 1987-07-31 | 1991-09-03 | Leendert M. Bijnagte | Apparatus and method for communicating textual and image information between a host computer and a remote display terminal |
US4942538A (en) * | 1988-01-05 | 1990-07-17 | Spar Aerospace Limited | Telerobotic tracker |
US4891761A (en) * | 1988-03-31 | 1990-01-02 | Mets, Inc. | Method for accurately updating positional information provided on a digital map |
US5247356A (en) * | 1992-02-14 | 1993-09-21 | Ciampa John A | Method and apparatus for mapping and measuring land |
US5414462A (en) * | 1993-02-11 | 1995-05-09 | Veatch; John W. | Method and apparatus for generating a comprehensive survey map |
US6037936A (en) * | 1993-09-10 | 2000-03-14 | Criticom Corp. | Computer vision system with a graphic user interface and remote camera control |
US5754850A (en) * | 1994-05-11 | 1998-05-19 | Realselect, Inc. | Real-estate method and apparatus for searching for homes in a search pool for exact and close matches according to primary and non-primary selection criteria |
US5682525A (en) * | 1995-01-11 | 1997-10-28 | Civix Corporation | System and methods for remotely accessing a selected group of items of interest from a database |
US5794216A (en) * | 1995-07-14 | 1998-08-11 | Brown; Timothy Robert | Methods and system for data acquisition in a multimedia real estate database |
US5852810A (en) * | 1996-01-29 | 1998-12-22 | Student Housing Network | Geographic specific information search system and method |
US6025844A (en) * | 1997-06-12 | 2000-02-15 | Netscape Communications Corporation | Method and system for creating dynamic link views |
US5864337A (en) * | 1997-07-22 | 1999-01-26 | Microsoft Corporation | Mehtod for automatically associating multimedia features with map views displayed by a computer-implemented atlas program |
US6323885B1 (en) * | 1998-09-18 | 2001-11-27 | Steven Paul Wiese | Real estate value map computer system |
US6307573B1 (en) * | 1999-07-22 | 2001-10-23 | Barbara L. Barros | Graphic-information flow method and system for visually analyzing patterns and relationships |
US6563529B1 (en) * | 1999-10-08 | 2003-05-13 | Jerry Jongerius | Interactive system for displaying detailed view and direction in panoramic images |
US7974930B2 (en) * | 2000-07-26 | 2011-07-05 | Pierce-Eislen, Inc. | Method and system for providing real estate information |
US6922701B1 (en) * | 2000-08-03 | 2005-07-26 | John A. Ananian | Generating cad independent interactive physical description remodeling, building construction plan database profile |
AU2001288502A1 (en) * | 2000-08-28 | 2002-03-13 | Isdi.Net, Llc | Value your home |
US7487114B2 (en) * | 2000-10-23 | 2009-02-03 | Costar Group, Inc. | System and method for associating aerial images, map features, and information |
US7184611B2 (en) * | 2000-12-26 | 2007-02-27 | Sony Corporation | Data recording apparatus and method, data reproducing apparatus and method, data recording and reproducing apparatus and method, and map image data format |
US7139794B2 (en) * | 2000-12-27 | 2006-11-21 | 3-D-V-U Israel (2000) Ltd. | System and methods for network image delivery with dynamic viewing frustum optimized for limited bandwidth communication channels |
US7171389B2 (en) * | 2001-06-01 | 2007-01-30 | Landnet Corporation | Identification, storage and display of land data on a website |
US20060197763A1 (en) * | 2002-02-11 | 2006-09-07 | Landnet Corporation | Document geospatial shape tagging, searching, archiving, and retrieval software |
CA2352844C (en) * | 2001-07-10 | 2009-11-17 | Geojet Information Solutions Inc. | 3-d map data visualization |
US7096428B2 (en) * | 2001-09-28 | 2006-08-22 | Fuji Xerox Co., Ltd. | Systems and methods for providing a spatially indexed panoramic video |
US20030101063A1 (en) * | 2001-09-28 | 2003-05-29 | Sexton Ashley K. | Real estate information system |
US20030064705A1 (en) * | 2001-10-02 | 2003-04-03 | Desiderio Piero L. | Communication of real estate information utilizing moving map GPS and wireless telecommunications to access MLS and other property specific information from remote databases |
US6636803B1 (en) * | 2001-11-30 | 2003-10-21 | Corus Home Realty | Real-estate information search and retrieval system |
US7092957B2 (en) * | 2002-01-18 | 2006-08-15 | Boundary Solutions Incorporated | Computerized national online parcel-level map data portal |
US6978206B1 (en) * | 2002-06-21 | 2005-12-20 | Infogation Corporation | Distributed navigation system |
US7424133B2 (en) * | 2002-11-08 | 2008-09-09 | Pictometry International Corporation | Method and apparatus for capturing, geolocating and measuring oblique images |
US8051089B2 (en) * | 2003-04-30 | 2011-11-01 | Hewlett-Packard Development Company, L.P. | Systems and methods for location-based real estate service |
US7475060B2 (en) * | 2003-05-09 | 2009-01-06 | Planeteye Company Ulc | Browsing user interface for a geo-coded media database |
US7986339B2 (en) * | 2003-06-12 | 2011-07-26 | Redflex Traffic Systems Pty Ltd | Automated traffic violation monitoring and reporting system with combined video and still-image data |
US7343232B2 (en) * | 2003-06-20 | 2008-03-11 | Geneva Aerospace | Vehicle control system including related methods and components |
JP4111885B2 (en) * | 2003-07-23 | 2008-07-02 | アルパイン株式会社 | Map search and display method and apparatus |
JP4447865B2 (en) * | 2003-08-01 | 2010-04-07 | ソニー株式会社 | Map display system, map data processing device, map display device, and map display method |
US7155336B2 (en) * | 2004-03-24 | 2006-12-26 | A9.Com, Inc. | System and method for automatically collecting images of objects at geographic locations and displaying same in online directories |
US20050288959A1 (en) * | 2004-06-16 | 2005-12-29 | David Eraker | Map-based search for real estate service providers |
US9213461B2 (en) * | 2004-06-16 | 2015-12-15 | Redfin Corporation | Web-based real estate mapping system |
US8751156B2 (en) * | 2004-06-30 | 2014-06-10 | HERE North America LLC | Method of operating a navigation system using images |
US20060041375A1 (en) * | 2004-08-19 | 2006-02-23 | Geographic Data Technology, Inc. | Automated georeferencing of digitized map images |
US8046306B2 (en) * | 2007-06-29 | 2011-10-25 | Zaio Corporation | System, method, and apparatus for property appraisals |
US20100122208A1 (en) * | 2007-08-07 | 2010-05-13 | Adam Herr | Panoramic Mapping Display |
-
2008
- 2008-02-22 US US12/036,197 patent/US8207964B1/en not_active Ceased
-
2012
- 2012-05-28 US US13/481,912 patent/US9384277B2/en not_active Expired - Fee Related
-
2013
- 2013-03-09 US US13/791,957 patent/US20130191359A1/en not_active Abandoned
- 2013-03-09 US US13/791,943 patent/US9311396B2/en not_active Expired - Fee Related
- 2013-03-09 US US13/791,938 patent/US20130201340A1/en not_active Abandoned
- 2013-03-09 US US13/791,951 patent/US20130191725A1/en not_active Abandoned
- 2013-03-09 US US13/791,947 patent/US20130191295A1/en not_active Abandoned
- 2013-03-09 US US13/791,945 patent/US20130188052A1/en not_active Abandoned
- 2013-03-09 US US13/791,955 patent/US8890866B2/en not_active Expired - Fee Related
- 2013-03-09 US US13/791,961 patent/US8558848B2/en not_active Expired - Fee Related
- 2013-03-09 US US13/791,953 patent/US9311397B2/en not_active Expired - Fee Related
- 2013-03-12 US US13/795,959 patent/US20130188008A1/en not_active Abandoned
- 2013-03-12 US US13/796,544 patent/US8902226B2/en not_active Expired - Fee Related
- 2013-03-14 US US13/830,164 patent/US20130201341A1/en not_active Abandoned
- 2013-04-11 US US13/861,336 patent/USRE45264E1/en not_active Expired - Fee Related
- 2013-09-10 US US14/023,028 patent/US20140019301A1/en not_active Abandoned
-
2014
- 2014-04-02 US US14/243,712 patent/US20140211015A1/en not_active Abandoned
-
2016
- 2016-07-01 US US15/200,784 patent/US20170094165A1/en not_active Abandoned
Patent Citations (79)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5684514A (en) | 1991-01-11 | 1997-11-04 | Advanced Interaction, Inc. | Apparatus and method for assembling content addressable video |
US5487116A (en) | 1993-05-25 | 1996-01-23 | Matsushita Electric Industrial Co., Ltd. | Vehicle recognition apparatus |
US5633946A (en) | 1994-05-19 | 1997-05-27 | Geospan Corporation | Method and apparatus for collecting and processing visual and spatial position information from a moving platform |
US7050102B1 (en) | 1995-01-31 | 2006-05-23 | Vincent Robert S | Spatial referenced photographic system with navigation arrangement |
US20080113672A1 (en) * | 1996-09-09 | 2008-05-15 | Tracbeam Llc | Multiple location estimators for wireless location |
US6237006B1 (en) | 1996-10-15 | 2001-05-22 | Mercury Interactive Corporation | Methods for graphically representing web sites and hierarchical node structures |
US6043837A (en) | 1997-05-08 | 2000-03-28 | Be Here Corporation | Method and apparatus for electronically distributing images from a panoptic camera system |
US6597818B2 (en) | 1997-05-09 | 2003-07-22 | Sarnoff Corporation | Method and apparatus for performing geo-spatial registration of imagery |
US6741790B1 (en) | 1997-05-29 | 2004-05-25 | Red Hen Systems, Inc. | GPS video mapping system |
US6798924B2 (en) | 1997-09-10 | 2004-09-28 | Ricoh Company, Ltd. | System and method for displaying an image indicating a positional relation between partially overlapping images |
US6111583A (en) * | 1997-09-29 | 2000-08-29 | Skyline Software Systems Ltd. | Apparatus and method for three-dimensional terrain rendering |
US6594600B1 (en) | 1997-10-24 | 2003-07-15 | Commissariat A L'energie Atomique | Method for calibrating the initial position and the orientation of one or several mobile cameras |
US6108728A (en) | 1997-11-14 | 2000-08-22 | Fujitsu Limited | Continuous recording system, method, and recording medium recording continuous recording program and capable of being read by computer |
US6504571B1 (en) | 1998-05-18 | 2003-01-07 | International Business Machines Corporation | System and methods for querying digital image archives using recorded parameters |
US7444003B2 (en) | 1998-10-23 | 2008-10-28 | Facet Technology Corporation | Method and apparatus for identifying objects depicted in a videostream |
US6625315B2 (en) | 1998-10-23 | 2003-09-23 | Facet Technology Corp. | Method and apparatus for identifying objects depicted in a videostream |
US7092548B2 (en) | 1998-10-23 | 2006-08-15 | Facet Technology Corporation | Method and apparatus for identifying objects depicted in a videostream |
US6985638B1 (en) | 1999-06-01 | 2006-01-10 | Shin Aoki | Efficient perspective transformation of partial images in composing a single image |
US6963657B1 (en) | 1999-09-24 | 2005-11-08 | Honda Giken Kogyo Kabushiki Kaisha | Object recognition system |
US6266615B1 (en) | 1999-09-27 | 2001-07-24 | Televigation, Inc. | Method and system for an interactive and real-time distributed navigation system |
US6525731B1 (en) | 1999-11-09 | 2003-02-25 | Ibm Corporation | Dynamic view-dependent texture mapping |
US6678394B1 (en) | 1999-11-30 | 2004-01-13 | Cognex Technology And Investment Corporation | Obstacle detection system |
US6366892B1 (en) | 2000-04-05 | 2002-04-02 | Titan Film Capitol Llc. | Method and system for automatically producing loan documents |
US6421003B1 (en) * | 2000-05-19 | 2002-07-16 | Sri International | Attitude determination using multiple baselines in a navigational positioning system |
US7457706B2 (en) * | 2000-08-16 | 2008-11-25 | Harris Corporation | 3-dimensional interactive image modeling system |
US6985929B1 (en) * | 2000-08-31 | 2006-01-10 | The United States Of America As Represented By The Secretary Of The Navy | Distributed object-oriented geospatial information distribution system and method thereof |
US20020047895A1 (en) | 2000-10-06 | 2002-04-25 | Bernardo Enrico Di | System and method for creating, storing, and utilizing composite images of a geographic location |
US7813596B2 (en) | 2000-10-06 | 2010-10-12 | Vederi, Llc | System and method for creating, storing and utilizing images of a geographic location |
US6895126B2 (en) | 2000-10-06 | 2005-05-17 | Enrico Di Bernardo | System and method for creating, storing, and utilizing composite images of a geographic location |
US7577316B2 (en) | 2000-10-06 | 2009-08-18 | Vederi, Llc | System and method for creating, storing and utilizing images of a geographic location |
US7805025B2 (en) | 2000-10-06 | 2010-09-28 | Vederi, Llc | System and method for creating, storing and utilizing images of a geographic location |
US7239760B2 (en) | 2000-10-06 | 2007-07-03 | Enrico Di Bernardo | System and method for creating, storing, and utilizing composite images of a geographic location |
US6904160B2 (en) | 2000-10-18 | 2005-06-07 | Red Hen Systems, Inc. | Method for matching geographic information with recorded images |
US7174301B2 (en) | 2000-10-23 | 2007-02-06 | Costar Group, Inc. | System and method for accessing geographic-based data |
US20020101426A1 (en) | 2001-01-31 | 2002-08-01 | Pioneer Corporation | Information display method |
US7573488B2 (en) | 2001-04-13 | 2009-08-11 | Abb Ab | System and method for organizing two and three dimensional image data |
US7181081B2 (en) | 2001-05-04 | 2007-02-20 | Legend Films Inc. | Image sequence enhancement system and method |
US7225207B1 (en) * | 2001-10-10 | 2007-05-29 | Google Inc. | Server for geospatially organized flat file data |
US6759979B2 (en) * | 2002-01-22 | 2004-07-06 | E-Businesscontrols Corp. | GPS-enhanced system and method for automatically capturing and co-registering virtual models of a site |
USRE41175E1 (en) * | 2002-01-22 | 2010-03-30 | Intelisum, Inc. | GPS-enhanced system and method for automatically capturing and co-registering virtual models of a site |
US7317558B2 (en) | 2002-03-28 | 2008-01-08 | Sanyo Electric Co., Ltd. | System and method for image processing of multiple images |
US20040061774A1 (en) | 2002-04-10 | 2004-04-01 | Wachtel Robert A. | Digital imaging system using overlapping images to formulate a seamless composite image and implemented using either a digital imaging sensor array |
US20030210806A1 (en) | 2002-05-07 | 2003-11-13 | Hitachi, Ltd. | Navigational information service with image capturing and sharing |
US7305365B1 (en) | 2002-06-27 | 2007-12-04 | Microsoft Corporation | System and method for controlling access to location information |
US8050521B2 (en) | 2002-07-27 | 2011-11-01 | Archaio, Llc | System and method for simultaneously viewing, coordinating, manipulating and interpreting three-dimensional and two-dimensional digital images of structures for providing true scale measurements and permitting rapid emergency information distribution |
US20040024709A1 (en) | 2002-08-05 | 2004-02-05 | Yu Paul D. | System and method for determining the identity of a party associated with a transaction |
US7599988B2 (en) * | 2002-08-05 | 2009-10-06 | Metacarta, Inc. | Desktop client interaction with a geographical text search system |
US7027054B1 (en) | 2002-08-14 | 2006-04-11 | Avaworks, Incorporated | Do-it-yourself photo realistic talking head creation system and method |
US6919892B1 (en) | 2002-08-14 | 2005-07-19 | Avaworks, Incorporated | Photo realistic talking head creation system and method |
US7139422B2 (en) | 2002-10-18 | 2006-11-21 | Leco Corporation | Indentation hardness test system |
US7324666B2 (en) | 2002-11-15 | 2008-01-29 | Whitegold Solutions, Inc. | Methods for assigning geocodes to street addressable entities |
US20040143543A1 (en) | 2003-01-17 | 2004-07-22 | Goldman Robert P. | Electronic real estate settlement |
US20040196282A1 (en) | 2003-02-14 | 2004-10-07 | Oh Byong Mok | Modeling and editing image panoramas |
US7777740B2 (en) | 2003-02-20 | 2010-08-17 | Binary Simplex, Inc. | Spatial decomposition methods using bit manipulation |
US20040177373A1 (en) | 2003-02-28 | 2004-09-09 | Tsuyoshi Kawabe | Graphic data distribution system and a data converter, and a graphic data distribution method |
US7619626B2 (en) | 2003-03-01 | 2009-11-17 | The Boeing Company | Mapping images from one or more sources into an image for display |
US20050270288A1 (en) * | 2003-03-05 | 2005-12-08 | Arcas Blaise A Y | Method for encoding and serving geospatial or other vector data as images |
US7254271B2 (en) | 2003-03-05 | 2007-08-07 | Seadragon Software, Inc. | Method for encoding and serving geospatial or other vector data as images |
US20050128212A1 (en) | 2003-03-06 | 2005-06-16 | Edecker Ada M. | System and method for minimizing the amount of data necessary to create a virtual three-dimensional environment |
US7467356B2 (en) | 2003-07-25 | 2008-12-16 | Three-B International Limited | Graphical user interface for 3d virtual display browser using virtual display windows |
US7171058B2 (en) | 2003-07-31 | 2007-01-30 | Eastman Kodak Company | Method and computer program product for producing an image of a desired aspect ratio |
US7698212B1 (en) | 2003-11-21 | 2010-04-13 | Peirson Chris A | Online settlement statement and funding control system and method |
US8042056B2 (en) | 2004-03-16 | 2011-10-18 | Leica Geosystems Ag | Browsers for large geometric data visualization |
US7659915B2 (en) | 2004-04-02 | 2010-02-09 | K-Nfb Reading Technology, Inc. | Portable reading device with mode processing |
US7999848B2 (en) | 2004-06-11 | 2011-08-16 | Stratech Systems Limited | Method and system for rail track scanning and foreign object detection |
US20060059073A1 (en) | 2004-09-15 | 2006-03-16 | Walzak Rebecca B | System and method for analyzing financial risk |
US8195386B2 (en) | 2004-09-28 | 2012-06-05 | National University Corporation Kumamoto University | Movable-body navigation information display method and movable-body navigation information display unit |
US20060155570A1 (en) | 2005-01-13 | 2006-07-13 | Jess Almeida | Aggregation and control of documents in the document repository using meta data and context information and creation of online info binder |
US20060155594A1 (en) | 2005-01-13 | 2006-07-13 | Jess Almeida | Adaptive step-by-step process with guided conversation logs for improving the quality of transaction data |
US7933395B1 (en) | 2005-06-27 | 2011-04-26 | Google Inc. | Virtual tour of user-defined paths in a geographic information system |
US7746343B1 (en) | 2005-06-27 | 2010-06-29 | Google Inc. | Streaming and interactive visualization of filled polygon data in a geographic information system |
US7702821B2 (en) | 2005-09-15 | 2010-04-20 | Eye-Fi, Inc. | Content-aware digital media storage device and methods of using the same |
US7747526B1 (en) | 2006-03-27 | 2010-06-29 | Fannie Mae | System and method for transferring mortgage loan servicing rights |
US7978970B2 (en) | 2006-09-29 | 2011-07-12 | Microscan Systems, Inc. | Systems and/or devices for providing diffuse light |
US20080120215A1 (en) | 2006-11-20 | 2008-05-22 | Wagner Gerald C | Systems and methods related to lifetime mortgages |
US7990394B2 (en) | 2007-05-25 | 2011-08-02 | Google Inc. | Viewing and navigating within panoramic images, and applications thereof |
US8180539B2 (en) | 2007-12-06 | 2012-05-15 | Denso Corporation | Location range setting apparatus, control method and controller for apparatus mounted in mobile object, and automotive air conditioner and control method for the same |
US20100004995A1 (en) | 2008-07-07 | 2010-01-07 | Google Inc. | Claiming Real Estate in Panoramic or 3D Mapping Environments for Advertising |
US8111307B2 (en) | 2008-10-25 | 2012-02-07 | Omnivision Technologies, Inc. | Defective color and panchromatic CFA image |
Non-Patent Citations (5)
Title |
---|
A. Akbarzadeh, et. al. "Towards Urban Reconstruction from Video", 3dPVT2006 Proceedings, Jun. 14, 2006. |
A. Roman, et. al, "Interactive Design of Multi-Perspective Images for Visualizing Urban Landscapes", IEEE Visualization 2004, Oct. 10-15, Austin Texas, pp. 537-544. |
Christian Fruh, Avideh Zakhor, An Automated Method for Large-Scale, Ground-Based City Model Acquisition, International Journal of Computer Vision, 5-24, 2004. |
Don Kimber, Jonathan Foote, and Surapong Lertsithichai , "FlyAbout: Spatially Indexed Panoramic Video". Multimedia '01 Proceedings of the ninth ACM international conference on Multimedia. New York 200, pp. 339-347. |
Mitsunori Yoshimura, Ryosuke Shibasaki, Tetsujo Anai, and Hirofumi Chikatsu , "Ground Based Sensor Integration for Spatial Data Acquisition and Database Development". International Archives of Photogrammetry and Remote Sensing. vol. XXXIII, Part B5. Amsterdam 2000, pp. 933-936. |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11470303B1 (en) | 2010-06-24 | 2022-10-11 | Steven M. Hoffberg | Two dimensional to three dimensional moving image converter |
US9230167B2 (en) | 2011-12-08 | 2016-01-05 | The Nielsen Company (Us), Llc. | Methods, apparatus, and articles of manufacture to measure geographical features using an image of a geographical location |
US9378509B2 (en) | 2012-05-09 | 2016-06-28 | The Nielsen Company (Us), Llc | Methods, apparatus, and articles of manufacture to measure geographical features using an image of a geographical location |
US9547866B2 (en) | 2013-03-14 | 2017-01-17 | The Nielsen Company (Us), Llc | Methods and apparatus to estimate demography based on aerial images |
US10885097B2 (en) | 2015-09-25 | 2021-01-05 | The Nielsen Company (Us), Llc | Methods and apparatus to profile geographic areas of interest |
Also Published As
Publication number | Publication date |
---|---|
US20130328858A9 (en) | 2013-12-12 |
US8890866B2 (en) | 2014-11-18 |
US20130191359A1 (en) | 2013-07-25 |
US20170094165A1 (en) | 2017-03-30 |
US20130201341A1 (en) | 2013-08-08 |
US8207964B1 (en) | 2012-06-26 |
US9311397B2 (en) | 2016-04-12 |
US20130191295A1 (en) | 2013-07-25 |
US20130191725A1 (en) | 2013-07-25 |
US9384277B2 (en) | 2016-07-05 |
US20130182108A1 (en) | 2013-07-18 |
US20140089811A1 (en) | 2014-03-27 |
US20130201340A1 (en) | 2013-08-08 |
US20130187949A1 (en) | 2013-07-25 |
US9311396B2 (en) | 2016-04-12 |
US20140019301A1 (en) | 2014-01-16 |
US20130188008A1 (en) | 2013-07-25 |
US8902226B2 (en) | 2014-12-02 |
US8558848B2 (en) | 2013-10-15 |
US20130191292A1 (en) | 2013-07-25 |
US20140211015A1 (en) | 2014-07-31 |
US20130191252A1 (en) | 2013-07-25 |
US20120293505A1 (en) | 2012-11-22 |
US20130188052A1 (en) | 2013-07-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
USRE45264E1 (en) | Methods and apparatus for generating three-dimensional image data models | |
US8078396B2 (en) | Methods for and apparatus for generating a continuum of three dimensional image data | |
US8666657B2 (en) | Methods for and apparatus for generating a continuum of three-dimensional image data | |
US7929800B2 (en) | Methods and apparatus for generating a continuum of image data | |
US11415986B2 (en) | Geocoding data for an automated vehicle | |
AU2015204838B2 (en) | Unmanned aircraft structure evaluation system and method | |
US10127721B2 (en) | Method and system for displaying and navigating an optimal multi-dimensional building model | |
CA3062310A1 (en) | Video data creation and management system | |
CN107609900A (en) | A kind of room streetscape pricing method based on oblique photograph and BIM | |
US11947354B2 (en) | Geocoding data for an automated vehicle | |
Stal et al. | Highly detailed 3D modelling of Mayan cultural heritage using an UAV |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
REMI | Maintenance fee reminder mailed | ||
LAPS | Lapse for failure to pay maintenance fees |