US20130188008A1 - Generating Video Drive-By Data Corresponding to a Geographic Location and Related Three Dimensional Models - Google Patents

Generating Video Drive-By Data Corresponding to a Geographic Location and Related Three Dimensional Models Download PDF

Info

Publication number
US20130188008A1
US20130188008A1 US13/795,959 US201313795959A US2013188008A1 US 20130188008 A1 US20130188008 A1 US 20130188008A1 US 201313795959 A US201313795959 A US 201313795959A US 2013188008 A1 US2013188008 A1 US 2013188008A1
Authority
US
United States
Prior art keywords
image
data
video
image data
camera
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/795,959
Inventor
William D. Meadow
Randall A. Gordie, Jr.
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
VISUAL REAL ESTATE Inc
Original Assignee
MV Patents LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US11/216,465 external-priority patent/US7389181B2/en
Application filed by MV Patents LLC filed Critical MV Patents LLC
Priority to US13/795,959 priority Critical patent/US20130188008A1/en
Publication of US20130188008A1 publication Critical patent/US20130188008A1/en
Assigned to MVP PORTFOLIO LLC reassignment MVP PORTFOLIO LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MV PATENTS LLC
Assigned to VISUAL REAL ESTATE, INC reassignment VISUAL REAL ESTATE, INC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MVP PORTFOLIO, LLC
Assigned to SiberLaw, LLP reassignment SiberLaw, LLP LIEN (SEE DOCUMENT FOR DETAILS). Assignors: Visual Real Estate, Inc.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/05Geographic models
    • G06F17/30265
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/29Geographical information databases
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/587Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using geographical or spatial information, e.g. location
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/7867Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title and artist information, manually generated time, location and usage information, user ratings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/787Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using geographical or spatial information, e.g. location
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9537Spatial or temporal dependent retrieval, e.g. spatiotemporal queries
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0278Product appraisal
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/06Buying, selling or leasing transactions
    • G06Q30/0601Electronic shopping [e-shopping]
    • G06Q30/0623Item investigation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/06Buying, selling or leasing transactions
    • G06Q30/0601Electronic shopping [e-shopping]
    • G06Q30/0623Item investigation
    • G06Q30/0625Directed, with specific intent or strategy
    • G06Q30/0629Directed, with specific intent or strategy for generating comparisons
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/16Real estate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/60Editing figures and text; Combining figures or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/20Finite element generation, e.g. wire-frame surface description, tesselation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4038Image mosaicing, e.g. composing plane images from plane sub-images
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/698Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/28Mobile studios
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/181Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources

Definitions

  • the present application relates in general to systems and methods of generating image data of a geographic location of interest.
  • the invention relates to generating video drive-by data corresponding to a selected geographic location.
  • the video drive-by data provides a street level view of a neighborhood surrounding the selected geographic location, and can be further combined with other data sources in a variety of applications.
  • the invention also relates to generating data descriptive of a continuum of images, such as those comprising a geographic setting.
  • MLS Multiple Listing Service
  • the MLS database typically includes one or more photographs and additional data regarding the property such as the number of bedrooms and the square footage.
  • More advanced systems may include short video clips constituting virtual tours of the property in question. While such systems may also show views of a property from the street, they do not give a user a sense of the neighborhood in which the property is located. Accordingly, in order to determine if the area surrounding the property of interest is desirable, a user may often be required to physically travel to the property in order to view the state of the neighborhood.
  • video drive-by data a complete on-line database of street-level video data, hereinafter referred to as “video drive-by” data, which can be made available to end users via the Internet.
  • the video data would be encoded with GPS coordinates as well as conventional street addresses and map coordinates to provide instant street level video of any location upon request.
  • GPS coordinates as well as conventional street addresses and map coordinates to provide instant street level video of any location upon request.
  • map coordinates to provide instant street level video of any location upon request.
  • a comp list typically includes average per square foot calculations of the similar properties that enables the creation of approximate value of the subject property.
  • a separate class of field appraisals includes both a Drive By Appraisal and a Full Interior Appraisal.
  • a Drive By Appraisal is where an appraiser drives by and shoots a photograph from the street of a subject property and it's comps. These photographs are linked to county parcel data and AVM calculations to provide a more accurate report to the mortgage company.
  • a mortgage banker In the case of three of four appraisal types, however, a mortgage banker must wait several days before receiving the information from an appraiser. If a system were available to allow quick access to all properties within a neighborhood, a mortgage banker could obtain photographs of a subject property and its comps and confirm that the comparables as selected by the AVM represents properties that have the same visual appearance of the subject property, thereby providing faster loan commitments and closings while at the same time lowering costs. Similarly, instant Drive By Appraisals could be provided thereby eliminating the costs and delays associated with an appraiser driving to manually photograph a subject property and its comps.
  • Photographs Images of geographic areas created on a local basis have been created in numerous formats. Photographs, movie cameras, video camera recorders, and more recently digital recorders have all been utilized to capture moving images of a geographic parcel. Photographs can be useful because they are easily associated with a particular real estate parcel; however, they are limited in the scope of the subject matter that they can include. In addition, photographs must be manually acquired and docketed in order to be associated with a property and subsequently retrieved. panoramas can increase the scope of the subject matter to a point, but are limited by a point of capture of the panoramic views.
  • Movie cameras, video recorders and digital recorders and other “motion pictures” provide for increased scope of image capture.
  • correlation of particular portions of a motion picture with a particular real estate parcel can be difficult if the motion picture includes more than one real estate parcel.
  • Aerial images, such as satellite pictures of geographic areas have also been captured, and specific parcels of land or landmarks can be located on the aerial images.
  • Substantial portions of ground level images have not been correlated with aerial images or with geopolitical indicators, in part, because there has not been any image vehicle capable to providing a format of high quality and accurate representations of street level images capable of being matched to macro level image data.
  • a system and method of providing video drive-by data is provided to enable a street level view of a neighborhood surrounding a selected geographic location.
  • the video drive-by data can be further combined with other data sources related to the geographic location.
  • the invention is based, in part, on providing: 1) accurate differential GPS data; 2) post processing of geo positioning signals to smooth curves in motion (splines); 3) highly accurate camera position and video frame position analysis processing to provide a calculation of the exact position of each video frame; 4) parcel data processing that analyses vector line data that is geo-coded with latitude and longitude values; 5) digital image photos processed with image superimposition algorithms; and 6) a database that includes video image files, parcel latitude and longitude data, and positioning data that is indexed to the video image files.
  • the invention enables the access to video images of any desired geographic location and its surrounding neighborhood, while relating such image data to other property related data such as property lines, landmarks, etc.
  • a system in a preferred embodiment, includes a video and data server farm.
  • the video and data server farm includes at least one video storage server that stores video image files containing video drive-by data that corresponds to a geographic location, a database server that processes a data query received from a user over the Internet that corresponds to a geographic location of interest, and an image server.
  • the database server identifies video image files stored in the video storage server that correspond to the geographic location of interest contained in the data query, and transfers the video image files over a pre-processing network to the image processing server.
  • the image processing server converts the video drive-by data to post-processed video data corresponding to a desired image format, and transfers the post-processed video data via a post-processing network to the Internet in response to the query.
  • a landing zone server is preferably provided to receive the video drive-by data from a portable memory device and permit the viewing and analysis of the video drive-by data prior to storage in the video storage server.
  • a map server is preferably provided to present a static image of an overhead view of the geographic location of interest.
  • the video drive-by data is preferably captured by a video capture system that includes a camera array that generates video image data of a neighborhood corresponding to a geographic location, a positioning unit that produces positioning data corresponding to the location of the camera array, and a processing unit that processes the video image data and positioning data to produce the video drive-by data.
  • the processing unit stores the video drive-by data on the portable memory device.
  • the camera array includes a plurality of camera units arranged to provide a 360 degree view of the neighborhood.
  • the processing unit preferably provides time stamp data and file pointer data that permits synchronization of video images to precise locations of individual frames.
  • processing means may be incorporated to provide a variety of geo-coded data layers in conjunction with the video drive-by data.
  • processing means can be provided for: creating geo-coded text, image and vector data which is superimposed onto the post-processed video data that is transferred to the Internet; linking latitude and longitude data with video drive-by data; calculating a camera position of designated individual frames of video drive-by data using position data points captured before and after a designated frame position; for examining changes of relative position in the horizontal position of designated landmarks in a sequence of video image frames to create fixed landmark data points; examining images that match common image patterns to enable the creation of an accurate vector map of roads; providing OCR processing to generate a database of recognized text that is geo-coded with latitude and longitude coordinates based on camera orientation and relative position; and selecting video drive-by data or a given street address by calculating a centroid of
  • Still further embodiments can include: processing means by which a sequence of video frames that are oriented toward a designated parcel and it's adjacent parcels can be stitched together to form a long horizontal image of a street which can be created with superimposed property county parcel data placed above the centroid of the parcel centered above the property; processing means for providing authenticated annotation of information or comments to a photo and or video; processing means for uploading a digital image photograph to the video and data server farm and comparing the uploaded photo to a library of images captured by the video capture system; and/or processing for comparing digital image photos of building structures to generate a set of numeric values representing line segments of sharp horizontal, vertical and angular lines that are stored as vectors each with common ratios of distances from each other, wherein the list of vector line segments can be searched with a high-speed database query to generate a list of possible matches of a subject property against a large database of properties with vector line data on file.
  • a system in accordance with the invention can offer invaluable assistance to local, county or state road and highway departments in planning new projects or tracking existing placements of signage, traffic signals, utility boxes, and other facilities and equipment locations. Administrators can preview and familiarize themselves with a targeted location in conjunction with assessing needs or assigning work details.
  • New layers of routing assistance can also be provided to emergency vehicle dispatching systems by including roadside views to allow rapid and accurate targeting of a destination.
  • the present invention can offer street level video into the neighborhood and directly to the street address, offering photographs of any property transmitted via wireless Internet to help the fire, police or rescue driver find his or her way.
  • the system can enhance online presentations with parcel line enhanced ortho imagery, property images and now with video. Further, with the system, unification of property records and county mapping can evolve to tie in to one-click access to street level video segments allowing a user to virtually drive down any street, while highlighting any property and its surrounding neighborhood with parcel lines drawn on the photograph.
  • tourism boards at any level can promote travel and recreation to a broader audience by integrating visuals and streaming video in presenting points of interest, accommodations, and dining as integrated layers to the local, county or state map that are unified to promotional content.
  • the video drive-by data generated by the system can offer compelling video segments highlighting parks, beaches, hunting, fishing, arena sports, museums and places of historical interest; while, also providing directions, drive maps, and links to information detailing places to stay, restaurants, rental agencies and more.
  • the system can be utilized to communicate the value and desirability of a new subdivision or location to potential builders and residents.
  • a property developer's web site can be switched on to enable parcel line enhanced satellite photography with access to street level video data of the newly paved roads of the development.
  • Web site visitors would have the freedom of taking a virtual tour of any selected parcel of property, looking around to note the scenery, tree line, and proximity to neighboring lots, thereby providing a true feel for the land within its natural surroundings.
  • Data and content can be integrated to any parcel with a ready feedback mechanism allowing viewers to inquire further or request an appointment.
  • property appraisers can gain new efficiencies at reduced costs by utilizing a video drive-by data archive of streaming street level video to select and capture freeze-frame video images of homes for sale needing an exterior photo.
  • the system's unification of property data, MLS records, mapping with the video resources allows any property to be quickly selected, analyzed, and compared to equivalent property listings, generating reports with active links to both data and visuals that can be emailed at a moments notice.
  • the present invention also relates to methods and apparatus of generating image data. More specifically, the present invention also relates to generating data descriptive of a continuum of three dimensional images, such as those comprising a geographic landscape. Accordingly, the present invention provides methods and apparatus for generating a continuum of three dimensional image data.
  • the continuum of three dimensional image data provides a street level representation of geographic locations. Two or more sets of image data are captured of a subject area, wherein each of the two or more images are captured from disparate points on a continuum
  • methods and apparatus for generating a three dimensional output that includes a continuum of image data sprayed over three-dimensional models.
  • the three-dimensional models can be representative of features captured by the image data, wherein image data can be captured at multiple disparate points along another continuum.
  • the user interface can also include multiple modalities of image data and statistical analysis of the image data.
  • the continuum of image data provides a two dimensional ribbon of street level views of geographic areas.
  • the ribbons are created by capturing two or more images of a subject, wherein each of the two or more images are captured from disparate points on a continuum. For example, each of the two or more images can be captured from a different position on a street as a vehicle drives down the street. Generally, the images will be captured from an orthogonal view of the subject.
  • Portions of the two or more images are then aligned in a dimension consistent with the continuum, wherein, according to our example the continuum includes the path of the vehicle. Therefore, the images would be aligned in a dimension consistent with the path of the vehicle.
  • a composite image of the subject is generated from the aligned portions of the two or more images.
  • positional data is recorded that is descriptive of a respective location of each of the two or more images.
  • the positional data can include, for example, latitude and longitude coordinates and can be used to associate the composite image with a particular portion of the subject. It is also within the scope of the invention to record the altitude of a camera used to capture the image data and thereby approximating the altitude of the subject matter of the image data.
  • a camera will be maintained approximately orthogonal to the subject captured and therefore, those embodiments will maintain the camera at an angle about between 75° and 105° in relation to the subject matter.
  • some embodiments can include various overlays of information on top of the continuum of image data. Overlays can include one or more of: metadata, data related to a composite image, and geospatial information.
  • FIG. 1 is a schematic diagram of a Video and Data Server Farm in accordance with a preferred embodiment of the invention
  • FIG. 2 is a schematic block diagram of a Video Capture System that supplies video data and corresponding synchronous GPS data to the Video and Data Server Farm illustrated in FIG. 1 ;
  • FIG. 3 is a process flow diagram illustrating the process of capturing the video data and synchronous GPS data utilized by the Video Capture System illustrated in FIG. 2 ;
  • FIG. 4 is a process flow diagram illustrating an efficient process in accordance with the present invention by which numerous field drivers with Video Capture Systems of the type illustrated in FIG. 2 can be continually supplied with hard disk drives to capture data;
  • FIG. 5 illustrates a web browser view with an address bar pointing to a web page that shows both an ortho image view of a mapserver and video pane in accordance with the invention
  • FIG. 6 is a process flow diagram for superimposing geo-coded text data over video captured by the Video Capture System of the type illustrated in FIG. 2 ;
  • FIG. 7 is a data flow diagram that illustrates the process by which an instant drive-by appraisal or Photo AVM (Automated Valuation Model) can be created using a Video drive-by database that contains geo-coded video;
  • An instant drive-by appraisal or Photo AVM Automated Valuation Model
  • FIG. 8 illustrates a top perspective of a Video Capture System as it drives past a structure
  • FIG. 9 illustrates a frame position interpolation in accordance with the present invention.
  • FIG. 10 shows a set of four illustrations which represent a single frame or portion of a window within an Internet Web Page in accordance with the invention
  • FIG. 11 illustrates a process by which images captured by a left facing and a left front facing camera in a Video Capture System can calculate and generate superimposed property
  • FIG. 12 illustrates a process to extract and create a point of interest database that is encoded with GPS data
  • FIG. 13 illustrates two ortho views of a Video Capture System as it drives in a direction of travel along a path past a selected parcel where one of six possible cameras may be viewing a subject property;
  • FIG. 14 illustrates a process by which an authenticated end user can submit data regarding a specific location using Video drive-by data as a geo-coding reference tool
  • FIG. 15 illustrates a process and visual references as to the calculations required to extract a photograph from a street number in accordance with the invention
  • FIG. 16 illustrates a novel method of finding a house from a photograph based on the invention
  • FIG. 17 illustrates one novel method of identifying a home from a large library of home photos with a high speed process in accordance with the invention
  • FIG. 18 illustrates a continuum from which image data sets capture a subject
  • FIG. 19 illustrates multiple images captured from disparate points along a continuum
  • FIG. 20 illustrates multiple image data sets that overlap subject matter
  • FIG. 21 illustrates a three-dimensional model generated form the image data sets
  • FIG. 22 illustrates a user interface generated from the image data sets and including multiple modalities of image data
  • FIG. 23 illustrates an exemplary user interface for traversing multiple modalities of image data
  • FIG. 24 illustrates an exemplary composite of image data set portions
  • FIG. 25 illustrates apparatus what may be used to implement the present invention
  • FIG. 26 illustrates basic geographic images captured from a camera at disparate points along a continuum and aligned with post processing
  • FIG. 27 illustrates an alignment of portions of two or more images captured from disparate points along a continuum to form a composite image
  • FIG. 28 illustrates multiple images captured from disparate points along a continuum and specific vertical slices that can be selected to create a composite image
  • FIG. 29 illustrates a block diagram of exemplary lens distortion factors and a preferred area of lens capture that can be selected in some embodiments of the present invention
  • FIG. 30 illustrates disparate points on a continuum with an orientation vector associated with each respective point
  • FIG. 31 illustrates a canyon formed by two continuums of image data according to the present invention.
  • FIG. 32 illustrates apparatus that can be used to implement some embodiments of the present invention.
  • FIG. 1 A Video & Data Server Farm 0100 in accordance with a preferred embodiment of the invention is illustrated in FIG. 1 .
  • the Video & Data Server Farm 0100 utilizes a set of servers to provide digital storage of video drive-by data, and processes the video drive-by data for delivery to an end user.
  • the video drive-by data preferably includes video data and Global Positioning System (GPS) data or “geo position data” generated by a Video Capture System 0201 illustrated in FIG. 2 .
  • GPS Global Positioning System
  • video drive-by data is preferably stored on an external hard disk drive (HDD) 0250 of the Video Capture System 0201 , and is then copied from the external HDD 0250 to a landing zone server 0130 of the Video & Data Server Farm 0100 for quality analysis and viewing via industry standard commercially available MPEG2 viewing software and file editors.
  • the Video drive-by data including the video data and GPS data, is transferred via a load network 0142 to be stored as video image files in video storage servers 0145 of the Video & Data Server Farm 0110 .
  • a user logs onto the Internet 0150 via an end user interface 0102 to access the video image files containing the converted video drive-by data via a data query that includes data related to a specific geographic location of interest.
  • the data query received by a web server 0105 and carried over a control network 0135 to a database server 0120 and a map server 0125 , which preferably calculates and presents a static image (for example a street map image) of an overhead view of the geographic location of interest.
  • the database server 0120 identifies the appropriate video image files corresponding to the location of interest from the video storage servers 0145 for transfer over a pre-processing network 0140 to image processing servers 0115 .
  • the image processing servers 0115 convert the original video drive-by data to one of many potential new image formats (depending on the particular application) which constitute Post-processed Video Data (PPVD).
  • the PPVD is then transferred to a video server 0110 over a post-processing network 0141 , which then forwards the PPVD data to the Internet 0150 in response to the query.
  • the video server 0110 can also concurrently transfer the PPVD back to the video storage servers 0145 , database servers 0120 and map server 0125 for future access from a video cache.
  • FIG. 2 is a schematic block diagram of the components in the Video Capture System 0201 .
  • a camera unit 0205 preferably containing six video cameras 0210 in a hexagonal array is provided in a camera housing assembly 0215 .
  • the camera housing assembly 0215 is mounted with differing configurations on top of a vehicle (not shown) preferably using commercially available roof racks and straps.
  • the electronics for converting and storing the video data out of the camera 205 are preferably provided in a passenger cab 0220 of the vehicle.
  • the Video Capture System 0201 is preferably powered from the vehicles electrical system, although a separate power source can be provided if so desired.
  • the camera 0205 feeds analog video signals 0227 along six cables in a cable harness 0225 which is fed through a small opening in a window of the vehicle.
  • the analog video signals 0227 are fed into six analog-to-digital (A/D) converters 0230 , which convert the analog video signals 0227 to compressed digital video data.
  • the compressed digital video data is then sent on data cables 0232 into a seven port USB hub 0235 .
  • the six streams of digital video data are fed via the hub 0235 into a laptop computer 0245 , which preferably writes the digital video data onto a portable storage device, for example the external HDD 0250 , simultaneously with the data from a GPS receiver 0240 . Accordingly, the data is stored with corresponding GPS synchronized coordinates in the external HDD 0250 as the Video drive-by data.
  • FIG. 3 is the block diagram illustrating the process of capturing the video data and synchronous GPS data to create the video drive-by data.
  • a set of programs is automatically launched when power is applied to the laptop computer 0245 .
  • the GPS capture program 0305 processes the data transmitted from the GPS receiver 0240 and writes 0307 this as time stamped data to the laptop computer's RAM 0330 and to a file on the external HDD 0250 .
  • the time stamp of the data file written contains the laptop computer's internal system clock time data as well as the GPS clock data.
  • a driver viewer application 0310 is initialized to enable the driver of the vehicle to confirm that the GPS receiver 0240 is working, and—as the video data is captured—the driver can preferably view segments of the video data on the display of the laptop computer 0245 to confirm the system is working properly.
  • the driver viewer application 0310 shows a map of roads with color codes representing areas that have already been surveyed in one color and streets that need to surveyed in a second color.
  • Video capture daemons 0315 are started to process the video drive-by data arriving via the USB hub to then write 0317 the six data files to the external HDD 0250 with the laptop computer's 0245 system time clock time stamp data and file pointer data added to the video file and a GPS log file.
  • the video capture daemons writes 0317 the file pointer data to the laptop computer's RAM 0330 for use by a master index file.
  • the master index file daemon 0340 is started which reads the GPS data, the six video file pointers values and then writes 0345 the master index file data to the external HDD 0250 .
  • the entire set of data is then copied to the landing zone server 0130 when the external HDD drives are sent to the Video & Data Server Farm 0100 for upload and processing as will be described in greater detail below.
  • FIG. 4 is a process flow diagram illustrating a preferred efficient process by which numerous field drivers with video capture systems can be continually supplied with hard disk drives to capture data.
  • the figure is comprised of three columns 0405 , 0410 and 0420 separated by dashed lines.
  • the left column 0405 entitled “Video & Data Server Farm” represents the location of the Video & Data Server Farm 0100 where all data is stored and subsequently delivered to the Internet 0150 .
  • the middle column 0410 entitled “In Transit Shipping” represents the movement of data in external hard disk drives 0250 via commercial shipping companies.
  • the right column 0420 entitled “Video Capture System Field Driver Residence” represents one of many locations where drivers in the field handle external hard disk drives for receiving, installing and shipping.
  • time movement is indicated along the vertical axis in a downward direction with the start of the process at the top indicated as Week 1.
  • Week 1 For example, the end of Week 1 is represented by a horizontal line 0419 .
  • Week 2 represented as the time between the lines 0419 and 0429 , a shipment 0443 of blank external hard disk drives 0445 is moved from Video & Data Server Farm 0405 to the Video Capture System Field Driver Residence 0420 .
  • the angle of the line in the downward direction to the right illustrates the time delay of several days as the external hard disk drives 0445 are physically moved by a commercial shipping company or private shipping means specifically set up for operation of the system.
  • the process starts with an initial shipment 0430 of a Video Capture System through In Transit Shipping process path 0432 to individual drivers in the field for installation in their respective vehicles.
  • a blank external HDD is then shipped in process path 0433 to the individual drivers in their respective Field Driver Residence for installation into the Video Capture System.
  • the external HDD 0250 is illustrated as several unique external hard disk drives as they change state and use as they move through various stages of processing.
  • the first external hard disk drive labeled as HDD 1 0440 is shipped blank to the Field Driver Residence where it is installed.
  • a driver drives along selected roads capturing video data and GPS data which is then written onto the HDD 1 as illustrated with the wavy line 0445 .
  • a second blank external hard disk drive HDD 2 0492 is shipped 0443 to the field driver where it sits in inventory waiting for use.
  • the driver ships 0450 the first full HDD 1 0491 to the Video & Data Server Farm for processing.
  • the first HDD 1 0491 is now re-labeled as HDD 1 - 4 in the landing zone upload process 0496 carried out at the Video & Data Server Farm.
  • the data is copied 0465 to the landing zone server and a GPS.LOG and a Master Index File data is loaded to the map server.
  • An operator is given an opportunity to review the data using video file viewing software to provide a visual quality review test process 0470 .
  • the quality review process 0470 is completed, the operator then erases 0475 the data from the external HDD and places the external HDD drive on the inventory shelf 0480 labeled as HDD 4 .
  • the field driver installs HDD 2 0492 into the video capture system and starts driving only a survey route to capture video data which is then stored 0455 . Later in Week 3, the field driver receives a blank HDD 3 0493 which is stored at the field driver residence until it is needed at the beginning of Week 4 0439 . The driver completes the capture process at the end of Week 3 0439 and ships 0463 HDD 2 0492 to Video & Data Server Farm 0405 , at which point the landing zone upload process is repeated 0498 .
  • the above-described process provides a simple yet efficient method of moving large amounts of data from the field driver residence back to the Video & Data Server Farm.
  • FIG. 5 illustrates a web browser 0505 view with an address bar 0510 pointing to a web page that shows both an ortho image view 0515 of a mapserver and video pane 0540 .
  • drive line layers 0520 are shown representing the GPS coordinates as a map layer on top of parcel lines 0525 with an end users mouse cursor 0530 pointing to the selected area of video that the user wants to see in the video pane 0540 .
  • a user can select a parcel 0535 , indicated with the double lines, and then the system can show a video view of the property 0545 in the video pane 0540 .
  • this type of interface a user can simultaneously view both conventional street map representations as well as the street level Video drive-by data generated by the system.
  • FIG. 6 is a process flow diagram of the processes implemented with the image processing servers 0115 required to superimpose geo-coded text data over video captured by the Video Capture System.
  • a video program 0610 determines available video segments that contain images where the property is in view. This is determined by using an open source programming tool GDAL which completes a range calculation from the latitude and longitude positions encoded in the video segments to the latitude and longitude positions of the subject property. The determination of if a parcel is in view is based on a variable distance that is initially set at approximately 300 feet from camera position to the nearest line segment of a parcel.
  • an extractor program 0620 (which is based on an open source programming tool MENCODER) converts the compressed MPEG2 data into lossless TIFF image files with each image frame file encoded for camera orientation and interpolated location.
  • a calculation using GDAL is made in a camera centroid position program 0630 to determine the exact latitude and longitude of a centroid of the parcel, which generally approximates the location of the structure.
  • another simple geometry calculation is made as to the orientation and location of the camera view that is pointed toward the centroid of the parcel.
  • the location of the centroid of the parcel and the distance to the parcel is then calculated for the frame.
  • a render program 0640 which runs on the image processing servers 0115 , is then provided (in this example implemented using open source programming tool suite OpenGL and GraphicsMagick) to calculate a location within the TIFF image file in which text or an icon can be placed based on the centroid.
  • a recompositing program 0650 (in this example implemented using an open source programming tool FFMPEG) then assembles each of the TIFF images together in time sequence to create a compressed video file. Finally, the video in MPEG2 format with the newly added text and icons added is streamed 0660 out to an end user using an open source Apache web server.
  • FIG. 7 is a data flow diagram that illustrates the process by which an instant drive-by appraisal or photo AVM (Automated Valuation Model) can be created using a video drive-by database that contains geo-coded video.
  • the diagram represents four discrete locations in three columns: the upper left column represents a Video & Data Server Farm 0701 ; the center column represents the Internet 0702 ; the right column represents an end user appraiser 0703 ; and the lower left column represents a lender 0704 .
  • data is transferred from end user appraiser 0703 to and from the Video & Data Server Farm 0701 and finally to the Lender 0704 via the Internet 0702 .
  • the process starts, for example, with the end user appraiser 0703 enters a street address 0710 for a subject property on a web form.
  • the street address data is then sent 0715 to the server hosting one of many commercially available AVM or Comparative Market Analysis Applications (CMAA) 0720 , which does a database search and selects a list of properties that are similar in price, size and nearby location to the subject property.
  • CMAA Comparative Market Analysis Applications
  • the list of comparables or comps is sent 0725 back to the end user appraiser 0703 who reviews and analyses various data fields for the comps (such as price, sale date, bedrooms, square feet, structure, lot size, etc.) from the full list of comps 0730 .
  • the selected list of comps 0735 is sent back to the Video & Data Server Farm 0701 where a select video program 0740 transfers video files from the image processing servers 0115 to a video-to-digital photo extractor program 0745 (implemented in the current example using open source programming tools MENCODER and DNoise 3 D). Digital photos of the subject property and selected comps are then sent 0750 to the end user appraiser 0703 , who is presented with a web form 0760 which contains a photo of the subject property 0762 and a set of photos of comps ( 0764 thru 0768 ).
  • the end user appraiser 0703 reviews the picture of the first comp 0764 and sees that it is approximately the same size and appearance and then clicks on the first check box 0784 to accept it as relevant and appropriate.
  • the third comp 0766 that is presented, it is too small and it is not selected 0786 .
  • the second 0765 and fourth comps 0767 are selected 0785 , 0787 and finally the fifth comp 0768 is not selected 0788 as being too large and not of appropriate visual quality as the subject.
  • the selection list is then sent back 0770 to the lender 0704 as the appraisal selected relevant comp properties report, which is then reviewed 0775 by the lending officer to assist in their financial decision process.
  • FIG. 8 illustrates a top view perspective of a video capture system in accordance with a further embodiment of the invention as it drives past a structure.
  • the line 0800 represents the track of the Video Capture System with one-second GPS time stamp intervals 0801 , 0802 .
  • the line 0800 represents the track of the Video Capture System with one-second GPS time stamp intervals 0801 , 0802 .
  • the right camera field of view 0820 has several easily identified high contrast landmarks in sight.
  • landmark represents any fixed object, coloration or structure that is easily identified in a digital image photograph, this can include (but is not limited to) light pole, mail box, edge of driveway, corner of building, roofline, front door, tree, bush, pond edge, awning, centerline on road, public sign, etc.
  • the landmarks are identified and labeled in the video frames with commercially available and open source programs that provide edge detection analysis of digital image photos.
  • the view of the left edge of a landmark tree trunk 0830 and its leaf-shading boundary 0831 is highlighted with an arrow 0835 .
  • the horizontal position of the tree within the field of view is indicated in line 0825 .
  • the camera has several other landmarks a window frame 0836 and a door 0837 that have virtual lines 0826 , 0827 drawn from the camera to the landmark. These virtual lines provide a point reference in the horizontal dimension with a video frame as to the angle by which this landmark is oriented from the camera.
  • a second set of landmark views virtual lines 0845 , 0846 and 0848 within the right camera field of view 0821 are available for simple geometry calculations.
  • the horizontal location of the tree trunk landmark 0835 is now in the extreme right most position of the video frame of 0815 camera position.
  • the position of each frame between the GPS one-second interval points can be calculated with time and speed between each. The assumptions are that no radical changes in speed or direction occur within the one second interval.
  • the change of horizontal location of the camera from frame position 5 0810 to frame position 29 0815 pointing to a fixed point 0835 enables a calculation of the distance to 0835 .
  • a precise set of relative positions can be calculated for landmarks in the video.
  • the newly created two dimensional data for the location of the landmarks can now be added to a GIS viewing software application such as the open source University of Minnesota MapSever.
  • FIG. 9 illustrates a process of frame position interpolation in accordance with the invention.
  • three line segments are drawn showing three scenarios of road travel to illustrate where large changes in velocity and direction within one second intervals can cause major errors in calculations of Video Capture System 0201 camera positions, if one is relying upon straight line travel from point to point.
  • a straight line travel 0901 three one second GPS points 0905 , 0906 and 0907 are logged.
  • a set of ten notch marks representing every third video frame are illustrated as 0908 and 0909 which are frames 4 and 7 respectively. In this case the actual camera position is nearly identical to the calculated camera position.
  • the second line segment 0902 illustrates the difference between actual travel 0915 of the Video Capture System on a windy road with a stop sign 0917 versus the calculated position of travel 0925 and 0925 versus the point to point travel 0926 .
  • the calculated position of travel is based upon using a fourth order polynomial function with parameters that the coordinates of four data points 0930 , 0931 , 0932 and 0933 .
  • This function calculates line segments illustrated as 0921 , 0922 , 0923 which have notch marks.
  • the first line segment 0921 it is drawn over the actual path illustrating a highly accurate estimation of the travel path.
  • a nearly straight line 0922 is calculated versus the actual turn.
  • the third line segment 0903 illustrates the calculated path 0940 of a Video Capture System as it navigates a cul-de-sac with is highly accurate in motion to the actual path versus the point to point path 0941 .
  • a spline polynomial method works well in that it creates smooth curves that fit the natural motion of drivers as they make constant G force turns with smooth turns on a steering wheel.
  • FIG. 10 illustrates an aerial transition to street process.
  • a set of four illustrations 1010 , 1020 , 1030 and 1040 are provided, which represent a single frame or portion of a window within an Internet web page.
  • the top three illustrations 1010 , 1020 and 1030 are created using an open source GIS Mapping program, such as University of Minnesota Map Server, to show parcel line boundaries.
  • the three sequential illustrations show the effect as a user is zooming twice from 1010 to 1020 and then 1020 to 1030 . Once the user has reached a designated zoom height as in the case of web page view 1030 where individual parcels can be clearly seen, addition image layers 1031 , 1032 and 1033 are added.
  • the centroid 1031 of a selected parcel is identified with the motion of travel 1032 of the video segment as the starting point of Video Capture System camera position, and direction of travel 0133 with the dashed line 0134 representing the field of view of the camera.
  • the fourth web page frame 1040 in the sequence illustrates the smooth transition from aerial to street view of a selected parcel. In this case a single frame is shown based on the position 1035 of the Video Capture System.
  • the centroid of the parcel of web page frame 1030 is illustrated as 1031 and in the fourth web page frame 1040 the same parcel centroid is now illustrated as 1041 providing the end user good visual cues to help them see where they are as they view a street level image.
  • the actual path of the Video Capture System is drawn on the video frames as 1042 as the drives past the designated parcel.
  • FIG. 11 illustrates a process utilizing real property vectors in accordance with the invention, by which images captured by a left facing and a left front facing camera in a Video Capture System can calculate and generate superimposed property lines on both video and exported digital image photos from the video drive-by data.
  • an ortho view 1200 of the travel of a Video Capture System along a road is shown with four discrete image frame capture points 1211 , 1221 , 1231 and 1241 . These image frame capture points are illustrated on the left hand side as 1210 , 1220 , 1230 and 1240 .
  • the camera In the case of camera position 1241 where the car is approaching the subject property 1249 and 1209 the camera has a view of the left side of the road 1203 and 1243 where the property line 1246 , 1206 and 1226 can be calculated as described above and shown on a digital image photo.
  • the second 1230 , third 1220 and fourth 1210 digital image photos show the changing perspectives of the Video Capture System camera as it moves past a subject property.
  • FIG. 12 illustrates a process to extract and create a point of interest database that is encoded with GPS data in accordance with the invention.
  • Video and Data Server Farm Image processing cluster of servers 0115 and the video frame to digital image photo extraction process 0620 any one of several dozen OCR programs can be run on the extracted digital image photos to build a results file of recognized text.
  • a Video Capture System's front left camera at position 1305 is viewing a sign 1310 and 1320 that contains recognizable text.
  • This text data is then written into a database that includes but is not limited to: the text characters, size and color of the text, relative position of the characters in the field of view and GPS coordinates of the video frame. Exact position in latitude and longitude of the text of the sign can be located as well as the relative orientation.
  • FIG. 13 illustrates two ortho views 1400 and 1450 of a Video Capture System as it drives in a direction of travel 1410 along a path 1415 past a selected parcel where one of six possible cameras may be viewing a subject property.
  • the hexagonal array 1405 indicates the range of view for each of the six cameras.
  • the Video Capture System at position 1401 has a subject property centroid 1420 in the field of view 1425 of the right front camera.
  • the field of view can be easily calculated by drawing a virtual line segment 1445 from the camera position 1401 to the centroid 1420 .
  • the right camera field of view 1475 is now containing the centroid 1470 of the subject property when the camera is in position 1451 , which is illustrated by the virtual line segment 1460 .
  • FIG. 14 illustrates a process by which an authenticated end user 1503 can submit data regarding a specific location using video drive-by data as geo-coding reference tool.
  • the figure has three columns: the left column represents a Video and Data Server Farm 1501 ; the middle column represents the Internet 1502 ; and on the right column represents an end user 1503 .
  • the process starts when the end user 1503 establishes credentials 1505 to authenticate with personal information such as their name, address and email with a credit card number, social security number, date of birth, drivers license number, etc. At that point they can select their login and password and submit 1506 to a common and commercially available registration application 1510 running on an Apache web server.
  • the user can then login 1515 and send a request to authenticate 1516 .
  • that user is authenticated 1522 , they are now granted access 1525 to the video server 1530 where the can do a series of queries 1526 (by street address or clicking on a map as shown in FIG. 5 or 7 ) where they obtain results 1532 in the form of video feeds.
  • the user When the user is ready to review and edit property data, they login and authenticate with the password. Using one of several modes of selecting a property, they view a particular parcel or area with a picture of the property.
  • the viewer application can show video with various information such as parcel data or property lines illustrated on the property.
  • the user can also select the video to present frames at a slower rate than 30 frames per second to allow careful viewing of the imagery.
  • they can annotate any landmark by clicking on the image of a property (such as the roofline of a house 0545 shown in FIG. 5 ). They can use the cursor to establish a point on an image as a designated landmark. The user can then tag the same landmark in one or more video frames to further clearly tag the image designated landmark.
  • the exact latitude and longitude coordinates can be determined (with an estimated error calculation included in the data).
  • that location data of the one or mouse clicks with a point, line segment or shading of area or attachment of an icon
  • the database includes individual record entries of each and every annotated data set to generate a detailed change log.
  • This log file can include but it is not limited to latitude and longitude of camera position, estimated accuracy of camera position at time of capture, time of original video capture as well as end user annotation time.
  • FIG. 15 illustrates a process and visual references as to the calculations required to extract a photograph from a street number.
  • the process starts with the loading of county parcel data that typically includes latitude and longitude coordinates of vector line segments that identify the individual parcel lines of given property.
  • Typical county parcel records include street name and number as part of the parcel number that is supplied as a separate database with the parcel number itself being the unique ID that is used to link street address to latitude and longitude encoded parcel lines.
  • This figure includes two diagrams 1605 and 1610 .
  • the top diagram 1605 shows a horizontally stitched together set of frames of video 1636 and 1641 as the left Camera is moving by a set of houses.
  • the stitching process can be accomplished by using any one of several dozen commercially available programs running on the video image server cluster 0115 after the video frames have been extracted as indicated in FIG. 6 .
  • the stitching program will be feed a set of digital image photos in this example Frame 1 1636 Frame 34 1641 , Frame 67 1646 and Frame 100 1651 are used to composite a complete wide horizontal view of the street.
  • Each second of video produces thirty individual image frames. Accordingly, in this example, 3.3 seconds of video were used from frame 1 to frame 100 .
  • the frame number to use when driving by properties is easily calculated by the speed data from the change in position of the GPS data points. Using techniques described above, text can be superimposed 1625 onto the digital image photo.
  • an end user is using a web page viewer 1615 to move a set of left 1621 and right 1620 control arrow button to select the image that evenly frames a subject property in the field of view.
  • a web page viewer 1615 By providing a long horizontal picture view 1605 in a square computer monitor viewer web page 1615 an end user can easily identify properties with large street numbers 1625 displayed as they are looking to identify a property.
  • FIG. 16 illustrates a novel method of finding a house from a photograph based on the invention.
  • FIG. 16 includes three columns: left column 1701 represents a Video and Data Server Farm, the middle column represents the Internet 1703 ; and the right column represents an end user.
  • the process starts with a user entering a web page 1705 where they can upload a digital image photograph 1706 of a property taken from the street.
  • the user can also input a zip code where they think the property is located to enable faster searching.
  • the Video and Data Server Farm the uploaded image is converted 1715 to the appropriate standard format of uncompressed bits representing each pixel.
  • the uncompressed image is then processed with one of several commercially available pattern recognition applications 1725 , where the uploaded image 1706 is compared to the street level digital image photographs extracted (for example those illustrated in FIGS. 13 and 15 ) from the video drive-by process.
  • the photos that have patterns matching the closest to the submitted 1706 photo are then extracted and added to a webserver to create a web page 1725 .
  • This web page is then delivered to the end user as a web page 1735 with the originally selected photo 1706 redisplayed as 1740 with additional extracted images 1745 , 1750 below.
  • the end user does the final human based pattern matching by selecting with a radio dialog 1755 the image that looks the closest and then presses the submit button 1760 .
  • This data selection is sent to the server 1765 where the server processes the selection 1770 to generate a web page that returns 1775 the street address and picture of the property 1780 .
  • FIG. 17 illustrates one novel method of identifying a home from a large library of home photos with a high speed process.
  • the process involves using one of many commercially available applications that generate edge detection patterns in an image.
  • an image of a house 1801 is shown with various lines that would be trivial for an edge detection algorithm to find properly.
  • horizontal lines such as the top of the front door 1815 , top of the roof 1816 , and roof line 1817 and vertical lines such as front door 1811 , left edge of the right window 1812 and right edge of the right window 1813 are easily identified.
  • the vertical edges are displayed as a set of bar codes that have a relative horizontal distance from each other.
  • the bar line 1821 is derived from the right side of the front door 1811 and the left side of the right window creates another bar line 1822 .
  • the horizontal edges that have been detected are displayed as bar lines.
  • bar line 1836 is detected on the top of roof line 1816
  • bar line 1837 is the roof line 1817
  • the bar line 1835 is the top of the door 1815
  • the bar line 1836 is the bottom of the right window sill 1816 .
  • the storage of the discrete bar lines as mathematical values with relative position data enables a numeric pattern match of values between photos taken at different distances from a subject property. In this way a set of numbers representing the largest bar lines as vectors can be stored and compared against a large numbers of properties with a simple sorting process.
  • the present invention may be applied to many applications in which the remote ability to view a specific location and its surrounding neighborhood is advantageous.
  • land owners, prospective land owners, insurance, mortgage, county appraisers need accurately to determine where property lines are located in relation to a property.
  • a physical visit to a property with a survey in hand is required to start the process of determining where the property lines are located.
  • a professional surveyor may be required to find the original iron stakes that are usually covered over with vegetation. Once the stakes are found, using long tape measures and surveying equipment, an accurate determination can be made of the locations of various structures and landmarks in relation to the property lines.
  • the present invention permits the merger of a unique set of data together to generate street level digital image photographs of properties that include superimposed illustrations of property lines based on original latitude and longitude data provided by the county property appraiser's office. This highly convenient method of illustrating survey data fits a wide variety of unmet market needs.
  • the present invention enables an end user the ability to select any street via an on-line web page and scroll along a view of images where the county parcel data is superimposed in the video stream. Images of property parcels as well as the camera orientation pointed toward the centroid of the parcel are provided along with property record data that can be processed to locate the centroid of the property and then display text data of sale price, owners name & street address above the image of the house.
  • property appraisers are mandated by many state and local governments to obtain photographs of properties that they tax. Currently some counties pay $10 to $50 to have a driver go to specific properties, take a digital image photo and email it to the appraiser. Also, some counties contract to have firms drive along streets shooting photos and storing them in a database that they deliver to the appraiser for a fee of $1 to $2 per parcel. In contrast, with the present invention, a county government could contract to view video of parcels in a far more economical manner by using our Video Capture System driven by their existing appraisal staff. The data they capture can then be added to the Video & Data Server Farm and is available for viewing in a 360° view not just in a single view of a property. From the video, dozens of individual photos are available for the property appraiser to view at any time and by default the centroid oriented view at a 90 degree angle can be shown first.
  • emergency service agencies have equipped their vehicles with high speed wireless data networks to augment their voice radio network.
  • the wireless data networks enable access to the Internet and the Virtual Private Network to provide information related to the location and details on the emergency call while in route.
  • This location data is typically a street address and a map with directions. Maps today generally do not have good positional accuracy for individual street addresses. So when emergency responders pull onto the street, they immediately commence looking for house numbers. When these numbers cannot be found especially at night, they resort to guessing and this consumes valuable time when seconds count.
  • a solution which has been implemented in several counties is to provide a database of street photos that are tied to parcel data.
  • This process is done by manually photographing properties and linking them to a database of street addresses that may or may not be encoded with latitude and longitude data from a GPS.
  • the present invention enables the automatic and low-cost capture of street level video which can be encoded with parcel data and GPS coordinates so that the Video & Data Server Farm can generate a long horizontal street level image that can be transmitted rapidly as a single image to an emergency vehicle in route which aids in the ability for the emergency responder to quickly decide at which house to stop.
  • the present invention includes the additional opportunity to present street level images of turns and intersections that enable users to “see” what an intersection looks like a few hundred feet before they arrive and need to turn.
  • the user could see that there is street level video available based on showing a double red line for example that the user could click on to see a street level image.
  • the user clicks on the red line they would immediately be presented with video from that exact latitude and longitude coordinate.
  • the Video & Data Server Farm has the latitude and longitude coordinate data of the camera as it drives by and in addition, and it has the latitude and longitude coordinate data of the centroid of the parcel that is for sale. Given this data it is easy to determine which camera position is actually pointed at and has in its field of view the centroid of the parcel.
  • homes are located with the foundation over the centroid and thus the automatic camera orientation method provided by the present invention will enable the automatic selection of the correct camera view to present to the user when they are “driving on-line”.
  • the present system allows the automation of finding a reasonably good view of a property, by utilizing the centroid image view automation process with an end user based process of “fine tuning” a view of a property.
  • the user viewing a stitched image of the subject and its next door neighbors, they can slide the bar right and left to simulate their movement of a photographer on the street moving right and left to get the image “just right”.
  • the system will store this optimal view coordinate data and also generate a full resolution digital image photo file for delivery to the user.
  • the servers would run one of several commonly available pattern matching processes against the images for a particular zip code and then present a set of properties that “look close” to the one the user submitted. Given the fact that humans are vastly superior to systems for pattern recognition, a list of 20 close matches out of 2000 potential homes in a given zip code can be easily viewed and selected by an end user in a couple of seconds. Once the user has selected the home that matches, they can then be presented with the street address and any other market comparable data they may wish to obtain.
  • methods and apparatus are provided for presenting a three-dimensional model of a subject, such as a geographic area.
  • the three-dimensional models can include a polygon based model format generated from a point cloud array and sprayed with actual image data of the subject, captured from disparate points along a continuum.
  • a user interface is presented with user interactive controls that allow a user to traverse the three-dimensional model.
  • a user location within the model will determine which portions of the three-dimensional model are viewed.
  • Some embodiments can include user interactive devices that allow a user to traverse the three-dimensional model in three directions: a) an essentially up and down vertical movement; b) an essentially side-to-side horizontal movement; and c) a depth movement into and out of the three-dimensional model.
  • Image data and three-dimensional polygon based models presented to an inventor at a given instance can be dependent upon the relative position selected by the user within the user interface.
  • icons, overlays and metadata can be presented to a user upon demand to provide further information relating to a given geographic area within view.
  • Still another aspect can include presentation of one or more geographic areas based upon criteria ascertainable from related data.
  • image data can be correlated with geospatial data so that a particular street address or Cartesian coordinate, such as a latitude/longitude location, is designated and specific image data is presented which correlates with the geospatial data.
  • the specific image data can be presented sprayed over three-dimensional structures generated from two-dimensional image data captured from that specific location.
  • some embodiments can also include indications on the user interface of user selectable data related to a particular location.
  • User selectable data can include, for example: cellular phone signal strength; wireless network data, including network name, location and signal strength; global position system satellite data and signal strength; radio signal strength; school district data, average yard size, property price range; building features; proximity to amenities; political boundaries; zoning restrictions; and almost any other measurable quantity.
  • Data can all be correlated according to a geographic location and presented in various combinations or upon demand.
  • the data can be presented in conjunction with or independent of, an image data continuum or image data set.
  • Video DriveByTM refers to street level video data captured in multiple angles encompassing a 360° view.
  • Video FlyByTM refers to Aerial/Satellite oblique (angular) view images with polygon line views.
  • RibbonViewTM refers to a film strip like view of properties that stands up at approximately 90° from a flat or angled aerial/satellite ortho image map and provides direct-on front images of properties to be displayed.
  • Flash Viewer refers to direct streaming of video to an online user via a web browser.
  • image data is captured from disparate points along a continuum.
  • the continuum 100 includes the path of a vehicle carrying a digital camera, or other image data capturing device.
  • Image data sets are captured at disparate points 101 - 106 along the continuum.
  • Some preferred embodiments include capturing each image data set at an angle that is generally orthogonal to the subject matter 107 .
  • Positional data and orientation of the camera capturing image data sets is recorded for each disparate point 101 - 106 along the continuum from which an image data set is captured.
  • Positional data can include any data indicative of where the subject matter of an image data set is located.
  • Some preferred embodiments of positional data include Cartesian coordinates that are context sensitive according to the mechanism used to generate coordinate data.
  • Positional data can be generated, for example, by an image data recording device, such as a camera, associated with a device for recording a global position (GPS device). Time stamps associated with image data sets and time stamps associated with the GPS data can be utilized to correlate the GPS location data with image data set recorded by the image data recording device.
  • an image data recording device such as a camera
  • GPS device global position
  • an altimeter can be used to record an altitude from which a camera records image data sets.
  • the altitude data can be associated with an image data set, for example, metadata correlated with the image data set.
  • Such embodiments can therefore include recordation of a latitude, longitude and altitude coordinate for a given image data set.
  • geospatial data such as latitude and longitude coordinates
  • GPS data can be time stamped and collected once every second.
  • GPS reception can be interrupted, depending upon location relative to large object, such as multistory buildings, or cold cover. Therefore, some additional embodiments can include an accelerometer for recording motion associated with a camera and a GPS unit operatively attached to the camera.
  • Data from the accelerometer can indicate movement of the camera.
  • Some accelerometers such as micro electro-mechanical system (MEMs) accelerometers can easily be incorporated into a camera system assembly attached to a vehicle.
  • MEMs micro electro-mechanical system
  • Use of multiple MEM accelerometers positioned to measure movement in four or more directions along an x-axis, y-axis, and z-axis in relation to a camera can also be used to calculate direction of movement.
  • the accelerometer can therefore be used to extrapolate a current position of the camera, based upon a last set of GPS geospatial data recorded.
  • Geospatial data can be used to indicate an initial geographic position.
  • a change in geospatial data can be additionally utilized to indicate velocity and direction of image data set capture.
  • Accelerometer data may also be used to indicate a velocity and direction of image data set capture.
  • Accelerometer data may also be used to indicate a velocity and direction of camera movement. Calculations of time elapsed at the indicated velocity (such as for example, the Kalman Filter) can yield a calculated position at a time of image capture, even if the time of image capture is between GPS readings.
  • one standard can include tracking a camera position with a GPS unit that records location at a rate of once per second.
  • the camera can record image data at a faster rate than once per second, such as, for example, one of: 12 images per second, 24 images per second or 29.97 images per second.
  • An initial camera position can be recorded that correlates with a GPS reading; subsequent image data capture will occur in between the one second GPS reading interval.
  • the camera position can be determined with a calculation based upon the velocity of camera movement supplied by the accelerometer and time elapsed since a previous GPS reading.
  • Still other embodiments can utilize optical flow methodology and visual odometry to facilitate calculations of a camera position and the velocity of a vehicle or person from which a series of image data sets are captured.
  • Visual odometry can be accomplished with a single omni-directional camera or with stereo cameras and is based upon processing that tracks the position of salient features in a series of feature sets and calculates movement based upon the relative positions of the features.
  • camera based simultaneous localization and mapping (SLAM) of visual image data can also be used to facilitate calculations of velocity of a change in position of a camera used to capture image data sets.
  • the velocity will be directly tied to the motion of a vehicle to which the camera is mounted, or a person carrying a camera rig.
  • Orientation of a camera can include a direction of image capture recorded by the camera. Orientation can be designated, for example, in relation to the cardinal directions, i.e., north, south, east and west. Any means available to record such a designation, such as an electronic compass, is within the scope of the present invention. However, it may be desirable to include a means to record the orientation with a greater degree of accuracy than is typically available through the use of an electronic compass.
  • orientation can be determined according to a fixed position of a camera in relation to the direction of travel of a vehicle (or person) used to transport the camera.
  • a plurality of cameras can be fixedly attached to a vehicle capturing Video DriveByTM data.
  • Each camera therefore maintains a constant direction of image capture in relation to the heading of the vehicle.
  • Mechanics of the camera such as, for example, lens parameters and shutter speed, can indicate a depth of field during camera image data capture.
  • Some embodiments can also include simultaneously capturing multiple image data sets and correlating two or more of the image data sets. Correlation can be accomplished via a time stamp or other chronological or synchronous.
  • the position of a camera can be combined with a direction of image capture and the depth of field of the camera to determine a location of image data captured by the camera at a particular instance in time.
  • the present invention can also include apparatus for utilizing echo location to determine a distance of an object from a camera capturing an image data set and storing a correlated distance with the image data set.
  • radar data can be correlated with a camera image data set to calculate the location of various objects captured by the camera.
  • a time stamp can also be combined with data to quantify a location for a particular image formed by the captured image data.
  • data used to calculate a location of an image is stored in a metadata file space associated with the image data.
  • some embodiments can store metadata in the exchangeable image file format (EXIF), TIFFTAGS or International Press Telecommunication Council (IPTC) formats.
  • Image data may be stored, for example in JPEG or TIFF formats.
  • other metadata formats can also be used.
  • the image data and metafile data are stored on an external data storage device, such as a hard disk drive operatively attached to the camera.
  • the data can be stored in the camera.
  • the metadata can include data descriptive of some or all of: date and time; camera settings such aperture, shutter speed and focal length; geospatial data from a GPS receiver unit; accelerometer data; inertial guidance system data; camera orientation; and camera fixed position related to vehicle travel.
  • image data sets 201 - 203 are illustrated which capture images of a subject 200 from different points along a continuum.
  • image data sets 304 - 307 overlap the subject matter captured. The overlap allows for features present in one image data set to be present in a second image data set, and preferably in a tertiary image data set.
  • FIG. 21 a block diagram illustrates three-dimensional structures 405 - 407 created from image data captured.
  • the structures can be created by generating a point cloud array from a plurality of features contained in the image data sets and converting the point cloud array to a polygon based model.
  • the polygon based model can be associated with a location relative to the image data set and image data can be sprayed over the polygon based model.
  • image data sprayed over the polygon based models includes the composite of image data.
  • a user can traverse the presentation of the composite of image data and the three-dimensional structures. Movement through the user interface can be accomplished with any known interaction device, such as, for example one or more of: a keyboard, mouse, video game controller Apple I Phone, digital cellular phone, remote controller, WiiMote and a joystick.
  • any known interaction device such as, for example one or more of: a keyboard, mouse, video game controller Apple I Phone, digital cellular phone, remote controller, WiiMote and a joystick.
  • the user interface can allow a user to move in three directions through a scene presented.
  • the directions can include an x coordinate providing a vertical dimension 401 , a y dimension providing a horizontal dimension 402 and a z coordinate providing a depth dimension 403 .
  • Three-dimensional modeling sprayed with actual image data allows the user to proceed in “off-track” paths that do not match a path created by the disparate points from which data is collected.
  • a time dimension can also be traversed, such as, for example, image data sets of a geospatial designation across time chronology such as a period of seasons or years.
  • Some embodiments can include a “take me there” time designation, wherein a user can indicate a time period and geospatial description.
  • a three-dimensional model can be generated with image data sets from the indicated time period for the designated geospatial area.
  • planes of image data 405 can change according to the perspective chosen by the user.
  • a three-dimensional structure of a house can be sprayed with image data included in the continuum composite of image data.
  • the actual image data will correlate with a surface of the three-dimensional structure facing the disparate points from which the image data was captured.
  • additional surface of the house 406 can also be sprayed with a texture and color captured by the actual image data. In this way, as a user moves in a z direction beyond front surface of the house, the user will be presented with additional surfaces of the house, with each surface having a color and texture extrapolated from the image data from the front of the house.
  • an object such as a tree 405 can also be created in three dimensions with color and texture based upon the image data composite.
  • user interactive device controls can be provided which allow a user to remove objects such as the tree 405 and the house 406 . Removal of an object will allow the user to see beyond the removed object into the scene provided. For example, if a tree 405 obscures the view of the front of a house 406 , activation of a user interactive control can remove the tree 405 allowing an unobstructed view of the three-dimensional model of the house 406 . If actual image data is not available for those portions of the house now viewable in the unobstructed view, then the present invention allows for color and texture data matching the known surfaces of the house to be sprayed over those portions now viewable.
  • Some embodiments can additionally allow a user to modify color and texture of three-dimensional models presented to them. Such modifications facilitate “what-if” scenarios. For example, a buyer contemplating purchasing a particular property can be presented with a three-dimensional model of the property with three-dimensional representations of objects on the property sprayed with actual image data captured of the property. According to some embodiments, the buyer can then modify the view presented to represent possible changes that may be made to the property. Changes may include, to continue the example, providing a brick facade over a painted surface, or removal of shrubbery, or addition of shrubbery. Addition of items can be provided from a menu of three-dimensional objects made available to the user. Color selections and textures can also be made available via a user selectable menu. In this fashion a unique blend of a composite continuum of image data and user defined image data can be presented, including three-dimensional modeled objects derived from the image data and user modifications to such models.
  • a user may be able to select a building and designate from a user interface a type and color of roofing the user would like to see illustrated for the building. Accordingly, a user may designate an architectural shingle in various hues, or a Spanish tile or steel roof and the present invention will spray the pattern and color selected onto a three-dimensional model of the building.
  • a user may upload a color or hue or texture to be sprayed upon a three-dimensional model of a structure.
  • the color and hue may also be stored and made available for subsequent application to portions of an image data set.
  • the present invention can also include generation of a user interface 500 with multiple representations of image data relating to a geographic area.
  • an aerial view of a geographic area 504 can be combined with one or more two-dimensional continuums of composite image data 505 - 506 .
  • the two-dimensional continuums of composite image data 505 - 506 can be generated according to methods and apparatus described in a related co-pending patent application by the same inventors.
  • one or more continuums of disparate points from which image data is captured can also be indicated on the aerial view 504 .
  • Three-dimensional models 510 - 503 can also be included in the user interface 500 and located on the aerial view according to geospatial data generated during image data capture and processing of the image data.
  • the three-dimensional models can be generated from multiple image data sets captured from disparate points along a continuum sprayed with the image data captured.
  • a user may locate a particular geographic area via different vehicles made available to a user.
  • a user can enter a geospatial designation, such as, for example, one or more of: a street address; a Cartesian coordinate; a “Favorite” name designating a geospatial area; a popular name for a geospatial area; a coded indication for a geospatial area, such as a multiple listing number or data file number; and metadata.
  • a “take me there” command can be executed to bring the user to the requested geospatial area.
  • the present invention provides a vehicle “cockpit” interface which allows a user to virtually “fly” over an aerial view and then transition into a virtual “dashboard” and “drive” mode along a street level two-dimensional continuum of actual image data, and in some preferred embodiments, a continuum of composite image data.
  • the user may also transition into a virtual three-dimensional modality which includes three-dimensional models sprayed with image data captured from a continuum of disparate points.
  • the three-dimensional models can be viewed from a virtual “drive by” mode or a virtual “walkabout.”
  • the virtual walkabout can allow a user to go off path and view the geographic area from amongst three-dimensional models sprayed with image data captured from a continuum of disparate points.
  • the user may manipulate the appearance of the three-dimensional models.
  • a take me there command can be additionally operative to designate a modality, such as the aerial view, two-dimensional ribbon view, or three-dimensional drive by or three-dimensional walk about view into which a user will enter.
  • a modality such as the aerial view, two-dimensional ribbon view, or three-dimensional drive by or three-dimensional walk about view into which a user will enter.
  • a user can be provided with tools that provide analysis of the image data shown. Analysis can include, for example, designation of a particular portion of a geographic area 508 .
  • the portion 508 can be representative of a real estate parcel, as indicated by municipal records, or some other designation.
  • a user may draw a polygon 508 onto an aerial image of a geographic area 500 and ask for analysis of the area contained within the polygon 508 .
  • Analysis can include, by way of example, a calculation of an area included within the polygon. Other analysis can include more sophisticated calculations. For example, it is within the scope of the invention to designate property boundaries, such as, through county record data and locate an image data of the property.
  • the image data can be used to generate a three-dimensional model of the property, including a model of a house within the property boundaries and other features, such as trees and utility demarcations.
  • the present invention may also be used to generate a size of a front yard, and, by extrapolation, the size of the side and back yards.
  • An approximate size of a house can also be calculated based the dimensions of the three-dimensional model generated.
  • Other features, such as how many stories, the house includes and the type of facade the house has may also be ascertained from analyzing the image data and the model.
  • Statistical representations can also be made, according to the analysis.
  • Statistical representations can provide a summary of characteristics of a single real estate parcel, or a designated area.
  • statistical analysis can include one or more of an average, median and high/low value designation for: the size of a property within a given neighborhood, the size of a front yard, how much road frontage each property includes, the average size of the houses within the designated area, the number of stories comprising the buildings within the area, siding types of buildings within the area, the size of side yards between houses, height of trees within the area, and almost any other data that may be calculated from the image data sets and the three-dimensional models.
  • the interface 600 can include a portion with an aerial view 601 and a polygon designation of geographic area of interest, such as for example, a particular real estate parcel or neighborhood location.
  • the interface may also include a portion with a two-dimensional continuum of image data 603 A.
  • the continuum of image data is preferably a composite of image data sets captured from disparate points.
  • Another portion 603 B can include image data sprayed over three-dimensional models 604 generated from point clouds of features captured by the image data sets.
  • image data sprayed over the three-dimensional polygon formats includes a composite image formed by aligning two or more of the image data sets.
  • the present invention creates a composite through alignment of portions of data from more than one data set. Alignment can be accomplished in image data processing.
  • image data processing the images 701 - 703 are aligned to form a composite image 700 .
  • the composite image 700 is essentially two dimensional image data arranged as a second continuum, or ribbon.
  • the second continuum includes ongoing image data 701 - 703 captured from the points defining the first continuum.
  • the series of points of image capture in the first continuum includes positions of a vehicle carrying an image capture device, such as a camera, as the vehicle traverses a path proximate to a geographic area.
  • the camera is positioned to capture image data of the geographic area.
  • Image data 701 - 703 is periodically captured as the vehicle traverses the path. The motion of the vehicle, combined with the periodic capture of image data 701 - 703 , thereby results in image data 701 - 703 being captured from disparate points along the first continuum.
  • a preferred embodiment includes capture of image data with a motion vector of the camera in space maintained generally orthogonal to a subject for which image data will be captured. Orthogonal capture of the image data can facilitate consistency for subsequent composite of portions of the image data captured. Therefore, data captured at an angle of between about 75° and 105° can provide data most easily assembled into a continuum of data. However, other angles may be used to create different effects. In addition, in some embodiments physical limitation may dictate the use of other angles. In such instances, post processing can be utilized to compensate for the angle of image data capture.
  • a continuous pictorial representation includes RibbonViewTM by Real Data Systems. RibbonViewTM correlates a ribbon of geographic image data with geospatial designations to facilitate identification of a particular geographic area, as discussed more fully below.
  • image capture processing can be accomplished in real time or via post image capture processing.
  • select portions 704 - 707 of two or more sets of captured image data are aligned to generate the composite image 708 .
  • some preferred embodiments include vertical slices of data 704 - 707 aligned in a horizontal plane to form the composite image 708 .
  • the length of a horizontal plane defining a composite image 708 is only limited by the length of a continuum along which points are defined and from which image data 704 - 707 is captured.
  • the use of only slices of data 704 - 707 from any particular captured image provides for a higher quality image 708 .
  • the quality is increased, for example, when a temporary obstruction, such as a passing car, person or animal, captured in one image data set, is only represented in a thin slice of a continuous ribbon 704 - 707 data.
  • alignment of multiple thin slices of image data 704 - 707 is facilitated from the perspective of which aberrations typical human sensory is capable of distinguishing. Numerous thin slices 704 - 707 are perceived as a smoother transition across the horizontal plane defining the composite image 708 removing a choppiness that may be experienced with the use of larger data images.
  • the present invention can include a uniform width of each slice of data 704 - 707 or a variable width.
  • the width of a particular slice may vary, for example, upon one or more of the velocity of a vehicle from which image data sets are captured, the sample rate of a camera used to capture an image data set 701 - 703 , the resolution of a picture comprising an image data set 701 - 703 and the path of a camera.
  • a high resolution image generated by a 2.1 mega pixel camera may have a 1600 by 1200 resolution and allow for a thinner slice 704 - 707 that includes a width of between about 5 to 700 pixels of an image data set.
  • Embodiments with very high quality can include a slice 704 - 707 of between 1 to 2 pixels.
  • some embodiments can include a slice 704 - 707 of an image data set 701 - 703 that includes a percentage of the image data set 701 - 703 . Accordingly, some preferred embodiments can include a slice 704 - 707 of between about 5% to about 12% of an image data set. Other preferred embodiments can include up to about 50% of an image data set. However, it should be understood that some embodiments include a slice 704 - 707 that constitutes any fraction of the complete image data set.
  • a database or other data processing mechanism can track each portion or slice 704 - 708 utilized to construct a continuum and associate the slice 704 - 708 with an original image 701 - 703 from which the slice 704 - 708 is derived.
  • User interactive devices can execute the retrieval of an entire original image 701 - 703 or a series of original images 701 - 703 upon request. In some instances, one or more original images 701 - 703 may provide detail not included in the composite image 708 .
  • a selected portion of an image data set may be determined by the physical attributes of the equipment used to capture an image data set.
  • a typical camera lens can impart some distortion to an image data set, as represented in the illustration by an elongated portion and a compressed portion.
  • Utilizing only a portion of an image data set, such as, for example, a center portion vertical slice, can minimize the effect of distortion introduced by a lens, or other source, to a composite image. Distortion is minimized when the composite image is made to include those portions of the image data set corresponding with the center of the lens.
  • Specific embodiments may likewise account for other aberrations that may be present in a set of image data.
  • the length of a composite image generated according to the present invention is limited only by the ability to capture image data from additional points on a continuum and store the captured image data for post processing.
  • the post processing allows for the alignment of portions of the image data compiled into a composite two-dimensional view that can continue so long as additional image data is made available to be added to it.
  • FIG. 25 illustrates a controller 800 that may be utilized to implement some embodiments of the present invention.
  • the controller 800 comprises a processor unit 810 , such as one or more processors, coupled to a communication device 820 configured to communicate via a communication network (not shown in FIG. 25 ).
  • the communication device 820 may be used to communicate, for example, with one or more online devices, such as a personal computer, laptop or a handheld device.
  • the processor 810 is also in communication with a storage device 830 .
  • the storage device 830 may comprise any appropriate information storage device, including combinations of magnetic storage devices (e.g., magnetic tape and hard disk drives), optical storage devices, and/or semiconductor memory devices such as Random Access Memory (RAM) devices and Read Only Memory (ROM) devices.
  • RAM Random Access Memory
  • ROM Read Only Memory
  • the storage device 830 can store a program 840 for controlling the processor 810 .
  • the processor 810 performs instructions of the program 840 , and thereby operates in accordance with the present invention.
  • the processor 810 may also cause the communication device 820 to transmit information, including, in some instances, control commands to operate apparatus to implement the processes described above.
  • the storage device 830 can additionally store related data in a database 830 A and database 830 B, as needed.
  • apparatus includes a video and data server farm.
  • the video and data server farm includes at least one video storage server that stores video image files containing video drive-by data that corresponds to a geographic location, a database server that processes a data query received from a user over the Internet that corresponds to a geographic location of interest, and an image server.
  • the database server identifies video image files stored in the video storage server that correspond to the geographic location of interest contained in the data query, and transfers the video image files over a pre-processing network to the image processing server.
  • the image processing server converts the video drive-by data to post-processed video data corresponding to a desired image format, and transfers the post-processed video data via post-processing network to the Internet response to the query.
  • a landing zone server can also be included which receives the video drive-by data from a portable memory device and permits the viewing and analysis of the video drive-by data prior to storage in the video storage server. Still further, a map server is preferably provided to present a static image an overhead view of the geographic location of interest.
  • Embodiments can also include one or more servers described above included in one or more physical units. Each server does not need to be a disparate apparatus. Still other embodiments can include one or more or the servers described above included in multiple physical units. Some embodiments can even include a single server, as described which includes multiple physical apparatus units at disparate locations.
  • a composite image 100 is formed by aligning two or more of the image data sets. Unlike stitching processes previously known, the present invention creates a composite through alignment of portions of data from more than one data set. Alignment can be accomplished in image data processing. Using image data processing, the images 101 - 103 are aligned to form a composite image 100 .
  • the composite image 100 is essentially two dimensional image data arranged as a second continuum, or ribbon.
  • the second continuum includes ongoing image data 101 - 103 captured from the points defining the first continuum.
  • the series of points of image capture in the first continuum includes positions of a vehicle carrying an image capture device, such as a camera, as the vehicle traverses a path proximate to a geographic area.
  • the camera is positioned to capture image data of the geographic area.
  • Image data 101 - 103 is periodically captured as the vehicle traverses the path. The motion of the vehicle, combined with the periodic capture of image data 101 - 103 , thereby results in image data 101 - 103 being captured from disparate points along the first continuum.
  • a preferred embodiment includes capture of image data with a camera maintained orthogonal to the subject. Orthogonal capture of the image data provides consistency for subsequent composite of portions of the image data captured. Therefore, data captured at an angle of between about 75° and 105° can provide data most easily assembled into a continuum of data. However, other angles may be used to create different effects.
  • a continuous pictorial representation includes RibbonViewTM by Real Data Systems. RibbonViewTM correlates a ribbon of geographic image data with geospatial designations to facilitate identification of a particular geographic area, as discussed more fully below.
  • image capture processing can be accomplished in real time or via post image capture processing.
  • select portions 104 - 107 of two or more sets of captured image data are aligned to generate the composite image 108 .
  • some preferred embodiments include vertical slices of data 104 - 107 aligned in a horizontal plane to form the composite image 108 .
  • the length of a horizontal plane defining a composite image 108 is only limited by the length of a continuum along which points are defined and from which image data 104 - 107 is captured.
  • the use of only slices of data 104 - 107 from any particular captured image provides for a higher quality image 108 .
  • the quality is increased, for example, when a temporary obstruction, such as a passing car, person or animal, captured in one image data set, is only represented in a thin slice of a continuous ribbon 104 - 107 data.
  • alignment of multiple thin slices of image data 104 - 107 is facilitated from the perspective of which aberrations typical human sensory is capable of distinguishing. Numerous thin slices 104 - 107 are perceived as a smoother transition across the horizontal plane defining the composite image 108 removing a choppiness that may be experienced with the use of larger data images.
  • the present invention can include a uniform width of each slice of data 104 - 107 or a variable width.
  • the width of a particular slice may vary, for example, upon one or more of the velocity of a vehicle from which image data sets are captured, the sample rate of a camera used to capture an image data set 101 - 103 , the resolution of a picture comprising an image data set 101 - 103 and the path of a camera.
  • a high resolution image generated by a 2.1 mega pixel camera may have a 1600 by 1200 resolution and allow for a thinner slice 104 - 107 that includes a width of between about 5 to 100 pixels of an image data set.
  • Embodiments with very high quality can include a slice 104 - 107 of between 1 to 2 pixels.
  • some embodiments can include a slice 104 - 107 of an image data set 101 - 103 that includes a percentage of the image data set 101 - 103 . Accordingly, some preferred embodiments can include a slice 104 - 107 of between about 5% to about 12% of an image data set. Other preferred embodiments can include up to about 50% of an image data set. However, it should be understood that some embodiments include a slice 104 - 107 that constitutes any fraction of the complete image data set.
  • each set of image data 201 - 203 represents data captured from a disparate point on a continuum proximate to the landscape captured.
  • a direction of travel along the continuum 200 is also illustrated.
  • Each data set 201 - 203 is divided into multiple slices of image data 204 - 209 .
  • a composite image 210 is generated from the multiple slices 204 - 209 , and in the exemplary case at hand, additional slices from additional image data sets.
  • a database or other data processing mechanism can track each portion or slice 204 - 208 utilized to construct a continuum and associate the slice 204 - 208 with an original image 201 - 203 from which the slice 204 - 208 is derived.
  • User interactive devices can execute the retrieval of an entire original image 201 - 203 or a series of original images 201 - 203 upon request. In some instances, one or more original images 201 - 203 may provide detail not included in the composite image 208 .
  • a selected portion 303 of an image data set 300 may be determined by the physical attributes of the equipment used to capture an image data set.
  • a typical camera lens can impart some distortion to an image data set, as represented in the illustration by an elongated portion 301 and a compressed portion 302 .
  • Utilizing only a portion 302 of an image data set 300 such as, for example, a center portion vertical slice 303 , can minimize the effect of distortion introduced by a lens, or other source, to a composite image 108 . Distortion is minimized when the composite image 108 is made to include those portions of the image data set corresponding with the center of the lens 303 .
  • Specific embodiments may likewise account for other aberrations that may be present in a set of image data.
  • positional data descriptive of a location of the subject matter of an image can also be generated by the image data recording device.
  • Positional data can include any data indicative of where the subject matter of an image is located.
  • Some preferred embodiments can include Cartesian coordinates that are context sensitive according to the mechanism used to generate coordinate data.
  • an image recording device such as a camera
  • a device for recording a global position such as a global positioning system (GPS) device or other equipment.
  • GPS global positioning system
  • Time stamps associated with image data and time stamps associated with the GPS data can be utilized to correlate the GPS location data with image data recorded by the camera.
  • an altimeter can be used to record an altitude from which a camera records image data sets.
  • the altitude data can be associated with an image data set, for example, metadata correlated with the image data set.
  • Such embodiments can therefore include recordation of a latitude, longitude and altitude coordinate for a given image data set.
  • geospatial data such as latitude and longitude coordinates
  • GPS data can be time stamped and collected once every second.
  • GPS reception can be interrupted, depending upon location relative to large object, such as multistory buildings, or cold cover. Therefore, some additional embodiments can include an accelerometer for recording motion associated with a camera and a GPS unit operatively attached to the camera.
  • Data from the accelerometer can indicate movement of the camera.
  • Some accelerometers such as micro electromechanical system (MEMs) accelerometers can easily be incorporated into a camera system assembly attached to a vehicle.
  • MEMs micro electromechanical system
  • Use of multiple MEM accelerometers positioned to measure movement in four or more directions along an x-axis, y-axis, and z-axis in relation to a camera can also be used to calculate direction of movement.
  • the accelerometer can therefore be used to extrapolate a current position of the camera, based upon a last set of GPS geospatial data recorded.
  • Geospatial data can be used to indicate an initial geographic position.
  • a change in geospatial data can be additionally utilized to indicate velocity and direction of image data set capture.
  • Accelerometer data may also be used to indicate a velocity and direction of image data set capture.
  • Accelerometer data may also be used to indicate a velocity and direction of camera movement. Calculations of time elapsed at the indicated velocity (such as for example, the Kalman Filter) can yield a calculated position at a time of image capture, even if the time of image capture is between GPS readings.
  • one standard can include tracking a camera position with a GPS unit that records location at a rate of once per second.
  • the camera can record image data at a faster rate than once per second, such as, for example, one of: 12 images per second, 24 images per second or 29.97 images per second.
  • An initial camera position can be recorded which correlates with a GPS reading, subsequent image data capture will occur in between the one second GPS reading interval.
  • the camera position can be determined with a calculation based upon the velocity of camera movement supplied by the accelerometer and time elapsed since a previous GPS reading.
  • Still other embodiments can utilize optical flow methodology and visual odometry to facilitate calculations of a camera position and the velocity of a vehicle or person from which a series of image data sets are captured.
  • Visual odometry can be accomplished with a single omni-directional camera or with stereo cameras, and is based upon processing which tracks the position of salient features in a series of feature sets and calculates movement based upon the relative positions of the features.
  • camera based simultaneous localization and mapping (SLAM) of visual image data can also be used to facilitate calculations of velocity of a change in position of a camera used to capture image data sets.
  • the velocity will be directly tied to the motion of a vehicle to which the camera is mounted, or a person carrying a camera rig.
  • Orientation of a camera can include a direction of image capture recorded by the camera. Orientation can be designated, for example, in relation to the cardinal directions, i.e. north, south, east and west. Any means available to record such a designation, such as an electronic compass, is within the scope of the present invention. However, it may be desirable to include a means to record the orientation with a greater degree of accuracy than is typically available through the use of an electronic compass.
  • orientation can be determined according to a fixed position of a camera in relation to the direction of travel of a vehicle (or person) used to transport the camera.
  • a plurality of cameras can be fixedly attached to a vehicle capturing Video DriveByTM data.
  • Each camera therefore maintains a constant direction of image capture in relation to the heading of the vehicle.
  • Mechanics of the camera such as, for example, lens parameters and shutter speed, can indicate a depth of field during camera image data capture.
  • Some embodiments can also include simultaneously capturing multiple image data sets and correlating two or more of the image data sets. Correlation can be accomplished via a time stamp or other chronological or synchronous.
  • the position of a camera can be combined with a direction of image capture and the depth of field of the camera, to determine a location of image data captured by the camera at a particular instance in time.
  • the present invention can also include apparatus for utilizing echo location to determine a distance of an object from a camera capturing an image data set and storing a correlated distance with the image data set.
  • radar data can be correlated with a camera image data set to calculate the location of various objects captured by the camera.
  • a time stamp can also be combined with data to quantify a location for a particular image formed by the captured image data.
  • data used to calculate a location of an image is stored in a metadata file space associated with the image data.
  • some embodiments can store metadata in the exchangeable image file format (EXIF), TIFFTAGS or International Press Telecommunication Council (IPTC) formats.
  • Image data may be stored, for example in JPEG or TIFF formats.
  • other metadata formats can also be used.
  • the image data and metafile data are stored on an external data storage device, such as a hard disk drive operatively attached to the camera.
  • the data can be stored in the camera.
  • the metadata can include data descriptive of some or all of: date and time; camera settings such aperture, shutter speed and focal length; geospatial data from a GPS receiver unit; accelerometer data; inertial guidance system data; camera orientation; and camera fixed position related to vehicle travel.
  • the RibbonViewTM illustrated 501 - 502 includes strips or ribbons of two-dimensional image data. Each strip 501 - 502 is generated via the capture of image data along a continuum 503 . As illustrated, the continuum 503 generally follows the roadway in the center of the illustration. Each composite image 501 - 502 is generated through the alignment of a portion of multiple images captured as data from disparate points along the continuum 503 . In some preferred embodiments, the present invention delivers two ribbons of composite images 501 - 502 with different views captured along the continuum. Each view is generated by a camera with a separate orientation.
  • the length of a composite image generated according to the present invention is limited only by the ability to capture image data from additional points on a continuum and store the captured image data for post processing.
  • the post processing allows for the alignment of portions of the image data compiled into a composite two-dimensional view that can continue so long as additional image data is made available to be added to it.
  • FIG. 32 illustrates a controller 600 that may be utilized to implement some embodiments of the present invention.
  • the controller 600 comprises a processor unit 610 , such as one or more processors, coupled to a communication device 620 configured to communicate via a communication network (not shown in FIG. 32 ).
  • the communication device 620 may be used to communicate, for example, with one or more online devices, such as a personal computer, laptop or a handheld device.
  • the processor 610 is also in communication with a storage device 630 .
  • the storage device 630 may comprise any appropriate information storage device, including combinations of magnetic storage devices (e.g., magnetic tape and hard disk drives), optical storage devices, and/or semiconductor memory devices such as Random Access Memory (RAM) devices and Read Only Memory (ROM) devices.
  • RAM Random Access Memory
  • ROM Read Only Memory
  • the storage device 630 can store a program 640 for controlling the processor 610 .
  • the processor 610 performs instructions of the program 640 , and thereby operates in accordance with the present invention.
  • the processor 610 may also cause the communication device 620 to transmit information, including, in some instances, control commands to operate apparatus to implement the processes described above.
  • the storage device 630 can additionally store related data in a database 630 A and database 630 B, as needed.
  • the camera array may be implemented with cameras that directly produce a digital output.
  • digital cameras may be employed instead of analog cameras with separate A/D converters.
  • the number of cameras in the array may also vary.
  • the preferred embodiment utilizes external hard disk drives, it will be understood that any type of portable memory device having sufficient capacity to store the video data may be utilized.
  • the various processing methods may be implemented utilizing the servers of the Video and Data Server Farm, the processing unit, i.e. the laptop computer, of the Video Capture System, and/or other processors linked thereto through the Internet or other network means, including various combinations of all three. Accordingly, other embodiments are within the scope of the following claims.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Strategic Management (AREA)
  • Data Mining & Analysis (AREA)
  • Finance (AREA)
  • Accounting & Taxation (AREA)
  • General Business, Economics & Management (AREA)
  • Economics (AREA)
  • Marketing (AREA)
  • Multimedia (AREA)
  • Development Economics (AREA)
  • Software Systems (AREA)
  • Tourism & Hospitality (AREA)
  • Library & Information Science (AREA)
  • Human Resources & Organizations (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Geometry (AREA)
  • Signal Processing (AREA)
  • Computer Graphics (AREA)
  • Remote Sensing (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Primary Health Care (AREA)
  • Game Theory and Decision Science (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Computer Hardware Design (AREA)
  • Human Computer Interaction (AREA)
  • Computational Linguistics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Artificial Intelligence (AREA)
  • Instructional Devices (AREA)
  • Processing Or Creating Images (AREA)

Abstract

Video drive-by data provides a street level view of a neighborhood surrounding a selected geographic location. A video and data server farm incorporates a video storage server that stores video image files containing video drive-by data corresponding to a geographic location, a database server that processes a data query received from a user over the Internet corresponding to a geographic location of interest, and an image processing server. In operation, the database server identifies video image files stored in the video storage server that correspond to the geographic location of interest contained in the data query and transfers the video image files over a pre-processing network to the image processing server. The image processing server converts the video drive-by data to post-processed video data corresponding to a desired image format and transfers the post-processed video data via a post-processing network to the Internet in response to the query.

Description

    CROSS REFERENCE TO RELATED APPLICATION
  • This application is a continuation of, and claims priority under 35 U.S.C. §120 from, nonprovisional U.S. patent application Ser. No. 13/481,912 entitled “Three Dimensional Image Data Models,” filed on May 28, 2012. Application Ser. No. 13/481,912, in turn, is a continuation of, and claims priority under 35 U.S.C. §120 from, nonprovisional U.S. patent application Ser. No. 12/036,197 entitled “Methods and Apparatus for Generating Three-Dimensional Image Data Models,” filed on Feb. 22, 2008, now U.S. Pat. No. 8,207,964. Application Ser. No. 12/036,197, in turn, is a continuation-in-part of, and claims priority under 35 U.S.C. §120 from, nonprovisional U.S. patent application Ser. No. 11/216,465 entitled “Apparatus and Method for Producing Video Drive-By Data Corresponding to a Geographic Location,” filed on Aug. 31, 2005, now U.S. Pat. No. 7,389,181. Application Ser. No. 11/216,465, in turn, claims priority under 35 U.S.C. §120 from provisional U.S. patent application Ser. No. 60/605,498 entitled “Apparatus and Method for Producing Video Drive-By Data Corresponding to a Geographic Location,” filed on Aug. 31, 2004. The subject matter of each of the foregoing documents is incorporated herein by reference.
  • TECHNICAL FIELD
  • The present application relates in general to systems and methods of generating image data of a geographic location of interest. Specifically, the invention relates to generating video drive-by data corresponding to a selected geographic location. The video drive-by data provides a street level view of a neighborhood surrounding the selected geographic location, and can be further combined with other data sources in a variety of applications. The invention also relates to generating data descriptive of a continuum of images, such as those comprising a geographic setting.
  • BACKGROUND INFORMATION
  • Companies and individuals looking for real estate to rent, buy, insure or finance have historically had to physically travel to the property in order to visually verify the condition of the property. With the advent of the Internet, it has become possible for individuals to view photographs of properties on-line in the comfort of their home or office. One such example is the MLS (Multiple Listing Service) database, which can be accessed by the public through Internet sites such as www.realtor.com. The MLS database typically includes one or more photographs and additional data regarding the property such as the number of bedrooms and the square footage. More advanced systems may include short video clips constituting virtual tours of the property in question. While such systems may also show views of a property from the street, they do not give a user a sense of the neighborhood in which the property is located. Accordingly, in order to determine if the area surrounding the property of interest is desirable, a user may often be required to physically travel to the property in order to view the state of the neighborhood.
  • In order to overcome the deficiencies of present systems of the type described above, it would be desirable to provide a system that would enable a user to view the surrounding neighborhood as well as the property itself. Such a system would enable the creation of a complete on-line database of street-level video data, hereinafter referred to as “video drive-by” data, which can be made available to end users via the Internet. Preferably, the video data would be encoded with GPS coordinates as well as conventional street addresses and map coordinates to provide instant street level video of any location upon request. For example, such a system would be extremely useful if a user could query the system by submitting GPS coordinates, street addresses or by “clicking” on a map overlay showing streets in the desired neighborhood.
  • The ability of a system to provide data associated with the neighborhood, as well as a particular property in question, would be extremely desirable in a number of government and commercial applications. Aiding in the process of appraising a property, for example, is just one application in which such a system can provide benefits. Mortgage lenders currently purchase several types of appraisal products as part of their financing decision process. The lowest cost appraisal is an automated valuation model (AVM), whereby county parcel data of subject property can be compared to the county parcel data of a list of nearby properties (called “comps” which is short for comparables) with a similar lot size, square footage, and amenities (such as waterfront access). A comp list typically includes average per square foot calculations of the similar properties that enables the creation of approximate value of the subject property. The next lowest cost appraisal is a Desk Review whereby an appraiser reviews the AVM data on behalf of the mortgage company and evaluates the relevance of the data presented in the AVM in conjunction with what they personally know regarding the area the subject is located in. A separate class of field appraisals includes both a Drive By Appraisal and a Full Interior Appraisal. A Drive By Appraisal is where an appraiser drives by and shoots a photograph from the street of a subject property and it's comps. These photographs are linked to county parcel data and AVM calculations to provide a more accurate report to the mortgage company. Finally, for a Full Interior Appraisal which is the most expensive and considered most accurate appraisal, an appraiser will tour the interior of a property taking numerous photographs, measuring and sketching a floor plan as well as notating the general quality of construction and upkeep, this data is then presented as a multi property report which includes tabular data of subject and comp data with comp photos.
  • In the case of three of four appraisal types, however, a mortgage banker must wait several days before receiving the information from an appraiser. If a system were available to allow quick access to all properties within a neighborhood, a mortgage banker could obtain photographs of a subject property and its comps and confirm that the comparables as selected by the AVM represents properties that have the same visual appearance of the subject property, thereby providing faster loan commitments and closings while at the same time lowering costs. Similarly, instant Drive By Appraisals could be provided thereby eliminating the costs and delays associated with an appraiser driving to manually photograph a subject property and its comps.
  • Images of geographic areas created on a local basis have been created in numerous formats. Photographs, movie cameras, video camera recorders, and more recently digital recorders have all been utilized to capture moving images of a geographic parcel. Photographs can be useful because they are easily associated with a particular real estate parcel; however, they are limited in the scope of the subject matter that they can include. In addition, photographs must be manually acquired and docketed in order to be associated with a property and subsequently retrieved. Panoramas can increase the scope of the subject matter to a point, but are limited by a point of capture of the panoramic views.
  • Movie cameras, video recorders and digital recorders and other “motion pictures” provide for increased scope of image capture. However, it is sometimes difficult to locate and view a particular portion of interest of images captured as motion pictures. In addition, correlation of particular portions of a motion picture with a particular real estate parcel can be difficult if the motion picture includes more than one real estate parcel. For example, it is relatively difficult to locate and view a particular real estate parcel in the context of its neighborhood setting, or particular aspects of its setting. Aerial images, such as satellite pictures of geographic areas have also been captured, and specific parcels of land or landmarks can be located on the aerial images.
  • In addition, methods and apparatus have been known for generating point clouds and generating representations of particular objects from processing of the point clouds.
  • However, prior to the present invention, there has not been a mechanism for generating a continuum of object representations based upon point cloud data. Nor has there been a mechanism for accurately correlating ground level images with substantial portions of an aerial image in a consistent and orderly format that allows for the identification of a particular parcel of land and provide both aerial and ground level views of the parcel, as well as a surrounding area of the parcel. Similarly, prior to the present invention, there has not been a method for correlating geopolitical indicators, such as property lot lines, or tax map parcels with aerial images and ground level video images.
  • Substantial portions of ground level images have not been correlated with aerial images or with geopolitical indicators, in part, because there has not been any image vehicle capable to providing a format of high quality and accurate representations of street level images capable of being matched to macro level image data.
  • SUMMARY
  • A system and method of providing video drive-by data is provided to enable a street level view of a neighborhood surrounding a selected geographic location. The video drive-by data can be further combined with other data sources related to the geographic location.
  • The invention is based, in part, on providing: 1) accurate differential GPS data; 2) post processing of geo positioning signals to smooth curves in motion (splines); 3) highly accurate camera position and video frame position analysis processing to provide a calculation of the exact position of each video frame; 4) parcel data processing that analyses vector line data that is geo-coded with latitude and longitude values; 5) digital image photos processed with image superimposition algorithms; and 6) a database that includes video image files, parcel latitude and longitude data, and positioning data that is indexed to the video image files. With these components, the invention enables the access to video images of any desired geographic location and its surrounding neighborhood, while relating such image data to other property related data such as property lines, landmarks, etc.
  • In a preferred embodiment, a system is provided that includes a video and data server farm. The video and data server farm includes at least one video storage server that stores video image files containing video drive-by data that corresponds to a geographic location, a database server that processes a data query received from a user over the Internet that corresponds to a geographic location of interest, and an image server. In operation, the database server identifies video image files stored in the video storage server that correspond to the geographic location of interest contained in the data query, and transfers the video image files over a pre-processing network to the image processing server. The image processing server converts the video drive-by data to post-processed video data corresponding to a desired image format, and transfers the post-processed video data via a post-processing network to the Internet in response to the query. A landing zone server is preferably provided to receive the video drive-by data from a portable memory device and permit the viewing and analysis of the video drive-by data prior to storage in the video storage server. Still further, a map server is preferably provided to present a static image of an overhead view of the geographic location of interest.
  • The video drive-by data is preferably captured by a video capture system that includes a camera array that generates video image data of a neighborhood corresponding to a geographic location, a positioning unit that produces positioning data corresponding to the location of the camera array, and a processing unit that processes the video image data and positioning data to produce the video drive-by data. The processing unit stores the video drive-by data on the portable memory device. In a preferred embodiment, the camera array includes a plurality of camera units arranged to provide a 360 degree view of the neighborhood. Further, the processing unit preferably provides time stamp data and file pointer data that permits synchronization of video images to precise locations of individual frames.
  • The above-described system can further be provided with a plurality of processing means to add additional features or permit utilization of the system for specific applications. For example, processing means may be incorporated to provide a variety of geo-coded data layers in conjunction with the video drive-by data. Similarly, processing means can be provided for: creating geo-coded text, image and vector data which is superimposed onto the post-processed video data that is transferred to the Internet; linking latitude and longitude data with video drive-by data; calculating a camera position of designated individual frames of video drive-by data using position data points captured before and after a designated frame position; for examining changes of relative position in the horizontal position of designated landmarks in a sequence of video image frames to create fixed landmark data points; examining images that match common image patterns to enable the creation of an accurate vector map of roads; providing OCR processing to generate a database of recognized text that is geo-coded with latitude and longitude coordinates based on camera orientation and relative position; and selecting video drive-by data or a given street address by calculating a centroid of a parcel and determining camera orientation and position to enable highly accurate delivery of video that is pointed toward the centroid of a parcel.
  • Still further embodiments can include: processing means by which a sequence of video frames that are oriented toward a designated parcel and it's adjacent parcels can be stitched together to form a long horizontal image of a street which can be created with superimposed property county parcel data placed above the centroid of the parcel centered above the property; processing means for providing authenticated annotation of information or comments to a photo and or video; processing means for uploading a digital image photograph to the video and data server farm and comparing the uploaded photo to a library of images captured by the video capture system; and/or processing for comparing digital image photos of building structures to generate a set of numeric values representing line segments of sharp horizontal, vertical and angular lines that are stored as vectors each with common ratios of distances from each other, wherein the list of vector line segments can be searched with a high-speed database query to generate a list of possible matches of a subject property against a large database of properties with vector line data on file.
  • The present invention can be utilized in a variety of applications. For example, a system in accordance with the invention can offer invaluable assistance to local, county or state road and highway departments in planning new projects or tracking existing placements of signage, traffic signals, utility boxes, and other facilities and equipment locations. Administrators can preview and familiarize themselves with a targeted location in conjunction with assessing needs or assigning work details.
  • New layers of routing assistance can also be provided to emergency vehicle dispatching systems by including roadside views to allow rapid and accurate targeting of a destination. Tied in directly to county GIS mapping, the present invention can offer street level video into the neighborhood and directly to the street address, offering photographs of any property transmitted via wireless Internet to help the fire, police or rescue driver find his or her way.
  • With the capability to fully integrate, refine and unify existing county property records and GIS mapping resources, the system can enhance online presentations with parcel line enhanced ortho imagery, property images and now with video. Further, with the system, unification of property records and county mapping can evolve to tie in to one-click access to street level video segments allowing a user to virtually drive down any street, while highlighting any property and its surrounding neighborhood with parcel lines drawn on the photograph.
  • Still further, local development agencies wanting to attract business and industrial relocations to their communities can insure an edge over competing community bids by offering one-click access to visually tour planned growth areas. Using integrated aerial mapping, property data and data generated by the system, users can scout available development areas, accessing integrated mapping layers to identify a development area's proximity to rail, highways, air and port facilities, while also including street level video footage of local amenities, residential neighborhoods, schools, recreation and other visual explorations.
  • Similarly, tourism boards at any level can promote travel and recreation to a broader audience by integrating visuals and streaming video in presenting points of interest, accommodations, and dining as integrated layers to the local, county or state map that are unified to promotional content. The video drive-by data generated by the system can offer compelling video segments highlighting parks, beaches, hunting, fishing, arena sports, museums and places of historical interest; while, also providing directions, drive maps, and links to information detailing places to stay, restaurants, rental agencies and more.
  • In non-governmental applications, the system can be utilized to communicate the value and desirability of a new subdivision or location to potential builders and residents. For example, a property developer's web site can be switched on to enable parcel line enhanced satellite photography with access to street level video data of the newly paved roads of the development. Web site visitors would have the freedom of taking a virtual tour of any selected parcel of property, looking around to note the scenery, tree line, and proximity to neighboring lots, thereby providing a true feel for the land within its natural surroundings. Data and content can be integrated to any parcel with a ready feedback mechanism allowing viewers to inquire further or request an appointment.
  • Similarly, property appraisers can gain new efficiencies at reduced costs by utilizing a video drive-by data archive of streaming street level video to select and capture freeze-frame video images of homes for sale needing an exterior photo. The system's unification of property data, MLS records, mapping with the video resources allows any property to be quickly selected, analyzed, and compared to equivalent property listings, generating reports with active links to both data and visuals that can be emailed at a moments notice.
  • The present invention also relates to methods and apparatus of generating image data. More specifically, the present invention also relates to generating data descriptive of a continuum of three dimensional images, such as those comprising a geographic landscape. Accordingly, the present invention provides methods and apparatus for generating a continuum of three dimensional image data. In some embodiments, the continuum of three dimensional image data provides a street level representation of geographic locations. Two or more sets of image data are captured of a subject area, wherein each of the two or more images are captured from disparate points on a continuum
  • In another embodiment, methods and apparatus are provided for generating a three dimensional output that includes a continuum of image data sprayed over three-dimensional models. The three-dimensional models can be representative of features captured by the image data, wherein image data can be captured at multiple disparate points along another continuum. The user interface can also include multiple modalities of image data and statistical analysis of the image data.
  • Yet other embodiments provide methods and apparatus for generating a continuum of image data. The continuum of image data provides a two dimensional ribbon of street level views of geographic areas. The ribbons are created by capturing two or more images of a subject, wherein each of the two or more images are captured from disparate points on a continuum. For example, each of the two or more images can be captured from a different position on a street as a vehicle drives down the street. Generally, the images will be captured from an orthogonal view of the subject.
  • Portions of the two or more images are then aligned in a dimension consistent with the continuum, wherein, according to our example the continuum includes the path of the vehicle. Therefore, the images would be aligned in a dimension consistent with the path of the vehicle. A composite image of the subject is generated from the aligned portions of the two or more images.
  • In some embodiments, positional data is recorded that is descriptive of a respective location of each of the two or more images. The positional data can include, for example, latitude and longitude coordinates and can be used to associate the composite image with a particular portion of the subject. It is also within the scope of the invention to record the altitude of a camera used to capture the image data and thereby approximating the altitude of the subject matter of the image data.
  • In some embodiments, a camera will be maintained approximately orthogonal to the subject captured and therefore, those embodiments will maintain the camera at an angle about between 75° and 105° in relation to the subject matter.
  • In another aspect, some embodiments can include various overlays of information on top of the continuum of image data. Overlays can include one or more of: metadata, data related to a composite image, and geospatial information.
  • Other advantages, features and applications of the invention will become apparent to those skilled in the art from the following detailed description of the preferred embodiments of the invention and the accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The invention will be described in detail with reference to certain preferred embodiments thereof and the accompanying drawings, wherein:
  • FIG. 1 is a schematic diagram of a Video and Data Server Farm in accordance with a preferred embodiment of the invention;
  • FIG. 2 is a schematic block diagram of a Video Capture System that supplies video data and corresponding synchronous GPS data to the Video and Data Server Farm illustrated in FIG. 1;
  • FIG. 3 is a process flow diagram illustrating the process of capturing the video data and synchronous GPS data utilized by the Video Capture System illustrated in FIG. 2;
  • FIG. 4 is a process flow diagram illustrating an efficient process in accordance with the present invention by which numerous field drivers with Video Capture Systems of the type illustrated in FIG. 2 can be continually supplied with hard disk drives to capture data;
  • FIG. 5 illustrates a web browser view with an address bar pointing to a web page that shows both an ortho image view of a mapserver and video pane in accordance with the invention;
  • FIG. 6 is a process flow diagram for superimposing geo-coded text data over video captured by the Video Capture System of the type illustrated in FIG. 2;
  • FIG. 7 is a data flow diagram that illustrates the process by which an instant drive-by appraisal or Photo AVM (Automated Valuation Model) can be created using a Video drive-by database that contains geo-coded video;
  • FIG. 8 illustrates a top perspective of a Video Capture System as it drives past a structure;
  • FIG. 9 illustrates a frame position interpolation in accordance with the present invention;
  • FIG. 10 shows a set of four illustrations which represent a single frame or portion of a window within an Internet Web Page in accordance with the invention;
  • FIG. 11 illustrates a process by which images captured by a left facing and a left front facing camera in a Video Capture System can calculate and generate superimposed property;
  • FIG. 12 illustrates a process to extract and create a point of interest database that is encoded with GPS data;
  • FIG. 13 illustrates two ortho views of a Video Capture System as it drives in a direction of travel along a path past a selected parcel where one of six possible cameras may be viewing a subject property;
  • FIG. 14 illustrates a process by which an authenticated end user can submit data regarding a specific location using Video drive-by data as a geo-coding reference tool;
  • FIG. 15 illustrates a process and visual references as to the calculations required to extract a photograph from a street number in accordance with the invention;
  • FIG. 16 illustrates a novel method of finding a house from a photograph based on the invention;
  • FIG. 17 illustrates one novel method of identifying a home from a large library of home photos with a high speed process in accordance with the invention;
  • FIG. 18 illustrates a continuum from which image data sets capture a subject;
  • FIG. 19 illustrates multiple images captured from disparate points along a continuum;
  • FIG. 20 illustrates multiple image data sets that overlap subject matter;
  • FIG. 21 illustrates a three-dimensional model generated form the image data sets;
  • FIG. 22 illustrates a user interface generated from the image data sets and including multiple modalities of image data;
  • FIG. 23 illustrates an exemplary user interface for traversing multiple modalities of image data;
  • FIG. 24 illustrates an exemplary composite of image data set portions;
  • FIG. 25 illustrates apparatus what may be used to implement the present invention;
  • FIG. 26 illustrates basic geographic images captured from a camera at disparate points along a continuum and aligned with post processing;
  • FIG. 27 illustrates an alignment of portions of two or more images captured from disparate points along a continuum to form a composite image;
  • FIG. 28 illustrates multiple images captured from disparate points along a continuum and specific vertical slices that can be selected to create a composite image;
  • FIG. 29 illustrates a block diagram of exemplary lens distortion factors and a preferred area of lens capture that can be selected in some embodiments of the present invention;
  • FIG. 30 illustrates disparate points on a continuum with an orientation vector associated with each respective point;
  • FIG. 31 illustrates a canyon formed by two continuums of image data according to the present invention; and
  • FIG. 32 illustrates apparatus that can be used to implement some embodiments of the present invention.
  • DETAILED DESCRIPTION
  • A Video & Data Server Farm 0100 in accordance with a preferred embodiment of the invention is illustrated in FIG. 1. The Video & Data Server Farm 0100 utilizes a set of servers to provide digital storage of video drive-by data, and processes the video drive-by data for delivery to an end user. The video drive-by data preferably includes video data and Global Positioning System (GPS) data or “geo position data” generated by a Video Capture System 0201 illustrated in FIG. 2. Specifically, video drive-by data is preferably stored on an external hard disk drive (HDD) 0250 of the Video Capture System 0201, and is then copied from the external HDD 0250 to a landing zone server 0130 of the Video & Data Server Farm 0100 for quality analysis and viewing via industry standard commercially available MPEG2 viewing software and file editors. The Video drive-by data, including the video data and GPS data, is transferred via a load network 0142 to be stored as video image files in video storage servers 0145 of the Video & Data Server Farm 0110.
  • In operation, a user logs onto the Internet 0150 via an end user interface 0102 to access the video image files containing the converted video drive-by data via a data query that includes data related to a specific geographic location of interest. The data query received by a web server 0105 and carried over a control network 0135 to a database server 0120 and a map server 0125, which preferably calculates and presents a static image (for example a street map image) of an overhead view of the geographic location of interest. The database server 0120 identifies the appropriate video image files corresponding to the location of interest from the video storage servers 0145 for transfer over a pre-processing network 0140 to image processing servers 0115. The image processing servers 0115 convert the original video drive-by data to one of many potential new image formats (depending on the particular application) which constitute Post-processed Video Data (PPVD). The PPVD is then transferred to a video server 0110 over a post-processing network 0141, which then forwards the PPVD data to the Internet 0150 in response to the query. In addition, if so desired, the video server 0110 can also concurrently transfer the PPVD back to the video storage servers 0145, database servers 0120 and map server 0125 for future access from a video cache.
  • FIG. 2 is a schematic block diagram of the components in the Video Capture System 0201. In a preferred embodiment, a camera unit 0205 preferably containing six video cameras 0210 in a hexagonal array is provided in a camera housing assembly 0215. The camera housing assembly 0215 is mounted with differing configurations on top of a vehicle (not shown) preferably using commercially available roof racks and straps. The electronics for converting and storing the video data out of the camera 205 are preferably provided in a passenger cab 0220 of the vehicle. The Video Capture System 0201 is preferably powered from the vehicles electrical system, although a separate power source can be provided if so desired. The camera 0205 feeds analog video signals 0227 along six cables in a cable harness 0225 which is fed through a small opening in a window of the vehicle. The analog video signals 0227 are fed into six analog-to-digital (A/D) converters 0230, which convert the analog video signals 0227 to compressed digital video data. The compressed digital video data is then sent on data cables 0232 into a seven port USB hub 0235. The six streams of digital video data are fed via the hub 0235 into a laptop computer 0245, which preferably writes the digital video data onto a portable storage device, for example the external HDD 0250, simultaneously with the data from a GPS receiver 0240. Accordingly, the data is stored with corresponding GPS synchronized coordinates in the external HDD 0250 as the Video drive-by data.
  • FIG. 3 is the block diagram illustrating the process of capturing the video data and synchronous GPS data to create the video drive-by data. A set of programs is automatically launched when power is applied to the laptop computer 0245. The GPS capture program 0305 processes the data transmitted from the GPS receiver 0240 and writes 0307 this as time stamped data to the laptop computer's RAM 0330 and to a file on the external HDD 0250. The time stamp of the data file written contains the laptop computer's internal system clock time data as well as the GPS clock data. A driver viewer application 0310 is initialized to enable the driver of the vehicle to confirm that the GPS receiver 0240 is working, and—as the video data is captured—the driver can preferably view segments of the video data on the display of the laptop computer 0245 to confirm the system is working properly. The driver viewer application 0310 shows a map of roads with color codes representing areas that have already been surveyed in one color and streets that need to surveyed in a second color.
  • Video capture daemons 0315 are started to process the video drive-by data arriving via the USB hub to then write 0317 the six data files to the external HDD 0250 with the laptop computer's 0245 system time clock time stamp data and file pointer data added to the video file and a GPS log file. In addition, the video capture daemons writes 0317 the file pointer data to the laptop computer's RAM 0330 for use by a master index file. The master index file daemon 0340 is started which reads the GPS data, the six video file pointers values and then writes 0345 the master index file data to the external HDD 0250. The entire set of data is then copied to the landing zone server 0130 when the external HDD drives are sent to the Video & Data Server Farm 0100 for upload and processing as will be described in greater detail below.
  • FIG. 4 is a process flow diagram illustrating a preferred efficient process by which numerous field drivers with video capture systems can be continually supplied with hard disk drives to capture data. The figure is comprised of three columns 0405, 0410 and 0420 separated by dashed lines. The left column 0405 entitled “Video & Data Server Farm” represents the location of the Video & Data Server Farm 0100 where all data is stored and subsequently delivered to the Internet 0150. The middle column 0410 entitled “In Transit Shipping” represents the movement of data in external hard disk drives 0250 via commercial shipping companies. The right column 0420 entitled “Video Capture System Field Driver Residence” represents one of many locations where drivers in the field handle external hard disk drives for receiving, installing and shipping.
  • In the illustrated example, time movement is indicated along the vertical axis in a downward direction with the start of the process at the top indicated as Week 1. For example, the end of Week 1 is represented by a horizontal line 0419. During Week 2, represented as the time between the lines 0419 and 0429, a shipment 0443 of blank external hard disk drives 0445 is moved from Video & Data Server Farm 0405 to the Video Capture System Field Driver Residence 0420. The angle of the line in the downward direction to the right illustrates the time delay of several days as the external hard disk drives 0445 are physically moved by a commercial shipping company or private shipping means specifically set up for operation of the system.
  • The process starts with an initial shipment 0430 of a Video Capture System through In Transit Shipping process path 0432 to individual drivers in the field for installation in their respective vehicles. A blank external HDD is then shipped in process path 0433 to the individual drivers in their respective Field Driver Residence for installation into the Video Capture System. The external HDD 0250 is illustrated as several unique external hard disk drives as they change state and use as they move through various stages of processing. For example, the first external hard disk drive labeled as HDD1 0440 is shipped blank to the Field Driver Residence where it is installed. During Week 2, a driver drives along selected roads capturing video data and GPS data which is then written onto the HDD1 as illustrated with the wavy line 0445. During Week 2, a second blank external hard disk drive HDD2 0492 is shipped 0443 to the field driver where it sits in inventory waiting for use. At the end of Week 2, the driver ships 0450 the first full HDD1 0491 to the Video & Data Server Farm for processing. At the beginning of Week 3 0429, the first HDD1 0491 is now re-labeled as HDD1-4 in the landing zone upload process 0496 carried out at the Video & Data Server Farm.
  • In the landing zone upload process 0496, the data is copied 0465 to the landing zone server and a GPS.LOG and a Master Index File data is loaded to the map server. An operator is given an opportunity to review the data using video file viewing software to provide a visual quality review test process 0470. Once the quality review process 0470 is completed, the operator then erases 0475 the data from the external HDD and places the external HDD drive on the inventory shelf 0480 labeled as HDD4.
  • At the beginning of Week 3, the field driver installs HDD2 0492 into the video capture system and starts driving only a survey route to capture video data which is then stored 0455. Later in Week 3, the field driver receives a blank HDD3 0493 which is stored at the field driver residence until it is needed at the beginning of Week 4 0439. The driver completes the capture process at the end of Week 3 0439 and ships 0463 HDD2 0492 to Video & Data Server Farm 0405, at which point the landing zone upload process is repeated 0498. The above-described process provides a simple yet efficient method of moving large amounts of data from the field driver residence back to the Video & Data Server Farm.
  • The processes for accessing and viewing data will now be described in greater detail. FIG. 5, for example, illustrates a web browser 0505 view with an address bar 0510 pointing to a web page that shows both an ortho image view 0515 of a mapserver and video pane 0540. In the ortho image view 0515, drive line layers 0520 are shown representing the GPS coordinates as a map layer on top of parcel lines 0525 with an end users mouse cursor 0530 pointing to the selected area of video that the user wants to see in the video pane 0540. In this way, a user can select a parcel 0535, indicated with the double lines, and then the system can show a video view of the property 0545 in the video pane 0540. With this type of interface, a user can simultaneously view both conventional street map representations as well as the street level Video drive-by data generated by the system.
  • FIG. 6 is a process flow diagram of the processes implemented with the image processing servers 0115 required to superimpose geo-coded text data over video captured by the Video Capture System. For a given parcel or street address, a video program 0610 determines available video segments that contain images where the property is in view. This is determined by using an open source programming tool GDAL which completes a range calculation from the latitude and longitude positions encoded in the video segments to the latitude and longitude positions of the subject property. The determination of if a parcel is in view is based on a variable distance that is initially set at approximately 300 feet from camera position to the nearest line segment of a parcel. Once a segment of video is selected that contains video frames within 300 feet, an extractor program 0620 (which is based on an open source programming tool MENCODER) converts the compressed MPEG2 data into lossless TIFF image files with each image frame file encoded for camera orientation and interpolated location.
  • For the target parcel, a calculation using GDAL is made in a camera centroid position program 0630 to determine the exact latitude and longitude of a centroid of the parcel, which generally approximates the location of the structure. Given the centroid of the parcel, another simple geometry calculation is made as to the orientation and location of the camera view that is pointed toward the centroid of the parcel. The location of the centroid of the parcel and the distance to the parcel is then calculated for the frame. A render program 0640, which runs on the image processing servers 0115, is then provided (in this example implemented using open source programming tool suite OpenGL and GraphicsMagick) to calculate a location within the TIFF image file in which text or an icon can be placed based on the centroid. This process is then repeated in each of the TIFF images which are then stored in each of the image files. A recompositing program 0650 (in this example implemented using an open source programming tool FFMPEG) then assembles each of the TIFF images together in time sequence to create a compressed video file. Finally, the video in MPEG2 format with the newly added text and icons added is streamed 0660 out to an end user using an open source Apache web server.
  • As stated above, one potential application for the above-described system is in the field of real estate appraisal. FIG. 7 is a data flow diagram that illustrates the process by which an instant drive-by appraisal or photo AVM (Automated Valuation Model) can be created using a video drive-by database that contains geo-coded video. The diagram represents four discrete locations in three columns: the upper left column represents a Video & Data Server Farm 0701; the center column represents the Internet 0702; the right column represents an end user appraiser 0703; and the lower left column represents a lender 0704. In operation, data is transferred from end user appraiser 0703 to and from the Video & Data Server Farm 0701 and finally to the Lender 0704 via the Internet 0702.
  • The process starts, for example, with the end user appraiser 0703 enters a street address 0710 for a subject property on a web form. The street address data is then sent 0715 to the server hosting one of many commercially available AVM or Comparative Market Analysis Applications (CMAA) 0720, which does a database search and selects a list of properties that are similar in price, size and nearby location to the subject property. The list of comparables or comps is sent 0725 back to the end user appraiser 0703 who reviews and analyses various data fields for the comps (such as price, sale date, bedrooms, square feet, structure, lot size, etc.) from the full list of comps 0730. The selected list of comps 0735 is sent back to the Video & Data Server Farm 0701 where a select video program 0740 transfers video files from the image processing servers 0115 to a video-to-digital photo extractor program 0745 (implemented in the current example using open source programming tools MENCODER and DNoise3D). Digital photos of the subject property and selected comps are then sent 0750 to the end user appraiser 0703, who is presented with a web form 0760 which contains a photo of the subject property 0762 and a set of photos of comps (0764 thru 0768).
  • In the above illustrated example, the end user appraiser 0703 reviews the picture of the first comp 0764 and sees that it is approximately the same size and appearance and then clicks on the first check box 0784 to accept it as relevant and appropriate. In the case of the third comp 0766 that is presented, it is too small and it is not selected 0786. The second 0765 and fourth comps 0767 are selected 0785, 0787 and finally the fifth comp 0768 is not selected 0788 as being too large and not of appropriate visual quality as the subject. The selection list is then sent back 0770 to the lender 0704 as the appraisal selected relevant comp properties report, which is then reviewed 0775 by the lending officer to assist in their financial decision process.
  • FIG. 8 illustrates a top view perspective of a video capture system in accordance with a further embodiment of the invention as it drives past a structure. The line 0800 represents the track of the Video Capture System with one-second GPS time stamp intervals 0801, 0802. Along the path from 0801 and 0805 there are 30 notch marks with one represented as 0805; each notch mark represents 1 of 30 video frames captured each second by a right facing video camera of the camera array. At frame 5 0810, the right camera field of view 0820 has several easily identified high contrast landmarks in sight. The term “landmark” represents any fixed object, coloration or structure that is easily identified in a digital image photograph, this can include (but is not limited to) light pole, mail box, edge of driveway, corner of building, roofline, front door, tree, bush, pond edge, awning, centerline on road, public sign, etc. The landmarks are identified and labeled in the video frames with commercially available and open source programs that provide edge detection analysis of digital image photos. The view of the left edge of a landmark tree trunk 0830 and its leaf-shading boundary 0831 is highlighted with an arrow 0835. The horizontal position of the tree within the field of view is indicated in line 0825. In addition, the camera has several other landmarks a window frame 0836 and a door 0837 that have virtual lines 0826, 0827 drawn from the camera to the landmark. These virtual lines provide a point reference in the horizontal dimension with a video frame as to the angle by which this landmark is oriented from the camera. As the Video Camera system travels along the road to another point in Frame 29 0815, which is one frame short of the second GPS time stamp 0802, a second set of landmark views virtual lines 0845, 0846 and 0848 within the right camera field of view 0821 are available for simple geometry calculations. The horizontal location of the tree trunk landmark 0835 is now in the extreme right most position of the video frame of 0815 camera position.
  • Using simple interpolation techniques, the position of each frame between the GPS one-second interval points can be calculated with time and speed between each. The assumptions are that no radical changes in speed or direction occur within the one second interval. Using simple geometry the change of horizontal location of the camera from frame position 5 0810 to frame position 29 0815 pointing to a fixed point 0835 enables a calculation of the distance to 0835. By combining the positions of the landmarks with the GPS data, a precise set of relative positions can be calculated for landmarks in the video. The newly created two dimensional data for the location of the landmarks can now be added to a GIS viewing software application such as the open source University of Minnesota MapSever.
  • FIG. 9 illustrates a process of frame position interpolation in accordance with the invention. In the illustrated example, three line segments are drawn showing three scenarios of road travel to illustrate where large changes in velocity and direction within one second intervals can cause major errors in calculations of Video Capture System 0201 camera positions, if one is relying upon straight line travel from point to point. In a straight line travel 0901 three one second GPS points 0905, 0906 and 0907 are logged. A set of ten notch marks representing every third video frame are illustrated as 0908 and 0909 which are frames 4 and 7 respectively. In this case the actual camera position is nearly identical to the calculated camera position.
  • The second line segment 0902 illustrates the difference between actual travel 0915 of the Video Capture System on a windy road with a stop sign 0917 versus the calculated position of travel 0925 and 0925 versus the point to point travel 0926. The calculated position of travel is based upon using a fourth order polynomial function with parameters that the coordinates of four data points 0930, 0931, 0932 and 0933. This function calculates line segments illustrated as 0921, 0922, 0923 which have notch marks. In the case of the first line segment 0921, it is drawn over the actual path illustrating a highly accurate estimation of the travel path. In the second line segment 0921, a nearly straight line 0922 is calculated versus the actual turn. In the worst case scenario of line segment three 0925 there is a large deviation from the actual path 0915 as a driver made a right turn at a stop sign (without stopping) within the one second interval. The calculation of 0925 is closer to the actual 0915 than the linear point to point 0926 line segment.
  • The third line segment 0903 illustrates the calculated path 0940 of a Video Capture System as it navigates a cul-de-sac with is highly accurate in motion to the actual path versus the point to point path 0941. A spline polynomial method works well in that it creates smooth curves that fit the natural motion of drivers as they make constant G force turns with smooth turns on a steering wheel.
  • FIG. 10 illustrates an aerial transition to street process. A set of four illustrations 1010, 1020, 1030 and 1040 are provided, which represent a single frame or portion of a window within an Internet web page. The top three illustrations 1010, 1020 and 1030 are created using an open source GIS Mapping program, such as University of Minnesota Map Server, to show parcel line boundaries. The three sequential illustrations show the effect as a user is zooming twice from 1010 to 1020 and then 1020 to 1030. Once the user has reached a designated zoom height as in the case of web page view 1030 where individual parcels can be clearly seen, addition image layers 1031, 1032 and 1033 are added. The centroid 1031 of a selected parcel is identified with the motion of travel 1032 of the video segment as the starting point of Video Capture System camera position, and direction of travel 0133 with the dashed line 0134 representing the field of view of the camera. Finally, the fourth web page frame 1040 in the sequence illustrates the smooth transition from aerial to street view of a selected parcel. In this case a single frame is shown based on the position 1035 of the Video Capture System. The centroid of the parcel of web page frame 1030 is illustrated as 1031 and in the fourth web page frame 1040 the same parcel centroid is now illustrated as 1041 providing the end user good visual cues to help them see where they are as they view a street level image. The actual path of the Video Capture System is drawn on the video frames as 1042 as the drives past the designated parcel.
  • FIG. 11 illustrates a process utilizing real property vectors in accordance with the invention, by which images captured by a left facing and a left front facing camera in a Video Capture System can calculate and generate superimposed property lines on both video and exported digital image photos from the video drive-by data. In this illustration, an ortho view 1200 of the travel of a Video Capture System along a road is shown with four discrete image frame capture points 1211, 1221, 1231 and 1241. These image frame capture points are illustrated on the left hand side as 1210, 1220, 1230 and 1240. In the case of camera position 1241 where the car is approaching the subject property 1249 and 1209 the camera has a view of the left side of the road 1203 and 1243 where the property line 1246, 1206 and 1226 can be calculated as described above and shown on a digital image photo. The second 1230, third 1220 and fourth 1210 digital image photos show the changing perspectives of the Video Capture System camera as it moves past a subject property.
  • FIG. 12 illustrates a process to extract and create a point of interest database that is encoded with GPS data in accordance with the invention. Using the Video and Data Server Farm Image processing cluster of servers 0115 and the video frame to digital image photo extraction process 0620, any one of several dozen OCR programs can be run on the extracted digital image photos to build a results file of recognized text. In this case, a Video Capture System's front left camera at position 1305 is viewing a sign 1310 and 1320 that contains recognizable text. This text data is then written into a database that includes but is not limited to: the text characters, size and color of the text, relative position of the characters in the field of view and GPS coordinates of the video frame. Exact position in latitude and longitude of the text of the sign can be located as well as the relative orientation.
  • FIG. 13 illustrates two ortho views 1400 and 1450 of a Video Capture System as it drives in a direction of travel 1410 along a path 1415 past a selected parcel where one of six possible cameras may be viewing a subject property. The hexagonal array 1405 indicates the range of view for each of the six cameras. The Video Capture System at position 1401 has a subject property centroid 1420 in the field of view 1425 of the right front camera. The field of view can be easily calculated by drawing a virtual line segment 1445 from the camera position 1401 to the centroid 1420. As the vehicle moves along the path 1465 in the second ortho view 1450 the right camera field of view 1475 is now containing the centroid 1470 of the subject property when the camera is in position 1451, which is illustrated by the virtual line segment 1460.
  • FIG. 14 illustrates a process by which an authenticated end user 1503 can submit data regarding a specific location using video drive-by data as geo-coding reference tool. The figure has three columns: the left column represents a Video and Data Server Farm 1501; the middle column represents the Internet 1502; and on the right column represents an end user 1503. The process starts when the end user 1503 establishes credentials 1505 to authenticate with personal information such as their name, address and email with a credit card number, social security number, date of birth, drivers license number, etc. At that point they can select their login and password and submit 1506 to a common and commercially available registration application 1510 running on an Apache web server. Once a new user is established 1512, the user can then login 1515 and send a request to authenticate 1516. When that user is authenticated 1522, they are now granted access 1525 to the video server 1530 where the can do a series of queries 1526 (by street address or clicking on a map as shown in FIG. 5 or 7) where they obtain results 1532 in the form of video feeds.
  • When the user is ready to review and edit property data, they login and authenticate with the password. Using one of several modes of selecting a property, they view a particular parcel or area with a picture of the property. The viewer application can show video with various information such as parcel data or property lines illustrated on the property. The user can also select the video to present frames at a slower rate than 30 frames per second to allow careful viewing of the imagery. At that point, where the end user sees a property, they can annotate any landmark by clicking on the image of a property (such as the roofline of a house 0545 shown in FIG. 5). They can use the cursor to establish a point on an image as a designated landmark. The user can then tag the same landmark in one or more video frames to further clearly tag the image designated landmark. By processing the relative location of the image in the video frames as the vehicle moves, the exact latitude and longitude coordinates can be determined (with an estimated error calculation included in the data). When the user has selected the point within the image, that location data of the one or mouse clicks with a (point, line segment or shading of area or attachment of an icon) can be submitted 1536 to the video server 1540 and submitted 1537 to the database server 1550 for permanent storage and retrieval by another user. The database includes individual record entries of each and every annotated data set to generate a detailed change log. This log file can include but it is not limited to latitude and longitude of camera position, estimated accuracy of camera position at time of capture, time of original video capture as well as end user annotation time.
  • FIG. 15 illustrates a process and visual references as to the calculations required to extract a photograph from a street number. The process starts with the loading of county parcel data that typically includes latitude and longitude coordinates of vector line segments that identify the individual parcel lines of given property. Typical county parcel records include street name and number as part of the parcel number that is supplied as a separate database with the parcel number itself being the unique ID that is used to link street address to latitude and longitude encoded parcel lines. This figure includes two diagrams 1605 and 1610. The top diagram 1605 shows a horizontally stitched together set of frames of video 1636 and 1641 as the left Camera is moving by a set of houses. The stitching process can be accomplished by using any one of several dozen commercially available programs running on the video image server cluster 0115 after the video frames have been extracted as indicated in FIG. 6. The stitching program will be feed a set of digital image photos in this example Frame 1 1636 Frame 34 1641, Frame 67 1646 and Frame 100 1651 are used to composite a complete wide horizontal view of the street. Each second of video produces thirty individual image frames. Accordingly, in this example, 3.3 seconds of video were used from frame 1 to frame 100. The frame number to use when driving by properties is easily calculated by the speed data from the change in position of the GPS data points. Using techniques described above, text can be superimposed 1625 onto the digital image photo.
  • As illustrated in diagram 1610, an end user is using a web page viewer 1615 to move a set of left 1621 and right 1620 control arrow button to select the image that evenly frames a subject property in the field of view. By providing a long horizontal picture view 1605 in a square computer monitor viewer web page 1615 an end user can easily identify properties with large street numbers 1625 displayed as they are looking to identify a property.
  • FIG. 16 illustrates a novel method of finding a house from a photograph based on the invention. FIG. 16 includes three columns: left column 1701 represents a Video and Data Server Farm, the middle column represents the Internet 1703; and the right column represents an end user. The process starts with a user entering a web page 1705 where they can upload a digital image photograph 1706 of a property taken from the street. The user can also input a zip code where they think the property is located to enable faster searching. Once they have entered the zip and attached the photograph, they can submit 1710 the data to the server. At the Video and Data Server Farm the uploaded image is converted 1715 to the appropriate standard format of uncompressed bits representing each pixel. The uncompressed image is then processed with one of several commercially available pattern recognition applications 1725, where the uploaded image 1706 is compared to the street level digital image photographs extracted (for example those illustrated in FIGS. 13 and 15) from the video drive-by process. The photos that have patterns matching the closest to the submitted 1706 photo are then extracted and added to a webserver to create a web page 1725. This web page is then delivered to the end user as a web page 1735 with the originally selected photo 1706 redisplayed as 1740 with additional extracted images 1745, 1750 below. The end user does the final human based pattern matching by selecting with a radio dialog 1755 the image that looks the closest and then presses the submit button 1760. This data selection is sent to the server 1765 where the server processes the selection 1770 to generate a web page that returns 1775 the street address and picture of the property 1780.
  • FIG. 17 illustrates one novel method of identifying a home from a large library of home photos with a high speed process. The process involves using one of many commercially available applications that generate edge detection patterns in an image. In this figure, an image of a house 1801 is shown with various lines that would be trivial for an edge detection algorithm to find properly. In order to process the data most efficiently, only horizontal and vertical edges would be selected. In this figure, horizontal lines such as the top of the front door 1815, top of the roof 1816, and roof line 1817 and vertical lines such as front door 1811, left edge of the right window 1812 and right edge of the right window 1813 are easily identified. In the second diagram 1802 the vertical edges are displayed as a set of bar codes that have a relative horizontal distance from each other. The bar line 1821 is derived from the right side of the front door 1811 and the left side of the right window creates another bar line 1822. In the third diagram 1803, the horizontal edges that have been detected are displayed as bar lines. In this diagram bar line 1836 is detected on the top of roof line 1816, and bar line 1837 is the roof line 1817, and the bar line 1835 is the top of the door 1815 and finally, the bar line 1836 is the bottom of the right window sill 1816. The storage of the discrete bar lines as mathematical values with relative position data enables a numeric pattern match of values between photos taken at different distances from a subject property. In this way a set of numbers representing the largest bar lines as vectors can be stored and compared against a large numbers of properties with a simple sorting process.
  • The present invention may be applied to many applications in which the remote ability to view a specific location and its surrounding neighborhood is advantageous. For example, land owners, prospective land owners, insurance, mortgage, county appraisers need accurately to determine where property lines are located in relation to a property. Today, a physical visit to a property with a survey in hand is required to start the process of determining where the property lines are located. In fact, many times a professional surveyor may be required to find the original iron stakes that are usually covered over with vegetation. Once the stakes are found, using long tape measures and surveying equipment, an accurate determination can be made of the locations of various structures and landmarks in relation to the property lines. In contrast, the present invention permits the merger of a unique set of data together to generate street level digital image photographs of properties that include superimposed illustrations of property lines based on original latitude and longitude data provided by the county property appraiser's office. This highly convenient method of illustrating survey data fits a wide variety of unmet market needs.
  • As another example, occasions exist where people as they drive past properties want to find out who owns the property and what they paid for it. The present invention enables an end user the ability to select any street via an on-line web page and scroll along a view of images where the county parcel data is superimposed in the video stream. Images of property parcels as well as the camera orientation pointed toward the centroid of the parcel are provided along with property record data that can be processed to locate the centroid of the property and then display text data of sale price, owners name & street address above the image of the house.
  • Further, property appraisers are mandated by many state and local governments to obtain photographs of properties that they tax. Currently some counties pay $10 to $50 to have a driver go to specific properties, take a digital image photo and email it to the appraiser. Also, some counties contract to have firms drive along streets shooting photos and storing them in a database that they deliver to the appraiser for a fee of $1 to $2 per parcel. In contrast, with the present invention, a county government could contract to view video of parcels in a far more economical manner by using our Video Capture System driven by their existing appraisal staff. The data they capture can then be added to the Video & Data Server Farm and is available for viewing in a 360° view not just in a single view of a property. From the video, dozens of individual photos are available for the property appraiser to view at any time and by default the centroid oriented view at a 90 degree angle can be shown first.
  • Still further, emergency service agencies have equipped their vehicles with high speed wireless data networks to augment their voice radio network. The wireless data networks enable access to the Internet and the Virtual Private Network to provide information related to the location and details on the emergency call while in route. This location data is typically a street address and a map with directions. Maps today generally do not have good positional accuracy for individual street addresses. So when emergency responders pull onto the street, they immediately commence looking for house numbers. When these numbers cannot be found especially at night, they resort to guessing and this consumes valuable time when seconds count. A solution which has been implemented in several counties is to provide a database of street photos that are tied to parcel data. This process is done by manually photographing properties and linking them to a database of street addresses that may or may not be encoded with latitude and longitude data from a GPS. The present invention enables the automatic and low-cost capture of street level video which can be encoded with parcel data and GPS coordinates so that the Video & Data Server Farm can generate a long horizontal street level image that can be transmitted rapidly as a single image to an emergency vehicle in route which aids in the ability for the emergency responder to quickly decide at which house to stop.
  • Today, on-line map users are presented with a top view of a set of streets that are annotated with color highlighting and indicators such as red stars that make it easy to identify where they need to go. The present invention includes the additional opportunity to present street level images of turns and intersections that enable users to “see” what an intersection looks like a few hundred feet before they arrive and need to turn. With the present invention, the user could see that there is street level video available based on showing a double red line for example that the user could click on to see a street level image. When the user clicks on the red line, they would immediately be presented with video from that exact latitude and longitude coordinate.
  • Today, home buyers turn their heads as they drive by properties that have for sale signs in the yard. Although there is no system on the market today for viewing street level video on-line, there is a need for the camera view to be oriented at a property as the 360 video capture system moves past the property. With the present invention, the Video & Data Server Farm has the latitude and longitude coordinate data of the camera as it drives by and in addition, and it has the latitude and longitude coordinate data of the centroid of the parcel that is for sale. Given this data it is easy to determine which camera position is actually pointed at and has in its field of view the centroid of the parcel. Typically homes are located with the foundation over the centroid and thus the automatic camera orientation method provided by the present invention will enable the automatic selection of the correct camera view to present to the user when they are “driving on-line”.
  • Today as appraisers and real estate sales people insurers, construction, repair, maintenance people are interested in specifying a location or a landmark for a property, there is no facility available on line whereby a user could login annotate an image and then store that for viewing by others. With the present invention, a user could access a video and/or a set of image files that include GPS coordinate data, camera orientation and optionally property line and parcel data superimposed over the image. With this image available, the invention provides a simple on-line viewing application that enables a user with his mouse to click on the image presented and then open a text/dialog box where they can add comments to the designated landmark within the photo. Given an authentication process where the comments made by the individuals can be stored within the database, a permanent log can be made and viewed by others relating to the property with accurate representation as to who made the comment.
  • Today, when photographers visit a property to take a street level photograph, they naturally take care as to locate the camera along the street and orient the camera within a few degrees left and right as to the field of view so the picture “looks good” and the house is “well centered”. The present system allows the automation of finding a reasonably good view of a property, by utilizing the centroid image view automation process with an end user based process of “fine tuning” a view of a property. With the user viewing a stitched image of the subject and its next door neighbors, they can slide the bar right and left to simulate their movement of a photographer on the street moving right and left to get the image “just right”. Once that user selects their favorite frame, the system will store this optimal view coordinate data and also generate a full resolution digital image photo file for delivery to the user.
  • Today when people view a photograph of a home that they are interested in buying on-line on a public website, they are not given the address of the property. Typically they are only provided with a photograph and the zip code of the property. In this regard a user is constrained from determining if the property is in a neighborhood they like or is near a particular set of amenities such as school, or church. For a buyer to locate the house, they would need to invest time to contact either the listing agent or they would need to agree to be represented by an agent to access the street address information. With the present invention, one could have an image of this home in the Video and Data Server Farm, thereby enabling a user to submit an image of the home to the Video and Data Server Farm to process and locate. The servers would run one of several commonly available pattern matching processes against the images for a particular zip code and then present a set of properties that “look close” to the one the user submitted. Given the fact that humans are vastly superior to systems for pattern recognition, a list of 20 close matches out of 2000 potential homes in a given zip code can be easily viewed and selected by an end user in a couple of seconds. Once the user has selected the home that matches, they can then be presented with the street address and any other market comparable data they may wish to obtain.
  • In another embodiment, methods and apparatus are provided for presenting a three-dimensional model of a subject, such as a geographic area. The three-dimensional models can include a polygon based model format generated from a point cloud array and sprayed with actual image data of the subject, captured from disparate points along a continuum.
  • A user interface is presented with user interactive controls that allow a user to traverse the three-dimensional model. A user location within the model will determine which portions of the three-dimensional model are viewed. Some embodiments can include user interactive devices that allow a user to traverse the three-dimensional model in three directions: a) an essentially up and down vertical movement; b) an essentially side-to-side horizontal movement; and c) a depth movement into and out of the three-dimensional model. Image data and three-dimensional polygon based models presented to an inventor at a given instance can be dependent upon the relative position selected by the user within the user interface.
  • In another aspect, icons, overlays and metadata can be presented to a user upon demand to provide further information relating to a given geographic area within view.
  • Still another aspect can include presentation of one or more geographic areas based upon criteria ascertainable from related data. For example, image data can be correlated with geospatial data so that a particular street address or Cartesian coordinate, such as a latitude/longitude location, is designated and specific image data is presented which correlates with the geospatial data. The specific image data can be presented sprayed over three-dimensional structures generated from two-dimensional image data captured from that specific location.
  • In additional aspects, some embodiments can also include indications on the user interface of user selectable data related to a particular location. User selectable data can include, for example: cellular phone signal strength; wireless network data, including network name, location and signal strength; global position system satellite data and signal strength; radio signal strength; school district data, average yard size, property price range; building features; proximity to amenities; political boundaries; zoning restrictions; and almost any other measurable quantity. Data can all be correlated according to a geographic location and presented in various combinations or upon demand. In addition, the data can be presented in conjunction with or independent of, an image data continuum or image data set.
  • DEFINITIONS
  • As used herein, Video DriveBy™ refers to street level video data captured in multiple angles encompassing a 360° view.
  • As used herein, Video FlyBy™ refers to Aerial/Satellite oblique (angular) view images with polygon line views.
  • As used herein, RibbonView™ refers to a film strip like view of properties that stands up at approximately 90° from a flat or angled aerial/satellite ortho image map and provides direct-on front images of properties to be displayed.
  • As used herein, Flash Viewer (Streaming Video) refers to direct streaming of video to an online user via a web browser.
  • Methods
  • According to the present invention, image data is captured from disparate points along a continuum. Referring now to FIG. 18, in some embodiments the continuum 100 includes the path of a vehicle carrying a digital camera, or other image data capturing device. Image data sets are captured at disparate points 101-106 along the continuum. Some preferred embodiments include capturing each image data set at an angle that is generally orthogonal to the subject matter 107.
  • Positional data and orientation of the camera capturing image data sets is recorded for each disparate point 101-106 along the continuum from which an image data set is captured. Positional data can include any data indicative of where the subject matter of an image data set is located. Some preferred embodiments of positional data include Cartesian coordinates that are context sensitive according to the mechanism used to generate coordinate data.
  • Positional data can be generated, for example, by an image data recording device, such as a camera, associated with a device for recording a global position (GPS device). Time stamps associated with image data sets and time stamps associated with the GPS data can be utilized to correlate the GPS location data with image data set recorded by the image data recording device.
  • In still another aspect, in some embodiments, an altimeter can be used to record an altitude from which a camera records image data sets. The altitude data can be associated with an image data set, for example, metadata correlated with the image data set. Such embodiments can therefore include recordation of a latitude, longitude and altitude coordinate for a given image data set. In addition, it is also within the scope of this invention to record a time of generation of an image data set and a depth of focus for an image data set.
  • According to the present invention, geospatial data, such as latitude and longitude coordinates, can be generated by the GPS and stored with image data generated by the camera. In some embodiments, GPS data can be time stamped and collected once every second. However, in some instances, GPS reception can be interrupted, depending upon location relative to large object, such as multistory buildings, or cold cover. Therefore, some additional embodiments can include an accelerometer for recording motion associated with a camera and a GPS unit operatively attached to the camera.
  • Data from the accelerometer can indicate movement of the camera. Some accelerometers, such as micro electro-mechanical system (MEMs) accelerometers can easily be incorporated into a camera system assembly attached to a vehicle. Use of multiple MEM accelerometers positioned to measure movement in four or more directions along an x-axis, y-axis, and z-axis in relation to a camera can also be used to calculate direction of movement. The accelerometer can therefore be used to extrapolate a current position of the camera, based upon a last set of GPS geospatial data recorded.
  • Geospatial data can be used to indicate an initial geographic position. A change in geospatial data can be additionally utilized to indicate velocity and direction of image data set capture. Accelerometer data may also be used to indicate a velocity and direction of image data set capture. Accelerometer data may also be used to indicate a velocity and direction of camera movement. Calculations of time elapsed at the indicated velocity (such as for example, the Kalman Filter) can yield a calculated position at a time of image capture, even if the time of image capture is between GPS readings.
  • For example, one standard can include tracking a camera position with a GPS unit that records location at a rate of once per second. The camera can record image data at a faster rate than once per second, such as, for example, one of: 12 images per second, 24 images per second or 29.97 images per second. An initial camera position can be recorded that correlates with a GPS reading; subsequent image data capture will occur in between the one second GPS reading interval. The camera position can be determined with a calculation based upon the velocity of camera movement supplied by the accelerometer and time elapsed since a previous GPS reading.
  • Still other embodiments can utilize optical flow methodology and visual odometry to facilitate calculations of a camera position and the velocity of a vehicle or person from which a series of image data sets are captured. Visual odometry can be accomplished with a single omni-directional camera or with stereo cameras and is based upon processing that tracks the position of salient features in a series of feature sets and calculates movement based upon the relative positions of the features. In some embodiments, camera based simultaneous localization and mapping (SLAM) of visual image data can also be used to facilitate calculations of velocity of a change in position of a camera used to capture image data sets. Typically, the velocity will be directly tied to the motion of a vehicle to which the camera is mounted, or a person carrying a camera rig.
  • Orientation of a camera can include a direction of image capture recorded by the camera. Orientation can be designated, for example, in relation to the cardinal directions, i.e., north, south, east and west. Any means available to record such a designation, such as an electronic compass, is within the scope of the present invention. However, it may be desirable to include a means to record the orientation with a greater degree of accuracy than is typically available through the use of an electronic compass.
  • Therefore, in some embodiments, orientation can be determined according to a fixed position of a camera in relation to the direction of travel of a vehicle (or person) used to transport the camera. For example, a plurality of cameras can be fixedly attached to a vehicle capturing Video DriveBy™ data. Each camera therefore maintains a constant direction of image capture in relation to the heading of the vehicle. Mechanics of the camera, such as, for example, lens parameters and shutter speed, can indicate a depth of field during camera image data capture. Some embodiments can also include simultaneously capturing multiple image data sets and correlating two or more of the image data sets. Correlation can be accomplished via a time stamp or other chronological or synchronous.
  • The position of a camera can be combined with a direction of image capture and the depth of field of the camera to determine a location of image data captured by the camera at a particular instance in time. The present invention can also include apparatus for utilizing echo location to determine a distance of an object from a camera capturing an image data set and storing a correlated distance with the image data set. For example, radar data can be correlated with a camera image data set to calculate the location of various objects captured by the camera. A time stamp can also be combined with data to quantify a location for a particular image formed by the captured image data.
  • In some embodiments of the present invention data used to calculate a location of an image is stored in a metadata file space associated with the image data. For example, some embodiments can store metadata in the exchangeable image file format (EXIF), TIFFTAGS or International Press Telecommunication Council (IPTC) formats. Image data may be stored, for example in JPEG or TIFF formats. However, other metadata formats can also be used. Typically, due to the size of data files that are generated during capture of Video DriveBy™ data, the image data and metafile data are stored on an external data storage device, such as a hard disk drive operatively attached to the camera. However, in some embodiments, the data can be stored in the camera.
  • As discussed above, the metadata can include data descriptive of some or all of: date and time; camera settings such aperture, shutter speed and focal length; geospatial data from a GPS receiver unit; accelerometer data; inertial guidance system data; camera orientation; and camera fixed position related to vehicle travel.
  • Referring now to FIG. 19, exemplary image data sets 201-203 are illustrated which capture images of a subject 200 from different points along a continuum. In some embodiments, as illustrated in FIG. 20, image data sets 304-307 overlap the subject matter captured. The overlap allows for features present in one image data set to be present in a second image data set, and preferably in a tertiary image data set.
  • Referring now to FIG. 21 a block diagram illustrates three-dimensional structures 405-407 created from image data captured. In some embodiments, the structures can be created by generating a point cloud array from a plurality of features contained in the image data sets and converting the point cloud array to a polygon based model. The polygon based model can be associated with a location relative to the image data set and image data can be sprayed over the polygon based model. In some preferred embodiments, image data sprayed over the polygon based models includes the composite of image data.
  • According to the present invention, a user can traverse the presentation of the composite of image data and the three-dimensional structures. Movement through the user interface can be accomplished with any known interaction device, such as, for example one or more of: a keyboard, mouse, video game controller Apple I Phone, digital cellular phone, remote controller, WiiMote and a joystick.
  • The user interface can allow a user to move in three directions through a scene presented. The directions can include an x coordinate providing a vertical dimension 401, a y dimension providing a horizontal dimension 402 and a z coordinate providing a depth dimension 403. Three-dimensional modeling sprayed with actual image data allows the user to proceed in “off-track” paths that do not match a path created by the disparate points from which data is collected.
  • In some embodiments, a time dimension can also be traversed, such as, for example, image data sets of a geospatial designation across time chronology such as a period of seasons or years. Some embodiments can include a “take me there” time designation, wherein a user can indicate a time period and geospatial description. According to the present invention, a three-dimensional model can be generated with image data sets from the indicated time period for the designated geospatial area.
  • As the user moves into a scene presented in the user interface, planes of image data 405 can change according to the perspective chosen by the user. For example, a three-dimensional structure of a house can be sprayed with image data included in the continuum composite of image data. The actual image data will correlate with a surface of the three-dimensional structure facing the disparate points from which the image data was captured. According to the present invention, additional surface of the house 406 can also be sprayed with a texture and color captured by the actual image data. In this way, as a user moves in a z direction beyond front surface of the house, the user will be presented with additional surfaces of the house, with each surface having a color and texture extrapolated from the image data from the front of the house.
  • In addition, an object such as a tree 405 can also be created in three dimensions with color and texture based upon the image data composite. In some embodiments, user interactive device controls can be provided which allow a user to remove objects such as the tree 405 and the house 406. Removal of an object will allow the user to see beyond the removed object into the scene provided. For example, if a tree 405 obscures the view of the front of a house 406, activation of a user interactive control can remove the tree 405 allowing an unobstructed view of the three-dimensional model of the house 406. If actual image data is not available for those portions of the house now viewable in the unobstructed view, then the present invention allows for color and texture data matching the known surfaces of the house to be sprayed over those portions now viewable.
  • Some embodiments can additionally allow a user to modify color and texture of three-dimensional models presented to them. Such modifications facilitate “what-if” scenarios. For example, a buyer contemplating purchasing a particular property can be presented with a three-dimensional model of the property with three-dimensional representations of objects on the property sprayed with actual image data captured of the property. According to some embodiments, the buyer can then modify the view presented to represent possible changes that may be made to the property. Changes may include, to continue the example, providing a brick facade over a painted surface, or removal of shrubbery, or addition of shrubbery. Addition of items can be provided from a menu of three-dimensional objects made available to the user. Color selections and textures can also be made available via a user selectable menu. In this fashion a unique blend of a composite continuum of image data and user defined image data can be presented, including three-dimensional modeled objects derived from the image data and user modifications to such models.
  • By way of non-limiting example, in some embodiments, a user may be able to select a building and designate from a user interface a type and color of roofing the user would like to see illustrated for the building. Accordingly, a user may designate an architectural shingle in various hues, or a Spanish tile or steel roof and the present invention will spray the pattern and color selected onto a three-dimensional model of the building. In still other embodiments, it is within the scope of some embodiments of the present invention to allow a user to upload a texture and color to be sprayed upon a three-dimensional model, or a portion of the three-dimensional model, such as a roof portion. According to such embodiments, a user may upload a color or hue or texture to be sprayed upon a three-dimensional model of a structure. The color and hue may also be stored and made available for subsequent application to portions of an image data set.
  • Referring now to FIG. 22, the present invention can also include generation of a user interface 500 with multiple representations of image data relating to a geographic area. For example, an aerial view of a geographic area 504 can be combined with one or more two-dimensional continuums of composite image data 505-506. The two-dimensional continuums of composite image data 505-506 can be generated according to methods and apparatus described in a related co-pending patent application by the same inventors. In addition, one or more continuums of disparate points from which image data is captured can also be indicated on the aerial view 504. Three-dimensional models 510-503 can also be included in the user interface 500 and located on the aerial view according to geospatial data generated during image data capture and processing of the image data.
  • As described in a related application, the three-dimensional models can be generated from multiple image data sets captured from disparate points along a continuum sprayed with the image data captured.
  • According to some embodiments of the present invention, a user may locate a particular geographic area via different vehicles made available to a user. For example, a user can enter a geospatial designation, such as, for example, one or more of: a street address; a Cartesian coordinate; a “Favorite” name designating a geospatial area; a popular name for a geospatial area; a coded indication for a geospatial area, such as a multiple listing number or data file number; and metadata. According to the present invention, a “take me there” command can be executed to bring the user to the requested geospatial area. Alternatively, the present invention provides a vehicle “cockpit” interface which allows a user to virtually “fly” over an aerial view and then transition into a virtual “dashboard” and “drive” mode along a street level two-dimensional continuum of actual image data, and in some preferred embodiments, a continuum of composite image data. The user may also transition into a virtual three-dimensional modality which includes three-dimensional models sprayed with image data captured from a continuum of disparate points. The three-dimensional models can be viewed from a virtual “drive by” mode or a virtual “walkabout.” The virtual walkabout can allow a user to go off path and view the geographic area from amongst three-dimensional models sprayed with image data captured from a continuum of disparate points. In addition, as discussed above, the user may manipulate the appearance of the three-dimensional models.
  • A take me there command can be additionally operative to designate a modality, such as the aerial view, two-dimensional ribbon view, or three-dimensional drive by or three-dimensional walk about view into which a user will enter.
  • In another aspect of the present invention, a user can be provided with tools that provide analysis of the image data shown. Analysis can include, for example, designation of a particular portion of a geographic area 508. The portion 508 can be representative of a real estate parcel, as indicated by municipal records, or some other designation. For example, a user may draw a polygon 508 onto an aerial image of a geographic area 500 and ask for analysis of the area contained within the polygon 508.
  • Analysis can include, by way of example, a calculation of an area included within the polygon. Other analysis can include more sophisticated calculations. For example, it is within the scope of the invention to designate property boundaries, such as, through county record data and locate an image data of the property. The image data can be used to generate a three-dimensional model of the property, including a model of a house within the property boundaries and other features, such as trees and utility demarcations. Based upon the location of the house within the property boundaries, the present invention may also be used to generate a size of a front yard, and, by extrapolation, the size of the side and back yards. An approximate size of a house can also be calculated based the dimensions of the three-dimensional model generated. Other features, such as how many stories, the house includes and the type of facade the house has may also be ascertained from analyzing the image data and the model.
  • Statistical representations can also be made, according to the analysis. Statistical representations can provide a summary of characteristics of a single real estate parcel, or a designated area. By way of example, statistical analysis can include one or more of an average, median and high/low value designation for: the size of a property within a given neighborhood, the size of a front yard, how much road frontage each property includes, the average size of the houses within the designated area, the number of stories comprising the buildings within the area, siding types of buildings within the area, the size of side yards between houses, height of trees within the area, and almost any other data that may be calculated from the image data sets and the three-dimensional models.
  • Referring now to FIG. 23, a view of an alternative exemplary user interface 600 according to the present invention is illustrated. The interface 600 can include a portion with an aerial view 601 and a polygon designation of geographic area of interest, such as for example, a particular real estate parcel or neighborhood location. The interface may also include a portion with a two-dimensional continuum of image data 603A. The continuum of image data is preferably a composite of image data sets captured from disparate points. Another portion 603B can include image data sprayed over three-dimensional models 604 generated from point clouds of features captured by the image data sets.
  • Referring now to FIG. 24, in some preferred embodiments, image data sprayed over the three-dimensional polygon formats includes a composite image formed by aligning two or more of the image data sets. Unlike stitching processes previously known, the present invention creates a composite through alignment of portions of data from more than one data set. Alignment can be accomplished in image data processing. Using image data processing, the images 701-703 are aligned to form a composite image 700. The composite image 700 is essentially two dimensional image data arranged as a second continuum, or ribbon. The second continuum includes ongoing image data 701-703 captured from the points defining the first continuum.
  • In some particular embodiments, the series of points of image capture in the first continuum includes positions of a vehicle carrying an image capture device, such as a camera, as the vehicle traverses a path proximate to a geographic area. The camera is positioned to capture image data of the geographic area. Image data 701-703 is periodically captured as the vehicle traverses the path. The motion of the vehicle, combined with the periodic capture of image data 701-703, thereby results in image data 701-703 being captured from disparate points along the first continuum.
  • A preferred embodiment includes capture of image data with a motion vector of the camera in space maintained generally orthogonal to a subject for which image data will be captured. Orthogonal capture of the image data can facilitate consistency for subsequent composite of portions of the image data captured. Therefore, data captured at an angle of between about 75° and 105° can provide data most easily assembled into a continuum of data. However, other angles may be used to create different effects. In addition, in some embodiments physical limitation may dictate the use of other angles. In such instances, post processing can be utilized to compensate for the angle of image data capture.
  • During image data processing, some or all of the images are aligned to form a composite image in the form of a continuous pictorial representation of the geographic area. One commercial embodiment of a continuous pictorial representation includes RibbonView™ by Real Data Systems. RibbonView™ correlates a ribbon of geographic image data with geospatial designations to facilitate identification of a particular geographic area, as discussed more fully below. In various embodiments, image capture processing can be accomplished in real time or via post image capture processing.
  • In some embodiments, select portions 704-707 of two or more sets of captured image data are aligned to generate the composite image 708. As illustrated, some preferred embodiments include vertical slices of data 704-707 aligned in a horizontal plane to form the composite image 708. Unlike a traditional photograph taken with a macro lens, according to the present invention, the length of a horizontal plane defining a composite image 708 is only limited by the length of a continuum along which points are defined and from which image data 704-707 is captured.
  • The use of only slices of data 704-707 from any particular captured image provides for a higher quality image 708. The quality is increased, for example, when a temporary obstruction, such as a passing car, person or animal, captured in one image data set, is only represented in a thin slice of a continuous ribbon 704-707 data. In addition, alignment of multiple thin slices of image data 704-707 is facilitated from the perspective of which aberrations typical human sensory is capable of distinguishing. Numerous thin slices 704-707 are perceived as a smoother transition across the horizontal plane defining the composite image 708 removing a choppiness that may be experienced with the use of larger data images.
  • The present invention can include a uniform width of each slice of data 704-707 or a variable width. The width of a particular slice may vary, for example, upon one or more of the velocity of a vehicle from which image data sets are captured, the sample rate of a camera used to capture an image data set 701-703, the resolution of a picture comprising an image data set 701-703 and the path of a camera. A high resolution image generated by a 2.1 mega pixel camera may have a 1600 by 1200 resolution and allow for a thinner slice 704-707 that includes a width of between about 5 to 700 pixels of an image data set. Embodiments with very high quality can include a slice 704-707 of between 1 to 2 pixels.
  • From a different perspective, some embodiments can include a slice 704-707 of an image data set 701-703 that includes a percentage of the image data set 701-703. Accordingly, some preferred embodiments can include a slice 704-707 of between about 5% to about 12% of an image data set. Other preferred embodiments can include up to about 50% of an image data set. However, it should be understood that some embodiments include a slice 704-707 that constitutes any fraction of the complete image data set.
  • It should be noted that although preferred embodiments may utilize vertical rectangular slices 704-709, the scope of the present invention is not limited by which portion of an image data set 701-703 is utilized to generate a composite image 270. Therefore, it is within the scope of the invention to use any portion of any image data set 701-703 to generate a composite image. Accordingly, slices of an image 701-703 other than vertical slices 704-709 may be apportioned and combined into a composite image 270. Slices may therefore include a slice angled at 60° or 75°, or other angle conducive to a particular application. In addition, it is also within the scope of the present invention to utilize irregular shaped portions of two or more image data sets 701-703 to generate a composite image 270.
  • In some embodiments, a database or other data processing mechanism, can track each portion or slice 704-708 utilized to construct a continuum and associate the slice 704-708 with an original image 701-703 from which the slice 704-708 is derived. User interactive devices can execute the retrieval of an entire original image 701-703 or a series of original images 701-703 upon request. In some instances, one or more original images 701-703 may provide detail not included in the composite image 708.
  • In some embodiments, a selected portion of an image data set may be determined by the physical attributes of the equipment used to capture an image data set. For example, a typical camera lens can impart some distortion to an image data set, as represented in the illustration by an elongated portion and a compressed portion. Utilizing only a portion of an image data set, such as, for example, a center portion vertical slice, can minimize the effect of distortion introduced by a lens, or other source, to a composite image. Distortion is minimized when the composite image is made to include those portions of the image data set corresponding with the center of the lens. Specific embodiments may likewise account for other aberrations that may be present in a set of image data.
  • It will be apparent to those schooled in the art that the length of a composite image generated according to the present invention is limited only by the ability to capture image data from additional points on a continuum and store the captured image data for post processing. The post processing allows for the alignment of portions of the image data compiled into a composite two-dimensional view that can continue so long as additional image data is made available to be added to it.
  • Apparatus
  • The teachings of the present invention may be implemented with any apparatus capable of embodying the innovative concepts described herein. Image capture can be accomplished, for example, via a digital camera capable of capturing 12 or more images per second. In addition, FIG. 25 illustrates a controller 800 that may be utilized to implement some embodiments of the present invention. The controller 800 comprises a processor unit 810, such as one or more processors, coupled to a communication device 820 configured to communicate via a communication network (not shown in FIG. 25). The communication device 820 may be used to communicate, for example, with one or more online devices, such as a personal computer, laptop or a handheld device.
  • The processor 810 is also in communication with a storage device 830. The storage device 830 may comprise any appropriate information storage device, including combinations of magnetic storage devices (e.g., magnetic tape and hard disk drives), optical storage devices, and/or semiconductor memory devices such as Random Access Memory (RAM) devices and Read Only Memory (ROM) devices.
  • The storage device 830 can store a program 840 for controlling the processor 810. The processor 810 performs instructions of the program 840, and thereby operates in accordance with the present invention. The processor 810 may also cause the communication device 820 to transmit information, including, in some instances, control commands to operate apparatus to implement the processes described above. The storage device 830 can additionally store related data in a database 830A and database 830B, as needed.
  • In some preferred embodiments, apparatus includes a video and data server farm. The video and data server farm includes at least one video storage server that stores video image files containing video drive-by data that corresponds to a geographic location, a database server that processes a data query received from a user over the Internet that corresponds to a geographic location of interest, and an image server. In operation, the database server identifies video image files stored in the video storage server that correspond to the geographic location of interest contained in the data query, and transfers the video image files over a pre-processing network to the image processing server. The image processing server converts the video drive-by data to post-processed video data corresponding to a desired image format, and transfers the post-processed video data via post-processing network to the Internet response to the query.
  • A landing zone server can also be included which receives the video drive-by data from a portable memory device and permits the viewing and analysis of the video drive-by data prior to storage in the video storage server. Still further, a map server is preferably provided to present a static image an overhead view of the geographic location of interest.
  • Embodiments can also include one or more servers described above included in one or more physical units. Each server does not need to be a disparate apparatus. Still other embodiments can include one or more or the servers described above included in multiple physical units. Some embodiments can even include a single server, as described which includes multiple physical apparatus units at disparate locations.
  • Referring now to FIG. 26, three images 101-103 are illustrated. Each of the images 101-103 is reproduced from image data captured from a disparate point on a continuum. As illustrated, a composite image 100 is formed by aligning two or more of the image data sets. Unlike stitching processes previously known, the present invention creates a composite through alignment of portions of data from more than one data set. Alignment can be accomplished in image data processing. Using image data processing, the images 101-103 are aligned to form a composite image 100. The composite image 100 is essentially two dimensional image data arranged as a second continuum, or ribbon. The second continuum includes ongoing image data 101-103 captured from the points defining the first continuum.
  • In some particular embodiments, the series of points of image capture in the first continuum includes positions of a vehicle carrying an image capture device, such as a camera, as the vehicle traverses a path proximate to a geographic area. The camera is positioned to capture image data of the geographic area. Image data 101-103 is periodically captured as the vehicle traverses the path. The motion of the vehicle, combined with the periodic capture of image data 101-103, thereby results in image data 101-103 being captured from disparate points along the first continuum.
  • A preferred embodiment includes capture of image data with a camera maintained orthogonal to the subject. Orthogonal capture of the image data provides consistency for subsequent composite of portions of the image data captured. Therefore, data captured at an angle of between about 75° and 105° can provide data most easily assembled into a continuum of data. However, other angles may be used to create different effects.
  • During image data processing, some or all of the images are aligned to form a composite image in the form of a continuous pictorial representation of the geographic area. One commercial embodiment of a continuous pictorial representation includes RibbonView™ by Real Data Systems. RibbonView™ correlates a ribbon of geographic image data with geospatial designations to facilitate identification of a particular geographic area, as discussed more fully below. In various embodiments, image capture processing can be accomplished in real time or via post image capture processing.
  • Referring now to FIG. 27, in some embodiments, select portions 104-107 of two or more sets of captured image data are aligned to generate the composite image 108. As illustrated, some preferred embodiments include vertical slices of data 104-107 aligned in a horizontal plane to form the composite image 108. Unlike a traditional photograph taken with a macro lens, according to the present invention, the length of a horizontal plane defining a composite image 108 is only limited by the length of a continuum along which points are defined and from which image data 104-107 is captured.
  • The use of only slices of data 104-107 from any particular captured image provides for a higher quality image 108. The quality is increased, for example, when a temporary obstruction, such as a passing car, person or animal, captured in one image data set, is only represented in a thin slice of a continuous ribbon 104-107 data. In addition, alignment of multiple thin slices of image data 104-107 is facilitated from the perspective of which aberrations typical human sensory is capable of distinguishing. Numerous thin slices 104-107 are perceived as a smoother transition across the horizontal plane defining the composite image 108 removing a choppiness that may be experienced with the use of larger data images.
  • The present invention can include a uniform width of each slice of data 104-107 or a variable width. The width of a particular slice may vary, for example, upon one or more of the velocity of a vehicle from which image data sets are captured, the sample rate of a camera used to capture an image data set 101-103, the resolution of a picture comprising an image data set 101-103 and the path of a camera. A high resolution image generated by a 2.1 mega pixel camera may have a 1600 by 1200 resolution and allow for a thinner slice 104-107 that includes a width of between about 5 to 100 pixels of an image data set. Embodiments with very high quality can include a slice 104-107 of between 1 to 2 pixels.
  • From a different perspective, some embodiments can include a slice 104-107 of an image data set 101-103 that includes a percentage of the image data set 101-103. Accordingly, some preferred embodiments can include a slice 104-107 of between about 5% to about 12% of an image data set. Other preferred embodiments can include up to about 50% of an image data set. However, it should be understood that some embodiments include a slice 104-107 that constitutes any fraction of the complete image data set.
  • Referring now to FIG. 28, three sets of image data 201-203 are illustrated, wherein each set of image data 201-203 represents data captured from a disparate point on a continuum proximate to the landscape captured. A direction of travel along the continuum 200 is also illustrated. Each data set 201-203 is divided into multiple slices of image data 204-209. A composite image 210 is generated from the multiple slices 204-209, and in the exemplary case at hand, additional slices from additional image data sets.
  • It should be noted that although preferred embodiments may utilize vertical rectangular slices 204-209, the scope of the present invention is not limited by which portion of an image data set 201-203 is utilized to generate a composite image 210. Therefore, it is within the scope of the invention to use any portion of any image data set 201-203 to generate a composite image. Accordingly, slices of an image 201-203 other than vertical slices 204-209 may be apportioned and combined into a composite image 210. Slices may therefore include a slice angled at 60° or 75°, or other angle conducive to a particular application. In addition, it is also within the scope of the present invention to utilize irregular shaped portions of two or more image data sets 201-203 to generate a composite image 210.
  • In some embodiments, a database or other data processing mechanism, can track each portion or slice 204-208 utilized to construct a continuum and associate the slice 204-208 with an original image 201-203 from which the slice 204-208 is derived. User interactive devices can execute the retrieval of an entire original image 201-203 or a series of original images 201-203 upon request. In some instances, one or more original images 201-203 may provide detail not included in the composite image 208.
  • Referring now to FIG. 29, in some embodiments, a selected portion 303 of an image data set 300 may be determined by the physical attributes of the equipment used to capture an image data set. For example, a typical camera lens can impart some distortion to an image data set, as represented in the illustration by an elongated portion 301 and a compressed portion 302. Utilizing only a portion 302 of an image data set 300, such as, for example, a center portion vertical slice 303, can minimize the effect of distortion introduced by a lens, or other source, to a composite image 108. Distortion is minimized when the composite image 108 is made to include those portions of the image data set corresponding with the center of the lens 303. Specific embodiments may likewise account for other aberrations that may be present in a set of image data.
  • In another aspect of the invention, and referring now to FIG. 30, positional data descriptive of a location of the subject matter of an image can also be generated by the image data recording device. Positional data can include any data indicative of where the subject matter of an image is located. Some preferred embodiments can include Cartesian coordinates that are context sensitive according to the mechanism used to generate coordinate data.
  • For example, an image recording device, such as a camera, can be associated with a device for recording a global position, such as a global positioning system (GPS) device or other equipment. Time stamps associated with image data and time stamps associated with the GPS data can be utilized to correlate the GPS location data with image data recorded by the camera.
  • In still another aspect, in some embodiments, an altimeter can be used to record an altitude from which a camera records image data sets. The altitude data can be associated with an image data set, for example, metadata correlated with the image data set. Such embodiments can therefore include recordation of a latitude, longitude and altitude coordinate for a given image data set. In addition, it is also within the scope of this invention to record a time of generation of an image data set and a depth of focus for an image data set.
  • According to the present invention, geospatial data, such as latitude and longitude coordinates, can be generated by the GPS and stored with image data generated by the camera. In some embodiments, GPS data can be time stamped and collected once every second. However, in some instances, GPS reception can be interrupted, depending upon location relative to large object, such as multistory buildings, or cold cover. Therefore, some additional embodiments can include an accelerometer for recording motion associated with a camera and a GPS unit operatively attached to the camera.
  • Data from the accelerometer can indicate movement of the camera. Some accelerometers, such as micro electromechanical system (MEMs) accelerometers can easily be incorporated into a camera system assembly attached to a vehicle. Use of multiple MEM accelerometers positioned to measure movement in four or more directions along an x-axis, y-axis, and z-axis in relation to a camera can also be used to calculate direction of movement. The accelerometer can therefore be used to extrapolate a current position of the camera, based upon a last set of GPS geospatial data recorded.
  • Geospatial data can be used to indicate an initial geographic position. A change in geospatial data can be additionally utilized to indicate velocity and direction of image data set capture. Accelerometer data may also be used to indicate a velocity and direction of image data set capture. Accelerometer data may also be used to indicate a velocity and direction of camera movement. Calculations of time elapsed at the indicated velocity (such as for example, the Kalman Filter) can yield a calculated position at a time of image capture, even if the time of image capture is between GPS readings.
  • For example, one standard can include tracking a camera position with a GPS unit that records location at a rate of once per second. The camera can record image data at a faster rate than once per second, such as, for example, one of: 12 images per second, 24 images per second or 29.97 images per second. An initial camera position can be recorded which correlates with a GPS reading, subsequent image data capture will occur in between the one second GPS reading interval. The camera position can be determined with a calculation based upon the velocity of camera movement supplied by the accelerometer and time elapsed since a previous GPS reading.
  • Still other embodiments can utilize optical flow methodology and visual odometry to facilitate calculations of a camera position and the velocity of a vehicle or person from which a series of image data sets are captured. Visual odometry can be accomplished with a single omni-directional camera or with stereo cameras, and is based upon processing which tracks the position of salient features in a series of feature sets and calculates movement based upon the relative positions of the features. In some embodiments, camera based simultaneous localization and mapping (SLAM) of visual image data can also be used to facilitate calculations of velocity of a change in position of a camera used to capture image data sets. Typically, the velocity will be directly tied to the motion of a vehicle to which the camera is mounted, or a person carrying a camera rig.
  • Orientation of a camera can include a direction of image capture recorded by the camera. Orientation can be designated, for example, in relation to the cardinal directions, i.e. north, south, east and west. Any means available to record such a designation, such as an electronic compass, is within the scope of the present invention. However, it may be desirable to include a means to record the orientation with a greater degree of accuracy than is typically available through the use of an electronic compass.
  • Therefore, in some embodiments, orientation can be determined according to a fixed position of a camera in relation to the direction of travel of a vehicle (or person) used to transport the camera. For example, a plurality of cameras can be fixedly attached to a vehicle capturing Video DriveBy™ data. Each camera therefore maintains a constant direction of image capture in relation to the heading of the vehicle. Mechanics of the camera, such as, for example, lens parameters and shutter speed, can indicate a depth of field during camera image data capture. Some embodiments can also include simultaneously capturing multiple image data sets and correlating two or more of the image data sets. Correlation can be accomplished via a time stamp or other chronological or synchronous.
  • The position of a camera can be combined with a direction of image capture and the depth of field of the camera, to determine a location of image data captured by the camera at a particular instance in time. The present invention can also include apparatus for utilizing echo location to determine a distance of an object from a camera capturing an image data set and storing a correlated distance with the image data set. For example, radar data can be correlated with a camera image data set to calculate the location of various objects captured by the camera. A time stamp can also be combined with data to quantify a location for a particular image formed by the captured image data.
  • In some embodiments of the present invention data used to calculate a location of an image is stored in a metadata file space associated with the image data. For example, some embodiments can store metadata in the exchangeable image file format (EXIF), TIFFTAGS or International Press Telecommunication Council (IPTC) formats. Image data may be stored, for example in JPEG or TIFF formats. However, other metadata formats can also be used. Typically, due to the size of data files that are generated during capture of Video DriveBy™ data, the image data and metafile data are stored on an external data storage device, such as a hard disk drive operatively attached to the camera. However, in some embodiments, the data can be stored in the camera.
  • As discussed above, the metadata can include data descriptive of some or all of: date and time; camera settings such aperture, shutter speed and focal length; geospatial data from a GPS receiver unit; accelerometer data; inertial guidance system data; camera orientation; and camera fixed position related to vehicle travel.
  • Referring now to FIG. 31, an illustration of some RibbonView™ embodiments of the present invention is illustrated. The RibbonView™ illustrated 501-502 includes strips or ribbons of two-dimensional image data. Each strip 501-502 is generated via the capture of image data along a continuum 503. As illustrated, the continuum 503 generally follows the roadway in the center of the illustration. Each composite image 501-502 is generated through the alignment of a portion of multiple images captured as data from disparate points along the continuum 503. In some preferred embodiments, the present invention delivers two ribbons of composite images 501-502 with different views captured along the continuum. Each view is generated by a camera with a separate orientation.
  • It will be apparent to those schooled in the art that the length of a composite image generated according to the present invention is limited only by the ability to capture image data from additional points on a continuum and store the captured image data for post processing. The post processing allows for the alignment of portions of the image data compiled into a composite two-dimensional view that can continue so long as additional image data is made available to be added to it.
  • Apparatus
  • The teachings of the present invention may be implemented with any apparatus capable of embodying the innovative concepts described herein. Image capture can be accomplished, for example, via a digital camera capable of capturing 12 or more images per second. In addition, FIG. 32 illustrates a controller 600 that may be utilized to implement some embodiments of the present invention. The controller 600 comprises a processor unit 610, such as one or more processors, coupled to a communication device 620 configured to communicate via a communication network (not shown in FIG. 32). The communication device 620 may be used to communicate, for example, with one or more online devices, such as a personal computer, laptop or a handheld device.
  • The processor 610 is also in communication with a storage device 630. The storage device 630 may comprise any appropriate information storage device, including combinations of magnetic storage devices (e.g., magnetic tape and hard disk drives), optical storage devices, and/or semiconductor memory devices such as Random Access Memory (RAM) devices and Read Only Memory (ROM) devices.
  • The storage device 630 can store a program 640 for controlling the processor 610. The processor 610 performs instructions of the program 640, and thereby operates in accordance with the present invention. The processor 610 may also cause the communication device 620 to transmit information, including, in some instances, control commands to operate apparatus to implement the processes described above. The storage device 630 can additionally store related data in a database 630A and database 630B, as needed.
  • CONCLUSION
  • The invention has been described with reference to certain preferred embodiments thereof. It will be understood, however, that modifications and variations are possible within the scope of the appended claims. For example, various methods or equipment may be used to implement the process steps described herein or to create a device according to the inventive concepts provided above and further described in the claims. In addition, various integration of components, as well as software and firmware can be implemented.
  • For example, while the illustrated embodiment utilizes a camera array containing analog cameras requiring digital conversation, the camera array may be implemented with cameras that directly produce a digital output. Also, digital cameras may be employed instead of analog cameras with separate A/D converters. The number of cameras in the array may also vary. Further, while the preferred embodiment utilizes external hard disk drives, it will be understood that any type of portable memory device having sufficient capacity to store the video data may be utilized. Still further, it will be understood that the various processing methods may be implemented utilizing the servers of the Video and Data Server Farm, the processing unit, i.e. the laptop computer, of the Video Capture System, and/or other processors linked thereto through the Internet or other network means, including various combinations of all three. Accordingly, other embodiments are within the scope of the following claims.

Claims (21)

1-21. (canceled)
22. A system comprising:
an image storage server that stores image files derived from drive-by image data acquired at street level;
a database server that processes a query relating to a geographic location of interest received from a user over a communications network; and
an image processing server, wherein the database server identifies image files stored in the image storage server that correspond to the geographic location of interest contained in the query and transfers the image files to the image processing server, and wherein the image processing server converts the drive-by image data to post-processed image data in a predetermined image format and transfers the post-processed image data via the communications network to the user in response to the query.
23. The system of claim 22, further comprising:
a landing zone server that receives the drive-by image data from a portable memory device and permits the viewing of the drive-by image data prior to storage in the image storage server.
24. The system of claim 22, further comprising:
a map server that presents a static image of an overhead view of the geographic location of interest.
25. The system of claim 22, wherein the drive-by image data includes individual street-level images.
26. The system of claim 22, wherein the drive-by image data is formatted as video data.
27. The system of claim 22, further comprising:
a camera that generates video data at the geographic location of interest;
a processing unit that calculates a camera position of a designated individual frame of the video data by capturing position data points before and after the designated individual frame was acquired.
28. The system of claim 22, wherein the image processing server stitches together a plurality of frames of video data to form a composite horizontal image of a street view.
29. The system of claim 22, further comprising:
a camera array that generates street-level images of the geographic location of interest;
a positioning unit that produces positioning data corresponding to a location of the camera array; and
a processing unit that processes the street-level images and the positioning data to generate the drive-by image data, wherein the processing unit stores the drive-by image data on a portable memory device.
30. The system of claim 29, wherein the camera array includes a plurality of camera units that generate street-level images from which a 360-degree image can be stitched together.
31. The system of claim 29, wherein the processing unit generates the drive-by image data by linking latitude and longitude data to the street-level images.
32. The system of claim 22, further comprising:
a database of recognized text generated by OCR processing the drive-by image data.
33. The system of claim 22, wherein the system identifies a street-level image acquired with a camera orientation directed towards a centroid of a parcel having a street address selected by the user.
34. The system of claim 22, wherein the image processing server stitches together a plurality of individual street-level images to form a composite horizontal image of a street view.
35. The system of claim 34, wherein the image processing server superimposes county parcel data adjacent to a centroid of each parcel depicted in the composite horizontal image.
36. A method comprising:
storing image files in a image storage server that are derived from drive-by image data coded with latitude and longitude values corresponding to the geographic locations at which the drive-by image data was acquired;
receiving a query from a user over a communications network, wherein the query relates to a geographic location of interest;
identifying those image files that correspond to the geographic location of interest;
transmitting the identified image files to an image processing server;
converting the identified image files to a predetermined image format; and
transmitting the identified image files in the predetermined image format over the communications network to the user.
37. The method of claim 36, wherein the predetermined image format is a 360-degree image stitched together from a plurality of image files corresponding to the geographic location of interest.
38. The method of claim 36, further comprising:
acquiring digital images using a camera array;
generating the latitude and longitude values corresponding to the location of the camera array;
generating the drive-by image data by combining the digital images with the latitude and longitude values; and
storing the drive-by image data on a portable memory device.
39. The method of claim 36, further comprising:
receiving an uploaded digital image from the user;
comparing the uploaded digital image to digital images contained in the image files in the image storage server;
generating a list of possible matches between the uploaded digital image and the digital images contained in the image files.
40. The method of claim 36, wherein the query designates the geographic location of interest by a street address, wherein the predetermined image format is a composite horizontal image of the street address stitched together from a plurality of digital images in the image files corresponding to the street address.
41. The method of claim 40, further comprising:
displaying the composite horizontal image with county parcel data adjacent to a parcel corresponding to the street address.
US13/795,959 2004-08-31 2013-03-12 Generating Video Drive-By Data Corresponding to a Geographic Location and Related Three Dimensional Models Abandoned US20130188008A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/795,959 US20130188008A1 (en) 2004-08-31 2013-03-12 Generating Video Drive-By Data Corresponding to a Geographic Location and Related Three Dimensional Models

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
US60549804P 2004-08-31 2004-08-31
US11/216,465 US7389181B2 (en) 2004-08-31 2005-08-31 Apparatus and method for producing video drive-by data corresponding to a geographic location
US12/036,197 US8207964B1 (en) 2008-02-22 2008-02-22 Methods and apparatus for generating three-dimensional image data models
US13/481,912 US9384277B2 (en) 2004-08-31 2012-05-28 Three dimensional image data models
US13/795,959 US20130188008A1 (en) 2004-08-31 2013-03-12 Generating Video Drive-By Data Corresponding to a Geographic Location and Related Three Dimensional Models

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US13/481,912 Continuation US9384277B2 (en) 2004-08-31 2012-05-28 Three dimensional image data models

Publications (1)

Publication Number Publication Date
US20130188008A1 true US20130188008A1 (en) 2013-07-25

Family

ID=46272907

Family Applications (18)

Application Number Title Priority Date Filing Date
US12/036,197 Ceased US8207964B1 (en) 2004-08-31 2008-02-22 Methods and apparatus for generating three-dimensional image data models
US13/481,912 Expired - Fee Related US9384277B2 (en) 2004-08-31 2012-05-28 Three dimensional image data models
US13/791,945 Abandoned US20130188052A1 (en) 2004-08-31 2013-03-09 Method of Providing Street View Data of a Comparable Real Estate Property
US13/791,953 Expired - Fee Related US9311397B2 (en) 2004-08-31 2013-03-09 Method and apparatus of providing street view data of a real estate property
US13/791,961 Expired - Fee Related US8558848B2 (en) 2004-08-31 2013-03-09 Wireless internet-accessible drive-by street view system and method
US13/791,938 Abandoned US20130201340A1 (en) 2004-08-31 2013-03-09 Method of viewing street View Data of a Real Estate Property
US13/791,951 Abandoned US20130191725A1 (en) 2004-08-31 2013-03-09 Method and Apparatus of Viewing Street View Data of a Real Estate Property
US13/791,947 Abandoned US20130191295A1 (en) 2004-08-31 2013-03-09 Method of Receiving a List of Links to Street View Data of Real Estate Properties
US13/791,957 Abandoned US20130191359A1 (en) 2004-08-31 2013-03-09 Method and Apparatus of Receiving a List of Links to Street View Data of Real Estate Properties
US13/791,955 Expired - Fee Related US8890866B2 (en) 2004-08-31 2013-03-09 Method and apparatus of providing street view data of a comparable real estate property
US13/791,943 Expired - Fee Related US9311396B2 (en) 2004-08-31 2013-03-09 Method of providing street view data of a real estate property
US13/796,544 Expired - Fee Related US8902226B2 (en) 2004-08-31 2013-03-12 Method for using drive-by image data to generate a valuation report of a selected real estate property
US13/795,959 Abandoned US20130188008A1 (en) 2004-08-31 2013-03-12 Generating Video Drive-By Data Corresponding to a Geographic Location and Related Three Dimensional Models
US13/830,164 Abandoned US20130201341A1 (en) 2004-08-31 2013-03-14 Generating a Point of Interest Database with Specific Geographic Locations Derived from Street-Level Image Data
US13/861,336 Expired - Fee Related USRE45264E1 (en) 2004-08-31 2013-04-11 Methods and apparatus for generating three-dimensional image data models
US14/023,028 Abandoned US20140019301A1 (en) 2004-08-31 2013-09-10 Internet-accessible drive-by street view system and method
US14/243,712 Abandoned US20140211015A1 (en) 2004-08-31 2014-04-02 Method and apparatus for providing image data of property
US15/200,784 Abandoned US20170094165A1 (en) 2004-08-31 2016-07-01 Method and apparatus for providing image data of property

Family Applications Before (12)

Application Number Title Priority Date Filing Date
US12/036,197 Ceased US8207964B1 (en) 2004-08-31 2008-02-22 Methods and apparatus for generating three-dimensional image data models
US13/481,912 Expired - Fee Related US9384277B2 (en) 2004-08-31 2012-05-28 Three dimensional image data models
US13/791,945 Abandoned US20130188052A1 (en) 2004-08-31 2013-03-09 Method of Providing Street View Data of a Comparable Real Estate Property
US13/791,953 Expired - Fee Related US9311397B2 (en) 2004-08-31 2013-03-09 Method and apparatus of providing street view data of a real estate property
US13/791,961 Expired - Fee Related US8558848B2 (en) 2004-08-31 2013-03-09 Wireless internet-accessible drive-by street view system and method
US13/791,938 Abandoned US20130201340A1 (en) 2004-08-31 2013-03-09 Method of viewing street View Data of a Real Estate Property
US13/791,951 Abandoned US20130191725A1 (en) 2004-08-31 2013-03-09 Method and Apparatus of Viewing Street View Data of a Real Estate Property
US13/791,947 Abandoned US20130191295A1 (en) 2004-08-31 2013-03-09 Method of Receiving a List of Links to Street View Data of Real Estate Properties
US13/791,957 Abandoned US20130191359A1 (en) 2004-08-31 2013-03-09 Method and Apparatus of Receiving a List of Links to Street View Data of Real Estate Properties
US13/791,955 Expired - Fee Related US8890866B2 (en) 2004-08-31 2013-03-09 Method and apparatus of providing street view data of a comparable real estate property
US13/791,943 Expired - Fee Related US9311396B2 (en) 2004-08-31 2013-03-09 Method of providing street view data of a real estate property
US13/796,544 Expired - Fee Related US8902226B2 (en) 2004-08-31 2013-03-12 Method for using drive-by image data to generate a valuation report of a selected real estate property

Family Applications After (5)

Application Number Title Priority Date Filing Date
US13/830,164 Abandoned US20130201341A1 (en) 2004-08-31 2013-03-14 Generating a Point of Interest Database with Specific Geographic Locations Derived from Street-Level Image Data
US13/861,336 Expired - Fee Related USRE45264E1 (en) 2004-08-31 2013-04-11 Methods and apparatus for generating three-dimensional image data models
US14/023,028 Abandoned US20140019301A1 (en) 2004-08-31 2013-09-10 Internet-accessible drive-by street view system and method
US14/243,712 Abandoned US20140211015A1 (en) 2004-08-31 2014-04-02 Method and apparatus for providing image data of property
US15/200,784 Abandoned US20170094165A1 (en) 2004-08-31 2016-07-01 Method and apparatus for providing image data of property

Country Status (1)

Country Link
US (18) US8207964B1 (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120330646A1 (en) * 2011-06-23 2012-12-27 International Business Machines Corporation Method For Enhanced Location Based And Context Sensitive Augmented Reality Translation
CN104361586A (en) * 2014-10-31 2015-02-18 国网上海市电力公司 Electric pole and tower maintenance and pre-warning system based on panoramas
US9245183B2 (en) 2014-06-26 2016-01-26 International Business Machines Corporation Geographical area condition determination
US10049589B1 (en) 2016-09-08 2018-08-14 Amazon Technologies, Inc. Obstacle awareness based guidance to clear landing space
US10121117B1 (en) 2016-09-08 2018-11-06 Amazon Technologies, Inc. Drone location signature filters
US20180352137A1 (en) * 2017-06-02 2018-12-06 L-Tron Corporation Event data and location-linked spherical imaging system
US10156631B2 (en) 2014-12-19 2018-12-18 Xidrone Systems, Inc. Deterrent for unmanned aerial systems
US10198955B1 (en) * 2016-09-08 2019-02-05 Amazon Technologies, Inc. Drone marker and landing zone verification
US10281570B2 (en) 2014-12-19 2019-05-07 Xidrone Systems, Inc. Systems and methods for detecting, tracking and identifying small unmanned systems such as drones
US10498955B2 (en) 2015-08-03 2019-12-03 Disney Enterprises, Inc. Commercial drone detection
CN110855906A (en) * 2018-07-24 2020-02-28 黑芝麻智能科技(上海)有限公司 Method for splicing optical images in 360-degree panoramic looking around by utilizing cameras and radars arranged around vehicle
US10907940B1 (en) 2017-12-12 2021-02-02 Xidrone Systems, Inc. Deterrent for unmanned aerial systems using data mining and/or machine learning for improved target detection and classification
TWI734037B (en) * 2018-09-28 2021-07-21 愛探極溫度行銷有限公司 Real estate holding and inheritance calculation management system
US11470303B1 (en) 2010-06-24 2022-10-11 Steven M. Hoffberg Two dimensional to three dimensional moving image converter

Families Citing this family (192)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7949191B1 (en) * 2007-04-04 2011-05-24 A9.Com, Inc. Method and system for searching for information on a network in response to an image query sent by a user from a mobile communications device
US9953459B2 (en) 2008-11-05 2018-04-24 Hover Inc. Computer vision database platform for a three-dimensional mapping system
US8606774B1 (en) * 2009-05-18 2013-12-10 Google Inc. Methods and systems for 3D shape retrieval
US8611592B2 (en) * 2009-08-26 2013-12-17 Apple Inc. Landmark identification using metadata
US9286624B2 (en) * 2009-09-10 2016-03-15 Google Inc. System and method of displaying annotations on geographic object surfaces
KR20110052124A (en) * 2009-11-12 2011-05-18 삼성전자주식회사 Method for generating and referencing panorama image and mobile terminal using the same
US9749597B2 (en) * 2010-08-10 2017-08-29 International Business Machines Corporation Precise target positioning in geographical imaging
US9317133B2 (en) * 2010-10-08 2016-04-19 Nokia Technologies Oy Method and apparatus for generating augmented reality content
US20120150573A1 (en) * 2010-12-13 2012-06-14 Omar Soubra Real-time site monitoring design
US8810598B2 (en) 2011-04-08 2014-08-19 Nant Holdings Ip, Llc Interference based augmented reality hosting platforms
US9552376B2 (en) 2011-06-09 2017-01-24 MemoryWeb, LLC Method and apparatus for managing digital files
WO2013128078A1 (en) 2012-02-29 2013-09-06 Nokia Corporation Method and apparatus for rendering items in a user interface
JP2015501984A (en) 2011-11-21 2015-01-19 ナント ホールディングス アイピー,エルエルシー Subscription bill service, system and method
US9002114B2 (en) 2011-12-08 2015-04-07 The Nielsen Company (Us), Llc Methods, apparatus, and articles of manufacture to measure geographical features using an image of a geographical location
US9501856B2 (en) * 2012-02-13 2016-11-22 Nokia Technologies Oy Method and apparatus for generating panoramic maps with elements of subtle movement
US9501700B2 (en) 2012-02-15 2016-11-22 Xactware Solutions, Inc. System and method for construction estimation using aerial images
US9378509B2 (en) 2012-05-09 2016-06-28 The Nielsen Company (Us), Llc Methods, apparatus, and articles of manufacture to measure geographical features using an image of a geographical location
US10296516B2 (en) * 2012-05-21 2019-05-21 Here Global B.V. Method and apparatus for navigation using multiple synchronized mobile devices
US8923893B2 (en) 2012-08-07 2014-12-30 Symbol Technologies, Inc. Real-time planogram generation and maintenance
US9154805B2 (en) * 2012-09-12 2015-10-06 Advanced Micro Devices, Inc. Video and image compression based on position of the image generating device
US9488489B2 (en) 2012-09-28 2016-11-08 Google Inc. Personalized mapping with photo tours
US10771748B2 (en) * 2012-11-27 2020-09-08 Metropolitan Life Insurance Co. System and method for interactive aerial imaging
US9071756B2 (en) 2012-12-11 2015-06-30 Facebook, Inc. Systems and methods for digital video stabilization via constraint-based rotation smoothing
US10924627B2 (en) * 2012-12-31 2021-02-16 Virtually Anywhere Content management for virtual tours
US9845636B2 (en) 2013-01-07 2017-12-19 WexEnergy LLC Frameless supplemental window for fenestration
US10883303B2 (en) 2013-01-07 2021-01-05 WexEnergy LLC Frameless supplemental window for fenestration
US9691163B2 (en) * 2013-01-07 2017-06-27 Wexenergy Innovations Llc System and method of measuring distances related to an object utilizing ancillary objects
US10196850B2 (en) 2013-01-07 2019-02-05 WexEnergy LLC Frameless supplemental window for fenestration
US20140253540A1 (en) * 2013-03-07 2014-09-11 Yoav DORI Method and system of incorporating real world objects into a virtual environment
US9082014B2 (en) 2013-03-14 2015-07-14 The Nielsen Company (Us), Llc Methods and apparatus to estimate demography based on aerial images
US9753950B2 (en) * 2013-03-15 2017-09-05 Pictometry International Corp. Virtual property reporting for automatic structure detection
US8874454B2 (en) * 2013-03-15 2014-10-28 State Farm Mutual Automobile Insurance Company Systems and methods for assessing a roof
US9756138B2 (en) * 2013-04-08 2017-09-05 Here Global B.V. Desktop application synchronization to process data captured on a mobile device
US10134194B2 (en) * 2013-07-17 2018-11-20 Evernote Corporation Marking up scenes using a wearable augmented reality device
US11721066B2 (en) 2013-07-23 2023-08-08 Hover Inc. 3D building model materials auto-populator
WO2015017855A1 (en) 2013-08-02 2015-02-05 Xactware Solutions, Inc. System and method for detecting features in aerial images using disparity mapping and segmentation techniques
US9903719B2 (en) * 2013-09-03 2018-02-27 Litel Instruments System and method for advanced navigation
US9208171B1 (en) * 2013-09-05 2015-12-08 Google Inc. Geographically locating and posing images in a large-scale image repository and processing framework
US9244940B1 (en) 2013-09-27 2016-01-26 Google Inc. Navigation paths for panorama
US9288368B2 (en) * 2013-10-08 2016-03-15 Delightfit, Inc. Video and map data synchronization for simulated athletic training
US9582516B2 (en) 2013-10-17 2017-02-28 Nant Holdings Ip, Llc Wide area augmented reality location-based services
WO2015061735A1 (en) 2013-10-25 2015-04-30 Hover Inc. Estimating dimensions of geo-referenced ground-level imagery using orthogonal imagery
US9836885B1 (en) 2013-10-25 2017-12-05 Appliance Computing III, Inc. Image-based rendering of real spaces
CN103914521B (en) * 2014-03-20 2015-02-25 北京建筑大学 Street view image storage method and device based on mixed tile pyramids
US9640223B2 (en) * 2014-03-27 2017-05-02 Tvu Networks Corporation Methods, apparatus and systems for time-based and geographic navigation of video content
US10327034B2 (en) 2014-03-27 2019-06-18 Tvu Networks Corporation Methods, apparatus and systems for exchange of video content
US9992246B2 (en) 2014-03-27 2018-06-05 Tvu Networks Corporation Methods, apparatus, and systems for instantly sharing video content on social media
US10062118B1 (en) 2014-04-02 2018-08-28 Liberty Mutual Insurance Company Concepts for providing an insurance quote
USD781317S1 (en) 2014-04-22 2017-03-14 Google Inc. Display screen with graphical user interface or portion thereof
USD781318S1 (en) 2014-04-22 2017-03-14 Google Inc. Display screen with graphical user interface or portion thereof
USD780777S1 (en) 2014-04-22 2017-03-07 Google Inc. Display screen with graphical user interface or portion thereof
US9972121B2 (en) 2014-04-22 2018-05-15 Google Llc Selecting time-distributed panoramic images for display
US9934222B2 (en) 2014-04-22 2018-04-03 Google Llc Providing a thumbnail image that follows a main image
US9189839B1 (en) 2014-04-24 2015-11-17 Google Inc. Automatically generating panorama tours
US10019761B2 (en) 2014-04-25 2018-07-10 State Farm Mutual Automobile Insurance Company System and method for virtual inspection of a structure
US10248731B1 (en) * 2014-05-16 2019-04-02 Corelogic Solutions, Llc System and method for linking data records for parcels
US9002647B1 (en) 2014-06-27 2015-04-07 Google Inc. Generating turn-by-turn direction previews
US9418472B2 (en) 2014-07-17 2016-08-16 Google Inc. Blending between street view and earth view
US20160042478A1 (en) * 2014-08-05 2016-02-11 Mastercard International Incorporated Methods and Systems for Verifying Images Associated With Offered Properties
US9724177B2 (en) 2014-08-19 2017-08-08 Align Technology, Inc. Viewfinder with real-time tracking for intraoral scanning
US9870596B2 (en) 2014-09-18 2018-01-16 International Business Machines Corporation Predicting community development trends
US9805061B2 (en) * 2014-11-18 2017-10-31 International Business Machines Corporation Image search for a location
US9830724B2 (en) * 2014-11-21 2017-11-28 Mastercard International Incorporated Detecting real estate vacancies using satellite imagery
US9851870B2 (en) * 2015-03-17 2017-12-26 Raytheon Company Multi-dimensional video navigation system and method using interactive map paths
US9924240B2 (en) 2015-05-01 2018-03-20 Google Llc Systems and methods for interactive video generation and rendering
US20170134815A1 (en) * 2015-05-06 2017-05-11 Joshua LOMELINO Systems for creating and distributing videos of real estate
JP6622990B2 (en) * 2015-06-29 2019-12-18 オリンパス株式会社 Information discrimination device, information discrimination method, and information discrimination program
AU2016315938B2 (en) * 2015-08-31 2022-02-24 Cape Analytics, Inc. Systems and methods for analyzing remote sensing imagery
US10885097B2 (en) 2015-09-25 2021-01-05 The Nielsen Company (Us), Llc Methods and apparatus to profile geographic areas of interest
US10220172B2 (en) 2015-11-25 2019-03-05 Resmed Limited Methods and systems for providing interface components for respiratory therapy
US9972125B2 (en) * 2015-12-16 2018-05-15 Google Inc. Split tile map rendering
CN105513127B (en) * 2015-12-25 2018-01-26 武汉大学 Shaft regularization three-dimensional modeling method and system based on density peaks cluster
US10352689B2 (en) 2016-01-28 2019-07-16 Symbol Technologies, Llc Methods and systems for high precision locationing with depth values
US10839569B2 (en) * 2016-02-03 2020-11-17 NorthStar Memorial Group LLC System for geospatial mapping of cemetery properties
CN105677899B (en) * 2016-02-04 2019-01-29 东南大学 A kind of enhanced vector numerical map production method comprising road direction of travel
US10204116B2 (en) * 2016-02-08 2019-02-12 Corelogic Solutions, Llc Image metadata enrichment generator
US10147224B2 (en) * 2016-02-16 2018-12-04 Samsung Electronics Co., Ltd. Method and apparatus for generating omni media texture mapping metadata
US10824320B2 (en) * 2016-03-07 2020-11-03 Facebook, Inc. Systems and methods for presenting content
EP3430591A4 (en) * 2016-03-16 2019-11-27 ADCOR Magnet Systems, LLC System for georeferenced, geo-oriented real time video streams
US10341617B2 (en) * 2016-03-23 2019-07-02 Purdue Research Foundation Public safety camera identification and monitoring system and method
US10506201B2 (en) * 2016-03-23 2019-12-10 Purdue Research Foundation Public safety camera identification and monitoring system and method
US10368037B2 (en) * 2016-03-23 2019-07-30 Purdue Research Foundation Public safety camera monitoring system and method
US10410441B2 (en) * 2016-05-16 2019-09-10 Wi-Tronix, Llc Real-time data acquisition and recording system viewer
CN106055610B (en) * 2016-05-25 2020-02-14 维沃移动通信有限公司 Voice information retrieval method and mobile terminal
US10643089B2 (en) 2016-10-13 2020-05-05 Ricoh Company, Ltd. Information processing system to obtain and manage images of a property
US10706839B1 (en) 2016-10-24 2020-07-07 United Services Automobile Association (Usaa) Electronic signatures via voice for virtual assistants' interactions
US11042161B2 (en) 2016-11-16 2021-06-22 Symbol Technologies, Llc Navigation control method and apparatus in a mobile automation system
JP6116746B1 (en) * 2016-11-17 2017-04-19 株式会社ネクスト Information processing apparatus, information processing method, and program
CN106713847A (en) * 2016-11-28 2017-05-24 天津商企生产力促进有限公司 Electromechanical integrated monitor based on virtual three-dimensional static scene
US20180150924A1 (en) * 2016-11-30 2018-05-31 Corelogic Solutions, Llc Computer-based photo re-use across property reports
US20180150923A1 (en) * 2016-11-30 2018-05-31 Corelogic Solutions, Llc Property study workflow system
US20180150925A1 (en) * 2016-11-30 2018-05-31 Corelogic Solutions, Llc Computer-based control of timing with which property review orders are processed
CN108462851A (en) * 2017-02-21 2018-08-28 北京视联动力国际信息技术有限公司 The method and mobile terminal of data processing, server, the second mobile terminal
US10360709B2 (en) * 2017-04-05 2019-07-23 Microsoft Technology Licensing, Llc Rendering images on map using orientations
GB2561882A (en) * 2017-04-27 2018-10-31 Nokia Technologies Oy Virtual reality viewport adaption
US10663590B2 (en) 2017-05-01 2020-05-26 Symbol Technologies, Llc Device and method for merging lidar data
US11093896B2 (en) 2017-05-01 2021-08-17 Symbol Technologies, Llc Product status detection system
US10726273B2 (en) 2017-05-01 2020-07-28 Symbol Technologies, Llc Method and apparatus for shelf feature and object placement detection from shelf images
US10591918B2 (en) 2017-05-01 2020-03-17 Symbol Technologies, Llc Fixed segmented lattice planning for a mobile automation apparatus
US11449059B2 (en) 2017-05-01 2022-09-20 Symbol Technologies, Llc Obstacle detection for a mobile automation apparatus
US10505057B2 (en) 2017-05-01 2019-12-10 Symbol Technologies, Llc Device and method for operating cameras and light sources wherein parasitic reflections from a paired light source are not reflected into the paired camera
US11367092B2 (en) 2017-05-01 2022-06-21 Symbol Technologies, Llc Method and apparatus for extracting and processing price text from an image set
DE112018002314T5 (en) 2017-05-01 2020-01-23 Symbol Technologies, Llc METHOD AND DEVICE FOR DETECTING AN OBJECT STATUS
US10949798B2 (en) 2017-05-01 2021-03-16 Symbol Technologies, Llc Multimodal localization and mapping for a mobile automation apparatus
US10628959B2 (en) 2017-05-03 2020-04-21 International Business Machines Corporation Location determination using street view images
WO2018201423A1 (en) 2017-05-05 2018-11-08 Symbol Technologies, Llc Method and apparatus for detecting and interpreting price label text
CN108882018B (en) 2017-05-09 2020-10-20 阿里巴巴(中国)有限公司 Video playing and data providing method in virtual scene, client and server
US10984182B2 (en) 2017-05-12 2021-04-20 Loveland Innovations, LLC Systems and methods for context-rich annotation and report generation for UAV microscan data
EP3631136A4 (en) 2017-05-30 2021-03-10 Wexenergy LLC Frameless supplemental window for fenestration
JP7199627B2 (en) 2017-05-31 2023-01-06 ウーブン プラネット ノース アメリカ,インコーポレイティド A vehicle system equipped with imaging equipment for high-definition near-real-time map generation
WO2019006189A1 (en) 2017-06-29 2019-01-03 Open Space Labs, Inc. Automated spatial indexing of images based on floorplan features
US11748830B2 (en) * 2017-08-11 2023-09-05 Tellurium Inc. Distributed ledger based system and method for the settlement and transfer of title to real estate
US10460465B2 (en) 2017-08-31 2019-10-29 Hover Inc. Method for generating roof outlines from lateral images
US10572763B2 (en) 2017-09-07 2020-02-25 Symbol Technologies, Llc Method and apparatus for support surface edge detection
US10521914B2 (en) 2017-09-07 2019-12-31 Symbol Technologies, Llc Multi-sensor object recognition system and method
CN107590854A (en) * 2017-09-15 2018-01-16 福建四创软件有限公司 Reservoir region three-dimensional live methods of exhibiting based on WEBGIS
US11200513B2 (en) 2017-10-13 2021-12-14 Carrier Corporation Real estate image analysis
US10936176B1 (en) * 2017-11-17 2021-03-02 United Services Automobile Association (Usaa) Systems and methods for interactive maps
RU2670429C1 (en) * 2017-11-24 2018-10-23 ООО "Ай Ти Ви групп" Systems and methods of tracking moving objects on video image
US11403816B2 (en) * 2017-11-30 2022-08-02 Mitsubishi Electric Corporation Three-dimensional map generation system, three-dimensional map generation method, and computer readable medium
CN107958441B (en) * 2017-12-01 2021-02-12 深圳市科比特航空科技有限公司 Image splicing method and device, computer equipment and storage medium
CN108182218B (en) * 2017-12-21 2021-11-19 深圳先进技术研究院 Video character recognition method and system based on geographic information system and electronic equipment
US10339384B2 (en) 2018-02-07 2019-07-02 Structionsite Inc. Construction photograph integration with 3D model images
US10791268B2 (en) 2018-02-07 2020-09-29 Structionsite Inc. Construction photograph integration with 3D model images
CN108600607A (en) * 2018-03-13 2018-09-28 上海网罗电子科技有限公司 A kind of fire-fighting panoramic information methods of exhibiting based on unmanned plane
CN108509191A (en) * 2018-03-16 2018-09-07 杭州聚秀科技有限公司 A kind of housing property management WEB graphic methods
US10832436B2 (en) 2018-04-05 2020-11-10 Symbol Technologies, Llc Method, system and apparatus for recovering label positions
US10823572B2 (en) 2018-04-05 2020-11-03 Symbol Technologies, Llc Method, system and apparatus for generating navigational data
US10740911B2 (en) 2018-04-05 2020-08-11 Symbol Technologies, Llc Method, system and apparatus for correcting translucency artifacts in data representing a support structure
US10809078B2 (en) 2018-04-05 2020-10-20 Symbol Technologies, Llc Method, system and apparatus for dynamic path generation
US11327504B2 (en) 2018-04-05 2022-05-10 Symbol Technologies, Llc Method, system and apparatus for mobile automation apparatus localization
US10467758B1 (en) 2018-07-13 2019-11-05 Structionsite Inc. Imagery-based construction progress tracking
CN108965718B (en) * 2018-08-03 2021-03-23 北京微播视界科技有限公司 Image generation method and device
CN109446281B (en) * 2018-09-05 2020-11-24 广州市城市规划勘测设计研究院 ArcGIS-based online intelligent mapping system construction method, system and mapping method
CN109389053B (en) * 2018-09-20 2021-08-06 同济大学 Method and system for detecting position information of vehicle to be detected around target vehicle
US11506483B2 (en) 2018-10-05 2022-11-22 Zebra Technologies Corporation Method, system and apparatus for support structure depth determination
US11010920B2 (en) 2018-10-05 2021-05-18 Zebra Technologies Corporation Method, system and apparatus for object detection in point clouds
US10944959B2 (en) 2018-11-12 2021-03-09 Open Space Labs, Inc. Automated spatial indexing of images to video
US11003188B2 (en) 2018-11-13 2021-05-11 Zebra Technologies Corporation Method, system and apparatus for obstacle handling in navigational path generation
US11090811B2 (en) 2018-11-13 2021-08-17 Zebra Technologies Corporation Method and apparatus for labeling of support structures
EP3881161A1 (en) 2018-11-14 2021-09-22 Cape Analytics, Inc. Systems, methods, and computer readable media for predictive analytics and change detection from remotely sensed imagery
US11079240B2 (en) 2018-12-07 2021-08-03 Zebra Technologies Corporation Method, system and apparatus for adaptive particle filter localization
US11416000B2 (en) 2018-12-07 2022-08-16 Zebra Technologies Corporation Method and apparatus for navigational ray tracing
US11100303B2 (en) 2018-12-10 2021-08-24 Zebra Technologies Corporation Method, system and apparatus for auxiliary label detection and association
US11015938B2 (en) 2018-12-12 2021-05-25 Zebra Technologies Corporation Method, system and apparatus for navigational assistance
US10731970B2 (en) 2018-12-13 2020-08-04 Zebra Technologies Corporation Method, system and apparatus for support structure detection
CN109740867B (en) * 2018-12-14 2021-06-11 广州供电局有限公司 Large-area power failure emergency repair scheme configuration method and device based on situation analysis
CA3028708A1 (en) 2018-12-28 2020-06-28 Zih Corp. Method, system and apparatus for dynamic loop closure in mapping trajectories
US10936178B2 (en) 2019-01-07 2021-03-02 MemoryWeb, LLC Systems and methods for analyzing and organizing digital photos and videos
CN109474645A (en) * 2019-01-29 2019-03-15 江苏省测绘工程院 A kind of application method of multi-source encryption GPS data
US10762690B1 (en) * 2019-02-11 2020-09-01 Apple Inc. Simulated overhead perspective images with removal of obstructions
CN109918466A (en) * 2019-03-08 2019-06-21 江西憶源多媒体科技有限公司 A kind of real-time map information overall situation rendering method based on video analysis
USD949159S1 (en) * 2019-06-02 2022-04-19 Apple Inc. Display screen or portion thereof with graphical user interface
US11080566B2 (en) 2019-06-03 2021-08-03 Zebra Technologies Corporation Method, system and apparatus for gap detection in support structures with peg regions
US11402846B2 (en) 2019-06-03 2022-08-02 Zebra Technologies Corporation Method, system and apparatus for mitigating data capture light leakage
US11662739B2 (en) 2019-06-03 2023-05-30 Zebra Technologies Corporation Method, system and apparatus for adaptive ceiling-based localization
US11151743B2 (en) 2019-06-03 2021-10-19 Zebra Technologies Corporation Method, system and apparatus for end of aisle detection
US11960286B2 (en) 2019-06-03 2024-04-16 Zebra Technologies Corporation Method, system and apparatus for dynamic task sequencing
US11200677B2 (en) 2019-06-03 2021-12-14 Zebra Technologies Corporation Method, system and apparatus for shelf edge detection
US11341663B2 (en) 2019-06-03 2022-05-24 Zebra Technologies Corporation Method, system and apparatus for detecting support structure obstructions
CN110599583B (en) * 2019-07-26 2022-03-18 深圳眸瞳科技有限公司 Unmanned aerial vehicle flight trajectory generation method and device, computer equipment and storage medium
US11971871B1 (en) * 2019-09-17 2024-04-30 Sketch Align, Inc. System controlled by data bearing records
WO2021055633A1 (en) * 2019-09-18 2021-03-25 Lopez Camilo System and method for generating a video
US20210133883A1 (en) 2019-11-04 2021-05-06 Tellus App, Inc. Decentralized architecture for property-backed vehicles and creation, publication, and distributed owner investment
US11094113B2 (en) 2019-12-04 2021-08-17 Geomni, Inc. Systems and methods for modeling structures using point clouds derived from stereoscopic image pairs
US11507103B2 (en) 2019-12-04 2022-11-22 Zebra Technologies Corporation Method, system and apparatus for localization-based historical obstacle handling
US11107238B2 (en) 2019-12-13 2021-08-31 Zebra Technologies Corporation Method, system and apparatus for detecting item facings
US11776140B2 (en) * 2019-12-16 2023-10-03 Insurance Services Office, Inc. Computer vision systems and methods for automatic alignment of parcels with geotagged aerial imagery
CN111240617B (en) * 2019-12-31 2020-09-25 飞燕航空遥感技术有限公司 Video delivery method and system based on three-dimensional map and environment monitoring method and system
US11146727B2 (en) * 2020-03-16 2021-10-12 Ke.Com (Beijing) Technology Co., Ltd. Method and device for generating a panoramic image
US11822333B2 (en) 2020-03-30 2023-11-21 Zebra Technologies Corporation Method, system and apparatus for data capture illumination control
EP3893482A1 (en) * 2020-04-07 2021-10-13 Parkling GmbH Transportable recording device for recording data for a located street panorama image
TWI744861B (en) * 2020-04-15 2021-11-01 財團法人工業技術研究院 Application service server, information processing method based on located region, and user equipment
WO2021231664A1 (en) * 2020-05-12 2021-11-18 Rocky Lee Systems and methods for interface-driven engagement
CN111666434B (en) * 2020-05-26 2021-11-02 武汉大学 Streetscape picture retrieval method based on depth global features
US11436812B2 (en) 2020-05-29 2022-09-06 Open Space Labs, Inc. Machine learning based object identification using scaled diagram and three-dimensional model
US11302034B2 (en) * 2020-07-09 2022-04-12 Tensorflight, Inc. Automated property inspections
US20220383436A1 (en) * 2020-07-09 2022-12-01 TensorFlight Poland sp. z o. o. Automated Property Inspections
US11450024B2 (en) 2020-07-17 2022-09-20 Zebra Technologies Corporation Mixed depth object detection
US12094494B2 (en) 2020-09-17 2024-09-17 Camilo Lopez System and method for generating and editing a video
CN112163412B (en) * 2020-09-30 2024-02-09 善恒展创(深圳)发展合伙企业(有限合伙) Data verification method and device, electronic equipment and storage medium
WO2022082007A1 (en) 2020-10-15 2022-04-21 Cape Analytics, Inc. Method and system for automated debris detection
US11593915B2 (en) 2020-10-21 2023-02-28 Zebra Technologies Corporation Parallax-tolerant panoramic image generation
US11392891B2 (en) 2020-11-03 2022-07-19 Zebra Technologies Corporation Item placement detection and optimization in material handling systems
US11847832B2 (en) 2020-11-11 2023-12-19 Zebra Technologies Corporation Object classification for autonomous navigation systems
CN112508784A (en) * 2020-11-20 2021-03-16 中南大学 Panoramic image method of planar object contour model based on image stitching
CN112730743A (en) * 2020-12-17 2021-04-30 中科三清科技有限公司 Interaction method and device for air quality mode forecast data
US11954882B2 (en) 2021-06-17 2024-04-09 Zebra Technologies Corporation Feature-based georegistration for mobile computing devices
US20240221140A1 (en) 2021-07-06 2024-07-04 Cape Analytics, Inc. System and method for property condition analysis
WO2023114027A1 (en) 2021-12-16 2023-06-22 Cape Analytics, Inc. System and method for change analysis
AU2023208758A1 (en) 2022-01-19 2024-06-20 Cape Analytics, Inc. System and method for object analysis
WO2024070532A1 (en) * 2022-09-26 2024-04-04 株式会社リコー Information processing device, information processing method, program, and information processing system

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120059720A1 (en) * 2004-06-30 2012-03-08 Musabji Adil M Method of Operating a Navigation System Using Images

Family Cites Families (119)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4396942A (en) * 1979-04-19 1983-08-02 Jackson Gates Video surveys
FR2461305B1 (en) * 1979-07-06 1985-12-06 Thomson Csf MAP INDICATOR SYSTEM MORE PARTICULARLY FOR AIR NAVIGATION
US4553206A (en) * 1983-10-03 1985-11-12 Wang Laboratories, Inc. Image storage and retrieval
US4635136A (en) * 1984-02-06 1987-01-06 Rochester Institute Of Technology Method and apparatus for storing a massive inventory of labeled images
NL8503378A (en) * 1985-12-06 1987-07-01 Theo Jogchum Poelstra NEW METHOD FOR BUILDING AND TRACKING DATA FILES FOR ROAD TRAFFIC.
US5032989A (en) * 1986-03-19 1991-07-16 Realpro, Ltd. Real estate search and location system and method
US4870576A (en) * 1986-03-19 1989-09-26 Realpro, Ltd. Real estate search and location system and method
CA1288516C (en) * 1987-07-31 1991-09-03 Leendert M. Bijnagte Apparatus and method for communicating textual and image information between a host computer and a remote display terminal
US4942538A (en) * 1988-01-05 1990-07-17 Spar Aerospace Limited Telerobotic tracker
US4891761A (en) * 1988-03-31 1990-01-02 Mets, Inc. Method for accurately updating positional information provided on a digital map
US5684514A (en) 1991-01-11 1997-11-04 Advanced Interaction, Inc. Apparatus and method for assembling content addressable video
US5247356A (en) * 1992-02-14 1993-09-21 Ciampa John A Method and apparatus for mapping and measuring land
US5414462A (en) * 1993-02-11 1995-05-09 Veatch; John W. Method and apparatus for generating a comprehensive survey map
US5487116A (en) 1993-05-25 1996-01-23 Matsushita Electric Industrial Co., Ltd. Vehicle recognition apparatus
US6037936A (en) * 1993-09-10 2000-03-14 Criticom Corp. Computer vision system with a graphic user interface and remote camera control
US5754850A (en) * 1994-05-11 1998-05-19 Realselect, Inc. Real-estate method and apparatus for searching for homes in a search pool for exact and close matches according to primary and non-primary selection criteria
DE69532126T2 (en) 1994-05-19 2004-07-22 Geospan Corp., Plymouth METHOD FOR COLLECTING AND PROCESSING VISUAL AND SPATIAL POSITION INFORMATION
US5682525A (en) * 1995-01-11 1997-10-28 Civix Corporation System and methods for remotely accessing a selected group of items of interest from a database
US6195122B1 (en) 1995-01-31 2001-02-27 Robert Vincent Spatial referenced photography
US5794216A (en) * 1995-07-14 1998-08-11 Brown; Timothy Robert Methods and system for data acquisition in a multimedia real estate database
US5852810A (en) * 1996-01-29 1998-12-22 Student Housing Network Geographic specific information search system and method
GB2337386B (en) * 1996-09-09 2001-04-04 Dennis J Dupray Location of a mobile station
US6144962A (en) 1996-10-15 2000-11-07 Mercury Interactive Corporation Visualization of web sites and hierarchical data structures
US6043837A (en) 1997-05-08 2000-03-28 Be Here Corporation Method and apparatus for electronically distributing images from a panoptic camera system
US6597818B2 (en) 1997-05-09 2003-07-22 Sarnoff Corporation Method and apparatus for performing geo-spatial registration of imagery
US6741790B1 (en) 1997-05-29 2004-05-25 Red Hen Systems, Inc. GPS video mapping system
US6025844A (en) * 1997-06-12 2000-02-15 Netscape Communications Corporation Method and system for creating dynamic link views
US5864337A (en) * 1997-07-22 1999-01-26 Microsoft Corporation Mehtod for automatically associating multimedia features with map views displayed by a computer-implemented atlas program
US6466701B1 (en) 1997-09-10 2002-10-15 Ricoh Company, Ltd. System and method for displaying an image indicating a positional relation between partially overlapping images
US6111583A (en) * 1997-09-29 2000-08-29 Skyline Software Systems Ltd. Apparatus and method for three-dimensional terrain rendering
FR2770317B1 (en) 1997-10-24 2000-12-08 Commissariat Energie Atomique METHOD FOR CALIBRATING THE ORIGINAL POSITION AND ORIENTATION OF ONE OR MORE MOBILE CAMERAS AND ITS APPLICATION TO MEASURING THE THREE-DIMENSIONAL POSITION OF FIXED OBJECTS
JP3788677B2 (en) 1997-11-14 2006-06-21 富士通株式会社 CONTINUOUS RECORDING SYSTEM AND METHOD, AND COMPUTER-READABLE RECORDING MEDIUM CONTAINING LONG TIME RECORDING PROGRAM
US6504571B1 (en) 1998-05-18 2003-01-07 International Business Machines Corporation System and methods for querying digital image archives using recorded parameters
US6323885B1 (en) * 1998-09-18 2001-11-27 Steven Paul Wiese Real estate value map computer system
US6266442B1 (en) * 1998-10-23 2001-07-24 Facet Technology Corp. Method and apparatus for identifying objects depicted in a videostream
JP3867883B2 (en) 1999-06-01 2007-01-17 株式会社リコー Image composition processing method, image composition processing apparatus, and recording medium
US6307573B1 (en) * 1999-07-22 2001-10-23 Barbara L. Barros Graphic-information flow method and system for visually analyzing patterns and relationships
JP3995846B2 (en) 1999-09-24 2007-10-24 本田技研工業株式会社 Object recognition device
US6266615B1 (en) 1999-09-27 2001-07-24 Televigation, Inc. Method and system for an interactive and real-time distributed navigation system
US6563529B1 (en) * 1999-10-08 2003-05-13 Jerry Jongerius Interactive system for displaying detailed view and direction in panoramic images
US6525731B1 (en) 1999-11-09 2003-02-25 Ibm Corporation Dynamic view-dependent texture mapping
US6678394B1 (en) 1999-11-30 2004-01-13 Cognex Technology And Investment Corporation Obstacle detection system
US6366892B1 (en) 2000-04-05 2002-04-02 Titan Film Capitol Llc. Method and system for automatically producing loan documents
US6421003B1 (en) * 2000-05-19 2002-07-16 Sri International Attitude determination using multiple baselines in a navigational positioning system
US7974930B2 (en) * 2000-07-26 2011-07-05 Pierce-Eislen, Inc. Method and system for providing real estate information
US6922701B1 (en) * 2000-08-03 2005-07-26 John A. Ananian Generating cad independent interactive physical description remodeling, building construction plan database profile
US6484101B1 (en) * 2000-08-16 2002-11-19 Imagelinks, Inc. 3-dimensional interactive image modeling system
US20020087389A1 (en) * 2000-08-28 2002-07-04 Michael Sklarz Value your home
US6985929B1 (en) * 2000-08-31 2006-01-10 The United States Of America As Represented By The Secretary Of The Navy Distributed object-oriented geospatial information distribution system and method thereof
US6895126B2 (en) 2000-10-06 2005-05-17 Enrico Di Bernardo System and method for creating, storing, and utilizing composite images of a geographic location
US6904160B2 (en) 2000-10-18 2005-06-07 Red Hen Systems, Inc. Method for matching geographic information with recorded images
US7174301B2 (en) 2000-10-23 2007-02-06 Costar Group, Inc. System and method for accessing geographic-based data
US7487114B2 (en) * 2000-10-23 2009-02-03 Costar Group, Inc. System and method for associating aerial images, map features, and information
US7184611B2 (en) * 2000-12-26 2007-02-27 Sony Corporation Data recording apparatus and method, data reproducing apparatus and method, data recording and reproducing apparatus and method, and map image data format
US7644131B2 (en) * 2000-12-27 2010-01-05 INOVO Limited Optimized image delivery over limited bandwidth communication channels
JP2002229890A (en) 2001-01-31 2002-08-16 Pioneer Electronic Corp Information displaying method
US20030206181A1 (en) 2001-04-13 2003-11-06 Abb Ab System and method for organizing two and three dimensional image data
AU2002305387B2 (en) 2001-05-04 2008-04-03 Legend Films, Llc Image sequence enhancement system and method
US7171389B2 (en) * 2001-06-01 2007-01-30 Landnet Corporation Identification, storage and display of land data on a website
US20060197763A1 (en) * 2002-02-11 2006-09-07 Landnet Corporation Document geospatial shape tagging, searching, archiving, and retrieval software
CA2352844C (en) * 2001-07-10 2009-11-17 Geojet Information Solutions Inc. 3-d map data visualization
US7096428B2 (en) * 2001-09-28 2006-08-22 Fuji Xerox Co., Ltd. Systems and methods for providing a spatially indexed panoramic video
US20030101063A1 (en) * 2001-09-28 2003-05-29 Sexton Ashley K. Real estate information system
US20030064705A1 (en) * 2001-10-02 2003-04-03 Desiderio Piero L. Communication of real estate information utilizing moving map GPS and wireless telecommunications to access MLS and other property specific information from remote databases
US7225207B1 (en) * 2001-10-10 2007-05-29 Google Inc. Server for geospatially organized flat file data
US6636803B1 (en) * 2001-11-30 2003-10-21 Corus Home Realty Real-estate information search and retrieval system
US7092957B2 (en) * 2002-01-18 2006-08-15 Boundary Solutions Incorporated Computerized national online parcel-level map data portal
US6759979B2 (en) * 2002-01-22 2004-07-06 E-Businesscontrols Corp. GPS-enhanced system and method for automatically capturing and co-registering virtual models of a site
JP3889650B2 (en) 2002-03-28 2007-03-07 三洋電機株式会社 Image processing method, image processing apparatus, computer program, and recording medium
WO2003087929A1 (en) 2002-04-10 2003-10-23 Pan-X Imaging, Inc. A digital imaging system
US20030210806A1 (en) 2002-05-07 2003-11-13 Hitachi, Ltd. Navigational information service with image capturing and sharing
US6978206B1 (en) * 2002-06-21 2005-12-20 Infogation Corporation Distributed navigation system
US7305365B1 (en) 2002-06-27 2007-12-04 Microsoft Corporation System and method for controlling access to location information
US8050521B2 (en) 2002-07-27 2011-11-01 Archaio, Llc System and method for simultaneously viewing, coordinating, manipulating and interpreting three-dimensional and two-dimensional digital images of structures for providing true scale measurements and permitting rapid emergency information distribution
US20040024709A1 (en) 2002-08-05 2004-02-05 Yu Paul D. System and method for determining the identity of a party associated with a transaction
JP2005535039A (en) * 2002-08-05 2005-11-17 メタカータ・インコーポレーテッド Interact with desktop clients with geographic text search systems
US7027054B1 (en) 2002-08-14 2006-04-11 Avaworks, Incorporated Do-it-yourself photo realistic talking head creation system and method
US6919892B1 (en) 2002-08-14 2005-07-19 Avaworks, Incorporated Photo realistic talking head creation system and method
US6996264B2 (en) 2002-10-18 2006-02-07 Leco Corporation Indentation hardness test system
US7424133B2 (en) * 2002-11-08 2008-09-09 Pictometry International Corporation Method and apparatus for capturing, geolocating and measuring oblique images
US7324666B2 (en) 2002-11-15 2008-01-29 Whitegold Solutions, Inc. Methods for assigning geocodes to street addressable entities
US20040143543A1 (en) 2003-01-17 2004-07-22 Goldman Robert P. Electronic real estate settlement
US20040196282A1 (en) 2003-02-14 2004-10-07 Oh Byong Mok Modeling and editing image panoramas
US7777740B2 (en) 2003-02-20 2010-08-17 Binary Simplex, Inc. Spatial decomposition methods using bit manipulation
US20040177373A1 (en) 2003-02-28 2004-09-09 Tsuyoshi Kawabe Graphic data distribution system and a data converter, and a graphic data distribution method
US7619626B2 (en) 2003-03-01 2009-11-17 The Boeing Company Mapping images from one or more sources into an image for display
US7254271B2 (en) * 2003-03-05 2007-08-07 Seadragon Software, Inc. Method for encoding and serving geospatial or other vector data as images
US20050128212A1 (en) 2003-03-06 2005-06-16 Edecker Ada M. System and method for minimizing the amount of data necessary to create a virtual three-dimensional environment
US8051089B2 (en) * 2003-04-30 2011-11-01 Hewlett-Packard Development Company, L.P. Systems and methods for location-based real estate service
US7475060B2 (en) * 2003-05-09 2009-01-06 Planeteye Company Ulc Browsing user interface for a geo-coded media database
US7986339B2 (en) * 2003-06-12 2011-07-26 Redflex Traffic Systems Pty Ltd Automated traffic violation monitoring and reporting system with combined video and still-image data
US7343232B2 (en) * 2003-06-20 2008-03-11 Geneva Aerospace Vehicle control system including related methods and components
JP4111885B2 (en) * 2003-07-23 2008-07-02 アルパイン株式会社 Map search and display method and apparatus
US7467356B2 (en) 2003-07-25 2008-12-16 Three-B International Limited Graphical user interface for 3d virtual display browser using virtual display windows
US7171058B2 (en) 2003-07-31 2007-01-30 Eastman Kodak Company Method and computer program product for producing an image of a desired aspect ratio
JP4447865B2 (en) * 2003-08-01 2010-04-07 ソニー株式会社 Map display system, map data processing device, map display device, and map display method
US7698212B1 (en) 2003-11-21 2010-04-13 Peirson Chris A Online settlement statement and funding control system and method
US8042056B2 (en) * 2004-03-16 2011-10-18 Leica Geosystems Ag Browsers for large geometric data visualization
US7155336B2 (en) * 2004-03-24 2006-12-26 A9.Com, Inc. System and method for automatically collecting images of objects at geographic locations and displaying same in online directories
US7659915B2 (en) 2004-04-02 2010-02-09 K-Nfb Reading Technology, Inc. Portable reading device with mode processing
WO2005120924A1 (en) 2004-06-11 2005-12-22 Stratech Systems Limited Method and system for rail track scanning and foreign object detection
US9213461B2 (en) * 2004-06-16 2015-12-15 Redfin Corporation Web-based real estate mapping system
US20050288959A1 (en) * 2004-06-16 2005-12-29 David Eraker Map-based search for real estate service providers
US20060041375A1 (en) * 2004-08-19 2006-02-23 Geographic Data Technology, Inc. Automated georeferencing of digitized map images
US20060059073A1 (en) 2004-09-15 2006-03-16 Walzak Rebecca B System and method for analyzing financial risk
US8195386B2 (en) 2004-09-28 2012-06-05 National University Corporation Kumamoto University Movable-body navigation information display method and movable-body navigation information display unit
US20060155594A1 (en) 2005-01-13 2006-07-13 Jess Almeida Adaptive step-by-step process with guided conversation logs for improving the quality of transaction data
US20060155570A1 (en) 2005-01-13 2006-07-13 Jess Almeida Aggregation and control of documents in the document repository using meta data and context information and creation of online info binder
US10198521B2 (en) * 2005-06-27 2019-02-05 Google Llc Processing ambiguous search requests in a geographic information system
US7702821B2 (en) 2005-09-15 2010-04-20 Eye-Fi, Inc. Content-aware digital media storage device and methods of using the same
US7747526B1 (en) 2006-03-27 2010-06-29 Fannie Mae System and method for transferring mortgage loan servicing rights
US7978970B2 (en) 2006-09-29 2011-07-12 Microscan Systems, Inc. Systems and/or devices for providing diffuse light
US7974917B2 (en) 2006-11-20 2011-07-05 Ibis Software Corporation Systems and methods related to lifetime mortgages
US7990394B2 (en) 2007-05-25 2011-08-02 Google Inc. Viewing and navigating within panoramic images, and applications thereof
US8046306B2 (en) * 2007-06-29 2011-10-25 Zaio Corporation System, method, and apparatus for property appraisals
US20100122208A1 (en) * 2007-08-07 2010-05-13 Adam Herr Panoramic Mapping Display
JP4990115B2 (en) 2007-12-06 2012-08-01 株式会社デンソー Position range setting device, control method and control device for moving object mounting device, and control method and control device for vehicle air conditioner
US20100004995A1 (en) 2008-07-07 2010-01-07 Google Inc. Claiming Real Estate in Panoramic or 3D Mapping Environments for Advertising
US8111307B2 (en) 2008-10-25 2012-02-07 Omnivision Technologies, Inc. Defective color and panchromatic CFA image

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120059720A1 (en) * 2004-06-30 2012-03-08 Musabji Adil M Method of Operating a Navigation System Using Images

Cited By (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11470303B1 (en) 2010-06-24 2022-10-11 Steven M. Hoffberg Two dimensional to three dimensional moving image converter
US20120330646A1 (en) * 2011-06-23 2012-12-27 International Business Machines Corporation Method For Enhanced Location Based And Context Sensitive Augmented Reality Translation
US9092674B2 (en) * 2011-06-23 2015-07-28 International Business Machines Corportion Method for enhanced location based and context sensitive augmented reality translation
US9245183B2 (en) 2014-06-26 2016-01-26 International Business Machines Corporation Geographical area condition determination
US9582719B2 (en) 2014-06-26 2017-02-28 International Business Machines Corporation Geographical area condition determination
CN104361586A (en) * 2014-10-31 2015-02-18 国网上海市电力公司 Electric pole and tower maintenance and pre-warning system based on panoramas
US10795010B2 (en) 2014-12-19 2020-10-06 Xidrone Systems, Inc. Systems and methods for detecting, tracking and identifying small unmanned systems such as drones
US10156631B2 (en) 2014-12-19 2018-12-18 Xidrone Systems, Inc. Deterrent for unmanned aerial systems
US10281570B2 (en) 2014-12-19 2019-05-07 Xidrone Systems, Inc. Systems and methods for detecting, tracking and identifying small unmanned systems such as drones
US10739451B1 (en) 2014-12-19 2020-08-11 Xidrone Systems, Inc. Systems and methods for detecting, tracking and identifying small unmanned systems such as drones
US10498955B2 (en) 2015-08-03 2019-12-03 Disney Enterprises, Inc. Commercial drone detection
US10198955B1 (en) * 2016-09-08 2019-02-05 Amazon Technologies, Inc. Drone marker and landing zone verification
US10388172B1 (en) * 2016-09-08 2019-08-20 Amazon Technologies, Inc. Obstacle awareness based guidance to clear landing space
US10049589B1 (en) 2016-09-08 2018-08-14 Amazon Technologies, Inc. Obstacle awareness based guidance to clear landing space
US10121117B1 (en) 2016-09-08 2018-11-06 Amazon Technologies, Inc. Drone location signature filters
US10922984B1 (en) 2016-09-08 2021-02-16 Amazon Technologies, Inc. Drone marker and landing zone verification
US11272090B2 (en) * 2017-06-02 2022-03-08 L-Tron Corporation Event data and location-linked spherical imaging system
US20180352137A1 (en) * 2017-06-02 2018-12-06 L-Tron Corporation Event data and location-linked spherical imaging system
US10907940B1 (en) 2017-12-12 2021-02-02 Xidrone Systems, Inc. Deterrent for unmanned aerial systems using data mining and/or machine learning for improved target detection and classification
CN110855906A (en) * 2018-07-24 2020-02-28 黑芝麻智能科技(上海)有限公司 Method for splicing optical images in 360-degree panoramic looking around by utilizing cameras and radars arranged around vehicle
TWI734037B (en) * 2018-09-28 2021-07-21 愛探極溫度行銷有限公司 Real estate holding and inheritance calculation management system

Also Published As

Publication number Publication date
US20130187949A1 (en) 2013-07-25
US20140211015A1 (en) 2014-07-31
US9384277B2 (en) 2016-07-05
US9311397B2 (en) 2016-04-12
USRE45264E1 (en) 2014-12-02
US20120293505A1 (en) 2012-11-22
US20130201340A1 (en) 2013-08-08
US20140019301A1 (en) 2014-01-16
US20130328858A9 (en) 2013-12-12
US8558848B2 (en) 2013-10-15
US20170094165A1 (en) 2017-03-30
US20140089811A1 (en) 2014-03-27
US8890866B2 (en) 2014-11-18
US9311396B2 (en) 2016-04-12
US20130191292A1 (en) 2013-07-25
US20130201341A1 (en) 2013-08-08
US20130188052A1 (en) 2013-07-25
US8207964B1 (en) 2012-06-26
US20130191359A1 (en) 2013-07-25
US8902226B2 (en) 2014-12-02
US20130191295A1 (en) 2013-07-25
US20130191725A1 (en) 2013-07-25
US20130191252A1 (en) 2013-07-25
US20130182108A1 (en) 2013-07-18

Similar Documents

Publication Publication Date Title
US9098870B2 (en) Internet-accessible real estate marketing street view system and method
US9311396B2 (en) Method of providing street view data of a real estate property
US7389181B2 (en) Apparatus and method for producing video drive-by data corresponding to a geographic location
AU2022203756B2 (en) Unmanned aircraft structure evaluation system and method
US11415986B2 (en) Geocoding data for an automated vehicle
US8078396B2 (en) Methods for and apparatus for generating a continuum of three dimensional image data
US8649610B2 (en) Methods and apparatus for auditing signage
US20080285886A1 (en) System For Displaying Images
US9529511B1 (en) System and method of generating a view for a point of interest
Ning et al. Exploring the vertical dimension of street view image based on deep learning: a case study on lowest floor elevation estimation
GB2424730A (en) Storage of road side images and retrieval through a map interface
Leberl et al. Automated photogrammetry for three-dimensional models of urban spaces
KR102467862B1 (en) A computer program for real estate management stored in a recording medium and a real estate management system using a computer program and an unmanned aerial vehicle
Leberl et al. 3D-models of the Human Habitat for the Internet.

Legal Events

Date Code Title Description
AS Assignment

Owner name: MVP PORTFOLIO LLC, FLORIDA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MV PATENTS LLC;REEL/FRAME:031733/0238

Effective date: 20130830

AS Assignment

Owner name: VISUAL REAL ESTATE, INC, FLORIDA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MVP PORTFOLIO, LLC;REEL/FRAME:032376/0428

Effective date: 20140307

AS Assignment

Owner name: SIBERLAW, LLP, NEW YORK

Free format text: LIEN;ASSIGNOR:VISUAL REAL ESTATE, INC.;REEL/FRAME:037155/0591

Effective date: 20151119

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION