CN105160684B - A kind of on-line automatic matching process for remotely sensing image geometric correction - Google Patents

A kind of on-line automatic matching process for remotely sensing image geometric correction Download PDF

Info

Publication number
CN105160684B
CN105160684B CN201510634043.XA CN201510634043A CN105160684B CN 105160684 B CN105160684 B CN 105160684B CN 201510634043 A CN201510634043 A CN 201510634043A CN 105160684 B CN105160684 B CN 105160684B
Authority
CN
China
Prior art keywords
image
matched
matching
image blocks
point
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201510634043.XA
Other languages
Chinese (zh)
Other versions
CN105160684A (en
Inventor
龙腾飞
焦伟利
何国金
王威
程博
张兆明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Remote Sensing and Digital Earth of CAS
Original Assignee
Institute of Remote Sensing and Digital Earth of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Remote Sensing and Digital Earth of CAS filed Critical Institute of Remote Sensing and Digital Earth of CAS
Priority to CN201510634043.XA priority Critical patent/CN105160684B/en
Publication of CN105160684A publication Critical patent/CN105160684A/en
Application granted granted Critical
Publication of CN105160684B publication Critical patent/CN105160684B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30181Earth observation

Landscapes

  • Image Processing (AREA)

Abstract

A kind of on-line automatic matching process for remotely sensing image geometric correction, is related to Remote Sensing Images Matching technical field.It is advantageous that making full use of the priori geological information of remote sensing image, by free or low cost network video map resource, equally distributed control point is obtained to high efficient and reliable.Using the initial imaging model of remote sensing image, image to be matched is divided into several regions by the number at the control point acquired as needed, and using a certain size image blocks as processing unit in each region;The approximate range for referring to image blocks is determined according to the range of image unit to be matched and initial imaging model, is then and resolution ratio similar in image unit to be matched by network downloading with reference to image blocks and by its resampling;SIFT matching, and excluding gross error point are carried out to image unit to be matched and with reference to image blocks, optimal matching points are finally selected from remaining match point and Least squares matching is carried out to it, correct its position coordinate.

Description

A kind of on-line automatic matching process for remotely sensing image geometric correction
Technical field
The present invention relates to a kind of practical remote sensing image automatic matching methods, can using network Aeronautics and Astronautics image Figure carries out on-line automatic matching to remote sensing image.It can be applied to the fields such as remote sensing, photogrammetric, mapping, image procossing.
Background technique
Image Auto-matching and geometric correction are the key links in Photogrammetry and Remote Sensing task, they are that image melts The basis of advanced application such as close, inlay, changing detection, map rejuvenation.Although in the past in decades to image Auto-matching Research is very more, but the Auto-matching of remote sensing image is still very challenging.Practical automatic matching method is answered This efficiency, robustness, in terms of the performance that has all had, but since remote sensing image data amount is big, scene is big, obtains item The features such as part is changeable, geometry deformation is complicated, existing method are difficult to take into account this several respect.In addition, the preparation with reference to image is also One difficult point of remote sensing image Auto-matching and geometric correction, especially high-resolution generally require very high with reference to the acquisition of image Cost.
In view of the limitation of existing automatic matching method, a kind of on-line automatic match party for remotely sensing image geometric correction Method has important practical value.
Summary of the invention
It is an object of the invention to solve the deficiencies in the prior art, propose a kind of for the fast of remotely sensing image geometric correction Speed, steady, accurately on-line automatic matching process, this method are capable of handling the remote sensing image of arbitrary size, utilize network video Map is automatically matched to a certain number of, well-distributed high-precision control point in a relatively short period of time, is directly used in remote sensing shadow The geometric accurate correction of picture.The advantage of this method essentially consists in the priori geological information for making full use of remote sensing image, by free or The network video map resource of low cost, obtains to high efficient and reliable equally distributed control point.
To solve the above problems, the present invention provides a kind of on-line automatic match parties for remotely sensing image geometric correction Method, the method comprising the steps of:
Remote sensing image to be matched is evenly dividing as several regions by the number at the control point S1. acquired as needed;
S2. if all imagery zones are processed, entire Image Matching process is completed;Otherwise, it starts to process next A imagery zone;
S3. current image region is divided into several image units by 256 pixels × 256 pixels size;
S4. if all image units in current image region are processed, mark the imagery zone that place is completed Reason, and it is transferred to step S2;Otherwise, next image unit is started to process;
S5: it is calculated according to the range of current image unit to be matched and initial imaging model and refers to shadow in network video map As the approximate range of block, then by network download it is corresponding with reference to image blocks and by its resampling be and image unit to be matched Similar resolution ratio;
S6: being matched using SIFT matching operator to image unit to be matched and with reference to image blocks, and excluding gross error point, If the step obtains 4 or more match points, step S7 is carried out, is otherwise transferred to step S4;
S7. optimal matching points are selected from the match point that step S6 is obtained and Least squares matching is carried out to it, are corrected This is added in matching result match point by the position coordinate of SIFT feature.
Wherein, step S5 further comprises:
S5.1 is according to the resolution ratio of image unit to be matched calculating network video map closest to level of zoom;
S5.2 calculates the corresponding width and height for referring to image blocks;
S5.3 calculates the latitude and longitude coordinates of the corresponding central point with reference to image blocks;
S5.4 sends static map service request and downloads corresponding with reference to image blocks;
S5.5 by the reference image blocks resampling of downloading be and resolution ratio similar in image unit to be matched.
Wherein, step S6 further comprises:
S6.1 carries out SIFT matching to image unit to be matched and with reference to image blocks;
S6.2 passes through dimensional constraints excluding gross error point;
S6.3 utilizes rotation angle restriction excluding gross error;
S6.4 constrains excluding gross error using RANSAC estimation similarity transformation;
S6.5 constrains excluding gross error point using affine transformation.
Detailed description of the invention
Fig. 1 is the on-line automatic matching process flow chart according to one embodiment of the present invention;
Fig. 2 is imagery zone and image unit schematic diagram in the method for the present invention;
Fig. 3 is according to acquisition network video map reference point shadow in the on-line automatic matching process of one embodiment of the present invention As block flow chart;
Fig. 4 be according in the on-line automatic matching process of one embodiment of the present invention to image unit to be matched and reference Image Block- matching flow chart;
Specific embodiment
On-line automatic matching process proposed by the present invention, is described with reference to the accompanying drawings as follows.
As shown in Figure 1, according to one embodiment of the present invention automatic matching method comprising steps of
Remote sensing image to be matched is evenly dividing as several regions by the number at the control point S1. acquired as needed, and Each imagery zone is labeled as untreated state, imagery zone schematic diagram is shown in Fig. 2;
S2. it if all imagery zones are processed, completes and terminates entire Image Matching process;Otherwise, beginning Manage next imagery zone;
S3. current image region is divided into several image units by 256 pixels × 256 pixels size, and will be each Image unit is denoted as untreated state, and image unit schematic diagram is shown in Fig. 2;
S4. if all image units in current image region are processed, mark the imagery zone that place is completed Reason, and it is transferred to step S2;Otherwise, next image unit is started to process;
S5. network video map is calculated according to the range of current image unit to be matched and initial imaging model (at present may be used Network video map include Google satellite image map, Bing Aerial image map, MapQuest satellite photomap and Mapbox satellite photomap) in reference to image blocks approximate range, then by network download it is corresponding refer to image blocks simultaneously It is and resolution ratio similar in image unit to be matched by its resampling;
S6. it is matched using SIFT matching operator to image unit to be matched and with reference to image blocks, and excluding gross error point, If the step obtains 4 or more match points, step S7 is carried out, is otherwise transferred to step S4;
S7. from the match point that step S6 is obtained, the maximum a pair of local contrast is chosen, and SIFT is matched Local geometric transformation model carries out Least squares matching to match point to this as initial value, and the point for SIFT feature of refining is sat This, is finally added in matching result match point, conditional equation formula (1) table of the Least squares matching to certain point by mark Show,
k1Is(xs, ys)+k2-Ir(xr, yr)=0 (1)
Wherein xs, ysFor image unit pixel coordinate to be matched, xr, yrTo refer to image blocks pixel coordinate,
xr=a0+a1xs+a2ys, yr=b0+b1xs+b2ys,
a0, a1, a2, b0, b1, b2For 6 geometric transformation parameters, k1, k2For 2 radiation transformation parameters,
Is(xs, ys) and Ir(xr, yr) be respectively image unit to be matched and with reference to image blocks gray value,
It treats in 11 pixels around match point × 11 pixels region each point and establishes error equation according to formula (1), then use Levenberg-Marquardt algorithm carries out optimization, optimal geometric transformation parameter is obtained, so that it is determined that after refining With a coordinate.
After image to be matched is divided into image unit, need according to the range of current image unit to be matched and initially at As model calculates the approximate range in network video map with reference to image blocks.Specifically, as shown in figure 3, step S5 is further wrapped It includes:
S5.1 is according to the resolution ratio of image unit to be matched calculating network video map closest to level of zoom, zoom-level Not according to the latitude and longitude coordinates and resolution ratio located where image unit to be matched using formula (2) calculating, at image unit place Longitude λ and latitude φ is calculated by the initial imaging model of image to be matched,
R in formulaearthRice is earth radius, and approximation is 6378137 meters,
GSD is the resolution ratio at image unit place,
N is level of zoom,
[] indicates to take the operation of closest integer;
S5.2 calculates the corresponding width width and height height for referring to image blocks: given level of zoom n, network video Map image coordinate x, y and longitude λ, the conversion method such as formula (3) of latitude φ and formula (4) are shown,
The latitude and longitude coordinates for being utilized respectively 4 vertex of image unit to be matched calculate corresponding 4 picture points by formula (3) Coordinate, then seeks the minimum circumscribed rectangle of this 4 points again, and the width and height of the rectangle are the width with reference to image blocks Width and height height;
S5.3 calculates the latitude and longitude coordinates λ of the corresponding central point with reference to image blocksrc, φrc, according to obtained in S5.2 most Small boundary rectangle calculates its center point image coordinate, then substitutes into formula (4) and calculates corresponding latitude and longitude coordinates λrc, φrc
S5.4 sends static map service request and downloads and corresponds to reference to image blocks, and static map service request is provided with unified The mode of source finger URL (URL) issues, and the static map service request format for commonly using network video map is as follows:
(1) Google satellite image map:
Https: //maps.googleapis.com/maps/api/staticmap? maptype=satellite& Zoom={ zoomLevel } &ce nter={ lat }, { lon } &size={ width } x { height } &key= {googleKey}
(2) Bing Aerial image map:
Http:// dev.virtualearth.net/RES T/v1/Imagery/Map/Aerial/ { lat }, { lon }/ { zoomLevel }? mapSize={ width }, { height } &key={ BingMapsKey }
(3) MapQuest satellite photomap:
Http:// www.mapquestapi.com/staticmap/v4/getmap? type=sat&zoom= { zoomLevel } &center={ lat }, { lon } &size={ width }, { height } &key={ mapquestKey }
(4) Mapbox satellite photomap:
Http:// api.tiles.mapbox.com/v4/mapbox.satellite/ { lon }, { lat }, { zoomLevel }/{ width } x { height } .p ng? access_token={ mapboxKey }
In above each URL, the content in " { } " is filled according to the calculated result of step S5.1, S5.2 and S5.3, specifically,
ZoomLevel is to be calculated in step S5.1 closest to level of zoom n,
Width, height are the width width and height height for the reference image blocks being calculated in step S5.2,
Lon, lat are the central point longitude λ for the reference image blocks being calculated in step S5.3rcWith latitude φrc,
GoogleKey, BingMapsKey, mapquestKey, mapboxKey are respectively the api of each Map Service of Network Key can be applied obtaining on corresponding website;
S5.5 by the reference image blocks resampling of downloading be and resolution ratio similar in image unit to be matched.
Downloading obtains after referring to image blocks accordingly, needs using SIFT matching operator to image unit to be matched and reference Image blocks are matched, and excluding gross error point.Specifically, as shown in figure 4, step S6 further comprises:
S6.1 from image unit to be matched and with reference to extraction SIFT feature in image blocks and calculates corresponding SIFT respectively Description vectors, using the Euclidean distance of SIFT description vectors as the distance metric of two SIFT features, to image list to be matched Each SIFT feature in member judges following two condition, meet alternative one or simultaneously meet 2 conditions i.e. as With candidate point:
(1) this into reference image blocks the ratio of each SIFT feature minimum range and time small distance less than 0.75;
(2) with reference in image blocks with the point apart from the smallest characteristic point other SIFT features into image unit to be matched Distance be not less than the distance of the point;
Its corresponding match point be with reference in image blocks with it apart from nearest SIFT feature;
S6.2 checks the SIFT scale coordinate ratio T of each candidate point and its match pointσIf 0.8≤Tσ≤ 1.25, then it protects It stays this to candidate point, otherwise rejects this to candidate point;
S6.3 calculates the SIFT principal direction differential seat angle of each candidate point and its match point, establishes intensity histogram with 10 degree for interval Figure, using the angle where histogram peak as image unit to be matched and with reference to the rotation angle θ between image blockspeak, then Check the SIFT principal direction differential seat angle Δ θ of each candidate point and its match point, if | Δ θ-θpeak|≤15 °, then retain this to time Otherwise reconnaissance rejects this to candidate point;
S6.4 estimates the similarity transformation relationship (5) between remaining candidate matches point using RANSAC algorithm, while rejecting not Meet the rough error point of the variation relation,
Wherein xs, ysFor image unit pixel coordinate to be matched, xr, yrTo refer to image blocks pixel coordinate, s and θ are similar The scale and rotation angle parameter of transformation, tx, tyRespectively translation parameters of the similarity transformation in the direction x and the direction y;
S6.5 calculates affine Transform Model to candidate point remaining in step S6.4,
Wherein xs, ysFor image unit pixel coordinate to be matched, xr, yrTo refer to image blocks pixel coordinate, a0, a1, a2, b0, b1, b2For 6 parameters of affine transformation;
Each pair of matching candidate point is checked using obtained affine Transform Model, if maximum residul difference is greater than 1 pixel, is picked Except this to match point after, recalculate affine Transform Model, and residual test is carried out to remaining candidate point;If maximum residul difference is not Greater than 1 pixel, then export remaining matching double points, end step S6.

Claims (1)

1. a kind of on-line automatic matching process for remotely sensing image geometric correction, it is characterized in that: include the following steps,
Remote sensing image to be matched is evenly dividing as several regions by the number at the control point that step S1. is acquired as needed;
If all imagery zones of step S2. are processed, entire Image Matching process is completed;Otherwise, it starts to process next A imagery zone;
Current image region is divided into several image units by 256 pixels × 256 pixels size by step S3.;
If all image units in step S4. current image region are processed, mark the imagery zone that place is completed Reason, and it is transferred to step S2;Otherwise, next image unit is started to process;
Step S5. calculates free or low cost network according to the range and initial imaging model of current image unit to be matched The approximate range of image blocks is referred in photomap, and corresponding reference image blocks are then downloaded by network and is by its resampling With resolution ratio similar in image unit to be matched;
Step S6. is matched using SIFT matching operator to image unit to be matched and with reference to image blocks, and excluding gross error Point carries out step S7 if the step obtains 4 or more match points, is otherwise transferred to step S4;
Step S7. selects optimal matching points from the match point that step S6 is obtained and carries out Least squares matching to it, amendment This is added in matching result match point by the position coordinate of SIFT feature;
Step S5 further comprises: S5.1 is according to the resolution ratio of image unit to be matched calculating network video map closest to contracting Put rank, level of zoom utilizes formula (2) calculating, image according to the latitude and longitude coordinates and resolution ratio located where image unit to be matched Longitude λ and latitude Φ at where unit are calculated by the initial imaging model of image to be matched,
R in formulaearthRice is earth radius, and approximation is 6378137 meters, and GSD is the resolution ratio at image unit place, and n is Level of zoom, [] indicate to take the operation of closest integer;
Step S5.2 calculates the corresponding width width and height height for referring to image blocks: given level of zoom n, network video Map image coordinate x, y and longitude λ, the conversion method such as formula (3) of latitude φ and formula (4) are shown,
4 picture point coordinates on 4 vertex are calculated by formula (3) using the latitude and longitude coordinates on 4 vertex of image unit to be matched, Then the minimum circumscribed rectangle of 4 picture points is sought again, and the width and height of the minimum circumscribed rectangle are the reference The width width and the height height of image blocks;
Step S5.3 calculates the longitude coordinate λ of the corresponding central point with reference to image blocksrcWith latitude coordinate φrc, obtained according in S5.2 To the minimum circumscribed rectangle calculate the central point image coordinate with reference to image blocks, then substitute into formula (4) calculate correspond to Longitude coordinate λrcWith latitude coordinate φrc
Step S5.4 sends static map service request and downloads and corresponds to reference to image blocks, and static map service request is provided with unified The mode of source finger URL (URL) issues;
Step S5.5 by the reference image blocks resampling of downloading be and resolution ratio similar in image unit to be matched.
CN201510634043.XA 2015-09-30 2015-09-30 A kind of on-line automatic matching process for remotely sensing image geometric correction Expired - Fee Related CN105160684B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510634043.XA CN105160684B (en) 2015-09-30 2015-09-30 A kind of on-line automatic matching process for remotely sensing image geometric correction

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510634043.XA CN105160684B (en) 2015-09-30 2015-09-30 A kind of on-line automatic matching process for remotely sensing image geometric correction

Publications (2)

Publication Number Publication Date
CN105160684A CN105160684A (en) 2015-12-16
CN105160684B true CN105160684B (en) 2019-01-18

Family

ID=54801526

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510634043.XA Expired - Fee Related CN105160684B (en) 2015-09-30 2015-09-30 A kind of on-line automatic matching process for remotely sensing image geometric correction

Country Status (1)

Country Link
CN (1) CN105160684B (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106408028B (en) * 2016-09-26 2020-04-14 珠海市测绘院 Urban and rural planning inspection surveying and mapping data processing method
CN106595598B (en) * 2016-12-21 2019-03-19 上海航天控制技术研究所 A kind of first optical remote sensing imaging method in permanent ground of wide visual field
CN108428220B (en) * 2018-03-05 2020-12-01 武汉大学 Automatic geometric correction method for ocean island reef area of remote sensing image of geostationary orbit satellite sequence
CN108492711A (en) * 2018-04-08 2018-09-04 黑龙江工业学院 A kind of drawing electronic map method and device
CN109242894B (en) * 2018-08-06 2021-04-09 广州视源电子科技股份有限公司 Image alignment method and system based on mobile least square method
CN109508674B (en) * 2018-11-13 2021-08-13 佳木斯大学 Airborne downward-looking heterogeneous image matching method based on region division
CN114004770B (en) * 2022-01-04 2022-04-26 成都国星宇航科技有限公司 Method and device for accurately correcting satellite space-time diagram and storage medium
CN116521927B (en) * 2023-06-30 2024-02-13 成都智遥云图信息技术有限公司 Remote sensing image matching method and system based on network map tiles

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103337052A (en) * 2013-04-17 2013-10-02 国家测绘地理信息局卫星测绘应用中心 Automatic geometric correction method for wide remote-sensing images

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2879791B1 (en) * 2004-12-16 2007-03-16 Cnes Epic METHOD FOR PROCESSING IMAGES USING AUTOMATIC GEOREFERENCING OF IMAGES FROM A COUPLE OF IMAGES TAKEN IN THE SAME FOCAL PLAN
CN102663680B (en) * 2012-03-06 2017-04-05 中国科学院对地观测与数字地球科学中心 Geometric image correction method based on region feature
CN103218783B (en) * 2013-04-17 2016-05-25 国家测绘地理信息局卫星测绘应用中心 Satellite remote sensing images fast geometric correcting method based on control point image database

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103337052A (en) * 2013-04-17 2013-10-02 国家测绘地理信息局卫星测绘应用中心 Automatic geometric correction method for wide remote-sensing images

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
An Automated Method for Geo-Referencing Satellite Images Using Google Tile Scheme;Khalil Al-Joburi;《World Applied Sciences Journal》;20141231;第32卷(第2期);第267-277页 *

Also Published As

Publication number Publication date
CN105160684A (en) 2015-12-16

Similar Documents

Publication Publication Date Title
CN105160684B (en) A kind of on-line automatic matching process for remotely sensing image geometric correction
CN107316325B (en) Airborne laser point cloud and image registration fusion method based on image registration
CN105761242B (en) Blind person walking positioning method based on computer binocular vision and inertial measurement
CN109029444A (en) One kind is based on images match and sterically defined indoor navigation system and air navigation aid
CN109523585B (en) Multisource remote sensing image feature matching method based on direction phase consistency
CN106709944B (en) Satellite remote sensing image registration method
KR20190026452A (en) A method of automatic geometric correction of digital elevation model made from satellite images and provided rpc
CN112750203B (en) Model reconstruction method, device, equipment and storage medium
CN112419374A (en) Unmanned aerial vehicle positioning method based on image registration
CN110660098A (en) Positioning method and device based on monocular vision
CN112200203A (en) Matching method of weak correlation speckle images in oblique field of view
Bruno et al. Accuracy assessment of 3d models generated from google street view imagery
Kim et al. Investigating applicability of unmanned aerial vehicle to the tidal flat zone
CN111693025A (en) Remote sensing image data generation method, system and equipment
CN113984039A (en) Method, device and system for correcting motion trail and storage medium
Zhang et al. Automatic fusion of hyperspectral images and laser scans using feature points
Duan et al. A combined image matching method for Chinese optical satellite imagery
KR101711575B1 (en) Method for Detecting of the Unified Control Points for RPC Adjustment of Satellite Image
Feng et al. Mountainous remote sensing images registration based on improved optical flow estimation
Tatar et al. Quasi-epipolar resampling of high resolution satellite stereo imagery for semi global matching
US20210110195A1 (en) Fast determinant of hessian filtering for image tiepoint candidate area assessment
CN113932793A (en) Three-dimensional coordinate positioning method and device, electronic equipment and storage medium
CN109696656A (en) Localization method and its system based on phase focusing
Alsubaie et al. The feasibility of 3D point cloud generation from smartphones
KR101775124B1 (en) System and method for automatic satellite image processing for improvement of location accuracy

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20190118

Termination date: 20190930