CN110942002A - Unmanned aerial vehicle aerial photography video frame positioning method based on rotation invariant perceptual hashing - Google Patents

Unmanned aerial vehicle aerial photography video frame positioning method based on rotation invariant perceptual hashing Download PDF

Info

Publication number
CN110942002A
CN110942002A CN201911129923.6A CN201911129923A CN110942002A CN 110942002 A CN110942002 A CN 110942002A CN 201911129923 A CN201911129923 A CN 201911129923A CN 110942002 A CN110942002 A CN 110942002A
Authority
CN
China
Prior art keywords
unmanned aerial
aerial vehicle
rotation
sequence
video frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911129923.6A
Other languages
Chinese (zh)
Other versions
CN110942002B (en
Inventor
印鉴
陈智聪
陈殷齐
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
National Sun Yat Sen University
Original Assignee
National Sun Yat Sen University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by National Sun Yat Sen University filed Critical National Sun Yat Sen University
Priority to CN201911129923.6A priority Critical patent/CN110942002B/en
Publication of CN110942002A publication Critical patent/CN110942002A/en
Application granted granted Critical
Publication of CN110942002B publication Critical patent/CN110942002B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/13Satellite images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence

Abstract

The invention provides an unmanned aerial vehicle aerial video frame positioning method based on rotation invariant perceptual hashing, which has certain translation and deformation (posing principle); differential encoding is used during encoding, so that illumination weather conditions are irrelevant; and the starting point-independent circular coding sequence is added to make the rotation invariance; the corner parts are not coded in the circular coding sequence because the corner parts have large difference due to different visual angles, thereby eliminating the different influences of other scenes of the corners in the rotation process. Experiments on real unmanned aerial vehicle videos show that compared with the prior perceptual hash method, the method successfully solves the problem that the traditional perceptual hash has no rotation invariant characteristic.

Description

Unmanned aerial vehicle aerial photography video frame positioning method based on rotation invariant perceptual hashing
Technical Field
The invention relates to the field of video image processing algorithms, in particular to an unmanned aerial vehicle aerial photography video frame positioning method based on rotation invariant perceptual hashing.
Background
In recent years, with the popularization of unmanned aerial vehicles, the use of unmanned aerial vehicle aerial videos to complete some tasks has received more and more attention from the industry. However, in the task of the unmanned aerial vehicle river channel abnormal cruising, the unmanned aerial vehicle videos in different time periods of the same river channel need to be positioned frame by frame, that is, a frame of one video at the same position on another video is found. Such a task has not been studied in a relevant place.
The related technology is based on deep learning picture retrieval, but is not suitable for application of unmanned aerial vehicle video frame positioning, because the deep learning picture retrieval aims at finding semantic matching between pictures, while semantic information of the whole video is basically the same and the deep learning is too time-consuming, and a short video has huge frame number. So we adopt Perceptual Hashing (Perceptual Hashing) as a simple and fast similar picture searching technique. Perceptual hashing is a type of one-way mapping from a multimedia data set to a perceptual summary set, i.e., a multimedia digital representation with the same perceptual content is uniquely mapped into a segment of digital summary, and perceptual robustness and security are satisfied. The perceptual hash provides safe and reliable technical support for information service modes such as multimedia content identification, retrieval, authentication and the like.
The Hash Functions are irreversible digital digests (digests) for extracting original data, have the characteristics of unidirectionality, vulnerability and the like, and can ensure the uniqueness and the non-tampering property of the original data. Various hash functions have been successfully applied in the fields of information retrieval and management, data authentication, and the like. However, the conventional hash function cannot satisfy the frame positioning function of two unmanned aerial vehicle aerial videos at different times, because even if videos shot at different times in the same place have some lens rotation differences, the hash mapping is incorrect due to the slight differences (the hash values are different by rotating the same picture).
Disclosure of Invention
The invention provides an unmanned aerial vehicle aerial video frame positioning method based on rotation invariant perceptual hashing, which eliminates different influences of other scenes of corners during rotation.
In order to achieve the technical effects, the technical scheme of the invention is as follows:
an unmanned aerial vehicle aerial photography video frame positioning method based on rotation invariant perceptual hashing comprises the following steps:
s1: calculating hash values of the video frame to be positioned, the reference frame and the video frame to be retrieved, the target frame, which are not changed in rotation;
s2: and comparing the reference frame and the target frame with all frames of the unmanned aerial vehicle video at another time to find a frame with the minimum difference in comparison with the hash value.
Further, the specific process of step S1 is:
s11: carrying out picture scaling on the reference frame and the target frame;
s12: carrying out gray-scale image conversion on the zoomed image;
s13: acquiring a circumferential coding sequence of the hash value from the gray level image;
s14: obtaining a binary form of the hash value of the picture processed in the step S13 so as to obtain a hash value with unchanged rotation;
s15: an information fingerprint is derived from the result of S14.
Further, the process of step S11 is:
the pictures are uniformly scaled to 8 × 8 pictures with 64 pixels, so that the operation has certain translation invariance, and when the translation amount is in a scaling range, the thumbnails are basically consistent.
Further, the process of step S12 is: and converting the scaled picture into a 256-step gray-scale image.
Further, the process of step S13 is:
encoding hash according to a circle mode to make the hash have rotation invariance, taking the length r (1-H/2) as the radius of the circle according to the length H of the picture, and taking the length of each radius as the mode x of polar coordinates (rcos theta + x)0,y=rsinθ+y0Calculating a coordinate sequence of each circle, wherein theta is calculated according to the ratio of 0-360°Each value of (x)0,y0) To scale the center of the picture, a circular sequence (x) is obtained when r is 11 1,y1 1)(x1 2,y1 2)(x1 3,y1 3) ... wherein xr iUnder the circumference of a circle with radius rThe ith non-repeating coordinate index calculated according to the degrees of 0 degrees, 1 degrees and 2 degrees to 360 degrees can be obtained1=I(x1 1,y1 1),I(x1 2,y1 2),I(x1 3,y1 3) ... a sequence of pixel values indexed by a circular coordinate, r such sequences can be obtained as S (the r-th sequence is S)r) Such a coding sequence also automatically removes the corner effect, because two frames with rotation differences in the actual captured video will be completely different in the corner portion.
Further, the process of obtaining the binary form of the hash value in step S14 is:
for each S in a differential mannerrCoding the sequence, traversing each pixel of the gray picture, recording the previous value as 1 when the previous value is larger than the current value, otherwise, recording the previous value as 0, comparing the last pixel value with a pixel, connecting the last pixel value and the pixel end to form a circumference, and finally obtaining r binary sequences as B, wherein the r-th sequence is Br
Further, the process of acquiring the rotation-invariant hash value in step S14 is:
handle BrPerforming cyclic shift, taking the sequence with the minimum binary sequence value after the cyclic shift as the final sequence of the circumference, and finally obtaining r binary sequences which are irrelevant to the starting point and are marked as Z, wherein the r sequence is Zr
Further, the process of step S15 is:
in the order of r from 1 to H/2, ZrCombined to form the final information fingerprint X ═ Z1Z2Z3...ZH/2
Further, the process of step S2 is:
and calculating the Hamming distance of the fingerprints of the reference frame and the target frame which need to be compared, wherein the larger the Hamming distance is, the more inconsistent the images are, otherwise, the smaller the Hamming distance is, the more similar the images are, when the distance is 0, the images are completely the same, all the video frames on the target frame are compared once, the first 20 frames with the minimum Hamming distance are found, and the frames with the highest similarity are the required frames when further SSIM comparison is carried out.
Compared with the prior art, the technical scheme of the invention has the beneficial effects that:
the rotation invariant perceptual hash of the invention has certain translation and no deformation by transforming the image into the thumbnail (posing principle); differential encoding is used during encoding, so that illumination weather conditions are irrelevant; and the starting point-independent circular coding sequence is added to make the rotation invariance; the corner parts are not coded in the circular coding sequence because the corner parts have large difference due to different visual angles, thereby eliminating the different influences of other scenes of the corners in the rotation process. Experiments on real unmanned aerial vehicle videos show that compared with the prior perceptual hash method, the method successfully solves the problem that the traditional perceptual hash has no rotation invariant characteristic.
Drawings
FIG. 1 is a schematic diagram of the algorithm rotation invariant coding scheme of the present invention;
FIG. 2 is a schematic diagram illustrating the computation of a rotation invariant perceptual hash according to the present invention;
FIG. 3 is a schematic view of a video frame alignment application of the present invention;
FIG. 4 is a classic Lena diagram;
FIG. 5 is a comparison graph of frame alignment application results for live video.
Detailed Description
The drawings are for illustrative purposes only and are not to be construed as limiting the patent;
for the purpose of better illustrating the embodiments, certain features of the drawings may be omitted, enlarged or reduced, and do not represent the size of an actual product;
it will be understood by those skilled in the art that certain well-known structures in the drawings and descriptions thereof may be omitted.
The technical solution of the present invention is further described below with reference to the accompanying drawings and examples.
As shown in fig. 1-2, an unmanned aerial vehicle aerial video frame positioning method based on rotation invariant perceptual hashing includes the following steps:
s1: calculating hash values of the video frame to be positioned, the reference frame and the video frame to be retrieved, the target frame, which are not changed in rotation;
s2: and comparing the reference frame and the target frame with all frames of the unmanned aerial vehicle video at another time to find a frame with the minimum difference in comparison with the hash value.
The specific process of step S1 is:
s11: carrying out picture scaling on the reference frame and the target frame;
s12: carrying out gray-scale image conversion on the zoomed image;
s13: acquiring a circumferential coding sequence of the hash value from the gray level image;
s14: obtaining a binary form of the hash value of the picture processed in the step S13 so as to obtain a hash value with unchanged rotation;
s15: an information fingerprint is derived from the result of S14.
The process of step S11 is:
the pictures are uniformly scaled to 8 × 8 pictures with 64 pixels, so that the operation has certain translation invariance, and when the translation amount is in a scaling range, the thumbnails are basically consistent.
The process of step S12 is: and converting the scaled picture into a 256-step gray-scale image.
As shown in fig. 3, the process of step S13 is:
encoding hash according to a circle mode to make the hash have rotation invariance, taking the length r (1-H/2) as the radius of the circle according to the length H of the picture, and taking the length of each radius as the mode x of polar coordinates (rcos theta + x)0,y=rsinθ+y0And calculating a coordinate sequence of each circle, wherein theta is a value of 0-360 degrees, (x)0,y0) To scale the center of the picture, a circular sequence (x) is obtained when r is 11 1,y1 1)(x1 2,y1 2)(x1 3,y1 3) ... wherein xr iRepresenting 0, 1, 2 DEG, 3 under a circle of radius rThe ith non-repeating coordinate index calculated in degrees of 60 deg., then S is obtained1=I(x1 1,y1 1),I(x1 2,y1 2),I(x1 3,y1 3) ... a sequence of pixel values indexed by a circular coordinate, r such sequences can be obtained as S (the r-th sequence is S)r) Such a coding sequence also automatically removes the corner effect, because two frames with rotation differences in the actual captured video will be completely different in the corner portion.
The process of obtaining the binary form of the hash value in step S14 is:
for each S in a differential mannerrCoding the sequence, traversing each pixel of the gray picture, recording the previous value as 1 when the previous value is larger than the current value, otherwise, recording the previous value as 0, comparing the last pixel value with a pixel, connecting the last pixel value and the pixel end to form a circumference, and finally obtaining r binary sequences as B, wherein the r-th sequence is Br
The process of acquiring the rotation-invariant hash value in step S14 is:
handle BrPerforming cyclic shift, taking the sequence with the minimum binary sequence value after the cyclic shift as the final sequence of the circumference, and finally obtaining r binary sequences which are irrelevant to the starting point and are marked as Z, wherein the r sequence is Zr
The process of step S15 is:
in the order of r from 1 to H/2, ZrCombined to form the final information fingerprint X ═ Z1Z2Z3...ZH/2
The process of step S2 is:
and calculating the Hamming distance of the fingerprints of the reference frame and the target frame which need to be compared, wherein the larger the Hamming distance is, the more inconsistent the images are, otherwise, the smaller the Hamming distance is, the more similar the images are, when the distance is 0, the images are completely the same, all the video frames on the target frame are compared once, the first 20 frames with the minimum Hamming distance are found, and the frames with the highest similarity are the required frames when further SSIM comparison is carried out.
The classical Lena graph and the image after the Lena graph is rotated (as shown in fig. 4) are selected experimentally and compared with the traditional hash method. The comparative results are shown in Table 1.
TABLE 1Lena graph after rotation, comparison with the traditional Hash method
Figure BDA0002278008940000051
Experiments show that our rotation-invariant hash is more robust than the hash in the conventional method in Lena's graph, regardless of the rotation.
Adopt actual unmanned aerial vehicle video of taking photo by plane. A video frame is randomly selected, and a traditional perceptual hash method and the like are used in the video of another time period. The whole process is the same as in fig. 3. Because there is no ground true, we do a subjective comparison on the finally searched frame. FIG. 5 shows a comparison of the frame alignment application results for live video.
From the results, it can be seen that our invention has a great improvement compared to the previous hash method, so that the frame with rotation can be accurately positioned. In fig. 5, since the same view angle frame does not exist in the whole video, the result of the hash search is very poor. But is found exactly in the method of the present application.
The same or similar reference numerals correspond to the same or similar parts;
the positional relationships depicted in the drawings are for illustrative purposes only and are not to be construed as limiting the present patent;
it should be understood that the above-described embodiments of the present invention are merely examples for clearly illustrating the present invention, and are not intended to limit the embodiments of the present invention. Other variations and modifications will be apparent to persons skilled in the art in light of the above description. And are neither required nor exhaustive of all embodiments. Any modification, equivalent replacement, and improvement made within the spirit and principle of the present invention should be included in the protection scope of the claims of the present invention.

Claims (9)

1. An unmanned aerial vehicle aerial photography video frame positioning method based on rotation invariant perceptual hashing is characterized by comprising the following steps:
s1: calculating hash values of the video frame to be positioned, the reference frame and the video frame to be retrieved, the target frame, which are not changed in rotation;
s2: and comparing the reference frame and the target frame with all frames of the unmanned aerial vehicle video at another time to find a frame with the minimum difference in comparison with the hash value.
2. The unmanned aerial vehicle aerial photography video frame positioning method based on rotation invariant perceptual hashing as claimed in claim 1, wherein the specific process of the step S1 is:
s11: carrying out picture scaling on the reference frame and the target frame;
s12: carrying out gray-scale image conversion on the zoomed image;
s13: acquiring a circumferential coding sequence of the hash value from the gray level image;
s14: obtaining a binary form of the hash value of the picture processed in the step S13 so as to obtain a hash value with unchanged rotation;
s15: an information fingerprint is derived from the result of S14.
3. The unmanned aerial vehicle aerial video frame positioning method based on rotation invariant perceptual hashing as claimed in claim 2, wherein the process of the step S11 is:
the pictures are uniformly scaled to 8 × 8 pictures with 64 pixels, so that the operation has certain translation invariance, and when the translation amount is in a scaling range, the thumbnails are basically consistent.
4. The unmanned aerial vehicle aerial video frame positioning method based on rotation invariant perceptual hashing as claimed in claim 3, wherein the process of the step S12 is: and converting the scaled picture into a 256-step gray-scale image.
5. The unmanned aerial vehicle aerial video frame positioning method based on rotation invariant perceptual hashing as claimed in claim 4, wherein the process of the step S13 is:
encoding hash according to a circle mode to make the hash have rotation invariance, taking the length r (1-H/2) as the radius of the circle according to the length H of the picture, and taking the length of each radius as the mode x of polar coordinates (rcos theta + x)0,y=rsinθ+y0And calculating a coordinate sequence of each circle, wherein theta is a value of 0-360 degrees, (x)0,y0) To scale the center of the picture, a circular sequence (x) is obtained when r is 11 1,y1 1)(x1 2,y1 2)(x1 3,y1 3) ... wherein xr iRepresenting the ith non-repeating coordinate index calculated according to the degrees of 0 DEG, 1 DEG, 2 DEG-360 DEG under the circumference with the radius r, then S can be obtained1=I(x1 1,y1 1),I(x1 2,y1 2),I(x1 3,y1 3) ... a sequence of pixel values indexed by a circular coordinate, r such sequences can be obtained as S (the r-th sequence is S)r) Such a coding sequence also automatically removes the corner effect, because two frames with rotation differences in the actual captured video will be completely different in the corner portion.
6. The unmanned aerial vehicle aerial video frame positioning method based on rotation invariant perceptual hashing as claimed in claim 5, wherein the process of obtaining the binary form of the hash value in the step S14 is:
for each S in a differential mannerrCoding the sequence, traversing each pixel of the gray picture, recording the previous value as 1 if the previous value is larger than the current value, otherwise, recording the previous value as 0, comparing the last pixel value with a pixel, connecting the last pixel value and the pixel end to form a circle, and finally obtaining r binary sequences as B, wherein the r-th sequenceIs Br
7. The unmanned aerial vehicle aerial photography video frame positioning method based on rotation invariant perceptual hashing as claimed in claim 6, wherein the process of obtaining the rotation invariant hash value in step S14 is:
handle BrPerforming cyclic shift, taking the sequence with the minimum binary sequence value after the cyclic shift as the final sequence of the circumference, and finally obtaining r binary sequences which are irrelevant to the starting point and are marked as Z, wherein the r sequence is Zr
8. The unmanned aerial vehicle aerial video frame positioning method based on rotation invariant perceptual hashing as claimed in claim 7, wherein the process of the step S15 is:
in the order of r from 1 to H/2, ZrCombined to form the final information fingerprint X ═ Z1Z2Z3...ZH/2
9. The unmanned aerial vehicle aerial video frame positioning method based on rotation invariant perceptual hashing as claimed in claim 8, wherein the process of the step S2 is:
and calculating the Hamming distance of the fingerprints of the reference frame and the target frame which need to be compared, wherein the larger the Hamming distance is, the more inconsistent the images are, otherwise, the smaller the Hamming distance is, the more similar the images are, when the distance is 0, the images are completely the same, all the video frames on the target frame are compared once, the first 20 frames with the minimum Hamming distance are found, and the frames with the highest similarity are the required frames when further SSIM comparison is carried out.
CN201911129923.6A 2019-11-18 2019-11-18 Unmanned aerial vehicle aerial video frame positioning method based on rotation invariant perception hash Active CN110942002B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911129923.6A CN110942002B (en) 2019-11-18 2019-11-18 Unmanned aerial vehicle aerial video frame positioning method based on rotation invariant perception hash

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911129923.6A CN110942002B (en) 2019-11-18 2019-11-18 Unmanned aerial vehicle aerial video frame positioning method based on rotation invariant perception hash

Publications (2)

Publication Number Publication Date
CN110942002A true CN110942002A (en) 2020-03-31
CN110942002B CN110942002B (en) 2023-11-07

Family

ID=69907730

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911129923.6A Active CN110942002B (en) 2019-11-18 2019-11-18 Unmanned aerial vehicle aerial video frame positioning method based on rotation invariant perception hash

Country Status (1)

Country Link
CN (1) CN110942002B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112434185A (en) * 2020-10-26 2021-03-02 国家广播电视总局广播电视规划院 Method, system, server and storage medium for searching similar video clips
CN112698661A (en) * 2021-03-22 2021-04-23 成都睿铂科技有限责任公司 Aerial survey data acquisition method, device and system for aircraft and storage medium
CN113704532A (en) * 2020-11-25 2021-11-26 天翼智慧家庭科技有限公司 Method and system for improving recall rate of picture retrieval

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105069042A (en) * 2015-07-23 2015-11-18 北京航空航天大学 Content-based data retrieval methods for unmanned aerial vehicle spying images
CN106126585A (en) * 2016-06-20 2016-11-16 北京航空航天大学 Unmanned plane image search method based on quality grading with the combination of perception Hash feature
CN107040790A (en) * 2017-04-01 2017-08-11 华南理工大学 A kind of video content certification and tampering location method based on many granularity Hash
CN109918537A (en) * 2019-01-18 2019-06-21 杭州电子科技大学 A kind of method for quickly retrieving of the ship monitor video content based on HBase
CN110427895A (en) * 2019-08-06 2019-11-08 李震 A kind of video content similarity method of discrimination based on computer vision and system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105069042A (en) * 2015-07-23 2015-11-18 北京航空航天大学 Content-based data retrieval methods for unmanned aerial vehicle spying images
CN106126585A (en) * 2016-06-20 2016-11-16 北京航空航天大学 Unmanned plane image search method based on quality grading with the combination of perception Hash feature
CN107040790A (en) * 2017-04-01 2017-08-11 华南理工大学 A kind of video content certification and tampering location method based on many granularity Hash
CN109918537A (en) * 2019-01-18 2019-06-21 杭州电子科技大学 A kind of method for quickly retrieving of the ship monitor video content based on HBase
CN110427895A (en) * 2019-08-06 2019-11-08 李震 A kind of video content similarity method of discrimination based on computer vision and system

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112434185A (en) * 2020-10-26 2021-03-02 国家广播电视总局广播电视规划院 Method, system, server and storage medium for searching similar video clips
CN112434185B (en) * 2020-10-26 2023-07-14 国家广播电视总局广播电视规划院 Method, system, server and storage medium for searching similar video clips
CN113704532A (en) * 2020-11-25 2021-11-26 天翼智慧家庭科技有限公司 Method and system for improving recall rate of picture retrieval
CN113704532B (en) * 2020-11-25 2024-04-26 天翼数字生活科技有限公司 Method and system for improving picture retrieval recall rate
CN112698661A (en) * 2021-03-22 2021-04-23 成都睿铂科技有限责任公司 Aerial survey data acquisition method, device and system for aircraft and storage medium

Also Published As

Publication number Publication date
CN110942002B (en) 2023-11-07

Similar Documents

Publication Publication Date Title
CN110942002A (en) Unmanned aerial vehicle aerial photography video frame positioning method based on rotation invariant perceptual hashing
Nakai et al. Use of affine invariants in locally likely arrangement hashing for camera-based document image retrieval
US8370338B2 (en) Large-scale asymmetric comparison computation for binary embeddings
JP5911578B2 (en) Method for encoding feature point position information of image, computer program, and mobile device
US8538164B2 (en) Image patch descriptors
US20110299770A1 (en) Performance of image recognition algorithms by pruning features, image scaling, and spatially constrained feature matching
CN108241645B (en) Image processing method and device
CN105335469A (en) Method and device for image matching and retrieving
EP3172700A1 (en) Improved method for fingerprint matching and camera identification, device and system
CN110599478A (en) Image area copying and pasting tampering detection method
CN105224619B (en) A kind of spatial relationship matching process and system suitable for video/image local feature
KR20140112635A (en) Feature Based Image Processing Apparatus and Method
Tsai et al. Mobile visual search using image and text features
KR101831587B1 (en) Camera-Angle Invariant Feature Extraction Method for Low Computation Number and Character Recognition
Tsai et al. Extent: Inferring image metadata from context and content
Min et al. Video copy detection using inclined video tomography and bag-of-visual-words
CN112131971A (en) Method for carrying out 256-dimensional binary quantization on 128-dimensional floating point type feature descriptor of HardNet
Ye et al. Multipurification of matching pairs based on ORB feature and PCB alignment case study
Gopakumar A survey on image splice forgery detection and localization techniques
KR101367821B1 (en) video identification method and apparatus using symmetric information of hierachical image blocks
Iida et al. Robust image identification for double-compressed jpeg images
CN115497099B (en) Single character image matching and identifying method based on circular scanning
CN115578734B (en) Single character image matching recognition method based on pyramid features
CN113642573B (en) Picture separation method based on grids
Pan et al. Recognizing characters with severe perspective distortion using hash tables and perspective invariants

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant