EP3092602A1 - Procédé de description de courbes planaires au moyen d'espaces d'échelles morphologiques - Google Patents

Procédé de description de courbes planaires au moyen d'espaces d'échelles morphologiques

Info

Publication number
EP3092602A1
EP3092602A1 EP14710049.9A EP14710049A EP3092602A1 EP 3092602 A1 EP3092602 A1 EP 3092602A1 EP 14710049 A EP14710049 A EP 14710049A EP 3092602 A1 EP3092602 A1 EP 3092602A1
Authority
EP
European Patent Office
Prior art keywords
distance
vector
morphological scale
silhouette
scale
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP14710049.9A
Other languages
German (de)
English (en)
Inventor
Erdem AKAGUNDUZ
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Aselsan Elektronik Sanayi ve Ticaret AS
Original Assignee
Aselsan Elektronik Sanayi ve Ticaret AS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Aselsan Elektronik Sanayi ve Ticaret AS filed Critical Aselsan Elektronik Sanayi ve Ticaret AS
Publication of EP3092602A1 publication Critical patent/EP3092602A1/fr
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/40Analysis of texture
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/469Contour-based spatial representations, e.g. vector-coding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/752Contour matching
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8543Content authoring using a description language, e.g. Multimedia and Hypermedia information coding Expert Group [MHEG], eXtensible Markup Language [XML]

Definitions

  • the invention relates to a method to construct a descriptor from binary silhouette images.
  • the input image is a black-and-white (binary) image including the silhouette (a complete closed region of white pixels) of an object in front of a complete black background.
  • the method calculates curvatures for both the Gaussian Scale-space and the morphological scale-space of the curve. Then these planar orientations are mapped to a color image, in order to describe the silhouette object.
  • This descriptor can be used for object recognition purposes. Background of the invention
  • the United States patent document US6711293 discloses a method to detect salient features on images, in which difference of Gaussians are used to construct a scale-space.
  • the present invention resembles the mentioned method in the sense that it performs feature extraction using a scale- space constructed from difference of Gaussians.
  • the present invention generates a scale-space using (a continuous mathematical representation of) closed curves and a curvature operator.
  • the mentioned method generates the scale-space using difference of Gaussians of image pixels.
  • the output of the mentioned method is a list of salient points on the image, whereas the output of the present invention is a list salient point on the contours of a silhouette specifically.
  • the United State patent application US2010080469 discloses system and method of generating feature descriptors for image identification.
  • Input image is Gaussian-blurred at different scales.
  • a difference of Gaussian space is obtained from differences of adjacent Gaussian-blurred images.
  • Key points are identified in the difference-of-Gaussian space.
  • primary sampling points are defined with three dimensional relative positions from key point and reaching into planes of different scales.
  • Secondary sampling points are identified for each primary sampling point.
  • Secondary image gradients are obtained between an image at a primary sampling point and images at secondary sampling points corresponding to this primary sampling point. Secondary image gradients form components of primary image gradients at primary sampling points.
  • Primary image gradients are concatenated to obtain a descriptor vector for input image.
  • Descriptor vector thus obtained is scale invariant and requires a number of additions equal to number of primary sampling points multiplied by a number of secondary sampling points.
  • US2013223730 discloses a feature descriptor extracting method in a feature descriptor extracting apparatus is provided.
  • the feature descriptor extracting method involves receiving an image from which a feature descriptor will be extracted, extracting a point at which a change in a pixel statistical value of the image is large as a feature point, and extracting a patch centered on the feature point, blocking the patch to calculate a statistical value of each of a plurality of patch blocks, calculating a morphological gradient by using a statistical value of the block-converted patch, and extracting a feature descriptor by using the morphological gradient in consideration of required feature descriptor complexity.
  • the United State patent application US20040184677 discloses a method detects silhouette edges in images. An ambient image is acquired of a scene with ambient light. A set of illuminated images is also acquired of the scene. Each illuminated image is acquired with a different light source illuminating the scene. The ambient image is combined with the set of illuminated to detect cast shadows, and silhouette edge pixels are located from the cast shadows.
  • the object of the invention is to provide a method to construct a descriptor from binary silhouette images.
  • Another object of the invention is to construct the orientations of all points of the silhouette in all morphological scale levels. Another object of the invention is to provide fast recognition with the learning distance vector.
  • Figure 1 is the flowchart of the method for describing planar curves using morphological scale spaces
  • Figure 2 is the flowchart of the checking the type of distance vector and calculating the distance
  • Figure 3 is the transaction for GSS of the curve to Orientation vector calculation to Orientation scale-space
  • a method for describing planar curves using morphological scale spaces (100) comprises the steps of;
  • step 102 - combining all local information, which are created in step 102 and step
  • step "checking the type of distance vector and calculating the distance (108)" comprises the sub-steps of;
  • the shape of an object is usually obtained via a segmentation operation which outputs a binary silhouette and/or a contour.
  • This contour is a closed planar curve sampled in pixel coordinates.
  • step 102 "sampling the arc-length of the curve by using continuous representation with the formula of parametric curve", a uniform-length parametrization is useful if a scale-space of the curve is to be constructed.
  • a continuous representation (B-spline) is used with the equation 1;
  • the silhouette is obtained via an active contours based method (i.e. as a result of an automatic or semi-automatic object segmentation operation), in which the curve is already defined with a parametric model (such as in Brigger et al. (2000) [1]), curve fitting step is not needed.
  • a parametric representation such as a continuous representation, it is very easy to uniformly sample the curve. If the r parameter is chosen uniformly between 0 to rmax, arcs of uniform length can be obtained. Each object contour is sampled into 512 numbers of points, which divide the curve into 512 equal length arcs. It is also possible to use affine-length parametrization (such as in Awrangjeb et al.
  • step 103 "constructing the orientation scale-space with the variable-scale Gaussian functions with parametric curve and orientation angle"; the orientation angle at a point is defined as the angle between the orientation vector and the x- axis, where the orientation vector is the unit vector perpendicular to the tangential line at that point;
  • ' x and ' y denote the first derivatives of the x and y components of the closed curve C(r) along the curve parameter r. Since O(r) can take values from 0 to 2 ⁇ radian; atan2 function (a two argument variation of the arctangent function that can distinguish diametrically opposite directions) is used. Consequently, the scale- space of a curve L(r, ⁇ ) is defined as:
  • L(r, ⁇ ) is the convolution of the variable-scale Gaussian function g(r, ⁇ ) ( ⁇ being the standard deviation) with the parametric curve C(r).
  • the orientation scale-space (OSS) 0(r, ⁇ ) can be defined as in equation
  • orientation angle values are stacked on top of each other and a (o- s)-by-(512) matrix of orientation angle values is obtained.
  • This matrix is called the orientation scale-space (OSS) and is depicted in the right column ( Figure 3).
  • step 104 "combining all local information, which are created in step 102 and step 103, and creating silhouette orientation images"; extracting orientation angle at a point provides local information.
  • all the local information should be combined in such a way that the representation carries all that the local pieces posses, while staying invariant under certain transformations.
  • step 105 "finding the minimum distance match for two silhouettes which are created in step 104", the distance D between two SOIs are calculated, the hue differences between the corresponding pixels (at most 0.5 along the hue circle) are accumulated and normalized; In the equation 7, the overall distance Da,b between two SOIs takes values from 0 to 1.
  • the radial SOI of two identical silhouettes with different starting points will be rotated versions of each other because for radial SOI the radial axis determines the parametric position r.
  • the first silhouette is the 20° rotated version of the other.
  • the best a obtained from equation 8 corresponds to the transformed silhouette whose hue channel is shifted by 20/360 (i.e. approximately 2 pixels shift when M is 32).
  • the rotation angle can be retrieved as accurately as the resolution of the SOI permits since positions are quantized into M.
  • an in- stabilized platform may experience in-plane rotation (camera roll).
  • the hue channel search can be limited to +/- 1/12 so that the rotation invariance capability is adjusted according to the needs of the problem. This way the computation burden is lightened as well.
  • step 106 "applying closing operation to the multiple levels of the silhouette's morphological scale-space and obtaining new scale-space which has the binary silhouette with operators with increasing size"; the silhouettes of the same class will have similar orientation distribution along their boundaries. Although this happens to be true for most of the cases, when silhouettes have small articulated parts or unexpected discontinuities, matching may not be performed.
  • the proposed representation is applied to multiple levels of the silhouette's morphological scale-space (MSS). This new scale-space is obtained simply by closing (dilation + erosion) the binary silhouette with operators of increasing size (Equation 9). The closing operation is applied on the binary image before the chain code is extracted.
  • MSS silhouette's morphological scale-space
  • the ⁇ operator denotes the morphological closing operation which is applied to binary silhouette B(x, y).
  • the structuring element f ( ⁇ , ⁇ ) is parametrized by pixel size o. At each MSS level, o is increased such that the closing operations affect a larger region. In our experiments o is k-20 pixels, where k is the MSS level starting from 0.
  • an extended distance feature vector By applying minimum distance formula to mutually corresponding levels of the MSS of the two silhouettes, an extended distance feature vector can be obtained:
  • Di a; b (a; r) denotes the distance between the SOIs of silhouettes a and b extracted from their corresponding ith MSS level, which ranges from 0 to m.
  • the distance feature vector is calculated between two silhouettes and search in rotation invariance dimension can be limited according to the needs of the problem, the computational complexity of this step is trivial compared to other methods in the literature that include dynamic programming and inner distance calculation (Ling and Jacobs, 2007 [4] ).
  • the mutual distance between two planar curves is defined by the vector Da;b.
  • a classifier can be trained which will cluster different categories of silhouettes. Since the vector to be learned is not a self-descriptor, but a mutual distance definition; these types of problems are referred as distance learning problems.
  • step 201 "if the distance is linear, then the weighed linear sum of the distance vector is calculated to obtain a scalar distance value", the weighted linear sum of the distance vector Da;b is calculated to obtain a scalar distance value da;b.
  • the cost function equation 11 is solved for a training set of distance vectors.
  • la;b is the label of the training vector da;b. If a and b have the same category, la;b is 0. It is 1 if they are not.
  • step 202 "if the distance is non-linear, then training an artifical neural network is used on the non-linear distance", using a linearly weighted sum of the distance vectors Da;b, the distance categories within the Da;b space are linearly separable.
  • this complex space that is constructed by using the Gaussian and morphological space-spaces of curves may be consisting of categories which are expectedly clustered in a nonlinear geometry. For this reason, it is logical to check the performance of a non-linear distance classifier and compare it with the linearly weighted model.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Databases & Information Systems (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Security & Cryptography (AREA)
  • Signal Processing (AREA)
  • Image Analysis (AREA)

Abstract

L'invention concerne un procédé de construction d'un descripteur à partir d'images de silhouette binaires. L'image d'entrée est une image en noir et blanc (binaire) comprenant la silhouette (une zone complètement fermée de pixels blancs) d'un objet (une cible) placé devant un arrière-plan entièrement noir. Le procédé calcule des courbures de l'espace d'échelle gaussien et de l'espace d'échelle morphologique de la courbe. Ces orientations planaires sont ensuite mappées sur une image en couleurs, afin de décrire l'objet silhouette. Ce descripteur peut être utilisé à des fins de reconnaissance d'objet.
EP14710049.9A 2014-01-10 2014-01-10 Procédé de description de courbes planaires au moyen d'espaces d'échelles morphologiques Withdrawn EP3092602A1 (fr)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/IB2014/058172 WO2015104585A1 (fr) 2014-01-10 2014-01-10 Procédé de description de courbes planaires au moyen d'espaces d'échelles morphologiques

Publications (1)

Publication Number Publication Date
EP3092602A1 true EP3092602A1 (fr) 2016-11-16

Family

ID=50277259

Family Applications (1)

Application Number Title Priority Date Filing Date
EP14710049.9A Withdrawn EP3092602A1 (fr) 2014-01-10 2014-01-10 Procédé de description de courbes planaires au moyen d'espaces d'échelles morphologiques

Country Status (3)

Country Link
EP (1) EP3092602A1 (fr)
KR (1) KR20160106113A (fr)
WO (1) WO2015104585A1 (fr)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111145228B (zh) * 2019-12-23 2023-05-26 西安电子科技大学 基于局部轮廓点与形状特征融合的异源图像配准方法

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6711293B1 (en) 1999-03-08 2004-03-23 The University Of British Columbia Method and apparatus for identifying scale invariant features in an image and use of same for locating an object in an image
US7206449B2 (en) 2003-03-19 2007-04-17 Mitsubishi Electric Research Laboratories, Inc. Detecting silhouette edges in images
US8363973B2 (en) 2008-10-01 2013-01-29 Fuji Xerox Co., Ltd. Descriptor for image corresponding point matching
EP2724295B1 (fr) * 2012-02-27 2017-03-22 Aselsan Elektronik Sanayi ve Ticaret Anonim Sirketi Système et procédé pour identifier des caractéristiques ne variant pas en fonction de l'échelle de contours d'objet sur des images
KR101912748B1 (ko) 2012-02-28 2018-10-30 한국전자통신연구원 확장성을 고려한 특징 기술자 생성 및 특징 기술자를 이용한 정합 장치 및 방법

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See references of WO2015104585A1 *

Also Published As

Publication number Publication date
KR20160106113A (ko) 2016-09-09
WO2015104585A1 (fr) 2015-07-16

Similar Documents

Publication Publication Date Title
Hughes et al. A deep learning framework for matching of SAR and optical imagery
Hassaballah et al. Image features detection, description and matching
Doush et al. Currency recognition using a smartphone: Comparison between color SIFT and gray scale SIFT algorithms
Schmid et al. Evaluation of interest point detectors
Oji An automatic algorithm for object recognition and detection based on ASIFT keypoints
Niu et al. Fast and effective Keypoint-based image copy-move forgery detection using complex-valued moment invariants
GB2550567A (en) Point Cloud Matching Method
Joshi et al. Recent advances in local feature detector and descriptor: a literature survey
Son et al. A multi-vision sensor-based fast localization system with image matching for challenging outdoor environments
Safdari et al. SIFT detector boosted by adaptive contrast threshold to improve matching robustness of remote sensing panchromatic images
Chen et al. Robust local feature descriptor for multisource remote sensing image registration
Zheng et al. A unified B-spline framework for scale-invariant keypoint detection
Paradowski et al. Local keypoints and global affine geometry: Triangles and ellipses for image fragment matching
Sujin et al. High-performance image forgery detection via adaptive SIFT feature extraction for low-contrast or small or smooth copy–move region images
Nanni et al. Ensemble of shape descriptors for shape retrieval and classification
Carro et al. Face recognition using SURF
Koutaki et al. Fast and high accuracy pattern matching using multi-stage refining eigen template
EP3092602A1 (fr) Procédé de description de courbes planaires au moyen d'espaces d'échelles morphologiques
Al-Azzawy Eigenface and SIFT for gender classification
Anggara et al. Integrated Colormap and ORB detector method for feature extraction approach in augmented reality
Jiang et al. Rotation and scale invariant shape context registration for remote sensing images with background variations
Mohapatra et al. Comparative study of corner and feature extractors for real-time object recognition in image processing
You et al. Pattern matching for industrial object recognition using geometry-based vector mapping descriptors
Cuong et al. SIFT Application Separates Motion Characteristics and Identifies Symbols on Tires
Li et al. A fast rotated template matching based on point feature

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20151007

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

DAX Request for extension of the european patent (deleted)
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20170301