WO2015104585A1 - Method for describing planar curves using morphological scale spaces - Google Patents

Method for describing planar curves using morphological scale spaces Download PDF

Info

Publication number
WO2015104585A1
WO2015104585A1 PCT/IB2014/058172 IB2014058172W WO2015104585A1 WO 2015104585 A1 WO2015104585 A1 WO 2015104585A1 IB 2014058172 W IB2014058172 W IB 2014058172W WO 2015104585 A1 WO2015104585 A1 WO 2015104585A1
Authority
WO
WIPO (PCT)
Prior art keywords
distance
vector
morphological scale
silhouette
scale
Prior art date
Application number
PCT/IB2014/058172
Other languages
French (fr)
Inventor
Erdem AKAGUNDUZ
Original Assignee
Aselsan Elektronik Sanayi Ve Ticaret Anonim Sirketi
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Aselsan Elektronik Sanayi Ve Ticaret Anonim Sirketi filed Critical Aselsan Elektronik Sanayi Ve Ticaret Anonim Sirketi
Priority to EP14710049.9A priority Critical patent/EP3092602A1/en
Priority to KR1020167021178A priority patent/KR20160106113A/en
Priority to PCT/IB2014/058172 priority patent/WO2015104585A1/en
Publication of WO2015104585A1 publication Critical patent/WO2015104585A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/40Analysis of texture
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/469Contour-based spatial representations, e.g. vector-coding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/752Contour matching
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8543Content authoring using a description language, e.g. Multimedia and Hypermedia information coding Expert Group [MHEG], eXtensible Markup Language [XML]

Definitions

  • the invention relates to a method to construct a descriptor from binary silhouette images.
  • the input image is a black-and-white (binary) image including the silhouette (a complete closed region of white pixels) of an object in front of a complete black background.
  • the method calculates curvatures for both the Gaussian Scale-space and the morphological scale-space of the curve. Then these planar orientations are mapped to a color image, in order to describe the silhouette object.
  • This descriptor can be used for object recognition purposes. Background of the invention
  • the United States patent document US6711293 discloses a method to detect salient features on images, in which difference of Gaussians are used to construct a scale-space.
  • the present invention resembles the mentioned method in the sense that it performs feature extraction using a scale- space constructed from difference of Gaussians.
  • the present invention generates a scale-space using (a continuous mathematical representation of) closed curves and a curvature operator.
  • the mentioned method generates the scale-space using difference of Gaussians of image pixels.
  • the output of the mentioned method is a list of salient points on the image, whereas the output of the present invention is a list salient point on the contours of a silhouette specifically.
  • the United State patent application US2010080469 discloses system and method of generating feature descriptors for image identification.
  • Input image is Gaussian-blurred at different scales.
  • a difference of Gaussian space is obtained from differences of adjacent Gaussian-blurred images.
  • Key points are identified in the difference-of-Gaussian space.
  • primary sampling points are defined with three dimensional relative positions from key point and reaching into planes of different scales.
  • Secondary sampling points are identified for each primary sampling point.
  • Secondary image gradients are obtained between an image at a primary sampling point and images at secondary sampling points corresponding to this primary sampling point. Secondary image gradients form components of primary image gradients at primary sampling points.
  • Primary image gradients are concatenated to obtain a descriptor vector for input image.
  • Descriptor vector thus obtained is scale invariant and requires a number of additions equal to number of primary sampling points multiplied by a number of secondary sampling points.
  • US2013223730 discloses a feature descriptor extracting method in a feature descriptor extracting apparatus is provided.
  • the feature descriptor extracting method involves receiving an image from which a feature descriptor will be extracted, extracting a point at which a change in a pixel statistical value of the image is large as a feature point, and extracting a patch centered on the feature point, blocking the patch to calculate a statistical value of each of a plurality of patch blocks, calculating a morphological gradient by using a statistical value of the block-converted patch, and extracting a feature descriptor by using the morphological gradient in consideration of required feature descriptor complexity.
  • the United State patent application US20040184677 discloses a method detects silhouette edges in images. An ambient image is acquired of a scene with ambient light. A set of illuminated images is also acquired of the scene. Each illuminated image is acquired with a different light source illuminating the scene. The ambient image is combined with the set of illuminated to detect cast shadows, and silhouette edge pixels are located from the cast shadows.
  • the object of the invention is to provide a method to construct a descriptor from binary silhouette images.
  • Another object of the invention is to construct the orientations of all points of the silhouette in all morphological scale levels. Another object of the invention is to provide fast recognition with the learning distance vector.
  • Figure 1 is the flowchart of the method for describing planar curves using morphological scale spaces
  • Figure 2 is the flowchart of the checking the type of distance vector and calculating the distance
  • Figure 3 is the transaction for GSS of the curve to Orientation vector calculation to Orientation scale-space
  • a method for describing planar curves using morphological scale spaces (100) comprises the steps of;
  • step 102 - combining all local information, which are created in step 102 and step
  • step "checking the type of distance vector and calculating the distance (108)" comprises the sub-steps of;
  • the shape of an object is usually obtained via a segmentation operation which outputs a binary silhouette and/or a contour.
  • This contour is a closed planar curve sampled in pixel coordinates.
  • step 102 "sampling the arc-length of the curve by using continuous representation with the formula of parametric curve", a uniform-length parametrization is useful if a scale-space of the curve is to be constructed.
  • a continuous representation (B-spline) is used with the equation 1;
  • the silhouette is obtained via an active contours based method (i.e. as a result of an automatic or semi-automatic object segmentation operation), in which the curve is already defined with a parametric model (such as in Brigger et al. (2000) [1]), curve fitting step is not needed.
  • a parametric representation such as a continuous representation, it is very easy to uniformly sample the curve. If the r parameter is chosen uniformly between 0 to rmax, arcs of uniform length can be obtained. Each object contour is sampled into 512 numbers of points, which divide the curve into 512 equal length arcs. It is also possible to use affine-length parametrization (such as in Awrangjeb et al.
  • step 103 "constructing the orientation scale-space with the variable-scale Gaussian functions with parametric curve and orientation angle"; the orientation angle at a point is defined as the angle between the orientation vector and the x- axis, where the orientation vector is the unit vector perpendicular to the tangential line at that point;
  • ' x and ' y denote the first derivatives of the x and y components of the closed curve C(r) along the curve parameter r. Since O(r) can take values from 0 to 2 ⁇ radian; atan2 function (a two argument variation of the arctangent function that can distinguish diametrically opposite directions) is used. Consequently, the scale- space of a curve L(r, ⁇ ) is defined as:
  • L(r, ⁇ ) is the convolution of the variable-scale Gaussian function g(r, ⁇ ) ( ⁇ being the standard deviation) with the parametric curve C(r).
  • the orientation scale-space (OSS) 0(r, ⁇ ) can be defined as in equation
  • orientation angle values are stacked on top of each other and a (o- s)-by-(512) matrix of orientation angle values is obtained.
  • This matrix is called the orientation scale-space (OSS) and is depicted in the right column ( Figure 3).
  • step 104 "combining all local information, which are created in step 102 and step 103, and creating silhouette orientation images"; extracting orientation angle at a point provides local information.
  • all the local information should be combined in such a way that the representation carries all that the local pieces posses, while staying invariant under certain transformations.
  • step 105 "finding the minimum distance match for two silhouettes which are created in step 104", the distance D between two SOIs are calculated, the hue differences between the corresponding pixels (at most 0.5 along the hue circle) are accumulated and normalized; In the equation 7, the overall distance Da,b between two SOIs takes values from 0 to 1.
  • the radial SOI of two identical silhouettes with different starting points will be rotated versions of each other because for radial SOI the radial axis determines the parametric position r.
  • the first silhouette is the 20° rotated version of the other.
  • the best a obtained from equation 8 corresponds to the transformed silhouette whose hue channel is shifted by 20/360 (i.e. approximately 2 pixels shift when M is 32).
  • the rotation angle can be retrieved as accurately as the resolution of the SOI permits since positions are quantized into M.
  • an in- stabilized platform may experience in-plane rotation (camera roll).
  • the hue channel search can be limited to +/- 1/12 so that the rotation invariance capability is adjusted according to the needs of the problem. This way the computation burden is lightened as well.
  • step 106 "applying closing operation to the multiple levels of the silhouette's morphological scale-space and obtaining new scale-space which has the binary silhouette with operators with increasing size"; the silhouettes of the same class will have similar orientation distribution along their boundaries. Although this happens to be true for most of the cases, when silhouettes have small articulated parts or unexpected discontinuities, matching may not be performed.
  • the proposed representation is applied to multiple levels of the silhouette's morphological scale-space (MSS). This new scale-space is obtained simply by closing (dilation + erosion) the binary silhouette with operators of increasing size (Equation 9). The closing operation is applied on the binary image before the chain code is extracted.
  • MSS silhouette's morphological scale-space
  • the ⁇ operator denotes the morphological closing operation which is applied to binary silhouette B(x, y).
  • the structuring element f ( ⁇ , ⁇ ) is parametrized by pixel size o. At each MSS level, o is increased such that the closing operations affect a larger region. In our experiments o is k-20 pixels, where k is the MSS level starting from 0.
  • an extended distance feature vector By applying minimum distance formula to mutually corresponding levels of the MSS of the two silhouettes, an extended distance feature vector can be obtained:
  • Di a; b (a; r) denotes the distance between the SOIs of silhouettes a and b extracted from their corresponding ith MSS level, which ranges from 0 to m.
  • the distance feature vector is calculated between two silhouettes and search in rotation invariance dimension can be limited according to the needs of the problem, the computational complexity of this step is trivial compared to other methods in the literature that include dynamic programming and inner distance calculation (Ling and Jacobs, 2007 [4] ).
  • the mutual distance between two planar curves is defined by the vector Da;b.
  • a classifier can be trained which will cluster different categories of silhouettes. Since the vector to be learned is not a self-descriptor, but a mutual distance definition; these types of problems are referred as distance learning problems.
  • step 201 "if the distance is linear, then the weighed linear sum of the distance vector is calculated to obtain a scalar distance value", the weighted linear sum of the distance vector Da;b is calculated to obtain a scalar distance value da;b.
  • the cost function equation 11 is solved for a training set of distance vectors.
  • la;b is the label of the training vector da;b. If a and b have the same category, la;b is 0. It is 1 if they are not.
  • step 202 "if the distance is non-linear, then training an artifical neural network is used on the non-linear distance", using a linearly weighted sum of the distance vectors Da;b, the distance categories within the Da;b space are linearly separable.
  • this complex space that is constructed by using the Gaussian and morphological space-spaces of curves may be consisting of categories which are expectedly clustered in a nonlinear geometry. For this reason, it is logical to check the performance of a non-linear distance classifier and compare it with the linearly weighted model.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Databases & Information Systems (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Security & Cryptography (AREA)
  • Signal Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a method to construct a descriptor from binary silhouette images. The input image is a black-and-white (binary) image including the silhouette (a complete closed region of white pixels) of an object (i.e. a target) in front of a complete black background. The method calculates curvatures for both the Gaussian Scale-space and the morphological scale-space of the curve. Then these planar orientations are mapped to a color image, in order to describe the silhouette object. This descriptor can be used for object recognition purposes.

Description

DESCRIPTION
Method for Describing Planar Curves Using Morphological Scale Spaces Field of the invention
The invention relates to a method to construct a descriptor from binary silhouette images. The input image is a black-and-white (binary) image including the silhouette (a complete closed region of white pixels) of an object in front of a complete black background. The method calculates curvatures for both the Gaussian Scale-space and the morphological scale-space of the curve. Then these planar orientations are mapped to a color image, in order to describe the silhouette object. This descriptor can be used for object recognition purposes. Background of the invention
The United States patent document US6711293, an application in the state of the art, discloses a method to detect salient features on images, in which difference of Gaussians are used to construct a scale-space. The present invention resembles the mentioned method in the sense that it performs feature extraction using a scale- space constructed from difference of Gaussians. However the present invention generates a scale-space using (a continuous mathematical representation of) closed curves and a curvature operator. However the mentioned method generates the scale-space using difference of Gaussians of image pixels. In addition, the output of the mentioned method is a list of salient points on the image, whereas the output of the present invention is a list salient point on the contours of a silhouette specifically.
The academic paper "Scale-based Description and Reconition of Planar Curves and Two-dimensional Shapes" (F.Mokhtarian and A.Macworth) discloses a method for constructing a representation for closed curves using a scale-space representation. The mentioned method resembles the present invention in the sense that it uses "circle of curvature" values to construct the scale-space. However the present method uses difference of the levels of the scale-space and unlike the mentioned method, identifies salient points over the curve, with their scale-information.
A patent application PCT/IB2012/050883 System and Method for Identifying Scale Invariant Features of Object Outlines on Images resembles the present application in the sense that it constructs a descriptor of the silhouette image using the orientations of the extracted feature points. However the present application also uses the morphological scale-space of the curve, thus introduces more informative description.
The United State patent application US2010080469 discloses system and method of generating feature descriptors for image identification. Input image is Gaussian-blurred at different scales. A difference of Gaussian space is obtained from differences of adjacent Gaussian-blurred images. Key points are identified in the difference-of-Gaussian space. For each key point, primary sampling points are defined with three dimensional relative positions from key point and reaching into planes of different scales. Secondary sampling points are identified for each primary sampling point. Secondary image gradients are obtained between an image at a primary sampling point and images at secondary sampling points corresponding to this primary sampling point. Secondary image gradients form components of primary image gradients at primary sampling points. Primary image gradients are concatenated to obtain a descriptor vector for input image. Descriptor vector thus obtained is scale invariant and requires a number of additions equal to number of primary sampling points multiplied by a number of secondary sampling points. The United State patent application US2013223730 discloses a feature descriptor extracting method in a feature descriptor extracting apparatus is provided. The feature descriptor extracting method involves receiving an image from which a feature descriptor will be extracted, extracting a point at which a change in a pixel statistical value of the image is large as a feature point, and extracting a patch centered on the feature point, blocking the patch to calculate a statistical value of each of a plurality of patch blocks, calculating a morphological gradient by using a statistical value of the block-converted patch, and extracting a feature descriptor by using the morphological gradient in consideration of required feature descriptor complexity. The United State patent application US20040184677 discloses a method detects silhouette edges in images. An ambient image is acquired of a scene with ambient light. A set of illuminated images is also acquired of the scene. Each illuminated image is acquired with a different light source illuminating the scene. The ambient image is combined with the set of illuminated to detect cast shadows, and silhouette edge pixels are located from the cast shadows.
Objects of the invention
The object of the invention is to provide a method to construct a descriptor from binary silhouette images.
Another object of the invention is to construct the orientations of all points of the silhouette in all morphological scale levels. Another object of the invention is to provide fast recognition with the learning distance vector.
Detailed description of the invention A method for describing planar curves using morphological scale spaces in order to fulfill the objects of the present invention is illustrated in the attached figures, where:
Figure 1 is the flowchart of the method for describing planar curves using morphological scale spaces
Figure 2 is the flowchart of the checking the type of distance vector and calculating the distance
Figure 3 is the transaction for GSS of the curve to Orientation vector calculation to Orientation scale-space
A method for describing planar curves using morphological scale spaces (100) comprises the steps of;
taking input data from the camera means and creating the curve from input data (101),
sampling the arc-length of the curve by using continuous representation with the formula of parametric curve (102),
constructing the orientation scale-space with the variable- scale Gaussian function with parametric curve and orientation angle (103),
- combining all local information, which are created in step 102 and step
103, and creating silhouette orientation images (104),
finding the minimum distance match for two silhouettes which are created in step 104 (105),
applying closing operation to the multiple levels of the silhouette's morphological scale-space and obtaining new scale-space which has the binary silhouette with operators with increasing size (106),
matching the calculations which are found in step 105 and step 106 (107), checking the type of distance vector and calculating the distance (108), sending output to the imaging means (109). In the method for describing planar curves using morphological scale spaces (100), the step "checking the type of distance vector and calculating the distance (108)" comprises the sub-steps of;
if the distance is linear, then the weighed linear sum of the distance vector is calculated to obtain a scalar distance value (201),
if the distance is non-linear, then training an artifical neural network is used on the non-linear distance (202).
The shape of an object is usually obtained via a segmentation operation which outputs a binary silhouette and/or a contour. This contour is a closed planar curve sampled in pixel coordinates. In step 102, "sampling the arc-length of the curve by using continuous representation with the formula of parametric curve", a uniform-length parametrization is useful if a scale-space of the curve is to be constructed. For this purpose a continuous representation (B-spline) is used with the equation 1;
^ (Eqn. 1)
In the equation 1, C(r) stands for the parametric curve, whereas Pi is the ith control point and Bi;k is the kth order basis function for the ith control point. If the equation is written in matrix form; (» = J{tr ■■ P
(Eqn. 2) In the equation 2, P is the N-by-2 control point matrix and J is the L-by-N basis matrix (for L number of pixels of the silhouette). Thus, by using the L number of pixels of the silhouette (i.e. C(r)) and by calculating basis function for L silhouette pixels (i.e. J(k, r) matrix), we can calculate the control points matrix P, which is our continuous representation. For each row of J matrix, the r parameter of each pixel must be known. For this purpose, first the chain code of the closed curve is extracted. The chain code carries the distance between two neighboring pixels (1 or 2 units). Starting from an arbitrary point, the r parameter is calculated using the arc-length (i.e. the chain code). We make pseudo-inversing to calculate the control points, assuming no ill-conditions with equation 3; (Eqn. 3)
If the silhouette is obtained via an active contours based method (i.e. as a result of an automatic or semi-automatic object segmentation operation), in which the curve is already defined with a parametric model (such as in Brigger et al. (2000) [1]), curve fitting step is not needed. Using a parametric representation such as a continuous representation, it is very easy to uniformly sample the curve. If the r parameter is chosen uniformly between 0 to rmax, arcs of uniform length can be obtained. Each object contour is sampled into 512 numbers of points, which divide the curve into 512 equal length arcs. It is also possible to use affine-length parametrization (such as in Awrangjeb et al. (2007) [2]); however since it is not the CSS (contour scale-space) but the GSS (Gaussian scale-space) that the method requires and since affine-length parametrization is more fragile under noise, arc- length parametrization is preferred. In addition, the proposed method performs better when the curves are sampled in arc-length.
In step 103, "constructing the orientation scale-space with the variable-scale Gaussian functions with parametric curve and orientation angle"; the orientation angle at a point is defined as the angle between the orientation vector and the x- axis, where the orientation vector is the unit vector perpendicular to the tangential line at that point;
(Eqn. 4)
In the equation 4, 'x and 'y denote the first derivatives of the x and y components of the closed curve C(r) along the curve parameter r. Since O(r) can take values from 0 to 2π radian; atan2 function (a two argument variation of the arctangent function that can distinguish diametrically opposite directions) is used. Consequently, the scale- space of a curve L(r, σ) is defined as:
In the equation 5, L(r, σ) is the convolution of the variable-scale Gaussian function g(r, σ) (σ being the standard deviation) with the parametric curve C(r). Similarly, the orientation scale-space (OSS) 0(r, σ) can be defined as in equation
(Eqn. 6) The initial curve is incrementally convolved with Gaussian to produce curves separated by a constant factor k in scale- space, shown stacked in the left column (Figure 3). Similar to Lowe (2004) [3], each octave of scale-space (i.e. doubling of σ) is divided into s number of intervals. Once a complete octave is constructed, the Gaussian curve that has twice the initial value of σ is re-sampled into half. The middle column shows the orientation vectors calculated for each sampled point at each interval of each octave. Consequently for higher octaves (o>l), the sequence of orientation angle values are up sampled into highest resolution (512). Then starting from the same point (r=0) at each interval, 512 orientation angle values are stacked on top of each other and a (o- s)-by-(512) matrix of orientation angle values is obtained. This matrix is called the orientation scale-space (OSS) and is depicted in the right column (Figure 3).
In the step 104, "combining all local information, which are created in step 102 and step 103, and creating silhouette orientation images"; extracting orientation angle at a point provides local information. In order to globally define the silhouette, all the local information should be combined in such a way that the representation carries all that the local pieces posses, while staying invariant under certain transformations.
In step 105, "finding the minimum distance match for two silhouettes which are created in step 104", the distance D between two SOIs are calculated, the hue differences between the corresponding pixels (at most 0.5 along the hue circle) are accumulated and normalized;
Figure imgf000009_0001
In the equation 7, the overall distance Da,b between two SOIs takes values from 0 to 1. SOIs are scale and resolution invariant. Hence, for a scaled or sampled version of the silhouette image, the curve fitting and arc-length parametrization steps virtually construct the same OSS. However the starting point invariance, that is to say, the uncertainty of the position of the first point r0=0 while fitting the curve and the rotation invariance under planar rotations must be handled before two SOIs can be matched. The radial SOI of two identical silhouettes with different starting points will be rotated versions of each other because for radial SOI the radial axis determines the parametric position r. Thus, we may satisfy starting point invariance by searching for a minimum distance match by rotating one of the SOIs.
When a silhouette is introduced with an in-plane rotation, theoretically the relative positions of the contour pixels do not change. However the orientation angles of all pixels are rotated with the same amount. Thus, the hue values for each pixel of the SOI change by the same amount along the hue circle. Since the hue values linearly map to orientation angles, by checking the hue shift between two SOIs, the amount of in-plane rotation can be retrieved. An in-plane rotation may affect the curve fitting algorithm and the starting point may probably change for a rotated version of the silhouette. For this reason, whenever a hue shift check is carried out, a starting point invariance search should also be applied. Thus the search becomes two dimensional, where both hue channel and the radial SOI is rotated in order to find the minimum distance match for two silhouettes with the equation 8;
I ^ £s¾, y¾ I = ergtn r)}
(Eqn. 8)
Two silhouettes and their radial SOIs are depicted. The first silhouette is the 20° rotated version of the other. Experiments show that for a 20 ° rotation, the best a obtained from equation 8, corresponds to the transformed silhouette whose hue channel is shifted by 20/360 (i.e. approximately 2 pixels shift when M is 32). The rotation angle can be retrieved as accurately as the resolution of the SOI permits since positions are quantized into M.
In some cases, limited robustness to rotation can be sufficient. For instance, an in- stabilized platform may experience in-plane rotation (camera roll). In that case, the hue channel search can be limited to +/- 1/12 so that the rotation invariance capability is adjusted according to the needs of the problem. This way the computation burden is lightened as well.
In step 106, "applying closing operation to the multiple levels of the silhouette's morphological scale-space and obtaining new scale-space which has the binary silhouette with operators with increasing size"; the silhouettes of the same class will have similar orientation distribution along their boundaries. Although this happens to be true for most of the cases, when silhouettes have small articulated parts or unexpected discontinuities, matching may not be performed. In order to overcome this problem, the proposed representation is applied to multiple levels of the silhouette's morphological scale-space (MSS). This new scale-space is obtained simply by closing (dilation + erosion) the binary silhouette with operators of increasing size (Equation 9). The closing operation is applied on the binary image before the chain code is extracted. MO
(Eqn. 9)
In the equation 9, the · operator denotes the morphological closing operation which is applied to binary silhouette B(x, y). The structuring element f (·, ·) is parametrized by pixel size o. At each MSS level, o is increased such that the closing operations affect a larger region. In our experiments o is k-20 pixels, where k is the MSS level starting from 0.
By applying minimum distance formula to mutually corresponding levels of the MSS of the two silhouettes, an extended distance feature vector can be obtained:
£¾ ¾(% ?¾), £i¾s f; rm m> rs
(Eqn. 10) In the equation 10, Di a; b (a; r) denotes the distance between the SOIs of silhouettes a and b extracted from their corresponding ith MSS level, which ranges from 0 to m.
In the step 108, "checking the type of distance vector and calculating the distance", the distance feature vector is calculated between two silhouettes and search in rotation invariance dimension can be limited according to the needs of the problem, the computational complexity of this step is trivial compared to other methods in the literature that include dynamic programming and inner distance calculation (Ling and Jacobs, 2007 [4] ). The mutual distance between two planar curves is defined by the vector Da;b. Using this vector, a classifier can be trained which will cluster different categories of silhouettes. Since the vector to be learned is not a self-descriptor, but a mutual distance definition; these types of problems are referred as distance learning problems. In the step 201, "if the distance is linear, then the weighed linear sum of the distance vector is calculated to obtain a scalar distance value", the weighted linear sum of the distance vector Da;b is calculated to obtain a scalar distance value da;b.
Figure imgf000012_0001
(Eqn. 11)
In order to estimate the optimum weight vector w, the cost function equation 11 is solved for a training set of distance vectors.
Figure imgf000012_0002
In the equation 12, la;b is the label of the training vector da;b. If a and b have the same category, la;b is 0. It is 1 if they are not.
In the step 202, "if the distance is non-linear, then training an artifical neural network is used on the non-linear distance", using a linearly weighted sum of the distance vectors Da;b, the distance categories within the Da;b space are linearly separable. However this complex space that is constructed by using the Gaussian and morphological space-spaces of curves may be consisting of categories which are expectedly clustered in a nonlinear geometry. For this reason, it is logical to check the performance of a non-linear distance classifier and compare it with the linearly weighted model. For this purpose, an artifical neural network with 3(m+l) input nodes (where (m+1) = 4 is the number of MSS layers in Equation 10), h number of hidden layer nodes and a single output layer node, is trained. References
[IJ.Brigger, P., Hoeg, J., Unser, M., 2000. B-spline snakes: A flexible tool for parametric contour detection. IEEE Transactions on Image Processing 9, 1484-1496.
[2]. Awrangjeb, M., Lu, G., Murshed, M., 2007. An affine resilient curvature scale-space corner detector, in: 32. IEEE Int. Conf . Acoustics, Speech and Signal Processing (ICASSP 2007), pp. 1233-1236.
[3]. Lowe, D., 2004. Distinctive image features from local scale-invariant keypoints. International Journal of Computer Vision 2, 91-110.
[4]. Ling, H., Jacobs, D.W., 2007. Shape classification using the inner- distance. IEEE Transactions on Pattern Analysis and Machine Intelligence 29, 286-299.

Claims

The method for describing planar curves using morphological scale spaces (100) essentially comprises the steps of;
- taking input data from the camera means and creating the curve from input data (101),
- sampling the arc-length of the curve by using continuous representation with the formula of parametric curve (102),
- constructing the orientation scale- space with the variable- scale Gaussian function with parametric curve and orientation angle (103),
- combining all local information, which are created in step 102 and step 103, and creating silhouette orientation images (104),
- finding the minimum distance match for two silhouettes which are created in step 104 (105),
- applying closing operation to the multiple levels of the silhouette's morphological scale-space and obtaining new scale-space which has the binary silhouette with operators with increasing size (106),
- matching the calculations which are found in step 105 and step 106 (107),
- checking the type of distance vector and calculating the distance (108),
- sending output to the imaging means (109).
In the method for describing planar curves using morphological scale spaces (100), the step "checking the type of distance vector and calculating the distance (108)" further comprises the sub-steps of;
- if the distance is linear, then the weighed linear sum of the distance vector is calculated to obtain a scalar distance value (201),
- if the distance is non-linear, then training an artifical neural network is used on the non-linear distance (202).
Method for describing planar curves using morphological scale spaces (100) according to claim 1, characterized in that continuous C(r} = P, - ½(r) representation is performed according to
formula.
4. Method for describing planar curves using morphological scale spaces (100) according to claim 1, characterized in that constructing the orientation angle at a point is defined as the angle between the orientation vector and the x-axis, where the orientation vector is the unit vector perpendicular to the tangential line at that point is performed according to formula.
5. Method for describing planar curves using morphological scale spaces (100) according to claim 1, characterized in that the hue differences between the corresponding pixels is performed according to
:>,., = 2 . y y
formula.
Method for describing planar curves using morphological scale spaces (100) according to claim 1, characterized in that finding the minimum distance match for two silhouettes is performed according to
Figure imgf000015_0001
formula.
Method for describing planar curves using morphological scale spaces (100) according to claim 1, characterized in that calculation of the weighed linear sum of the distance vector is performed according to
m
Figure imgf000015_0002
formula. Method for describing planar curves using morphological scale spaces (100) according to claim 1, characterized in that estimation the optimum weight vector is performed according to
= argmm (daJf - l^
formula.
PCT/IB2014/058172 2014-01-10 2014-01-10 Method for describing planar curves using morphological scale spaces WO2015104585A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP14710049.9A EP3092602A1 (en) 2014-01-10 2014-01-10 Method for describing planar curves using morphological scale spaces
KR1020167021178A KR20160106113A (en) 2014-01-10 2014-01-10 Method for describing planar curves using morphological scale spaces
PCT/IB2014/058172 WO2015104585A1 (en) 2014-01-10 2014-01-10 Method for describing planar curves using morphological scale spaces

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/IB2014/058172 WO2015104585A1 (en) 2014-01-10 2014-01-10 Method for describing planar curves using morphological scale spaces

Publications (1)

Publication Number Publication Date
WO2015104585A1 true WO2015104585A1 (en) 2015-07-16

Family

ID=50277259

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2014/058172 WO2015104585A1 (en) 2014-01-10 2014-01-10 Method for describing planar curves using morphological scale spaces

Country Status (3)

Country Link
EP (1) EP3092602A1 (en)
KR (1) KR20160106113A (en)
WO (1) WO2015104585A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111145228A (en) * 2019-12-23 2020-05-12 西安电子科技大学 Heterogeneous image registration method based on local contour point and shape feature fusion

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6711293B1 (en) 1999-03-08 2004-03-23 The University Of British Columbia Method and apparatus for identifying scale invariant features in an image and use of same for locating an object in an image
US20040184677A1 (en) 2003-03-19 2004-09-23 Ramesh Raskar Detecting silhouette edges in images
US20100080469A1 (en) 2008-10-01 2010-04-01 Fuji Xerox Co., Ltd. Novel descriptor for image corresponding point matching
US20130223730A1 (en) 2012-02-28 2013-08-29 Electronics And Telecommunications Research Institute Scalable feature descriptor extraction and matching method and system
WO2013128233A1 (en) * 2012-02-27 2013-09-06 Aselsan Elektronik Sanayi Ve Ticaret Anonim Sirketi System and method for identifying scale invariant features of object outlines on images

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6711293B1 (en) 1999-03-08 2004-03-23 The University Of British Columbia Method and apparatus for identifying scale invariant features in an image and use of same for locating an object in an image
US20040184677A1 (en) 2003-03-19 2004-09-23 Ramesh Raskar Detecting silhouette edges in images
US20100080469A1 (en) 2008-10-01 2010-04-01 Fuji Xerox Co., Ltd. Novel descriptor for image corresponding point matching
WO2013128233A1 (en) * 2012-02-27 2013-09-06 Aselsan Elektronik Sanayi Ve Ticaret Anonim Sirketi System and method for identifying scale invariant features of object outlines on images
US20130223730A1 (en) 2012-02-28 2013-08-29 Electronics And Telecommunications Research Institute Scalable feature descriptor extraction and matching method and system

Non-Patent Citations (8)

* Cited by examiner, † Cited by third party
Title
A.C. JALBA ET AL: "Shape representation and recognition through morphological curvature scale spaces", IEEE TRANSACTIONS ON IMAGE PROCESSING, vol. 15, no. 2, 1 February 2006 (2006-02-01), pages 331 - 341, XP055129248, ISSN: 1057-7149, DOI: 10.1109/TIP.2005.860606 *
AWRANGJEB, M.; LU, G.; MURSHED, M.: "An affine resilient curvature scale-space corner detector", IEEE INT. CONF. ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2007, vol. 32, 2007, pages 1233 - 1236
BEBIS G ET AL: "Curvature scale-space-driven object recognition with an indexing scheme based on artificial neural networks", PATTERN RECOGNITION, ELSEVIER, GB, vol. 32, no. 7, 1 July 1999 (1999-07-01), pages 1175 - 1201, XP004169470, ISSN: 0031-3203, DOI: 10.1016/S0031-3203(98)00159-9 *
BEN K. JANG ET AL: "Morphological Scale Space for 2D Shape Smoothing", COMPUTER VISION AND IMAGE UNDERSTANDING, vol. 70, no. 2, 1 May 1998 (1998-05-01), pages 121 - 141, XP055129250, ISSN: 1077-3142, DOI: 10.1006/cviu.1997.0626 *
BRIGGER, P.; HOEG, J.; UNSER, M.: "B-spline snakes: A flexible tool for parametric contour detection", IEEE TRANSACTIONS ON IMAGE PROCESSING, vol. 9, 2000, pages 1484 - 1496, XP000963522, DOI: doi:10.1109/83.862624
ERDEM AKAGUNDUZ: "Scale invariant sillhouette features", SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2013 21ST, IEEE, 24 April 2013 (2013-04-24), pages 1 - 4, XP032423323, ISBN: 978-1-4673-5562-9, DOI: 10.1109/SIU.2013.6531586 *
LING, H.; JACOBS, D.W.: "Shape classification using the inner- distance", IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, vol. 29, 2007, pages 286 - 299, XP011147678, DOI: doi:10.1109/TPAMI.2007.41
LOWE, D.: "Distinctive image features from local scale-invariant keypoints", INTERNATIONAL JOURNAL OF COMPUTER VISION, vol. 2, 2004, pages 91 - 110

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111145228A (en) * 2019-12-23 2020-05-12 西安电子科技大学 Heterogeneous image registration method based on local contour point and shape feature fusion
CN111145228B (en) * 2019-12-23 2023-05-26 西安电子科技大学 Heterologous image registration method based on fusion of local contour points and shape features

Also Published As

Publication number Publication date
EP3092602A1 (en) 2016-11-16
KR20160106113A (en) 2016-09-09

Similar Documents

Publication Publication Date Title
Hassaballah et al. Image features detection, description and matching
Hughes et al. A deep learning framework for matching of SAR and optical imagery
Doush et al. Currency recognition using a smartphone: Comparison between color SIFT and gray scale SIFT algorithms
Schmid et al. Evaluation of interest point detectors
Oji An automatic algorithm for object recognition and detection based on ASIFT keypoints
Jaberi et al. Accurate and robust localization of duplicated region in copy–move image forgery
Niu et al. Fast and effective Keypoint-based image copy-move forgery detection using complex-valued moment invariants
GB2550567A (en) Point Cloud Matching Method
Joshi et al. Recent advances in local feature detector and descriptor: a literature survey
CN111199558A (en) Image matching method based on deep learning
Safdari et al. SIFT detector boosted by adaptive contrast threshold to improve matching robustness of remote sensing panchromatic images
Paradowski et al. Local keypoints and global affine geometry: Triangles and ellipses for image fragment matching
Zheng et al. A unified B-spline framework for scale-invariant keypoint detection
Nanni et al. Ensemble of shape descriptors for shape retrieval and classification
Carro et al. Face recognition using SURF
EP3092602A1 (en) Method for describing planar curves using morphological scale spaces
Sujin et al. High-performance image forgery detection via adaptive SIFT feature extraction for low-contrast or small or smooth copy–move region images
Wu et al. An accurate feature point matching algorithm for automatic remote sensing image registration
Anggara et al. Integrated Colormap and ORB detector method for feature extraction approach in augmented reality
Akagündüz Shape recognition using orientational and morphological scale‐spaces of curvatures
Al-Azzawy Eigenface and SIFT for gender classification
Jiang et al. Rotation and scale invariant shape context registration for remote sensing images with background variations
You et al. Pattern matching for industrial object recognition using geometry-based vector mapping descriptors
Li et al. A fast rotated template matching based on point feature
Xing et al. A robust line matching method based on local appearance descriptor and neighboring geometric attributes

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14710049

Country of ref document: EP

Kind code of ref document: A1

REEP Request for entry into the european phase

Ref document number: 2014710049

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2014710049

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 20167021178

Country of ref document: KR

Kind code of ref document: A