WO2022185156A1 - Marker for artificial neural networks, related computer-implemented method for the recognition and interpretation and related system - Google Patents
Marker for artificial neural networks, related computer-implemented method for the recognition and interpretation and related system Download PDFInfo
- Publication number
- WO2022185156A1 WO2022185156A1 PCT/IB2022/051625 IB2022051625W WO2022185156A1 WO 2022185156 A1 WO2022185156 A1 WO 2022185156A1 IB 2022051625 W IB2022051625 W IB 2022051625W WO 2022185156 A1 WO2022185156 A1 WO 2022185156A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- marker
- angular portion
- elements
- fact
- key point
- Prior art date
Links
- 239000003550 marker Substances 0.000 title claims abstract description 64
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 45
- 238000000034 method Methods 0.000 title claims description 26
- 239000003086 colorant Substances 0.000 claims abstract description 8
- 239000011159 matrix material Substances 0.000 claims abstract description 8
- 238000001514 detection method Methods 0.000 claims description 6
- 238000013529 biological neural network Methods 0.000 description 2
- 238000005094 computer simulation Methods 0.000 description 2
- 210000002569 neuron Anatomy 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000001154 acute effect Effects 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K19/00—Record carriers for use with machines and with at least a part designed to carry digital markings
- G06K19/06—Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code
- G06K19/06009—Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code with optically detectable marking
- G06K19/06037—Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code with optically detectable marking multi-dimensional coding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K19/00—Record carriers for use with machines and with at least a part designed to carry digital markings
- G06K19/06—Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code
- G06K19/06009—Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code with optically detectable marking
- G06K19/06046—Constructional details
- G06K19/0614—Constructional details the marking being selective to wavelength, e.g. color barcode or barcodes only visible under UV or IR
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/14—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
- G06K7/1404—Methods for optical code recognition
- G06K7/1408—Methods for optical code recognition the method being specifically adapted for the type of code
- G06K7/1417—2D bar codes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/14—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
- G06K7/1404—Methods for optical code recognition
- G06K7/1439—Methods for optical code recognition including a method step for retrieval of the optical code
- G06K7/1443—Methods for optical code recognition including a method step for retrieval of the optical code locating of the code in an image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/14—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
- G06K7/1404—Methods for optical code recognition
- G06K7/1439—Methods for optical code recognition including a method step for retrieval of the optical code
- G06K7/1456—Methods for optical code recognition including a method step for retrieval of the optical code determining the orientation of the optical code with respect to the reader and correcting therefore
Definitions
- the present invention relates to a marker for artificial neural networks, a related computer-implemented method for the recognition and interpretation, and a related system.
- markers also called fiducial markers
- fiducial markers are well known which are expressed as images and are recognized by electronic systems to provide information such as, e.g., in the case of augmented reality, a transition from 2D to 3D space.
- the marker-based recognition is based on the detection, within an acquired image, of some characteristic patterns detectable according to the strong contrast of the edges (usually square or round black on a white background) compared to the rest of the image. On the basis of the representation of the edges and of the figure inside the edges it is possible to classify the markers in different subgroups. Usually the recognition of the edges within the image is achieved by distinguishing the areas of the image itself where the gradients of pixels change dramatically from one value to another. The discrimination between one marker and another is carried out on the basis of an encoding of the value enclosed between the identified edges.
- fiducial markers which are of the type of square binary markers.
- the marker may be distorted by the perspective of the image and, therefore, difficult to recognize and/or decode.
- An artificial neural network is a computational model composed of artificial “neurons”, loosely inspired by the simplification of a biological neural network and is used to solve Artificial Intelligence engineering problems related to different technological fields such as computer science, electronics or other subjects.
- Neural networks cause computers to be able to solve problems independently and improve their capabilities.
- the neural network is “trained” so that it can provide results that are as reliable as possible.
- a neural network must, therefore, be able to autonomously process the information related to the positioning and the distortion of the marker.
- the main aim of the present invention is to devise a marker for artificial neural networks which is uniquely recognizable within an image by an artificial neural network.
- Another object of the present invention is to devise a marker for artificial neural networks, a related computer-implemented method for the recognition and interpretation and a related system which enable rapid and efficient decoding of the marker by an artificial neural network regardless of its placement within an image.
- Another object of the present invention is to devise a marker for artificial neural networks, a related computer-implemented method for the recognition and interpretation, and a related system which allow the mentioned drawbacks of the prior art to be overcome within a simple, rational, easy and effective to use as well as affordable solution.
- FIGS. 1 and 2 are schematic representations of a marker according to the invention
- Figure 3 is a schematic representation illustrating the steps of the method according to the invention.
- reference numeral 1 globally indicates a marker for artificial neural networks.
- artificial neural network refers to a computational model composed of artificial “neurons”, loosely inspired by the simplification of a biological neural network.
- the marker 1 comprises a plurality of elements 2 defining a matrix structure and provided with at least one inner area 3 the elements 2 of which define a matrix code.
- the inner area 3 comprises a plurality of elements 2 of white or black color.
- Each of these elements 2 may be associated with a value of 0 or 1, depending on its color.
- the inner area 3 encodes a binary code 4 for the representation of a predefined piece of information.
- the inner area 3 is decodable by a computer system capable of converting the binary code 4 into the aforementioned information.
- the marker 1 comprises at least one outer edge 6 to the inner area 3 comprising four angular portions 7, 8, 9, 10 of different colors.
- An artificial neural network in fact, is able to autonomously process the information and provide a univocal result. To achieve this result, the neural network must be “trained” so that it can provide results that are as reliable as possible.
- the angular portions 7, 8, 9, 10 of different colors surrounding the inner area 3 uniquely identify the marker 1 and make the identification thereof extremely fast and unequivocal.
- the outer edge 6 is also defined by a plurality of elements 2 arranged externally to the inner area 3.
- the outer edge 6 has a thickness equal to one of the elements 2.
- the elements 2 of which the outer edge 6 is composed are arranged side by side on a single level.
- the angular portions 7, 8, 9, 10 are each defined by three elements 2, arranged in an “L” pattern and forming the outer corners of the marker 1.
- the angular portions 7, 8, 9, 10 comprise: a first angular portion 7, the elements 2 of which are green; a second angular portion 8, the elements 2 of which are blue; a third angular portion 9, the elements 2 of which are white; a fourth angular portion 10, the elements 2 of which are red.
- colors are defined according to the RGB (Red Green Blue) model in which the color green is represented by the code (0, 255, 0), the color blue is represented by the code (0, 0, 255), the color white is represented by the code (255, 255, 255) and the color red is represented by the code (255, 0, 0).
- RGB Red Green Blue
- Each side portion 11 is bounded between two angular portions 7, 8, 9, 10.
- the inner area 3 represents a precise succession of elements 2 which encode a particular code.
- the arrangement of the angular portions 7, 8, 9, 10 allows the orientation of marker 1 to be determined and the inner area 3 to be correctly decoded.
- the first angular portion 7 is arranged in the upper left corner; the second angular portion 8 is arranged in the upper right corner; the third angular portion 9 is arranged in the lower right corner; the fourth angular portion 10 is arranged in the lower left corner.
- the inner edge 12 allows the angular portions 7, 8, 9, 10 to be effectively recognized thanks to the chromatic contrast between the various colors of the latter and the black color of the inner edge itself.
- the inner edge 12 in turn has a thickness equal to one of the elements 2.
- the elements 2 of which the inner edge 12 is composed are arranged side by side on a single level.
- the inner edge 12 has a different thickness. Consequently, if the inner area 3 is of the type of an S x S matrix, the marker 1 has a matrix structure (S+4) x (S+4).
- the present invention also relates to a computer- implemented method for the recognition and interpretation of a marker for artificial neural networks as described above.
- the present method is, therefore, performed by means of a computerized recognition and interpretation system.
- the method according to the invention comprises at least one step of acquisition of an image 13 containing at least one marker 1.
- the image 13 may contain a plurality of markers 1.
- the acquisition of the image 13 is performed by means of an image acquisition unit 14.
- the image acquisition unit 14 may comprise at least one of either a camera or a video camera.
- the image acquisition unit 14 may be of the fixed type, i.e., stably arranged in a predefined area, or of the movable type, such as e.g. a camera/video camera of a smart-phone, tablet, or the like.
- the method comprises one step of detection of the marker 1 within the image 13 by means of a first neural network 15.
- the first neural network 15 is configured to recognize the marker 1 within the image 13.
- the image 13 is analyzed in order to find sub-areas of the image plane in which the possible presence of markers is estimated.
- the first neural network 15 is configured to recognize at least one of: shape, color, pattern.
- the method then comprises one step of determination of the orientation of the marker 1 by means of a second neural network 16.
- the step of determination is performed in order to determine the arrangement of the marker 1 within the image 13, i.e. to determine whether it is rotated with respect to the readout position.
- the method comprises at least one step of cutting the marker 1 with respect to a background of the image 13 and at least one step of resizing of the cut-out marker 1.
- the cutting step is performed in such a way as to isolate the marker 1 from elements in the image 13 which are unrelated to the marker itself.
- the resizing step is performed so as to give all the cut-out markers 1 the same predefined size.
- the cutting step and the resizing step are performed by means of a cutting and resizing unit 17.
- the determination step comprises at least one step of locating at least one key point 18, 19, 20, 21 in the cut-out marker 1.
- the determination step comprises a locating step of a plurality of key points 18, 19, 20, 21.
- the key points 18, 19, 20, 21 comprise: a first key point 18 corresponding to the first angular portion 7; a second key point 19 corresponding to the second angular portion 8; a third key point 20 corresponding to the third angular portion 9; a fourth key point 21 corresponding to the fourth angular portion 10.
- each key point 18, 19, 20, 21 corresponds to a point of the marker 1 wherein the relevant angular portion 7, 8, 9, 10 meets the inner edge 12.
- the second neural network 16 is configured to determine the coordinates of the key points 18, 19, 20, 21 according to a Cartesian system. In other words, the second neural network 16 assigns the coordinates (x,y) to each key point 18, 19, 20, 21 so as to establish the relevant position thereof and to determine the possible rotation of the marker 1 with respect to the readout position.
- the method comprises a step of decoding of the marker 1 by means of a third neural network 22.
- the third neural network 22 is configured to convert the inner area 3 into a binary code 4.
- the method comprises a step of rotation of the cut-out marker 1 so that: the first key point 18 is positioned in the upper left corner; the second key point 19 is positioned in the upper right corner; the third key point 20 is positioned in the lower right corner; the fourth key point 21 is positioned in the lower left corner.
- the rotation step allows the marker 1 to be placed in the readout position.
- the method also comprises a step of rectification of the cut-out marker 1, by means of homographic transformation.
- the rectification step is performed so as to correct any distortion of the cut-out marker 1, caused by the perspective of the image 13, and to cut it again to remove those parts which are not relevant for the next decoding step.
- the rotation step and the rectification step are performed by means of a rotation and rectifying unit 23.
- the decoding step comprises at least one step of associating each white element 2 of the inner area 3 with a binary value of 1 and each black element 2 of the inner area 3 with a binary value of 0, in order to generate a binary code 4.
- the method Prior to the detection step of the marker 1 , the method comprises a training step of at least one of the first neural network 15, the second neural network 16 and the third neural network 22.
- the training step comprises: a first training step by means of an artificial dataset; and a second training step by means of the artificial dataset combined with a real dataset.
- the training step is performed for all three neural networks 15, 16, 22.
- the artificial dataset is automatically created starting from random images 13 on which a random number of markers 1 are superimposed with random positions, angles, perspective, illumination, saturation, transparency, etc., by means of image creation software.
- the neural networks 15, 16, 22 are, therefore, trained on this artificial dataset to obtain a first working version of the recognition and interpretation system.
- the real dataset is created from real images 13 where the real markers 1 have been printed, cut out and placed realistically in the scene.
- Semi-automatic annotation involves that given a real image 13, the position of such markers 1 and, in particular the position of the relevant key points 18, 19, 20, 21, is first estimated by the first version of the system and then corrected (if necessary) manually by a human annotator.
- the second training step is performed by means of the first version of the system on a final dataset obtained by the union of the artificial dataset and the real dataset, to obtain a final version of the recognition and interpretation system.
- the present invention also relates to a recognition and interpretation system of markers for artificial neural networks.
- the system comprises: at least one image acquisition unit 14 configured to acquire at least one image 13 containing at least one marker 1 as described above; and at least one processing and control unit 15, 16, 17, 22, 23 configured to perform the method as described above.
- the image acquisition unit 14 may comprise at least one of either a camera or a video camera.
- the image acquisition unit 14 may be of the fixed type, i.e., stably arranged in a predefined area, or of the movable type, such as a camera/video camera of a smart-phone, tablet, or the like.
- the processing and control unit 15, 16, 17, 22, 23 comprises: at least a first neural network 15 configured to perform the detection step; at least one cutting and resizing unit 17 configured to perform the cutting step and the resizing step; at least a second neural network 16 configured to perform the determination step; at least one rotation and rectifying unit 23 configured to perform the rotation step and the rectification step; and at least a third neural network 22 configured to perform the decoding step.
- the described invention achieves the intended objects and in particular the fact is underlined that the present marker, thanks to the presence of an outer edge provided with angular portions of different colors and of an inner edge in black color, is univocally recognizable inside an image by an artificial neural network.
- the marker, method and system according to the invention allow for fast and efficient decoding of the marker by an artificial neural network regardless of its placement within an image.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Electromagnetism (AREA)
- Health & Medical Sciences (AREA)
- Toxicology (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Processing (AREA)
- Radar Systems Or Details Thereof (AREA)
- Image Analysis (AREA)
Abstract
The marker (1) for artificial neural networks comprises a plurality of elements (2) defining a matrix structure and provided with one inner area (3) the elements (2) of which define a matrix code and with at least one outer edge (6) to the inner area (3) comprising four angular portions (7, 8, 9, 10) of different colors.
Description
MARKER FOR ARTIFICIAL NEURAL NETWORKS, RELATED COMPUTER-IMPLEMENTED METHOD FOR THE RECOGNITION AND INTERPRETATION AND RELATED SYSTEM
Technical Field
The present invention relates to a marker for artificial neural networks, a related computer-implemented method for the recognition and interpretation, and a related system.
Background Art
In the electronic industry, the use of markers, also called fiducial markers, is well known which are expressed as images and are recognized by electronic systems to provide information such as, e.g., in the case of augmented reality, a transition from 2D to 3D space.
The marker-based recognition is based on the detection, within an acquired image, of some characteristic patterns detectable according to the strong contrast of the edges (usually square or round black on a white background) compared to the rest of the image. On the basis of the representation of the edges and of the figure inside the edges it is possible to classify the markers in different subgroups. Usually the recognition of the edges within the image is achieved by distinguishing the areas of the image itself where the gradients of pixels change dramatically from one value to another. The discrimination between one marker and another is carried out on the basis of an encoding of the value enclosed between the identified edges.
Among the various types of fiducial markers are known, for example, the ArUco markers which are of the type of square binary markers.
The main problems ascertained in the use of markers are due to the fact that the identification and reading of the marker itself depends unequivocally on the quality of the acquired image.
In addition, the marker may be distorted by the perspective of the image and, therefore, difficult to recognize and/or decode.
It follows that the algorithms for their recognition must be programmed ad hoc in order to allow a rapid and effective decoding.
This issue is even more acute in the field of artificial neural networks.
An artificial neural network is a computational model composed of artificial “neurons”, loosely inspired by the simplification of a biological neural network and is used to solve Artificial Intelligence engineering problems related to different technological fields such as computer science, electronics or other subjects.
Neural networks cause computers to be able to solve problems independently and improve their capabilities.
To achieve this, the neural network is “trained” so that it can provide results that are as reliable as possible. A neural network must, therefore, be able to autonomously process the information related to the positioning and the distortion of the marker.
The difficulties encountered in the detection of markers in images make the markers themselves susceptible to improvement.
Description of the Invention
The main aim of the present invention is to devise a marker for artificial neural networks which is uniquely recognizable within an image by an artificial neural network.
Another object of the present invention is to devise a marker for artificial neural networks, a related computer-implemented method for the recognition and interpretation and a related system which enable rapid and efficient decoding of the marker by an artificial neural network regardless of its placement within an image.
Another object of the present invention is to devise a marker for artificial neural networks, a related computer-implemented method for the recognition and interpretation, and a related system which allow the mentioned drawbacks of the prior art to be overcome within a simple, rational, easy and effective to use as well as affordable solution.
The aforementioned objects are achieved by the present marker for artificial neural networks having the characteristics of claim 1.
Brief Description of the Drawings
Other characteristics and advantages of the present invention will become more apparent from the description of a preferred, but not exclusive, embodiment of a marker for artificial neural networks, a related computer-implemented method for the recognition and interpretation, and a related system, illustrated by way of an indicative, yet non-limiting example in the accompanying tables of drawings wherein:
Figures 1 and 2 are schematic representations of a marker according to the invention;
Figure 3 is a schematic representation illustrating the steps of the method according to the invention.
Embodiments of the Invention
With particular reference to these figures, reference numeral 1 globally indicates a marker for artificial neural networks.
In the context of the present disclosure, the expression “artificial neural network” refers to a computational model composed of artificial “neurons”, loosely inspired by the simplification of a biological neural network.
The marker 1 comprises a plurality of elements 2 defining a matrix structure and provided with at least one inner area 3 the elements 2 of which define a matrix code.
In particular, the inner area 3 comprises a plurality of elements 2 of white or black color.
Each of these elements 2 may be associated with a value of 0 or 1, depending on its color. In other words, the inner area 3 encodes a binary code 4 for the representation of a predefined piece of information.
The inner area 3 is decodable by a computer system capable of converting the binary code 4 into the aforementioned information.
According to the invention, the marker 1 comprises at least one outer edge 6 to the inner area 3 comprising four angular portions 7, 8, 9, 10 of different colors. The presence of an outer edge 6 with angular portions 7, 8, 9, 10 of different colors allows the marker 1 to result easily recognizable also by an artificial neural network.
An artificial neural network, in fact, is able to autonomously process the information and provide a univocal result. To achieve this result, the neural network must be “trained” so that it can provide results that are as reliable as possible.
The angular portions 7, 8, 9, 10 of different colors surrounding the inner area 3 uniquely identify the marker 1 and make the identification thereof extremely fast and unequivocal.
The outer edge 6 is also defined by a plurality of elements 2 arranged externally to the inner area 3.
The outer edge 6 has a thickness equal to one of the elements 2. In other words, the elements 2 of which the outer edge 6 is composed are arranged side by side on a single level.
It cannot, however, be ruled out that the outer edge 6 has a different thickness.
In the embodiment shown in the figures, the angular portions 7, 8, 9, 10 are each defined by three elements 2, arranged in an “L” pattern and forming the outer corners of the marker 1.
It cannot, however, be ruled out that the angular portions 7, 8, 9, 10 are composed of a different number of elements 2.
Usefully, the angular portions 7, 8, 9, 10 comprise: a first angular portion 7, the elements 2 of which are green; a second angular portion 8, the elements 2 of which are blue; a third angular portion 9, the elements 2 of which are white; a fourth angular portion 10, the elements 2 of which are red.
More in detail, colors are defined according to the RGB (Red Green Blue) model in which the color green is represented by the code (0, 255, 0), the color blue is represented by the code (0, 0, 255), the color white is represented by the code (255, 255, 255) and the color red is represented by the code (255, 0, 0).
The outer edge 6 also comprises a plurality of side portions 11 the elements 2 of which are white in color (RGB=(255, 255, 255)).
Each side portion 11 is bounded between two angular portions 7, 8, 9, 10.
In a predefined readout position, the inner area 3 represents a precise succession
of elements 2 which encode a particular code.
Should the marker 1 be positioned rotated with respect to the readout position and the inner area 3 be decoded according to this rotated position, it would inevitably provide an incorrect code and, consequently, incorrect information. The arrangement of the angular portions 7, 8, 9, 10 allows the orientation of marker 1 to be determined and the inner area 3 to be correctly decoded.
In the readout position: the first angular portion 7 is arranged in the upper left corner; the second angular portion 8 is arranged in the upper right corner; the third angular portion 9 is arranged in the lower right corner; the fourth angular portion 10 is arranged in the lower left corner.
The marker 1 also comprises at least one inner edge 12, positioned between the inner area 3 and the outer edge 6, the elements 2 of which are black (RGB=(0, 0, 0)).
The inner edge 12 allows the angular portions 7, 8, 9, 10 to be effectively recognized thanks to the chromatic contrast between the various colors of the latter and the black color of the inner edge itself.
The inner edge 12 in turn has a thickness equal to one of the elements 2. Similarly to the outer edge 6, the elements 2 of which the inner edge 12 is composed are arranged side by side on a single level.
It cannot, however, be ruled out that the inner edge 12 has a different thickness. Consequently, if the inner area 3 is of the type of an S x S matrix, the marker 1 has a matrix structure (S+4) x (S+4).
According to a further aspect, the present invention also relates to a computer- implemented method for the recognition and interpretation of a marker for artificial neural networks as described above.
The present method is, therefore, performed by means of a computerized recognition and interpretation system.
The method according to the invention comprises at least one step of acquisition of an image 13 containing at least one marker 1.
More in detail, the image 13 may contain a plurality of markers 1.
The acquisition of the image 13 is performed by means of an image acquisition unit 14.
The image acquisition unit 14 may comprise at least one of either a camera or a video camera.
The image acquisition unit 14 may be of the fixed type, i.e., stably arranged in a predefined area, or of the movable type, such as e.g. a camera/video camera of a smart-phone, tablet, or the like.
Next, the method comprises one step of detection of the marker 1 within the image 13 by means of a first neural network 15.
The first neural network 15 is configured to recognize the marker 1 within the image 13.
In particular, the image 13 is analyzed in order to find sub-areas of the image plane in which the possible presence of markers is estimated.
More in detail, the first neural network 15 is configured to recognize at least one of: shape, color, pattern.
The method then comprises one step of determination of the orientation of the marker 1 by means of a second neural network 16.
In more detail, the step of determination is performed in order to determine the arrangement of the marker 1 within the image 13, i.e. to determine whether it is rotated with respect to the readout position.
Conveniently, prior to the determination step, the method comprises at least one step of cutting the marker 1 with respect to a background of the image 13 and at least one step of resizing of the cut-out marker 1.
The cutting step is performed in such a way as to isolate the marker 1 from elements in the image 13 which are unrelated to the marker itself.
The resizing step is performed so as to give all the cut-out markers 1 the same predefined size.
The cutting step and the resizing step are performed by means of a cutting and resizing unit 17.
Once the display of the marker 1 has been optimized, the determination step is performed.
Advantageously, the determination step comprises at least one step of locating at least one key point 18, 19, 20, 21 in the cut-out marker 1.
More specifically, the determination step comprises a locating step of a plurality of key points 18, 19, 20, 21.
The key points 18, 19, 20, 21 comprise: a first key point 18 corresponding to the first angular portion 7; a second key point 19 corresponding to the second angular portion 8; a third key point 20 corresponding to the third angular portion 9; a fourth key point 21 corresponding to the fourth angular portion 10. Specifically, each key point 18, 19, 20, 21 corresponds to a point of the marker 1 wherein the relevant angular portion 7, 8, 9, 10 meets the inner edge 12.
The second neural network 16 is configured to determine the coordinates of the key points 18, 19, 20, 21 according to a Cartesian system. In other words, the second neural network 16 assigns the coordinates (x,y) to each key point 18, 19, 20, 21 so as to establish the relevant position thereof and to determine the possible rotation of the marker 1 with respect to the readout position.
Finally, the method comprises a step of decoding of the marker 1 by means of a third neural network 22.
The third neural network 22 is configured to convert the inner area 3 into a binary code 4.
Conveniently, prior to the decoding step, the method comprises a step of rotation of the cut-out marker 1 so that: the first key point 18 is positioned in the upper left corner; the second key point 19 is positioned in the upper right corner; the third key point 20 is positioned in the lower right corner; the fourth key point 21 is positioned in the lower left corner.
Substantially, the rotation step allows the marker 1 to be placed in the readout position.
The method also comprises a step of rectification of the cut-out marker 1, by means of homographic transformation.
The rectification step is performed so as to correct any distortion of the cut-out
marker 1, caused by the perspective of the image 13, and to cut it again to remove those parts which are not relevant for the next decoding step.
The rotation step and the rectification step are performed by means of a rotation and rectifying unit 23.
The decoding step comprises at least one step of associating each white element 2 of the inner area 3 with a binary value of 1 and each black element 2 of the inner area 3 with a binary value of 0, in order to generate a binary code 4.
Prior to the detection step of the marker 1 , the method comprises a training step of at least one of the first neural network 15, the second neural network 16 and the third neural network 22.
The training step comprises: a first training step by means of an artificial dataset; and a second training step by means of the artificial dataset combined with a real dataset.
In more detail, the training step is performed for all three neural networks 15, 16, 22.
The artificial dataset is automatically created starting from random images 13 on which a random number of markers 1 are superimposed with random positions, angles, perspective, illumination, saturation, transparency, etc., by means of image creation software.
These images 13 are annotated fully automatically, the positions of the markers 1 on the image plane and the coordinates of the relevant key points 18, 19, 20, 21 being known to the image creation software.
The neural networks 15, 16, 22 are, therefore, trained on this artificial dataset to obtain a first working version of the recognition and interpretation system.
The real dataset is created from real images 13 where the real markers 1 have been printed, cut out and placed realistically in the scene.
Then, there is a semi-automatic annotation of the real dataset, performed by means of the first version of the system.
Semi-automatic annotation involves that given a real image 13, the position of such markers 1 and, in particular the position of the relevant key points 18, 19,
20, 21, is first estimated by the first version of the system and then corrected (if necessary) manually by a human annotator.
Finally, the second training step is performed by means of the first version of the system on a final dataset obtained by the union of the artificial dataset and the real dataset, to obtain a final version of the recognition and interpretation system.
According to a third aspect, the present invention also relates to a recognition and interpretation system of markers for artificial neural networks.
The system comprises: at least one image acquisition unit 14 configured to acquire at least one image 13 containing at least one marker 1 as described above; and at least one processing and control unit 15, 16, 17, 22, 23 configured to perform the method as described above.
Specifically, the image acquisition unit 14 may comprise at least one of either a camera or a video camera.
The image acquisition unit 14 may be of the fixed type, i.e., stably arranged in a predefined area, or of the movable type, such as a camera/video camera of a smart-phone, tablet, or the like.
The processing and control unit 15, 16, 17, 22, 23 comprises: at least a first neural network 15 configured to perform the detection step; at least one cutting and resizing unit 17 configured to perform the cutting step and the resizing step; at least a second neural network 16 configured to perform the determination step; at least one rotation and rectifying unit 23 configured to perform the rotation step and the rectification step; and at least a third neural network 22 configured to perform the decoding step.
It has in practice been ascertained that the described invention achieves the intended objects and in particular the fact is underlined that the present marker, thanks to the presence of an outer edge provided with angular portions of different colors and of an inner edge in black color, is univocally recognizable
inside an image by an artificial neural network.
Furthermore, the marker, method and system according to the invention allow for fast and efficient decoding of the marker by an artificial neural network regardless of its placement within an image.
Claims
1) Marker (1) for artificial neural networks, comprising a plurality of elements (2) defining a matrix structure and provided with at least one inner area (3) the elements (2) of which define a matrix code, characterized by the fact that it comprises at least one outer edge (6) to said inner area (3) comprising four angular portions (7, 8, 9, 10) of different colors.
2) Marker (1) according to claim 1, characterized by the fact that said angular portions (7, 8, 9, 10) comprise: a first angular portion (7), the elements (2) of which are green; a second angular portion (8), the elements (2) of which are blue; a third angular portion (9), the elements (2) of which are white; a fourth angular portion (10), the elements (2) of which are red.
3) Marker (1) according to claim 2, characterized by the fact that: said first angular portion (7) is arranged in the upper left corner; said second angular portion (8) is arranged in the upper right corner; said third angular portion (9) is arranged in the lower right corner; said fourth angular portion (10) is arranged in the lower left corner.
4) Marker (1) according to one or more of the preceding claims, characterized by the fact that it comprises at least one inner edge (12), positioned between said inner area (3) and said outer edge (6), the elements (2) of which are black.
5) Marker (1) according to one or more of the preceding claims, characterized by the fact that said inner area (3) comprises a plurality of elements (2) of white or black color.
6) Computer-implemented method for the recognition and interpretation of a marker (1) according to one or more of the preceding claims, characterized by the fact that it comprises the following steps: acquisition of an image (13) containing at least one marker (1); detection of said marker (1) inside said image (13) by means of a first neural network (15); determination of the orientation of said marker (1) by means of a second
neural network (16); decoding of said marker (1) by means of a third neural network (22).
7) Method according to claim 6, characterized by the fact that, prior to said determination step, it comprises at least one step of cutting said marker (1) with respect to a background of said image (13) and at least one step of resizing of said cut-out marker (1).
8) Method according to claim 6 or 7, characterized by the fact that said determination step comprises at least one step of locating at least one key point (18, 19, 20, 21) in said cut-out marker (1), wherein: a first key point (18) corresponds to said first angular portion (7); a second key point (19) corresponds to said second angular portion (8); a third key point (20) corresponds to said third angular portion (9); a fourth key point (21) corresponds to said fourth angular portion (10).
9) Method according to one or more of claims 6 to 8, characterized by the fact that, prior to said decoding step, it comprises a step of rotation of said cut-out marker (1) so that: said first key point (18) is positioned in the upper left corner; said second key point (19) is positioned in the upper right corner; said third key point (20) is positioned in the lower right corner; said fourth key point (21) is positioned in the lower left corner.
10) Method according to one or more of claims 6 to 9, characterized by the fact that said decoding step comprises at least one step of associating each white element (2) of said inner area (3) with a binary value of 1 and each black element (2) of said inner area (3) with a binary value of 0, in order to generate a binary code (4).
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
IT102021000004982A IT202100004982A1 (en) | 2021-03-03 | 2021-03-03 | MARKER FOR ARTIFICIAL NEURAL NETWORKS, RELATED METHOD IMPLEMENTED BY COMPUTER RECOGNITION AND INTERPRETATION AND RELATED SYSTEM |
IT102021000004982 | 2021-03-03 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022185156A1 true WO2022185156A1 (en) | 2022-09-09 |
Family
ID=76159721
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB2022/051625 WO2022185156A1 (en) | 2021-03-03 | 2022-02-24 | Marker for artificial neural networks, related computer-implemented method for the recognition and interpretation and related system |
Country Status (2)
Country | Link |
---|---|
IT (1) | IT202100004982A1 (en) |
WO (1) | WO2022185156A1 (en) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120039529A1 (en) * | 2010-08-14 | 2012-02-16 | Rujan Entwicklung Und Forschung Gmbh | Producing, Capturing and Using Visual Identification Tags for Moving Objects |
US20190294936A1 (en) * | 2016-12-20 | 2019-09-26 | Universidad De Alicante | Method for detecting and recognizing long-range high-density visual markers |
US20190384954A1 (en) * | 2018-06-18 | 2019-12-19 | Abbyy Production Llc | Detecting barcodes on images |
-
2021
- 2021-03-03 IT IT102021000004982A patent/IT202100004982A1/en unknown
-
2022
- 2022-02-24 WO PCT/IB2022/051625 patent/WO2022185156A1/en unknown
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120039529A1 (en) * | 2010-08-14 | 2012-02-16 | Rujan Entwicklung Und Forschung Gmbh | Producing, Capturing and Using Visual Identification Tags for Moving Objects |
US20190294936A1 (en) * | 2016-12-20 | 2019-09-26 | Universidad De Alicante | Method for detecting and recognizing long-range high-density visual markers |
US20190384954A1 (en) * | 2018-06-18 | 2019-12-19 | Abbyy Production Llc | Detecting barcodes on images |
Also Published As
Publication number | Publication date |
---|---|
IT202100004982A1 (en) | 2022-09-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP4868530B2 (en) | Image recognition device | |
CN110400278B (en) | Full-automatic correction method, device and equipment for image color and geometric distortion | |
CN102667810A (en) | Face recognition in digital images | |
CN105009170A (en) | Object identification device, method, and storage medium | |
CN110472625B (en) | Chinese chess piece visual identification method based on Fourier descriptor | |
US20180253852A1 (en) | Method and device for locating image edge in natural background | |
CN110114781B (en) | Method for detecting and identifying remote high density visual indicia | |
CN108960280B (en) | Picture similarity detection method and system | |
CN103049731A (en) | Decoding method for point-distributed color coding marks | |
WO2021004414A1 (en) | Coding pattern, coding and reading methods for same, calibration board, and calibration method | |
CN112184826B (en) | Calibration plate and calibration method | |
CN113129397B (en) | Decoding method of parallelogram coding mark based on graphic geometric relation | |
CN112991517B (en) | Three-dimensional reconstruction method for texture image coding and decoding automatic matching | |
CN112270716B (en) | Decoding and positioning method for artificial visual landmarks | |
WO2022185156A1 (en) | Marker for artificial neural networks, related computer-implemented method for the recognition and interpretation and related system | |
CN116110069A (en) | Answer sheet identification method and device based on coding mark points and relevant medium thereof | |
CN109858474B (en) | Detection and identification method for transformer oil surface temperature controller | |
Tybusch et al. | Color-based and recursive fiducial marker for augmented reality | |
CN112219207A (en) | Method for optically identifying a marking | |
JP2014225754A (en) | Marker embedding device, marker detection device, marker embedding method, marker detection method, and program | |
CN116824183B (en) | Image feature matching method and device based on multiple feature descriptors | |
CN114222143B (en) | Encoding and decoding method based on tiny image | |
CN117095055A (en) | Robustness marking block based on hybrid tag and positioning method thereof | |
CN117437381A (en) | Self-adaptive rectangular jigsaw method and system based on multi-classification neural network and directivity description sub-algorithm | |
CN105654467A (en) | Shadow detection method based on color constancy characteristic and visual attention mechanism |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 22710439 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205 DATED 17.01.2024) |