WO2022265262A1 - Procédé d'extraction de données pour l'entraînement d'intelligence artificielle basé sur des mégadonnées, et programme informatique enregistré sur support d'enregistrement pour l'exécuter - Google Patents

Procédé d'extraction de données pour l'entraînement d'intelligence artificielle basé sur des mégadonnées, et programme informatique enregistré sur support d'enregistrement pour l'exécuter Download PDF

Info

Publication number
WO2022265262A1
WO2022265262A1 PCT/KR2022/007663 KR2022007663W WO2022265262A1 WO 2022265262 A1 WO2022265262 A1 WO 2022265262A1 KR 2022007663 W KR2022007663 W KR 2022007663W WO 2022265262 A1 WO2022265262 A1 WO 2022265262A1
Authority
WO
WIPO (PCT)
Prior art keywords
data
learning
artificial intelligence
image
annotation
Prior art date
Application number
PCT/KR2022/007663
Other languages
English (en)
Korean (ko)
Inventor
김도훈
Original Assignee
주식회사 인피닉
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 주식회사 인피닉 filed Critical 주식회사 인피닉
Publication of WO2022265262A1 publication Critical patent/WO2022265262A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/55Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/166Editing, e.g. inserting or deleting
    • G06F40/169Annotation, e.g. comment data or footnotes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/02Knowledge representation; Symbolic representation

Definitions

  • the present invention relates to the processing of artificial intelligence (AI) learning data. More specifically, it relates to a method for extracting artificial intelligence learning data that meets needs based on preformed big data and a computer program recorded on a recording medium to execute the method.
  • AI artificial intelligence
  • AI Artificial intelligence
  • machine learning refers to learning to optimize parameters with given data using a model composed of multiple parameters. Such machine learning is classified into supervised learning, unsupervised learning, and reinforcement learning according to the form of learning data.
  • designing data for artificial intelligence (AI) learning proceeds in the steps of data structure design, data collection, data refinement, data processing, data expansion, and data verification.
  • data structure design is performed through ontology definition, classification system definition, and the like.
  • Data collection is performed by collecting data through direct filming, web crawling, or associations/professional organizations.
  • Data purification is performed by removing redundant data from collected data and de-identifying personal information.
  • Data processing is performed by performing annotation and inputting metadata.
  • Data extension is performed by performing ontology mapping and supplementing or extending the ontology as needed.
  • data verification is performed by verifying validity according to the set target quality using various verification tools.
  • Data processing is also referred to as data labeling. More specifically, the annotation work of data processing is performed by a worker processing a bounding box for an object included in an image.
  • the bounding box is an area for specifying the position and shape of an object included in the image.
  • image information may include a file name, file size, and image size.
  • Copyright information may include information about the copyright holder of the image.
  • the photographing condition information may include resolution, bit value, aperture transmittance, exposure time, ISO sensitivity, focal length, aperture value, angle of view, white balance, and RGB depth.
  • Object information may include annotation type, coordinate value, area size, class name, tag item, clipping, major classification, middle classification, small classification, and instance information.
  • the environment information may include a shooting date and time, a shooting location, and weather information.
  • the number and type of information that can be input in the metadata input process of data processing may vary depending on the purpose of artificial intelligence (AI) learning.
  • environment information may further include the type of road on which an image is captured, road surface information, traffic congestion information, and the like.
  • One object of the present invention is to provide a method capable of extracting data for artificial intelligence (AI) learning that meets a specific purpose using metadata from big data.
  • AI artificial intelligence
  • Another object of the present invention is to provide a computer program recorded on a recording medium to execute a method capable of generating data for artificial intelligence (AI) learning that meets a specific purpose from big data.
  • AI artificial intelligence
  • the present invention proposes a method for generating artificial intelligence (AI) learning data that meets a specific purpose from big data using metadata.
  • the method includes the steps of distributing, by a learning data generating device, a plurality of images to be annotated;
  • a plurality of annotation devices specify the coordinates of an object included in each image for the distributed plurality of images, and generate coordinates of the specified object in the image and metadata for the image. and generating an annotation work result by including the generated metadata; and receiving, by the learning data generating device, the plurality of annotation job results, classifying the plurality of annotation job results based on the metadata, and generating data for artificial intelligence (AI) learning. can do.
  • big data is built using the annotation work result classified based on the metadata, and when a requested value is input from the outside, the big data is mined.
  • An annotation work result corresponding to the requested value may be extracted by mining or filtering, and the extracted annotation work result may be packaged to generate the artificial intelligence (AI) learning data.
  • the requested value may be a value indicating the type of information to be included in the AI learning data.
  • the big data includes a hash table composed of a plurality of buckets according to the type of the metadata, and each of the buckets includes a link to an annotation work result. It may be configured to include more than one slot (slot) of.
  • the step of generating the AI learning data is to search for one or more slots by substituting the requested value into a hash function previously set in correspondence with the hash table, and to search for one or more slots. You can extract one or more annotation work outputs through the included links.
  • the link included in the slot is maintained.
  • the capacity of the hash table may be increased by readjusting the home bucket of the slot.
  • the data structure is such that the type of information not included is included in the metadata.
  • the plurality of images may be redistributed to the plurality of annotation devices.
  • the big data is divided into a plurality of layers, and in the process of constructing the big data, based on the job evaluation grade given to the user of the annotation device, the A layer to which the annotation work result belongs may be determined, and the annotation work result may be extracted from each layer according to a ratio corresponding to the request value input from the outside.
  • AI artificial intelligence
  • the step of generating the annotation work result when the plurality of images correspond to images continuously photographed by the same camera, environmental information previously given to an image photographed in advance among the plurality of images is followed by It is possible to generate metadata for an image taken later by including it as environment information of a captured image.
  • the environment information may include information on the date and time of image capture, location, weather information, road type, road surface, and traffic congestion.
  • the step of generating the annotation work result if there are a plurality of previously photographed images, environment information previously assigned to an image including an object to which the same tracking identifier is assigned among the plurality of images is selected. It can be included as environment information of images taken later.
  • the tracking ID may be a unique identifier of an object assigned to track the object.
  • the step of generating the annotation work result is based on a result of comparing the background area obtained by removing the area occupied by the object to which the same ID is assigned for each of the preceding and succeeding images, It may be determined whether or not the captured image can be applied as the environment information of the captured image by following the environment information given to the captured image.
  • the present invention proposes a computer program recorded on a recording medium to execute the verification method as described above.
  • the computer program may include a memory; transceiver; and a processor configured to process instructions resident in the memory.
  • the computer program may further include distributing, by the processor, a plurality of images to be annotated through the transceiver; receiving, by the processor, the plurality of annotation job results through the transceiver, the annotation job result including coordinates of an object in the image for each of the plurality of images and metadata for the image; and generating data for artificial intelligence (AI) learning by the processor classifying the plurality of annotation work results based on the metadata, which may be a computer program recorded on a recording medium.
  • AI artificial intelligence
  • AI learning data when AI learning data is requested from a customer who wants to learn AI, after that, data structure design, data collection, data refinement, data processing, It took a relatively long time to expand data, verify data, generate artificial intelligence (AI) learning data based on the results, and deliver it to customers.
  • AI artificial intelligence
  • AI artificial intelligence
  • FIG. 1 and 2 are configuration diagrams of an artificial intelligence learning system according to various embodiments of the present invention.
  • FIG. 3 is a logical configuration diagram of an apparatus for generating learning data according to an embodiment of the present invention.
  • FIG. 4 is a logical configuration diagram of an annotation device according to an embodiment of the present invention.
  • FIG. 5 is a hardware configuration diagram of an annotation device according to an embodiment of the present invention.
  • FIG. 6 is a signal flow diagram for explaining a method of generating data for artificial intelligence (AI) learning according to an embodiment of the present invention.
  • FIG. 7 is a conceptual diagram for explaining the concept of generating data for artificial intelligence (AI) learning according to an embodiment of the present invention.
  • FIG. 8 is an exemplary view for explaining the concept of a tracking ID according to an embodiment of the present invention.
  • FIG. 9 is an exemplary diagram for explaining a user interface (UI) for checking a tracking ID according to an embodiment of the present invention.
  • UI user interface
  • 10 and 11 are exemplary diagrams for explaining a process of determining environment information in consecutive images according to some embodiments of the present invention.
  • 12 and 13 are exemplary diagrams for explaining a process of automatically determining environment information according to some embodiments of the present invention.
  • UI 14 is an exemplary diagram for explaining a user interface (UI) for inspecting metadata according to an embodiment of the present invention.
  • 15 is a flowchart illustrating an operation process of an annotation device according to an embodiment of the present invention.
  • first and second used in this specification may be used to describe various components, but the components should not be limited by the terms. These terms are only used for the purpose of distinguishing one component from another. For example, a first element may be termed a second element, and similarly, a second element may be termed a first element, without departing from the scope of the present invention.
  • AI artificial intelligence
  • the present invention intends to propose various means capable of more easily processing a large number of images and more easily verifying the processed data.
  • the artificial intelligence learning system includes a learning data generating device 100, one or more annotation devices 200-1, 200-2, ..., 200-n; 200 , It may be configured to include a learning data verification device 300 and an artificial intelligence learning device 400.
  • the artificial intelligence learning system includes one or more annotation devices (200-a, 200-b, ..., 200-m; 200) and a learning data verification device ( 300-a, 300-b, ..., 300-m; 300) consisting of a plurality of groups (Group-a, Group-b ..., Group-m), the learning data generating device 100 and the artificial intelligence learning device ( 400) may be configured.
  • components of the artificial intelligence learning system are merely functionally distinct elements, two or more components are integrated and implemented in the actual physical environment, or one component is implemented in the actual physical environment. may be implemented separately from each other.
  • the learning data generating device 100 is a device that can be used to design and generate artificial intelligence (AI) learning data.
  • AI artificial intelligence
  • the learning data generating device 100 is basically a device that is distinguished from the learning data verifying device 300, but in an actual physical environment, the learning data generating device 100 and the learning data verifying device 300 are integrated into one device. It may be integrated and implemented.
  • the learning data generating device 100 preemptively generates artificial intelligence (AI) data before receiving a request for artificial intelligence (AI) learning data from the artificial intelligence learning device 400. It is possible to build big data that can generate data for learning.
  • AI artificial intelligence
  • the learning data generating device 100 receives a request value according to the customer's needs from the artificial intelligence learning device 400, it mines or filters the preemptively built big data to generate artificial intelligence (AI). Data for learning may be generated, and the generated data for artificial intelligence (AI) learning may be transmitted to the artificial intelligence learning device 400 .
  • AI artificial intelligence
  • the learning data generation device 100 having such characteristics transmits and receives data with the annotation device 200, the learning data verification device 300, and the artificial intelligence learning device 400, and performs calculations based on the transmitted and received data. Any device that can do this is acceptable.
  • the learning data generating device 100 may be any one of a fixed computing device such as a desktop, workstation, or server, but is not limited thereto.
  • a detailed configuration and operation of the learning data generating device 100 will be described later with reference to FIG. 3 .
  • the annotation device 200 is a device that can be used to annotate images distributed by the learning data generating device 100.
  • the annotation device 200 tracks a specific object from each image distributed by the learning data generating device 100 (ie, an annotation target image).
  • a tracking identifier can be assigned.
  • the annotation device 200 may automatically assign environment information to each image distributed by the learning data generating device 100.
  • the annotation device 200 having such characteristics may be any device capable of transmitting and receiving data to and from the learning data generating device 100 and the learning data verifying device 300 and performing calculations based on the transmitted and received data.
  • the annotation device 100 may be a stationary computing device such as a desktop, workstation, or server, or a smart phone, a laptop, a tablet, a phablet, or a portable multimedia player.
  • PMP Portable Multimedia Player
  • PDAs personal digital assistants
  • E-book reader e-book readers
  • annotation device 200 may be a device in which an annotation worker performs annotation through a clouding service.
  • the learning data verification device 300 is a device that can be used to verify artificial intelligence (AI) learning data. That is, the learning data verification device 300 verifies whether the annotation task result generated by the annotation device 200 meets a pre-set target quality or whether the annotation task result is valid for artificial intelligence (AI) learning. It is a device that can
  • the learning data verification device 300 may receive an annotation job result from the annotation device 200 .
  • the annotation work result may include coordinates of an object specified by an annotation worker and metadata about an image or object.
  • the metadata of the annotation work result may include image information, copyright information, photographing condition information, specific object information, and environment information, but is not limited thereto.
  • Such an annotation work result may have a JSON (Java Script Object Notation) file format, but is not limited thereto.
  • the learning data verification device 300 may inspect the received annotation work result. To this end, the learning data verification device 300 may perform inspection using a script for annotation work results.
  • the script is a code for verifying whether or not the target quality previously set for the annotation work result meets or whether the data is valid.
  • the learning data verification device 300 can provide a user interface (UI) that can more easily inspect the tracking ID and environment variables of objects included in the annotation work result. .
  • UI user interface
  • the learning data verification apparatus 300 outputs a user interface (UI) capable of verifying the metadata of the annotation work result, a cropped image of an object to which the same tracking ID is assigned from images included in the image group.
  • UI user interface
  • the cropped images collected on the user interface (UI) may be sequentially arranged and output according to the order of the images taken.
  • the image group is a set of annotated images. Such image groups may be formed for each annotation device 200 or for each project.
  • the image group includes images taken consecutively in time series by the same camera as the camera that took the image to be annotated, and other images arranged adjacent to the camera that took the image to be annotated. Images spatially continuously photographed by a camera may be included, but are not limited thereto.
  • the cropped image is an image obtained by cropping a part of an image that is an annotation work target based on coordinates set to specify an object by annotation.
  • the learning data verification apparatus 300 compares patterns of RGB (Red, Green, Blue) values and patterns of edges of objects included in each of the cropped images to crop images of objects to which the same tracking ID is assigned. In cross-contrast, it is possible to self-inspect whether the same tracking ID is assigned to each object.
  • RGB Red, Green, Blue
  • the learning data verification apparatus 300 cross-contrasts the RGB value pattern and the edge pattern of objects to which different tracking IDs have been assigned, targeting cropped images of objects to which different tracking IDs have been assigned. It is also possible to check whether different tracking IDs are assigned to each other.
  • the pattern of RGB values means a pattern of a ratio of RGB values to pixels in an image and an arrangement order of RGB values.
  • the pattern of the edge means a pattern for the number of pixels included in an enclosure closed by the edge extracted from the image and the relative position where the edge is located in the image.
  • the learning data verification device 300 determines that it is incorrect to assign the same tracking ID or to assign different tracking IDs
  • the tracking ID that is determined to be incorrect is displayed on the user interface (UI).
  • Information about the assigned object can be output.
  • the learning data verification device 300 outputs a user interface (UI) in which cropped images of a plurality of objects to which the same tracking ID is assigned are successively arranged, and the result of self-checking whether the tracking ID of the objects is correctly assigned
  • UI user interface
  • the user of the training data verification device 300 can easily check and inspect a plurality of objects determined to be the same at a glance.
  • the learning data verification apparatus 300 in outputting a user interface (UI) capable of verifying the metadata of the annotation work result, the learning data verification apparatus 300 has a figure having a preset color corresponding to a value included in each metadata. ), each of the images included in the image group may be graphicalized and included in the user interface (UI).
  • UI user interface
  • the learning data verification apparatus 300 may express each of the images included in the image group as a single line having a color corresponding to a value included in metadata.
  • the training data verification apparatus 300 may construct bar-shaped graphics by continuously arranging each line segment in the direction of the width of the line segment according to the sequence of images captured.
  • the training data verification device 300 may output the configured bar-shaped graphic by including it in the user interface (UI).
  • UI user interface
  • the learning data verification apparatus 300 outputs a user interface (UI) in which a plurality of images are simply graphic according to the value of the metadata, so that the user of the learning data verification apparatus 300 can view the metadata of the plurality of images.
  • UI user interface
  • the learning data verification device 300 may transmit annotation work results and inspection results received from the annotation devices 200 to the learning data generating device 100 .
  • the learning data verification device 300 having the above characteristics can transmit and receive data with the annotation device 200 and the learning data generating device 100 and perform calculations based on the transmitted and received data. Any device may be acceptable.
  • the learning data verification device 300 may be any one of a fixed computing device such as a desktop, workstation, or server, but is not limited thereto.
  • the artificial intelligence learning device 400 is a device that can be used to develop artificial intelligence (AI).
  • the artificial intelligence learning device 400 provides the learning data generating device 100 with a request value including requirements that AI learning data must satisfy in order for AI to achieve its development purpose. can transmit
  • the artificial intelligence learning device 400 may receive artificial intelligence (AI) learning data from the learning data generating device 100 .
  • the artificial intelligence learning device 400 may perform machine learning on artificial intelligence (AI) to be developed using the received artificial intelligence (AI) learning data.
  • the artificial intelligence learning device 400 may be any device capable of transmitting and receiving data to and from the learning data generating device 100 and performing calculations using the transmitted and received data.
  • the artificial intelligence learning device 400 may be any one of a fixed computing device such as a desktop, workstation, or server, but is not limited thereto.
  • the learning data generating device 100, one or more annotation devices 200, the learning data verification device 300, and the artificial intelligence learning device 400 are directly connected to each other through a security line and a public wired communication network.
  • data may be transmitted and received using a network in which one or more of the mobile communication networks are combined.
  • public wired communication networks may include Ethernet, x Digital Subscriber Line (xDSL), Hybrid Fiber Coax (HFC), and Fiber To The Home (FTTH). It may be, but is not limited thereto.
  • xDSL Digital Subscriber Line
  • HFC Hybrid Fiber Coax
  • FTTH Fiber To The Home
  • CDMA Code Division Multiple Access
  • WCDMA Wideband CDMA
  • HSPA High Speed Packet Access
  • LTE Long Term Evolution
  • 5th generation mobile telecommunication may be included, but is not limited thereto.
  • FIG. 3 is one of the present invention in the examples It is a logical configuration diagram of the learning data generating device according to FIG.
  • the learning data generating device 300 includes a communication unit 105, an input/output unit 110, a storage unit 115, a data structure design unit 120, a data collection and refinement unit 125, and a data It may include a processing unit 130, a big data management unit 135, and a learning data generation unit 140.
  • the components of the learning data generating device 100 are merely functionally distinct elements, two or more components are integrated and implemented in an actual physical environment, or one component is integrated with each other in an actual physical environment. It could be implemented separately.
  • the communication unit 105 may transmit/receive data with one or more of the annotation device 200, the learning data verification device 300, and the artificial intelligence learning device 400.
  • the communication unit 105 may transmit a plurality of images to be annotated to the annotation device 200 .
  • the communication unit 105 may receive annotation work results and inspection results from the learning data verification device 300 .
  • the communication unit 105 may receive a request value from the artificial intelligence learning device 400 .
  • the required value is a value related to requirements that AI learning data must satisfy in order for AI to achieve its development purpose. That is, the requested value is a value indicating the type of information to be included in the data for AI learning.
  • the communication unit 105 may transmit artificial intelligence (A) learning data to the artificial intelligence learning device 400 .
  • the input/output unit 110 may receive a signal from a user through a user interface (UI) or output an operation result to the outside.
  • UI user interface
  • the input/output unit 110 may receive a control signal for designing a data structure for artificial intelligence (AI) learning from a user.
  • the input/output unit 110 may receive data (ie, images) for artificial intelligence (AI) learning from the outside.
  • the input/output unit 110 may receive an input of an allocation amount for distributing annotation work to a plurality of annotation devices 200 from a user.
  • the input/output unit 110 may output state information of the big data managed by the big data management unit 135 . Also, the input/output unit 110 may output the requested value received from the artificial intelligence learning device 400 through the communication unit 105 .
  • the storage unit 115 may store data for generating artificial intelligence (AI) learning data.
  • AI artificial intelligence
  • the storage unit 115 may store a data structure designed by the data structure design unit 120 .
  • the storage unit 115 may store data (ie, images) for artificial intelligence (AI) learning collected by the data collection and refinement unit 125 .
  • the storage unit 115 may store annotation work results received from the annotation device 200 or the learning data verification device 300 .
  • the storage unit 115 may store information about big data being generated and managed by the big data management unit 135 .
  • the storage unit 115 according to another embodiment of the present invention may store information about big data being generated and managed by the big data management unit 135 through an external storage device.
  • the data structure design unit 120 may design a data structure for artificial intelligence (AI) learning.
  • AI artificial intelligence
  • the data structure design unit 120 may design a data structure for artificial intelligence (AI) learning based on user control input through the input/output unit 110 .
  • the data structure design unit 120 may define an ontology for artificial intelligence (AI) learning and a data classification system for artificial intelligence (AI) learning based on user control.
  • AI artificial intelligence
  • the user of the learning data generating device 100 can control the structure of data for artificial intelligence (AI) learning to have versatility in consideration of the needs of the main business partner or customer.
  • the data collection and refinement unit 125 may collect and refine data for artificial intelligence (AI) learning based on the data structure designed by the data structure design unit 120 .
  • AI artificial intelligence
  • the data collection and refinement unit 125 may receive an image from the outside through the input/output unit 110 based on the designed data structure.
  • the data collection and refinement unit 125 may collect images by performing web crawling through the communication unit 105 .
  • the data collection and refinement unit 125 may download an image from an external institution's device having image data.
  • the data collection and refinement unit 125 may remove redundant or extremely similar images from among collected images. Also, the data collection and refinement unit 125 may de-identify personal information included in the collected images.
  • the data processing unit 130 may process the collected and refined data (ie, the image to be annotated) using the annotation device 200 .
  • the data processing unit 130 distributes the collected and refined data (ie, images to be annotated) to the plurality of annotation devices 200 based on the quota set through the input/output unit 110 can be sent.
  • the data processing unit 130 may receive annotation work results directly from the annotation device 200 or receive annotation work results and inspection results from the learning data verification device 300 through the communication unit 105 .
  • the big data management unit 135 may construct and manage big data related to a plurality of annotation work results based on metadata.
  • the big data management unit 135 performs a plurality of annotation tasks received from the annotation device 200 or the learning data verification device 300 through the data processing unit 130 based on the metadata included in the annotation task result. Results can be categorized.
  • the metadata included in the annotation work result may include image information, copyright information, shooting condition information, annotated object information, and environment information, but is not limited thereto.
  • Image information included in metadata may include a file name, file size, and image size.
  • Copyright information may include information about the copyright holder of the image.
  • the photographing condition information may include resolution, bit value, aperture transmittance, exposure time, ISO sensitivity, focal length, aperture value, angle of view, white balance, and RGB depth.
  • Object information may include annotation type, coordinate value, area size, class name, tag item, clipping, major classification, middle classification, small classification, and instance information.
  • the environmental information may include shooting date and time, shooting location, weather information, road type, road surface information, and traffic jam information.
  • the number and type of information that may be included in metadata is not limited thereto.
  • the big data management unit 135 may build big data using a plurality of annotation work results classified based on metadata.
  • the big data management unit 135 may include a plurality of annotation work results received after the big data is pre-constructed in the pre-constructed big data.
  • the big data constructed by the big data management unit 135 may include a hash table.
  • a hash table included in big data may be composed of a plurality of buckets according to the type of metadata.
  • Each of the plurality of buckets constituting the hash table may include one or more slots.
  • each slot included in the bucket may include a link to an annotation work result.
  • the big data management unit 1335 in the process of constructing big data, when the load density of the hash table exceeds a preset critical density, maintains the links included in each slot and stores the metadata.
  • the capacity of the hash table can be increased by doubling the number of buckets according to the type of , and then readjusting only the home bucket of each slot.
  • the big data constructed by the big data management unit 135 may be partitioned into a plurality of layers.
  • a layer of big data may be determined, and a corresponding annotation work result may be added to the determined layer of big data.
  • the learning data generating unit 140 may generate artificial intelligence (AI) learning data using the big data built by the big data management unit 135 .
  • AI artificial intelligence
  • the learning data generator 140 mines or filters the built big data to obtain the requested value. Annotation work results corresponding to can be extracted.
  • the required value is a value related to requirements that AI learning data must satisfy in order for AI to achieve its development purpose. That is, the requested value is a value indicating the type of information to be included in the data for AI learning.
  • the learning data generating unit 140 may search for one or more slots by substituting a requested value into a hash function set in advance corresponding to a hash table of big data.
  • the training data generator 140 may extract one or more annotation work results through links respectively included in one or more searched slots.
  • the learning data generation unit 140 may extract annotation work results from each layer of the big data composed of a plurality of layers according to a preset ratio corresponding to the requested value.
  • the learning data generation unit 140 may generate AI learning data by packaging the extracted annotation work result. And, the learning data generation unit 140 may transmit the generated artificial intelligence (AI) learning data to the artificial intelligence learning device 400 through the communication unit 105 .
  • AI artificial intelligence
  • the learning data generator 140 allows the type of information not included in the metadata to be included in the metadata,
  • the data structure may be redesigned through the data structure design unit 120 .
  • the learning data generation unit 140 may redistribute the plurality of images to the plurality of annotation devices 200 through the data processing unit 130 to redistribute the annotation work.
  • FIG. 4 is a logical configuration diagram of an annotation device according to an embodiment of the present invention.
  • the annotation device 200 includes a communication unit 205, an input/output unit 210, a storage unit 215, an object specification unit 220, a metadata generation unit 225, and a result generation unit ( 230) may be configured.
  • the components of the annotation device 200 are merely functionally distinct elements, two or more components are integrated and implemented in an actual physical environment, or one component is separated from each other in an actual physical environment. could be implemented.
  • the communication unit 205 may transmit/receive data with the learning data generating device 100 and the learning data verifying device 300 .
  • the communication unit 205 may receive a plurality of images distributed by the learning data generating device 100 .
  • the image is an image that is a target of annotation work for artificial intelligence (AI) learning.
  • AI artificial intelligence
  • Such a plurality of images may be received individually or collectively.
  • the communication unit 205 may transmit the annotation work result to the learning data verification device 300 .
  • Such an annotation work result may have a JSON file format, but is not limited thereto.
  • the input/output unit 210 may receive a signal from a user through a user interface (UI) or output an operation result to the outside.
  • UI user interface
  • the user means a person who performs annotation work.
  • Such a user may be referred to as a worker, performer, labeler, data labeler, and the like, but is not limited thereto.
  • the input/output unit 210 may output an image to be an annotation work.
  • the input/output unit 110 may receive a control signal for setting a bounding box from a user. Also, the input/output unit 110 may overlay and output a bounding box on the image.
  • the input/output unit 110 may receive a control signal for setting metadata of an image from a user.
  • the storage unit 215 may store data required for annotation work.
  • the storage unit 215 may store an image received through the communication unit 205 . Also, the storage unit 215 may store project properties, image properties, or user properties received through the communication unit 205 .
  • the object specifying unit 220 may specify an object included in an image to be annotated.
  • the object specifying unit 220 may load an image to be an annotation work into memory.
  • the object specifying unit 220 may select a tool according to a user's control signal input through the input/output unit 210 .
  • the tool is a tool for setting a bounding box that specifies one or more objects included in the image.
  • the object specifying unit 220 may receive coordinates through the selected tool according to a control signal input through the input/output unit 210 .
  • the object specifying unit 220 may specify an object included in the image by setting a bounding box based on the input coordinates.
  • the bounding box is an area for specifying an object to be learned by artificial intelligence (AI) among objects included in the image.
  • AI artificial intelligence
  • Such a bounding box may have a rectangle or polygon shape, but is not limited thereto.
  • the object specification unit 220 receives two coordinates from the user through the input/output unit 210, and has the input two coordinates as the coordinates of the upper left vertex and the coordinates of the lower right vertex in the image.
  • An object included in an image can be specified by setting a bounding box based on a rectangle.
  • the two coordinates may be set by the user inputting one type of input signal twice (eg, mouse click) or by the user inputting two types of input signal once (eg, mouse drag). It may, but is not limited thereto.
  • the metadata generating unit 225 may generate metadata for an image that is a target of annotation work.
  • the metadata may include image information, copyright information, shooting condition information, annotated object information, and environment information, but is not limited thereto.
  • Image information included in metadata may include a file name, file size, and image size.
  • Copyright information may include information about the copyright holder of the image.
  • the photographing condition information may include resolution, bit value, aperture transmittance, exposure time, ISO sensitivity, focal length, aperture value, angle of view, white balance, and RGB depth.
  • Object information may include annotation type, coordinate value, area size, class name, tag item, clipping, major classification, middle classification, small classification, and instance information.
  • the environmental information may include shooting date and time, shooting location, weather information, road type, road surface information, and traffic jam information.
  • the number and type of information that may be included in metadata is not limited thereto.
  • the metadata generator 225 may set specific values to be included in the metadata according to a control signal input from the user, but characteristically, the metadata generator 225 responds to the user's control signal. It is possible to automatically determine or assign the object's tracking ID and environment information without relying on it. In other words, the metadata generator 225 may generate metadata by including coordinates in the image of the specified object, an automatically assigned tracking ID, and automatically determined environment information.
  • the metadata generation unit 225 may automatically assign a tracking ID to the object specified by the object specification unit 220 .
  • the tracking ID is a unique identifier of an object assigned to track an object from an image group.
  • An image group is a set of annotated images.
  • the image group includes images taken consecutively in time series by the same camera as the camera that took the image to be annotated, and other images arranged adjacent to the camera that took the image to be annotated. Images spatially continuously photographed by a camera may be included, but are not limited thereto.
  • the metadata generating unit 225 may assign a tracking ID based on the color of the object. Specifically, the metadata generating unit 225 may identify a pattern of RGB values for an area occupied by an object in an image. The metadata generation unit 225 may generate an identifier corresponding to the identified pattern of RGB values and assign it as the tracking ID of the object specified by the object specification unit 220 .
  • the pattern of RGB values means a pattern of ratios of RGB values to pixels and arrangement order of RGB values.
  • the metadata generating unit 225 before generating a new identifier, determines the pattern of RGB values for objects specified from the images included in the image group and the object specified by the object specifying unit 220. Similar objects may be searched for by comparing the identified patterns of RGB values. Also, when a similar object is found from the images included in the image group, the metadata generator 225 assigns the tracking ID assigned to the searched similar object to the tracking ID of the object specified by the object specifying unit 220. can do.
  • the metadata generator 225 may assign a tracking ID based on the shape of an object. Specifically, the metadata generating unit 225 may extract an edge of an image (edge detection). The metadata generation unit 225 may identify only the edge of the object specified by the object specification unit 220 from all edges of the extracted image. The metadata generating unit 225 may generate an identifier corresponding to the edge pattern of the identified object and assign it as the tracking ID of the object specified by the object specifying unit 220 .
  • the pattern of the edge means a pattern for the number of pixels included in the area closed by the edge extracted from the image and the relative position of the edge within the image.
  • the metadata generation unit 225 identifies edge patterns for objects specified from the images included in the image group and the object identified by the object specification unit 220 before newly generating the identifier. Similar objects may be searched for by comparing the edge patterns of objects with each other. Also, when a similar object is found from the images included in the image group, the metadata generator 225 assigns the tracking ID assigned to the searched similar object to the tracking ID of the object specified by the object specifying unit 220. can do.
  • the metadata generator 225 may assign a tracking ID based on symbols or characters included in objects. Specifically, the metadata generating unit 225 may identify whether symbols or characters are included in the area occupied by the object in the image. The metadata generation unit 225 targets objects specified from the images included in the image group, and similar objects including the same symbol or character as the symbol or character identified from the object specified by the object specifying unit 220. can be searched for. Also, when a similar object is found from the images included in the image group, the metadata generator 225 assigns the tracking ID assigned to the searched similar object to the tracking ID of the object specified by the object specifying unit 220. can do.
  • the metadata generating unit 225 may automatically determine environment information about an image that is a target of annotation work.
  • the environment information refers to information related to a place or time point at which an image was captured by a camera.
  • Such environmental information may include, but is not limited to, information on the date and time of image capture, location, weather information, road type, road surface, and traffic congestion. It doesn't.
  • the metadata generation unit 225 compares an existing image with an image to be annotated, and converts environmental information given to the existing image to the environment of the image to be annotated. information can be applied.
  • the metadata generation unit 225 compares the existing image with the image to be annotated, and determines whether environmental information given to the existing image can be applied as environmental information of the image to be annotated. can determine whether
  • the existing image means an image taken prior to the image to be annotated by the same camera as the camera that captured the image to be annotated.
  • the metadata generation unit 225 assigns the same tracking ID as the object specified by the object specification unit 220 among the plurality of images. You can select an image that contains an object that has been deleted as an existing image.
  • the metadata generation unit 225 selects an image whose shooting time is closest to the image to be annotated from among the plurality of images. You can also choose by image.
  • the metadata generating unit 225 may identify a background area occupied by an object to which the same tracking ID as an object of the existing image is removed from the image to be annotated.
  • the metadata generating unit 225 may identify a background area occupied by an object to which the same tracking ID as the object of the image to be annotated is removed from the selected existing image.
  • the metadata generation unit 225 converts environment information pre-assigned to the existing image to a target of annotation based on a result of comparing the background area of the image to be annotated with the background area of the existing image. It can be determined whether or not it is applicable as environment information of the image to be applied.
  • the metadata generating unit 225 is an area occupied by an object to which the same ID has been assigned to each of the previously photographed image (i.e., the previous image) and the subsequent photographed image (annotation target image).
  • the background areas from which are removed are compared with each other, and based on the comparison result, it is determined whether or not the environmental information previously assigned to the previously photographed image can be applied as the environment information of the subsequent photographed image.
  • the metadata generation unit 225 determines the existing image based on the pattern of RGB values for the background area of the image to be annotated and the pattern of RGB values for the background area of the existing image. It is possible to determine whether pre-given environmental information can be applied as environmental information of an image to be annotated.
  • the metadata generation unit 225 is based on the pattern of the edge of the background area of the image to be annotated and the pattern of the edge of the background area of the existing image, the environment given to the existing image. It is possible to determine whether the information can be applied as environmental information of an image to be annotated.
  • the metadata generator 225 may set the average brightness value or average gamma value of the background area of the image to be annotated, and the average brightness value or average brightness value of the background area of the existing image. Based on the average gamma value, it may be determined whether environmental information given to an existing image can be applied as environmental information of an image to be annotated.
  • the metadata generation unit 225 when it is determined that the metadata generation unit 225 can apply the environmental information previously assigned to the existing image as the environmental information of the image to be annotated, the metadata generation unit 225 annotates the environmental information previously assigned to the existing image. Metadata can be created by including it as the environment information of the target image.
  • the metadata generation unit 225 assigns previously given environment information to the preceding image among the plurality of images to be photographed. By including it as environment information of the image, it is possible to automatically generate metadata for images taken later.
  • the metadata generation unit 225 cannot apply the environment information already given to the existing image as the environmental information of the image to be annotated, one or more sets in the image to be annotated. Based on the characteristics of the reference point, environmental information about an image to be annotated can be automatically determined.
  • the metadata generating unit 225 may set one or more reference points in an image to be annotated.
  • a number of images used in machine learning have a relatively similar composition so that artificial intelligence (AI) can achieve a specific purpose.
  • a reference point according to an embodiment of the present invention may be one or more points in an image set to correspond to the learning purpose of artificial intelligence (AI).
  • the metadata generator 225 may determine environment information of an image to be annotated based on the color, brightness, or gamma of each pixel of one or more set reference points.
  • the metadata generator 225 may extract an edge of an image to be annotated, and divide the image into a plurality of regions based on the extracted edge.
  • the metadata generating unit 225 may set one or more reference points for each of the plurality of divided areas.
  • the metadata generation unit 225 may identify a background area by removing an area occupied by an object from an image to be annotated based on the extracted edge.
  • the metadata generating unit 225 may determine the orientation of the object in the image based on the class assigned to the object. For example, when the class assigned to the object is a vehicle, the metadata generation unit 225 may determine the direction in which the wheels of the object are heading in the image as the ground.
  • the metadata generator 225 may identify a ground area and an air area from among a plurality of areas included in the identified background area, based on the determined object directionality.
  • the ground area is an area closest to the ground among a plurality of areas divided by edges.
  • the aerial area is an area furthest from the ground among a plurality of areas divided by edges.
  • the metadata generation unit 225 determines road surface information constituting the environment information based on the pixels of the reference point set in the ground area, and determines the environment information based on the pixels of the reference point set in the air area. It is possible to determine constituting weather information.
  • the metadata generator 225 may set one or more reference points for each of the four line segments constituting the border of the image to be annotated.
  • the metadata generating unit 225 may determine the orientation of the object in the image based on the class assigned to the object.
  • the metadata generation unit 225 may identify a ground segment and an air segment among four line segments based on the determined object direction.
  • the ground segment is a line segment closest to the ground among four line segments constituting the edge of the image.
  • the mid-air segment is the farthest line segment from the ground among the four line segments constituting the edge of the image.
  • the metadata generator 225 determines road surface information constituting environmental information based on pixels of a reference point set in a ground segment, and weather information constituting environmental information based on pixels of a reference point set in an aerial segment can decide
  • the metadata generation unit 225 sets a reference point at the same location for each of the image to be annotated and the existing image included in the image group, and the pixel of the reference point set in the image to be annotated is compared with the existing image. Contrasting the pixels of reference points identically set in the image, if the color, brightness, and gamma values of the pixels are within the preset allowable range, the preset environment information of the existing image will be determined as the environment information of the image to be annotated. may be If the difference between the color, brightness, and gamma values of the pixels of the reference point set identically in the image to be annotated and the existing image is out of the allowable range, the metadata generator 225 sets the environment information manually. A message informing that it should be done may be output through a user interface (UI).
  • UI user interface
  • the result generation unit 230 may generate an annotation work result based on the metadata generated by the metadata generation unit 225 .
  • the annotation work result may have a JSON file format, but is not limited thereto.
  • the result generation unit 230 may directly transmit the generated annotation work result to the learning data generation device 100 or to the learning data verification device 300 through the communication unit 205 .
  • the annotation device 200 includes a processor 250, a memory 255, a transceiver 260, an input/output device 265, and a data bus. , 270) and storage (Storage, 275).
  • the processor 250 may implement operations and functions of the annotation device 200 based on instructions according to the software 280a in which the method according to the embodiments of the present invention is resident in the memory 255 .
  • Software 280a in which a method according to embodiments of the present invention is implemented may be loaded in the memory 255 .
  • the transceiver 260 may transmit and receive data to and from the learning data generating device 100 and the learning data verifying device 300 .
  • the input/output device 265 may receive data necessary for the operation of the annotation device 200 and output an image, a bounding box, and metadata to be an annotation work.
  • the data bus 270 is connected to the processor 250, the memory 255, the transceiver 260, the input/output device 265, and the storage 275, and is a movement path for transferring data between each component. role can be fulfilled.
  • the storage 275 stores an application programming interface (API), a library file, a resource file, etc. necessary for the execution of the software 280a in which the method according to the embodiments of the present invention is implemented. can be saved
  • the storage 275 may store software 280b in which a method according to embodiments of the present invention is implemented. Also, the storage 275 may store information necessary for performing a method according to embodiments of the present invention.
  • the software (280a, 280b) for implementing the method of assigning and verifying the tracking ID resides in the memory 255 or stored in the storage 275
  • the processor 250 is artificial intelligence (AI) Step of specifying an object included in an image that is a target of annotation work for learning, step of giving a tracking ID to the specified object by the processor 250, and step of the processor 250 specifying the image of the specified object It may be a computer program recorded on a recording medium to execute the step of generating metadata for the image by including the inside coordinates and the assigned tracking ID.
  • AI artificial intelligence
  • the software (280a, 280b) for implementing the method of generating and checking metadata resident in the memory 255 or stored in the storage 275 is the processor 250 artificial intelligence (AI) Specifying an object included in an image that is a target of annotation work for learning, the processor 250 compares the image with an existing image, and converts environment information given to the existing image into environment information of the image. Determining whether or not it is applicable, and if the processor 250 determines that the environment information given to the existing image is applicable as the environment information of the image, the coordinates of the specified object in the image and the existing image It may be a computer program recorded on a recording medium to execute the step of generating metadata for the image by including the given environment information.
  • AI artificial intelligence
  • the software (280a, 280b) for implementing a method for automatically generating metadata residing in the memory 255 or stored in the storage 275 is provided by the processor 250 using artificial intelligence (AI) Specifying an object included in an image that is a target of annotation work for learning, the processor 250 setting one or more reference points in the image, and based on the color, brightness, or gamma of a pixel of the set reference point
  • AI artificial intelligence
  • the recording medium It can be a recorded computer program.
  • the processor 250 may include an Application-Specific Integrated Circuit (ASIC), another chipset, a logic circuit, and/or a data processing device.
  • the memory 255 may include read-only memory (ROM), random access memory (RAM), flash memory, a memory card, a storage medium, and/or other storage devices.
  • the transceiver 260 may include a baseband circuit for processing wired/wireless signals.
  • the input/output device 265 includes an input device such as a keyboard, a mouse, and/or a joystick, and a Liquid Crystal Display (LCD), an Organic LED (OLED), and/or a liquid crystal display (LCD).
  • an image output device such as an active matrix OLED (AMOLED) may include a printing device such as a printer or a plotter.
  • AMOLED active matrix OLED
  • a module may reside in memory 255 and be executed by processor 250 .
  • the memory 255 may be internal or external to the processor 250 and may be connected to the processor 250 by various well-known means.
  • Each component shown in FIG. 5 may be implemented by various means, eg, hardware, firmware, software, or a combination thereof.
  • one embodiment of the present invention includes one or more Application Specific Integrated Circuits (ASICs), Digital Signal Processors (DSPs), Digital Signal Processing Devices (DSPDs), Programmable Logic Devices (PLDs), FPGAs ( Field Programmable Gate Arrays), processors, controllers, microcontrollers, microprocessors, etc.
  • ASICs Application Specific Integrated Circuits
  • DSPs Digital Signal Processors
  • DSPDs Digital Signal Processing Devices
  • PLDs Programmable Logic Devices
  • FPGAs Field Programmable Gate Arrays
  • processors controllers, microcontrollers, microprocessors, etc.
  • an embodiment of the present invention is implemented in the form of a module, procedure, function, etc. that performs the functions or operations described above, and is stored on a recording medium readable through various computer means.
  • the recording medium may include program commands, data files, data structures, etc. alone or in combination.
  • Program instructions recorded on the recording medium may be those specially designed and configured for the present invention, or those known and usable to those skilled in computer software.
  • recording media include magnetic media such as hard disks, floppy disks and magnetic tapes, optical media such as CD-ROMs (Compact Disk Read Only Memory) and DVDs (Digital Video Disks), floptical It includes hardware devices specially configured to store and execute program instructions, such as magneto-optical media, such as a floptical disk, and ROM, RAM, flash memory, and the like. Examples of program instructions may include high-level language codes that can be executed by a computer using an interpreter or the like as well as machine language codes generated by a compiler. These hardware devices may be configured to operate as one or more pieces of software to perform the operations of the present invention, and vice versa.
  • the learning data generation apparatus 100 of the artificial intelligence learning system may design a data structure for artificial intelligence (AI) learning (S105).
  • AI artificial intelligence
  • the learning data generating device 100 may define an ontology for AI learning and a classification system of data for AI learning based on a user's control.
  • the user of the learning data generating device 100 can control the structure of data for artificial intelligence (AI) learning to have versatility in consideration of the needs of the main business partner or customer.
  • AI artificial intelligence
  • the learning data generating device 100 may collect and refine data for artificial intelligence (AI) learning based on the designed data structure (S110).
  • AI artificial intelligence
  • the learning data generating device 100 may receive images from the outside, collect images by performing web crawling, or download images from an external institution's device that holds image data.
  • the learning data generation device 100 may remove duplicate or extremely similar images from among the collected images and de-identify personal information included in the images.
  • the learning data generating device 100 may distribute and transmit the collected and refined data (ie, the image to be annotated) to a plurality of annotation devices 200 (S115).
  • Each of the plurality of annotation devices 200 of the artificial intelligence learning system may specify an object included in an image received from the learning data generating device 100 and to be annotated ( S120).
  • the annotation device 200 may receive coordinates through a tool according to a user's control signal, and set a bounding box based on the input coordinates to specify an object included in the image.
  • the bounding box is an area for specifying an object to be learned by artificial intelligence (AI) among objects included in the image.
  • the annotation device 200 may generate metadata for an image to be an annotation work (S125).
  • the annotation device 200 may set specific values to be included in metadata according to a control signal input from a user, but the tracking ID and environment of the object are not dependent on the user's control signal. Information can be determined or assigned automatically.
  • the annotation device 200 may generate an annotation work result based on the generated metadata and transmit the generated annotation work result to the learning data verification device 300 (S130).
  • the learning data verification apparatus 300 of the artificial intelligence learning system may inspect annotation work results received from the annotation apparatuses 200 (S135).
  • the learning data verification device 300 may perform inspection using a script for annotation work results.
  • the script is a code for verifying whether or not the target quality previously set for the annotation work result meets or whether the data is valid.
  • the learning data verification device 300 can provide a user interface (UI) that can more easily inspect tracking IDs and environment variables of objects included in annotation work results. .
  • UI user interface
  • the learning data verification device 300 may transmit annotation work results and inspection results to the learning data generating device 100 (S140).
  • the learning data generating apparatus 100 may build big data about a plurality of annotation work results based on the metadata of the annotation work results (S145).
  • the learning data generating device 100 may receive a data request including a request value from the artificial intelligence learning device 400 (S150).
  • the required value is a value related to requirements that AI learning data must satisfy in order for AI to achieve its development purpose. That is, the requested value is a value indicating the type of information to be included in the data for AI learning.
  • the learning data generating apparatus 100 may mine or filter the constructed big data to extract an annotation work result corresponding to a requested value.
  • the learning data generating device 100 may generate artificial intelligence (AI) learning data by packaging the extracted annotation work result (S155).
  • AI artificial intelligence
  • the learning data generating device 100 may transmit the generated artificial intelligence (AI) learning data to the artificial intelligence learning device 400 .
  • AI artificial intelligence

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Library & Information Science (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

L'invention concerne un procédé par lequel des données pour l'entraînement d'intelligence artificielle basé sur des mégadonnées peuvent être extraites. Le procédé peut comprendre une étape à laquelle un dispositif de génération de données d'entraînement reçoit une pluralité de résultats de tâche d'annotation, classifie la pluralité de résultats de tâche d'annotation sur la base de métadonnées, et génère des données pour l'entraînement d'intelligence artificielle (IA).
PCT/KR2022/007663 2021-06-17 2022-05-30 Procédé d'extraction de données pour l'entraînement d'intelligence artificielle basé sur des mégadonnées, et programme informatique enregistré sur support d'enregistrement pour l'exécuter WO2022265262A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR1020210078671A KR102310600B1 (ko) 2021-06-17 2021-06-17 빅데이터 기반의 인공지능 학습용 데이터 추출 방법 및 이를 실행하기 위하여 기록매체에 기록된 컴퓨터 프로그램
KR10-2021-0078671 2021-06-17

Publications (1)

Publication Number Publication Date
WO2022265262A1 true WO2022265262A1 (fr) 2022-12-22

Family

ID=78150823

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2022/007663 WO2022265262A1 (fr) 2021-06-17 2022-05-30 Procédé d'extraction de données pour l'entraînement d'intelligence artificielle basé sur des mégadonnées, et programme informatique enregistré sur support d'enregistrement pour l'exécuter

Country Status (2)

Country Link
KR (1) KR102310600B1 (fr)
WO (1) WO2022265262A1 (fr)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102310600B1 (ko) * 2021-06-17 2021-10-13 주식회사 인피닉 빅데이터 기반의 인공지능 학습용 데이터 추출 방법 및 이를 실행하기 위하여 기록매체에 기록된 컴퓨터 프로그램
KR102592561B1 (ko) * 2021-10-27 2023-10-23 (주) 경성테크놀러지 인공지능 학습용 데이터 구축 시스템

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040236761A1 (en) * 2003-05-23 2004-11-25 Hans-Joachim Both File system storage
KR20190088094A (ko) * 2017-12-29 2019-07-26 주식회사 아임클라우드 자동 학습 데이터 생성을 통한 모델 학습 시스템 및 방법
KR20200042629A (ko) * 2018-10-16 2020-04-24 주식회사 키센스 인공지능 학습을 위한 모바일 기기의 터치 기반 어노테이션과 이미지 생성 방법 및 그 장치
US20200152316A1 (en) * 2018-11-09 2020-05-14 Lunit Inc. Method for managing annotation job, apparatus and system supporting the same
US20200202171A1 (en) * 2017-05-14 2020-06-25 Digital Reasoning Systems, Inc. Systems and methods for rapidly building, managing, and sharing machine learning models
KR102310600B1 (ko) * 2021-06-17 2021-10-13 주식회사 인피닉 빅데이터 기반의 인공지능 학습용 데이터 추출 방법 및 이를 실행하기 위하여 기록매체에 기록된 컴퓨터 프로그램

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040236761A1 (en) * 2003-05-23 2004-11-25 Hans-Joachim Both File system storage
US20200202171A1 (en) * 2017-05-14 2020-06-25 Digital Reasoning Systems, Inc. Systems and methods for rapidly building, managing, and sharing machine learning models
KR20190088094A (ko) * 2017-12-29 2019-07-26 주식회사 아임클라우드 자동 학습 데이터 생성을 통한 모델 학습 시스템 및 방법
KR20200042629A (ko) * 2018-10-16 2020-04-24 주식회사 키센스 인공지능 학습을 위한 모바일 기기의 터치 기반 어노테이션과 이미지 생성 방법 및 그 장치
US20200152316A1 (en) * 2018-11-09 2020-05-14 Lunit Inc. Method for managing annotation job, apparatus and system supporting the same
KR102310600B1 (ko) * 2021-06-17 2021-10-13 주식회사 인피닉 빅데이터 기반의 인공지능 학습용 데이터 추출 방법 및 이를 실행하기 위하여 기록매체에 기록된 컴퓨터 프로그램

Also Published As

Publication number Publication date
KR102310600B1 (ko) 2021-10-13

Similar Documents

Publication Publication Date Title
WO2022265262A1 (fr) Procédé d'extraction de données pour l'entraînement d'intelligence artificielle basé sur des mégadonnées, et programme informatique enregistré sur support d'enregistrement pour l'exécuter
WO2023120831A1 (fr) Procédé de désidentification et programme informatique enregistré sur un support d'enregistrement en vue de son exécution
EP3735662A1 (fr) Procédé de réalisation d'apprentissage d'un réseau neuronal profond et appareil associé
WO2020045848A1 (fr) Système et procédé pour le diagnostic d'une maladie à l'aide d'un réseau neuronal effectuant une segmentation
WO2020138745A1 (fr) Procédé de traitement d'image, appareil, dispositif électronique et support d'informations lisible par ordinateur
WO2010062063A2 (fr) Procédé et système pour prévenir une utilisation illicite liée à un logiciel de navigation
WO2019059725A1 (fr) Procédé et dispositif de fourniture de service de réalité augmentée
WO2021091022A1 (fr) Système d'apprentissage automatique et procédé de fonctionnement pour système d'apprentissage automatique
WO2021153969A1 (fr) Procédés et systèmes de gestion du traitement d'un réseau neuronal à travers des processeurs hétérogènes
EP3241102A1 (fr) Système électronique doté d'un mécanisme de gestion d'accès, et son procédé de fonctionnement
WO2020091207A1 (fr) Procédé, appareil et programme informatique pour compléter une peinture d'une image et procédé, appareil et programme informatique pour entraîner un réseau neuronal artificiel
WO2021010671A9 (fr) Système de diagnostic de maladie et procédé pour réaliser une segmentation au moyen d'un réseau neuronal et d'un bloc non localisé
WO2021080175A1 (fr) Procédé de traitement de contenu
EP3922036A1 (fr) Appareil et procédé de génération d'image
WO2021172921A1 (fr) Appareil d'imagerie médicale et procédé de traitement d'images médicales
WO2023282466A1 (fr) Procédé de collecte de données d'apprentissage utilisant une prévisualisation laser, et programme informatique enregistré sur un support d'enregistrement pour l'exécuter
WO2020032560A2 (fr) Système et procédé de génération de résultats de diagnostic
EP3803797A1 (fr) Procédés et systèmes de réalisation d'opérations de modification sur un support
WO2023054832A1 (fr) Procédé d'annotation pour désigner facilement un objet et programme informatique enregistré dans un support d'enregistrement afin de mettre en œuvre le procédé
WO2023022321A1 (fr) Serveur d'apprentissage distribué et procédé d'apprentissage distribué
WO2015174623A1 (fr) Dispositif et procédé pour traitement d'entrée
KR102356914B1 (ko) 메타데이터 자동 생성 방법 및 이를 실행하기 위하여 기록매체에 기록된 컴퓨터 프로그램
WO2023054833A1 (fr) Procédé d'augmentation de données pour apprentissage automatique, et programme informatique enregistré dans un support d'enregistrement pour l'exécution de celui-ci
KR102313942B1 (ko) 연속된 이미지의 객체 추적 정보를 포함하는 어노테이션 방법 및 이를 실행하기 위하여 기록매체에 기록된 컴퓨터 프로그램
WO2020204287A1 (fr) Appareil d'affichage et procédé de traitement d'image associé

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22825188

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 22825188

Country of ref document: EP

Kind code of ref document: A1