WO2019039758A1 - Procédé de génération et d'apprentissage de réseau neuronal amélioré - Google Patents

Procédé de génération et d'apprentissage de réseau neuronal amélioré Download PDF

Info

Publication number
WO2019039758A1
WO2019039758A1 PCT/KR2018/008621 KR2018008621W WO2019039758A1 WO 2019039758 A1 WO2019039758 A1 WO 2019039758A1 KR 2018008621 W KR2018008621 W KR 2018008621W WO 2019039758 A1 WO2019039758 A1 WO 2019039758A1
Authority
WO
WIPO (PCT)
Prior art keywords
input
layer
neural network
data
generating
Prior art date
Application number
PCT/KR2018/008621
Other languages
English (en)
Korean (ko)
Inventor
김광민
Original Assignee
주식회사 수아랩
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 주식회사 수아랩 filed Critical 주식회사 수아랩
Publication of WO2019039758A1 publication Critical patent/WO2019039758A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Definitions

  • the present invention relates to a neural network, and more particularly, to a neural network in which learning efficiency is improved by processing two or more input data simultaneously.
  • Deep learning is an algorithm for extracting and utilizing features from learning data using deep neural networks (DNN).
  • DNN deep neural networks
  • Each layer of the neural network receives the information extracted from the lower layer and generates more abstract information using this information to transmit to the upper layer.
  • Multiple layers of neural networks can extract very high-level features through multiple abstractions. High-level features include more information than low-level features, but are also robust to variations, so high and stable performance can be achieved if recognition is performed using them.
  • CNN convolutional neural network
  • the origins of the convolution neural network are the Neocognitron, which Fukushima imitated in the 1980s. Then, in the 1990s, LeCun began to be widely used in real-world problems by successfully applying a gradient-based learning algorithm.
  • the excellent performance of convolutional neural networks has attracted the explosive interest of many researchers, and research is actively being conducted to further improve the convolution neural network or to apply it to new problems.
  • the present invention provides a method for generating an improved neural network that enhances learning efficiency by simultaneously processing two or more inputs.
  • the present invention also provides a method for simultaneously learning two or more kinds of inputs to learn a neural network efficiently.
  • a method of generating a neural network model implemented by a computing device comprises generating a first input layer to which a first input is applied, step; And generating a hidden layer to which an output of the first input layer is applied, wherein the hidden layer includes a second input layer to which a second input is applied;
  • the method of generating a neural network model including the neural network model may be provided.
  • a method for learning a neural network model implemented by a computing device comprises the steps of: (a) generating a training data set comprising a first input and a second input, receiving a training dataset; Applying the first input to a first input layer; And applying the second input to a second input layer, the second input layer being included in a hidden layer to which the output of the first input layer is applied; And a method for learning a neural network model.
  • a computing device comprising: at least one processor; And a memory for storing instructions executable on the one or more processors; The one or more processors generating a first input layer to which a first input is applied; And generating a hidden layer to which the output of the first input layer is applied, the hidden layer including a second input layer to which a second input is applied;
  • a computing device can be provided.
  • a computer program stored in a computer-readable storage medium including encoded instructions when executed by one or more processors of a computer system, Cause the processors to perform operations to create a neural network model, the operations comprising: generating a first input layer to which a first input is applied; And generating a hidden layer to which the output of the first input layer is applied, the hidden layer including a second input layer to which a second input is applied;
  • two or more types of inputs can be simultaneously processed to create an improved neural network with enhanced learning efficiency. Also, according to the present disclosure, two or more kinds of inputs can be simultaneously processed, and the neural network can be efficiently learned.
  • 1 is a view for explaining the object of the present invention.
  • FIG. 2 shows a structure of a neural network model according to an embodiment of the present invention.
  • FIG. 3 is a diagram for explaining a method of generating a neural network model according to an embodiment of the present invention.
  • FIG. 4 is a diagram for explaining a method for learning a neural network model according to an embodiment of the present invention.
  • FIG. 5 illustrates an exemplary block diagram of a neural network module in accordance with an embodiment of the present invention.
  • FIG. 6 shows a block diagram of an exemplary computing device for implementing a method for generating a neural network model and a method for learning a neural network, in accordance with an embodiment of the present invention.
  • the terms " an embodiment, “ “ an embodiment, “ “ an embodiment, “ “ an embodiment, etc. are intended to indicate that any aspect or design described is better or worse than other aspects or designs.
  • the terms 'component,' 'module,' 'system,' 'interface,' and the like generally refer to a computer-related entity and include, for example, hardware, And software.
  • 1 is a view for explaining the object of the present invention.
  • a neural network may be a recognition model that mimics the computational capabilities of a biological system using a large number of artificial neurons connected by a link.
  • the neural network may be implemented in software, hardware, or a combination thereof.
  • a neural network may be referred to as an artificial neural network.
  • Neural networks can utilize artificial neurons that simplify the function of biological neurons.
  • An artificial neuron can be referred to as a node.
  • Artificial neurons can be interconnected via a connection line with connection weights.
  • a connection weight is a specific value of a connection line and may be referred to as a synapse weight or connection strength.
  • Neuronal networks can perform human cognitive or learning processes through artificial neurons.
  • the neural network can be utilized in the field of determining input data by using a computer.
  • a neural network is used to identify an object existing in a video image, to determine the position of an object, to detect and identify a lesion in a medical image, to understand contents of a voice using voice data, And extracting information for extracting information from the contents of the document.
  • the present invention is not limited to this, and the neural network can be utilized in various fields.
  • the structure of the neural network can be generally determined in advance, and the connection weight according to the connection between the nodes can be calculated to an appropriate value by using data already known to be correct to which class belongs. Data that are known to be correct in this way may be referred to as 'training data', and the process of determining the connection weights may be referred to as 'training'.
  • the neural network for which the connection weight is determined may predict which class the input data belongs to and output the predicted value, which may be referred to as a " test " process.
  • a &quot test &quot
  • process test &quot
  • the above-described processes can be referred to as various terms.
  • the neural network may be useful for the neural network to learn additional data of a different type from the learning data. More specifically, the neural network can learn additional features that are difficult to learn using only one type of learning data, using additional data. For example, it may be useful for a neural network to learn discrete data associated with learning data. Here, the discrete data may include data useful for identifying abnormal data among the learning data.
  • the neural network when the neural network is used to check whether defects have occurred in the steel in the steel manufacturing process, the neural network can learn a plurality of steel images 100 that are determined to be defective as learning data .
  • the steel image 100 may be divided into a plurality of image patches 110 and used for learning to reduce the amount of calculation required for learning.
  • the steel image 100 may be divided into a plurality of image patches 110 having a pixel size of 20x20 and used for learning.
  • the plurality of image patches 110 may have respective coordinate values.
  • defects of steel may exist in an image patch whose coordinate values are (7, 4), (7, 5).
  • the coordinate value of the image patch 110 may be provided as learning data so that it can learn which coordinate value the image patch (i.e., abnormal data) has. This can increase the learning efficiency of the neural network.
  • the neural network model of the present invention may include a first input layer 220 to which a first input is applied and a second input layer 270 to which a second input is applied.
  • the first input 200 applied to the first input layer 220 may include general learning data. More specifically, the first input 200 may include learning data that has been used in learning of a conventional neural network. For example, the first input 200 may include image data, sound data, image data, character data, and the like. However, the present invention is not limited to this, and the first input may include various data.
  • the second input 210 applied to the second input layer 270 may include a different type of data than the first input 200.
  • the second input may include data capable of learning characteristics that are difficult to learn with only the first input 200.
  • the second input may be applied with discrete data related to the first input.
  • the coordinate value of the image patch 110 may be applied.
  • the technical features of the present invention can be used to improve learning efficiency by providing more data to the neural network.
  • the neural network model of the present invention has a structure for simultaneously processing two kinds of data which are different from each other, so that efficient learning is possible.
  • FIG. 2 shows a structure of a neural network model according to an embodiment of the present invention.
  • the neural network model may comprise a plurality of layers.
  • the neural network model may include a first input layer 220, a hidden layer 230, and an output layer (not shown).
  • the first input layer 220 may receive a first input for performing learning and transmit the first input to the hidden layer 230.
  • the output layer may transmit the first input to the hidden layer 230 based on data received from the hidden layer 230, The output of the model can be generated.
  • the hidden layer 230 is located between the first input layer 220 and the output layer and can change learning data transmitted through the first input layer 220 to a value that is easy to predict.
  • the layers included in the first input layer 220, the hidden layer 230, and the output layer may include a plurality of nodes.
  • a node included in the first input layer 220 is referred to as a first input node
  • a node included in the hidden layer 230 is referred to as a hidden node
  • a node may be referred to as an output node.
  • the first input nodes included in the first input layer 220 and the hidden nodes included in the hidden layer 230 may be connected to each other through connection links having connection weights.
  • the hidden nodes included in the hidden layer 230 and the output nodes included in the output layer may be connected to each other through connection lines having connection weights.
  • the hidden layer 230 may include a plurality of layers.
  • a neural network in which the hidden layer 230 includes a plurality of layers may be referred to as a deep neural network. Learning a deep neural network can be referred to as deep learning.
  • the hidden layer 230 may include a plurality of convolution layers 240, a pooling layer 250, and a fully connected layer 260 and 270.
  • Convolution layer 240 may refer to a layer having a structure in which neighboring nodes in the previous layer are connected to nodes in the next layer.
  • the pooling layer 250 may refer to a layer having a structure capable of selecting and integrating a part at a node of a previous layer.
  • the pulley connected layers 260 and 280 may refer to a layer having a structure in which the nodes included in the layer are connected to all nodes included in the previous layer.
  • the above-mentioned layers are exemplary, and some of the layers may be changed or omitted, and additional layers may be included in the hidden layer 230.
  • some layers included in the hidden layer 230 may be included in the first input layer 220 or the output layer as the case may be.
  • the hidden layer 230 may include a second input layer 270.
  • the second input layer 270 may be disposed in parallel with any one of the pulley connected layers included in the hidden layer 230. Also, the second input layer 270 may be applied with the second input 210, and may transmit the output to the upper layer.
  • the neural network model of the present invention may include a first input layer 220, which is a first input, and a second input layer 270, to which a second input is applied.
  • the first input 200 applied to the first input layer 220 may include general learning data. More specifically, the first input 200 may include learning data that has been used in learning of a conventional neural network. For example, the first input 200 may include image data, sound data, image data, character data, and the like. However, the present invention is not limited to this, and the first input may include various data.
  • the second input 210 applied to the second input layer 270 may include a different type of data than the first input 200. More specifically, the second input may include data that can learn features that are difficult to learn with only the first input 200. For example, the second input may include data that is applied to discrete data associated with the first input . In one embodiment, when a plurality of image patches 110 are applied to the first input layer 220 as a first input 200, 2 input 210, the coordinate value of the image patch 110 may be applied.
  • the technical features of the present invention can be used to improve learning efficiency by providing more data to the neural network.
  • the neural network model of the present invention has a structure for simultaneously processing two kinds of data which are different from each other, so that efficient learning is possible.
  • FIG. 3 is a diagram for explaining a method of generating a neural network model according to an embodiment of the present invention.
  • the neural network model may include a first input layer 220, a hidden layer 230, and an output layer.
  • the above-mentioned layers are exemplary, and some of the layers may be changed or omitted, and additional layers may be included in the neural network model.
  • the layers may refer to multiple layers.
  • the neural network model may be a convolution neural network.
  • a convolutional neural network is a feedforward artificial neural network (a network in which data is handed over from the lower layer to the upper layer) tied in a manner that responds to overlapping regions of individual neurons, Lt; / RTI >
  • a method of generating a neural network model may include generating (300) a first input layer 220 to which a first input 200 is applied.
  • the first input layer 220 may be generated as the lowest layer of the neural network model.
  • the first input layer 220 may receive the first input 200.
  • the first input layer 220 may communicate the received first input 200 to the hidden layer 230 for use in learning and testing.
  • the first input 200 may be transformed to be processed in the hidden layer 230 in the first input layer 220.
  • the first input 200 may also be preprocessed to improve learning efficiency.
  • the first input 200 may be preprocessed such that the noise component is removed or upscaled.
  • the present invention is not limited to this, and the first input layer 220 may be formed as a lower layer or a higher layer of the neural network model, and may perform various operations.
  • the first input 200 applied to the first input layer 220 may include general learning data. More specifically, the first input 200 may include learning data that has been used in learning of a conventional neural network. For example, the first input 200 may include image data, sound data, image data, character data, and the like. However, the present invention is not limited to this, and the first input may include various data.
  • a method of generating a neural network model may include generating 310 a hidden layer 230 to which the output of the first input layer 220 is applied.
  • the hidden layer 230 may include a second input layer 270 to which a second input 210 is applied.
  • the hidden layer 230 may include a plurality of layers.
  • the hidden layer 230 may include at least one convolution layer 240, at least one pulling layer 250, at least one pulley connected layer 260, 280, and a second input layer 270 .
  • the plurality of layers included in the hidden layer 230 may be arranged in series or in parallel with each other.
  • the present invention is not limited thereto, and the hidden layer 230 may include various layers.
  • the hidden layer 230 may receive the output of the first input layer 220 and extract and classify the features. More specifically, the convolution layer 240 can extract features from the hidden layer 230 by abstracting the input data. For example, the convolution layer 240 may apply a filter to a certain area of data (e.g., 5x5 or 3x3) to extract features. These filters can be changed according to learning.
  • the pulling layer 250 may perform a sub-sampling process. The sub-sampling process can take a maximum value or take an average value for the input data.
  • the neural network model may obtain a global feature while passing the first input 200 through the multiple layers of convolution layer 240 and pooling layer 250. However, the present invention is not limited to this, and the hidden layer can perform various operations.
  • the pulley connected layers 260 and 280 are generally positioned above the convolution layer 240 and the pulling layer 250 and perform operations of sorting using extracted features while passing through the convolution layer and the pooling layer .
  • the pulley connected layers 260 and 280 may be formed in multiple layers.
  • the pulley connected layers 260 and 280 may be formed in two layers as shown in FIG.
  • the present invention is not limited to this, and the pulley-connected layers 260 and 280 may be formed in three or more layers.
  • the second input layer 270 may receive the second input 210.
  • the second input layer 270 may be used to classify the received second input 210 with the extracted features as the first input 200 passes through the convolution layer 240 and the pulling layer 250 can do.
  • the second input layer 270 may be created above the convolution layer 240 and the pooling layer 250.
  • the second input layer 270 may be formed at various positions on the hidden layer 230.
  • the second input layer may receive the second input directly.
  • the second input layer can receive the second input directly without going through another layer.
  • the present invention is not limited to this, and the second input layer can receive the second input through another layer.
  • the second input layer may receive its output as an input after the second input is applied to another layer.
  • the node to which the second input is applied can be connected to only a small number of nodes included in the upper layer.
  • a node to which a second input is applied may be connected to only one node included in the upper layer.
  • the present invention is not limited to this, and a node to which a second input is applied may be connected to two or more nodes included in an upper layer.
  • the node through which the second input is passed may not be connected to the node used to extract the characteristics of the first input. Therefore, the second input may be rarely used in the process of extracting the characteristic of the first input, or may be used only in a part of the process.
  • the second input layer 270 may be disposed in parallel with the first pulley-connected layer 260 included in the pulley-connected layers 260 and 280.
  • the neural network may make the second input applied to the second input layer 270 available for classification operations.
  • the second input layer 270 may include a convolution layer 240 and an upper layer of the pulling layer 250 such that the second input 210 is not used in the process of extracting the characteristics of the first input 200.
  • the pulley connected layers 260 and 280 are generally positioned above the convolution layer 240 and the pulling layer 250 to perform the task of classifying extracted features.
  • the second input layer 270 may be disposed in parallel with the first pulley connected layer 260 to allow the second input 210 to be used to perform the classification operation.
  • the second input layer may be disposed in parallel with any of the convolution layer 240 and the pulling layer 250 and the second input 210 may have a characteristic of the first input 200 Can be used in the extraction process.
  • the first pulley connected layer 260 may be the lowest layer among the pulley connected layers 260 and 280.
  • the second input 210 is applied after the feature of the first input 200 is extracted rather than in the process of extracting the feature of the first input 200.
  • the pulley connected layers 260 and 280 may be formed in multiple layers for efficient classification generally. Accordingly, the second input layer 270 is arranged in parallel to the lowermost layer among the plurality of pulley-connected layers, and is used for starting the operation of classifying the second input 210 using the extracted feature .
  • the present invention is not limited thereto. In some cases, the second input layer 270 may be disposed in a lower layer or a higher layer among the pulley-connected layers in parallel.
  • the second input layer 270 and the first pulley connected layer may be connected to the second pulley connected layer.
  • the second pulley connected layer may be a layer located on top of the second input layer 270 and the first pulley connected layer.
  • the second input 210 may be applied to multiple layers of pulley-connected layers with features extracted from the lower layer and used for classification operations.
  • the second input layer 270 and the first pulley connected layer can be coupled to deliver the extracted features of the first input 200 and the second input 210 to the second pulley connected layer.
  • the second pulley connected layer can then process the received data and pass it back to the multi-layer pulley connected or output layer.
  • the second input layer 270 and the first pulley connected layer 260 may be connected to various kinds of layers.
  • the second input layer 270 and the first pulley connected layer may be coupled to a 1x1 convolution layer.
  • the 1x1 convolution layer can perform the task of classifying extracted features, such as pulley connected layers 260 and 280, but is not limited thereto.
  • the first input 200 and the second input 210 may comprise different types of data.
  • the first input 200 applied to the first input layer 220 may include general learning data. More specifically, the first input 200 may include learning data that has been used in learning of a conventional neural network.
  • the first input 200 may include image data, sound data, image data, character data, and the like.
  • the present invention is not limited to this, and the first input may include various data.
  • the second input 210 applied to the second input layer 270 may include a different kind of data than the first input 200.
  • the second input may include data capable of learning characteristics that are difficult to learn with only the first input 200.
  • the second input may be applied with discrete data related to the first input.
  • the coordinate value of the image patch 110 may be applied.
  • the second input 210 may include discrete data related to the first input 200. More specifically, the discrete data associated with the first input 200 can be used to learn characteristics that are difficult to learn with only the first input. In one embodiment, the discrete data may include data useful for identifying normal and abnormal data in the training data applied to the first input 200. [ Specifically, in the case where the first input 200 is a plurality of image patches 110 in which the steel image 100 is divided, the discrete data related to the first input 200 may include defects (I.e., normal data), and data related to a cause of occurrence of abnormal data that can be used to identify an image patch in which a defect exists (i.e., abnormal data).
  • defects I.e., normal data
  • data related to a cause of occurrence of abnormal data that can be used to identify an image patch in which a defect exists
  • discrete data may include coordinate values of image patches or the like that can be used to identify locations where defects predominantly occur in steel, or steel images that may be used to identify defects in certain processes May include data such as the time, place, process information, etc. at which the apparatus 100 is photographed, and may include data such as temperature, humidity, manufacturing apparatus, etc., which can be used to identify defects in any manufacturing environment.
  • the discrete data may include data indicative of the characteristics of the training data applied to the first input 200. More specifically, when the first input 200 is speech data, the discrete data related to the first input 200 may include the sex, age, key (s) of the speaker that can be used to learn the characteristics of the speech data according to the speaker , Weight, and the like, and may include a recording location, a recording time, a recording device, etc., which can be used to learn characteristics of voice data according to the surrounding environment.
  • the discrete data associated with the first input 200 may include various data.
  • the second input 210 which includes discrete data related to the first input 200, can be usefully used to learn the first input 200.
  • the second input may be used as additional information to learn the first input 200 and may be used to learn additional features of the first input 200 associated with the second input.
  • the first input 200 is the partitioned image patch 110 of the steel image 100 and the second input 210 is the coordinate value of the image patch 110
  • the position where the defects mainly occur in the steel image 100 can be learned by using the coordinate values of the defects 110.
  • the neural network model can learn the characteristics of speech data according to the speaker's gender and age.
  • the neural network model of the present invention has a structure for simultaneously processing two kinds of data which are different from each other, thereby enabling efficient learning.
  • FIG. 4 is a diagram for explaining a method for learning a neural network model according to an embodiment of the present invention.
  • a method for learning a neural network model includes receiving (400) a learning data set that includes a first input (200) and a second input (210), a first input layer (410) applying a first input (200) to the second input layer (270) and applying (420) a second input to a second input layer (270).
  • the second input layer 270 may be included in the hidden layer 230 to which the output of the first input layer 220 is applied.
  • the neural network model of the present invention can be learned by various methods.
  • neural network models can be learned through supervised learning.
  • the supervised learning is a technique of inputting the learning data and the corresponding output data together into the convolutional neural network and updating the weight so that the output data corresponding to the learning data is output.
  • the neural network model of the present invention can update connection weights through back-propagation learning.
  • the error-domain propagation learning estimates an error through forward computation on a given learning data, and then estimates an error in a reverse direction starting from the output layer and toward the first input layer 220 and the second input layer 270 It can be a way to update the connection weights in the direction of reducing the error while propagating one error.
  • the present invention is not limited to this, and the neural network can be learned through Unsupervised Learning or Reinforcement Learning, or may be learned by other learning methods.
  • the learning data set of the present invention may include a first input 200 and a second input 210.
  • the first input 200 can be applied to the first input layer 220 and the second input 210 can be applied to the second input layer 270 included in the hidden layer 230 .
  • the second input layer 270 may be disposed in parallel with the first pulley connected layer included in the pulley connected layer.
  • the first pulley-connected layer may be the lowest layer among the pulley-connected layers.
  • the second input layer 270 and the first pulley connected layer may be connected to a second pulley connected layer, wherein the second pulley connected layer includes a second input layer 270 and the first pulley connected layer
  • the layer may be a layer located at the top of the layer.
  • the first input 200 may include general types of data used as learning data of a neural network.
  • the first input 200 may include image data, sound data, image data, character data, and the like.
  • the first input 200 may include various types of data that can be used for learning of the neural network model.
  • the second input may include a different type of data than the first input 200 applied to the first input layer 220.
  • the second input 210 may include other types of data such as sound data, image data, character data, and the like.
  • the second input 210 may include data of the same kind as that of the first input 200, and may also include various other types of data.
  • the second input 210 may include discrete data related to the first input 200.
  • the discrete data associated with the first input 200 may be data including information regarding the first input 200.
  • the discrete data associated with the first input 200 is the coordinate value of the image patch, The time, place and temperature at which the camera 100 is photographed, process information, and the like.
  • the discrete data associated with the first input 200 may be the gender, age, recording location, recording time, recording device, etc. of the speaker.
  • the second input 210 which includes discrete data related to the first input 200, can be usefully used to learn and test the first input 200.
  • the first input 200 is the partitioned image patch 110 of the steel image 100 and the second input 210 is the coordinate value of the image patch 110
  • the position where the defects mainly occur in the steel image 100 can be learned by using the coordinate values of the defects 110.
  • the neural network model can learn the characteristics of speech data according to the speaker's gender and age.
  • the neural network model of the present invention has a structure for simultaneously processing two kinds of data which are different from each other, thereby enabling efficient learning.
  • FIG. 5 shows an exemplary block diagram of a neural network module in accordance with an embodiment of the present invention.
  • the neural network module 600 may include a neural network generator 610 and a neural network learner 620.
  • the neural network generating unit 610 may include a first input layer generating unit 611 and a hidden layer generating unit 612.
  • the hidden layer generating unit 612 may include a second input layer generating unit 613.
  • the components of neural network module 600 shown in FIG. 5 are exemplary, some of which may be omitted, or additional components may be present.
  • the first input layer generator 611 may generate the first input layer 220 to which the first input 200 is applied.
  • the hidden layer generating unit 612 may generate the hidden layer 230 by generating the convolution layer 240, the pulling layer 250, the pulley connected layers 260 and 280, and other layers.
  • the second input layer generating unit 613 may generate the second input layer 270.
  • the neural network learning unit 620 may receive a learning data set that includes a first input 200 and a second input 210. [ The neural network learning unit 620 applies a first input 200 to the first input layer 220 and a second input 210 to the second input layer 270 in order to learn a neural network model. can do.
  • FIG. 6 shows a block diagram of an exemplary computing device for implementing a method for generating a neural network model and a method for learning a neural network, in accordance with an embodiment of the present invention.
  • program modules include routines, programs, components, data structures, etc. that perform particular tasks or implement particular abstract data types. It will also be appreciated by those skilled in the art that the methods of the present invention may be practiced with other computer systems, such as single-processor or multiprocessor computer systems, minicomputers, mainframe computers, as well as personal computers, handheld computing devices, microprocessor-based or programmable consumer electronics, And may operate in conjunction with one or more associated devices).
  • the described embodiments of the invention may also be practiced in distributed computing environments where certain tasks are performed by remote processing devices connected through a communications network.
  • program modules may be located in both local and remote memory storage devices.
  • Computers typically include a variety of computer readable media. Any medium accessible by a computer may be a computer-readable medium, which may include volatile and non-volatile media, transitory and non-transitory media, removable and non-removable media, Removable media.
  • computer readable media can comprise computer readable storage media and computer readable transmission media.
  • Computer-readable storage media includes both volatile and nonvolatile media, both temporary and non-volatile media, both removable and non-removable, implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data Media.
  • the computer-readable storage medium may be any form of storage device such as RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital video disk (DVD) or other optical disk storage, magnetic cassette, magnetic tape, Apparatus, or any other medium that can be used to store the desired information that can be accessed by a computer.
  • Computer readable transmission media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism, It includes all information delivery media.
  • modulated data signal refers to a signal that has one or more of its characteristics set or changed to encode information in the signal.
  • the transmit and receive (communication) media include wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of any of the above described media are also intended to be included within the scope of computer readable transmission media.
  • Apparatus 702 includes a processing unit 704, a system memory 706, and a system bus 708.
  • the system bus 708 couples system components, including but not limited to, system memory 706 to the processing unit 704.
  • the processing unit 704 may be any of a variety of commercially available processors. Dual processors and other multiprocessor architectures may also be used as the processing unit 704.
  • the system bus 708 may be any of several types of bus structures that may additionally be interconnected to a local bus using any of the memory bus, peripheral bus, and various commercial bus architectures.
  • the system memory 706 includes a read only memory (ROM) 710 and a random access memory (RAM) 712.
  • the basic input / output system (BIOS) is stored in a non-volatile memory 710, such as a ROM, EPROM, or EEPROM, which can be used to assist in transferring information between components within the computing device 702, It contains basic routines.
  • the RAM 712 may also include a high speed RAM such as static RAM for caching data.
  • the computing device 702 may also be configured for external use within an appropriate chassis (not shown), such as an internal hard disk drive (HDD) 714 (e.g., EIDE, SATA) (E.g., read from or write to a removable diskette 718), and an optical disk drive 720 (e.g., a CD-ROM For reading disc 722 or reading from or writing to other high capacity optical media such as DVD).
  • HDD hard disk drive
  • SATA Serial Advanced Technology Attachment
  • optical disk drive 720 e.g., a CD-ROM For reading disc 722 or reading from or writing to other high capacity optical media such as DVD.
  • the hard disk drive 714, magnetic disk drive 716 and optical disk drive 720 are connected to the system bus 708 by a hard disk drive interface 724, a magnetic disk drive interface 726 and an optical drive interface 728, respectively.
  • the interface 724 for external drive implementation includes at least one or both of USB (Universal Serial Bus) and IEEE 1394 interface technologies.
  • drives and their associated computer-readable storage media provide nonvolatile storage of data, data structures, computer-executable instructions, and the like.
  • the drives and media correspond to storing any data in a suitable digital format.
  • a removable optical medium such as a HDD, a removable magnetic disk, and a CD or a DVD
  • a zip drive, magnetic cassette, flash memory card, Or the like may also be used in the exemplary operating environment and any such medium may include computer-executable instructions for carrying out the methods of the present invention.
  • a number of program modules including an operating system 730, one or more application programs 732, other program modules 734, and program data 736, may be stored in the drive 732 and the RAM 712.
  • the program module 734 may include a neural network module (not shown) for generating, learning, and testing a neural network model. All or a portion of the operating system, applications, modules, and / or data may also be cached in RAM 712. It will be appreciated that the present invention may be implemented in a variety of commercially available operating systems or combinations of operating systems.
  • a user may enter commands and information into computing device 702 via one or more wired / wireless input devices, e.g., a pointing device such as keyboard 738 and mouse 740.
  • a pointing device such as keyboard 738 and mouse 740.
  • Other input devices may include a microphone, IR remote control, joystick, game pad, stylus pen, touch screen, etc.
  • input device interface 742 that is coupled to the system bus 708, but may be a parallel port, an IEEE 1394 serial port, a game port, a USB port, an IR interface, ≪ / RTI > and so forth.
  • a monitor 744 or other type of display device is also connected to the system bus 708 via an interface, such as a video adapter 746.
  • the computer typically includes other peripheral output devices (not shown) such as speakers, printers,
  • Computing device 702 may operate in a networked environment using logical connections to one or more remote computers, such as remote computer (s) 748 via wired and / or wireless communication.
  • the remote computer (s) 748 may be a workstation, a server computer, a router, a personal computer, a portable computer, a microprocessor-based entertainment device, a peer device or other conventional network node, But for the sake of simplicity, only memory storage device 750 is shown.
  • the logical connections depicted include a wired / wireless connection to a local area network (LAN) 752 and / or a larger network, e.g., a wide area network (WAN)
  • LAN local area network
  • WAN wide area network
  • the computing device 702 When used in a LAN networking environment, the computing device 702 is connected to the local network 752 via a wired and / or wireless communication network interface or adapter 756.
  • the adapter 756 may facilitate wired or wireless communication to the LAN 752 and the LAN 752 also includes a wireless access point installed therein to communicate with the wireless adapter 756.
  • the computing device 702 When used in a WAN networking environment, the computing device 702 may include a modem 758, or may be connected to a communications server on the WAN 754, or to establish communications over the WAN 754, And other means.
  • a modem 758 which may be an internal or external and a wired or wireless device, is connected to the system bus 708 via a serial port interface 742.
  • program modules depicted relative to computing device 702, or portions thereof, may be stored in remote memory / storage device 750. It will be appreciated that the network connections shown are exemplary and other means of establishing a communication link between the computers may be used.
  • the computing device 702 may be any wireless device or entity that is deployed and operated in wireless communication, such as a printer, a scanner, a desktop and / or portable computer, a portable data assistant (PDA) To any associated equipment or location, and to communication with the phone.
  • PDA portable data assistant
  • the communication may be a predefined structure, such as in a conventional network, or simply an ad hoc communication between at least two devices.
  • Wi-Fi Wireless Fidelity
  • Wi-Fi is a wireless technology such as a cell phone that allows such devices, e.g., computers, to transmit and receive data indoors and outdoors, i. E. Anywhere within the coverage area of a base station.
  • Wi-Fi networks use a wireless technology called IEEE 802.6 (a, b, g, etc.) to provide a secure, reliable and high-speed wireless connection.
  • Wi-Fi can be used to connect computers to each other, the Internet, and a wired network (using IEEE 802.3 or Ethernet).
  • the Wi-Fi network can operate in unlicensed 2.4 and 5 GHz wireless bands, for example, at 6 Mbps (802.6a) or 54 Mbps (802.6b) data rates, or in products containing both bands have.
  • the various embodiments presented herein may be implemented as a method, apparatus, or article of manufacture using standard programming and / or engineering techniques.
  • article of manufacture includes a computer program, carrier, or media accessible from any computer-readable device.
  • the computer-readable medium can be a magnetic storage device (e.g., a hard disk, a floppy disk, a magnetic strip, etc.), an optical disk (e.g., CD, DVD, etc.), a smart card, But are not limited to, devices (e. G., EEPROM, cards, sticks, key drives, etc.).
  • machine-readable medium includes, but is not limited to, various other mediums capable of storing and / or retaining instruction (s) and / or data.
  • the present invention can be used for learning of a neural network utilized in a field for discriminating input data using a computing device.

Abstract

la présente invention concerne, selon un mode de réalisation donné à titre d'exemple, un procédé de génération d'un modèle de réseau neuronal mis en œuvre par un dispositif informatique, le procédé comprenant les étapes consistant : à générer une première couche d'entrée à laquelle une première entrée est appliquée; et à générer une couche cachée à laquelle une sortie de la première couche d'entrée est appliquée, la couche cachée comprenant une seconde couche d'entrée à laquelle une seconde entrée est appliquée. Selon un mode de réalisation donné à titre d'exemple de la présente invention, un procédé d'apprentissage d'un modèle de réseau neuronal mis en œuvre par un dispositif informatique peut être proposé, le procédé comprenant les étapes consistant : à recevoir un ensemble de données d'apprentissage comprenant une première entrée et une seconde entrée; à appliquer la première entrée à une première couche d'entrée; et à appliquer la seconde entrée à une seconde couche d'entrée, la seconde couche d'entrée étant incluse dans une couche cachée à laquelle une sortie de la première couche d'entrée est appliquée.
PCT/KR2018/008621 2017-08-25 2018-07-30 Procédé de génération et d'apprentissage de réseau neuronal amélioré WO2019039758A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR20170107981 2017-08-25
KR10-2017-0107981 2017-08-25

Publications (1)

Publication Number Publication Date
WO2019039758A1 true WO2019039758A1 (fr) 2019-02-28

Family

ID=65438959

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2018/008621 WO2019039758A1 (fr) 2017-08-25 2018-07-30 Procédé de génération et d'apprentissage de réseau neuronal amélioré

Country Status (1)

Country Link
WO (1) WO2019039758A1 (fr)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100169256A1 (en) * 2006-05-19 2010-07-01 Inha-Industry Partnership Institute Separate Learning System and Method Using Two-Layered Neural Network Having Target Values for Hidden Nodes
KR20160063965A (ko) * 2014-11-27 2016-06-07 삼성전자주식회사 뉴럴 네트워크 구조 확장 방법, 디멘션 리덕션 방법, 및 그 방법을 이용한 장치
US20160162781A1 (en) * 2013-07-26 2016-06-09 Isis Innovation Ltd. Method of training a neural network
JP2017049907A (ja) * 2015-09-04 2017-03-09 国立研究開発法人情報通信研究機構 ニューラルネットワーク、その学習方法、及びコンピュータプログラム

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100169256A1 (en) * 2006-05-19 2010-07-01 Inha-Industry Partnership Institute Separate Learning System and Method Using Two-Layered Neural Network Having Target Values for Hidden Nodes
US20160162781A1 (en) * 2013-07-26 2016-06-09 Isis Innovation Ltd. Method of training a neural network
KR20160063965A (ko) * 2014-11-27 2016-06-07 삼성전자주식회사 뉴럴 네트워크 구조 확장 방법, 디멘션 리덕션 방법, 및 그 방법을 이용한 장치
JP2017049907A (ja) * 2015-09-04 2017-03-09 国立研究開発法人情報通信研究機構 ニューラルネットワーク、その学習方法、及びコンピュータプログラム

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
SOO HYUN BAE: "ACOUSTIC SCENE CLASSIFICATION USING PARALLEL COMBINATION OF LSTM AND CNN", DETECTION AND CLASSIFICATION OF ACOUSTIC SCENES AND EVENTS 2016, 3 September 2016 (2016-09-03), pages 1 - 5, XP055579570, Retrieved from the Internet <URL:http://www.cs.tut.fi/sgn/arg/dcase2016/documents/challenge_technical_reports/DCASE2016_Bae_1025.pdf> *

Similar Documents

Publication Publication Date Title
KR101916347B1 (ko) 딥러닝 기반 이미지 비교 장치, 방법 및 컴퓨터 판독가능매체에 저장된 컴퓨터 프로그램
Jauro et al. Deep learning architectures in emerging cloud computing architectures: Recent development, challenges and next research trend
WO2018217019A1 (fr) Dispositif de détection d&#39;un code malveillant variant sur la base d&#39;un apprentissage de réseau neuronal, procédé associé, et support d&#39;enregistrement lisible par ordinateur dans lequel un programme d&#39;exécution dudit procédé est enregistré
WO2020005049A1 (fr) Procédé d&#39;apprentissage pour réseau neuronal artificiel
WO2019039757A1 (fr) Dispositif et procédé de génération de données d&#39;apprentissage et programme informatique stocké dans un support d&#39;enregistrement lisible par ordinateur
CN110990631A (zh) 视频筛选方法、装置、电子设备和存储介质
WO2022005091A1 (fr) Procédé et appareil de lecture de l&#39;âge d&#39;un os
WO2022164191A1 (fr) Système et procédé d&#39;hyperpersonnalisation basée sur un micro-genre avec apprentissage automatique multimodal
KR102078133B1 (ko) 데이터의 어노말리 감지 방법
WO2019098418A1 (fr) Procédé et dispositif d&#39;apprentissage de réseau neuronal
WO2020032506A1 (fr) Système de détection de vision et procédé de détection de vision l&#39;utilisant
WO2021107422A1 (fr) Procédé de surveillance de charge non intrusive utilisant des données de consommation d&#39;énergie
WO2022149696A1 (fr) Procédé de classification utilisant un modèle d&#39;apprentissage profond
WO2022075678A2 (fr) Appareil et procédé de détection de symptômes anormaux d&#39;un véhicule basés sur un apprentissage auto-supervisé en utilisant des données pseudo-normales
KR20190065989A (ko) 훈련 데이터를 생성하는 방법, 장치 및 컴퓨터-판독가능 저장 매체에 저장된 컴퓨터 프로그램
WO2024085366A1 (fr) Procédé de mise en oeuvre de mouvement de caméra au moyen d&#39;une caméra virtuelle
WO2019039758A1 (fr) Procédé de génération et d&#39;apprentissage de réseau neuronal amélioré
WO2023101417A1 (fr) Procédé permettant de prédire une précipitation sur la base d&#39;un apprentissage profond
WO2022114600A1 (fr) Procédé fondé sur l&#39;imagerie médicale pour détecter une lésion de substance blanche
Kaya et al. Binary classification of criminal tools from the images of the case using CNN
Vdovjak et al. Modern CNNs Comparison for Fire Detection in RGB Images
KR20210157795A (ko) 모델 학습 방법
WO2023219237A1 (fr) Procédé basé sur l&#39;intelligence artificielle pour évaluer une doublure
WO2024029669A1 (fr) Système de reconstruction d&#39;image antagoniste et procédé de reconstruction d&#39;image antagoniste
WO2023075351A1 (fr) Procédé d&#39;apprentissage d&#39;intelligence artificielle pour robot industriel

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18847980

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18847980

Country of ref document: EP

Kind code of ref document: A1