WO2020097795A1 - Image processing method, apparatus and device, and storage medium and program product - Google Patents

Image processing method, apparatus and device, and storage medium and program product Download PDF

Info

Publication number
WO2020097795A1
WO2020097795A1 PCT/CN2018/115252 CN2018115252W WO2020097795A1 WO 2020097795 A1 WO2020097795 A1 WO 2020097795A1 CN 2018115252 W CN2018115252 W CN 2018115252W WO 2020097795 A1 WO2020097795 A1 WO 2020097795A1
Authority
WO
WIPO (PCT)
Prior art keywords
convolution
convolution kernels
image
preset
processing
Prior art date
Application number
PCT/CN2018/115252
Other languages
French (fr)
Chinese (zh)
Inventor
阙灿
白涛
Original Assignee
北京比特大陆科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京比特大陆科技有限公司 filed Critical 北京比特大陆科技有限公司
Priority to CN201880098341.3A priority Critical patent/CN112913253A/en
Priority to PCT/CN2018/115252 priority patent/WO2020097795A1/en
Publication of WO2020097795A1 publication Critical patent/WO2020097795A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications

Definitions

  • This application relates to the field of image processing, for example, to an image processing method, device, device, storage medium, and program product.
  • Deep convolutional network is considered to be the most effective machine learning algorithm at present, and it is widely used in the technical field of image processing, such as image detection, classification, recognition and so on.
  • the image features can be extracted through the convolution layer, the extracted features are compressed through the pooling layer, the main features are extracted, and then the extracted main features are sent to the fully connected layer for calculation Output value.
  • the pooling layer in the prior art processes image features, two methods are used: average pooling and maximum pooling.
  • the average pooling refers to calculating the average value of the image area as the pooled value of the area;
  • Maximum pooling refers to selecting the maximum value of the image area as the value after pooling the area.
  • An embodiment of the present disclosure provides an image processing method, including:
  • the second quantity is greater than or equal to the first quantity.
  • An embodiment of the present disclosure also provides an image processing apparatus, including:
  • An obtaining module configured to obtain an image to be processed, and performing convolution processing on the image to be processed according to a convolution layer in a neural network to obtain a first number of image features
  • a processing module configured to perform convolution processing on the image features according to a second number of preset convolution kernels to obtain a first number of processing results
  • a determining module configured to input the processing result into a fully connected layer of the neural network, so that the fully connected layer determines an output result according to the processing result;
  • the second quantity is greater than or equal to the first quantity.
  • An embodiment of the present disclosure also provides a computer including the above-mentioned image processing device.
  • An embodiment of the present disclosure also provides a computer-readable storage medium that stores computer-executable instructions that are configured to perform the above-described image processing method.
  • An embodiment of the present disclosure also provides a computer program product.
  • the computer program product includes a computer program stored on a computer-readable storage medium.
  • the computer program includes program instructions. When the program instructions are executed by a computer, the The computer executes the image processing method described above.
  • An embodiment of the present disclosure also provides an electronic device, including:
  • At least one processor At least one processor
  • a memory communicatively connected to the at least one processor; wherein,
  • the memory stores instructions executable by the at least one processor.
  • the at least one processor executes the above-mentioned image processing method.
  • the present disclosure provides an image processing method, device, equipment, storage medium, and program product, including: acquiring an image to be processed, and performing convolution processing on the image to be processed according to a convolutional layer in a neural network to obtain a first number of image features; Perform convolution processing on the image features according to the second number of preset convolution kernels to obtain the first number of processing results; input the processing results into the fully connected layer of the neural network, so that the fully connected layer determines the output result according to the processing results; wherein, The second quantity is greater than or equal to the first quantity.
  • the pooling layer in the prior art is not used to process the image features, but a preset volume is set Convolution kernel, which performs convolution processing on image features by a preset convolution kernel to obtain a processing result including more feature information, and then fuses the processing result by a fully connected layer to determine an output result with higher accuracy.
  • FIG. 1 is an exemplary embodiment showing a structure diagram of a neural network for processing pictures in the prior art
  • FIG. 2 is a flowchart of an image processing method according to an exemplary embodiment of the present invention.
  • FIG. 3 is a flowchart of an image processing method according to another exemplary embodiment of the present invention.
  • FIG. 4 is a structural diagram of an image processing apparatus according to an exemplary embodiment of the present invention.
  • FIG. 5 is a structural diagram of an image processing apparatus shown in another exemplary embodiment of the present invention.
  • Fig. 6 is a structural diagram of an electronic device according to an exemplary embodiment of the present invention.
  • FIG. 1 is an exemplary embodiment showing a structure diagram of a neural network for processing pictures in the prior art.
  • the image is first subjected to 34-layer convolution to extract image features, and these features are averaged by the average pooling layer, and then the fully connected layer outputs the fused features. For example, if a feature output includes 9 feature values, the average pooling layer will average these nine feature values, and then input the average value to the fully connected layer. Assuming that a total of 512 features are included, the average pooling layer outputs 512 averages to the fully connected layer. This causes these picture features to be averaged, causing the accuracy of the fully connected layer to recognize pictures based on the picture features.
  • the conventional average pooling layer or the maximum pooling layer is not set, but the extracted image features are processed in the form of convolution kernels to avoid the average picture feature or the extraction of image features.
  • the maximum value leads to the problem that the accuracy of the final recognition picture decreases.
  • Fig. 2 is a flowchart of an image processing method according to an exemplary embodiment of the present invention.
  • the image processing method provided in this embodiment includes:
  • Step 101 Obtain an image to be processed, and perform convolution processing on the image to be processed according to the convolutional layer in the neural network to obtain a first number of image features.
  • the method provided in this embodiment may be executed by an electronic device with a computing function, for example, a computer, a mobile phone, a tablet computer, or the like.
  • the method provided in this embodiment may be packaged in software, and then the method provided in this embodiment may be executed through the software.
  • the method provided in this embodiment may also be set in a background server, where the background server is used to process images, and the processing result may be output through front-end software.
  • the electronic device that executes the method provided in this embodiment can acquire the image to be processed, and the image can be in a picture format (such as jpg, png, tif, gif, etc.) or a video format (such as RMVB, AVI, WMV, MPG, etc.). If the image acquired by the electronic device is in a picture format, the picture can be directly processed. If the image acquired by the electronic device is in a video format, several pictures can be extracted from the video in units of frames, and then the pictures are processed.
  • a picture format such as jpg, png, tif, gif, etc.
  • a video format such as RMVB, AVI, WMV, MPG, etc.
  • the server can obtain the image to be processed from a preset database, or the user can input the image in the front-end terminal for image processing, and then the front-end terminal Send the image to the server, so that the server can obtain the image to be processed.
  • the above-mentioned preset database may be set in the background server or in other devices, for example, it may be a removable storage device, or it may be a cloud database.
  • the electronic device used for image processing may also be a user terminal.
  • the user uploads an image in the terminal or specifies an image in the terminal.
  • the method provided in this embodiment is provided in the terminal, and the user upload or Specify the image and process it.
  • the method provided in this embodiment processes the image based on the neural network.
  • the image to be processed may be convoluted based on the convolution layer set in the neural network to obtain the first number of image features.
  • the convolutional layer of the neural network may include one convolutional layer or multiple convolutional layers. These convolutional layers can be used to extract image features and output a first number of image features. The value of the first number is related to the number of convolution kernels set in the convolutional layer. Assuming that the image to be processed is a color image, it has three channels of R, G, and B when it is input to the convolution layer. Assuming that 5 convolution kernels are set in the convolution layer, the 5 convolution kernels are used to compare R , G, and B channels of images are convolved. When a convolution kernel is used to process one channel of images, one image feature can be output, and a convolution kernel is used to perform R, G, and B channel images. The image features output by the convolution process are superimposed to obtain the image features of the convolution kernel image to be processed by convolution extraction, and then the image features of 5 images to be processed can be output.
  • Step 102 Perform convolution processing on image features according to a second number of preset convolution kernels to obtain a first number of processing results.
  • a second number of preset convolution kernels may be set for processing image features output by the convolution layer.
  • the second quantity is greater than or equal to the first quantity.
  • the specific second quantity may be an integer multiple of the first quantity. For example, if the first quantity is 512, the second quantity may also be 512, or may be a value such as 512 ⁇ 512.
  • the preset number of convolution kernels can be set according to requirements, which is not limited in this embodiment.
  • preset convolution kernels may also be trained to determine the weight values inside these convolution kernels.
  • the fully connected layer can calculate the data results and compare it with the known data results in the training data , And then adjust the weight value in the convolutional layer and the preset convolution kernel according to the comparison result.
  • the weight value in the preset convolution kernel is initially a random value. After training and learning, the weight value can be adjusted. When the ratio of the data result output by the fully connected layer to the known data result is greater than the allowable value, then The training of the preset convolution kernel can be stopped. Training rules can also be added during training to adjust the weight values in the convolution kernels so that the preset convolution kernels are not the same.
  • the first image feature of the image to be processed can be convoluted using the preset convolution kernel after training, so as to output an accurate processing result.
  • a convolution kernel corresponding to the image feature can be used to perform convolution processing on the image feature.
  • the specific convolution process is the same as the conventional convolution method. For example, if the size of the image feature is 3 ⁇ 3, the convolution kernel The size of is 3 ⁇ 3, then multiply the image feature value of the corresponding position and the weight value in the convolution kernel, and then superpose the product to obtain the processing result.
  • an image feature is The corresponding convolution kernel is The resulting processing result is 4.
  • the size of the preset convolution kernel is the same as the size of the image feature, specifically the size of the two, there is no need to set the offset parameter, otherwise, the offset parameter needs to be set.
  • the size of the preset convolution kernel is the same as the size of the image feature, if both are 3 ⁇ 3, the preset convolution kernel can complete the convolution processing of the image feature without moving. If the two are different, for example, the preset convolution kernel is 3 ⁇ 3, and the image feature size is 3 ⁇ 4, then the preset convolution kernel needs to move one step in the direction of the row, and perform two convolutions and operations to complete Convolution of image features.
  • the size of the image feature is related to the size of the convolution kernel in the convolution layer. Therefore, the size of the image feature can be determined first according to the size of the convolution kernel set in the convolution layer, and then according to the size of the image feature Preset the size of the convolution kernel, and train the preset convolution kernel.
  • the image features can be directly convoluted according to the preset convolution kernel, and the first number of processing results can be obtained.
  • the second quantity is greater than the first quantity, the convolution result can be further processed to obtain the processing result.
  • the first number is 512, including a total of 512 ⁇ 512 convolution kernels
  • the preset convolution kernels corresponding to each image feature may be 512, and the image may be checked with the corresponding preset convolution kernels The feature is convolved to obtain the convolution result, and then the convolution result is superimposed to obtain the processing result corresponding to the image feature, and finally 512 processing results are obtained.
  • the values included in each image feature can be comprehensively considered, compared with directly averaging the values in the image features or directly taking a feature
  • the maximum value of is used as the processing result, and the resulting processing result has a smaller loss in feature value, so that the data input to the fully connected layer includes more detailed image features, making the output result of the fully connected layer determined based on the input data more accurate.
  • the number of fully connected input data in the prior art is the same as the number of features output by the convolution layer. Therefore, in the method provided in this embodiment, after processing image features according to a preset convolution kernel, the first For a large number of processing results, these processing results can be directly input to the fully connected layer without changing the structure of the fully connected layer in the prior art.
  • Step 103 Input the processing result into the fully connected layer of the neural network, so that the fully connected layer determines the output result according to the processing result.
  • the processing result obtained by performing convolution processing on the image features by the preset convolution kernel can be input to the fully connected layer of the neural network, and then the fully connected layer determines the final output result. Because the processing result of the input fully connected layer includes richer image feature information, the accuracy of the output result determined by the fully connected layer is higher, and the result of image processing is more accurate.
  • the image features of the image to be processed can be extracted through the convolutional layer, and these features are independent.
  • the preset image convolution kernel can reduce and compress the extracted image features, for example, reduce the dimension of 3 ⁇ 3 image features For a value, the results after dimensionality reduction are also independent. Therefore, the fully connected layer needs to combine all the results to determine the output result of the image.
  • the output result may be the result of image recognition, classification, and detection.
  • the principle of the fully connected layer in the prior art may be used.
  • the method provided in this embodiment is used to process an image.
  • the method is performed by a device provided with the method provided in this embodiment, and the device is usually implemented in hardware and / or software.
  • An embodiment of the present disclosure provides an image processing method, including: acquiring an image to be processed, and performing convolution processing on the image to be processed according to a convolution layer in a neural network to obtain a first number of image features; and a second number of presets
  • the convolution kernel performs convolution processing on the image features to obtain the first number of processing results; the processing results are input to the fully connected layer of the neural network, so that the fully connected layer determines the output result according to the processing results; wherein, the second number is greater than or equal to the first Quantity.
  • the image features are not processed using the pooling layer in the prior art, but a preset convolution kernel is set, and the image is checked by the preset convolution kernel
  • the features are subjected to convolution processing to obtain a processing result that includes more feature information, and the processing result is fused by the fully connected layer to determine an output result with higher accuracy.
  • Fig. 3 is a flowchart of an image processing method according to another exemplary embodiment of the present invention.
  • the image processing method provided in this embodiment includes:
  • Step 201 Obtain an image to be processed, and perform convolution processing on the image to be processed according to the convolutional layer in the neural network to obtain a first number of image features.
  • step 201 and step 101 are similar, and will not be repeated here.
  • the image features of the image to be processed are extracted, the image features can be processed based on the preset convolution kernel.
  • step 2021 is performed; if the first number is N, the preset convolution kernels are N groups, and each group includes N preset volumes
  • step 2022 is performed; if the first number is N, M groups of preset convolution kernels are set in total, and each group includes N preset convolution kernels, then step 2024 is performed.
  • the size of the preset convolution kernel is the same as the size of the image feature.
  • Each preset convolution kernel has a corresponding image feature, and the preset convolution kernel has the same size as its corresponding image feature, specifically referring to the same size of data, for example, the image feature is m ⁇ n, that is, the image feature includes m rows And n columns of feature values, the size of the corresponding preset convolution kernel is also m ⁇ n, which includes m rows and n columns of weight values.
  • the offset parameter of the convolution kernel is the step size that is used when the convolution kernel is used to perform the convolution calculation on the image features. For example, if the offset parameter is 1, when the convolution kernel is used to process the image features, the second convolution calculation will Receptive field, move 1 pixel to the right or down based on the first calculation.
  • Step 2021 Perform convolution calculation according to each preset convolution kernel and corresponding image features to obtain a first number of processing results.
  • each image feature is calculated by the convolutional layer on the entire image to be processed. Therefore, each image feature represents a type of feature of the entire image. You can set different types of preset convolution kernels, and use the corresponding preset convolution kernels and image features to perform convolution calculations.
  • each preset convolution kernel and its corresponding image feature are convoluted to obtain N processing results.
  • Step 2022 Perform convolution calculation on each preset convolution kernel in each group of convolution kernels and corresponding image features to obtain N dimensionality reduction feature values corresponding to each group of convolution kernels.
  • N groups of preset convolution kernels may be set, and each group includes N preset convolution kernels.
  • each preset convolution kernel corresponds to one image feature, that is, N preset convolution kernels included in a group correspond to N image features one-to-one.
  • N preset convolution kernels in a group and corresponding N image features can be convoluted to obtain N dimensionality reduction feature values.
  • step 2023 N dimension reduction feature values corresponding to each group of convolution kernels are superimposed to obtain N processing results.
  • the number of features output by the convolutional layer is the same as the number of data of the fully connected input. If all the processing results of the N groups are directly input to the fully connected layer, the calculation amount of the fully connected layer will be excessive. Therefore, N dimensionality reduction feature values corresponding to a group of convolution kernels can be superimposed to obtain one processing result, then N groups of preset convolution kernels can obtain N processing results.
  • the image features are convolved with the preset convolution kernel to reduce the dimensionality of the image features, so that the dimensionality reduction feature values do not lose the original image feature meaning, that is, each dimensionality reduction feature value can still Represents a type of feature of a picture to be processed.
  • the N dimensionality reduction feature values in a set of convolution kernels are added to obtain a processing result, which is used to represent the overall characteristics of the image to be processed, and setting N sets of preset convolution kernels can be based on N sets of preset convolution kernels To determine the N overall features of the image to be processed.
  • Step 2024 Perform convolution calculation on each preset convolution kernel in each group of convolution kernels and corresponding image features to obtain N dimensionality reduction feature values corresponding to each group of convolution kernels.
  • M groups of preset convolution kernels may also be set, and each group includes N preset convolution kernels.
  • the N preset convolution kernels included in each group have a one-to-one correspondence with the image features, and the N convolution kernels included in each group can be used to perform convolution calculation with the image features to obtain N dimensionality reduction feature values . If M sets of preset convolution kernels are set, a total of M ⁇ N dimension reduction feature values can be obtained.
  • Step 2025 the N dimension reduction feature values corresponding to each group of convolution kernels are equally divided into T groups, and each group of dimension reduction feature values are superimposed to obtain T processing results corresponding to each group of convolution kernels.
  • the product of T and M is N.
  • the N dimensionality reduction feature values corresponding to each group of preset convolution kernels can be scored as T groups.
  • Each dimension-reduction feature value can represent a type of feature of the image to be processed, and the N dimension-reduction feature values are divided into T groups, then the T-type features can be divided into one group, and then the dimension-reduction features in each group Values are superimposed to obtain T processing results.
  • T processing results can be obtained, and if M sets of preset convolution kernels are set, T ⁇ M processing results, that is, N processing results can be obtained.
  • Each dimensionality reduction feature value can represent a type of feature information of the image to be processed.
  • T groups of dimensionality reduction feature values can be obtained, and then the features of each group can be superimposed to obtain T images to be processed Comprehensive feature information.
  • T ⁇ M comprehensive feature information can be obtained.
  • step 203 may be executed.
  • Step 203 Input the processing result into the fully connected layer of the neural network, so that the fully connected layer determines the output result according to the processing result.
  • step 203 and step 103 are similar, and will not be repeated here.
  • the preset convolution kernel before using the preset convolution kernel to process the image to be processed, the preset convolution kernel needs to be trained to determine the weight value inside each preset convolution kernel.
  • the weight value in the preset convolution kernel needs to be randomly generated. During the training process, the weight value of the preset convolution kernel will also be adjusted. After the weight value is randomly generated, or After adjusting the weight value every / several times, the method provided in this embodiment may further include:
  • the convolution kernel is normalized according to the vector to enhance the difference between the convolution kernels.
  • the vector corresponding to the preset convolution kernel can be determined according to the weight value. Then determine whether each weight value is similar according to the vector. If it is similar, adjust the weight value in the corresponding preset convolution kernel to make the two vectors have a certain difference.
  • the weight value in the preset convolution kernel can be determined as a one-dimensional vector. For example, if the convolution kernel is 3 ⁇ 3, a total of 9 weight values can be included. These weight values can be spliced in the order of rows. That is, the first row, the second row, and the third row are spliced in sequence, and a one-dimensional vector can be obtained. The corresponding one-dimensional vector can be generated according to the weight value of each preset convolution kernel.
  • the inner product of every two vectors can be calculated, and whether the two vectors are similar according to the inner product result can be determined.
  • the inner product is equal to the product of the magnitude of the two vectors and the cosine of the angle. It can be considered that the smaller the inner product, the greater the angle between the two vectors, and the greater the difference between the two. Therefore, it can be determined whether the two are similar based on the inner product of the vector corresponding to the preset convolution kernel.
  • Fig. 4 is a structural diagram of an image processing apparatus according to an exemplary embodiment of the present invention.
  • the image processing apparatus provided in this embodiment includes:
  • the obtaining module 41 is configured to obtain an image to be processed, and perform convolution processing on the image to be processed according to a convolutional layer in a neural network to obtain a first number of image features;
  • the processing module 42 is configured to perform convolution processing on the image features according to a second number of preset convolution kernels to obtain a first number of processing results;
  • the determining module 43 is configured to input the processing result into the fully connected layer of the neural network, so that the fully connected layer determines the output result according to the processing result;
  • the second quantity is greater than or equal to the first quantity.
  • the image processing device includes an acquisition module for acquiring an image to be processed, and performing a convolution process on the image to be processed according to a convolutional layer in a neural network to obtain a first number of image features; a processing module is used for The second number of preset convolution kernels perform convolution processing on the image features to obtain the first number of processing results; the determination module is used to input the processing results into the fully connected layer of the neural network, so that the fully connected layer determines the output according to the processing results Results; where the second quantity is greater than or equal to the first quantity.
  • the image features are not processed using the pooling layer in the prior art, but a preset convolution kernel is set, and the image is checked by the preset convolution kernel
  • the features are subjected to convolution processing to obtain a processing result that includes more feature information, and the processing result is fused by the fully connected layer to determine an output result with higher accuracy.
  • Fig. 5 is a structural diagram of an image processing apparatus according to another exemplary embodiment of the present invention.
  • the size of the preset convolution kernel is the same as the size of the image feature.
  • the second quantity is the same as the first quantity
  • the processing module 42 includes:
  • the first processing unit 421 is configured to perform convolution calculation according to each of the preset convolution kernels and the corresponding image features to obtain a first number of the processing results.
  • the second number of preset convolution kernels includes N groups of preset convolution kernels, and each group includes N preset convolution kernels, where N is equal to the first number;
  • the processing module 42 includes a second processing unit 422 for:
  • N dimensionality reduction feature values corresponding to each set of convolution kernels are superimposed to obtain N processing results.
  • the second number of preset convolution kernels includes M groups of convolution kernels, and each group includes N convolution kernels, where N is equal to the first number;
  • the processing module 42 includes a third processing unit 423 for:
  • N dimensionality reduction feature values corresponding to each group of convolution kernels are equally divided into T groups, and the dimensionality reduction feature values of each group are superimposed to obtain T processing results corresponding to each group of convolution kernels;
  • the product of T and M is N.
  • each of the preset convolution kernels does not include offset parameters.
  • the device provided in this embodiment further includes an adjustment module 44 for:
  • the convolution kernel is normalized according to the vector to enhance the difference between the convolution kernels.
  • the adjustment module 44 is specifically used to:
  • An embodiment of the present disclosure also provides a computer including the above-mentioned image processing device.
  • An embodiment of the present disclosure also provides a computer-readable storage medium that stores computer-executable instructions that are configured to perform the above-described image processing method.
  • An embodiment of the present disclosure also provides a computer program product.
  • the computer program product includes a computer program stored on a computer-readable storage medium.
  • the computer program includes program instructions. When the program instructions are executed by a computer, the The computer executes the above image processing method.
  • the aforementioned computer-readable storage medium may be a transient computer-readable storage medium or a non-transitory computer-readable storage medium.
  • Fig. 6 is a structural diagram of an electronic device according to an exemplary embodiment of the present invention.
  • An embodiment of the present disclosure also provides an electronic device, whose structure is shown in FIG. 6, the electronic device includes:
  • At least one processor (processor) 60 one processor 60 is taken as an example in FIG. 6; and the memory (memory) 61 may further include a communication interface (Communication Interface) 62 and a bus 63.
  • the processor 60, the communication interface 62, and the memory 61 can complete communication with each other through the bus 63.
  • the communication interface 62 may be used for information transmission.
  • the processor 60 may call logic instructions in the memory 61 to execute the image processing method of the above-mentioned embodiment.
  • logic instructions in the aforementioned memory 61 can be implemented in the form of software functional units and sold or used as an independent product, and can be stored in a computer-readable storage medium.
  • the memory 61 is a computer-readable storage medium that can be used to store software programs and computer-executable programs, such as program instructions / modules corresponding to the methods in the embodiments of the present disclosure.
  • the processor 60 executes functional applications and data processing by running software programs, instructions, and modules stored in the memory 61, that is, implementing the image processing method in the above method embodiments.
  • the memory 61 may include a storage program area and a storage data area, wherein the storage program area may store an operating system and application programs required for at least one function; the storage data area may store data created according to the use of a terminal device and the like.
  • the memory 61 may include a high-speed random access memory, and may also include a non-volatile memory.
  • the technical solutions of the embodiments of the present disclosure may be embodied in the form of software products, which are stored in a storage medium and include one or more instructions to make a computer device (which may be a personal computer, server, or network) Equipment, etc.) to perform all or part of the steps of the method described in the embodiments of the present disclosure.
  • the aforementioned storage medium may be a non-transitory storage medium, including: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disk or optical disk, etc.
  • a medium that can store program codes may also be a transient storage medium.
  • first, second, etc. may be used in this application to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another.
  • the first element can be called the second element, and likewise, the second element can be called the first element, as long as all occurrences of the "first element” are consistently renamed and all occurrences of The “second component” can be renamed consistently.
  • the first element and the second element are both elements, but they may not be the same element.
  • the various aspects, implementations, implementations, or features in the described embodiments can be used alone or in any combination.
  • Various aspects in the described embodiments may be implemented by software, hardware, or a combination of software and hardware.
  • the described embodiments may also be embodied by a computer-readable medium that stores computer-readable code including instructions executable by at least one computing device.
  • the computer-readable medium can be associated with any data storage device capable of storing data, which can be read by a computer system.
  • Computer-readable media used for examples may include read-only memory, random access memory, CD-ROM, HDD, DVD, magnetic tape, optical data storage devices, and the like.
  • the computer-readable medium may also be distributed in computer systems connected through a network, so that computer-readable codes can be stored and executed in a distributed manner.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

Embodiments of the present invention relate to an image processing method, apparatus and device, and a storage medium and a program product. The method comprises: acquiring an image to be processed, convolving the image to be processed according to a convolution layer in a neutral network to obtain a first number of image features; checking the image features for convolution according to a second number of preset convolution kernels to obtain a first number of processing results; inputting the processing results into a full connection layer of the neutral network so that the full connection layer determines an output result according to the processing results; wherein, the second number is greater than the first number. In the solutions provided by the embodiments of the present invention, after the image features are extracted through the convolution layer, the image features are not processed by a pooling layer in the prior art, instead, the preset convolution kernels are set to check the image features for convolution to obtain processing results containing more feature information, then the processing results are fused by the full connection layer to determine a more accurate output result.

Description

图像处理方法、装置、设备、存储介质及程序产品Image processing method, device, equipment, storage medium and program product 技术领域Technical field
本申请涉及图像处理领域,例如涉及一种图像处理方法、装置、设备、存储介质及程序产品。This application relates to the field of image processing, for example, to an image processing method, device, device, storage medium, and program product.
背景技术Background technique
深度卷积网络被认为是目前最有效的机器学习算法,广泛应用于图像处理的技术领域,例如图像的检测、分类、识别等等。Deep convolutional network is considered to be the most effective machine learning algorithm at present, and it is widely used in the technical field of image processing, such as image detection, classification, recognition and so on.
目前,通过神经网络对图像进行处理时,可以通过卷积层提取图像特征,通过池化层对提取的特征进行压缩,提取其中的主要特征,再将提取的主要特征发送给全连接层,计算出输出值。现有技术中的池化层对图像特征进行处理时,采用的方式是平均池化和最大池化两种方式,平均池化是指计算图像区域的平均值作为该区域池化后的值;最大池化是指选图像区域的最大值作为该区域池化后的值。At present, when processing an image through a neural network, the image features can be extracted through the convolution layer, the extracted features are compressed through the pooling layer, the main features are extracted, and then the extracted main features are sent to the fully connected layer for calculation Output value. When the pooling layer in the prior art processes image features, two methods are used: average pooling and maximum pooling. The average pooling refers to calculating the average value of the image area as the pooled value of the area; Maximum pooling refers to selecting the maximum value of the image area as the value after pooling the area.
但是,现有技术中的这种提取所有的特征取平均值或最大值的处理方式,会导致提取的特征差异性不明显,造成神经网络在图像处理过程中输出的结果不准确的问题。However, in the prior art, the processing method of extracting all the features to take the average value or the maximum value will cause the difference in the extracted features to be insignificant, resulting in the problem of inaccurate output results of the neural network during the image processing process.
发明内容Summary of the invention
本公开实施例提供了一种图像处理方法,包括:An embodiment of the present disclosure provides an image processing method, including:
获取待处理图像,根据神经网络中的卷积层对所述待处理图像进行卷积处理,得到第一数量的图像特征;Obtain the image to be processed, and perform convolution on the image to be processed according to the convolution layer in the neural network to obtain a first number of image features;
根据第二数量的预设卷积核对所述图像特征进行卷积处理,得到第一数量的处理结果;Performing convolution processing on the image features according to a second number of preset convolution kernels to obtain a first number of processing results;
将所述处理结果输入所述神经网络的全连接层,以使所述全连接层根据所述处理结果确定输出结果;Input the processing result into the fully connected layer of the neural network, so that the fully connected layer determines the output result according to the processing result;
其中,所述第二数量大于等于所述第一数量。Wherein, the second quantity is greater than or equal to the first quantity.
本公开实施例还提供了一种图像处理装置,包括:An embodiment of the present disclosure also provides an image processing apparatus, including:
获取模块,用于获取待处理图像,根据神经网络中的卷积层对所述待处理图像进行卷积处理,得到第一数量的图像特征;An obtaining module, configured to obtain an image to be processed, and performing convolution processing on the image to be processed according to a convolution layer in a neural network to obtain a first number of image features;
处理模块,用于根据第二数量的预设卷积核对所述图像特征进行卷积处理,得到第一数量的处理结果;A processing module, configured to perform convolution processing on the image features according to a second number of preset convolution kernels to obtain a first number of processing results;
确定模块,用于将所述处理结果输入所述神经网络的全连接层,以使所述全连接层根据所述处理结果确定输出结果;A determining module, configured to input the processing result into a fully connected layer of the neural network, so that the fully connected layer determines an output result according to the processing result;
其中,所述第二数量大于等于所述第一数量。Wherein, the second quantity is greater than or equal to the first quantity.
本公开实施例还提供了一种计算机,包含上述的图像处理装置。An embodiment of the present disclosure also provides a computer including the above-mentioned image processing device.
本公开实施例还提供了一种计算机可读存储介质,存储有计算机可执行指令,所述计算机可执行指令设置为执行上述的图像处理方法。An embodiment of the present disclosure also provides a computer-readable storage medium that stores computer-executable instructions that are configured to perform the above-described image processing method.
本公开实施例还提供了一种计算机程序产品,所述计算机程序产品包括存储在计算机可读存储介质上的计算机程序,所述计算机程序包括程序指令,当所述程序指令被计算机执行时,使所述计算机执行上述的图像处理方法。An embodiment of the present disclosure also provides a computer program product. The computer program product includes a computer program stored on a computer-readable storage medium. The computer program includes program instructions. When the program instructions are executed by a computer, the The computer executes the image processing method described above.
本公开实施例还提供了一种电子设备,包括:An embodiment of the present disclosure also provides an electronic device, including:
至少一个处理器;以及At least one processor; and
与所述至少一个处理器通信连接的存储器;其中,A memory communicatively connected to the at least one processor; wherein,
所述存储器存储有可被所述至少一个处理器执行的指令,所述指令被所述至少一个处理器执行时,使所述至少一个处理器执行上述的图像处理方法。The memory stores instructions executable by the at least one processor. When the instructions are executed by the at least one processor, the at least one processor executes the above-mentioned image processing method.
本公开提供一种图像处理方法、装置、设备、存储介质及程序产品,包括:获取待处理图像,根据神经网络中的卷积层对待处理图像进行卷积处理,得到第一数量的图像特征;根据第二数量的预设卷积核对图像特征进行卷积处理,得到第一数量的处理结果;将处理结果输入神经网络的全连接层,以使全连接层根据处理结果确定输出结果;其中,第二数量大于等于第一数量。本实施例提供的方法、装置、设备、存储介质及程序产品中,在通过卷积层提取图像特征后,不采用现有技术中的池化层对图像特征进行处理,而是设置预设卷积核,由预设卷积核对图像特征进行卷积处理,得到包括更多特征 信息的处理结果,再由全连接层对处理结果进行融合,从而确定出精度更高的输出结果。The present disclosure provides an image processing method, device, equipment, storage medium, and program product, including: acquiring an image to be processed, and performing convolution processing on the image to be processed according to a convolutional layer in a neural network to obtain a first number of image features; Perform convolution processing on the image features according to the second number of preset convolution kernels to obtain the first number of processing results; input the processing results into the fully connected layer of the neural network, so that the fully connected layer determines the output result according to the processing results; wherein, The second quantity is greater than or equal to the first quantity. In the method, device, device, storage medium, and program product provided in this embodiment, after the image features are extracted through the convolution layer, the pooling layer in the prior art is not used to process the image features, but a preset volume is set Convolution kernel, which performs convolution processing on image features by a preset convolution kernel to obtain a processing result including more feature information, and then fuses the processing result by a fully connected layer to determine an output result with higher accuracy.
附图说明BRIEF DESCRIPTION
一个或多个实施例通过与之对应的附图进行示例性说明,这些示例性说明和附图并不构成对实施例的限定,附图中具有相同参考数字标号的元件示为类似的元件,附图不构成比例限制,并且其中:One or more embodiments are exemplified by the corresponding drawings. These exemplary descriptions and the drawings do not constitute a limitation on the embodiments. Elements with the same reference numerals in the drawings are shown as similar elements. The drawings do not constitute a proportional limitation, and among them:
图1为一示例性实施例示出现有技术中用于对图片进行处理的神经网络结构图;FIG. 1 is an exemplary embodiment showing a structure diagram of a neural network for processing pictures in the prior art;
图2为本发明一示例性实施例示出的图像处理方法的流程图;2 is a flowchart of an image processing method according to an exemplary embodiment of the present invention;
图3为本发明另一示例性实施例示出的图像处理方法的流程图;FIG. 3 is a flowchart of an image processing method according to another exemplary embodiment of the present invention;
图4为本发明一示例性实施例示出的图像处理装置的结构图;4 is a structural diagram of an image processing apparatus according to an exemplary embodiment of the present invention;
图5为本发明另一示例性实施例示出的图像处理装置的结构图;5 is a structural diagram of an image processing apparatus shown in another exemplary embodiment of the present invention;
图6为本发明一示例性实施例示出的电子设备的结构图。Fig. 6 is a structural diagram of an electronic device according to an exemplary embodiment of the present invention.
具体实施方式detailed description
为了能够更加详尽地了解本公开实施例的特点与技术内容,下面结合附图对本公开实施例的实现进行详细阐述,所附附图仅供参考说明之用,并非用来限定本公开实施例。在以下的技术描述中,为方便解释起见,通过多个细节以提供对所披露实施例的充分理解。然而,在没有这些细节的情况下,一个或多个实施例仍然可以实施。在其它情况下,为简化附图,熟知的结构和装置可以简化展示。In order to understand the features and technical contents of the embodiments of the present disclosure in more detail, the following describes the implementation of the embodiments of the present disclosure in detail with reference to the drawings. The accompanying drawings are for reference only and are not intended to limit the embodiments of the present disclosure. In the following technical description, for convenience of explanation, various details are provided to provide a sufficient understanding of the disclosed embodiments. However, without these details, one or more embodiments can still be implemented. In other cases, to simplify the drawings, well-known structures and devices can be simplified.
图1为一示例性实施例示出现有技术中用于对图片进行处理的神经网络结构图。FIG. 1 is an exemplary embodiment showing a structure diagram of a neural network for processing pictures in the prior art.
如图1所示,以现有技术中的ResNet为例,图片先经过34层卷积提取图像特征,这些特征经过平均池化层取平均值后,再由全连接层输出融合后的特征。例如,输出的一个特征共包括9个特征值,则平均池化层会将这九 个特征值进行平均,然后将平均值输入到全连接层。假设共包括512个特征,则平均池化层向全连接层输出512个平均值。这就导致这些图片特征被平均化,造成全连接层根据图片特征识别图片的精度降低。As shown in FIG. 1, taking ResNet in the prior art as an example, the image is first subjected to 34-layer convolution to extract image features, and these features are averaged by the average pooling layer, and then the fully connected layer outputs the fused features. For example, if a feature output includes 9 feature values, the average pooling layer will average these nine feature values, and then input the average value to the fully connected layer. Assuming that a total of 512 features are included, the average pooling layer outputs 512 averages to the fully connected layer. This causes these picture features to be averaged, causing the accuracy of the fully connected layer to recognize pictures based on the picture features.
本实施例公开的方案中,不设置常规的平均池化层或最大池化层,而是采用卷积核的方式对提取的图像特征进行处理,避免之间平均图片特征或提取图片特征中的最大值,导致最终识别图片的精度下降的问题。In the solution disclosed in this embodiment, the conventional average pooling layer or the maximum pooling layer is not set, but the extracted image features are processed in the form of convolution kernels to avoid the average picture feature or the extraction of image features. The maximum value leads to the problem that the accuracy of the final recognition picture decreases.
图2为本发明一示例性实施例示出的图像处理方法的流程图。Fig. 2 is a flowchart of an image processing method according to an exemplary embodiment of the present invention.
如图2所示,本实施例提供的图像处理方法包括:As shown in FIG. 2, the image processing method provided in this embodiment includes:
步骤101,获取待处理图像,根据神经网络中的卷积层对待处理图像进行卷积处理,得到第一数量的图像特征。Step 101: Obtain an image to be processed, and perform convolution processing on the image to be processed according to the convolutional layer in the neural network to obtain a first number of image features.
其中,可以由具备计算功能的电子设备执行本实施例提供的方法,例如是计算机、手机、平板电脑等。可以将本实施例提供的方法封装在软件中,进而通过软件执行本实施例提供的方法。Among them, the method provided in this embodiment may be executed by an electronic device with a computing function, for example, a computer, a mobile phone, a tablet computer, or the like. The method provided in this embodiment may be packaged in software, and then the method provided in this embodiment may be executed through the software.
具体的,还可以将本实施例提供的方法设置在后台服务器中,该后台服务器用于处理图像,可以通过前端的软件输出处理结果。Specifically, the method provided in this embodiment may also be set in a background server, where the background server is used to process images, and the processing result may be output through front-end software.
进一步的,执行本实施例提供的方法的电子设备可以获取待处理图像,该图像可以是图片格式(如jpg,png,tif,gif等),也可以是视频格式(如RMVB、AVI、WMV、MPG等)。若电子设备获取的图像是图片格式,则可以直接对该图片进行处理。若电子设备获取的图像是视频格式,则可以以帧为单位从视频中提取出若干张图片,再对图片进行处理。Further, the electronic device that executes the method provided in this embodiment can acquire the image to be processed, and the image can be in a picture format (such as jpg, png, tif, gif, etc.) or a video format (such as RMVB, AVI, WMV, MPG, etc.). If the image acquired by the electronic device is in a picture format, the picture can be directly processed. If the image acquired by the electronic device is in a video format, several pictures can be extracted from the video in units of frames, and then the pictures are processed.
实际应用时,若电子设备是用于图像处理的后台服务器,则该服务器可以从预设数据库中获取待处理图像,还可以由用户在用于图像处理的前端终端中输入图像,再由前端终端将图像发送给该服务器,从而使服务器获取待处理图像。上述预设数据库可以设置在后台服务器中,也可以设置在其他设备中,例如可以是可移动的存储装置,还可以是云端数据库。In practical applications, if the electronic device is a background server for image processing, the server can obtain the image to be processed from a preset database, or the user can input the image in the front-end terminal for image processing, and then the front-end terminal Send the image to the server, so that the server can obtain the image to be processed. The above-mentioned preset database may be set in the background server or in other devices, for example, it may be a removable storage device, or it may be a cloud database.
其中,用于图像处理的电子设备也可以是用户终端,用户在终端中上传图像,或者在终端中指定图像,该终端中设置有本实施例提供的方法,能够根据用户的操作获取用户上传或指定的图像,并对其进行处理。The electronic device used for image processing may also be a user terminal. The user uploads an image in the terminal or specifies an image in the terminal. The method provided in this embodiment is provided in the terminal, and the user upload or Specify the image and process it.
具体的,本实施例提供的方法基于神经网络对图像进行处理。在获取完待处理图像后,可以基于神经网络中设置的卷积层对待处理图像进行卷积处理,得到第一数量的图像特征。Specifically, the method provided in this embodiment processes the image based on the neural network. After acquiring the image to be processed, the image to be processed may be convoluted based on the convolution layer set in the neural network to obtain the first number of image features.
进一步的,神经网络的卷积层可以包括一个卷积层,也可以包括多个卷积层。可以通过这些卷积层提取图像特征,并输出第一数量的图像特征。第一数量的值与卷积层中设置的卷积核数量相关。假设待处理图像是彩色图像,则在其输入卷积层时具有R、G、B三个通道,假设卷积层中设置有5个卷积核,则通过这5个卷积核分别对R、G、B三个通道的图像进行卷积处理,使用一个卷积核对一个通道的图像进行处理时,能够输出1个图像特征,将一个卷积核对R、G、B三个通道的图像进行卷积处理输出的图像特征进行叠加,就得到该卷积核待处理图像进行卷积提取的图像特征,进而能够输出5个待处理图像的图像特征。Further, the convolutional layer of the neural network may include one convolutional layer or multiple convolutional layers. These convolutional layers can be used to extract image features and output a first number of image features. The value of the first number is related to the number of convolution kernels set in the convolutional layer. Assuming that the image to be processed is a color image, it has three channels of R, G, and B when it is input to the convolution layer. Assuming that 5 convolution kernels are set in the convolution layer, the 5 convolution kernels are used to compare R , G, and B channels of images are convolved. When a convolution kernel is used to process one channel of images, one image feature can be output, and a convolution kernel is used to perform R, G, and B channel images. The image features output by the convolution process are superimposed to obtain the image features of the convolution kernel image to be processed by convolution extraction, and then the image features of 5 images to be processed can be output.
步骤102,根据第二数量的预设卷积核对图像特征进行卷积处理,得到第一数量的处理结果。Step 102: Perform convolution processing on image features according to a second number of preset convolution kernels to obtain a first number of processing results.
实际应用时,可以设置第二数量的预设卷积核,用于对卷积层输出的图像特征进行处理。第二数量大于等于第一数量,具体第二数量可以是第一数量的整数倍,例如,若第一数量是512,则第二数量也可以是512,还可以是512×512等数值。In actual application, a second number of preset convolution kernels may be set for processing image features output by the convolution layer. The second quantity is greater than or equal to the first quantity. The specific second quantity may be an integer multiple of the first quantity. For example, if the first quantity is 512, the second quantity may also be 512, or may be a value such as 512 × 512.
其中,可以根据需求设置预设卷积核的数量,本实施例不对此进行限制。Among them, the preset number of convolution kernels can be set according to requirements, which is not limited in this embodiment.
具体的,在使用本实施例提供的方法对待处理图片进行处理前,还可以对预设卷积核进行训练,从而确定这些卷积核内部的权重值。Specifically, before the image to be processed is processed using the method provided in this embodiment, preset convolution kernels may also be trained to determine the weight values inside these convolution kernels.
进一步的,可以预先搭建卷积层、预设卷积核以及全连接层,并将训练数据输入卷积层,从而使卷积层对训练数据进行处理,提取数据特征,再将数据特征输入预设卷积核,通过预设卷积核度数据特征进行卷积处理,并将处理结果输出至全连接层,全连接层可以计算出数据结果,并与训练数据中的已知数据结果比对,再根据比对结果调整卷积层以及预设卷积核中的权重值。Further, you can build a convolutional layer, a preset convolution kernel, and a fully connected layer in advance, and input the training data into the convolutional layer, so that the convolutional layer processes the training data, extracts the data features, and then enters the data features into the Set up a convolution kernel, perform convolution processing with preset convolution kernel data features, and output the processing results to the fully connected layer. The fully connected layer can calculate the data results and compare it with the known data results in the training data , And then adjust the weight value in the convolutional layer and the preset convolution kernel according to the comparison result.
实际应用时,预设卷积核内的权重值最初为随机值,经过训练学习,能够调整其中的权重值,当全连接层输出的数据结果与已知数据结果一致的比 例大于允许值,则可以停止对预设卷积核的训练。还可以在训练时加入训练规则,用于调整卷积核内的权重值,使得各个预设卷积核之间不相同。In actual application, the weight value in the preset convolution kernel is initially a random value. After training and learning, the weight value can be adjusted. When the ratio of the data result output by the fully connected layer to the known data result is greater than the allowable value, then The training of the preset convolution kernel can be stopped. Training rules can also be added during training to adjust the weight values in the convolution kernels so that the preset convolution kernels are not the same.
其中,在对待处理图像进行处理时,可以使用训练完毕的预设卷积核对待处理图片的第一图像特征进行卷积处理,从而输出准确的处理结果。Wherein, when processing the image to be processed, the first image feature of the image to be processed can be convoluted using the preset convolution kernel after training, so as to output an accurate processing result.
具体的,可以使用与图像特征对应的卷积核对图像特征进行卷积处理,具体的卷积过程与现有技术的卷积方式相同,例如,图像特征的尺寸是3×3的,卷积核的尺寸是3×3,则将对应位置的图像特征值以及卷积核内的权重值相乘,再将乘积叠加,得到处理结果。假设一个图像特征为
Figure PCTCN2018115252-appb-000001
与其对应的卷积核为
Figure PCTCN2018115252-appb-000002
得到的处理结果为4。
Specifically, a convolution kernel corresponding to the image feature can be used to perform convolution processing on the image feature. The specific convolution process is the same as the conventional convolution method. For example, if the size of the image feature is 3 × 3, the convolution kernel The size of is 3 × 3, then multiply the image feature value of the corresponding position and the weight value in the convolution kernel, and then superpose the product to obtain the processing result. Suppose an image feature is
Figure PCTCN2018115252-appb-000001
The corresponding convolution kernel is
Figure PCTCN2018115252-appb-000002
The resulting processing result is 4.
进一步的,若预设卷积核的大小与图像特征的大小相同,具体指二者的尺寸大小,则无需设置偏置参数,否则,需要设置偏置参数。若预设卷积核的大小与图像特征的大小相同,如都是3×3的,则预设卷积核不需要移动,就能够完成对图像特征的卷积处理。若二者不同,例如预设卷积核是3×3的,图像特征大小是3×4的,则预设卷积核需要在行的方向移动一步,进行两次卷及操作,才能够完成对图像特征的卷积处理。Further, if the size of the preset convolution kernel is the same as the size of the image feature, specifically the size of the two, there is no need to set the offset parameter, otherwise, the offset parameter needs to be set. If the size of the preset convolution kernel is the same as the size of the image feature, if both are 3 × 3, the preset convolution kernel can complete the convolution processing of the image feature without moving. If the two are different, for example, the preset convolution kernel is 3 × 3, and the image feature size is 3 × 4, then the preset convolution kernel needs to move one step in the direction of the row, and perform two convolutions and operations to complete Convolution of image features.
实际应用时,图像特征的尺寸大小与卷积层中的卷积核尺寸相关,因此,可以根据卷积层中设置的卷积核尺寸先确定图像特征的尺寸,再根据图像特征的尺寸大小确定预设卷积核的大小,并对预设卷积核进行训练。In actual application, the size of the image feature is related to the size of the convolution kernel in the convolution layer. Therefore, the size of the image feature can be determined first according to the size of the convolution kernel set in the convolution layer, and then according to the size of the image feature Preset the size of the convolution kernel, and train the preset convolution kernel.
其中,若第一数量与第二数量相同,则可以直接根据预设卷积核对图像特征进行卷积操作,就能够得到第一数量的处理结果。若第二数量大于第一数量,则可以进一步的对卷积结果进行处理,再得到处理结果。例如,若第一数量是512,共包括512×512个卷积核,则其中与每个图像特征对应的预设卷积核可以是512个,可以用与之对应的预设卷积核对图像特征进行卷积处理得到卷积结果,再将卷积结果叠加,得到该图像特征对应的处理结果,最终得到512个处理结果。Wherein, if the first number is the same as the second number, the image features can be directly convoluted according to the preset convolution kernel, and the first number of processing results can be obtained. If the second quantity is greater than the first quantity, the convolution result can be further processed to obtain the processing result. For example, if the first number is 512, including a total of 512 × 512 convolution kernels, the preset convolution kernels corresponding to each image feature may be 512, and the image may be checked with the corresponding preset convolution kernels The feature is convolved to obtain the convolution result, and then the convolution result is superimposed to obtain the processing result corresponding to the image feature, and finally 512 processing results are obtained.
通过预设卷积核对图像特征进行处理,能够将每个图像特征中包括的值进行综合考虑,相较于现有技术中直接对图像特征中的值取平均值,或是直 接取一个特征中的最大值作为处理结果,得到的处理结果在特征值上损失更小,从而输入全连接层的数据包括了更详细的图像特征,使得全连接层根据输入数据确定的输出结果精度更高。By processing the image features through preset convolution kernels, the values included in each image feature can be comprehensively considered, compared with directly averaging the values in the image features or directly taking a feature The maximum value of is used as the processing result, and the resulting processing result has a smaller loss in feature value, so that the data input to the fully connected layer includes more detailed image features, making the output result of the fully connected layer determined based on the input data more accurate.
具体的,现有技术中的全连接输入数据数量,与卷积层输出的特征数量相同,因此,本实施例提供的方法中,根据预设卷积核对图像特征进行处理后,能够得到第一数量的处理结果,可以直接将这些处理结果输入到全连接层,不需要改变现有技术中的全连接层的结构。Specifically, the number of fully connected input data in the prior art is the same as the number of features output by the convolution layer. Therefore, in the method provided in this embodiment, after processing image features according to a preset convolution kernel, the first For a large number of processing results, these processing results can be directly input to the fully connected layer without changing the structure of the fully connected layer in the prior art.
步骤103,将处理结果输入神经网络的全连接层,以使全连接层根据处理结果确定输出结果。Step 103: Input the processing result into the fully connected layer of the neural network, so that the fully connected layer determines the output result according to the processing result.
可以将预设卷积核对图像特征进行卷积处理得到的处理结果输入神经网络的全连接层,再由全连接层确定最终的输出结果。由于输入全连接层的处理结果包括了更丰富的图像特征信息,使得全连接层确定的输出结果精度更高,对图像处理的结果更为准确。The processing result obtained by performing convolution processing on the image features by the preset convolution kernel can be input to the fully connected layer of the neural network, and then the fully connected layer determines the final output result. Because the processing result of the input fully connected layer includes richer image feature information, the accuracy of the output result determined by the fully connected layer is higher, and the result of image processing is more accurate.
通过卷积层能够提取出待处理图像的图像特征,这些特征之间是独立的,通过预设卷积核可以对提取的图像特征进行降维、压缩,例如将3×3的图像特征降维为一个值,降维后的结果之间也是独立,因此,需要全连接层结合所有的结果,确定出图像的输出结果。该输出结果可以是图像的识别、分类、检测结果。本实施例提供的方法中,可以采用现有技术中的全连接层原理。The image features of the image to be processed can be extracted through the convolutional layer, and these features are independent. The preset image convolution kernel can reduce and compress the extracted image features, for example, reduce the dimension of 3 × 3 image features For a value, the results after dimensionality reduction are also independent. Therefore, the fully connected layer needs to combine all the results to determine the output result of the image. The output result may be the result of image recognition, classification, and detection. In the method provided in this embodiment, the principle of the fully connected layer in the prior art may be used.
本实施例提供的方法用于对图像进行处理,该方法由设置有本实施例提供的方法的设备执行,该设备通常以硬件和/或软件的方式来实现。The method provided in this embodiment is used to process an image. The method is performed by a device provided with the method provided in this embodiment, and the device is usually implemented in hardware and / or software.
本公开实施例提供了一种图像处理方法,包括:获取待处理图像,根据神经网络中的卷积层对待处理图像进行卷积处理,得到第一数量的图像特征;根据第二数量的预设卷积核对图像特征进行卷积处理,得到第一数量的处理结果;将处理结果输入神经网络的全连接层,以使全连接层根据处理结果确定输出结果;其中,第二数量大于等于第一数量。本实施例提供的方法中,在通过卷积层提取图像特征后,不采用现有技术中的池化层对图像特征进行处理,而是设置预设卷积核,由预设卷积核对图像特征进行卷积处理,得到包括更多特征信息的处理结果,再由全连接层对处理结果进行融合,从而确定出精度更高的输出结果。An embodiment of the present disclosure provides an image processing method, including: acquiring an image to be processed, and performing convolution processing on the image to be processed according to a convolution layer in a neural network to obtain a first number of image features; and a second number of presets The convolution kernel performs convolution processing on the image features to obtain the first number of processing results; the processing results are input to the fully connected layer of the neural network, so that the fully connected layer determines the output result according to the processing results; wherein, the second number is greater than or equal to the first Quantity. In the method provided in this embodiment, after the image features are extracted through the convolution layer, the image features are not processed using the pooling layer in the prior art, but a preset convolution kernel is set, and the image is checked by the preset convolution kernel The features are subjected to convolution processing to obtain a processing result that includes more feature information, and the processing result is fused by the fully connected layer to determine an output result with higher accuracy.
图3为本发明另一示例性实施例示出的图像处理方法的流程图。Fig. 3 is a flowchart of an image processing method according to another exemplary embodiment of the present invention.
如图3所示,本实施例提供的图像处理方法,包括:As shown in FIG. 3, the image processing method provided in this embodiment includes:
步骤201,获取待处理图像,根据神经网络中的卷积层对待处理图像进行卷积处理,得到第一数量的图像特征。Step 201: Obtain an image to be processed, and perform convolution processing on the image to be processed according to the convolutional layer in the neural network to obtain a first number of image features.
步骤201与步骤101的具体原理和实现方式类似,此处不再赘述。The specific principles and implementations of step 201 and step 101 are similar, and will not be repeated here.
提取到待处理图像的图像特征后,可以基于预设的卷积核对图像特征进行处理。After the image features of the image to be processed are extracted, the image features can be processed based on the preset convolution kernel.
若设置的预设卷积核数量第二数量与第一数量相同,则执行步骤2021;若第一数量为N,设置的预设卷积核为N组,且每组包括N个预设卷积核,则执行步骤2022;若第一数量为N,共设置M组预设卷积核,每组包括N个预设卷积核,则执行步骤2024。If the second number of preset convolution kernels is the same as the first number, step 2021 is performed; if the first number is N, the preset convolution kernels are N groups, and each group includes N preset volumes For the accumulation kernel, step 2022 is performed; if the first number is N, M groups of preset convolution kernels are set in total, and each group includes N preset convolution kernels, then step 2024 is performed.
本实施例提供的方法中,预设卷积核的大小与图像特征的大小相同。每个预设卷积核具有对应的图像特征,该预设卷积核与其对应的图像特征的大小相同,具体是指数据的尺寸相同,例如图像特征是m×n,即图像特征包括m行、n列特征值,则与其对应的预设卷积核的大小也是m×n,其内部包括m行、n列权重值。In the method provided in this embodiment, the size of the preset convolution kernel is the same as the size of the image feature. Each preset convolution kernel has a corresponding image feature, and the preset convolution kernel has the same size as its corresponding image feature, specifically referring to the same size of data, for example, the image feature is m × n, that is, the image feature includes m rows And n columns of feature values, the size of the corresponding preset convolution kernel is also m × n, which includes m rows and n columns of weight values.
在使用预设卷积核对图像特征进行处理时,由于二者大小相同,因此,只需要通过预设卷积核对图像特征进行一次卷积计算,就能够实现对图像特征的处理,无需设置预设卷积核的偏置参数。偏置参数是使用卷积核对图像特征进行卷积计算时移动的步长,例如,偏置参数是1,则使用卷积核对图像特征进行处理时,第二次卷积计算时卷积核的感受野,在第一次计算时的基础上向右或向下移动1个像素。When using preset convolution kernels to process image features, because the two are the same size, you only need to perform a convolution calculation on the image features through the preset convolution kernels to process the image features without setting a preset The offset parameter of the convolution kernel. The offset parameter is the step size that is used when the convolution kernel is used to perform the convolution calculation on the image features. For example, if the offset parameter is 1, when the convolution kernel is used to process the image features, the second convolution calculation will Receptive field, move 1 pixel to the right or down based on the first calculation.
步骤2021,根据每个预设卷积核与对应的图像特征进行卷积计算,得到第一数量的处理结果。Step 2021: Perform convolution calculation according to each preset convolution kernel and corresponding image features to obtain a first number of processing results.
其中,若预设卷积核的设置数量与图像特征的数量相同,则一个预设卷积核与一个图像特征对应。每个图像特征都是通过卷积层对整个待处理图像进行计算得到的,因此,每个图像特征都代表了整张图像的一类特点。可以设置不同类别的预设卷积核,并使用相应的预设卷积核与图像特征进行卷积 计算。Wherein, if the number of preset convolution kernels is the same as the number of image features, then one preset convolution kernel corresponds to one image feature. Each image feature is calculated by the convolutional layer on the entire image to be processed. Therefore, each image feature represents a type of feature of the entire image. You can set different types of preset convolution kernels, and use the corresponding preset convolution kernels and image features to perform convolution calculations.
假设预设卷积核的数量是N,图像特征数量也是N,那么每个预设卷积核与其相应的图像特征进行卷积计算,能够得到N个处理结果。Assuming that the number of preset convolution kernels is N and the number of image features is also N, then each preset convolution kernel and its corresponding image feature are convoluted to obtain N processing results.
步骤2022,将每组卷积核中的每个预设卷积核与对应的图像特征进行卷积计算,得到每组卷积核对应的N个降维特征值。Step 2022: Perform convolution calculation on each preset convolution kernel in each group of convolution kernels and corresponding image features to obtain N dimensionality reduction feature values corresponding to each group of convolution kernels.
本实施例提供的另一种实施方式中,若第一数量为N,可以设置N组预设卷积核,且每组包括N个预设卷积核。每组预设卷积核中,每个预设卷积核与一个图像特征对应,即一组中包括的N个预设卷积核,与N个图像特征一一对应。In another implementation provided by this embodiment, if the first number is N, N groups of preset convolution kernels may be set, and each group includes N preset convolution kernels. In each group of preset convolution kernels, each preset convolution kernel corresponds to one image feature, that is, N preset convolution kernels included in a group correspond to N image features one-to-one.
此时,可以将一组中的N个预设卷积核与对应的N个图像特征进行卷积计算,得到N个降维特征值。At this time, N preset convolution kernels in a group and corresponding N image features can be convoluted to obtain N dimensionality reduction feature values.
步骤2023,将每组卷积核对应的N个降维特征值进行叠加,得到N个处理结果。In step 2023, N dimension reduction feature values corresponding to each group of convolution kernels are superimposed to obtain N processing results.
一般情况下,卷积层输出的特征数量与全连接输入的数据数量相同,若直接将N组的处理结果都输入全连接层,会导致全连接层计算量过大。因此,可以将一组卷积核对应的N个降维特征值进行叠加,得到一个处理结果,那么N组预设卷积核则可以得到N个处理结果。In general, the number of features output by the convolutional layer is the same as the number of data of the fully connected input. If all the processing results of the N groups are directly input to the fully connected layer, the calculation amount of the fully connected layer will be excessive. Therefore, N dimensionality reduction feature values corresponding to a group of convolution kernels can be superimposed to obtain one processing result, then N groups of preset convolution kernels can obtain N processing results.
实际应用时,将图像特征与预设卷积核进行卷积计算,能够对图像特征进行降维,并使降维特征值不损失原有的图像特征含义,即每个降维特征值仍然能够表示待处理图片的一类特征。将一组卷积核中的N个降维特征值相加得到处理结果,用于表示待处理图像的整体特征,而设置N组预设卷积核,则能够基于N组预设卷积核,确定出待处理图像的N个整体特征。In actual application, the image features are convolved with the preset convolution kernel to reduce the dimensionality of the image features, so that the dimensionality reduction feature values do not lose the original image feature meaning, that is, each dimensionality reduction feature value can still Represents a type of feature of a picture to be processed. The N dimensionality reduction feature values in a set of convolution kernels are added to obtain a processing result, which is used to represent the overall characteristics of the image to be processed, and setting N sets of preset convolution kernels can be based on N sets of preset convolution kernels To determine the N overall features of the image to be processed.
步骤2024,将每组卷积核中的每个预设卷积核与对应的图像特征进行卷积计算,得到每组卷积核对应的N个降维特征值。Step 2024: Perform convolution calculation on each preset convolution kernel in each group of convolution kernels and corresponding image features to obtain N dimensionality reduction feature values corresponding to each group of convolution kernels.
在另一种实施方式中,若图像特征的数量是N,还可以设置M组预设卷积核,每组包括N个预设卷积核。In another embodiment, if the number of image features is N, M groups of preset convolution kernels may also be set, and each group includes N preset convolution kernels.
与上述实施方式相似,每组包括的N个预设卷积核与图像特征一一对应,可以使用每组包括的N个卷积核与图像特征进行卷积计算,得到N个降维特 征值。若设置有M组预设卷积核,则共可以得到M×N个降维特征值。Similar to the above embodiment, the N preset convolution kernels included in each group have a one-to-one correspondence with the image features, and the N convolution kernels included in each group can be used to perform convolution calculation with the image features to obtain N dimensionality reduction feature values . If M sets of preset convolution kernels are set, a total of M × N dimension reduction feature values can be obtained.
步骤2025,将每组卷积核对应的N个降维特征值平均分为T组,将每组降维特征值进行叠加得到每组卷积核对应的T个处理结果。Step 2025, the N dimension reduction feature values corresponding to each group of convolution kernels are equally divided into T groups, and each group of dimension reduction feature values are superimposed to obtain T processing results corresponding to each group of convolution kernels.
其中,T与M的乘积为N。Among them, the product of T and M is N.
可以将每组预设卷积核对应的N个降维特征值评分为T组。每个降维特征值可以表示待处理图像的一类特征,将N个降维特征值分为T组,则可以将其中的T类特征分为一组,再将每组中的降维特征值进行叠加,即得到T个处理结果。对于一组预设卷积核能够得到T个处理结果,若设置有M组预设卷积核,则可以得到T×M个处理结果,也就是N个处理结果。The N dimensionality reduction feature values corresponding to each group of preset convolution kernels can be scored as T groups. Each dimension-reduction feature value can represent a type of feature of the image to be processed, and the N dimension-reduction feature values are divided into T groups, then the T-type features can be divided into one group, and then the dimension-reduction features in each group Values are superimposed to obtain T processing results. For a set of preset convolution kernels, T processing results can be obtained, and if M sets of preset convolution kernels are set, T × M processing results, that is, N processing results can be obtained.
每个降维特征值能够代表待处理图像一类特征信息,将降维特征值分组后,能够得到T组降维特征值,再将每组的特征进行叠加,能够得到T个待处理图像的综合特征信息。设置M组预设卷积核,则能够得到T×M个综合特征信息。Each dimensionality reduction feature value can represent a type of feature information of the image to be processed. After grouping the dimensionality reduction feature values, T groups of dimensionality reduction feature values can be obtained, and then the features of each group can be superimposed to obtain T images to be processed Comprehensive feature information. By setting M sets of preset convolution kernels, T × M comprehensive feature information can be obtained.
步骤2021或2023或2025执行完毕后,可以执行步骤203。After step 2021 or 2023 or 2025 is executed, step 203 may be executed.
步骤203,将处理结果输入神经网络的全连接层,以使全连接层根据处理结果确定输出结果。Step 203: Input the processing result into the fully connected layer of the neural network, so that the fully connected layer determines the output result according to the processing result.
步骤203与步骤103的具体原理和实现方式类似,此处不再赘述。The specific principles and implementations of step 203 and step 103 are similar, and will not be repeated here.
可选的,本实施例提供的方法中,在使用预设卷积核对待处理图像进行处理前,需要对预设卷积核进行训练,从而确定各个预设卷积核内部的权重值。Optionally, in the method provided in this embodiment, before using the preset convolution kernel to process the image to be processed, the preset convolution kernel needs to be trained to determine the weight value inside each preset convolution kernel.
在训练预设卷积核时,需要先随机生成预设卷积核内的权重值,在训练过程中,还会对预设卷积核的权重值进行调整,在随机生成权重值后,或每次/若干次调整权重值后,本实施例提供的方法还可以包括:When training the preset convolution kernel, the weight value in the preset convolution kernel needs to be randomly generated. During the training process, the weight value of the preset convolution kernel will also be adjusted. After the weight value is randomly generated, or After adjusting the weight value every / several times, the method provided in this embodiment may further include:
根据预设卷积核确定对应的向量;Determine the corresponding vector according to the preset convolution kernel;
根据向量对卷积核进行正规化处理,以使卷积核之间的差异性增强。The convolution kernel is normalized according to the vector to enhance the difference between the convolution kernels.
在权重值发生变化后,可以根据权重值确定预设卷积核对应的向量。再根据向量确定各个权重值是否相似,若相似,则调整对应的预设卷积核内的 权重值,使得两个向量间具有一定的差异性。After the weight value changes, the vector corresponding to the preset convolution kernel can be determined according to the weight value. Then determine whether each weight value is similar according to the vector. If it is similar, adjust the weight value in the corresponding preset convolution kernel to make the two vectors have a certain difference.
其中,可以将预设卷积核内的权重值确定为一维向量,例如,若卷积核是3×3的,则共包括9个权重值,可以将这些权重值按照行的顺序拼接,即第一行、第二行、第三行的顺序拼接,就能够得到一个一维向量。可以根据每个预设卷积核的权重值生成其相应的一维向量。Among them, the weight value in the preset convolution kernel can be determined as a one-dimensional vector. For example, if the convolution kernel is 3 × 3, a total of 9 weight values can be included. These weight values can be spliced in the order of rows. That is, the first row, the second row, and the third row are spliced in sequence, and a one-dimensional vector can be obtained. The corresponding one-dimensional vector can be generated according to the weight value of each preset convolution kernel.
具体的,可以计算每两个向量的内积,根据内积结果确定这两个向量是否相近。内积等于两个向量大小与夹角余弦值的乘积,可以认为内积越小,两个向量间的夹角越大,那么二者间差异越大。因此,可以基于预设卷积核对应的向量内积,确定二者是否相似。Specifically, the inner product of every two vectors can be calculated, and whether the two vectors are similar according to the inner product result can be determined. The inner product is equal to the product of the magnitude of the two vectors and the cosine of the angle. It can be considered that the smaller the inner product, the greater the angle between the two vectors, and the greater the difference between the two. Therefore, it can be determined whether the two are similar based on the inner product of the vector corresponding to the preset convolution kernel.
预设卷积核之间具有明显差异性,能够从不同角度对图像特征进行压缩,从而提高图像特征的表现力,提高图片处理精度。There are obvious differences between preset convolution kernels, which can compress image features from different angles, thereby improving the expressiveness of image features and improving the accuracy of picture processing.
图4为本发明一示例性实施例示出的图像处理装置的结构图。Fig. 4 is a structural diagram of an image processing apparatus according to an exemplary embodiment of the present invention.
如图4所示,本实施例提供的图像处理装置,包括:As shown in FIG. 4, the image processing apparatus provided in this embodiment includes:
获取模块41,用于获取待处理图像,根据神经网络中的卷积层对所述待处理图像进行卷积处理,得到第一数量的图像特征;The obtaining module 41 is configured to obtain an image to be processed, and perform convolution processing on the image to be processed according to a convolutional layer in a neural network to obtain a first number of image features;
处理模块42,用于根据第二数量的预设卷积核对所述图像特征进行卷积处理,得到第一数量的处理结果;The processing module 42 is configured to perform convolution processing on the image features according to a second number of preset convolution kernels to obtain a first number of processing results;
确定模块43,用于将所述处理结果输入所述神经网络的全连接层,以使所述全连接层根据所述处理结果确定输出结果;The determining module 43 is configured to input the processing result into the fully connected layer of the neural network, so that the fully connected layer determines the output result according to the processing result;
其中,所述第二数量大于等于所述第一数量。Wherein, the second quantity is greater than or equal to the first quantity.
本实施例提供的图像处理装置,包括获取模块,用于获取待处理图像,根据神经网络中的卷积层对待处理图像进行卷积处理,得到第一数量的图像特征;处理模块,用于根据第二数量的预设卷积核对图像特征进行卷积处理,得到第一数量的处理结果;确定模块,用于将处理结果输入神经网络的全连接层,以使全连接层根据处理结果确定输出结果;其中,第二数量大于等于第一数量。本实施例提供的装置中,在通过卷积层提取图像特征后,不采用现有技术中的池化层对图像特征进行处理,而是设置预设卷积核,由预设卷积核对图像特征进行卷积处理,得到包括更多特征信息的处理结果,再由全 连接层对处理结果进行融合,从而确定出精度更高的输出结果。The image processing device provided in this embodiment includes an acquisition module for acquiring an image to be processed, and performing a convolution process on the image to be processed according to a convolutional layer in a neural network to obtain a first number of image features; a processing module is used for The second number of preset convolution kernels perform convolution processing on the image features to obtain the first number of processing results; the determination module is used to input the processing results into the fully connected layer of the neural network, so that the fully connected layer determines the output according to the processing results Results; where the second quantity is greater than or equal to the first quantity. In the device provided in this embodiment, after the image features are extracted through the convolution layer, the image features are not processed using the pooling layer in the prior art, but a preset convolution kernel is set, and the image is checked by the preset convolution kernel The features are subjected to convolution processing to obtain a processing result that includes more feature information, and the processing result is fused by the fully connected layer to determine an output result with higher accuracy.
本实施例提供的图像处理装置的具体原理和实现方式均与图2所示的实施例类似,此处不再赘述。The specific principles and implementations of the image processing apparatus provided in this embodiment are similar to the embodiment shown in FIG. 2 and will not be repeated here.
图5为本发明另一示例性实施例示出的图像处理装置的结构图。Fig. 5 is a structural diagram of an image processing apparatus according to another exemplary embodiment of the present invention.
如图5所示,在上述实施例的基础上,本实施例提供的图像处理装置,所述预设卷积核的大小与所述图像特征的大小相同。As shown in FIG. 5, on the basis of the foregoing embodiment, in the image processing apparatus provided in this embodiment, the size of the preset convolution kernel is the same as the size of the image feature.
可选的,所述第二数量与所述第一数量相同;Optionally, the second quantity is the same as the first quantity;
所述处理模块42包括:The processing module 42 includes:
第一处理单元421,用于根据每个所述预设卷积核与对应的所述图像特征进行卷积计算,得到第一数量的所述处理结果。The first processing unit 421 is configured to perform convolution calculation according to each of the preset convolution kernels and the corresponding image features to obtain a first number of the processing results.
可选的,所述第二数量的预设卷积核包括N组预设卷积核,每组包括N个预设卷积核,其中,N等于所述第一数量;Optionally, the second number of preset convolution kernels includes N groups of preset convolution kernels, and each group includes N preset convolution kernels, where N is equal to the first number;
所述处理模块42包括第二处理单元422,用于:The processing module 42 includes a second processing unit 422 for:
将每组卷积核中的每个所述预设卷积核与对应的所述图像特征进行卷积计算,得到每组卷积核对应的N个降维特征值;Performing convolution calculation on each of the preset convolution kernels in each group of convolution kernels and the corresponding image features to obtain N dimensionality reduction feature values corresponding to each group of convolution kernels;
将每组卷积核对应的N个降维特征值进行叠加,得到N个所述处理结果。N dimensionality reduction feature values corresponding to each set of convolution kernels are superimposed to obtain N processing results.
可选的,所述第二数量的预设卷积核包括M组卷积核,每组包括N个卷积核,其中,N等于所述第一数量;Optionally, the second number of preset convolution kernels includes M groups of convolution kernels, and each group includes N convolution kernels, where N is equal to the first number;
所述处理模块42包括第三处理单元423,用于:The processing module 42 includes a third processing unit 423 for:
将每组卷积核中的每个所述预设卷积核与对应的所述图像特征进行卷积计算,得到每组卷积核对应的N个降维特征值;Performing convolution calculation on each of the preset convolution kernels in each group of convolution kernels and the corresponding image features to obtain N dimensionality reduction feature values corresponding to each group of convolution kernels;
将每组卷积核对应的N个降维特征值平均分为T组,将每组所述降维特征值进行叠加得到每组卷积核对应的T个所述处理结果;N dimensionality reduction feature values corresponding to each group of convolution kernels are equally divided into T groups, and the dimensionality reduction feature values of each group are superimposed to obtain T processing results corresponding to each group of convolution kernels;
其中,T与M的乘积为N。Among them, the product of T and M is N.
可选的,每个所述预设卷积核不包含偏置参数。Optionally, each of the preset convolution kernels does not include offset parameters.
可选的,本实施例提供的装置,还包括调整模块44,用于:Optionally, the device provided in this embodiment further includes an adjustment module 44 for:
根据所述预设卷积核确定对应的向量;Determine the corresponding vector according to the preset convolution kernel;
根据所述向量对所述卷积核进行正规化处理,以使所述卷积核之间的差异性增强。The convolution kernel is normalized according to the vector to enhance the difference between the convolution kernels.
可选的,所述调整模块44具体用于:Optionally, the adjustment module 44 is specifically used to:
确定两个所述向量的内积,根据所述内积确定所述向量对应的所述预设卷积核是否相似,若是,则调整所述预设卷积核。Determine the inner product of the two vectors, and determine whether the preset convolution kernel corresponding to the vector is similar according to the inner product, and if so, adjust the preset convolution kernel.
本实施例提供的图像处理装置的具体原理和实现方式均与图3所示的实施例类似,此处不再赘述。The specific principle and implementation of the image processing apparatus provided in this embodiment are similar to the embodiment shown in FIG. 3, and details are not described herein again.
本公开实施例还提供了一种计算机,包含上述的图像处理装置。An embodiment of the present disclosure also provides a computer including the above-mentioned image processing device.
本公开实施例还提供了一种计算机可读存储介质,存储有计算机可执行指令,所述计算机可执行指令设置为执行上述图像处理方法。An embodiment of the present disclosure also provides a computer-readable storage medium that stores computer-executable instructions that are configured to perform the above-described image processing method.
本公开实施例还提供了一种计算机程序产品,所述计算机程序产品包括存储在计算机可读存储介质上的计算机程序,所述计算机程序包括程序指令,当所述程序指令被计算机执行时,使所述计算机执行上述图像处理方法。An embodiment of the present disclosure also provides a computer program product. The computer program product includes a computer program stored on a computer-readable storage medium. The computer program includes program instructions. When the program instructions are executed by a computer, the The computer executes the above image processing method.
上述的计算机可读存储介质可以是暂态计算机可读存储介质,也可以是非暂态计算机可读存储介质。The aforementioned computer-readable storage medium may be a transient computer-readable storage medium or a non-transitory computer-readable storage medium.
图6为本发明一示例性实施例示出的电子设备的结构图。Fig. 6 is a structural diagram of an electronic device according to an exemplary embodiment of the present invention.
本公开实施例还提供了一种电子设备,其结构如图6所示,该电子设备包括:An embodiment of the present disclosure also provides an electronic device, whose structure is shown in FIG. 6, the electronic device includes:
至少一个处理器(processor)60,图6中以一个处理器60为例;和存储器(memory)61,还可以包括通信接口(Communication Interface)62和总线63。其中,处理器60、通信接口62、存储器61可以通过总线63完成相互间的通信。通信接口62可以用于信息传输。处理器60可以调用存储器61中的逻辑指令,以执行上述实施例的图像处理方法。At least one processor (processor) 60, one processor 60 is taken as an example in FIG. 6; and the memory (memory) 61 may further include a communication interface (Communication Interface) 62 and a bus 63. The processor 60, the communication interface 62, and the memory 61 can complete communication with each other through the bus 63. The communication interface 62 may be used for information transmission. The processor 60 may call logic instructions in the memory 61 to execute the image processing method of the above-mentioned embodiment.
此外,上述的存储器61中的逻辑指令可以通过软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。In addition, the logic instructions in the aforementioned memory 61 can be implemented in the form of software functional units and sold or used as an independent product, and can be stored in a computer-readable storage medium.
存储器61作为一种计算机可读存储介质,可用于存储软件程序、计算机 可执行程序,如本公开实施例中的方法对应的程序指令/模块。处理器60通过运行存储在存储器61中的软件程序、指令以及模块,从而执行功能应用以及数据处理,即实现上述方法实施例中的图像处理方法。The memory 61 is a computer-readable storage medium that can be used to store software programs and computer-executable programs, such as program instructions / modules corresponding to the methods in the embodiments of the present disclosure. The processor 60 executes functional applications and data processing by running software programs, instructions, and modules stored in the memory 61, that is, implementing the image processing method in the above method embodiments.
存储器61可包括存储程序区和存储数据区,其中,存储程序区可存储操作系统、至少一个功能所需的应用程序;存储数据区可存储根据终端设备的使用所创建的数据等。此外,存储器61可以包括高速随机存取存储器,还可以包括非易失性存储器。The memory 61 may include a storage program area and a storage data area, wherein the storage program area may store an operating system and application programs required for at least one function; the storage data area may store data created according to the use of a terminal device and the like. In addition, the memory 61 may include a high-speed random access memory, and may also include a non-volatile memory.
本公开实施例的技术方案可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括一个或多个指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本公开实施例所述方法的全部或部分步骤。而前述的存储介质可以是非暂态存储介质,包括:U盘、移动硬盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、磁碟或者光盘等多种可以存储程序代码的介质,也可以是暂态存储介质。The technical solutions of the embodiments of the present disclosure may be embodied in the form of software products, which are stored in a storage medium and include one or more instructions to make a computer device (which may be a personal computer, server, or network) Equipment, etc.) to perform all or part of the steps of the method described in the embodiments of the present disclosure. The aforementioned storage medium may be a non-transitory storage medium, including: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disk or optical disk, etc. A medium that can store program codes may also be a transient storage medium.
当用于本申请中时,虽然术语“第一”、“第二”等可能会在本申请中使用以描述各元件,但这些元件不应受到这些术语的限制。这些术语仅用于将一个元件与另一个元件区别开。比如,在不改变描述的含义的情况下,第一元件可以叫做第二元件,并且同样第,第二元件可以叫做第一元件,只要所有出现的“第一元件”一致重命名并且所有出现的“第二元件”一致重命名即可。第一元件和第二元件都是元件,但可以不是相同的元件。When used in this application, although the terms "first", "second", etc. may be used in this application to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another. For example, without changing the meaning of the description, the first element can be called the second element, and likewise, the second element can be called the first element, as long as all occurrences of the "first element" are consistently renamed and all occurrences of The "second component" can be renamed consistently. The first element and the second element are both elements, but they may not be the same element.
本申请中使用的用词仅用于描述实施例并且不用于限制权利要求。如在实施例以及权利要求的描述中使用的,除非上下文清楚地表明,否则单数形式的“一个”(a)、“一个”(an)和“所述”(the)旨在同样包括复数形式。类似地,如在本申请中所使用的术语“和/或”是指包含一个或一个以上相关联的列出的任何以及所有可能的组合。另外,当用于本申请中时,术语“包括”(comprise)及其变型“包括”(comprises)和/或包括(comprising)等指陈述的特征、整体、步骤、操作、元素,和/或组件的存在,但不排除一个或一个以上其它特征、整体、步骤、操作、元素、组件和/或这些的分组的存在或添加。The terms used in this application are only used to describe the embodiments and are not used to limit the claims. As used in the description of the embodiments and claims, unless the context clearly indicates otherwise, the singular forms "a", "an" and "said" are intended to include plural forms as well . Similarly, the term "and / or" as used in this application is meant to include any and all possible combinations of one or more associated lists. In addition, when used in this application, the term "comprise" and its variations "comprises" and / or includes etc. refer to the stated features, wholes, steps, operations, elements, and / or The presence of components does not exclude the presence or addition of one or more other features, wholes, steps, operations, elements, components, and / or groups of these.
所描述的实施例中的各方面、实施方式、实现或特征能够单独使用或以任意组合的方式使用。所描述的实施例中的各方面可由软件、硬件或软硬件的结合实现。所描述的实施例也可以由存储有计算机可读代码的计算机可读介质体现,该计算机可读代码包括可由至少一个计算装置执行的指令。所述计算机可读介质可与任何能够存储数据的数据存储装置相关联,该数据可由计算机系统读取。用于举例的计算机可读介质可以包括只读存储器、随机存取存储器、CD-ROM、HDD、DVD、磁带以及光数据存储装置等。所述计算机可读介质还可以分布于通过网络联接的计算机系统中,这样计算机可读代码就可以分布式存储并执行。The various aspects, implementations, implementations, or features in the described embodiments can be used alone or in any combination. Various aspects in the described embodiments may be implemented by software, hardware, or a combination of software and hardware. The described embodiments may also be embodied by a computer-readable medium that stores computer-readable code including instructions executable by at least one computing device. The computer-readable medium can be associated with any data storage device capable of storing data, which can be read by a computer system. Computer-readable media used for examples may include read-only memory, random access memory, CD-ROM, HDD, DVD, magnetic tape, optical data storage devices, and the like. The computer-readable medium may also be distributed in computer systems connected through a network, so that computer-readable codes can be stored and executed in a distributed manner.
上述技术描述可参照附图,这些附图形成了本申请的一部分,并且通过描述在附图中示出了依照所描述的实施例的实施方式。虽然这些实施例描述的足够详细以使本领域技术人员能够实现这些实施例,但这些实施例是非限制性的;这样就可以使用其它的实施例,并且在不脱离所描述的实施例的范围的情况下还可以做出变化。比如,流程图中所描述的操作顺序是非限制性的,因此在流程图中阐释并且根据流程图描述的两个或两个以上操作的顺序可以根据若干实施例进行改变。作为另一个例子,在若干实施例中,在流程图中阐释并且根据流程图描述的一个或一个以上操作是可选的,或是可删除的。另外,某些步骤或功能可以添加到所公开的实施例中,或两个以上的步骤顺序被置换。所有这些变化被认为包含在所公开的实施例以及权利要求中。The above technical description may refer to the accompanying drawings, which form a part of the present application, and the description shows an implementation according to the described embodiments in the drawings. Although these embodiments are described in sufficient detail to enable those skilled in the art to implement these embodiments, these embodiments are non-limiting; so that other embodiments can be used without departing from the scope of the described embodiments Changes can also be made under circumstances. For example, the sequence of operations described in the flowchart is non-limiting, so the sequence of two or more operations explained in the flowchart and described according to the flowchart may be changed according to several embodiments. As another example, in several embodiments, one or more operations illustrated in the flowchart and described in accordance with the flowchart are optional or may be deleted. In addition, certain steps or functions may be added to the disclosed embodiments, or two or more steps may be replaced in sequence. All these changes are considered to be included in the disclosed embodiments and claims.
另外,上述技术描述中使用术语以提供所描述的实施例的透彻理解。然而,并不需要过于详细的细节以实现所描述的实施例。因此,实施例的上述描述是为了阐释和描述而呈现的。上述描述中所呈现的实施例以及根据这些实施例所公开的例子是单独提供的,以添加上下文并有助于理解所描述的实施例。上述说明书不用于做到无遗漏或将所描述的实施例限制到本公开的精确形式。根据上述教导,若干修改、选择适用以及变化是可行的。在某些情况下,没有详细描述为人所熟知的处理步骤以避免不必要地影响所描述的实施例。In addition, terminology is used in the above technical description to provide a thorough understanding of the described embodiments. However, no excessively detailed details are required to implement the described embodiments. Therefore, the above description of the embodiments is presented for explanation and description. The embodiments presented in the above description and the examples disclosed according to these embodiments are provided separately to add context and help to understand the described embodiments. The above description is not intended to be without omission or to limit the described embodiments to the precise form of this disclosure. Based on the above teachings, several modifications, choices and changes are possible. In some cases, well-known processing steps are not described in detail to avoid unnecessarily affecting the described embodiments.

Claims (20)

  1. 一种图像处理方法,其特征在于,包括:An image processing method, characterized in that it includes:
    获取待处理图像,根据神经网络中的卷积层对所述待处理图像进行卷积处理,得到第一数量的图像特征;Obtain the image to be processed, and perform convolution on the image to be processed according to the convolution layer in the neural network to obtain a first number of image features;
    根据第二数量的预设卷积核对所述图像特征进行卷积处理,得到第一数量的处理结果;Performing convolution processing on the image features according to a second number of preset convolution kernels to obtain a first number of processing results;
    将所述处理结果输入所述神经网络的全连接层,以使所述全连接层根据所述处理结果确定输出结果;Input the processing result into the fully connected layer of the neural network, so that the fully connected layer determines the output result according to the processing result;
    其中,所述第二数量大于等于所述第一数量。Wherein, the second quantity is greater than or equal to the first quantity.
  2. 根据权利要求1所述的方法,其特征在于,所述预设卷积核的大小与所述图像特征的大小相同。The method according to claim 1, wherein the size of the preset convolution kernel is the same as the size of the image feature.
  3. 根据权利要求1或2所述的方法,其特征在于,所述第二数量与所述第一数量相同;The method according to claim 1 or 2, wherein the second quantity is the same as the first quantity;
    所述根据预设卷积核对所述图像特征进行卷积处理,得到第一数量的处理结果,包括:The convolution processing of the image features according to a preset convolution kernel to obtain a first number of processing results includes:
    根据每个所述预设卷积核与对应的所述图像特征进行卷积计算,得到第一数量的所述处理结果。Performing convolution calculation according to each of the preset convolution kernels and the corresponding image features to obtain a first number of the processing results.
  4. 根据权利要求1或2所述的方法,其特征在于,所述第二数量的预设卷积核包括N组预设卷积核,每组包括N个预设卷积核,其中,N等于所述第一数量;The method according to claim 1 or 2, wherein the second number of preset convolution kernels includes N groups of preset convolution kernels, and each group includes N preset convolution kernels, where N is equal to The first quantity;
    所述根据预设卷积核对所述图像特征进行卷积处理,得到第一数量的处理结果,包括:The convolution processing of the image features according to a preset convolution kernel to obtain a first number of processing results includes:
    将每组卷积核中的每个所述预设卷积核与对应的所述图像特征进行卷积计算,得到每组卷积核对应的N个降维特征值;Performing convolution calculation on each of the preset convolution kernels in each group of convolution kernels and the corresponding image features to obtain N dimensionality reduction feature values corresponding to each group of convolution kernels;
    将每组卷积核对应的N个降维特征值进行叠加,得到N个所述处理结果。N dimensionality reduction feature values corresponding to each set of convolution kernels are superimposed to obtain N processing results.
  5. 根据权利要求1或2所述的方法,其特征在于,所述第二数量的预设卷积核包括M组卷积核,每组包括N个卷积核,其中,N等于所述第一数量;The method according to claim 1 or 2, wherein the second number of preset convolution kernels includes M groups of convolution kernels, and each group includes N convolution kernels, where N is equal to the first Quantity
    所述根据预设卷积核对所述图像特征进行卷积处理,得到第一数量的处 理结果,包括:The convolution processing of the image features according to a preset convolution kernel to obtain a first number of processing results includes:
    将每组卷积核中的每个所述预设卷积核与对应的所述图像特征进行卷积计算,得到每组卷积核对应的N个降维特征值;Performing convolution calculation on each of the preset convolution kernels in each group of convolution kernels and the corresponding image features to obtain N dimensionality reduction feature values corresponding to each group of convolution kernels;
    将每组卷积核对应的N个降维特征值平均分为T组,将每组所述降维特征值进行叠加得到每组卷积核对应的T个所述处理结果;N dimensionality reduction feature values corresponding to each group of convolution kernels are equally divided into T groups, and the dimensionality reduction feature values of each group are superimposed to obtain T processing results corresponding to each group of convolution kernels;
    其中,T与M的乘积为N。Among them, the product of T and M is N.
  6. 根据权利要求1或2所述的方法,其特征在于,每个所述预设卷积核不包含偏置参数。The method according to claim 1 or 2, wherein each of the preset convolution kernels does not include an offset parameter.
  7. 根据权利要求1所述的方法,其特征在于,还包括:The method according to claim 1, further comprising:
    根据所述预设卷积核确定对应的向量;Determine the corresponding vector according to the preset convolution kernel;
    根据所述向量对所述卷积核进行正规化处理,以使所述卷积核之间的差异性增强。The convolution kernel is normalized according to the vector to enhance the difference between the convolution kernels.
  8. 根据权利要求7所述的方法,其特征在于,所述根据所述向量对所述卷积核进行正规化处理包括:The method according to claim 7, wherein the normalizing the convolution kernel according to the vector comprises:
    确定两个所述向量的内积,根据所述内积确定所述向量对应的所述预设卷积核是否相似,若是,则调整所述预设卷积核。Determine the inner product of the two vectors, and determine whether the preset convolution kernel corresponding to the vector is similar according to the inner product, and if so, adjust the preset convolution kernel.
  9. 一种图像处理装置,其特征在于,包括:An image processing device, characterized in that it includes:
    获取模块,用于获取待处理图像,根据神经网络中的卷积层对所述待处理图像进行卷积处理,得到第一数量的图像特征;An obtaining module, configured to obtain an image to be processed, and performing convolution processing on the image to be processed according to a convolution layer in a neural network to obtain a first number of image features;
    处理模块,用于根据第二数量的预设卷积核对所述图像特征进行卷积处理,得到第一数量的处理结果;A processing module, configured to perform convolution processing on the image features according to a second number of preset convolution kernels to obtain a first number of processing results;
    确定模块,用于将所述处理结果输入所述神经网络的全连接层,以使所述全连接层根据所述处理结果确定输出结果;A determining module, configured to input the processing result into a fully connected layer of the neural network, so that the fully connected layer determines an output result according to the processing result;
    其中,所述第二数量大于等于所述第一数量。Wherein, the second quantity is greater than or equal to the first quantity.
  10. 根据权利要求9所述的装置,其特征在于,所述预设卷积核的大小与所述图像特征的大小相同。The apparatus according to claim 9, wherein the size of the preset convolution kernel is the same as the size of the image feature.
  11. 根据权利要求9或10所述的装置,其特征在于,所述第二数量与所述第一数量相同;The device according to claim 9 or 10, wherein the second quantity is the same as the first quantity;
    所述处理模块包括:The processing module includes:
    第一处理单元,用于根据每个所述预设卷积核与对应的所述图像特征进行卷积计算,得到第一数量的所述处理结果。The first processing unit is configured to perform convolution calculation according to each of the preset convolution kernels and the corresponding image features to obtain a first number of the processing results.
  12. 根据权利要9或10所述的装置,其特征在于,所述第二数量的预设卷积核包括N组预设卷积核,每组包括N个预设卷积核,其中,N等于所述第一数量;The apparatus according to claim 9 or 10, wherein the second number of preset convolution kernels includes N groups of preset convolution kernels, and each group includes N preset convolution kernels, where N is equal to The first quantity;
    所述处理模块包括第二处理单元,用于:The processing module includes a second processing unit for:
    将每组卷积核中的每个所述预设卷积核与对应的所述图像特征进行卷积计算,得到每组卷积核对应的N个降维特征值;Performing convolution calculation on each of the preset convolution kernels in each group of convolution kernels and the corresponding image features to obtain N dimensionality reduction feature values corresponding to each group of convolution kernels;
    将每组卷积核对应的N个降维特征值进行叠加,得到N个所述处理结果。N dimensionality reduction feature values corresponding to each set of convolution kernels are superimposed to obtain N processing results.
  13. 根据权利要求9或10所述的装置,其特征在于,所述第二数量的预设卷积核包括M组卷积核,每组包括N个卷积核,其中,N等于所述第一数量;The apparatus according to claim 9 or 10, wherein the second number of preset convolution kernels includes M groups of convolution kernels, and each group includes N convolution kernels, where N is equal to the first Quantity
    所述处理模块包括第三处理单元,用于:The processing module includes a third processing unit for:
    将每组卷积核中的每个所述预设卷积核与对应的所述图像特征进行卷积计算,得到每组卷积核对应的N个降维特征值;Performing convolution calculation on each of the preset convolution kernels in each group of convolution kernels and the corresponding image features to obtain N dimensionality reduction feature values corresponding to each group of convolution kernels;
    将每组卷积核对应的N个降维特征值平均分为T组,将每组所述降维特征值进行叠加得到每组卷积核对应的T个所述处理结果;N dimensionality reduction feature values corresponding to each group of convolution kernels are equally divided into T groups, and the dimensionality reduction feature values of each group are superimposed to obtain T processing results corresponding to each group of convolution kernels;
    其中,T与M的乘积为N。Among them, the product of T and M is N.
  14. 根据权利要求9或10所述的装置,其特征在于,每个所述预设卷积核不包含偏置参数。The apparatus according to claim 9 or 10, wherein each of the preset convolution kernels does not include an offset parameter.
  15. 根据权利要求9所述的装置,其特征在于,还包括调整模块,用于:The device according to claim 9, further comprising an adjustment module for:
    根据所述预设卷积核确定对应的向量;Determine the corresponding vector according to the preset convolution kernel;
    根据所述向量对所述卷积核进行正规化处理,以使所述卷积核之间的差异性增强。The convolution kernel is normalized according to the vector to enhance the difference between the convolution kernels.
  16. 根据权利要求15所述的装置,其特征在于,所述调整模块具体用于:The device according to claim 15, wherein the adjustment module is specifically configured to:
    确定两个所述向量的内积,根据所述内积确定所述向量对应的所述预设卷积核是否相似,若是,则调整所述预设卷积核。Determine the inner product of the two vectors, and determine whether the preset convolution kernel corresponding to the vector is similar according to the inner product, and if so, adjust the preset convolution kernel.
  17. 一种计算机,其特征在于,包含权利要求9-16任一项所述的装置。A computer, characterized by comprising the device according to any one of claims 9-16.
  18. 一种电子设备,其特征在于,包括:An electronic device, characterized in that it includes:
    至少一个处理器;以及At least one processor; and
    与所述至少一个处理器通信连接的存储器;其中,A memory communicatively connected to the at least one processor; wherein,
    所述存储器存储有可被所述至少一个处理器执行的指令,所述指令被所述至少一个处理器执行时,使所述至少一个处理器执行权利要求1-8任一项所述的方法。The memory stores instructions executable by the at least one processor, and when the instructions are executed by the at least one processor, causes the at least one processor to perform the method of any one of claims 1-8 .
  19. 一种计算机可读存储介质,其特征在于,存储有计算机可执行指令,所述计算机可执行指令设置为执行权利要求1-8任一项所述的方法。A computer-readable storage medium, characterized in that computer-executable instructions are stored, and the computer-executable instructions are configured to perform the method of any one of claims 1-8.
  20. 一种计算机程序产品,其特征在于,所述计算机程序产品包括存储在计算机可读存储介质上的计算机程序,所述计算机程序包括程序指令,当所述程序指令被计算机执行时,使所述计算机执行权利要求1-8任一项所述的方法。A computer program product, characterized in that the computer program product includes a computer program stored on a computer-readable storage medium, the computer program includes program instructions, and when the program instructions are executed by a computer, the computer Performing the method of any one of claims 1-8.
PCT/CN2018/115252 2018-11-13 2018-11-13 Image processing method, apparatus and device, and storage medium and program product WO2020097795A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201880098341.3A CN112913253A (en) 2018-11-13 2018-11-13 Image processing method, apparatus, device, storage medium, and program product
PCT/CN2018/115252 WO2020097795A1 (en) 2018-11-13 2018-11-13 Image processing method, apparatus and device, and storage medium and program product

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2018/115252 WO2020097795A1 (en) 2018-11-13 2018-11-13 Image processing method, apparatus and device, and storage medium and program product

Publications (1)

Publication Number Publication Date
WO2020097795A1 true WO2020097795A1 (en) 2020-05-22

Family

ID=70731002

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/115252 WO2020097795A1 (en) 2018-11-13 2018-11-13 Image processing method, apparatus and device, and storage medium and program product

Country Status (2)

Country Link
CN (1) CN112913253A (en)
WO (1) WO2020097795A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114399828B (en) * 2022-03-25 2022-07-08 深圳比特微电子科技有限公司 Training method of convolution neural network model for image processing
CN115987511B (en) * 2023-03-07 2023-05-23 北京数牍科技有限公司 Image reasoning method, device, electronic equipment and computer readable storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6782135B1 (en) * 2000-02-18 2004-08-24 Conexant Systems, Inc. Apparatus and methods for adaptive digital video quantization
US6917703B1 (en) * 2001-02-28 2005-07-12 Nevengineering, Inc. Method and apparatus for image analysis of a gabor-wavelet transformed image using a neural network
CN107563290A (en) * 2017-08-01 2018-01-09 中国农业大学 A kind of pedestrian detection method and device based on image
CN107871306A (en) * 2016-09-26 2018-04-03 北京眼神科技有限公司 Method and device for denoising picture
CN107968962A (en) * 2017-12-12 2018-04-27 华中科技大学 A kind of video generation method of the non-conterminous image of two frames based on deep learning
CN108376386A (en) * 2018-03-23 2018-08-07 深圳天琴医疗科技有限公司 A kind of construction method and device of the super-resolution model of image

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9536293B2 (en) * 2014-07-30 2017-01-03 Adobe Systems Incorporated Image assessment using deep convolutional neural networks
CN106295566B (en) * 2016-08-10 2019-07-09 北京小米移动软件有限公司 Facial expression recognizing method and device
CN107316013B (en) * 2017-06-14 2020-04-07 西安电子科技大学 Hyperspectral image classification method based on NSCT (non-subsampled Contourlet transform) and DCNN (data-to-neural network)
CN108259997B (en) * 2018-04-02 2019-08-23 腾讯科技(深圳)有限公司 Image correlation process method and device, intelligent terminal, server, storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6782135B1 (en) * 2000-02-18 2004-08-24 Conexant Systems, Inc. Apparatus and methods for adaptive digital video quantization
US6917703B1 (en) * 2001-02-28 2005-07-12 Nevengineering, Inc. Method and apparatus for image analysis of a gabor-wavelet transformed image using a neural network
CN107871306A (en) * 2016-09-26 2018-04-03 北京眼神科技有限公司 Method and device for denoising picture
CN107563290A (en) * 2017-08-01 2018-01-09 中国农业大学 A kind of pedestrian detection method and device based on image
CN107968962A (en) * 2017-12-12 2018-04-27 华中科技大学 A kind of video generation method of the non-conterminous image of two frames based on deep learning
CN108376386A (en) * 2018-03-23 2018-08-07 深圳天琴医疗科技有限公司 A kind of construction method and device of the super-resolution model of image

Also Published As

Publication number Publication date
CN112913253A (en) 2021-06-04

Similar Documents

Publication Publication Date Title
TWI721510B (en) Method, apparatus and storage medium for binocular image depth estimation
US9697416B2 (en) Object detection using cascaded convolutional neural networks
CN110473137B (en) Image processing method and device
CN109284733B (en) Shopping guide negative behavior monitoring method based on yolo and multitask convolutional neural network
US11222211B2 (en) Method and apparatus for segmenting video object, electronic device, and storage medium
WO2018036293A1 (en) Image segmentation method, apparatus, and fully convolutional network system
US20180285683A1 (en) Methods and apparatus for image salient object detection
CN109840477B (en) Method and device for recognizing shielded face based on feature transformation
US20230085605A1 (en) Face image processing method, apparatus, device, and storage medium
WO2022121485A1 (en) Image multi-tag classification method and apparatus, computer device, and storage medium
WO2021003936A1 (en) Image segmentation method, electronic device, and computer-readable storage medium
US11615612B2 (en) Systems and methods for image feature extraction
CN103744974B (en) Method and device for selecting local interest points
CN111339884A (en) Image recognition method and related equipment and device
WO2020097795A1 (en) Image processing method, apparatus and device, and storage medium and program product
CN111914908A (en) Image recognition model training method, image recognition method and related equipment
CN111597933A (en) Face recognition method and device
CN112115295A (en) Video image detection method and device and electronic equipment
WO2017202086A1 (en) Image screening method and device
CN110717929A (en) Image target detection method, device and storage medium
CN111598176A (en) Image matching processing method and device
KR102123835B1 (en) System and method for image registration based on adaptive classification
CN111738272A (en) Target feature extraction method and device and electronic equipment
CN114612926A (en) Method and device for counting number of people in scene, computer equipment and storage medium
CN112424787B (en) Method and device for extracting image key points

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18940142

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 09/09/2021)

122 Ep: pct application non-entry in european phase

Ref document number: 18940142

Country of ref document: EP

Kind code of ref document: A1