WO2020062191A1 - Procédé, appareil et dispositif de traitement d'image - Google Patents

Procédé, appareil et dispositif de traitement d'image Download PDF

Info

Publication number
WO2020062191A1
WO2020062191A1 PCT/CN2018/108891 CN2018108891W WO2020062191A1 WO 2020062191 A1 WO2020062191 A1 WO 2020062191A1 CN 2018108891 W CN2018108891 W CN 2018108891W WO 2020062191 A1 WO2020062191 A1 WO 2020062191A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
frame
network model
resolution
sample
Prior art date
Application number
PCT/CN2018/108891
Other languages
English (en)
Chinese (zh)
Inventor
谭文伟
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to PCT/CN2018/108891 priority Critical patent/WO2020062191A1/fr
Priority to CN201880093293.9A priority patent/CN112088393B/zh
Publication of WO2020062191A1 publication Critical patent/WO2020062191A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting

Definitions

  • the present invention relates to image processing technology, and in particular, to an image processing method, device, and device.
  • multimedia information (picture information or video information) has become a mainstream multimedia file.
  • the terminal needs high-speed broadband to transmit high-resolution multimedia information, which will greatly increase the cost of information exchange between the two sides of the interactive terminal. Therefore, users usually convert high-resolution multimedia information to low-resolution multimedia information, and then send the low-resolution multimedia information to other terminals, which reduces the interaction cost.
  • the receiving terminal After receiving the low-resolution multimedia information, the receiving terminal needs to restore the low-resolution multimedia information to high-resolution multimedia information in order to obtain more detailed information. In practice, it has been found that the quality of the high-resolution multimedia information restored is poor.
  • the invention provides an image processing method, device and equipment, which improve the accuracy of converting a low-resolution image into a high-resolution image, so as to improve the quality of the high-resolution image.
  • an embodiment of the present invention provides an image packet processing method.
  • the method includes: acquiring a target image requiring super-resolution processing; and inputting the target image to a super-scoring network model for processing to obtain a high-resolution image.
  • Image wherein the network parameters of the super-segmented network model are obtained by adjusting multi-frame sample images and the semantic feature maps corresponding to the sample images in each frame, and the semantic feature maps are semantically identified through the image semantic network model owned.
  • the super-segment network model is a semantic-enhanced network model, that is, a semantic-enhanced super-segment network model can convert a low-resolution image into a semantic-enhanced high-resolution image, and a semantic-enhanced high-resolution image can be provided. More detailed feature information can provide high-resolution edge structure information, which improves the quality of high-resolution images.
  • an error of the super-scoring network model is determined according to the multi-frame sample image and a semantic feature map corresponding to each of the sample images; when the error is greater than a preset error value, the super-scoring network is determined.
  • the network parameters of the model are adjusted.
  • the network parameters of the ultra-scoring network model may be adjusted according to the multi-frame sample image and the semantic feature map corresponding to each sample image.
  • obtain a high-resolution sub-image and a low-resolution sub-image corresponding to each frame of the multi-frame sample images input each frame of the target sub-image into the image semantic network model, and perform semantic recognition to obtain each frame.
  • the semantic feature image corresponding to the sample image, and the target sub-image is a high-resolution sub-image or a low-resolution sub-image corresponding to any of the sample images in the multi-frame sample image;
  • a high-resolution sub-image of the sample image and a superimposed image of the semantic feature image of the sample image are used as reference images, and a low-resolution sub-image of the sample image is used as a training sample.
  • the super image is calculated based on the reference image and the training sample image.
  • the error of the sub-network model is in order to obtain the super-sub-network model with lower error.
  • the image processing device may set weights for the images output by the image semantic network model, in order to obtain an incapable super-segment network model to meet the different image needs of the user, that is, the larger the weight value, it indicates the semantics in the superimposed image
  • the weight value indicates the semantics in the superimposed image
  • the image semantic network model includes a multi-layer neural network, and the target sub-image is input into the image semantic network model.
  • the multi-layer neural network included in the image semantic network model performs semantic recognition and outputs multiple information.
  • Frame candidate feature images, each layer of the neural network outputs a frame of candidate feature images; performing grayscale processing on the candidate feature images of each frame to obtain a grayscale image; determining parameter values of the grayscale images of each frame, and
  • the gray image with the highest value is used as the semantic feature image of the sample image corresponding to the target sub-image, and the parameter value is determined according to the sharpness of the gray image and / or the amount of information provided by the gray image .
  • a candidate image with a higher definition and / or a larger amount of information is selected as a semantic feature image from the multi-frame candidate feature images to improve the quality of the semantic feature images, and further, to improve the super-scoring network model Performance for processing high-resolution images.
  • a super-scoring network model that matches the type of the target image can be selected to process the target image.
  • an embodiment of the present invention provides an image processing apparatus having a function of realizing the behavior in the implementation manner of the first aspect.
  • This function can be realized by hardware, and can also be implemented by hardware executing corresponding software.
  • the hardware or software includes one or more modules corresponding to the above functions, and the modules may be software and / or hardware.
  • the implementation of the image processing apparatus can be referred to the method for the first aspect. Implementation manners, duplicates are not repeated.
  • an embodiment of the present invention provides an electronic device.
  • the electronic device includes: a memory configured to store one or more programs; and a processor configured to call a program stored in the memory to implement the first aspect described above.
  • a processor configured to call a program stored in the memory to implement the first aspect described above.
  • FIG. 1 is a schematic flowchart of an image processing method according to an embodiment of the present invention
  • FIG. 2 is a schematic flowchart of an image processing method according to an embodiment of the present invention.
  • FIG. 3 is a schematic diagram of another super-segment network model and an image semantic network model according to an embodiment of the present invention.
  • FIG. 4 is a schematic structural diagram of an image processing apparatus according to an embodiment of the present invention.
  • FIG. 5 is a schematic structural diagram of an electronic device according to an embodiment of the present invention.
  • the picture processing device may be set in any electronic device and used for performing a high-resolution picture conversion operation on a picture.
  • the electronic device includes, but is not limited to, smart mobile devices (such as mobile phones, PDAs, media players, etc.), wearable devices, headsets, personal computers, server computers, handheld or laptop devices, and so on.
  • FIG. 1 is a schematic flowchart of an image processing method according to an embodiment of the present invention. The method may be executed by an image processing apparatus. The specific explanation of the image processing apparatus is as described above. As shown in FIG. 1, the image processing method may include the following steps.
  • the image processing apparatus may obtain a target image requiring super-resolution processing from a local database, or download a target image requiring super-resolution processing from a network.
  • the target image refers to an image with a resolution lower than a preset resolution value, and the target image may refer to a captured image or any frame image in a captured video.
  • the network parameters of the super-segmented network model are adjusted according to the multi-frame sample images and the semantic feature maps corresponding to the sample images in each frame, and the semantic feature maps are obtained through the semantic recognition of the image semantic network model.
  • the image processing device can input the target image into a super-scoring network model for processing to obtain a high-resolution image to improve the quality of the high-resolution image.
  • the high-resolution image may refer to an image with a resolution greater than a preset resolution value.
  • the high-resolution image may provide users with more detailed feature information and edge structure information.
  • the super-segmentation network model and the image semantic network model can be constituted by a convolutional neural network.
  • a convolutional neural network there are usually multiple convolutional layers, and each convolutional layer includes multiple convolutional kernels. It is three-dimensional and contains data in three dimensions of C, H, and W. C, H, and W represent the depth, height, and width of the data, respectively.
  • a convolution kernel is essentially a combination of a series of weights. By adjusting the weight of the convolution kernel in the super-segmented network model, the image conversion error of the super-segmented network model can be reduced. error.
  • the network parameter of the super-segment network model refers to the weight of the convolution kernel in the super-segment network model.
  • the image processing device may preprocess the target image, and input the pre-processed target image to a super-scoring network model for processing to obtain a high-resolution image.
  • preprocessing includes cropping the target image to extract areas that are of interest to the target image, such as cropping out the face of a person; or preprocessing includes scaling the target image to obtain a suitable image. The size of the super network model.
  • the image processing apparatus may obtain the type of the target image, determine a super-scoring network model that matches the type of the target image, and input the target image to a super-scoring network model that matches the type of the target image for processing To get high-resolution images.
  • the image processing device can obtain the type of the target image and classify it according to the content included in the target image.
  • the type of the target image includes the person image type and the scene image type. Or animal image types, which are classified according to the state of the target image, and the type of the target image includes a static image type or a dynamic image type.
  • a super-scoring network model matching the type of the target image is determined, and the target image is input to the super-scoring network model matching the type of the target image for processing to obtain a high-resolution image.
  • the target image is a person image type
  • a super network model matching the person image type is obtained, and the target image is input into the matched super network model for processing to obtain a high-resolution image.
  • the network parameters of the matched super-segmentation network model are adjusted through multiple frames including a person sample image and a semantic feature image corresponding to each frame of the person sample image.
  • the image processing device may train different types of super-scoring network models according to different types of sample images and semantic feature images corresponding to the sample images, for example, using multiple frames of sample images including animals and each frame of sample images
  • Corresponding semantic feature images are super-network models suitable for processing images including animals.
  • the network parameters of the super-segmented network model are adjusted based on a large number of sample images and the semantic feature images of each frame of sample images, the semantic feature images contain detailed feature information of the sample images. And edge structure information. Therefore, the high-resolution image obtained through the super-scoring network model can provide more detailed feature information and provide high-resolution edge structure information, which improves the quality of high-resolution images.
  • FIG. 2 is a schematic flowchart of an image processing method according to an embodiment of the present invention.
  • the method may be executed by an image processing apparatus.
  • the specific explanation of the image processing apparatus is as described above.
  • the difference between the embodiment of the present invention and the embodiment described in FIG. 1 is that the embodiment of the present invention calculates the error of the super-scoring network model by using multiple frame sample images and the semantic feature images of each frame sample image.
  • the error is greater than a preset error value
  • An embodiment of the present invention is shown in FIG. 2.
  • the image processing method may include the following steps.
  • the image processing apparatus may determine the error of the super-scoring network model according to the multi-frame sample image and the semantic feature map corresponding to each sample image.
  • step S201 includes steps S11 to S15. .
  • the target sub-image of each frame into the image semantic network model and perform semantic recognition to obtain the semantic feature image corresponding to the sample image of each frame.
  • the target sub-image is a high-resolution corresponding to any sample image in the multi-frame sample image. Rate sub-image or low-resolution sub-image.
  • the low-resolution sub-images of each frame are input into the super-scoring network model and processed to obtain high-resolution feature images of the sample images of each frame.
  • the image processing device may perform sampling processing on each frame of the multi-frame sample image to obtain a low-resolution sub-image corresponding to each frame of the sample image, and perform enhancement processing on each frame of the sample image to obtain each frame.
  • the low-resolution sub-images of each frame are input to the super-scoring network model for processing to obtain high-resolution feature images of the sample images of each frame, and the target sub-images of each frame are input to the image semantic network model for semantic recognition to obtain each frame.
  • the semantic feature image includes detailed feature information and edge structure information of the sample image.
  • the superimposed high-resolution sub-image of the sample image and the semantic feature image of the corresponding sample image are superimposed to obtain a superimposed image.
  • the superimposed image is a semantic-enhanced high-resolution image.
  • the superimposed image and the corresponding sample image of each frame are superimposed.
  • the high-resolution feature images of the two images are compared to obtain the degree of difference between the high-resolution feature image of the sample image and the superimposed image of the corresponding sample image.
  • the greater the degree of difference the smaller the similarity between the high-resolution feature image obtained by the super-segment network model and the superimposed image (that is, the semantic-enhanced high-resolution image), that is, the high-resolution feature is obtained by the super-segment network model.
  • the quality of the image is poor; on the contrary, the smaller the difference, it indicates that the similarity between the high-resolution feature image obtained by the super-segmented network model and the superimposed image (that is, the high-resolution image with enhanced semantics) is greater, that is, The quality of the high-resolution feature images obtained by the sub-network model is better. Therefore, the sum of differences can be calculated, and the sum of differences can be used as the error of the super network model.
  • the error of the super network model refers to the error that the super network model converts the image into a high-resolution image.
  • the quality of the high-resolution images processed by the sub-network model is poor; on the contrary, the smaller the error, it indicates that the quality of the high-resolution images processed by the super-network model is better.
  • each convolutional layer includes N k * k convolution kernels, where N can be any of [20 100] Integer, k can be 3 or 5.
  • the image processing device can obtain high-resolution sub-images and low-resolution sub-images of each frame of sample images in the N-frame sample images, and input the low-resolution sub-images of each frame of sample images into the super-scoring network model for processing.
  • the high-resolution feature image corresponding to the frame sample image is extracted with feature information of each frame of the high-resolution feature image, and is identified as f W (x j ), where x j represents the j-th sample image.
  • the target sub-image is input into the super-scoring network and S operation is performed to obtain a semantic feature image.
  • the semantic feature image is superimposed with a high-resolution sub-image to obtain a superimposed image.
  • the feature information of the superimposed image is extracted and identified as f s (y j ) + z j , y j is the target image corresponding to the j-th sample image, f s (y j ) represents the feature information of the semantic feature image of the target image corresponding to the j-th sample image, and z j is the height of the j-frame sample image.
  • Feature information of the resolution sub-image is extracted and identified as f s (y j ) + z j , y j is the target image corresponding to the j-th sample image, f s (y j ) represents the feature information of the semantic feature image of the target image corresponding to the j-th sample image, and z j is the height of the j-frame sample image.
  • the feature information of the high-resolution feature image of each frame is compared with the feature information of the corresponding superimposed image to determine the degree of difference between the high-resolution feature image of the sample image and the superimposed image of the corresponding sample image;
  • the difference sum is described, and the difference sum is used as an error of the super-scoring network model, and identified as W.
  • the error of the super-segmentation network can be expressed by equation (1).
  • MSE (f W (x j ), fs (y j ) + z j ) represents the feature information of the superimposed image of the j-th sample image and the feature of the high-resolution feature image of the j-th sample image. The degree of difference in information.
  • the image processing apparatus may set weights for the images output by the image semantic network model, and process the semantic feature images of the sample images in each frame according to the weights to obtain processed semantic feature maps.
  • the high-resolution sub-image of the sample image and the processed semantic feature image corresponding to the sample image are superimposed to obtain a superimposed image.
  • the image processing device can set weights for the image output by the image semantic network model according to the scene or according to the needs of the user, and process the semantic feature images corresponding to each frame sample image according to the weights to obtain the processed semantic feature images.
  • the super-resolution sub-image of the image is superimposed with the processed semantic feature image of the corresponding sample image to obtain a superimposed image.
  • the larger the weight value the more information the semantic feature image provides in the superimposed image, the higher the definition of the superimposed image, and further, the high-resolution image output by the super-network model is closer to the semantic feature image; otherwise, the weight The smaller the value, the less the information provided by the semantic feature image in the superimposed image, the lower the clarity of the superimposed image, and the closer the high-resolution image output by the super-scoring network model is to the target sub-image.
  • the weight set for the image output by the semantic network model is ⁇ .
  • the semantic feature image corresponding to each frame of the sample image is processed to obtain the processed semantic feature image, and the height of each frame of the sample image is high.
  • the resolution sub-image is superimposed with the processed semantic feature image of the corresponding sample image to obtain a superimposed image, and the feature information of the superimposed image is extracted. It can be identified as ⁇ f s (y j ) + z j , and ⁇ f s (y j ) is the processed image.
  • the feature information of the semantic feature image, and further, the error of the super-scoring network can be expressed by equation (2).
  • step S12 includes: inputting a target sub-image into the image semantic network model, and performing multi-layer neural network recognition through a multi-layer neural network included in the image semantic network model to output multiple frames of candidate feature images, and each layer of the neural network output One frame of candidate feature image, performing grayscale processing on each frame of the candidate feature image to obtain a grayscale image, determining a parameter value of each grayscale image, and using the grayscale image with the largest parameter value as the corresponding target sub-image For the semantic feature image of the sample image, the parameter value is determined according to the sharpness of the grayscale image and / or the amount of information provided by the grayscale image.
  • the image processing apparatus may input a target sub-image into the image semantic network model, and perform multi-layer neural network included in the image semantic network model to perform semantic recognition and output multiple frames of candidate feature images.
  • step S202 Determine whether the error is less than or equal to a preset error value.
  • the image processing device can determine whether the error is less than or equal to the preset error value. When the error is less than or equal to the preset error value, it indicates that the super-scoring network model can output a high-quality high-resolution image, and step S204 can be performed; otherwise, when the error If it is greater than the preset error value, it indicates that the super-scoring network model cannot output a high-quality high-resolution image, and step S205 may be performed.
  • the error When the error is greater than the preset error value, adjust the network parameters of the super network model and repeat S201.
  • the error of the execution of the super network model is less than or equal to the preset error value, so that the super network model can output high-quality High resolution image.
  • the error of the super-segmentation network model is smaller than the preset error, it indicates that the super-segmentation network model can output a high-quality high-resolution image, and the image processing device can obtain a target image that needs to be super-resolution-processed.
  • the target image is input to a super-scoring network model for processing to obtain a high-resolution image.
  • the target image is input to the super-scoring network model for processing to obtain a high-resolution image, so that more detailed feature information and higher-definition edge feature information can be obtained from the high-resolution image.
  • the network parameters of the super-segmented network model are adjusted according to a large number of sample images and the semantic feature images of each frame of the sample images, the semantic feature images contain detailed feature information of the sample images And edge structure information. Therefore, the high-resolution image obtained through the super-scoring network model can provide more detailed feature information and provide high-resolution edge structure information, which improves the quality of high-resolution images.
  • FIG. 4 is a schematic structural diagram of an image processing apparatus according to an embodiment of the present invention.
  • the image processing apparatus described in this embodiment includes:
  • the obtaining module 401 is configured to obtain a target image that needs to be subjected to super-resolution processing.
  • a processing module 402 configured to input the target image into a super-scoring network model for processing to obtain a high-resolution image
  • the network parameters of the super-segmented network model are adjusted according to the multi-frame sample images and the semantic feature maps corresponding to the sample images in each frame, and the semantic feature maps are obtained through semantic recognition of the image semantic network model.
  • a determining module 403 is configured to determine an error of the super-scoring network model according to the multi-frame sample images and a semantic feature map corresponding to each of the sample images.
  • An adjustment module 404 is configured to adjust network parameters of the ultra-scoring network model when the error is greater than a preset error value.
  • the determining module 403 is specifically configured to obtain a high-resolution sub-image and a low-resolution sub-image corresponding to each frame of the multi-frame sample images; input each frame of the target sub-image into the image semantic network model for semantics
  • the semantic feature image corresponding to each of the sample images is identified, and the target sub-image is a high-resolution sub-image or a low-resolution sub-image corresponding to any one of the multi-frame sample images;
  • the low-resolution sub-images are input into the super-scoring network model and processed to obtain high-resolution feature images of the sample image in each frame; the high-resolution sub-images of the sample image in each frame and the semantic features of the corresponding sample image Superimpose the images to obtain a superimposed image; determine the difference between the high-resolution feature image of each sample image and the superimposed image of the corresponding sample image; calculate the sum of the differences, and use the sum of the differences as the error.
  • a setting module 405 is configured to set a weight for an image output by the image semantic network model.
  • the processing module 402 is further configured to process the semantic feature image of the sample image of each frame according to the weights to obtain a processed semantic feature map.
  • the determining module 403 is specifically configured to superimpose the high-resolution sub-image of the sample image of each frame and the processed semantic feature image of the corresponding sample image to obtain a superimposed image.
  • the determining module 403 is specifically configured to input the target sub-image into the image semantic network model, and perform multi-layer neural network included in the image semantic network model to perform semantic recognition and output multiple frame candidate feature images, each layer
  • the neural network outputs a frame of candidate feature images; performing grayscale processing on each frame of the candidate feature images to obtain a grayscale image; determining a parameter value of the grayscale image of each frame, and using the grayscale image with the largest parameter value as the grayscale image
  • the parameter value of the semantic feature of the sample image corresponding to the target sub-image is determined according to the sharpness of the gray image and / or the amount of information provided by the gray image.
  • the obtaining module 401 is further configured to obtain a type of the target image; and determine a super-scoring network model that matches the type of the target image.
  • the processing module 402 is configured to input the target image into a super-scoring network model matching the type of the target image for processing to obtain a high-resolution image.
  • the network parameters of the super-segmented network model are adjusted based on a large number of sample images and the semantic feature images of each frame of the sample images, the semantic feature images contain detailed feature information of the sample images. And edge structure information. Therefore, the high-resolution image obtained through the super-scoring network model can provide more detailed feature information and provide high-resolution edge structure information, which improves the quality of high-resolution images.
  • FIG. 5 is a schematic structural diagram of an electronic device according to an embodiment of the present invention.
  • the electronic device includes a processor 501, a memory 502, a communication interface 503, and a power source 504.
  • the processor 501, the memory 502, the communication interface 503, and the power source 504 are connected to each other through a bus.
  • the processor 501 may be one or more CPUs. In the case where the processor 501 is a CPU, the CPU may be a single-core CPU or a multi-core CPU.
  • the processor 501 and the processor 501 may include a modem for The signal received by the transceiver 805 is subjected to modulation or demodulation processing.
  • the memory 502 includes, but is not limited to, RAM, ROM), EPROM, and CD-ROM.
  • the memory 502 is used to store instructions, an operating system, various applications, and data.
  • the communication interface 503 is connected to a forwarding plane device or other control plane devices.
  • the communication interface 503 includes multiple interfaces, which are respectively connected to multiple terminals or connected to a forwarding plane device.
  • the communication interface 503 may be a wired interface, a wireless interface, or a combination thereof.
  • the wired interface may be, for example, an Ethernet interface.
  • the Ethernet interface can be an optical interface, an electrical interface, or a combination thereof.
  • the wireless interface may be, for example, a wireless local area network (English: wireless local area network, abbreviation: WLAN) interface, a cellular network interface, or a combination thereof.
  • WLAN wireless local area network
  • the power supply 504 is configured to supply power to a control plane device.
  • the memory 502 is also used to store program instructions.
  • the processor 501 may call the program instructions stored in the memory 502 to implement the image processing method as shown in the foregoing embodiments of the present application.
  • control plane device provided in the embodiment of the present invention is similar to that of the method embodiment of the present invention. Therefore, the implementation and beneficial effects of the control plane device can be referred to as well as the beneficial effects. More details.
  • the present invention also provides a computer-readable storage medium on which a computer program is stored.
  • a computer program is stored.
  • the implementation of the present invention also provides a computer program product.
  • the computer program product includes a non-volatile computer-readable storage medium storing a computer program.
  • the computer program executes the corresponding embodiments of FIG. 1 and FIG. 2 described above.
  • the steps of the image processing method in FIG. 1, and the implementation and beneficial effects of the computer program product for solving the problem refer to the implementation and beneficial effects of the image processing method in FIG. 1 and FIG. 2 described above.

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

L'invention concerne un procédé, un appareil et un dispositif de traitement d'image. Le procédé consiste à : acquérir une image cible qui nécessite un traitement à traite haute résolution ; et entrer l'image cible dans un modèle de réseau à très haute résolution pour un traitement afin d'obtenir une image à haute résolution, les paramètres de réseau du modèle de réseau à très haute résolution étant obtenus en réalisant un ajustement en fonction des multiples trames d'une image d'échantillon et d'une carte de caractéristiques sémantiques correspondant à chaque trame de l'image d'échantillon, et la carte de caractéristiques sémantiques étant obtenue au moyen d'un modèle de réseau sémantique d'images effectuant une reconnaissance sémantique, ce qui permet d'améliorer la qualité de l'image à haute résolution obtenue.
PCT/CN2018/108891 2018-09-29 2018-09-29 Procédé, appareil et dispositif de traitement d'image WO2020062191A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2018/108891 WO2020062191A1 (fr) 2018-09-29 2018-09-29 Procédé, appareil et dispositif de traitement d'image
CN201880093293.9A CN112088393B (zh) 2018-09-29 2018-09-29 图像处理方法、装置及设备

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2018/108891 WO2020062191A1 (fr) 2018-09-29 2018-09-29 Procédé, appareil et dispositif de traitement d'image

Publications (1)

Publication Number Publication Date
WO2020062191A1 true WO2020062191A1 (fr) 2020-04-02

Family

ID=69952653

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/108891 WO2020062191A1 (fr) 2018-09-29 2018-09-29 Procédé, appareil et dispositif de traitement d'image

Country Status (2)

Country Link
CN (1) CN112088393B (fr)
WO (1) WO2020062191A1 (fr)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111932463A (zh) * 2020-08-26 2020-11-13 腾讯科技(深圳)有限公司 图像处理方法、装置、设备及存储介质
CN112016542A (zh) * 2020-05-08 2020-12-01 珠海欧比特宇航科技股份有限公司 城市积涝智能检测方法及系统
US20210209732A1 (en) * 2020-06-17 2021-07-08 Beijing Baidu Netcom Science And Technology Co., Ltd. Face super-resolution realization method and apparatus, electronic device and storage medium
CN113592709A (zh) * 2021-02-19 2021-11-02 腾讯科技(深圳)有限公司 图像超分处理方法、装置、设备及存储介质
CN116612466A (zh) * 2023-07-20 2023-08-18 腾讯科技(深圳)有限公司 基于人工智能的内容识别方法、装置、设备及介质
CN116883236A (zh) * 2023-05-22 2023-10-13 阿里巴巴(中国)有限公司 图像超分方法以及图像数据处理方法

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080162561A1 (en) * 2007-01-03 2008-07-03 International Business Machines Corporation Method and apparatus for semantic super-resolution of audio-visual data
CN105793891A (zh) * 2013-11-30 2016-07-20 夏普株式会社 用于确定高分辨率输出图像的方法和设备
CN107169450A (zh) * 2017-05-15 2017-09-15 中国科学院遥感与数字地球研究所 一种高分辨率遥感图像的场景分类方法及系统
CN108428212A (zh) * 2018-01-30 2018-08-21 中山大学 一种基于双拉普拉斯金字塔卷积神经网络的图像放大方法
CN108537731A (zh) * 2017-12-29 2018-09-14 西安电子科技大学 基于压缩多尺度特征融合网络的图像超分辨率重建方法

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10929977B2 (en) * 2016-08-25 2021-02-23 Intel Corporation Coupled multi-task fully convolutional networks using multi-scale contextual information and hierarchical hyper-features for semantic image segmentation
CN106780363B (zh) * 2016-11-21 2019-07-23 北京金山安全软件有限公司 一种图片处理方法、装置及电子设备
CN108229455B (zh) * 2017-02-23 2020-10-16 北京市商汤科技开发有限公司 物体检测方法、神经网络的训练方法、装置和电子设备

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080162561A1 (en) * 2007-01-03 2008-07-03 International Business Machines Corporation Method and apparatus for semantic super-resolution of audio-visual data
CN105793891A (zh) * 2013-11-30 2016-07-20 夏普株式会社 用于确定高分辨率输出图像的方法和设备
CN107169450A (zh) * 2017-05-15 2017-09-15 中国科学院遥感与数字地球研究所 一种高分辨率遥感图像的场景分类方法及系统
CN108537731A (zh) * 2017-12-29 2018-09-14 西安电子科技大学 基于压缩多尺度特征融合网络的图像超分辨率重建方法
CN108428212A (zh) * 2018-01-30 2018-08-21 中山大学 一种基于双拉普拉斯金字塔卷积神经网络的图像放大方法

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112016542A (zh) * 2020-05-08 2020-12-01 珠海欧比特宇航科技股份有限公司 城市积涝智能检测方法及系统
US20210209732A1 (en) * 2020-06-17 2021-07-08 Beijing Baidu Netcom Science And Technology Co., Ltd. Face super-resolution realization method and apparatus, electronic device and storage medium
US11710215B2 (en) * 2020-06-17 2023-07-25 Beijing Baidu Netcom Science And Technology Co., Ltd. Face super-resolution realization method and apparatus, electronic device and storage medium
CN111932463A (zh) * 2020-08-26 2020-11-13 腾讯科技(深圳)有限公司 图像处理方法、装置、设备及存储介质
CN111932463B (zh) * 2020-08-26 2023-05-30 腾讯科技(深圳)有限公司 图像处理方法、装置、设备及存储介质
CN113592709A (zh) * 2021-02-19 2021-11-02 腾讯科技(深圳)有限公司 图像超分处理方法、装置、设备及存储介质
CN113592709B (zh) * 2021-02-19 2023-07-25 腾讯科技(深圳)有限公司 图像超分处理方法、装置、设备及存储介质
CN116883236A (zh) * 2023-05-22 2023-10-13 阿里巴巴(中国)有限公司 图像超分方法以及图像数据处理方法
CN116883236B (zh) * 2023-05-22 2024-04-02 阿里巴巴(中国)有限公司 图像超分方法以及图像数据处理方法
CN116612466A (zh) * 2023-07-20 2023-08-18 腾讯科技(深圳)有限公司 基于人工智能的内容识别方法、装置、设备及介质
CN116612466B (zh) * 2023-07-20 2023-09-29 腾讯科技(深圳)有限公司 基于人工智能的内容识别方法、装置、设备及介质

Also Published As

Publication number Publication date
CN112088393B (zh) 2022-09-23
CN112088393A (zh) 2020-12-15

Similar Documents

Publication Publication Date Title
WO2020062191A1 (fr) Procédé, appareil et dispositif de traitement d'image
CN110555795B (zh) 高解析度风格迁移
CN110660037B (zh) 图像间脸部交换的方法、装置、系统和计算机程序产品
CN109493350B (zh) 人像分割方法及装置
CN107330439B (zh) 一种图像中物体姿态的确定方法、客户端及服务器
US20220222776A1 (en) Multi-Stage Multi-Reference Bootstrapping for Video Super-Resolution
CN112823379B (zh) 用于训练机器学习模型的方法和装置、用于视频风格转移的装置
CN111507333B (zh) 一种图像矫正方法、装置、电子设备和存储介质
US20210256663A1 (en) Image processing method and apparatus, computer device, and storage medium
US11481683B1 (en) Machine learning models for direct homography regression for image rectification
CN112990219B (zh) 用于图像语义分割的方法和装置
CN110428382B (zh) 一种用于移动终端的高效视频增强方法、装置和存储介质
WO2023035531A1 (fr) Procédé de reconstruction à super-résolution pour image de texte et dispositif associé
CN113487618B (zh) 人像分割方法、装置、电子设备及存储介质
US20230343119A1 (en) Captured document image enhancement
CN104200498B (zh) 融合Cortex‑A7的实时视频超分辨率处理方法
WO2024027583A1 (fr) Procédé et appareil de traitement d'images, et dispositif électronique ainsi que support de stockage lisible
CN114863539A (zh) 一种基于特征融合的人像关键点检测方法及系统
CN114830168A (zh) 图像重建方法、电子设备和计算机可读存储介质
TW202240535A (zh) 用於影像處理的方法及系統
WO2022016996A1 (fr) Dispositif et procédé de traitement d'image, appareil électronique et support de stockage lisible par ordinateur
TWI711004B (zh) 圖片處理方法和裝置
CN116934591A (zh) 多尺度特征提取的图像拼接方法、装置、设备及存储介质
CN116310615A (zh) 图像处理方法、装置、设备及介质
US20240119633A1 (en) Method for converting near infrared image to rgb image and apparatus for same

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18934890

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18934890

Country of ref document: EP

Kind code of ref document: A1