WO2023202570A1 - Procédé de traitement d'image et appareil de traitement d'image, dispositif électronique et support de stockage lisible - Google Patents

Procédé de traitement d'image et appareil de traitement d'image, dispositif électronique et support de stockage lisible Download PDF

Info

Publication number
WO2023202570A1
WO2023202570A1 PCT/CN2023/088947 CN2023088947W WO2023202570A1 WO 2023202570 A1 WO2023202570 A1 WO 2023202570A1 CN 2023088947 W CN2023088947 W CN 2023088947W WO 2023202570 A1 WO2023202570 A1 WO 2023202570A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
area
target area
information
processing
Prior art date
Application number
PCT/CN2023/088947
Other languages
English (en)
Chinese (zh)
Inventor
任帅
Original Assignee
维沃移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 维沃移动通信有限公司 filed Critical 维沃移动通信有限公司
Publication of WO2023202570A1 publication Critical patent/WO2023202570A1/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/04Context-preserving transformations, e.g. by using an importance map
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/77Retouching; Inpainting; Scratch removal
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Definitions

  • This application belongs to the field of image processing technology, and specifically relates to an image processing method and processing device, electronic equipment and readable storage media.
  • Processing methods in related technologies such as erasing specific areas through image restoration, only use the texture information of the erased area, resulting in poor erasure effects and leaving obvious repair traces on the picture. If you use deep learning to process pictures, you can reduce repair traces, but it is easy to erase other areas in the picture that are similar to specific areas, and the processing effect is not good.
  • the purpose of the embodiments of the present application is to provide an image processing method and processing device, electronic equipment and readable storage medium, which can solve the problem of poor processing effect of erasing image content in related technologies.
  • embodiments of the present application provide an image processing method, including:
  • the image information of the preset area in the first image is erased to obtain an erased second image, where the size of the second image is the same as the first image, and the preset area includes the target area;
  • a processed target image is generated based on the area image and the first image.
  • an image processing device including:
  • a positioning module used to locate the target area in the first image and obtain the position information of the target area
  • the erasing module is used to erase the image information of the preset area in the first image through the image processing model to obtain an erased second image, where the size of the second image is the same as the first image, and the preset area includes target area;
  • An interception module configured to intercept an area image corresponding to the target area in the second image based on the location information
  • a processing module configured to generate a processed target image based on the area image and the first image.
  • embodiments of the present application provide an electronic device, including a processor and a memory.
  • the memory stores programs or instructions that can be run on the processor.
  • the program or instructions are executed by the processor, the method of the first aspect is implemented. step.
  • embodiments of the present application provide a readable storage medium that stores a program or instructions, and when the program or instructions are executed by a processor, the steps of the method in the first aspect are implemented.
  • inventions of the present application provide a chip.
  • the chip includes a processor and a communication interface.
  • the communication interface is coupled to the processor.
  • the processor is used to run programs or instructions to implement the steps of the method in the first aspect. .
  • embodiments of the present application provide a computer program product, the program product is stored in a storage medium, and the program product is executed by at least one processor to implement the method as described in the first aspect.
  • an embodiment of the present application provides an image processing device, including the device being configured to perform the method described in the first aspect.
  • the image area that needs to be processed is first positioned and the location information is recorded. Then, image processing is performed on the first image as a whole through the adversarial model.
  • the image processing steps are based on image recognition technology and deep learning technology.
  • the model automatically identifies the parts that need to be processed, and based on the global information of the first image, the parts that need to be processed are Erase the preset area and keep the image of the erased preset area consistent with the overall image.
  • the area image at the same position is intercepted, and the area image is combined with the original first image. , that is, only the parts of the second image that the user needs to eliminate are selected to process the original image. Therefore, the final target image retains the consistency of the overall image and the content of the target area that needs to be processed. Erasing is performed while ensuring that other similar areas in the image will not be erased by mistake, which improves the processing efficiency when erasing specific content in the image.
  • Figure 1 shows one of the flowcharts of an image processing method according to an embodiment of the present application
  • Figure 2 shows one of the schematic diagrams of an image processing method according to an embodiment of the present application
  • Figure 3 shows the second schematic diagram of the image processing method according to the embodiment of the present application.
  • Figure 4 shows the third schematic diagram of the image processing method according to the embodiment of the present application.
  • Figure 5 shows a schematic structural diagram of an image processing model according to an embodiment of the present application
  • Figure 6 shows the second flowchart of the image processing method according to the embodiment of the present application.
  • Figure 7 shows a structural block diagram of an image processing device according to an embodiment of the present application.
  • Figure 8 shows a structural block diagram of an electronic device according to an embodiment of the present application.
  • FIG. 9 is a schematic diagram of the hardware structure of an electronic device implementing an embodiment of the present application.
  • first, second, etc. in the description and claims of this application are used to distinguish similar objects and are not used to describe a specific order or sequence. It is to be understood that the figures so used are interchangeable under appropriate circumstances so that the embodiments of the present application can be practiced in orders other than those illustrated or described herein, and that "first,”"second,” etc. are distinguished
  • the object is usually A category does not limit the number of objects.
  • the first object can be one or multiple.
  • “and/or” in the description and claims indicates at least one of the connected objects, and the character “/" generally indicates that the related objects are in an "or” relationship.
  • Figure 1 shows one of the flowcharts of the image processing method according to the embodiment of the present application. As shown in Figure 1, the method includes:
  • Step 102 locate the target area in the first image and obtain the location information of the target area
  • Step 104 Erase the image information of the preset area in the first image through the image processing model to obtain the erased second image;
  • step 104 the size of the second image is the same as the first image, and the preset area includes the target area;
  • Step 106 According to the position information, intercept the area image corresponding to the target area in the second image;
  • Step 108 Generate a processed target image based on the area image and the first image.
  • the text area in the image can be Position and record the location information corresponding to the area.
  • the location information may be coordinate information.
  • the original image of the first image is input into the preset image processing model.
  • the image processing model can perform the first image processing according to the type of information that the user needs to erase. Automatically identify the target information to be erased in an image, and process the entire first image based on the global information of the target image.
  • Figure 2 shows one of the schematic diagrams of an image processing method according to an embodiment of the present application.
  • the first image 200 includes text information 202, and the user needs to erase the text information 202.
  • OCR Optical Character Recognition
  • the area where the text information 202 is located in the first image 200 is marked, that is, the target area.
  • Domain 204 simultaneously records the coordinate information of the target area 204, thereby recording the location information of the target area 204.
  • the first image is input to a preset image processing model.
  • the image processing model can automatically identify the image information content that the user wants to erase, such as text information, and erase it based on the global information of the target image, including color information, pixel information, etc.
  • FIG 3 shows the second schematic diagram of the image processing method according to the embodiment of the present application.
  • the image processing model processes the entire first image 300 and processes several pre-images that are identified as containing text. All areas are erased.
  • the image processing model performs erasure processing on both preset areas.
  • the first preset area 302 contains the text information that the user needs to erase, and the second preset area 304 contains Because the characteristics of the QR code information and text information are close, the QR code information is mistakenly recognized as text and erased.
  • the image processing model outputs the second image, according to the identified position information of the target area, in the second image, the corresponding area image is intercepted according to the same coordinates, and the area image is intercepted according to the coordinates of the target area. , superimposed on the original image of the first image, thereby covering the target area of the first image, thereby generating the target image.
  • Figure 4 shows the third schematic diagram of the image processing method according to the embodiment of the present application. As shown in Figure 4, on the target image 400, the text information in the target area 402 is erased, while the QR code 404 is retained.
  • the embodiment of the present application intercepts the area image at the same position in the second image erased by the image processing model, and combines the area image with the original first image, that is, Only the parts of the second image that the user needs to eliminate are selected to process the original image. Therefore, the final target image has the content of the target area that needs to be processed while retaining the coordination of the overall image. Erase, while ensuring that other similar areas in the image will not be erased by mistake, improving the processing efficiency when erasing specific content in the image.
  • generating a processed target image based on the area image and the first image includes: covering the area image on the target area according to the position information to obtain the target image.
  • the intercepted regional image processed by the image processing model is overlaid on the original first image according to the coordinate information of the target area, so that the image content in the target area is completely replaced by area image, so that in the replaced target image, only the area that needs to be erased is replaced by the processed image, and the content of the target area that needs to be processed is erased while retaining the consistency of the overall image. It also ensures that other similar areas in the image will not be erased by mistake, which improves the processing efficiency when erasing specific content in the image.
  • the method before erasing the image information of the preset area in the first image, the method further includes:
  • the second training image is an image obtained by removing preset image information from the first training image
  • Erase the image information of the preset area in the first image including:
  • a preset adversarial model is trained to obtain a trained image processing model. Specifically, first, a first image is collected, and the first image is manually processed to generate a second image. The first image is an original image, and the second image is a second image obtained by erasing preset areas such as text content through image editing or image modification software.
  • a preset generative adversarial network is trained to perform end-to-end erasing of text in the entire image.
  • the specific method is to build a network model.
  • the unet network structure is used.
  • the image is first down-sampled to obtain the image semantic information, and then the image is up-sampled and restored to the original size to obtain the image output.
  • the unet a U-shaped network for dense prediction segmentation
  • a layer of lightweight unet results are added for further processing to obtain the final processing result.
  • the discriminant network includes a dual-scale network.
  • the first scale network includes several cascaded convolutions, in which the step size of the convolution can be set to 1.
  • the pooling layer is not introduced to ensure The image resolution is not reduced.
  • the input of the second layer of scale network is obtained by downsampling the output of the first layer network twice.
  • the ground truth of the dual-scale network is obtained from the erased image and the image that is downsampled twice as much as the erased image.
  • the output result of the generation network is input to the discriminant network, and the difference between the currently generated erased image and the original annotated erased image (i.e., the second image) is judged through the discriminant network, and is used as the reverse loss (loss) Propagate the optimized network parameters and finally obtain the optimized network structure.
  • FIG. 5 shows a schematic structural diagram of an image processing model according to an embodiment of the present application.
  • the image processing model 500 includes a 2-layer unet network structure, that is, a first network 502 and a second network 504.
  • the first image 506 to be processed is input into the first network 502, and the specific information therein is processed through the first network 502. Erase.
  • the processed third image 508 is obtained.
  • the third image 508 is down-sampled to obtain a fourth image 510 with reduced resolution, thereby obtaining the image semantic definition.
  • the fourth image 510 obtained by down-sampling is used as the second image.
  • layer network that is, the input of the second network 504.
  • the fifth image 512 output by the second network 504 is upsampled and restored to its original size to obtain the final second image 514.
  • the content contained in the first image is the content that the user needs to erase. If the user needs to erase the text information in the image, the first image contains the text information. If the user needs to erase the face, the first image contains face information.
  • This application trains the image recognition model so that the image recognition model can The concentrated image content is erased, and the corresponding content area in the first image is erased, so that the erased image can maintain the coordination of the overall image and improve image processing efficiency.
  • the target area is a character image area
  • locating the target area in the first image includes:
  • the target area specifically includes a character image area, that is to say, the user needs to erase the character area in the first image.
  • OCR optical character recognition
  • the first image is first preprocessed.
  • a denoising algorithm is used to remove noise on the first image.
  • the text or characters are obtained through the OCR detection algorithm, and the coordinate information is located.
  • the OCR detection algorithm can obtain character detection frames in various scenarios such as horizontal, vertical, and curved.
  • a four-point coordinate frame can be used to represent them; while for curved irregular character detection frames, an eight-point coordinate frame can be used. If there is no text information or character information in the current picture, an empty character coordinate box will be returned.
  • the character detection frame is marked with coordinate information, and the coordinate information refers to the coordinates of the character detection frame in the first image.
  • the target area is located in the first image through the coordinate information of the character detection frame, so that after the second image is obtained through the image processing model, the processed area image is overlaid on the target area according to the coordinate information, so that the generated In the target image, it not only ensures the coordination of the overall image, but also effectively erases the content of the target area that needs to be processed. It also ensures that other similar areas in the image will not be mistakenly erased, which improves the efficiency of erasing specific areas in the image. content processing efficiency.
  • Figure 6 shows the second flowchart of the image processing method according to the embodiment of the present application. As shown in Figure 6, the method includes:
  • Step 602 Perform text positioning on the original image and record the text coordinate information in the original image
  • step 602 the image is first preprocessed, and a denoising algorithm is used to reduce the noise on the image. Remove. Then train the OCR detection model and obtain text positioning information through the OCR detection algorithm.
  • the OCR detection algorithm can obtain text detection frames in various scenarios such as horizontal, vertical, and curved. Horizontal and vertical text boxes are represented by a four-point coordinate frame, while curved text boxes are represented by an eight-point coordinate frame. If there is no text information in the current picture, an empty text coordinate box will be returned.
  • Step 604 Collect paired data and train a generative adversarial network to perform text erasure
  • step 604 pairs of ⁇ original pictures, text-erased pictures> are collected, and model training is performed using a generative adversarial method to obtain an image processing model.
  • the original image is first collected, and the corresponding text-erased image is obtained through PS, thereby obtaining a pair of ⁇ original image, text-erased image>.
  • a generative adversarial network is trained to perform end-to-end erasure of entire image text.
  • the specific method is to build a network model.
  • the unet network structure is used.
  • the image is first down-sampled to obtain image semantic information, and then the image is up-sampled. Restore the original size to obtain the image output.
  • a layer of lightweight unet is added after the unet structure for further processing to obtain the final processing result.
  • the discriminant network consists of a dual-scale network.
  • the first scale network consists of several convolution cascades.
  • the step size of the convolution is set to 1.
  • the pooling layer is not introduced to ensure the resolution of the image. does not decrease, then add a layer of the same scale network after the first scale network, but the input of the second layer of scale network is obtained by downsampling twice the output of the first layer of network.
  • the ground truth of the dual-scale network is obtained from the erased image and the image that is downsampled twice as much as the erased image.
  • the output result of the generation network is input into the discriminant network to determine the difference between the currently generated erased image and the original annotated erased image, and is used as loss backpropagation to optimize the network parameters, and finally the optimized network structure is obtained.
  • Step 606 Use the optimized model to infer the original image to obtain the erased image result
  • step 606 the optimized generative adversarial model is used to infer the original image to obtain the erased image result.
  • the specific method is to input the original image into the trained model.
  • the discriminator needs to be removed and only the generator is retained. That is, the output image obtained after passing through the generator is the wiped Image result after division.
  • Step 608 Map the obtained text coordinate information to the erased picture, and crop out the erased area where the text coordinates are located;
  • step 608 map the text coordinate information obtained in the step to the obtained erased picture result, and cut out the erased area where the text coordinates are located; considering that some areas that are very similar to text, such as fences and textures, , flowers, grass, etc. are easily erased as text. Therefore, in order to ensure that this information is not mistakenly scrawled, it is necessary to map the text coordinate information recorded from the original image to the acquired erased image. Because the coordinates are non-rectangular frames, it is necessary to Set a mask image as large as the input image. The mask image is originally pure black. Set the area where the coordinate box is located to white to get the erased area where the text coordinates that need to be cropped are located.
  • Step 610 Paste the erased area back to the original image to obtain the final text-erased image.
  • step 610 the cropped erasure area is pasted back to the original image to obtain the final text erasure image.
  • the specific method is to obtain the mask image corresponding to the text area that needs to be erased. At this time, you only need to paste the erased area corresponding to the pure white area in the mask image back to the original image, so that you can get the final result.
  • the required text is erased from the picture, thus avoiding the accidental painting of areas that are very similar to the text.
  • an image processing device is provided.
  • Figure 7 shows a structural block diagram of the image processing device according to an embodiment of the present application. As shown in Figure 7, the image processing device 700 includes:
  • Positioning module 702 is used to locate the target area in the first image and obtain the location information of the target area
  • the erasing module 704 is used to erase the image information of the preset area in the first image through the image processing model to obtain an erased second image, where the size of the second image is the same as the first image, and the preset area Include target areas;
  • the interception module 706 is used to intercept the area image corresponding to the target area in the second image according to the location information
  • the processing module 708 is used to generate a processed target image according to the area image and the first image.
  • the embodiment of the present application intercepts the area image at the same position in the second image erased by the image processing model, and combines the area image with the original first image, that is, Only the parts of the second image that the user needs to eliminate are selected to process the original image. Therefore, the final target image has the content of the target area that needs to be processed while retaining the coordination of the overall image. Erase, while ensuring that other similar areas in the image will not be erased by mistake, improving the processing efficiency when erasing specific content in the image.
  • the image processing device further includes: a covering module, configured to overlay the area image on the target area according to the location information to obtain the target image.
  • the embodiment of the present application also overlays the intercepted regional image processed by the image processing model on the original first image according to the coordinate information of the target area, thereby making the image content in the target area , completely replaced by a regional image, so that in the replaced target image, only the area that needs to be erased is replaced by the processed image. While retaining the consistency of the overall image, the target area that needs to be processed is improved. The content is erased while ensuring that other similar areas in the image will not be accidentally erased, which improves the processing efficiency when erasing specific content in the image.
  • the processing device further includes:
  • An acquisition module configured to acquire a first training image and a second training image, where the second training image is an image obtained by removing preset image information from the first training image;
  • a training module used to train a preset model through the first training image and the second training image to obtain a trained image processing model, where the image processing model includes a first network and a second network;
  • the erasing module is also used to erase the first image through the first network to obtain the processed third image;
  • the sampling module is used to downsample the third image to obtain the fourth image
  • the erasing module is also used to erase the fourth image through the second network to obtain the processed fifth image
  • the sampling module is also used to upsample the fifth image to obtain the second image.
  • This application trains the image recognition model so that the image recognition model can The concentrated image content is erased, and the corresponding content area in the first image is erased, so that the erased image can maintain the coordination of the overall image and improve image processing efficiency.
  • the target area is a character image area
  • the processing device further includes:
  • a recognition module configured to perform optical character recognition on the first image and obtain a character detection frame in the first image
  • the positioning module is also used to locate the target area in the first image based on the coordinate information of the character detection frame.
  • the embodiment of the present application locates the target area in the first image, so that after the second image is obtained through the image processing model, the processed area image is overlaid on the target area according to the coordinate information, so that in the generated target image, It not only ensures the coordination of the overall image, but also effectively erases the content of the target area that needs to be processed. It also ensures that other similar areas in the image will not be mistakenly erased, improving the processing when erasing specific content in the image. efficiency.
  • the image processing device in the embodiment of the present application may be an electronic device or a component in the electronic device, such as an integrated circuit or a chip.
  • the electronic device may be a terminal or other devices other than the terminal.
  • the electronic device can be a mobile phone, a tablet computer, a notebook computer, a handheld computer, a vehicle-mounted electronic device, a mobile internet device (Mobile Internet Device, MID), or augmented reality (AR)/virtual reality (VR).
  • the image processing device in the embodiment of the present application may be a device with an operating system.
  • the operating system can be an Android operating system, an iOS operating system, or other possible operating systems, which are not specifically limited in the embodiments of this application.
  • the image processing device provided by the embodiments of the present application can implement various processes implemented by the above method embodiments. To avoid duplication, they will not be described again here.
  • the embodiment of the present application also provides an electronic device.
  • Figure 8 shows a structural block diagram of the electronic device according to the embodiment of the present application.
  • the electronic device 800 includes a processor 802, a memory 804, and a storage device 800.
  • the electronic devices in the embodiments of the present application include the above-mentioned mobile electronic devices and non-mobile electronic devices.
  • FIG. 9 is a schematic diagram of the hardware structure of an electronic device implementing an embodiment of the present application.
  • the electronic device 900 includes but is not limited to: radio frequency unit 901, network module 902, audio output unit 903, input unit 904, sensor 905, display unit 906, user input unit 907, interface unit 908, memory 909, processor 910 and other components .
  • the electronic device 900 may also include a power supply (such as a battery) that supplies power to various components.
  • the power supply may be logically connected to the processor 910 through a power management system, thereby managing charging, discharging, and function through the power management system. Consumption management and other functions.
  • the structure of the electronic device shown in Figure 9 does not constitute a limitation on the electronic device.
  • the electronic device may include more or less components than shown in the figure, or combine certain components, or arrange different components, which will not be described again here. .
  • the processor 910 is used to locate the target area in the first image and obtain the location information of the target area;
  • the image information of the preset area in the first image is erased to obtain an erased second image, where the size of the second image is the same as the first image, and the preset area includes the target area;
  • a processed target image is generated based on the area image and the first image.
  • the embodiment of the present application intercepts the area image at the same position in the second image erased by the image processing model, and combines the area image with the original first image, that is, Only the parts of the second image that the user needs to eliminate are selected to process the original image. Therefore, the final target image retains the consistency of the overall image. In this case, the content of the target area that needs to be processed is erased, while ensuring that other similar areas in the image will not be erased by mistake, which improves the processing efficiency when erasing specific content in the image.
  • the processor 910 is also configured to overlay the area image on the target area according to the location information to obtain the target image.
  • the embodiment of the present application also overlays the intercepted regional image processed by the image processing model on the original first image according to the coordinate information of the target area, thereby making the image content in the target area , completely replaced by a regional image, so that in the replaced target image, only the area that needs to be erased is replaced by the processed image. While retaining the consistency of the overall image, the target area that needs to be processed is improved. The content is erased while ensuring that other similar areas in the image will not be accidentally erased, which improves the processing efficiency when erasing specific content in the image.
  • the processor 910 is also configured to obtain a first training image and a second training image, where the second training image is image data obtained after erasing a preset area in the first training image;
  • Erase the image information of the preset area in the first image including:
  • This application trains the image recognition model, so that the image recognition model can erase the corresponding content area in the first image according to the image content in the training set, and can make the erased image maintain the coordination of the overall image. , improve image processing efficiency.
  • the target area is a character image area
  • the processor 910 is further configured to perform optical character recognition on the first image, and obtain a character detection frame in the first image;
  • the embodiment of the present application locates the target area in the first image, so as to obtain the target area through the image processing model. After arriving at the second image, based on the coordinate information, the processed area image is overlaid on the target area, so that the generated target image not only ensures the coordination of the overall image, but also adjusts the content of the target area that needs to be processed. In order to effectively erase, it also ensures that other similar areas in the image will not be accidentally erased, and improves the processing efficiency when erasing specific content in the image.
  • the input unit 904 may include a graphics processor (Graphics Processing Unit, GPU) 9041 and a microphone 9042.
  • the graphics processor 9041 is responsible for the image capture device (GPU) in the video capture mode or the image capture mode. Process the image data of still pictures or videos obtained by cameras (such as cameras).
  • the display unit 906 may include a display panel 9061, which may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like.
  • the user input unit 907 includes a touch panel 9071 and at least one of other input devices 9072 .
  • Touch panel 9071 also known as touch screen.
  • the touch panel 9071 may include two parts: a touch detection device and a touch controller.
  • Other input devices 9072 may include but are not limited to physical keyboards, function keys (such as volume control keys, switch keys, etc.), trackballs, mice, and joysticks, which will not be described again here.
  • Memory 909 can be used to store software programs as well as various data.
  • the memory 909 may mainly include a first storage area for storing programs or instructions and a second storage area for storing data, wherein the first storage area may store an operating system, an application program or instructions required for at least one function (such as a sound playback function, Image playback function, etc.) etc.
  • memory 909 may include volatile memory or nonvolatile memory, or memory 909 may include both volatile and nonvolatile memory.
  • the non-volatile memory can be read-only memory (Read-Only Memory, ROM), programmable read-only memory (Programmable ROM, PROM), erasable programmable read-only memory (Erasable PROM, EPROM), electrically removable memory.
  • Volatile memory can be random access memory (Random Access Memory, RAM), static random access memory (Static RAM, SRAM), dynamic random access memory (Dynamic RAM, DRAM), synchronous dynamic random access memory (Synchronous DRAM, SDRAM), double data rate synchronous dynamic random access memory (Double Data Rate SDRAM, DDRSDRAM), enhanced synchronous dynamic random access memory (Enhanced SDRAM, ESDRAM), synchronous link dynamic random access memory (Synch link DRAM) , SLDRAM) and Direct Rambus RAM (DRRAM).
  • Memory 909 in embodiments of the present application includes, but is not limited to, these and any other suitable types of memory.
  • the processor 910 may include one or more processing units; optionally, the processor 910 integrates an application processor and a modem processor, where the application processor mainly handles operations related to the operating system, user interface, application programs, etc., Modem processors mainly process wireless communication signals, such as baseband processors. It can be understood that the above modem processor may not be integrated into the processor 910.
  • Embodiments of the present application also provide a readable storage medium.
  • Programs or instructions are stored on the readable storage medium.
  • the program or instructions are executed by a processor, each process of the above method embodiments is implemented and the same technology can be achieved. The effect will not be described here to avoid repetition.
  • the processor is the processor in the electronic device described in the above embodiment.
  • the readable storage media includes computer-readable storage media, such as computer read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disks or optical disks, etc.
  • An embodiment of the present application further provides a chip.
  • the chip includes a processor and a communication interface.
  • the communication interface is coupled to the processor.
  • the processor is used to run programs or instructions to implement various processes of the above method embodiments. , and can achieve the same technical effect, so to avoid repetition, they will not be described again here.
  • chips mentioned in the embodiments of this application may also be called system-on-chip, system-on-a-chip, system-on-a-chip or system-on-chip, etc.
  • Embodiments of the present application provide a computer program product.
  • the program product is stored in a storage medium.
  • the program product is executed by at least one processor to implement the processes of the above method embodiments and can achieve the same technical effect. To avoid repetition, we will not go into details here.
  • the methods of the above embodiments can be implemented by means of software plus the necessary general hardware platform. Of course, it can also be implemented by hardware, but in many cases the former is better. implementation.
  • the technical solution of the present application can be embodied in the form of a computer software product that is essentially or contributes to the existing technology.
  • the computer software product is stored in a storage medium (such as ROM/RAM, disk , optical disk), including several instructions to cause a terminal (which can be a mobile phone, computer, server, or network device, etc.) to execute the methods described in various embodiments of this application.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Editing Of Facsimile Originals (AREA)
  • Processing Or Creating Images (AREA)

Abstract

La présente invention concerne un procédé de traitement d'image, un appareil de traitement d'image, un dispositif électronique et un support de stockage lisible. Le procédé de traitement d'image comprend : le positionnement d'une région cible dans une première image pour obtenir des informations de position de la région cible (102) ; l'effacement d'informations d'image d'une région prédéfinie dans la première image pour obtenir une seconde image après l'effacement (104), la taille de la seconde image étant la même que celle de la première image, et la région prédéfinie comprenant la région cible ; selon les informations de position, le découpage dans la seconde image d'une image de région correspondant à la région cible (106) ; et, selon l'image de région et la première image, la génération d'une image cible après le traitement (108).
PCT/CN2023/088947 2022-04-21 2023-04-18 Procédé de traitement d'image et appareil de traitement d'image, dispositif électronique et support de stockage lisible WO2023202570A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210420012.4 2022-04-21
CN202210420012.4A CN114792285A (zh) 2022-04-21 2022-04-21 图像处理方法和处理装置、电子设备和可读存储介质

Publications (1)

Publication Number Publication Date
WO2023202570A1 true WO2023202570A1 (fr) 2023-10-26

Family

ID=82461312

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/088947 WO2023202570A1 (fr) 2022-04-21 2023-04-18 Procédé de traitement d'image et appareil de traitement d'image, dispositif électronique et support de stockage lisible

Country Status (2)

Country Link
CN (1) CN114792285A (fr)
WO (1) WO2023202570A1 (fr)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114792285A (zh) * 2022-04-21 2022-07-26 维沃移动通信有限公司 图像处理方法和处理装置、电子设备和可读存储介质
CN115797783A (zh) * 2023-02-01 2023-03-14 北京有竹居网络技术有限公司 无障碍信息的生成方法、装置、电子设备和存储介质

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108805789A (zh) * 2018-05-29 2018-11-13 厦门市美亚柏科信息股份有限公司 一种基于对抗神经网络去除水印的方法、装置、设备及可读介质
CN109035147A (zh) * 2018-08-10 2018-12-18 Oppo广东移动通信有限公司 图像处理方法及装置、电子装置、存储介质和计算机设备
CN110163194A (zh) * 2019-05-08 2019-08-23 腾讯科技(深圳)有限公司 一种图像处理方法、装置及存储介质
CN111260757A (zh) * 2018-12-03 2020-06-09 马上消费金融股份有限公司 一种图像处理方法、装置及终端设备
WO2021129466A1 (fr) * 2019-12-26 2021-07-01 Oppo广东移动通信有限公司 Procédé de détection de filigrane, dispositif, terminal et support de stockage
CN113538273A (zh) * 2021-07-13 2021-10-22 荣耀终端有限公司 图像处理方法及图像处理装置
CN114049280A (zh) * 2021-11-25 2022-02-15 广州华多网络科技有限公司 图像消除修复方法及其装置、设备、介质、产品
CN114792285A (zh) * 2022-04-21 2022-07-26 维沃移动通信有限公司 图像处理方法和处理装置、电子设备和可读存储介质

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108805789A (zh) * 2018-05-29 2018-11-13 厦门市美亚柏科信息股份有限公司 一种基于对抗神经网络去除水印的方法、装置、设备及可读介质
CN109035147A (zh) * 2018-08-10 2018-12-18 Oppo广东移动通信有限公司 图像处理方法及装置、电子装置、存储介质和计算机设备
CN111260757A (zh) * 2018-12-03 2020-06-09 马上消费金融股份有限公司 一种图像处理方法、装置及终端设备
CN110163194A (zh) * 2019-05-08 2019-08-23 腾讯科技(深圳)有限公司 一种图像处理方法、装置及存储介质
WO2021129466A1 (fr) * 2019-12-26 2021-07-01 Oppo广东移动通信有限公司 Procédé de détection de filigrane, dispositif, terminal et support de stockage
CN113538273A (zh) * 2021-07-13 2021-10-22 荣耀终端有限公司 图像处理方法及图像处理装置
CN114049280A (zh) * 2021-11-25 2022-02-15 广州华多网络科技有限公司 图像消除修复方法及其装置、设备、介质、产品
CN114792285A (zh) * 2022-04-21 2022-07-26 维沃移动通信有限公司 图像处理方法和处理装置、电子设备和可读存储介质

Also Published As

Publication number Publication date
CN114792285A (zh) 2022-07-26

Similar Documents

Publication Publication Date Title
US10924676B2 (en) Real-time visual effects for a live camera view
WO2023202570A1 (fr) Procédé de traitement d'image et appareil de traitement d'image, dispositif électronique et support de stockage lisible
US10593023B2 (en) Deep-learning-based automatic skin retouching
US9020183B2 (en) Tagging images with labels
US10678410B2 (en) Browser-based image processing
US20100054601A1 (en) Image Tagging User Interface
US10346032B2 (en) Controlling display object on display screen
WO2019109900A1 (fr) Procédé et dispositif d'édition vidéo, et terminal mobile intelligent
CN110633116A (zh) 截图处理方法及装置
CN103327246A (zh) 一种多媒体拍摄处理方法、装置及智能终端
US11003467B2 (en) Visual history for content state changes
WO2016018682A1 (fr) Traitement d'image pour identifier un objet à insérer dans un document
WO2024055194A1 (fr) Procédé de génération d'objet virtuel, ainsi que procédé d'apprentissage de codec et appareil associé
CN115437736A (zh) 一种笔记记录方法和装置
KR101947553B1 (ko) 객체 기반 동영상 편집 장치 및 방법
CN115967854A (zh) 拍照方法、装置及电子设备
US11481940B2 (en) Structural facial modifications in images
CN107872730A (zh) 一种视频中的嵌入内容的获取方法和装置
CN114363521B (zh) 图像处理方法、装置及电子设备
US20230326094A1 (en) Integrating overlaid content into displayed data via graphics processing circuitry and processing circuitry using a computing memory and an operating system memory
CN116033094A (zh) 视频编辑方法及装置
CN115240260A (zh) 图像处理方法及其装置
CN116975337A (zh) 图像搜索方法、装置、电子设备和可读存储介质
CN118192864A (zh) 截屏方法、装置及电子设备
CN117149019A (zh) 文件编辑方法、终端、电子设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23791225

Country of ref document: EP

Kind code of ref document: A1