WO2023125374A1 - Procédé et appareil de traitement d'image, dispositif électronique et support de stockage - Google Patents

Procédé et appareil de traitement d'image, dispositif électronique et support de stockage Download PDF

Info

Publication number
WO2023125374A1
WO2023125374A1 PCT/CN2022/141815 CN2022141815W WO2023125374A1 WO 2023125374 A1 WO2023125374 A1 WO 2023125374A1 CN 2022141815 W CN2022141815 W CN 2022141815W WO 2023125374 A1 WO2023125374 A1 WO 2023125374A1
Authority
WO
WIPO (PCT)
Prior art keywords
style
image
target
model
trained
Prior art date
Application number
PCT/CN2022/141815
Other languages
English (en)
Chinese (zh)
Inventor
白须
Original Assignee
北京字跳网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京字跳网络技术有限公司 filed Critical 北京字跳网络技术有限公司
Publication of WO2023125374A1 publication Critical patent/WO2023125374A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks

Definitions

  • Embodiments of the present disclosure relate to the technical field of image processing, for example, to an image processing method, device, electronic equipment, and storage medium.
  • Image style transfer can be understood as rendering an image into an image with a specific artistic style.
  • Image style transfer images in the related art are mostly implemented by texture synthesis.
  • a style transfer model is trained to convert the image into a certain style based on the style transfer model.
  • the style transfer model in the related art cannot process images with different subject attributes, resulting in poor effect of the obtained style transfer image, which further affects the effect of user experience.
  • the present disclosure provides an image processing method, device, electronic equipment, and storage medium, so as to obtain a special effect image of a target style type and improve the display richness of image content.
  • an embodiment of the present disclosure provides an image processing method, the method including:
  • the target special effect image is displayed in the image display area.
  • an embodiment of the present disclosure further provides an image processing device, which includes:
  • the image acquisition module to be processed is configured to acquire the image to be processed including the target subject
  • the special effect image determination module is configured to input the image to be processed and the subject attributes of the target subject into the target style conversion model to obtain a target special effect image that converts the target subject into a target style type;
  • the image display module is configured to display the target special effect image in the image display area.
  • an embodiment of the present disclosure further provides an electronic device, and the electronic device includes:
  • processors one or more processors
  • storage means configured to store one or more programs
  • the one or more processors When the one or more programs are executed by the one or more processors, the one or more processors implement the image processing method according to any one of the embodiments of the present disclosure.
  • the embodiments of the present disclosure further provide a storage medium containing computer-executable instructions, and the computer-executable instructions are used to execute any one of the image processing methods described in the embodiments of the present disclosure when executed by a computer processor.
  • FIG. 1 is a schematic flowchart of an image processing method provided by an embodiment of the present disclosure
  • FIG. 2 is a schematic flowchart of an image processing method provided by another embodiment of the present disclosure.
  • FIG. 3 is a schematic flowchart of an image processing method provided by another embodiment of the present disclosure.
  • FIG. 4 is a schematic structural diagram of a target model to be trained provided by an embodiment of the present disclosure
  • FIG. 5 is a schematic flowchart of an image processing method provided by another embodiment of the present disclosure.
  • FIG. 6 is a schematic structural diagram of an image processing device provided by an embodiment of the present disclosure.
  • FIG. 7 is a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure.
  • the term “comprise” and its variations are open-ended, ie “including but not limited to”.
  • the term “based on” is “based at least in part on”.
  • the term “one embodiment” means “at least one embodiment”; the term “another embodiment” means “at least one further embodiment”; the term “some embodiments” means “at least some embodiments.” Relevant definitions of other terms will be given in the description below.
  • the disclosed technical solution can be applied to any scene of image style conversion, for example, converting a captured static image into an image of a certain theme style, the theme style can be Japanese style, Korean style, or designed by a designer Any theme style. Apply it in the scene of special effects video shooting, for example, convert a certain user in the captured screen or all users in the entire screen into a video with a certain theme style.
  • the style type may match the style of the makeup, and at the same time, the screen to which the user belongs may also be converted to match the style of the makeup.
  • users in each video frame to be processed can be displayed according to a corresponding style theme, or the entire video frame can be converted into a certain theme style.
  • Fig. 1 is a schematic flow chart of an image processing method provided by an embodiment of the present disclosure.
  • the embodiment of the present disclosure is applicable to the situation of converting an image frame into a target style type in any image display scene supported by the Internet.
  • the method can be executed by an image processing device, and the device can be realized in the form of software and/or hardware, for example, realized by electronic equipment, and the electronic equipment can be a mobile terminal, a PC terminal or a server, etc.
  • the scene of arbitrary image display is usually implemented by the cooperation of the client and the server.
  • the method provided in this embodiment can be executed by the server, the client, or the cooperation of the client and the server.
  • the method includes:
  • the device for executing the image processing method provided by the embodiments of the present disclosure may be integrated into application software supporting image processing functions, and the software may be installed in electronic equipment, for example, the electronic equipment may be a mobile terminal or a PC terminal, etc.
  • the application software may be a type of software for image/video processing, and its specific application software will not be described here one by one, as long as the image/video processing can be realized. It can also be a specially developed application program to realize the addition and display of special effects in the software, or it can be integrated in the corresponding page, and the user can realize the special effect addition process through the integrated page in the PC terminal.
  • the image to be processed may be an image collected based on the application software.
  • the image including the target subject may be captured in real time based on the application software.
  • the associated video frame or image processing is the style type consistent with the target style transfer type. It is also possible to set a corresponding special effect based on the style type, and after detecting that the user triggers the special effect, all the captured images may be converted into the corresponding style type.
  • the scene of image shooting or video shooting there may be multiple subjects in the captured image.
  • all users captured in the frame may be used as target subjects. It is also possible to mark which user or multiple users are the target subject before adding the special effect, and correspondingly, when the image to be processed is collected and determined to include the target subject in the image to be processed, it is processed.
  • the style theme conversion control can be triggered.
  • the image to be processed can be collected based on the camera device deployed on the terminal device.
  • the target subject may or may not be included in the image to be processed.
  • An image randomly obtained from a webpage may also be used as the image to be processed.
  • the image to be processed can be converted into a target special effect image consistent with the target style type.
  • the image to be processed is an image captured randomly or downloaded, and may also be an image captured in real time.
  • the target style conversion model is pre-trained and used to convert the image to be processed into a model of the corresponding style type.
  • the target style transfer model can be a GAN network based on the res structure.
  • the target special effect image is an image obtained after being processed by the target style transfer model.
  • the target subject in the target special effect image may be a target theme style, or an image obtained by converting the entire image to be processed into a corresponding target theme style type.
  • the image style output by the target style conversion model is consistent with the style type of the training samples used when training the target style conversion model. For example, if the style type of the style image in the training sample is style A, then the target style type is style A, and the target style conversion model outputs an image of style A; if the style type of the style image in the training sample is style B, then, The output of the target style transfer model is a B-style image. That is, the style type output by the target style transfer model is consistent with the style type used during model training.
  • the subject attribute can be the gender attribute or style type attribute of the target subject.
  • the gender attribute can be male or female
  • the style type attribute can be a preset style type.
  • the target style conversion model wants to achieve Output images of a certain style type under multiple style types, and you can define the style type and gender type corresponding to different tags in the alpha channel.
  • the image to be processed may be processed into an image corresponding to a corresponding style type according to the gender attribute and/or style attribute of the image to be processed.
  • the reason for defining the subject attributes is to avoid the need to train a model suitable for different genders in related technologies, which takes up memory, and realize that after the image to be processed is acquired, the subject attribute of the target subject in the image to be processed can be identified , and then input the subject attribute as information of the alpha channel into the target style transfer model, so as to perform style transfer processing on the image content of different gender attributes in different images to be processed based on a target style transfer model.
  • you want to implement multiple style types you need to train models of different style types, and there are cases where a model cannot implement multiple style type conversions.
  • the input of the model to be trained must not only include the image that needs to be converted to the image style type, but also need to edit the label value of the alpha channel, so that you can execute this The conversion model of the target special effect diagram of the technical scheme.
  • the subject attribute of the target object in the image to be processed may be determined based on the subject attribute identification module deployed in the terminal. Input the subject attributes and the image to be processed into the target style conversion model to obtain the target special effect image that converts the target subject into the target style type, or converts the entire image to be processed into the target special effect image of the target style type.
  • the target style transfer model before inputting the image to be processed into the target style transfer model, it further includes: scaling the image to be processed into an image of a certain size, and the size may be 384*384 pixels.
  • the above-mentioned processing method can be performed on the obtained samples.
  • the image display area can be understood as an area where the target special effect image is displayed.
  • the target image may be displayed on the display interface.
  • the image display area can be divided into two areas according to the principle of left and right, or the principle of up and down.
  • One area shows the target special effect image
  • the other area shows the original acquired image to be processed.
  • target video is to be formed, special effect processing can be performed on the image to be processed collected in real time to obtain the target special effect image.
  • a plurality of target special effect images are sequentially spliced to obtain a target video.
  • the target style type includes Japanese style, Korean style, ancient costume style, comic style, or at least one of multiple preset style types to be selected.
  • Period costume styles can include styles from any dynasty.
  • the subject attribute of the target subject in the image to be processed can be determined, and the subject attribute and the image to be processed can be used as the input of the pre-trained target style conversion model Parameters, to obtain the target special effect image that converts the target subject in the image to be processed into the target style type, and the image to be processed can be converted into the style type, thereby improving the user experience.
  • Fig. 2 is a schematic flow chart of an image processing method provided by another embodiment of the present disclosure.
  • a target style conversion model can be trained to obtain the target style conversion model.
  • Described method as shown in Figure 2 comprises:
  • the third original image including facial information may be collected in a real environment, or the third original image including facial information downloaded from a webpage, or may be a randomly generated facial image based on a facial image generation model.
  • the image to be used is an image after processing the style of the third original image, for example, it may be an image designed by a designer and corresponding to a certain style type.
  • the third-style image is an image of one or more styles that are hand-painted from the original image. Style types can be various. Various style types are used as the style types to be selected.
  • the image obtained after converting the third original image into a certain style is used as the image to be used. That is, the image to be used is an image of a certain style.
  • image cropping processing can be used.
  • the cropping process can be understood as aligning the nose tip and the center of the eyes as reference points, or aligning the center of the chin and the center of the eyes as reference points. Adjust the display ratio of the facial image in the display interface, thereby expanding the cropping range of the image, so that the entire head (including hair) and facial images are displayed on the display interface, and at the same time, images with corresponding background information are also included.
  • the image generation model to be trained may be a styleganv2 model, and the model parameters in the model are default values at this time.
  • the third output image is an image of a certain style randomly generated based on the image generation model to be trained. That is, the style type of the third output image is indeterminate.
  • Gaussian noise is randomly sampled Gaussian noise.
  • the image generation model to be trained processes the Gaussian noise to obtain third output images of different styles.
  • the image generation model to be trained is trained to obtain the target image generation model, which can generate sample data of different style types, so that a model capable of realizing conversion of different style types can be obtained based on the sample data training.
  • S230 Process the third output image and the third style image based on the first discriminator, determine a loss value, and correct model parameters in the image generation model to be trained based on the loss value.
  • the input of the first discriminator is the third output image and the third style image.
  • the first discriminator is configured to determine a loss value between the third output image and the third style image.
  • the model parameters in the image generation model to be trained can be corrected according to the loss value.
  • the target image generation model is an image generation model obtained through final training. Repeat the above steps based on multiple training samples until the convergence of the loss function is detected, and use the image generation model obtained in this case as the target image generation model.
  • S250 Process Gaussian noise based on the target image generation model to obtain the second style image.
  • the randomly sampled Gaussian noise can be processed based on the target image generation model, so as to obtain the second style image for training the first style transfer model.
  • the number of second style images can be as large as possible.
  • style types of the second style images may be the same or different.
  • the style types of the second style images can be as many and rich as possible.
  • the second style image can be processed on the basis of retaining the second style image.
  • the expression editing model may be a model for adding facial expressions to the target subject in the second style image.
  • the facial expression can be an open mouth, and the size of the mouth opening is different; expressions such as smiling and laughing, and the specific expressions are not specifically limited in this embodiment.
  • the second style image can be input into the expression editing model on the basis of retaining the second style image, so as to add expressions to users in the second style image, and obtain expression A second style image with a change in content. Based on the original second-style image and the second-style image with the expression added, training samples for training the first-style model can be obtained.
  • the target image generation model can be obtained based on the training, and the Gaussian noise can be processed to obtain the second style image under various style types; the expression can be added to the second style image based on the expression editing model to obtain the expression The second style image with changed content, so as to obtain the first style conversion model based on the second style image training, and then determine the training samples for training the target style conversion model based on the first style conversion model, so as to obtain the target style conversion model based on the training samples , to avoid the uneven quality of samples in related technologies, resulting in the inability to effectively perform style conversion for the original image, and to perform style conversion on the image to be processed, thereby improving the matching degree between the converted image and the user, Improved user experience.
  • Fig. 3 is a schematic flow chart of an image processing method provided by another embodiment of the present disclosure.
  • the second style image is obtained based on the target image generation model, it can be based on the second style image and the real environment
  • the second original image collected in the first style transfer model is trained to obtain the first style transfer model.
  • Described method as shown in Figure 3 comprises:
  • S310 Determine at least one second style image.
  • S320 Construct a target model to be trained including a style processing model to be trained, a target discriminator, and a target style comparer.
  • target discriminator and target style comparer are pre-trained models.
  • the target model to be trained includes a style processing model to be trained, a target discriminator and a target style comparer.
  • the output of the style processing model to be trained is the input of the target discriminator and the target style comparer respectively, based on the output of the target discriminator and the target style comparer, the model parameters in the target model to be trained are corrected to obtain the target to be used Model.
  • the style processing model to be trained includes: a style feature extractor to be trained, a content feature extractor to be trained, a feature fusion unit to be trained, and a compiler to be trained.
  • the style model to be trained is a GAN (Generative Adversarial Network, Generative Adversarial Network) model based on the starganv2 structure. This model is primarily set up to generate batches of unpaired data, i.e. second style images.
  • GAN Geneative Adversarial Network, Generative Adversarial Network
  • the input of the target model to be trained is two images, one image is the image collected in the real environment, that is, the second original image; the other image is an image of a certain style generated based on the image generation model.
  • the style types of the multiple second style images may be the same or different.
  • the style type corresponding to the second style image may be used as the style type to be selected.
  • the target model to be used is a model obtained through training based on the second style image and the second original image.
  • the specific training method of the target model to be trained based on the second original image and the second style image may be as follows: by combining and processing the second style image and the second original image, multiple second training samples are obtained; Wherein, the second training sample includes a second original image and a second style image; for the second training sample, the content splicing feature of the second original image is obtained based on the content feature extractor to be trained, and based on the style to be trained
  • the feature extractor obtains the style mosaic features of the second style image, performs fusion processing on the content mosaic features and style mosaic features based on the feature fusion model to be trained, obtains the fusion features, and inputs the fusion features into the compiler to be trained to obtain the actual output image; Input the actual output image and the second style image into the target discriminator to determine the first loss value; input the actual output image and the second style image into the target style comparison device to determine the style loss value; based on the first loss value and the style loss value, modify the model parameters in the style processing model to be trained
  • the second style image and the second original image may be randomly combined to obtain a plurality of second training samples.
  • Each training sample includes a second style image and a second original image.
  • the processing method for each second training sample is the same, and the processing of one of the training samples is taken as an example for introduction.
  • the target model to be trained For example, input the second original image and the second style image into the target model to be trained.
  • the content mosaic feature and style mosaic feature are fused to obtain the fusion feature.
  • Compile and process the fusion features based on the compiler to be trained to obtain the actual output image.
  • the ideal actual output image should include the image content of the second original image and the style features of the second style image.
  • the model parameters in the style processing model to be trained are default values, there are some differences between the obtained actual image and the ideally obtained image.
  • the target discriminator and target style in the target model to be trained Comparator for processing. Input the actual output image and the second style image corresponding to the actual output image into the target discriminator to obtain the first loss value; at the same time, input the actual output image and the second style image corresponding to the actual output image into For the target style comparator, determine the style loss value. Based on the first loss value and the style loss value, model parameters in the style processing model to be trained can be corrected. Take the convergence of the loss function in the style processing model to be trained as the training target, and obtain the target model to be used.
  • the second original image is A
  • the second style image is B.
  • the image content of the original image A is obtained based on the content feature extractor to be trained, and at the same time, the style features of the style image B are obtained based on the style feature extractor to be trained; the image content and style features are spliced based on the feature fusion model to be trained, and the actual Output the fused features corresponding to image C. Compile and process the fusion feature based on the compiler to be trained to obtain the actual output image C.
  • the model parameters in the style processing model to be trained in the target model to be trained are corrected until the loss function in the style processing model to be trained converges, and the target model to be used is obtained.
  • the target discriminator and the target style comparer in the target to-be-used model can be eliminated, that is, only the trained style processing model to be trained is retained, so as to obtain the to-be-used style model.
  • an original image and an image of a preferred style type can be randomly input into the style conversion model to be used to obtain a style consistent with a certain style type. image, at this point, the content in the image is consistent with the content in the original image.
  • S350 Determine a reference image of the target style type, and determine the first style conversion model based on the reference image and the style model to be used.
  • the target style type is the style type finally selected according to the preference of the user.
  • the reference image is an image consistent with the target style type.
  • the reference image can be bound with the style model to be used, so that after an original image is input, the style features of the reference image are extracted based on the style model to be used, and the image content and style features of the first original image are fused to obtain Images that match the style type.
  • the model after binding the reference image and the style model to be used is used as the first style conversion model.
  • the model structure in the first style conversion model is relatively complex, when it is deployed on a mobile terminal device, there may be insufficient computing power. Based on this, the first style conversion model can be deployed on the server , so that the server can perform style conversion processing on the image.
  • the technical solutions of the embodiments of the present disclosure can generate second style images of various styles based on the target image generation model, and perform training on the target model to be trained based on the second style image and the original image to obtain the target model to be used, and
  • the target model to be used and the pre-selected image of a certain style type are packaged and processed to obtain the first style conversion model.
  • the first style conversion model can convert the input original image into a target special effect image consistent with the packaged style type, so that various collected images can be processed, and the convenience of sample acquisition and image content processing is improved. sex.
  • Fig. 5 is a schematic flowchart of an image processing method provided by another embodiment of the present disclosure.
  • the first training sample can be constructed based on the first style conversion model , and then process the to-be-trained style conversion model based on the first training sample to obtain a target style conversion model.
  • a target style conversion model For an example implementation, please refer to the detailed description of the technical solution. Wherein, technical terms that are the same as or corresponding to those in the foregoing embodiments will not be repeated here.
  • the method includes:
  • the first style conversion model cannot be directly deployed on terminal devices due to its high requirements on computing power. Based on this, corresponding training samples can be constructed based on the first style conversion model. The trained target style transfer model can be deployed on the terminal device.
  • the first original image may be an image collected by a camera device in a real environment, or an image generated based on a certain image generation model. In order to improve the accuracy of model training, as many first original images as possible can be obtained.
  • the first original image may or may not have a corresponding style.
  • brightness changes may be performed on the first original image, for example, brightness adjustment may be performed on the entire image, or brightness changes may be performed only on the face image in the original image. This allows for a random brightness correction to make the lighting conditions of the trained mesh more random.
  • the face image pixels in the image can be extracted, and the brightness of the face pixels can be brightened.
  • the style type of the first style image is consistent with the target style type.
  • the first style image is generated based on the pre-trained first style transfer model.
  • the style type corresponding to the first style conversion model is bound to the pre-bound image style type, that is, according to the user's preference for multiple style images, one of the style types of images can be selected and the pre-trained target model Binding, so as to obtain the first style conversion model.
  • the first style conversion model can be deployed on the server, so that after receiving the image to be processed, the bound image and the image to be processed can be processed based on the first style conversion model to obtain the target special effect image of the target style type, and display on the client side.
  • the calculation amount of this model is very large, and it is not suitable for deployment on terminal devices. Therefore, corresponding training samples can be obtained based on the first style conversion model, and then the above-mentioned target style conversion model can be obtained by training based on the corresponding training samples.
  • the collected first original image is processed based on the pre-trained first style conversion model to obtain the first style image consistent with the style type of the first style conversion model. Based on this method, multiple training samples are obtained.
  • obtaining the corresponding first style image based on the first original image may be: performing content extraction on the current first original image based on the content feature extractor to obtain image content features; based on the style feature The extractor extracts style features from a preset reference style image that is consistent with the target style type to obtain image style features; based on the feature fuser, the image content features and the image style features are fused, Obtaining features to be compiled; obtaining a first style image corresponding to the current first original image based on the processing of the features to be compiled by the compiler.
  • the content of the first original image is obtained based on the content feature extractor
  • the style feature of the reference image is obtained based on the style feature extractor.
  • the image content and style features are fused to obtain the fused features.
  • the first original image may be subjected to style conversion processing to obtain a first style image of the first original image under the target style type. Based on the first original image and the corresponding first style image, a first training sample is determined.
  • the model parameters in the style transfer model to be trained are default values.
  • the object attribute may be the gender attribute of the target subject in the original image.
  • the style transfer model to be trained may be trained based on a plurality of first training samples to obtain a target style transfer model.
  • processing manner for each first training sample is the same, and the processing of one of the training samples may be used as an example for introduction.
  • the first training sample includes a corresponding original image and a first style image corresponding to the first original image.
  • the gender attribute of the target object in the first original image can be determined based on the corresponding algorithm, and the gender attribute and the first original image are input into the style conversion model to be trained, which will be compared with the second
  • the first style image corresponding to the original image is used as the output of the style conversion model to be trained, and the style conversion model to be trained is trained to obtain the target style conversion model.
  • the size of the image needs to be scaled to a certain size, so as to improve the efficiency of the model for image processing.
  • the label whose gender attribute is female is set to 0
  • the label whose gender attribute is male is set to 1
  • four channels are constructed to train a target style conversion model, which can be used for different gender attributes The image is processed.
  • the technical solution of the embodiment of the present disclosure can train the target style conversion model deployed on the terminal device, so that when the image to be processed is collected, the style type conversion can be performed on the image based on the client, which improves the convenience of image processing.
  • FIG. 6 is a schematic structural diagram of an image processing device provided by an embodiment of the present disclosure.
  • the device includes: an image acquisition module 510 to be processed, a special effect image determination module 520 and an image display module 530 .
  • the to-be-processed image acquisition module 510 is set to acquire the to-be-processed image including the target subject;
  • the special effect image determination module 520 is set to input the subject attribute of the to-be-processed image and the target subject into the target style conversion model , to obtain the target special effect image converted from the target subject into the target style type;
  • the image display module 530 is configured to display the target special effect image in the image display area.
  • the device includes:
  • the first training sample acquisition module is configured to acquire a plurality of first training samples; wherein, the first training samples include a first original image and a first style image consistent with the target style type; the first The style image is generated by the first style transfer model;
  • the first training module is configured to use the first original image in the current first training sample and the object attribute of the object to be processed as the input of the style conversion model to be trained for the first training sample, and use the current first training sample
  • the first style image in the training sample is used as the output of the style model to be trained to obtain the target style conversion model through training; wherein the object attribute matches the subject attribute.
  • the first training sample acquisition module includes:
  • a first original image acquisition unit configured to acquire at least one first original image
  • the first style image acquisition unit is configured to stylize the current first original image based on the first style conversion model for the first original image, to obtain a first style image consistent with the target style type;
  • the first training sample acquisition unit is configured to determine the plurality of first training samples based on the first original image and the corresponding first style image.
  • the first style conversion model includes a style feature extractor, a content feature extractor, a feature fusion device and a compiler, and the first style image acquisition unit is set to:
  • the device includes:
  • the first model construction unit is configured to construct a target model to be trained including a style processing model to be trained, a target discriminator, and a target style comparer; wherein, the target discriminator and the target style comparer are pre-trained
  • the model determination unit to be used is configured to train the target model to be trained according to at least one second style image of at least one style type to be selected and at least one second original image to obtain the target model to be used ; Wherein, the second style image is determined based on the target image generation model; the to-be-used style model determining unit is configured to use the to-be-trained style processing model trained in the target to-be-used model as the to-be-used style model;
  • the first style model determination unit is configured to determine a reference image corresponding to the target style type, and determine the first style conversion model based on the reference image and the style model to be used; wherein, the The target style type is one of the at least one style type to be selected.
  • the style processing model to be trained includes: a style feature extractor to be trained, a content feature extractor to be trained, a feature fusion device to be trained, and a compiler to be trained, and the model determination unit to be used Set as:
  • a plurality of second training samples are obtained by randomly combining the second style image and the second original image; wherein, the second training samples include a second original image and a second style image; for the second training samples, obtaining the content splicing features of the second original image based on the content feature extractor to be trained, and obtaining the style splicing features of the second style image based on the style feature extractor to be trained, and obtaining the splicing features of the second style image based on the features to be trained
  • the fusion model fuses the content splicing feature and the style splicing feature to obtain the fusion feature, and inputs the fusion feature into the compiler to be trained to obtain an actual output image; combines the actual output image and the second Inputting the style image into the target discriminator to determine a first loss value, and inputting the actual output image and the second style image into the target style comparer to determine a style loss value; based on the The first loss value and the style loss value are used to modify the model parameters in the style processing model to be
  • the first style model determining unit is configured to determine a target style type from at least one style type to be selected; obtain a reference image consistent with the target style type, and The reference image is packaged with the style model to be used to obtain the first style conversion model.
  • the device also includes:
  • the third style image acquiring unit is configured to acquire a third original image in an image to be used under the style type to be selected, and cut the image to be used to obtain a third style image;
  • the third output image acquisition unit is configured to input Gaussian noise into the image generation model to be trained to obtain a third output image
  • a model parameter correction unit configured to process the third output image and the third style image based on the first discriminator, determine a loss value, and correct the model parameters in the image generation model to be trained based on the loss value ;
  • the target image generation model determination unit is configured to use the convergence of the loss function in the image generation model to be trained as the training target to obtain the target image generation model.
  • the device includes:
  • the second style image acquisition unit is configured to process Gaussian noise based on the target image generation model to obtain a second style image
  • the second style image updating unit is configured to add expressions to the second style image based on the expression editing model generated through pre-training, and update the second style image.
  • the target style type includes Japanese style, Korean style, ancient costume style, comic style or multiple preset style types to be selected.
  • the subject attribute of the target subject in the image to be processed can be determined, and the subject attribute and the image to be processed can be used as the input of the pre-trained target style conversion model Parameters, to obtain the target special effect image that converts the target subject in the image to be processed into the target style type, and the image to be processed can be converted into the style type, thereby improving the user experience.
  • the image processing device provided by the embodiment of the present disclosure can execute the image processing method provided by any embodiment of the present disclosure, and has corresponding functional modules and beneficial effects for executing the method.
  • FIG. 7 is a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure.
  • the terminal equipment in the embodiment of the present disclosure may include but not limited to such as mobile phone, notebook computer, digital broadcast receiver, PDA (personal digital assistant), PAD (tablet computer), PMP (portable multimedia player), vehicle terminal (such as mobile terminals such as car navigation terminals) and fixed terminals such as digital TVs, desktop computers and the like.
  • the electronic device shown in FIG. 7 is only an example, and should not limit the functions and application scope of the embodiments of the present disclosure.
  • an electronic device 500 may include a processing device (such as a central processing unit, a graphics processing unit, etc.) 501, which may be randomly accessed according to a program stored in a read-only memory (ROM) 502 or loaded from a storage device 508.
  • ROM read-only memory
  • RAM random access memory
  • various appropriate actions and processes are executed by programs in the memory (RAM) 503 .
  • RAM random access memory
  • various programs and data necessary for the operation of the electronic device 500 are also stored.
  • the processing device 501, ROM 502, and RAM 503 are connected to each other through a bus 504.
  • An edit/output (I/O) interface 505 is also connected to the bus 504 .
  • the following devices can be connected to the I/O interface 505: input devices 506 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.; including, for example, a liquid crystal display (LCD), speaker, vibration an output device 507 such as a computer; a storage device 508 including, for example, a magnetic tape, a hard disk, etc.; and a communication device 509.
  • the communication means 509 may allow the electronic device 500 to perform wireless or wired communication with other devices to exchange data. While FIG. 7 shows electronic device 500 having various means, it should be understood that implementing or having all of the means shown is not a requirement. More or fewer means may alternatively be implemented or provided.
  • embodiments of the present disclosure include a computer program product, which includes a computer program carried on a non-transitory computer readable medium, where the computer program includes program code for executing the method shown in the flowchart.
  • the computer program may be downloaded and installed from a network via communication means 509, or from storage means 508, or from ROM 502.
  • the processing device 501 When the computer program is executed by the processing device 501, the above-mentioned functions defined in the methods of the embodiments of the present disclosure are performed.
  • the electronic device provided by the embodiment of the present disclosure belongs to the same idea as the image processing method provided by the above embodiment, and the technical details not described in detail in this embodiment can be referred to the above embodiment, and this embodiment has the same benefits as the above embodiment Effect.
  • An embodiment of the present disclosure provides a computer storage medium, on which a computer program is stored, and when the program is executed by a processor, the image processing method provided in the foregoing embodiments is implemented.
  • the computer-readable medium mentioned above in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium or any combination of the two.
  • a computer readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or device, or any combination thereof. More specific examples of computer-readable storage media may include, but are not limited to, electrical connections with one or more wires, portable computer diskettes, hard disks, random access memory (RAM), read-only memory (ROM), erasable Programmable read-only memory (EPROM or flash memory), optical fiber, portable compact disk read-only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
  • a computer-readable storage medium may be any tangible medium that contains or stores a program that can be used by or in conjunction with an instruction execution system, apparatus, or device.
  • a computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave carrying computer-readable program code therein. Such propagated data signals may take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • a computer-readable signal medium may also be any computer-readable medium other than a computer-readable storage medium, which can transmit, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device .
  • Program code embodied on a computer readable medium may be transmitted by any appropriate medium, including but not limited to wires, optical cables, RF (radio frequency), etc., or any suitable combination of the above.
  • the client and the server can communicate using any currently known or future network protocols such as Hypertext Transfer Protocol (HyperText Transfer Protocol, HTTP), and can communicate with digital data in any form or medium
  • HTTP Hypertext Transfer Protocol
  • Examples of communication networks include local area networks ("LANs”), wide area networks ("WANs”), internetworks (e.g., the Internet), and peer-to-peer networks (e.g., ad hoc peer-to-peer networks), as well as any currently known or future developed network of.
  • the above-mentioned computer-readable medium may be included in the above-mentioned electronic device, or may exist independently without being incorporated into the electronic device.
  • the above-mentioned computer-readable medium carries one or more programs, and when the above-mentioned one or more programs are executed by the electronic device, the electronic device:
  • the target special effect image is displayed in the image display area.
  • Computer program code for carrying out operations of the present disclosure may be written in one or more programming languages, or combinations thereof, including but not limited to object-oriented programming languages—such as Java, Smalltalk, C++, and Includes conventional procedural programming languages - such as the "C" language or similar programming languages.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer can be connected to the user computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computer (such as through an Internet service provider). Internet connection).
  • LAN local area network
  • WAN wide area network
  • Internet service provider such as AT&T, MCI, Sprint, EarthLink, MSN, GTE, etc.
  • each block in a flowchart or block diagram may represent a module, program segment, or portion of code that contains one or more logical functions for implementing specified executable instructions.
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or they may sometimes be executed in the reverse order, depending upon the functionality involved.
  • each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented by a dedicated hardware-based system that performs the specified functions or operations , or may be implemented by a combination of dedicated hardware and computer instructions.
  • the units involved in the embodiments described in the present disclosure may be implemented by software or by hardware. Wherein, the name of the unit does not constitute a limitation of the unit itself under certain circumstances, for example, the first obtaining unit may also be described as "a unit for obtaining at least two Internet Protocol addresses".
  • FPGAs Field Programmable Gate Arrays
  • ASICs Application Specific Integrated Circuits
  • ASSPs Application Specific Standard Products
  • SOCs System on Chips
  • CPLD Complex Programmable Logical device
  • a machine-readable medium may be a tangible medium that may contain or store a program for use by or in conjunction with an instruction execution system, apparatus, or device.
  • a machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
  • a machine-readable medium may include, but is not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, apparatus, or devices, or any suitable combination of the foregoing.
  • machine-readable storage media would include one or more wire-based electrical connections, portable computer discs, hard drives, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM or flash memory), optical fiber, compact disk read only memory (CD-ROM), optical storage, magnetic storage, or any suitable combination of the foregoing.
  • RAM random access memory
  • ROM read only memory
  • EPROM or flash memory erasable programmable read only memory
  • CD-ROM compact disk read only memory
  • magnetic storage or any suitable combination of the foregoing.
  • Example 1 provides an image processing method, the method including:
  • the target special effect image is displayed in the image display area.
  • Example 2 provides an image processing method, the method further includes:
  • the first training samples include a first original image and a first style image consistent with the target style type; the first style image is generated by the first style conversion model of;
  • the first original image in the current first training sample and the object attribute of the object to be processed are used as the input of the style conversion model to be trained, and the first style in the current first training sample is An image, as the output of the style model to be trained, is trained to obtain the target style conversion model;
  • the object attribute matches the subject attribute.
  • Example 3 provides an image processing method, wherein,
  • the acquisition of multiple first training samples includes:
  • For the first original image stylize the current first original image based on the first style conversion model to obtain a first style image consistent with the target style type;
  • the plurality of first training samples is determined based on the first original image and the corresponding first style image.
  • Example 4 provides an image processing method, wherein the first style conversion model includes a style feature extractor, a content feature extractor, a feature fusion unit, and a compiler , the stylization processing of the current first original image based on the first style conversion model to obtain a first style image consistent with the target style type, including:
  • Example 5 provides an image processing method, the method further includes:
  • a target model to be trained including a style processing model to be trained, a target discriminator, and a target style comparer; wherein, the target discriminator and the target style comparer are pre-trained;
  • the constructed target model to be trained is trained to obtain the target model to be used; wherein the second style image is determined based on the target image generation model;
  • style processing model to be trained trained in the target model to be used as the style model to be used;
  • the target style type is the at least one One of the style types to be selected.
  • Example 6 provides an image processing method, wherein the style processing model to be trained includes: a style feature extractor to be trained, a content feature extractor to be trained, The training feature fusion device and the compiler to be trained, the target model to be trained is trained according to the multiple second style images of at least one style type to be selected, and at least one second original image, and the target to be trained model is obtained.
  • Use models including:
  • a plurality of second training samples are obtained by randomly combining the second style image and the second original image; wherein, the second training samples include a second original image and a second style image;
  • the content splicing feature of the second original image is obtained based on the content feature extractor to be trained, and the style splicing feature of the second style image is obtained based on the style feature extractor to be trained, based on the
  • the feature fusion model to be trained fuses the content splicing features and the style splicing features to obtain the fusion features, and inputs the fusion features into the compiler to be trained to obtain the actual output image;
  • Example 7 provides an image processing method, wherein the determining a reference image corresponding to the target style type is based on the reference image and the determined Describe the style model to be used, and determine the first style conversion model, including:
  • Example 8 provides an image processing method, the method further includes:
  • Gaussian noise is input in the image generation model to be trained, obtains the 3rd output image
  • Example 9 provides an image processing method, the method further includes:
  • Example 10 provides an image processing method, wherein the target style type includes Japanese style, Korean style, ancient costume style, comic style, or a variety of preset Select a style type.
  • Example Eleven provides an image processing device, which includes:
  • the image acquisition module to be processed is configured to acquire the image to be processed including the target subject
  • the special effect image determination module is configured to input the image to be processed and the subject attributes of the target subject into the target style conversion model to obtain a target special effect image that converts the target subject into a target style type;
  • the image display module is configured to display the target special effect image in the image display area.
  • the technical solution of the embodiment of the present disclosure realizes the conversion of the image to be processed into an image with the target theme style, which improves the richness of the image display content, the novelty of the theme style, and the adaptability between the theme style and the user.

Abstract

La présente invention concerne un procédé et un appareil de traitement d'image, un dispositif électronique et un support de stockage. Le procédé consiste à : acquérir une image à traiter comprenant un sujet cible (S110) ; entrer l'image à traiter et un attribut de sujet du sujet cible dans un modèle de conversion de style cible pour obtenir une image d'effet spécial cible dans laquelle le sujet cible a été converti en un type de style cible (S120) ; et afficher l'image d'effet spécial cible dans une zone d'affichage d'image (S130).
PCT/CN2022/141815 2021-12-29 2022-12-26 Procédé et appareil de traitement d'image, dispositif électronique et support de stockage WO2023125374A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202111641158.3 2021-12-29
CN202111641158.3A CN114331820A (zh) 2021-12-29 2021-12-29 图像处理方法、装置、电子设备及存储介质

Publications (1)

Publication Number Publication Date
WO2023125374A1 true WO2023125374A1 (fr) 2023-07-06

Family

ID=81017138

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/141815 WO2023125374A1 (fr) 2021-12-29 2022-12-26 Procédé et appareil de traitement d'image, dispositif électronique et support de stockage

Country Status (2)

Country Link
CN (1) CN114331820A (fr)
WO (1) WO2023125374A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117611434A (zh) * 2024-01-17 2024-02-27 腾讯科技(深圳)有限公司 模型训练方法、图像风格转换方法、装置及电子设备

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114331820A (zh) * 2021-12-29 2022-04-12 北京字跳网络技术有限公司 图像处理方法、装置、电子设备及存储介质
CN114866706A (zh) * 2022-06-01 2022-08-05 北京字跳网络技术有限公司 图像处理方法、装置、电子设备及存储介质
CN115145442A (zh) * 2022-06-07 2022-10-04 杭州海康汽车软件有限公司 一种环境图像的显示方法、装置、车载终端及存储介质
CN114926326A (zh) * 2022-06-28 2022-08-19 北京字跳网络技术有限公司 图像处理方法、装置、电子设备及存储介质
CN116126182A (zh) * 2022-09-08 2023-05-16 北京字跳网络技术有限公司 特效处理方法、装置、电子设备及存储介质
CN115249221A (zh) * 2022-09-23 2022-10-28 阿里巴巴(中国)有限公司 图像处理方法、装置和云设备

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109146825A (zh) * 2018-10-12 2019-01-04 深圳美图创新科技有限公司 摄影风格转换方法、装置及可读存储介质
CN110598781A (zh) * 2019-09-05 2019-12-20 Oppo广东移动通信有限公司 图像处理方法、装置、电子设备及存储介质
US20200134797A1 (en) * 2018-10-31 2020-04-30 Boe Technology Group Co., Ltd. Image style conversion method, apparatus and device
CN111402112A (zh) * 2020-03-09 2020-07-10 北京字节跳动网络技术有限公司 图像处理方法、装置、电子设备及计算机可读介质
CN111784567A (zh) * 2020-07-03 2020-10-16 北京字节跳动网络技术有限公司 用于转换图像的方法、装置、电子设备和计算机可读介质
CN113780326A (zh) * 2021-03-02 2021-12-10 北京沃东天骏信息技术有限公司 一种图像处理方法、装置、存储介质及电子设备
CN114331820A (zh) * 2021-12-29 2022-04-12 北京字跳网络技术有限公司 图像处理方法、装置、电子设备及存储介质

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108205813B (zh) * 2016-12-16 2022-06-03 微软技术许可有限责任公司 基于学习网络的图像风格化
US10891723B1 (en) * 2017-09-29 2021-01-12 Snap Inc. Realistic neural network based image style transfer
CN108961198B (zh) * 2018-07-09 2021-06-08 中国海洋大学 一种多风格生成对抗网络的水下图像合成方法及其应用
CN110830706A (zh) * 2018-08-08 2020-02-21 Oppo广东移动通信有限公司 图像处理方法和装置、存储介质、电子设备
CN111583097A (zh) * 2019-02-18 2020-08-25 北京三星通信技术研究有限公司 图像处理方法、装置、电子设备及计算机可读存储介质
CN109949213B (zh) * 2019-03-15 2023-06-16 百度在线网络技术(北京)有限公司 用于生成图像的方法和装置
US11625576B2 (en) * 2019-11-15 2023-04-11 Shanghai United Imaging Intelligence Co., Ltd. Systems and methods for image style transformation
CN112669308A (zh) * 2021-01-06 2021-04-16 携程旅游信息技术(上海)有限公司 基于风格迁移的图像生成方法、系统、设备及存储介质
CN113705302A (zh) * 2021-03-17 2021-11-26 腾讯科技(深圳)有限公司 图像生成模型的训练方法、装置、计算机设备及存储介质
CN113850712A (zh) * 2021-09-03 2021-12-28 北京达佳互联信息技术有限公司 图像风格转换模型的训练方法、图像风格转换方法及装置

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109146825A (zh) * 2018-10-12 2019-01-04 深圳美图创新科技有限公司 摄影风格转换方法、装置及可读存储介质
US20200134797A1 (en) * 2018-10-31 2020-04-30 Boe Technology Group Co., Ltd. Image style conversion method, apparatus and device
CN110598781A (zh) * 2019-09-05 2019-12-20 Oppo广东移动通信有限公司 图像处理方法、装置、电子设备及存储介质
CN111402112A (zh) * 2020-03-09 2020-07-10 北京字节跳动网络技术有限公司 图像处理方法、装置、电子设备及计算机可读介质
CN111784567A (zh) * 2020-07-03 2020-10-16 北京字节跳动网络技术有限公司 用于转换图像的方法、装置、电子设备和计算机可读介质
CN113780326A (zh) * 2021-03-02 2021-12-10 北京沃东天骏信息技术有限公司 一种图像处理方法、装置、存储介质及电子设备
CN114331820A (zh) * 2021-12-29 2022-04-12 北京字跳网络技术有限公司 图像处理方法、装置、电子设备及存储介质

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117611434A (zh) * 2024-01-17 2024-02-27 腾讯科技(深圳)有限公司 模型训练方法、图像风格转换方法、装置及电子设备
CN117611434B (zh) * 2024-01-17 2024-05-07 腾讯科技(深圳)有限公司 模型训练方法、图像风格转换方法、装置及电子设备

Also Published As

Publication number Publication date
CN114331820A (zh) 2022-04-12

Similar Documents

Publication Publication Date Title
WO2023125374A1 (fr) Procédé et appareil de traitement d'image, dispositif électronique et support de stockage
US20220239882A1 (en) Interactive information processing method, device and medium
WO2023125361A1 (fr) Procédé et appareil de génération de caractère, dispositif électronique et support de stockage
CN111669502B (zh) 目标对象显示方法、装置及电子设备
CN111399729A (zh) 图像的绘制方法、装置、可读介质和电子设备
US20230421716A1 (en) Video processing method and apparatus, electronic device and storage medium
WO2023125379A1 (fr) Procédé et appareil de génération de caractère, dispositif électronique et support de stockage
WO2019227429A1 (fr) Procédé, dispositif, appareil, terminal, serveur associes à la génération d' un contenu multimédia
WO2023109842A1 (fr) Procédé et appareil de présentation d'image, et dispositif électronique et support de stockage
WO2023093897A1 (fr) Procédé et appareil de traitement d'image, dispositif électronique et support de stockage
WO2023083142A1 (fr) Procédé et appareil de segmentation de phrases, support de stockage et dispositif électronique
WO2023138560A1 (fr) Procédé et appareil de génération d'images stylisées, dispositif électronique et support de stockage
WO2022142875A1 (fr) Procédé et appareil de traitement d'image, dispositif électronique et support de stockage
WO2023045710A1 (fr) Procédés et appareils d'affichage et de correspondance multimédia, dispositif et support
WO2023040749A1 (fr) Procédé et appareil de traitement d'image, dispositif électronique et support de stockage
WO2022171024A1 (fr) Procédé et appareil d'affichage d'images, dispositif et support
US20240119082A1 (en) Method, apparatus, device, readable storage medium and product for media content processing
US11818491B2 (en) Image special effect configuration method, image recognition method, apparatus and electronic device
WO2023109829A1 (fr) Procédé et appareil de traitement d'image, dispositif électronique et support de stockage
WO2022233223A1 (fr) Procédé et appareil d'assemblage d'image, dispositif et support
WO2023138549A1 (fr) Procédé et appareil de traitement d'image, et dispositif électronique et support de stockage
WO2023273697A1 (fr) Procédé et appareil de traitement d'image, procédé et appareil d'entraînement de modèle, dispositif électronique et support
WO2023138498A1 (fr) Procédé et appareil de génération d'image stylisée, dispositif électronique et support de stockage
CN112785669B (zh) 一种虚拟形象合成方法、装置、设备及存储介质
JP2023538825A (ja) ピクチャのビデオへの変換の方法、装置、機器および記憶媒体

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22914628

Country of ref document: EP

Kind code of ref document: A1