WO2021259185A1 - Image processing method and apparatus, device, and readable storage medium - Google Patents

Image processing method and apparatus, device, and readable storage medium Download PDF

Info

Publication number
WO2021259185A1
WO2021259185A1 PCT/CN2021/101166 CN2021101166W WO2021259185A1 WO 2021259185 A1 WO2021259185 A1 WO 2021259185A1 CN 2021101166 W CN2021101166 W CN 2021101166W WO 2021259185 A1 WO2021259185 A1 WO 2021259185A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
target
image
target object
input
Prior art date
Application number
PCT/CN2021/101166
Other languages
French (fr)
Chinese (zh)
Inventor
陈露兰
Original Assignee
维沃移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 维沃移动通信有限公司 filed Critical 维沃移动通信有限公司
Publication of WO2021259185A1 publication Critical patent/WO2021259185A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04845Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04847Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4038Image mosaicing, e.g. composing plane images from plane sub-images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content

Definitions

  • This application relates to the field of image technology, and in particular to an image processing method, device, device, and readable storage medium.
  • the embodiments of the present application provide an image processing method, device, equipment, and readable storage medium to solve the problem of complicated operations when generating a video.
  • an embodiment of the present application provides an image processing method.
  • the image processing method includes the following steps:
  • the first video is a video obtained by amplifying the expression or action of the target object in the second video
  • the second video is recorded for the target object during the shooting of the target image Video.
  • an image processing device including:
  • the first receiving module is configured to receive the first input of the user to the target object in the target image when the target image is displayed;
  • the first play module is configured to play the first video corresponding to the target object in response to the first input;
  • the first video is a video obtained by amplifying the expression or action of the target object in the second video
  • the second video is recorded for the target object during the shooting of the target image Video.
  • an embodiment of the present application provides an electronic device that includes a processor, a memory, and a program or instruction that is stored on the memory and can run on the processor.
  • the program or instruction is The processor implements the steps of the method described in the first aspect when executed.
  • an embodiment of the present application provides a readable storage medium, on which a program or instruction is stored, and when the program or instruction is executed by a processor, the steps of the method described in the first aspect are implemented .
  • an embodiment of the present application provides a chip, the chip includes a processor and a communication interface, the communication interface is coupled with the processor, and the processor is used to run a program or an instruction to implement the chip as in the first aspect The method described.
  • the first input of the user to the target object in the target image is received, and in response to the first input, the first video corresponding to the target object is played ,
  • the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is recorded for the target object during the shooting of the target image Video.
  • FIG. 1 is one of the flowcharts of the image processing method provided by the embodiment of the present application.
  • FIG. 2 is the second flowchart of the image processing method provided by the embodiment of the present application.
  • FIGS 3 to 4 are respectively schematic diagrams of display interfaces provided by embodiments of the present application.
  • FIG. 5 is a schematic diagram of a preprocessed video of a target object generated in the implementation of this application.
  • FIGS. 6 to 9 are schematic diagrams of display interfaces provided by embodiments of the present application.
  • Fig. 10(a) and Fig. 10(b) are schematic diagrams of the display interfaces provided by the embodiments of the present application.
  • FIG. 11 is the third flowchart of the image processing method provided by the embodiment of the present application.
  • Figures 12(a) and 12(b) are schematic diagrams of display interfaces provided by embodiments of the present application.
  • FIG(a) and 13(b) are schematic diagrams of display interfaces provided by embodiments of the present application.
  • 15(a) and 15(b) are schematic diagrams of display interfaces provided by embodiments of the present application.
  • 16(a) and 16(b) are schematic diagrams of display interfaces provided by an embodiment of the present application.
  • Figure 17 is a structural diagram of an image processing device provided by an embodiment of the present application.
  • FIG. 18 is one of the structural diagrams of an electronic device provided by an embodiment of the present application.
  • FIG. 19 is the second structural diagram of the electronic device provided by the embodiment of the present application.
  • first and second in the specification and claims of this application are used to distinguish similar objects, but not to describe a specific sequence or sequence. It should be understood that the data used in this way can be interchanged under appropriate circumstances, so that the embodiments of the present application can be implemented in an order other than those illustrated or described here, and the "first", “second”, etc. are distinguished
  • the objects are usually of one type, and the number of objects is not limited.
  • the first object may be one or multiple.
  • “and/or” in the description and claims means at least one of the connected objects, and the character “/” generally means that the associated objects before and after are in an "or” relationship.
  • FIG. 1 is a flowchart of an image processing method provided by an embodiment of the present application. As shown in FIG. 1, it includes the following steps:
  • Step 101 In the case of displaying a target image, receive a user's first input of a target object in the target image.
  • the target image may be an image captured by an electronic device when implementing the embodiments of the present application, may also be an image stored in an album of the electronic device, or may be an image corresponding to a shooting preview interface.
  • the first input may be touch input, click input, etc., for example.
  • the target object is not limited in the embodiment of the present application, and may be, for example, a certain person or a certain object.
  • Step 102 In response to the first input, play a first video corresponding to the target object.
  • the first video is a video obtained by amplifying the expression or action of the target object in the second video
  • the second video is recorded for the target object during the shooting of the target image Video. Since the second video is a video recorded for the target object during the shooting of the target image, in this case, the obtained image and video can reflect the action information of the target object. Therefore, in this way , Can make the video obtained more in line with the needs of users.
  • the video action amplification processing technology is based on signal processing technology, which treats the data of each pixel in the video as a time sequence, and amplifies a specific frequency domain signal through signal processing to obtain an action-amplified video.
  • Enlarging the video through video amplification technology can make those small changes that are hard to detect by the human eye amplify into an obvious look.
  • AI Artificial Intelligence
  • the user’s area of interest can be intercepted to obtain the emoticon package. Allow users to easily and conveniently obtain emoticons based on micro-expressions.
  • the length of the first video may be about 1s-2s.
  • the above method can be applied to electronic devices, such as: mobile phones, tablet computers (Tablet Personal Computer), laptop computers (Laptop Computer), personal digital assistants (personal digital assistant, PDA), mobile Internet devices (Mobile Internet Device, MID) or Wearable Device (Wearable Device), etc.
  • electronic devices such as: mobile phones, tablet computers (Tablet Personal Computer), laptop computers (Laptop Computer), personal digital assistants (personal digital assistant, PDA), mobile Internet devices (Mobile Internet Device, MID) or Wearable Device (Wearable Device), etc.
  • the first input of the user to the target object in the target image is received, and in response to the first input, the first video corresponding to the target object is played ,
  • the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is recorded for the target object during the shooting of the target image Video.
  • the image processing method of the embodiment of the present application may further include:
  • the multiple target frames of the second video include the target object
  • an area image is respectively captured from the multiple target frames, and each of the area images includes the target object. Then, a plurality of the regional images are spliced to obtain a pre-processed video. Afterwards, a video action amplification process is performed on the preprocessed video to obtain the first video.
  • the AI technology can be used to identify the video frame of the second video, and identify whether the target object is included. If the target object is included, it can be used as the target frame. Wherein, the target object may be one or multiple.
  • the display positions of the captured regional images may be the same.
  • the first display position of the target object in the first target frame is determined through intelligent recognition.
  • the first display position can be understood as the display position of the target object on the display screen of the electronic device. Since the size and pixel size of the display screen are known, if the target object is recognized, the first display position can be determined.
  • a rectangular frame or the like can be used to identify the display position of the target object in the image.
  • what is identified by the rectangular frame needs to include at least the face of the target object, or may also include the overall outline of the target object.
  • the region images corresponding to the first display position can be respectively intercepted from multiple target image frames. That is to say, the display position of the region image intercepted from different frames on the display screen is also the first display position.
  • the first video is a video that contains information about the zoom-in of the micro-expression action.
  • the first video obtained based on the video action magnification processing includes not only the magnification information of the micro-expression action, but also the magnification information of the movement such as limbs.
  • the region image is intercepted from the target frame including the target object, and then the region image is spliced to form a preprocessed video, and the video action is enlarged. In this way, since it is the first video obtained by processing the regional image, the processing speed is faster.
  • the image processing method of the embodiment of the present application may further include:
  • the target object may be one or multiple, and the obtained first video may include one or more If the target object includes multiple objects, the user can select the corresponding object as needed.
  • the target object includes at least two objects
  • the user's comments on the first object in the target image may be received
  • the second input obtain the video corresponding to the first object from the first video.
  • the second input may be an operation such as a click or a touch.
  • the multiple target frames of the first video include the first object
  • the multiple The area images are respectively captured in the target frame, and each of the area images includes the first object, and then the area images are spliced to obtain the video of the first object.
  • the target image can also be recognized to obtain the recognition result. Then, according to the recognition result, the object in the target image is highlighted.
  • the target image is intelligently recognized, and the object in the target image is highlighted according to the recognition result, for example, the location of the target object is displayed normally, and other locations of the target image are displayed blurred.
  • the method further includes receiving a third input from the user;
  • a dynamic image corresponding to the target object is obtained based on the first video corresponding to the target object.
  • the third input may be a touch input, a click input, and so on.
  • the dynamic image may be an image in the gif format, so that the first video can be quickly generated into a dynamic image, such as generating an emoticon package, which can increase interest.
  • Fig. 2 is a flowchart of an image processing method provided by an embodiment of the present application.
  • the micro-expression during the shooting of the subject is acquired.
  • an enlarged version of the video based on its micro-expression action can be generated.
  • the video may be an emoticon package or the like.
  • Step 201 Obtain an image of the target object.
  • the target object is a person as an example for description.
  • the terminal enters the camera preview interface.
  • the video generation function can be turned on. For example, when the user's operation on 21 in FIG. 3 is detected, as shown in FIG. 4, the video generation function can be turned on.
  • Step 202 Obtain a video of the target object.
  • the photo When the photo is taken, a photo is generated and stored in the album. And, the video recording function is turned on while taking a picture, and a video is obtained, that is, the video is recorded at the same time during the picture.
  • the time of the video can be very short, for example, about 1s-2s. In the embodiment of the present application, only one photo is displayed in the album, and the video content is not displayed.
  • Step 203 Generate a preprocessed video of the target object.
  • the location of the target person in the video can be identified through AI detection. Based on the interception of the image of the target person at the same position in each frame of the video, the video of the target person is obtained. If there are multiple people in the video, you can capture multiple videos.
  • the intercepted video is processed to generate the pre-processed video after the action is amplified, that is, those micro expressions in the original intercepted video that are not easily detectable by the naked eye, such as slightly lifting the corner of the mouth, slightly squeezing the eyebrows, and slightly opening the eyes Wait, in the new video, the amplitude of those micro-expression movements can be enlarged due to the processing of the video action amplification technology, which is obvious.
  • the position of the person in the corresponding image is recognized through AI, and the pre-processed video (the video containing the information of the micro-expression action enlargement) is one-to-one correspondence with the position of the person in the image.
  • the album will store a photo and a video corresponding to the location of the person, as well as their corresponding information.
  • Step 204 Preview the generated preprocessed video.
  • a preview function can be provided.
  • the target object that the user desires to preview is determined, and a video of the object is obtained.
  • the terminal enters the album video preview mode.
  • the album video preview mode is turned on, in response to the user's input, on the image obtained in step 201, a rectangular dashed frame is displayed at the position of the recognized person according to the recognition result of the AI.
  • Different characters have different colors of the dashed frame, and the terminal can display rectangular buttons with the same color as the rectangular dashed frame, as shown in the buttons 23 and 24 in Figure 7, that is, the buttons 23 and 24 of different colors are used to distinguish the interface The objects within the dashed box.
  • the user can long press any position in a certain dotted frame 25.
  • FIG 8 in response to the user's input, except for the content of the dotted frame, other parts of the image become semi-transparent, and the content of the character corresponding to the character’s zoomed-in information video including micro-expression actions is played once in the dotted frame. .
  • the user can observe the micro-expression of the person in the dashed frame at the moment of taking the photo, and the micro-expression actions have been processed to become obvious, and the user can clearly see the changes in the micro-actions that are usually difficult to notice.
  • the photo content is restored from the translucent interface when the emoji package was previewed to the video preview mode interface, as shown in Figure 9. If the user wants to preview again, he can press and hold the area in the dashed rectangle again.
  • Step 205 Generate a video of the target object.
  • the video preview can be exited, and the normal photo preview mode can be restored.
  • the micro-expression during the shooting of the person is acquired through the video action amplification technology and the AI recognition technology.
  • the user can specify a certain character to generate an enlarged version of the video based on its micro-expression, and the user does not need to perform additional video processing, which simplifies the operation and improves the user experience.
  • Fig. 11 is a flowchart of an image processing method provided by an embodiment of the present application.
  • the micro-expression during the shooting of the person is acquired.
  • an expression package based on the enlarged version of the micro expression action can be generated.
  • the user can set the location of the emoticon package so that the generated emoticon package can meet the needs of more users and enhance the user interaction experience.
  • the video may be an emoticon package or the like.
  • Step 1101 Obtain an image of the target object.
  • Step 1102 Obtain a video of the target object.
  • steps 1101 and 1102 can refer to the description of the foregoing steps 201 and 202.
  • Step 1103 Generate a preprocessed video of the target object.
  • step 203 the recorded video is processed through the video action amplification technology. After the processing is completed, a new video is generated after all the micro actions in the video are amplified by the action, that is, the micro expression action amplification information video.
  • Step 1104 Preview the generated preprocessed video.
  • the terminal enters the video preview mode.
  • the video preview mode is turned on, in response to the user's operation, on the image obtained in step 201, a rectangular dashed frame is displayed at the position of the recognized person according to the recognition result of the AI.
  • Different characters have different colors of the dashed frame, and the terminal can display rectangular buttons 27 and 28 with the same color as the rectangular dashed frame, that is, the buttons 27 and 28 of different colors are used to distinguish the objects in the dashed frame on the interface.
  • the electronic device may display a button 29. When the user finishes adjusting the dashed frame, click button 29 to determine the final position of the dashed frame.
  • the user can preview the micro-expression action once to zoom in on the information video content by long pressing the rectangular button in Fig. 13(a) corresponding to the color of the dashed rectangular frame.
  • Figure 13(b) in response to the user's input, except for the content of the dashed frame, other parts of the image become semi-transparent.
  • the video content of the micro-expression action magnification information corresponds to the position of the dotted rectangular box on the photo. It is the micro-emoji action magnification information video obtained after processing by applying the video action magnification technology (that is, the result of step 1103). ) Carry out the corresponding position tailoring generation. That is, after the user operates the dotted rectangular frame, the information of the person included in the rectangular frame can be recognized in response to the user's operation, and then, from the video obtained in step 1103, the video of the location of the person is intercepted .
  • the photo content is restored from the translucent interface when the emoji package was previewed to the video preview mode interface. If the user wants to preview again, he can press and hold the area in the dashed rectangle again.
  • the user can also adjust the position of the dashed rectangular frame.
  • the dashed frame can be adjusted upward.
  • the dashed frame is adjusted to encompass the entire face. If the user only wants to preview the zooming action of the micro-expression of a certain eye, the dashed frame is adjusted to the position of a certain eye.
  • Step 1105 Generate a video of the target object.
  • the user can click the rectangular button below the photo with the same color as the rectangle dotted line of the previewed micro-emoji action zoomed-in video.
  • the corresponding micro-expression action zoomed in on the video content will generate the video (such as an emoticon package in gif format) and save it in the album, and prompt the save success (as shown in Figure 16(b)).
  • the video preview can be exited in response to the user's input, and the normal photo preview mode can be restored.
  • the micro-expression during the shooting of the person is acquired through the video action amplification technology and the AI recognition technology.
  • the user can specify a certain character to generate an action-enlarged emoticon package based on its micro-expression, and the user does not need to perform additional video processing, which simplifies the operation and improves the user experience.
  • the user can set the location of the emoticon package so that the generated emoticon package can meet the needs of more users and enhance the user interaction experience.
  • the execution subject of the image processing method provided in the embodiments of the present application may be an image processing device, or a control module for executing the image processing method in the image processing device.
  • an image processing method executed by an image processing apparatus is taken as an example to illustrate the image processing apparatus provided in the embodiment of the present application.
  • FIG. 17 is a structural diagram of an image processing device provided by an embodiment of the present application. As shown in FIG. 17, the image processing device 1700 includes:
  • the first receiving module 1701 is configured to receive a user's first input of the target object in the target image when the target image is displayed; the first playing module 1702 is configured to respond to the first input to play and The first video corresponding to the target object; wherein, the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is a video in the target image The video recorded for the target object during the shooting process.
  • the device further includes:
  • the first interception module is configured to separately intercept regional images from the multiple target frames when the multiple target frames of the second video include the target object, and each of the regional images includes the target;
  • the first splicing module is used for splicing a plurality of the regional images to obtain a pre-processed video
  • the first processing module is configured to perform video action amplifying processing on the pre-processed video to obtain the first video.
  • the device further includes:
  • the first extraction module is configured to extract at least two images in the second video, wherein the image content of the at least one image includes the micro-actions or micro-expressions of the target object;
  • the second processing module is configured to perform video action magnification processing on the micro-actions or micro-expressions in the at least two images;
  • the third processing module is configured to splice the at least two images after the video action enlargement processing is performed to obtain the first video.
  • the device further includes:
  • a second receiving module configured to receive a second input of a user on the first object in the target image when the target object includes at least two objects;
  • the first obtaining module is configured to obtain a video corresponding to the first object from the first video in response to the second input.
  • the first obtaining module includes:
  • the first interception submodule is configured to separately intercept regional images from the multiple target frames when the multiple target frames of the first video include the first object, and each of the regional images includes The first object;
  • the first splicing sub-module is used for splicing the regional images to obtain the video of the first object.
  • the device further includes:
  • the recognition module is used to recognize the target image to obtain a recognition result
  • the fourth processing module is configured to highlight the object in the target image according to the recognition result.
  • the device further includes:
  • the third receiving module is used to receive the third input of the user
  • the fifth processing module is configured to obtain a dynamic image corresponding to the target object based on the first video corresponding to the target object in response to the third input.
  • the first input of the user to the target object in the target image is received, and in response to the first input, the first video corresponding to the target object is played ,
  • the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is recorded for the target object during the shooting of the target image Video.
  • the image processing device in the embodiment of the present application may be a device, or a component, an integrated circuit, or a chip in a terminal.
  • the device can be a mobile electronic device or a non-mobile electronic device.
  • the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a handheld computer, a vehicle electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook, or a personal digital assistant (personal digital assistant). assistant, PDA), etc.
  • Non-mobile electronic devices can be servers, network attached storage (NAS), personal computers (PC), televisions (television, TV), teller machines or self-service machines, etc., this application The embodiments are not specifically limited.
  • the image processing device in the embodiment of the present application may be a device with an operating system.
  • the operating system may be an Android operating system, an ios operating system, or other possible operating systems, which are not specifically limited in the embodiment of the present application.
  • the image processing apparatus provided by the embodiment of the present application can implement the various processes implemented by the method embodiments in FIG. 1 to FIG. 16. In order to avoid repetition, details are not described herein again.
  • an embodiment of the present application further provides an electronic device 1800, including a processor 1801, a memory 1802, and programs or instructions that are stored in the memory 1802 and run on the processor 1801,
  • an electronic device 1800 including a processor 1801, a memory 1802, and programs or instructions that are stored in the memory 1802 and run on the processor 1801,
  • the program or instruction is executed by the processor 1801
  • each process of the above-mentioned image processing method embodiment is realized, and the same technical effect can be achieved. To avoid repetition, details are not repeated here.
  • the electronic devices in the embodiments of the present application include the above-mentioned mobile electronic devices and non-mobile electronic devices.
  • FIG. 19 is a schematic diagram of the hardware structure of an electronic device that implements an embodiment of the present application.
  • the electronic device 1900 includes, but is not limited to: a radio frequency unit 1901, a network module 1902, an audio output unit 1903, an input unit 1904, a sensor 1905, a display unit 1906, a user input unit 1907, an interface unit 1908, a memory 1909, a processor 1910, etc. part.
  • the electronic device 1900 may also include a power source (such as a battery) for supplying power to various components.
  • the power source may be logically connected to the processor 1910 through a power management system, so that the power management system can manage charging, discharging, and power management. Consumption management and other functions.
  • the structure of the electronic device shown in FIG. 19 does not constitute a limitation on the electronic device.
  • the electronic device may include more or fewer components than those shown in the figure, or some components may be combined, or different component arrangements, which will not be repeated here. .
  • the radio frequency unit 1901 is configured to receive a user's first input to the target object in the target image when the target image is displayed; the processor 1910 is configured to respond to the first input and play and The first video corresponding to the target object; wherein, the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is a shot of the target image The video recorded for the target object during the process.
  • the first input of the user to the target object in the target image is received, and in response to the first input, the first video corresponding to the target object is played ,
  • the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is recorded for the target object during the shooting of the target image Video.
  • the processor 1910 is configured to, when the multiple target frames of the second video include the target object, respectively intercept regional images from the multiple target frames, and each of the regional images is Including the target object; splicing a plurality of the regional images to obtain a pre-processed video; performing video action amplification processing on the pre-processed video to obtain the first video.
  • the processor 1910 is configured to extract at least two images in the second video, where the image content of the at least two images includes micro-actions or micro-expressions of the target object;
  • the micro-actions or micro-expressions in the at least two images are subjected to video action enlargement processing; the at least two images after the video action enlargement processing are spliced to obtain the first video.
  • the processor 1910 is configured to receive a second input from the user to the first object in the target image when the target object includes at least two objects; in response to the second input, from Obtain the video corresponding to the first object from the first video.
  • the processor 1910 is configured to, when the multiple target frames of the first video include the first object, respectively intercept regional images from the multiple target frames, each of the regional images Both include the first object; stitching the region images to obtain a video of the first object.
  • the processor 1910 is configured to recognize the target image to obtain a recognition result; according to the recognition result, highlight an object in the target image.
  • the processor 1910 is configured to receive a third input of the user; in response to the third input, obtain a dynamic image corresponding to the target object based on the first video corresponding to the target object.
  • the input unit 1904 may include a graphics processing unit (GPU) 19041 and a microphone 19042.
  • the graphics processor 19041 is paired by the image capture device ( For example, the image data of the still picture or video obtained by the camera) is processed.
  • the display unit 1906 may include a display panel 19061, and the display panel 19061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like.
  • the user input unit 1907 includes a touch panel 19071 and other input devices 19072.
  • the touch panel 19071 is also called a touch screen.
  • the touch panel 19071 may include two parts, a touch detection device and a touch controller.
  • Other input devices 19072 may include, but are not limited to, a physical keyboard, function keys (such as volume control buttons, switch buttons, etc.), trackball, mouse, and joystick, which will not be repeated here.
  • the memory 1909 may be used to store software programs and various data, including but not limited to application programs and operating systems.
  • the processor 1910 may integrate an application processor and a modem processor.
  • the application processor mainly processes an operating system, a user interface, and an application program, and the modem processor mainly processes wireless communication. It can be understood that the foregoing modem processor may not be integrated into the processor 1910.
  • the embodiment of the present application also provides a readable storage medium with a program or instruction stored on the readable storage medium.
  • the program or instruction is executed by a processor, each process of the above-mentioned image processing method embodiment is realized, and the same technology can be achieved. The effect, in order to avoid repetition, will not be repeated here.
  • the readable storage medium such as read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk, or optical disk, etc.
  • An embodiment of the present application further provides a chip, the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is used to run a program or an instruction to implement each process of the above method embodiment , And can achieve the same technical effect, in order to avoid repetition, I will not repeat it here.
  • chips mentioned in the embodiments of the present application may also be referred to as system-level chips, system-on-chips, system-on-chips, or system-on-chips.
  • the technical solution of this application essentially or the part that contributes to the existing technology can be embodied in the form of a software product, and the computer software product is stored in a storage medium (such as ROM/RAM, magnetic disk, The optical disc) includes several instructions to make a terminal (which can be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) execute the methods described in the various embodiments of the present application.
  • a terminal which can be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • User Interface Of Digital Computer (AREA)
  • Studio Devices (AREA)

Abstract

Disclosed in the present application are an image processing method and apparatus, a device, and a readable storage medium. The method comprises: when displaying a target image, receiving a first input of a user for a target subject in the target image; and, in response to the first input, playing a first video corresponding to the target subject, the first video being a video obtained by performing enlargement processing on the expression or action of the target subject in a second video, and the second video being a video of the target subject recorded during the filming process of the target image.

Description

图像处理方法、装置、设备及可读存储介质Image processing method, device, equipment and readable storage medium
相关申请的交叉引用Cross-references to related applications
本申请主张在2020年6月24日在中国提交的中国专利申请号No.202010595589.X的优先权,其全部内容通过引用包含于此。This application claims the priority of Chinese Patent Application No. 202010595589.X filed in China on June 24, 2020, the entire content of which is incorporated herein by reference.
技术领域Technical field
本申请涉及图像技术领域,尤其涉及一种图像处理方法、装置、设备及可读存储介质。This application relates to the field of image technology, and in particular to an image processing method, device, device, and readable storage medium.
背景技术Background technique
移动终端等电子设备的使用越来越普及。目前,在利用移动终端等生成具有放大表情或者动作的视频时,一般需要用户录制视频或者拍摄多张照片,再通过视频处理软件对视频或者图片进行处理。但是,按照这种方式生成视频的操作复杂。The use of electronic devices such as mobile terminals is becoming more and more popular. At present, when using a mobile terminal or the like to generate a video with a magnified expression or action, it is generally necessary for the user to record a video or take multiple photos, and then process the video or picture through video processing software. However, the operation of generating a video in this way is complicated.
发明内容Summary of the invention
本申请实施例提供一种图像处理方法、装置、设备及可读存储介质,以解决生成视频时操作复杂的问题。The embodiments of the present application provide an image processing method, device, equipment, and readable storage medium to solve the problem of complicated operations when generating a video.
第一方面,本申请实施例提供了一种图像处理方法,图像处理方法,包括以下步骤:In the first aspect, an embodiment of the present application provides an image processing method. The image processing method includes the following steps:
在显示目标图像的情况下,接收用户对所述目标图像中的目标对象的第一输入;In the case of displaying the target image, receiving a user's first input on the target object in the target image;
响应于所述第一输入,播放与所述目标对象对应的第一视频;In response to the first input, play a first video corresponding to the target object;
其中,所述第一视频为对第二视频中的所述目标对象的表情或动作进行放大处理得到的视频,所述第二视频为在所述目标图像的拍摄过程中针对所述目标对象录制的视频。Wherein, the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is recorded for the target object during the shooting of the target image Video.
第二方面,本申请实施例提供了一种图像处理装置,包括:In the second aspect, an embodiment of the present application provides an image processing device, including:
第一接收模块,用于在显示目标图像的情况下,接收用户对所述目标图 像中的目标对象的第一输入;The first receiving module is configured to receive the first input of the user to the target object in the target image when the target image is displayed;
第一播放模块,用于响应于所述第一输入,播放与所述目标对象对应的第一视频;The first play module is configured to play the first video corresponding to the target object in response to the first input;
其中,所述第一视频为对第二视频中的所述目标对象的表情或动作进行放大处理得到的视频,所述第二视频为在所述目标图像的拍摄过程中针对所述目标对象录制的视频。Wherein, the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is recorded for the target object during the shooting of the target image Video.
第三方面,本申请实施例提供了一种电子设备,该电子设备包括处理器、存储器及存储在所述存储器上并可在所述处理器上运行的程序或指令,所述程序或指令被所述处理器执行时实现如第一方面所述的方法的步骤。In a third aspect, an embodiment of the present application provides an electronic device that includes a processor, a memory, and a program or instruction that is stored on the memory and can run on the processor. The program or instruction is The processor implements the steps of the method described in the first aspect when executed.
第四方面,本申请实施例提供了一种可读存储介质,所述可读存储介质上存储程序或指令,所述程序或指令被处理器执行时实现如第一方面所述的方法的步骤。In a fourth aspect, an embodiment of the present application provides a readable storage medium, on which a program or instruction is stored, and when the program or instruction is executed by a processor, the steps of the method described in the first aspect are implemented .
第五方面,本申请实施例提供了一种芯片,所述芯片包括处理器和通信接口,所述通信接口和所述处理器耦合,所述处理器用于运行程序或指令,实现如第一方面所述的方法。In a fifth aspect, an embodiment of the present application provides a chip, the chip includes a processor and a communication interface, the communication interface is coupled with the processor, and the processor is used to run a program or an instruction to implement the chip as in the first aspect The method described.
在本申请实施例中,在显示目标图像的情况下,接收用户对所述目标图像中的目标对象的第一输入,响应于所述第一输入,播放与所述目标对象对应的第一视频,而所述第一视频为对第二视频中的所述目标对象的表情或动作进行放大处理得到的视频,所述第二视频为在所述目标图像的拍摄过程中针对所述目标对象录制的视频。由此可以看出,在生成包括表情或者动作放大的视频时,无需调用其他的视频处理软件,从而简化了生成视频时的操作。In the embodiment of the present application, when the target image is displayed, the first input of the user to the target object in the target image is received, and in response to the first input, the first video corresponding to the target object is played , And the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is recorded for the target object during the shooting of the target image Video. It can be seen that there is no need to call other video processing software when generating a video that includes expressions or magnified actions, which simplifies the operation when generating a video.
附图说明Description of the drawings
图1是本申请实施例提供的图像处理方法的流程图之一;FIG. 1 is one of the flowcharts of the image processing method provided by the embodiment of the present application;
图2是本申请实施例提供的图像处理方法的流程图之二;FIG. 2 is the second flowchart of the image processing method provided by the embodiment of the present application;
图3至图4分别是本申请实施例提供的显示界面示意图;Figures 3 to 4 are respectively schematic diagrams of display interfaces provided by embodiments of the present application;
图5是本申请实施中生成目标对象的预处理视频的示意图;FIG. 5 is a schematic diagram of a preprocessed video of a target object generated in the implementation of this application;
图6至图9是本申请实施例提供的显示界面示意图;6 to 9 are schematic diagrams of display interfaces provided by embodiments of the present application;
图10(a)和图10(b)本申请实施例提供的显示界面示意图;Fig. 10(a) and Fig. 10(b) are schematic diagrams of the display interfaces provided by the embodiments of the present application;
图11是本申请实施例提供的图像处理方法的流程图之三;FIG. 11 is the third flowchart of the image processing method provided by the embodiment of the present application;
图12(a)和图12(b)是本申请实施例提供的显示界面示意图;Figures 12(a) and 12(b) are schematic diagrams of display interfaces provided by embodiments of the present application;
图13(a)和图13(b)是本申请实施例提供的显示界面示意图;13(a) and 13(b) are schematic diagrams of display interfaces provided by embodiments of the present application;
图14(a)和图14(b)是本申请实施例提供的显示界面示意图;14(a) and 14(b) are schematic diagrams of display interfaces provided by embodiments of the present application;
图15(a)和图15(b)是本申请实施例提供的显示界面示意图;15(a) and 15(b) are schematic diagrams of display interfaces provided by embodiments of the present application;
图16(a)和图16(b)是本申请实施例提供的显示界面示意图;16(a) and 16(b) are schematic diagrams of display interfaces provided by an embodiment of the present application;
图17是本申请实施例提供的图像处理装置的结构图;Figure 17 is a structural diagram of an image processing device provided by an embodiment of the present application;
图18是本申请实施例提供的电子设备的结构图之一;FIG. 18 is one of the structural diagrams of an electronic device provided by an embodiment of the present application;
图19是本申请实施例提供的电子设备的结构图之二。FIG. 19 is the second structural diagram of the electronic device provided by the embodiment of the present application.
具体实施方式detailed description
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。The technical solutions in the embodiments of the present application will be described clearly and completely in conjunction with the accompanying drawings in the embodiments of the present application. Obviously, the described embodiments are part of the embodiments of the present application, rather than all of them. Based on the embodiments in this application, all other embodiments obtained by those of ordinary skill in the art without creative work shall fall within the protection scope of this application.
本申请的说明书和权利要求书中的术语“第一”、“第二”等是用于区别类似的对象,而不用于描述特定的顺序或先后次序。应该理解这样使用的数据在适当情况下可以互换,以便本申请的实施例能够以除了在这里图示或描述的那些以外的顺序实施,且“第一”、“第二”等所区分的对象通常为一类,并不限定对象的个数,例如第一对象可以是一个,也可以是多个。此外,说明书以及权利要求中“和/或”表示所连接对象的至少其中之一,字符“/”,一般表示前后关联对象是一种“或”的关系。The terms "first" and "second" in the specification and claims of this application are used to distinguish similar objects, but not to describe a specific sequence or sequence. It should be understood that the data used in this way can be interchanged under appropriate circumstances, so that the embodiments of the present application can be implemented in an order other than those illustrated or described here, and the "first", "second", etc. are distinguished The objects are usually of one type, and the number of objects is not limited. For example, the first object may be one or multiple. In addition, "and/or" in the description and claims means at least one of the connected objects, and the character "/" generally means that the associated objects before and after are in an "or" relationship.
下面结合附图,通过具体的实施例及其应用场景对本申请实施例提供的图像处理方法进行详细地说明。The image processing method provided by the embodiments of the present application will be described in detail below with reference to the accompanying drawings, through specific embodiments and application scenarios thereof.
参见图1,图1是本申请实施例提供的图像处理方法的流程图,如图1所示,包括以下步骤:Referring to FIG. 1, FIG. 1 is a flowchart of an image processing method provided by an embodiment of the present application. As shown in FIG. 1, it includes the following steps:
步骤101、在显示目标图像的情况下,接收用户对所述目标图像中的目标对象的第一输入。Step 101: In the case of displaying a target image, receive a user's first input of a target object in the target image.
其中,所述目标图像可以是在执行本申请实施例的时候利用电子设备拍 摄的图像,还可以是存储在电子设备的相册中的图像,也可以是拍摄预览界面对应的图像。Wherein, the target image may be an image captured by an electronic device when implementing the embodiments of the present application, may also be an image stored in an album of the electronic device, or may be an image corresponding to a shooting preview interface.
其中,所述第一输入例如可以是触摸输入、点击输入等等。所述目标对象在本申请实施例中不做限定,例如可以是某个人,或者某个物体等。Wherein, the first input may be touch input, click input, etc., for example. The target object is not limited in the embodiment of the present application, and may be, for example, a certain person or a certain object.
步骤102、响应于所述第一输入,播放与所述目标对象对应的第一视频。Step 102: In response to the first input, play a first video corresponding to the target object.
其中,所述第一视频为对第二视频中的所述目标对象的表情或动作进行放大处理得到的视频,所述第二视频为在所述目标图像的拍摄过程中针对所述目标对象录制的视频。由于所述第二视频为在所述目标图像的拍摄过程中针对所述目标对象录制的视频,在这种情况下,获得的图像和视频可以反映目标对象的动作信息,因此,通过这种方式,可使获得的视频更符合用户的需求。Wherein, the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is recorded for the target object during the shooting of the target image Video. Since the second video is a video recorded for the target object during the shooting of the target image, in this case, the obtained image and video can reflect the action information of the target object. Therefore, in this way , Can make the video obtained more in line with the needs of users.
其中,视频动作放大处理技术基于信号处理技术,将视频中每个像素的数据看作时间序列,并且通过信号处理放大特定的频域信号,从而得到动作放大的视频。Among them, the video action amplification processing technology is based on signal processing technology, which treats the data of each pixel in the video as a time sequence, and amplifies a specific frequency domain signal through signal processing to obtain an action-amplified video.
通过视频放大技术对视频进行放大处理,能够让那些人眼难以察觉的微小变化放大成显而易见的样子,通过AI(Artificial Intelligence,人工智能)识别技术截取用户感兴趣的区域即可得到表情包,从而让用户简单便捷的获得基于微表情的表情包。Enlarging the video through video amplification technology can make those small changes that are hard to detect by the human eye amplify into an obvious look. Through AI (Artificial Intelligence) recognition technology, the user’s area of interest can be intercepted to obtain the emoticon package. Allow users to easily and conveniently obtain emoticons based on micro-expressions.
其中,所述第一视频的长度可为1s-2s左右。Wherein, the length of the first video may be about 1s-2s.
在本申请实施例中,上述方法可以应用于电子设备,例如:手机、平板电脑(Tablet Personal Computer)、膝上型电脑(Laptop Computer)、个人数字助理(personal digital assistant,PDA)、移动上网装置(Mobile Internet Device,MID)或可穿戴式设备(Wearable Device)等。In the embodiments of this application, the above method can be applied to electronic devices, such as: mobile phones, tablet computers (Tablet Personal Computer), laptop computers (Laptop Computer), personal digital assistants (personal digital assistant, PDA), mobile Internet devices (Mobile Internet Device, MID) or Wearable Device (Wearable Device), etc.
在本申请实施例中,在显示目标图像的情况下,接收用户对所述目标图像中的目标对象的第一输入,响应于所述第一输入,播放与所述目标对象对应的第一视频,而所述第一视频为对第二视频中的所述目标对象的表情或动作进行放大处理得到的视频,所述第二视频为在所述目标图像的拍摄过程中针对所述目标对象录制的视频。由此可以看出,在生成包括表情或者动作放大的视频时,无需调用其他的视频处理软件,从而简化了生成视频时的操作。In the embodiment of the present application, when the target image is displayed, the first input of the user to the target object in the target image is received, and in response to the first input, the first video corresponding to the target object is played , And the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is recorded for the target object during the shooting of the target image Video. It can be seen that there is no need to call other video processing software when generating a video that includes expressions or magnified actions, which simplifies the operation when generating a video.
可选的,在步骤102之前,本申请实施例的图像处理方法还可包括:Optionally, before step 102, the image processing method of the embodiment of the present application may further include:
在所述第二视频的多个目标帧包括所述目标对象的情况下,分别从所述多个目标帧中截取区域图像,每个所述区域图像均包括所述目标对象。然后,对多个所述区域图像进行拼接,得到预处理视频。之后,对所述预处理视频进行视频动作放大处理,得到所述第一视频。In the case where the multiple target frames of the second video include the target object, an area image is respectively captured from the multiple target frames, and each of the area images includes the target object. Then, a plurality of the regional images are spliced to obtain a pre-processed video. Afterwards, a video action amplification process is performed on the preprocessed video to obtain the first video.
具体的,在上述过程中,可通过AI技术对第二视频的视频帧进行识别,识别其中是否包括目标对象。如果包括目标对象,则可作为目标帧。其中,所述目标对象可以是一个,也可以是多个。Specifically, in the above process, the AI technology can be used to identify the video frame of the second video, and identify whether the target object is included. If the target object is included, it can be used as the target frame. Wherein, the target object may be one or multiple.
可选地,在本申请实施例中,不同的目标帧中,截取的区域图像的显示位置可相同。以目标对象为一个为例,通过智能识别确定出该目标对象在第一目标帧中的第一显示位置。其中,该第一显示位置可以理解为该目标对象在电子设备的显示屏上的显示位置。由于显示屏的尺寸、像素大小已知,所以,如果识别出了该目标对象,即可确定该第一显示位置。为清楚的标识该目标对象,可利用矩形框等标识,将该目标对象在图像中的显示位置标识出来。可选的,该矩形框标识出来的至少需要包括目标对象的脸部,或者,还可包括目标对象的整体轮廓。Optionally, in this embodiment of the present application, in different target frames, the display positions of the captured regional images may be the same. Taking the target object as an example, the first display position of the target object in the first target frame is determined through intelligent recognition. Wherein, the first display position can be understood as the display position of the target object on the display screen of the electronic device. Since the size and pixel size of the display screen are known, if the target object is recognized, the first display position can be determined. In order to clearly identify the target object, a rectangular frame or the like can be used to identify the display position of the target object in the image. Optionally, what is identified by the rectangular frame needs to include at least the face of the target object, or may also include the overall outline of the target object.
在确定了第一显示位置后,可从多个目标图像帧中,分别截取该第一显示位置对应的区域图像。也就是说,从不同帧中截取的区域图像在显示屏上的显示位置也是该第一显示位置。After the first display position is determined, the region images corresponding to the first display position can be respectively intercepted from multiple target image frames. That is to say, the display position of the region image intercepted from different frames on the display screen is also the first display position.
通过对所述预处理视频进行视频动作放大处理,以目标对象为人为例,那些肉眼不易察觉的微表情,比如微提嘴角,微挤眉,眼睛微睁大等,在经过视频动作放大技术的处理,那些微表情动作幅度得以放大,显而易见。此时,第一视频即含有微表情动作放大信息的视频。By performing video action zoom processing on the pre-processed video, taking the target object as an example, those micro expressions that are not easily detectable by the naked eye, such as slightly lifting the corners of the mouth, slightly squeezing the eyebrows, slightly widening the eyes, etc. Processing, the amplitude of those micro-expression movements can be magnified, which is obvious. At this time, the first video is a video that contains information about the zoom-in of the micro-expression action.
当然,在本申请实施例中,基于视频动作放大处理获得的第一视频中,不仅仅包括微表情动作的放大信息,还可包括肢体等动作的放大信息。Of course, in the embodiment of the present application, the first video obtained based on the video action magnification processing includes not only the magnification information of the micro-expression action, but also the magnification information of the movement such as limbs.
通过以上方式,从包括目标对象的目标帧中截取区域图像,再利用区域图像拼接形成预处理视频,并对其进行视频动作放大处理。通过这种方式,由于是通过对区域图像进行处理而得到的第一视频,因此,其处理速度较快。Through the above method, the region image is intercepted from the target frame including the target object, and then the region image is spliced to form a preprocessed video, and the video action is enlarged. In this way, since it is the first video obtained by processing the regional image, the processing speed is faster.
可选的,在步骤102之前,本申请实施例的图像处理方法还可包括:Optionally, before step 102, the image processing method of the embodiment of the present application may further include:
提取所述第二视频中的至少两张图像,其中,所述至少两张图像的图像内容中包括所述目标对象的微动作或微表情。然后,对所述至少两张图像中的微动作或者微表情进行视频动作放大处理。之后,对进行视频动作放大处理后的所述至少两张图像进行拼接,得到所述第一视频。通过这种方式,由于是对至少两张图像进行处理得到第一视频,可选地,目标对象可以是一个,也可以是多个,所获得的第一视频中可包括一个,也可以包括多个的对象,在目标对象包括多个的情况下,用户可以根据需要选择对应的对象。Extracting at least two images in the second video, wherein the image content of the at least two images includes micro-actions or micro-expressions of the target object. Then, the micro-action or micro-expression in the at least two images is subjected to video action enlargement processing. After that, the at least two images after the video action enlargement processing are spliced to obtain the first video. In this way, since the first video is obtained by processing at least two images, optionally, the target object may be one or multiple, and the obtained first video may include one or more If the target object includes multiple objects, the user can select the corresponding object as needed.
在所述目标对象包括至少两个对象的情况下,为便于用户从获得的第一视频中选择所需的内容,获得了第一视频之后,可接收用户对所述目标图像中的第一对象的第二输入,并响应于所述第二输入,从所述第一视频中获取所述第一对象对应的视频。其中,所述第二输入可以是点击、触摸等操作。In the case where the target object includes at least two objects, in order to facilitate the user to select the desired content from the obtained first video, after the first video is obtained, the user's comments on the first object in the target image may be received And in response to the second input, obtain the video corresponding to the first object from the first video. Wherein, the second input may be an operation such as a click or a touch.
具体的,在从所述第一视频中获取所述第一对象对应的视频的过程中,在所述第一视频的多个目标帧包括所述第一对象的情况下,从所述多个目标帧中分别截取区域图像,每个所述区域图像均包括所述第一对象,然后,将所述区域图像进行拼接,得到所述第一对象的视频。Specifically, in the process of acquiring the video corresponding to the first object from the first video, in the case where the multiple target frames of the first video include the first object, from the multiple The area images are respectively captured in the target frame, and each of the area images includes the first object, and then the area images are spliced to obtain the video of the first object.
这样,先对微动作或微表情进行放大处理后,再由用户选择对应的视频,In this way, after the micro-action or micro-expression is zoomed in, the user can select the corresponding video.
在以上实施例的基础上,还可对所述目标图像进行识别,得到识别结果。然后,根据识别结果,突出显示所述目标图像中的对象。例如,对目标图像进行智能识别,根据识别结果,突出显示所述目标图像中的对象,例如,将目标对象所在位置正常显示,目标图像的其他位置模糊显示等。通过以上方式,可便于用户对目标对象的选择。On the basis of the above embodiment, the target image can also be recognized to obtain the recognition result. Then, according to the recognition result, the object in the target image is highlighted. For example, the target image is intelligently recognized, and the object in the target image is highlighted according to the recognition result, for example, the location of the target object is displayed normally, and other locations of the target image are displayed blurred. Through the above method, it is convenient for the user to select the target object.
可选地,在播放与目标对象对应的第一视频之后,还包括接收用户的第三输入;Optionally, after the first video corresponding to the target object is played, the method further includes receiving a third input from the user;
响应于第三输入,基于与目标对象对应的第一视频得到目标对象对应的动态图像。In response to the third input, a dynamic image corresponding to the target object is obtained based on the first video corresponding to the target object.
可选地,第三输入可以是触摸输入、点击输入等等。Optionally, the third input may be a touch input, a click input, and so on.
动态图像可以是gif格式的图像,这样可以将第一视频快速地生成动态图像,比如生成表情包,可以增加趣味性。The dynamic image may be an image in the gif format, so that the first video can be quickly generated into a dynamic image, such as generating an emoticon package, which can increase interest.
图2是本申请实施例提供的图像处理方法的流程图。在本申请实施例中, 基于视频动作放大技术以及AI识别技术,获取对象拍摄过程中的微表情。可对于指定的某个对象,生成基于其微表情动作放大版的视频。其中,所述视频可以是表情包等。Fig. 2 is a flowchart of an image processing method provided by an embodiment of the present application. In the embodiment of the present application, based on the video action amplification technology and the AI recognition technology, the micro-expression during the shooting of the subject is acquired. For a specified object, an enlarged version of the video based on its micro-expression action can be generated. Wherein, the video may be an emoticon package or the like.
如图2所示,包括以下步骤:As shown in Figure 2, it includes the following steps:
步骤201、获取目标对象的图像。Step 201: Obtain an image of the target object.
在本申请实施例中,以目标对象为人为例进行描述。In the embodiments of the present application, the target object is a person as an example for description.
在此步骤中,开启拍摄模式。如图3所示,当启动拍照功能时,终端进入拍照预览界面。响应于用户的输入,可开启生成视频功能。例如,当检测到用户对图3中21的操作,如图4所示,可开启生成视频功能。In this step, turn on the shooting mode. As shown in Figure 3, when the camera function is activated, the terminal enters the camera preview interface. In response to the user's input, the video generation function can be turned on. For example, when the user's operation on 21 in FIG. 3 is detected, as shown in FIG. 4, the video generation function can be turned on.
步骤202、获取目标对象的视频。Step 202: Obtain a video of the target object.
当拍照后,生成一张照片并存储在相册中。并且,在拍照的同时开启视频录制功能,获得一段视频,也就是说,该视频是在拍照过程中同时录制的。该视频的时间可以很短,例如1s-2s左右。在本申请实施例中,相册中仅显示一张照片,并不显示该视频内容。When the photo is taken, a photo is generated and stored in the album. And, the video recording function is turned on while taking a picture, and a video is obtained, that is, the video is recorded at the same time during the picture. The time of the video can be very short, for example, about 1s-2s. In the embodiment of the present application, only one photo is displayed in the album, and the video content is not displayed.
步骤203、生成目标对象的预处理视频。Step 203: Generate a preprocessed video of the target object.
结合图5所示,获得视频后,可通过AI检测识别出视频中的目标人物的位置。基于对视频的每帧图像中同一位置处目标人物的图像的截取,获得目标人物的视频。若视频中有多个人物,可以截取多个视频。然后,通过视频动作放大技术,将截取的视频进行处理,生成动作放大后的预处理视频,即原本截取的视频中那些肉眼不易察觉的微表情,比如微提嘴角,微挤眉,眼睛微睁大等,在新的视频中因经过视频动作放大技术的处理,那些微表情动作幅度得以放大,显而易见。As shown in Figure 5, after the video is obtained, the location of the target person in the video can be identified through AI detection. Based on the interception of the image of the target person at the same position in each frame of the video, the video of the target person is obtained. If there are multiple people in the video, you can capture multiple videos. Then, through the video action amplification technology, the intercepted video is processed to generate the pre-processed video after the action is amplified, that is, those micro expressions in the original intercepted video that are not easily detectable by the naked eye, such as slightly lifting the corner of the mouth, slightly squeezing the eyebrows, and slightly opening the eyes Wait, in the new video, the amplitude of those micro-expression movements can be enlarged due to the processing of the video action amplification technology, which is obvious.
如上生成视频后,通过AI识别对应图像中该人物的位置,同时将预处理视频(含有微表情动作放大信息的视频)与图像中的人物位置进行一一对应。此时,相册处将存储一张照片以及对应人物位置的视频,以及二者的对应信息。After the video is generated as above, the position of the person in the corresponding image is recognized through AI, and the pre-processed video (the video containing the information of the micro-expression action enlargement) is one-to-one correspondence with the position of the person in the image. At this time, the album will store a photo and a video corresponding to the location of the person, as well as their corresponding information.
步骤204、预览生成的预处理视频。Step 204: Preview the generated preprocessed video.
当生成预处理视频后,可提供预览功能。此时,响应于用户的输入而确定用户期望预览的目标对象,并获得该对象的视频。When the pre-processed video is generated, a preview function can be provided. At this time, in response to the user's input, the target object that the user desires to preview is determined, and a video of the object is obtained.
如图6所示,当用户进入相册需要查看视频拍摄模式下拍摄所得的照片时,可点击图6中的按键22。响应于用户的输入,终端进入相册视频预览模式。当相册视频预览模式开启时,响应于用户的输入,在步骤201中获得的图像上,根据AI的识别结果,在识别出的人物的位置显示矩形虚线框。不同人物的虚线框颜色不同,并且终端上可显示与矩形虚线框颜色一样的矩形按键,如图7中的按键23和按键24所示,也即,利用不同颜色的按键23和按键24区别界面中虚线框内的对象。As shown in FIG. 6, when the user enters the album and needs to view the photos taken in the video shooting mode, he can click the button 22 in FIG. 6. In response to the user's input, the terminal enters the album video preview mode. When the album video preview mode is turned on, in response to the user's input, on the image obtained in step 201, a rectangular dashed frame is displayed at the position of the recognized person according to the recognition result of the AI. Different characters have different colors of the dashed frame, and the terminal can display rectangular buttons with the same color as the rectangular dashed frame, as shown in the buttons 23 and 24 in Figure 7, that is, the buttons 23 and 24 of different colors are used to distinguish the interface The objects within the dashed box.
此时,用户可长按某一个虚线框25中的任意位置。如图8所示,响应于用户的输入,图像中除了该虚线框的内容,其他部分均变成半透明,并且在该虚线框中播放一次该人物对应的包括微表情动作放大信息视频的内容。此时用户即可观察到该虚线框中人物拍照时刻的微表情,并且微表情动作都经过处理变得显而易见,用户可明显的看到那些平时难以注意到的微动作变化。At this time, the user can long press any position in a certain dotted frame 25. As shown in Figure 8, in response to the user's input, except for the content of the dotted frame, other parts of the image become semi-transparent, and the content of the character corresponding to the character’s zoomed-in information video including micro-expression actions is played once in the dotted frame. . At this point, the user can observe the micro-expression of the person in the dashed frame at the moment of taking the photo, and the micro-expression actions have been processed to become obvious, and the user can clearly see the changes in the micro-actions that are usually difficult to notice.
一次预览结束后,照片内容由表情包预览时半透明界面恢复到视频预览模式的界面,如图9所示。若用户想再次预览,可再次长按该矩形虚线框中的区域。After a preview is over, the photo content is restored from the translucent interface when the emoji package was previewed to the video preview mode interface, as shown in Figure 9. If the user wants to preview again, he can press and hold the area in the dashed rectangle again.
步骤205、生成目标对象的视频。Step 205: Generate a video of the target object.
若用户结束对当前微表情动作放大视频内容的预览后,如图10(a)所示,可点击照片下方与所预览微表情动作放大视频的矩形虚线颜色一致的矩形按键。响应于该操作,对应的微表情动作放大视频内容生成视频(例如生成gif格式的表情包)保存于相册中,并且提示保存成功(如图10(b)所示)。If the user finishes the preview of the current micro-emoji action zoomed-in video content, as shown in Figure 10(a), click the rectangular button below the photo with the same color as the rectangle dotted line of the previewed micro-emoji action zoomed video. In response to this operation, the corresponding micro-expression action magnifies the video content to generate a video (for example, generating an emoticon package in gif format) and saves it in the album, and prompts that the saving is successful (as shown in FIG. 10(b)).
当用户再次点击按键22,则响应于用户的输入可以退出视频预览,恢复到正常照片预览模式。When the user clicks the button 22 again, in response to the user's input, the video preview can be exited, and the normal photo preview mode can be restored.
在以上的实施例中,通过视频动作放大技术以及AI识别技术,获取人物拍摄过程中的微表情。用户可指定某个人物,生成基于其微表情的动作放大版的视频,并且用户不需要进行额外的视频处理,简化了操作,提升了用户的体验。In the above embodiment, the micro-expression during the shooting of the person is acquired through the video action amplification technology and the AI recognition technology. The user can specify a certain character to generate an enlarged version of the video based on its micro-expression, and the user does not need to perform additional video processing, which simplifies the operation and improves the user experience.
图11是本申请实施例提供的图像处理方法的流程图。在本申请实施例中,基于视频动作放大技术以及AI识别技术,获取人物拍摄过程中的微表情。可对于指定的某个人物,生成基于其微表情动作放大版的表情包。同时,用户 可对表情包的位置进行设置,使生成的表情包满足更多用户的需求,提升用户交互体验。其中,所述视频可以是表情包等。Fig. 11 is a flowchart of an image processing method provided by an embodiment of the present application. In the embodiment of the present application, based on the video action amplification technology and the AI recognition technology, the micro-expression during the shooting of the person is acquired. For a specified character, an expression package based on the enlarged version of the micro expression action can be generated. At the same time, the user can set the location of the emoticon package so that the generated emoticon package can meet the needs of more users and enhance the user interaction experience. Wherein, the video may be an emoticon package or the like.
如图11所示,包括以下步骤:As shown in Figure 11, it includes the following steps:
步骤1101、获取目标对象的图像。Step 1101: Obtain an image of the target object.
步骤1102、获取目标对象的视频。Step 1102: Obtain a video of the target object.
其中,步骤1101和1102的具体描述可参照前述步骤201和202的描述。Among them, the detailed description of steps 1101 and 1102 can refer to the description of the foregoing steps 201 and 202.
步骤1103、生成目标对象的预处理视频。Step 1103: Generate a preprocessed video of the target object.
与步骤203不同的是,在此步骤中,通过视频动作放大技术将录制的视频进行处理,处理完成后生成视频中所有微动作被动作放大后新的视频,即微表情动作放大信息视频。The difference from step 203 is that in this step, the recorded video is processed through the video action amplification technology. After the processing is completed, a new video is generated after all the micro actions in the video are amplified by the action, that is, the micro expression action amplification information video.
步骤1104、预览生成的预处理视频。Step 1104: Preview the generated preprocessed video.
如图12(a)所示,当用户进入相册需要查看视频拍摄模式下拍摄所得的照片时,可点击图12(a)中的按键26。响应于用户的输入,终端进入视频预览模式。当视频预览模式开启时,响应于用户的操作,在步骤201中获得的图像上,根据AI的识别结果,在识别出的人物的位置显示矩形虚线框。不同人物的虚线框颜色不同,并且终端上可显示与矩形虚线框颜色一样的矩形按键27,28,也即,利用不同颜色的按键27和按键28区别界面中虚线框内的对象。同时,如图12(b)所示,电子设备可显示按键29。当用户调整完虚线框时,点击按键29即可确定虚线框的最终位置。As shown in Fig. 12(a), when the user enters the album and needs to view the photos taken in the video shooting mode, he can click the button 26 in Fig. 12(a). In response to the user's input, the terminal enters the video preview mode. When the video preview mode is turned on, in response to the user's operation, on the image obtained in step 201, a rectangular dashed frame is displayed at the position of the recognized person according to the recognition result of the AI. Different characters have different colors of the dashed frame, and the terminal can display rectangular buttons 27 and 28 with the same color as the rectangular dashed frame, that is, the buttons 27 and 28 of different colors are used to distinguish the objects in the dashed frame on the interface. At the same time, as shown in FIG. 12(b), the electronic device may display a button 29. When the user finishes adjusting the dashed frame, click button 29 to determine the final position of the dashed frame.
此时,用户可通过长按图13(a)与虚线矩形框颜色对应的矩形按键,来预览一次微表情动作放大信息视频内容。如图13(b)所示,响应于用户的输入,图像中除了该虚线框的内容,其他部分均变成半透明。At this time, the user can preview the micro-expression action once to zoom in on the information video content by long pressing the rectangular button in Fig. 13(a) corresponding to the color of the dashed rectangular frame. As shown in Figure 13(b), in response to the user's input, except for the content of the dashed frame, other parts of the image become semi-transparent.
当用户长按了某个虚线框时,该微表情动作放大信息视频内容与照片处虚线矩形框位置对应,是通过应用视频动作放大技术处理后得到的微表情动作放大信息视频(即步骤1103所得)进行对应位置剪裁生成。也就是说,当用户对该虚线矩形框进行操作之后,可响应于用户的操作而识别该矩形框所包括的人物的信息,然后,从步骤1103获得的视频中,截取该人物所在位置的视频。When the user long presses a dashed box, the video content of the micro-expression action magnification information corresponds to the position of the dotted rectangular box on the photo. It is the micro-emoji action magnification information video obtained after processing by applying the video action magnification technology (that is, the result of step 1103). ) Carry out the corresponding position tailoring generation. That is, after the user operates the dotted rectangular frame, the information of the person included in the rectangular frame can be recognized in response to the user's operation, and then, from the video obtained in step 1103, the video of the location of the person is intercepted .
一次预览结束后,照片内容由表情包预览时半透明界面恢复到视频预览 模式的界面。若用户想再次预览,可再次长按该矩形虚线框中的区域。After a preview is over, the photo content is restored from the translucent interface when the emoji package was previewed to the video preview mode interface. If the user wants to preview again, he can press and hold the area in the dashed rectangle again.
此外,用户还可调整虚线矩形框的位置。如图14(a)和(b)所示,比如用户想预览整个脸部的微表情的放大动作,则用户可触摸虚线框并移动。如图14(a)所示,例如可向上调整虚线框。响应于用户的操作,将虚线框调整到包围整个脸部,若用户仅想预览某个眼睛的微表情的放大动作,则将虚线框调整到某个眼睛的位置。In addition, the user can also adjust the position of the dashed rectangular frame. As shown in Figure 14 (a) and (b), for example, if the user wants to preview the zoom-in action of the micro-expression of the entire face, the user can touch the dotted frame and move it. As shown in Fig. 14(a), for example, the dashed frame can be adjusted upward. In response to the user's operation, the dashed frame is adjusted to encompass the entire face. If the user only wants to preview the zooming action of the micro-expression of a certain eye, the dashed frame is adjusted to the position of a certain eye.
当用户调整完对当前矩形框位置满意可点击如图15(a)中的按键29,即矩形框位置调整确定键,退出编辑模式,如图15(b)所示。When the user is satisfied with the position of the current rectangular frame after adjusting, click the button 29 in Figure 15(a), that is, the confirmation button for adjusting the position of the rectangular frame to exit the editing mode, as shown in Figure 15(b).
步骤1105、生成目标对象的视频。Step 1105: Generate a video of the target object.
若用户结束对当前微表情动作放大视频内容的预览后,如图16(a)所示,可点击照片下方与所预览微表情动作放大视频的矩形虚线颜色一致的矩形按键。响应于该操作,对应的微表情动作放大视频内容将生成视频(如gif格式的表情包)保存于相册中,并且提示保存成功(如图16(b)所示)。If the user finishes the preview of the current micro-emoji action zoomed-in video content, as shown in Figure 16(a), the user can click the rectangular button below the photo with the same color as the rectangle dotted line of the previewed micro-emoji action zoomed-in video. In response to this operation, the corresponding micro-expression action zoomed in on the video content will generate the video (such as an emoticon package in gif format) and save it in the album, and prompt the save success (as shown in Figure 16(b)).
当用户再次点击按键25,则可响应于用户的输入可以退出视频预览,恢复到正常照片预览模式。When the user clicks the button 25 again, the video preview can be exited in response to the user's input, and the normal photo preview mode can be restored.
在以上的实施例中,通过视频动作放大技术以及AI识别技术,获取人物拍摄过程中的微表情。用户可指定某个人物,生成基于其微表情的动作放大版表情包,并且用户不需要进行额外的视频处理,简化了操作,提升了用户的体验。同时,用户可对表情包的位置进行设置,使生成的表情包满足更多用户的需求,提升用户交互体验。In the above embodiment, the micro-expression during the shooting of the person is acquired through the video action amplification technology and the AI recognition technology. The user can specify a certain character to generate an action-enlarged emoticon package based on its micro-expression, and the user does not need to perform additional video processing, which simplifies the operation and improves the user experience. At the same time, the user can set the location of the emoticon package so that the generated emoticon package can meet the needs of more users and enhance the user interaction experience.
需要说明的是,本申请实施例提供的图像处理方法,执行主体可以为图像处理装置,或者该图像处理装置中的用于执行图像处理方法的控制模块。本申请实施例中以图像处理装置执行图像处理方法为例,说明本申请实施例提供的图像处理装置。It should be noted that the execution subject of the image processing method provided in the embodiments of the present application may be an image processing device, or a control module for executing the image processing method in the image processing device. In the embodiment of the present application, an image processing method executed by an image processing apparatus is taken as an example to illustrate the image processing apparatus provided in the embodiment of the present application.
参见图17,图17是本申请实施例提供的图像处理装置的结构图,如图17所示,图像处理装置1700包括:Referring to FIG. 17, FIG. 17 is a structural diagram of an image processing device provided by an embodiment of the present application. As shown in FIG. 17, the image processing device 1700 includes:
第一接收模块1701,用于在显示目标图像的情况下,接收用户对所述目标图像中的目标对象的第一输入;第一播放模块1702,用于响应于所述第一输入,播放与所述目标对象对应的第一视频;其中,所述第一视频为对第二 视频中的所述目标对象的表情或动作进行放大处理得到的视频,所述第二视频为在所述目标图像的拍摄过程中针对所述目标对象录制的视频。The first receiving module 1701 is configured to receive a user's first input of the target object in the target image when the target image is displayed; the first playing module 1702 is configured to respond to the first input to play and The first video corresponding to the target object; wherein, the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is a video in the target image The video recorded for the target object during the shooting process.
可选的,所述装置还包括:Optionally, the device further includes:
第一截取模块,用于在所述第二视频的多个目标帧包括所述目标对象的情况下,分别从所述多个目标帧中截取区域图像,每个所述区域图像均包括所述目标对象;The first interception module is configured to separately intercept regional images from the multiple target frames when the multiple target frames of the second video include the target object, and each of the regional images includes the target;
第一拼接模块,用于对多个所述区域图像进行拼接,得到预处理视频;The first splicing module is used for splicing a plurality of the regional images to obtain a pre-processed video;
第一处理模块,用于对所述预处理视频进行视频动作放大处理,得到所述第一视频。The first processing module is configured to perform video action amplifying processing on the pre-processed video to obtain the first video.
可选的,所述装置还包括:Optionally, the device further includes:
第一提取模块,用于提取所述第二视频中的至少两张图像,其中,所述至少一张图像的图像内容中包括所述目标对象的微动作或微表情;The first extraction module is configured to extract at least two images in the second video, wherein the image content of the at least one image includes the micro-actions or micro-expressions of the target object;
第二处理模块,用于对所述至少两张图像中的微动作或者微表情进行视频动作放大处理;The second processing module is configured to perform video action magnification processing on the micro-actions or micro-expressions in the at least two images;
第三处理模块,用于对进行视频动作放大处理后的所述至少两张图像进行拼接,得到所述第一视频。The third processing module is configured to splice the at least two images after the video action enlargement processing is performed to obtain the first video.
可选的,所述装置还包括:Optionally, the device further includes:
第二接收模块,用于在所述目标对象包括至少两个对象的情况下,接收用户对所述目标图像中的第一对象的第二输入;A second receiving module, configured to receive a second input of a user on the first object in the target image when the target object includes at least two objects;
第一获取模块,用于响应于所述第二输入,从所述第一视频中获取所述第一对象对应的视频。The first obtaining module is configured to obtain a video corresponding to the first object from the first video in response to the second input.
可选的,所述第一获取模块包括:Optionally, the first obtaining module includes:
第一截取子模块,用于在所述第一视频的多个目标帧包括所述第一对象的情况下,从所述多个目标帧中分别截取区域图像,每个所述区域图像均包括所述第一对象;The first interception submodule is configured to separately intercept regional images from the multiple target frames when the multiple target frames of the first video include the first object, and each of the regional images includes The first object;
第一拼接子模块,用于将所述区域图像进行拼接,得到所述第一对象的视频。The first splicing sub-module is used for splicing the regional images to obtain the video of the first object.
可选的,所述装置还包括:Optionally, the device further includes:
识别模块,用于对所述目标图像进行识别,得到识别结果;The recognition module is used to recognize the target image to obtain a recognition result;
第四处理模块,用于根据所述识别结果,突出显示所述目标图像中的对象。The fourth processing module is configured to highlight the object in the target image according to the recognition result.
可选的,所述装置还包括:Optionally, the device further includes:
第三接收模块,用于接收用户的第三输入;The third receiving module is used to receive the third input of the user;
第五处理模块,用于响应于所述第三输入,基于所述与所述目标对象对应的第一视频得到所述目标对象对应的动态图像。The fifth processing module is configured to obtain a dynamic image corresponding to the target object based on the first video corresponding to the target object in response to the third input.
在本申请实施例中,在显示目标图像的情况下,接收用户对所述目标图像中的目标对象的第一输入,响应于所述第一输入,播放与所述目标对象对应的第一视频,而所述第一视频为对第二视频中的所述目标对象的表情或动作进行放大处理得到的视频,所述第二视频为在所述目标图像的拍摄过程中针对所述目标对象录制的视频。由此可以看出,利用本申请实施例,在生成包括表情或者动作放大的视频时,无需调用其他的视频处理软件,从而简化了生成视频时的操作。In the embodiment of the present application, when the target image is displayed, the first input of the user to the target object in the target image is received, and in response to the first input, the first video corresponding to the target object is played , And the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is recorded for the target object during the shooting of the target image Video. It can be seen from this that, by using the embodiments of the present application, when generating a video that includes expressions or magnified actions, there is no need to call other video processing software, which simplifies the operation when generating a video.
本申请实施例中的图像处理装置可以是装置,也可以是终端中的部件、集成电路、或芯片。该装置可以是移动电子设备,也可以为非移动电子设备。示例性的,移动电子设备可以为手机、平板电脑、笔记本电脑、掌上电脑、车载电子设备、可穿戴设备、超级移动个人计算机(ultra-mobile personal computer,UMPC)、上网本或者个人数字助理(personal digital assistant,PDA)等,非移动电子设备可以为服务器、网络附属存储器(Network Attached Storage,NAS)、个人计算机(personal computer,PC)、电视机(television,TV)、柜员机或者自助机等,本申请实施例不作具体限定。The image processing device in the embodiment of the present application may be a device, or a component, an integrated circuit, or a chip in a terminal. The device can be a mobile electronic device or a non-mobile electronic device. Exemplarily, the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a handheld computer, a vehicle electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook, or a personal digital assistant (personal digital assistant). assistant, PDA), etc. Non-mobile electronic devices can be servers, network attached storage (NAS), personal computers (PC), televisions (television, TV), teller machines or self-service machines, etc., this application The embodiments are not specifically limited.
本申请实施例中的图像处理装置可以为具有操作系统的装置。该操作系统可以为安卓(Android)操作系统,可以为ios操作系统,还可以为其他可能的操作系统,本申请实施例不作具体限定。The image processing device in the embodiment of the present application may be a device with an operating system. The operating system may be an Android operating system, an ios operating system, or other possible operating systems, which are not specifically limited in the embodiment of the present application.
本申请实施例提供的图像处理装置能够实现图1至图16的方法实施例实现的各个过程,为避免重复,这里不再赘述。The image processing apparatus provided by the embodiment of the present application can implement the various processes implemented by the method embodiments in FIG. 1 to FIG. 16. In order to avoid repetition, details are not described herein again.
可选的,如图18所示,本申请实施例还提供一种电子设备1800,包括处理器1801,存储器1802,存储在存储器1802上并可在所述处理器1801上运行的程序或指令,该程序或指令被处理器1801执行时实现上述图像处理方法 实施例的各个过程,且能达到相同的技术效果,为避免重复,这里不再赘述。Optionally, as shown in FIG. 18, an embodiment of the present application further provides an electronic device 1800, including a processor 1801, a memory 1802, and programs or instructions that are stored in the memory 1802 and run on the processor 1801, When the program or instruction is executed by the processor 1801, each process of the above-mentioned image processing method embodiment is realized, and the same technical effect can be achieved. To avoid repetition, details are not repeated here.
需要说明的是,本申请实施例中的电子设备包括上述所述的移动电子设备和非移动电子设备。It should be noted that the electronic devices in the embodiments of the present application include the above-mentioned mobile electronic devices and non-mobile electronic devices.
图19为实现本申请实施例的一种电子设备的硬件结构示意图。该电子设备1900包括但不限于:射频单元1901、网络模块1902、音频输出单元1903、输入单元1904、传感器1905、显示单元1906、用户输入单元1907、接口单元1908、存储器1909、以及处理器1910等部件。FIG. 19 is a schematic diagram of the hardware structure of an electronic device that implements an embodiment of the present application. The electronic device 1900 includes, but is not limited to: a radio frequency unit 1901, a network module 1902, an audio output unit 1903, an input unit 1904, a sensor 1905, a display unit 1906, a user input unit 1907, an interface unit 1908, a memory 1909, a processor 1910, etc. part.
本领域技术人员可以理解,电子设备1900还可以包括给各个部件供电的电源(比如电池),电源可以通过电源管理系统与处理器1910逻辑相连,从而通过电源管理系统实现管理充电、放电、以及功耗管理等功能。图19中示出的电子设备结构并不构成对电子设备的限定,电子设备可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件布置,在此不再赘述。Those skilled in the art can understand that the electronic device 1900 may also include a power source (such as a battery) for supplying power to various components. The power source may be logically connected to the processor 1910 through a power management system, so that the power management system can manage charging, discharging, and power management. Consumption management and other functions. The structure of the electronic device shown in FIG. 19 does not constitute a limitation on the electronic device. The electronic device may include more or fewer components than those shown in the figure, or some components may be combined, or different component arrangements, which will not be repeated here. .
其中,射频单元1901,用于在显示目标图像的情况下,接收用户对所述目标图像中的目标对象的第一输入;处理器1910,用于响应于所述第一输入,播放与所述目标对象对应的第一视频;其中,所述第一视频为对第二视频中的所述目标对象的表情或动作进行放大处理得到的视频,所述第二视频为在所述目标图像的拍摄过程中针对所述目标对象录制的视频。Wherein, the radio frequency unit 1901 is configured to receive a user's first input to the target object in the target image when the target image is displayed; the processor 1910 is configured to respond to the first input and play and The first video corresponding to the target object; wherein, the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is a shot of the target image The video recorded for the target object during the process.
在本申请实施例中,在显示目标图像的情况下,接收用户对所述目标图像中的目标对象的第一输入,响应于所述第一输入,播放与所述目标对象对应的第一视频,而所述第一视频为对第二视频中的所述目标对象的表情或动作进行放大处理得到的视频,所述第二视频为在所述目标图像的拍摄过程中针对所述目标对象录制的视频。由此可以看出,利用本申请实施例,在生成包括表情或者动作放大的视频时,无需调用其他的视频处理软件,从而简化了生成视频时的操作。In the embodiment of the present application, when the target image is displayed, the first input of the user to the target object in the target image is received, and in response to the first input, the first video corresponding to the target object is played , And the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is recorded for the target object during the shooting of the target image Video. It can be seen from this that, by using the embodiments of the present application, when generating a video that includes expressions or magnified actions, there is no need to call other video processing software, which simplifies the operation when generating a video.
可选的,处理器1910,用于在所述第二视频的多个目标帧包括所述目标对象的情况下,分别从所述多个目标帧中截取区域图像,每个所述区域图像均包括所述目标对象;对多个所述区域图像进行拼接,得到预处理视频;对所述预处理视频进行视频动作放大处理,得到所述第一视频。Optionally, the processor 1910 is configured to, when the multiple target frames of the second video include the target object, respectively intercept regional images from the multiple target frames, and each of the regional images is Including the target object; splicing a plurality of the regional images to obtain a pre-processed video; performing video action amplification processing on the pre-processed video to obtain the first video.
可选的,处理器1910,用于提取所述第二视频中的至少两张图像,其中, 所述至少两张图像的图像内容中包括所述目标对象的微动作或微表情;对所述至少两张图像中的微动作或者微表情进行视频动作放大处理;对进行视频动作放大处理后的所述至少两张图像进行拼接,得到所述第一视频。Optionally, the processor 1910 is configured to extract at least two images in the second video, where the image content of the at least two images includes micro-actions or micro-expressions of the target object; The micro-actions or micro-expressions in the at least two images are subjected to video action enlargement processing; the at least two images after the video action enlargement processing are spliced to obtain the first video.
可选的,处理器1910,用于在所述目标对象包括至少两个对象的情况下,接收用户对所述目标图像中的第一对象的第二输入;响应于所述第二输入,从所述第一视频中获取所述第一对象对应的视频。Optionally, the processor 1910 is configured to receive a second input from the user to the first object in the target image when the target object includes at least two objects; in response to the second input, from Obtain the video corresponding to the first object from the first video.
可选的,处理器1910,用于在所述第一视频的多个目标帧包括所述第一对象的情况下,从所述多个目标帧中分别截取区域图像,每个所述区域图像均包括所述第一对象;将所述区域图像进行拼接,得到所述第一对象的视频。Optionally, the processor 1910 is configured to, when the multiple target frames of the first video include the first object, respectively intercept regional images from the multiple target frames, each of the regional images Both include the first object; stitching the region images to obtain a video of the first object.
可选的,处理器1910,用于对所述目标图像进行识别,得到识别结果;根据所述识别结果,突出显示所述目标图像中的对象。Optionally, the processor 1910 is configured to recognize the target image to obtain a recognition result; according to the recognition result, highlight an object in the target image.
可选的,处理器1910,用于接收用户的第三输入;响应于所述第三输入,基于所述与所述目标对象对应的第一视频得到所述目标对象对应的动态图像。Optionally, the processor 1910 is configured to receive a third input of the user; in response to the third input, obtain a dynamic image corresponding to the target object based on the first video corresponding to the target object.
应理解的是,本申请实施例中,输入单元1904可以包括图形处理器(Graphics Processing Unit,GPU)19041和麦克风19042,图形处理器19041对在视频捕获模式或图像捕获模式中由图像捕获装置(如摄像头)获得的静态图片或视频的图像数据进行处理。显示单元1906可包括显示面板19061,可以采用液晶显示器、有机发光二极管等形式来配置显示面板19061。用户输入单元1907包括触控面板19071以及其他输入设备19072。触控面板19071,也称为触摸屏。触控面板19071可包括触摸检测装置和触摸控制器两个部分。其他输入设备19072可以包括但不限于物理键盘、功能键(比如音量控制按键、开关按键等)、轨迹球、鼠标、操作杆,在此不再赘述。存储器1909可用于存储软件程序以及各种数据,包括但不限于应用程序和操作系统。处理器1910可集成应用处理器和调制解调处理器,其中,应用处理器主要处理操作系统、用户界面和应用程序等,调制解调处理器主要处理无线通信。可以理解的是,上述调制解调处理器也可以不集成到处理器1910中。It should be understood that, in the embodiment of the present application, the input unit 1904 may include a graphics processing unit (GPU) 19041 and a microphone 19042. The graphics processor 19041 is paired by the image capture device ( For example, the image data of the still picture or video obtained by the camera) is processed. The display unit 1906 may include a display panel 19061, and the display panel 19061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. The user input unit 1907 includes a touch panel 19071 and other input devices 19072. The touch panel 19071 is also called a touch screen. The touch panel 19071 may include two parts, a touch detection device and a touch controller. Other input devices 19072 may include, but are not limited to, a physical keyboard, function keys (such as volume control buttons, switch buttons, etc.), trackball, mouse, and joystick, which will not be repeated here. The memory 1909 may be used to store software programs and various data, including but not limited to application programs and operating systems. The processor 1910 may integrate an application processor and a modem processor. The application processor mainly processes an operating system, a user interface, and an application program, and the modem processor mainly processes wireless communication. It can be understood that the foregoing modem processor may not be integrated into the processor 1910.
本申请实施例还提供一种可读存储介质,可读存储介质上存储有程序或指令,该程序或指令被处理器执行时实现上述图像处理方法实施例的各个过程,且能达到相同的技术效果,为避免重复,这里不再赘述。The embodiment of the present application also provides a readable storage medium with a program or instruction stored on the readable storage medium. When the program or instruction is executed by a processor, each process of the above-mentioned image processing method embodiment is realized, and the same technology can be achieved. The effect, in order to avoid repetition, will not be repeated here.
其中,所述可读存储介质,如只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等。Wherein, the readable storage medium, such as read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk, or optical disk, etc.
本申请实施例另提供了一种芯片,所述芯片包括处理器和通信接口,所述通信接口和所述处理器耦合,所述处理器用于运行程序或指令,实现上述方法实施例的各个过程,且能达到相同的技术效果,为避免重复,这里不再赘述。An embodiment of the present application further provides a chip, the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is used to run a program or an instruction to implement each process of the above method embodiment , And can achieve the same technical effect, in order to avoid repetition, I will not repeat it here.
应理解,本申请实施例提到的芯片还可以称为系统级芯片、系统芯片、芯片系统或片上系统芯片等。It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-level chips, system-on-chips, system-on-chips, or system-on-chips.
需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者装置不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者装置所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括该要素的过程、方法、物品或者装置中还存在另外的相同要素。此外,需要指出的是,本申请实施方式中的方法和装置的范围不限按示出或讨论的顺序来执行功能,还可包括根据所涉及的功能按基本同时的方式或按相反的顺序来执行功能,例如,可以按不同于所描述的次序来执行所描述的方法,并且还可以添加、省去、或组合各种步骤。另外,参照某些示例所描述的特征可在其他示例中被组合。It should be noted that in this article, the terms "include", "include" or any other variants thereof are intended to cover non-exclusive inclusion, so that a process, method, article or device including a series of elements not only includes those elements, It also includes other elements that are not explicitly listed, or elements inherent to the process, method, article, or device. If there are no more restrictions, the element defined by the sentence "including a..." does not exclude the existence of other identical elements in the process, method, article, or device that includes the element. In addition, it should be pointed out that the scope of the methods and devices in the embodiments of the present application is not limited to performing functions in the order shown or discussed, and may also include performing functions in a substantially simultaneous manner or in reverse order according to the functions involved. The functions are performed, for example, the described method may be performed in an order different from the described order, and various steps may also be added, omitted, or combined. In addition, features described with reference to certain examples may be combined in other examples.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端(可以是手机,计算机,服务器,空调器,或者网络设备等)执行本申请各个实施例所述的方法。Through the description of the above implementation manners, those skilled in the art can clearly understand that the above-mentioned embodiment methods can be implemented by means of software plus the necessary general hardware platform. Of course, it can also be implemented by hardware, but in many cases the former is better.的实施方式。 Based on this understanding, the technical solution of this application essentially or the part that contributes to the existing technology can be embodied in the form of a software product, and the computer software product is stored in a storage medium (such as ROM/RAM, magnetic disk, The optical disc) includes several instructions to make a terminal (which can be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) execute the methods described in the various embodiments of the present application.
上面结合附图对本申请的实施例进行了描述,但是本申请并不局限于上述的具体实施方式,上述的具体实施方式仅仅是示意性的,而不是限制性的,本领域的普通技术人员在本申请的启示下,在不脱离本申请宗旨和权利要求所保护的范围情况下,还可做出很多形式,均属于本申请的保护之内。The embodiments of the present application are described above with reference to the accompanying drawings, but the present application is not limited to the above-mentioned specific embodiments. The above-mentioned specific embodiments are only illustrative and not restrictive. Those of ordinary skill in the art are Under the enlightenment of this application, many forms can be made without departing from the purpose of this application and the scope of protection of the claims, all of which fall within the protection of this application.

Claims (20)

  1. 一种图像处理方法,包括以下步骤:An image processing method includes the following steps:
    在显示目标图像的情况下,接收用户对所述目标图像中的目标对象的第一输入;In the case of displaying the target image, receiving a user's first input on the target object in the target image;
    响应于所述第一输入,播放与所述目标对象对应的第一视频;In response to the first input, play a first video corresponding to the target object;
    其中,所述第一视频为对第二视频中的所述目标对象的表情或动作进行放大处理得到的视频,所述第二视频为在所述目标图像的拍摄过程中针对所述目标对象录制的视频。Wherein, the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is recorded for the target object during the shooting of the target image Video.
  2. 根据权利要求1所述的方法,其中,在所述播放与所述目标对象对应的第一视频之前,所述方法还包括:The method according to claim 1, wherein, before the playing the first video corresponding to the target object, the method further comprises:
    在所述第二视频的多个目标帧包括所述目标对象的情况下,分别从所述多个目标帧中截取区域图像,每个所述区域图像均包括所述目标对象;In the case that the multiple target frames of the second video include the target object, respectively intercepting area images from the multiple target frames, and each of the area images includes the target object;
    对多个所述区域图像进行拼接,得到预处理视频;Splicing a plurality of the regional images to obtain a pre-processed video;
    对所述预处理视频进行视频动作放大处理,得到所述第一视频。Perform video action amplification processing on the preprocessed video to obtain the first video.
  3. 根据权利要求1所述的方法,其中,在所述播放与所述目标对象对应的第一视频之前,所述方法还包括:The method according to claim 1, wherein, before the playing the first video corresponding to the target object, the method further comprises:
    提取所述第二视频中的至少两张图像,其中,所述至少两张图像的图像内容中包括所述目标对象的微动作或微表情;Extracting at least two images in the second video, wherein the image content of the at least two images includes micro-actions or micro-expressions of the target object;
    对所述至少两张图像中的微动作或者微表情进行视频动作放大处理;Performing video action magnification processing on the micro-actions or micro-expressions in the at least two images;
    对进行视频动作放大处理后的所述至少两张图像进行拼接,得到所述第一视频。Splicing the at least two images after the video action enlargement processing is performed to obtain the first video.
  4. 根据权利要求3所述的方法,其中,在所述得到所述第一视频之后,所述方法包括:The method according to claim 3, wherein, after said obtaining the first video, the method comprises:
    在所述目标对象包括至少两个对象的情况下,接收用户对所述目标图像中的第一对象的第二输入;In a case where the target object includes at least two objects, receiving a second input from the user to the first object in the target image;
    响应于所述第二输入,从所述第一视频中获取所述第一对象对应的视频。In response to the second input, a video corresponding to the first object is obtained from the first video.
  5. 根据权利要求4所述的方法,其中,所述从所述第一视频中获取所述第一对象的视频,包括:The method according to claim 4, wherein said obtaining the video of the first object from the first video comprises:
    在所述第一视频的多个目标帧包括所述第一对象的情况下,从所述多个目标帧中分别截取区域图像,每个所述区域图像均包括所述第一对象;In the case that the multiple target frames of the first video include the first object, respectively intercepting regional images from the multiple target frames, and each of the regional images includes the first object;
    将所述区域图像进行拼接,得到所述第一对象的视频。Splicing the regional images to obtain the video of the first object.
  6. 根据权利要求1所述的方法,还包括:The method according to claim 1, further comprising:
    对所述目标图像进行识别,得到识别结果;Recognizing the target image to obtain a recognition result;
    根据所述识别结果,突出显示所述目标图像中的对象。According to the recognition result, the object in the target image is highlighted.
  7. 根据权利要求1所述的方法,其中,在所述播放与所述目标对象对应的第一视频之后,还包括:The method according to claim 1, wherein after the playing the first video corresponding to the target object, the method further comprises:
    接收用户的第三输入;Receive the user's third input;
    响应于所述第三输入,基于所述与所述目标对象对应的第一视频得到所述目标对象对应的动态图像。In response to the third input, a dynamic image corresponding to the target object is obtained based on the first video corresponding to the target object.
  8. 一种图像处理装置,包括:An image processing device, including:
    第一接收模块,用于在显示目标图像的情况下,接收用户对所述目标图像中的目标对象的第一输入;The first receiving module is configured to receive the user's first input of the target object in the target image when the target image is displayed;
    第一播放模块,用于响应于所述第一输入,播放与所述目标对象对应的第一视频;The first play module is configured to play the first video corresponding to the target object in response to the first input;
    其中,所述第一视频为对第二视频中的所述目标对象的表情或动作进行放大处理得到的视频,所述第二视频为在所述目标图像的拍摄过程中针对所述目标对象录制的视频。Wherein, the first video is a video obtained by amplifying the expression or action of the target object in the second video, and the second video is recorded for the target object during the shooting of the target image Video.
  9. 根据权利要求8所述的装置,还包括:The device according to claim 8, further comprising:
    第一截取模块,用于在所述第二视频的多个目标帧包括所述目标对象的情况下,分别从所述多个目标帧中截取区域图像,每个所述区域图像均包括所述目标对象;The first interception module is configured to separately intercept regional images from the multiple target frames when the multiple target frames of the second video include the target object, and each of the regional images includes the target;
    第一拼接模块,用于对多个所述区域图像进行拼接,得到预处理视频;The first splicing module is used for splicing a plurality of the regional images to obtain a pre-processed video;
    第一处理模块,用于对所述预处理视频进行视频动作放大处理,得到所述第一视频。The first processing module is configured to perform video action amplifying processing on the pre-processed video to obtain the first video.
  10. 根据权利要求8所述的装置,还包括:The device according to claim 8, further comprising:
    第一提取模块,用于提取所述第二视频中的至少两张图像,其中,所述至少一张图像的图像内容中包括所述目标对象的微动作或微表情;The first extraction module is configured to extract at least two images in the second video, wherein the image content of the at least one image includes the micro-actions or micro-expressions of the target object;
    第二处理模块,用于对所述至少两张图像中的微动作或者微表情进行视频动作放大处理;The second processing module is configured to perform video action magnification processing on the micro-actions or micro-expressions in the at least two images;
    第三处理模块,用于对进行视频动作放大处理后的所述至少两张图像进行拼接,得到所述第一视频。The third processing module is configured to splice the at least two images after the video action enlargement processing is performed to obtain the first video.
  11. 根据权利要求10所述的装置,还包括:The device according to claim 10, further comprising:
    第二接收模块,用于在所述目标对象包括至少两个对象的情况下,接收用户对所述目标图像中的第一对象的第二输入;A second receiving module, configured to receive a second input of a user on the first object in the target image when the target object includes at least two objects;
    第一获取模块,用于响应于所述第二输入,从所述第一视频中获取所述第一对象对应的视频。The first obtaining module is configured to obtain a video corresponding to the first object from the first video in response to the second input.
  12. 根据权利要求11所述的装置,其中,所述第一获取模块包括:The apparatus according to claim 11, wherein the first obtaining module comprises:
    第一截取子模块,用于在所述第一视频的多个目标帧包括所述第一对象的情况下,从所述多个目标帧中分别截取区域图像,每个所述区域图像均包括所述第一对象;The first interception submodule is configured to separately intercept regional images from the multiple target frames when the multiple target frames of the first video include the first object, and each of the regional images includes The first object;
    第一拼接子模块,用于将所述区域图像进行拼接,得到所述第一对象的视频。The first splicing sub-module is used for splicing the regional images to obtain the video of the first object.
  13. 根据权利要求8所述的装置,还包括:The device according to claim 8, further comprising:
    识别模块,用于对所述目标图像进行识别,得到识别结果;The recognition module is used to recognize the target image to obtain a recognition result;
    第四处理模块,用于根据所述识别结果,突出显示所述目标图像中的对象。The fourth processing module is configured to highlight the object in the target image according to the recognition result.
  14. 根据权利要求8所述的装置,还包括:The device according to claim 8, further comprising:
    第三接收模块,用于接收用户的第三输入;The third receiving module is used to receive the third input of the user;
    第五处理模块,用于响应于所述第三输入,基于所述与所述目标对象对应的第一视频得到所述目标对象对应的动态图像。The fifth processing module is configured to obtain a dynamic image corresponding to the target object based on the first video corresponding to the target object in response to the third input.
  15. 一种电子设备,包括:存储器、处理器及存储在存储器上并可在处理器上运行的程序或指令,所述处理器执行所述程序或指令时实现如权利要求1至7任一项所述的图像处理方法中的步骤。An electronic device comprising: a memory, a processor, and a program or instruction stored on the memory and capable of being run on the processor. The processor executes the program or instruction as described in any one of claims 1 to 7 The steps in the image processing method described above.
  16. 一种可读存储介质,所述可读存储介质上存储程序或指令,所述程序或指令被处理器执行时实现如权利要求1至7任一项所述的图像处理方法中的步骤。A readable storage medium storing a program or instruction on the readable storage medium, and when the program or instruction is executed by a processor, the steps in the image processing method according to any one of claims 1 to 7 are realized.
  17. 一种计算机程序产品,所述计算机程序产品被至少一个处理器执行以实现如权利要求1-7任一项所述的图像处理方法。A computer program product, which is executed by at least one processor to implement the image processing method according to any one of claims 1-7.
  18. 一种图像处理装置,被配置成用于执行如权利要求1-7任一项所述的图像处理方法。An image processing device configured to execute the image processing method according to any one of claims 1-7.
  19. 一种电子设备,被配置成用于执行如权利要求1-7任一项所述的图像处理方法。An electronic device configured to execute the image processing method according to any one of claims 1-7.
  20. 一种芯片,所述芯片包括处理器和通信接口,所述通信接口和所述处理器耦合,所述处理器用于运行程序或指令,实现如权利要求1-7任一项所述的图像处理方法。A chip comprising a processor and a communication interface, the communication interface is coupled with the processor, and the processor is used to run a program or an instruction to implement the image processing according to any one of claims 1-7 method.
PCT/CN2021/101166 2020-06-24 2021-06-21 Image processing method and apparatus, device, and readable storage medium WO2021259185A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010595589.X 2020-06-24
CN202010595589.XA CN111722775A (en) 2020-06-24 2020-06-24 Image processing method, device, equipment and readable storage medium

Publications (1)

Publication Number Publication Date
WO2021259185A1 true WO2021259185A1 (en) 2021-12-30

Family

ID=72569189

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/101166 WO2021259185A1 (en) 2020-06-24 2021-06-21 Image processing method and apparatus, device, and readable storage medium

Country Status (2)

Country Link
CN (1) CN111722775A (en)
WO (1) WO2021259185A1 (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111722775A (en) * 2020-06-24 2020-09-29 维沃移动通信(杭州)有限公司 Image processing method, device, equipment and readable storage medium
CN112199016B (en) * 2020-09-30 2023-02-21 北京字节跳动网络技术有限公司 Image processing method, image processing device, electronic equipment and computer readable storage medium
CN112653920B (en) * 2020-12-18 2022-05-24 北京字跳网络技术有限公司 Video processing method, device, equipment and storage medium
CN113067983B (en) * 2021-03-29 2022-11-15 维沃移动通信(杭州)有限公司 Video processing method and device, electronic equipment and storage medium
CN114339365A (en) * 2021-12-23 2022-04-12 广东维沃软件技术有限公司 Video playing method and device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150296317A1 (en) * 2014-04-15 2015-10-15 Samsung Electronics Co., Ltd. Electronic device and recording method thereof
CN106504280A (en) * 2016-10-17 2017-03-15 努比亚技术有限公司 A kind of method and terminal for browsing video
CN110301136A (en) * 2017-02-17 2019-10-01 Vid拓展公司 The system and method for selective object of interest scaling are carried out in streamed video
CN110933488A (en) * 2018-09-19 2020-03-27 传线网络科技(上海)有限公司 Video editing method and device
CN111722775A (en) * 2020-06-24 2020-09-29 维沃移动通信(杭州)有限公司 Image processing method, device, equipment and readable storage medium

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20140101394A (en) * 2011-12-28 2014-08-19 인텔 코포레이션 Systems, methods and apparatus for providing content based on a collection of images
CN104917666B (en) * 2014-03-13 2019-08-06 腾讯科技(深圳)有限公司 A kind of method and apparatus making personalized dynamic expression
US11042588B2 (en) * 2014-04-24 2021-06-22 Nokia Technologies Oy Apparatus, method, and computer program product for video enhanced photo browsing
JP2018517984A (en) * 2015-06-15 2018-07-05 トムソン ライセンシングThomson Licensing Apparatus and method for video zoom by selecting and tracking image regions
CN105578275A (en) * 2015-12-16 2016-05-11 小米科技有限责任公司 Video display method and apparatus
CN105718198A (en) * 2016-01-20 2016-06-29 广东欧珀移动通信有限公司 Generation method of dynamic image and mobile terminal
EP3465684A1 (en) * 2016-05-27 2019-04-10 IMINT Image Intelligence AB System and method for a zoom function

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150296317A1 (en) * 2014-04-15 2015-10-15 Samsung Electronics Co., Ltd. Electronic device and recording method thereof
CN106504280A (en) * 2016-10-17 2017-03-15 努比亚技术有限公司 A kind of method and terminal for browsing video
CN110301136A (en) * 2017-02-17 2019-10-01 Vid拓展公司 The system and method for selective object of interest scaling are carried out in streamed video
CN110933488A (en) * 2018-09-19 2020-03-27 传线网络科技(上海)有限公司 Video editing method and device
CN111722775A (en) * 2020-06-24 2020-09-29 维沃移动通信(杭州)有限公司 Image processing method, device, equipment and readable storage medium

Also Published As

Publication number Publication date
CN111722775A (en) 2020-09-29

Similar Documents

Publication Publication Date Title
WO2021259185A1 (en) Image processing method and apparatus, device, and readable storage medium
US11114130B2 (en) Method and device for processing video
CN111541845B (en) Image processing method and device and electronic equipment
WO2019134516A1 (en) Method and device for generating panoramic image, storage medium, and electronic apparatus
EP3547218B1 (en) File processing device and method, and graphical user interface
WO2021238943A1 (en) Gif picture generation method and apparatus, and electronic device
WO2021136136A1 (en) Screenshot method and electronic device
CN111432265B (en) Method for processing video pictures, related device and storage medium
US20150332091A1 (en) Device and method of processing image
WO2021073579A1 (en) Method for capturing scrolling screenshot and terminal device
EP3975046B1 (en) Method and apparatus for detecting occluded image and medium
JP2023554519A (en) Electronic document editing method and device, computer equipment and program
CN112911147B (en) Display control method, display control device and electronic equipment
CN112672061B (en) Video shooting method and device, electronic equipment and medium
US12019669B2 (en) Method, apparatus, device, readable storage medium and product for media content processing
WO2023284632A1 (en) Image display method and apparatus, and electronic device
CN111726676B (en) Image generation method, display method, device and equipment based on video
JP2024501558A (en) Display control methods, devices, electronic devices and media
CN112954046A (en) Information sending method, information sending device and electronic equipment
CN113794831B (en) Video shooting method, device, electronic equipment and medium
WO2022134390A1 (en) Labeling method and apparatus, electronic device, and storage medium
KR20230061519A (en) Screen capture methods, devices and electronics
WO2023011300A1 (en) Method and apparatus for recording facial expression of video viewer
WO2022151687A1 (en) Group photo image generation method and apparatus, device, storage medium, computer program, and product
CN112165584A (en) Video recording method, video recording device, electronic equipment and readable storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21829860

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21829860

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 21829860

Country of ref document: EP

Kind code of ref document: A1