CN113077470B - Method, system, device and medium for cutting horizontal and vertical screen conversion picture - Google Patents

Method, system, device and medium for cutting horizontal and vertical screen conversion picture Download PDF

Info

Publication number
CN113077470B
CN113077470B CN202110324841.8A CN202110324841A CN113077470B CN 113077470 B CN113077470 B CN 113077470B CN 202110324841 A CN202110324841 A CN 202110324841A CN 113077470 B CN113077470 B CN 113077470B
Authority
CN
China
Prior art keywords
picture
video
determining
key
dynamic information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110324841.8A
Other languages
Chinese (zh)
Other versions
CN113077470A (en
Inventor
曾荣
徐蕾
吴三阳
王伟
陆赞信
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
iMusic Culture and Technology Co Ltd
Original Assignee
iMusic Culture and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by iMusic Culture and Technology Co Ltd filed Critical iMusic Culture and Technology Co Ltd
Priority to CN202110324841.8A priority Critical patent/CN113077470B/en
Publication of CN113077470A publication Critical patent/CN113077470A/en
Application granted granted Critical
Publication of CN113077470B publication Critical patent/CN113077470B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The invention provides a method, a system, a device and a storage medium for cutting a horizontal and vertical screen conversion picture, wherein the method comprises the following steps: acquiring a horizontal screen video file, and separating the horizontal screen video file to obtain a video picture and an audio file; acquiring a face position in a video picture, determining a face image, and segmenting the face image to obtain dynamic information; matching with the audio file according to the dynamic information, and determining key characters according to a matching result; the method separates the video picture and the audio file, divides the dynamic information of the human face in the video picture, and performs characteristic matching on the dynamic information obtained by division and the audio file to locate the speaker in the video, and divides the key character and the non-key character on a section of video containing a plurality of pictures, thereby keeping the picture of the key character as much as possible and cutting the picture of the non-key character to obtain better cutting effect, and can be widely applied to the technical field of video processing.

Description

Method, system, device and medium for cutting horizontal and vertical screen conversion picture
Technical Field
The invention relates to the technical field of video processing, in particular to a method, a system, a device and a storage medium for cutting a horizontal and vertical screen conversion picture.
Background
In video products, two types of video forms exist, namely, a complete video form: horizontal version video and vertical version video. Generally, the content of the horizontal version video is higher in richness and more in content types, and due to the time length, a common user is used to watch the mobile phone transversely to obtain better film watching experience; the vertical video is more interesting, shorter in time and more focused and concise in plot. Users are generally more inclined to view in portrait situations at more fragmented times and in scenes. Under the condition of not carrying out any processing, the video picture is directly zoomed, the vertical version video is displayed under the horizontal screen state or the horizontal version video is displayed under the vertical screen state, and a large amount of black edges appear.
At present, the method for automatically converting horizontal and vertical screen videos through a video processing technology is a high cost performance method, the method has good balance between cost and film viewing experience, massive videos can be automatically converted with extremely low cost, complete information of the videos can be kept as much as possible through the video processing technology, and better film viewing experience is provided for users.
But in some more specific cases, for example, multiple faces appear in a picture of a tv show or movie and a long period of conversation is conducted; if the key area of the video is determined only according to the position of the face during the cropping process, confusion occurs, for example, the person in the cropped picture does not speak, but the speaking person does not appear in the cropped picture; or the area of the key elements in the video is too large, so that the clipping algorithm balances the occupation of a plurality of key elements, and a clipping effect with worse quality is formed.
Disclosure of Invention
In view of the above, to at least partially solve one of the above technical problems, an embodiment of the present invention is directed to a method for clipping a horizontal-vertical screen conversion picture, which can retain the picture of a key character as much as possible and clip the picture of a non-key character to obtain a better clipping effect; meanwhile, the embodiment also provides a system, a device and a computer readable storage medium for correspondingly realizing the method.
In a first aspect, a technical solution of the present application provides a method for clipping a horizontal-vertical screen conversion picture, which includes the steps of:
acquiring a horizontal screen video file, and separating the horizontal screen video file to obtain a video picture and an audio file;
acquiring the face position in the video picture, determining a face image, and segmenting the face image to obtain dynamic information;
matching the dynamic information with the audio file, and determining key figures according to matching results;
and cutting the video picture according to the key character to obtain a vertical screen picture.
In a feasible embodiment of the present application, the obtaining a face position in the video frame, determining a face image, and segmenting the face image to obtain dynamic information includes:
determining the face image according to the face position, generating a gray level image of the face image, and extracting a feature map of the gray level image;
predicting to obtain key points according to the gray level image, generating a similarity transformation matrix of the key points, and determining a key point hot spot diagram according to the similarity transformation matrix;
and iterating and determining the dynamic information of the human face through a feedforward neural network according to the gray level image, the feature map and the key point hot spot map.
In a possible embodiment of the present disclosure, the matching the dynamic information with the audio file and determining a key person according to a matching result include:
coding the dynamic information to obtain a first feature vector;
coding the audio file to obtain a second feature vector;
splicing the first feature vector and the second feature vector, and outputting a matching probability through a convolutional neural network;
and determining the face position with the highest score in the matching probability as the key figure.
In a possible embodiment of the present disclosure, the encoding according to the dynamic information to obtain a first feature vector includes:
determining an average value of pixels in the dynamic information through principal component analysis;
calculating a covariance matrix of the average value of the pixels, and determining a pixel eigenvector according to the covariance matrix;
and projecting the dynamic information according to the pixel characteristic vector to obtain the first characteristic vector.
In a possible embodiment of the present disclosure, the encoding the audio file to obtain a second feature vector includes:
increasing the high-frequency part of the audio file to obtain a pre-emphasis signal, and framing the pre-emphasis signal to obtain a single-frame signal;
windowing the single-frame signal, and obtaining a frequency domain signal through fast Fourier transform;
and determining the energy of the frequency domain signal through a Mel filter bank, determining a Mel frequency cepstrum coefficient, and obtaining a second eigenvector through the Mel frequency cepstrum coefficient.
In a feasible embodiment of the present disclosure, the cropping the video picture according to the key character to obtain a portrait picture includes:
determining that the key character does not exist in the horizontal screen video file, keeping the picture height in the horizontal screen video file unchanged, determining the picture width according to a preset picture proportion, and determining a cutting area according to the picture height and the picture width;
and clipping according to the clipping area to obtain the vertical screen picture.
In a feasible embodiment of the present disclosure, the cropping the video picture according to the key character to obtain a portrait picture includes:
determining that the key character exists in the horizontal screen video file, and determining the cropping area;
and controlling the cutting area to slide in the picture of the horizontal screen video file, determining that the face of the key figure is positioned in the cutting area, and cutting the picture to obtain the vertical screen picture.
In a second aspect, a technical solution of the present invention further provides a software system for clipping a horizontal screen and a vertical screen transition picture, including: the system comprises a sound and picture separation module, a video and audio file acquisition module and a video and audio file acquisition module, wherein the sound and picture separation module is used for acquiring a horizontal screen video and audio file and separating the horizontal screen video and audio file from the horizontal screen video and audio file;
the characteristic cutting module is used for acquiring the face position in the video picture, determining a face image and segmenting the face image to obtain dynamic information;
the character matching module is used for matching the dynamic information with the audio file and determining key characters according to a matching result;
and the picture cutting module is used for cutting the video picture according to the key character to obtain a vertical screen picture.
In a third aspect, the present invention further provides a clipping device for horizontally and vertically switching pictures, including:
at least one processor;
at least one memory for storing at least one program;
when the at least one program is executed by the at least one processor, the at least one processor executes the cropping method of the landscape and portrait screen conversion picture in the first aspect.
In a fourth aspect, the present invention also provides a storage medium, in which a processor-executable program is stored, and the processor-executable program is used for executing the method in the first aspect when being executed by a processor.
Advantages and benefits of the present invention will be set forth in part in the description which follows, and in part will be obvious from the description, or may be learned by practice of the invention:
according to the technical scheme, the video picture and the audio file are separated, the dynamic information of the face is segmented in the video picture, the dynamic information obtained by segmentation is subjected to feature matching with the audio file, the speaker in the video is positioned, and the key character and the non-key character are distinguished on a section of video containing a multi-person picture, so that the picture of the key character can be kept as far as possible, the picture of the non-key character is cut, and a better cutting effect is obtained.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present application, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a flowchart illustrating steps of a method for cropping a horizontal/vertical screen transition picture according to an embodiment of the present disclosure;
FIG. 2 is a schematic view of a landscape screen before cropping in an embodiment of the present application;
FIG. 3 is a schematic diagram of a clipped portrait screen in an embodiment of the present application.
Detailed Description
Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like or similar reference numerals refer to the same or similar elements or elements having the same or similar function throughout. The embodiments described below with reference to the accompanying drawings are illustrative only for the purpose of explaining the present invention, and are not to be construed as limiting the present invention. The step numbers in the following embodiments are provided only for convenience of illustration, the order between the steps is not limited at all, and the execution order of each step in the embodiments can be adapted according to the understanding of those skilled in the art.
In the embodiments of the present application, the horizontal video file or the video frame mentioned in the embodiments refers to a case where the frame size of the frame is larger than the frame width, for example, the ratio is 21:9, 16:9, etc.; the vertical screen video file or video picture refers to the situation that the picture length of the picture is smaller than the picture width, such as 9:16, 3:4, and the like.
In a first aspect, as shown in fig. 1, the technical solution of the present application provides an embodiment of a method for clipping a horizontal-vertical screen conversion picture, where the method includes steps S100-S400:
s100, acquiring the horizontal screen video file, and separating the horizontal screen video file to obtain a video picture and an audio file.
Specifically, an audio file is derived from the acquired cross screen video and audio through the existing video clipping method or video clipping tool, and the original video file is directly used as a video for face changing to perform subsequent processing. Typically, the exported audio file is an MP3 format file.
S200, acquiring a face position in a video picture, determining a face image, and segmenting the face image to obtain dynamic information;
the dynamic information refers to a part of the human face, such as lips, cheeks, and chin, which is obviously different from the beginning and the end of speaking. Specifically, the face position of each frame in the video is obtained through a face recognition algorithm. In the embodiment, YOLOv3(You Only Look one) is adopted to extract the face position, YOLOv3 is an object detection algorithm based on a deep convolutional neural network, and a high-precision face detection model is obtained by public face data training; in practice WIDER FACE (a face detection reference data set) is used for training. Carrying out frame-by-frame identification on the face in the video picture by training the obtained face detection model; after the face position in the picture is determined, extracting a face image at the position, namely the face image contains face information, registering face key points of the face image through a DAN (deep Alignment network) algorithm, connecting key points of a chin, a lip and two cheek parts through a key point rule to obtain a face key point combination in each frame of picture, and collecting the key point combinations of each frame of video picture to obtain dynamic information. The embodiment determines the character of the person speaking in each frame of the video through the dynamic information of the face. In addition, besides the dynamic information obtained by dividing the face image, the remaining key points are used as static information of the face image, that is, the static information is partial features of the face with little difference between the start and stop of speaking, such as head contour, ears, nose, forehead, eyebrows, etc.; when the character of the person speaking in each frame of the video is determined through the dynamic information of the face, the face information can also be corrected through the static information.
S300, matching the dynamic information with the audio file, and determining key characters according to a matching result;
specifically, the dynamic information of each face in each frame of image and the corresponding audio are input into the discrimination model, and the result of matching the dynamic information with the audio is obtained through the discrimination model, for example, when the matching result is a specific score, the face corresponding to the dynamic information is determined to be a key character, i.e., a speaker, in the retaining wall frame image according to the highest score of matching between the audio and the face.
S400, clipping the video picture according to the key characters to obtain a vertical screen picture.
Specifically, after key characters in each frame of video pictures are determined, clipping is carried out on the original video pictures according to the size of a preset clipping picture, the clipping process is repeated, the video obtained after clipping of each frame is spliced, a converted portrait video is generated, and original audio files are added according to the time axis of the video pictures; it will be appreciated that the cropped picture size should be a picture-to-picture ratio that satisfies the portrait screen.
In some possible embodiments, the step S200 of acquiring a face position in a video frame, determining a face image, and segmenting the face image to obtain dynamic information may further include the step S210-S230 of subdividing:
s210, determining a face image according to the face position, generating a gray level image of the face image, and extracting a feature map of the gray level image;
specifically, gray level processing is carried out on a face image determined according to the face position to obtain a radian image, and a characteristic diagram of the gray level image is extracted according to a Feed Forward NN (Feed Forward NN) in a DAN network architecture; the calculation process of the feature map comprises the following steps: the feature matrix of 1 × 3136 is output, a matrix transformation is performed to obtain a 56 × 56 matrix, and then up-sampling is performed to obtain a matrix of the same size as the input grayscale image, for example, 112 × 112.
S220, obtaining key points according to the gray level image prediction, generating a similarity transformation matrix of the key points, and determining a key point hot spot diagram according to the similarity transformation matrix;
specifically, the positions of a first batch of key points are obtained through prediction by a standard key point template provided by a DAN network architecture, and a similar change matrix from the first batch of key points to the standard key point template is obtained through calculation of Connection Layers (connections Layers) in the DAN network architecture; the input gray level image can be corrected through the similarity change matrix, and the key points are transformed to obtain a key point hot spot diagram, and the same is true. The feature map is obtained by feature extraction through a next layer of Feed Forward NN, wherein the hot spot map of the key points is a calculation mode of central attenuation, the values at the key points are maximum, and the values are smaller when the key points are farther away.
And S230, iterating through a feedforward neural network according to the gray level image, the feature map and the key point hot spot map to determine dynamic information of the face.
Specifically, a gray image, a feature map and a key point hot spot map are input to a Feed Forward NN for iteration to obtain a new key point hot spot map, after the iteration is finished, the key points of the face in the gray image are obtained, and the key points in each frame of video picture are combined to obtain the dynamic information of the face.
In some optional embodiments, step S300, matching with the audio file according to the dynamic information, and determining the key character according to the matching result includes more subdivided steps S310-S330:
s310, coding the dynamic information to obtain a first feature vector;
specifically, in the embodiment, the dynamic information may be encoded into a real number vector with a length of 1024 by an encoder through an encoder of the dynamic information of the face and inputting the dynamic information of the face image, that is, the first feature vector.
S320, coding the audio file to obtain a second feature vector;
specifically, the audio file is divided into frames according to a time axis, each frame of the audio is subjected to feature extraction by using an MFCC (Mel-Frequency cepstrum coefficient) algorithm, and features output by the MFCC are encoded by an encoder to obtain a real number vector with 1024 lengths, namely a second feature vector.
S330, splicing the first feature vector and the second feature vector, and outputting the matching probability through a convolutional neural network; determining the face position with the highest score in the matching probability as a key figure;
specifically, the feature vectors obtained in the steps S310 and S320 are spliced to form a vector of 2048 length, and the probability of yes or no prediction result is output through the trained convolutional neural network.
In some possible embodiments, the step of encoding according to the dynamic information to obtain the first feature vector S310 may include more detailed steps S311 to S313:
s311, determining the average value of the pixels in the dynamic information through principal component analysis;
specifically, in this embodiment, through Principal Component Analysis (PCA), an average value of each pixel of all the dynamic information is first calculated to obtain an average value matrix, where the number of rows of the matrix is 1, and the number of columns of the matrix is M, where M is the image size of the dynamic information.
S312, calculating a covariance matrix of the average value of the pixels, and determining a pixel feature vector according to the covariance matrix;
specifically, the rows of the matrix obtained in step S311 are copied, so that the number of rows of the matrix is consistent with the number N of samples of the dynamic information, and each pixel according to the dynamic information is converted into an original data matrix of N rows and M columns, where M is the image size of the dynamic information. And calculating to obtain a covariance matrix through the original data matrix and the mean matrix. And calculating an eigenvalue matrix and an eigenvector matrix by using the covariance matrix, and sequencing the eigenvalues from large to small. The eigenvalue matrix is N × 1, the eigenvector matrix is N × N, and each row represents an eigenvector corresponding to a certain eigenvalue. And normalized for each row of the feature vector matrix.
S313, projecting the dynamic information according to the pixel characteristic vector to obtain a first characteristic vector;
specifically, according to the normalized eigenvector matrix, dynamic information pixels of each face image of the dynamic information are projected to a PCA space, and a projection vector of a single face image is obtained, namely the first eigenvector.
In some possible embodiments, the step of encoding the audio file into the second feature vector S320 may include the more subdivided steps S321-S323:
s321, improving the high-frequency part of the audio file to obtain a pre-emphasis signal, and framing the pre-emphasis signal to obtain a single-frame signal;
specifically, the voice signal in the audio file is preprocessed, and the voice signal is pre-emphasized through a high-pass filter, wherein the pre-emphasis is to improve the high-frequency part to flatten the spectrum of the signal, so that the spectrum can be obtained by the same signal-to-noise ratio in the whole frequency band from low frequency to high frequency. Meanwhile, the method is also used for eliminating the vocal cords and lip effects in the generation process, compensating the high-frequency part of the voice signal which is restrained by the pronunciation system, and highlighting the formants of the high frequency. Then the signal is framed, and n (n is a positive integer) sampling points are grouped into an observation unit, which is called a frame. In the embodiment, N is 256 or 512, and covers about 20-30 ms. To avoid excessive variation between two adjacent frames, an overlap region is formed between two adjacent frames, where the overlap region includes m sampling points, typically m is about 1/2 or 1/3 of n.
S322, windowing the single-frame signal, and obtaining a frequency domain signal through fast Fourier transform;
specifically, since speech is constantly changing over a long range and cannot be processed without fixed characteristics, a signal tile of each frame is substituted into a window function, and the value outside the window is set to 0, so as to eliminate signal discontinuity which may be caused at both ends of each frame. The commonly used window functions include a square window, a hamming window, a hanning window, and the like, and in the embodiment, the hamming window is adopted according to the frequency domain characteristics of the window function. Each frame is multiplied by a hamming window to increase the continuity of the left and right ends of the frame. Since the signal is usually difficult to see by the transformation in the time domain, it is usually observed by transforming it into an energy distribution in the frequency domain, and different energy distributions can represent the characteristics of different voices. After multiplication by the hamming window, each frame must also undergo a fast fourier transform to obtain the energy distribution over the spectrum. And carrying out fast Fourier transform on each frame signal subjected to framing and windowing to obtain the frequency spectrum of each frame. And the power spectrum of the voice signal is obtained by taking the modulus square of the frequency spectrum of the voice signal.
S323, determining the energy of the frequency domain signal through a Mel filter bank, determining a Mel frequency cepstrum coefficient, and obtaining a second eigenvector through the Mel frequency cepstrum coefficient.
Specifically, an energy spectrum is obtained through calculation according to the power spectrum, the energy spectrum passes through a group of Mel-scale triangular filter banks, and a filter bank with M filters is defined, wherein the number of the filters is close to that of critical bands, and in the embodiment, the adopted filter is a triangular filter. The frequency spectrum is smoothed through filtering, the effect of harmonic waves is eliminated, and the formants of the original voice are highlighted. A speech recognition system featuring MFCC is not affected by the difference in pitch of the input speech. And then calculating the logarithmic energy output by each filter bank, obtaining an MFCC coefficient through discrete cosine transform, and coding through a coder to obtain a second eigenvector.
In some possible embodiments, the step S400 of cropping the video frame according to the key character to obtain the portrait frame may include the steps S410 or S420:
s410, determining that key characters do not exist in the horizontal screen video file, keeping the picture height in the horizontal screen video file unchanged, determining the picture width according to a preset picture proportion, determining a cutting area according to the picture height and the picture width, and cutting according to the cutting area to obtain a vertical screen picture;
or the like, or, alternatively,
s420, determining key characters in the horizontal screen video file, determining a cutting area, controlling the cutting area to slide in the picture of the horizontal screen video file, determining the face of the key character to be located in the cutting area, and cutting the picture to obtain a vertical screen picture.
Specifically, the size of the cropping area is determined first, the size of the cropping area is determined by the original video resolution and the cropping target proportion, the cropping principle is to keep the pictures of the original video as much as possible, for example, the original video resolution is 1920 × 1080 (width: height), and the cropping target proportion is wide: when the video is 9:16 high, in order to keep the original video pictures as many as possible, the height is not cut, and the width is cut to 1920 × 1080/16 × 9 (607.5), so that the size of the cut area is 607.5 × 1080 (width: high);
as shown in fig. 2, the horizontal screen image before clipping is a schematic diagram, clipping is performed according to the determined key speaker, in the clipping process, in the embodiment, a clipping frame of 607.5:1080 is slid from left to back on an original image of 1920:1080, when the face of the key speaker is in the middle of the clipping frame, the position of the clipping frame is the required clipping position, as shown in fig. 3, the clipping is a schematic diagram of the vertical screen image obtained after clipping.
In a second aspect, the present application provides a cropping system for a landscape-portrait screen transformation picture, which is used in the method in the first aspect, and includes:
the sound and picture separation module is used for acquiring the horizontal screen video and audio file and separating the horizontal screen video and audio file to obtain a video picture and an audio file;
the characteristic cutting module is used for acquiring the face position in the video picture, determining a face image and segmenting the face image to obtain dynamic information;
the character matching module is used for matching the dynamic information with the audio file and determining key characters according to a matching result;
and the picture cutting module is used for cutting the video picture according to the key character to obtain a vertical screen picture.
In a third aspect, the present application further provides a clipping device for horizontal and vertical screen conversion pictures, which includes at least one processor; at least one memory for storing at least one program; when the at least one program is executed by the at least one processor, the at least one processor executes a cropping method of a landscape-portrait screen converted picture as in the first aspect.
An embodiment of the present invention further provides a storage medium storing a program, where the program is executed by a processor to implement the method in the first aspect.
From the above specific implementation process, it can be concluded that the technical solution provided by the present invention has the following advantages or advantages compared to the prior art:
the invention provides a technical scheme for clipping a horizontal and vertical screen conversion picture based on speaker positioning, and aims to distinguish key characters from non-key characters on a section of video containing a multi-person picture by positioning speakers in the video, so that the picture of the key characters can be kept as far as possible and the picture of the non-key characters can be clipped, and a better clipping effect can be obtained.
In alternative embodiments, the functions/acts noted in the block diagrams may occur out of the order noted in the operational illustrations. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality/acts involved. Furthermore, the embodiments presented and described in the flow charts of the present invention are provided by way of example in order to provide a more thorough understanding of the technology. The disclosed methods are not limited to the operations and logic flows presented herein. Alternative embodiments are contemplated in which the order of various operations is changed and in which sub-operations described as part of larger operations are performed independently.
Furthermore, although the present invention is described in the context of functional modules, it should be understood that, unless otherwise stated to the contrary, one or more of the functions and/or features may be integrated in a single physical device and/or software module, or one or more of the functions and/or features may be implemented in a separate physical device or software module. It will also be appreciated that a detailed discussion of the actual implementation of each module is not necessary for an understanding of the present invention. Rather, the actual implementation of the various functional modules in the apparatus disclosed herein will be understood within the ordinary skill of an engineer, given the nature, function, and internal relationship of the modules. Accordingly, those skilled in the art can, using ordinary skill, practice the invention as set forth in the claims without undue experimentation. It is also to be understood that the specific concepts disclosed are merely illustrative of and not intended to limit the scope of the invention, which is defined by the appended claims and their full scope of equivalents.
The logic and/or steps represented in the flowcharts or otherwise described herein, e.g., an ordered listing of executable instructions that can be considered to implement logical functions, can be embodied in any computer-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions.
In the description herein, references to the description of the term "one embodiment," "some embodiments," "an example," "a specific example," or "some examples," etc., mean that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the invention. In this specification, the schematic representations of the terms used above do not necessarily refer to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples.
While embodiments of the invention have been shown and described, it will be understood by those of ordinary skill in the art that: various changes, modifications, substitutions and alterations can be made to the embodiments without departing from the principles and spirit of the invention, the scope of which is defined by the claims and their equivalents.
While the preferred embodiments of the present invention have been illustrated and described, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (9)

1. A method for cutting a horizontal and vertical screen conversion picture is characterized by comprising the following steps:
acquiring a horizontal screen video file, and separating the horizontal screen video file to obtain a video picture and an audio file;
acquiring the face position in the video picture, determining a face image, and segmenting the face image to obtain dynamic information;
matching the dynamic information with the audio file, and determining key figures according to matching results;
clipping the video picture according to the key character to obtain a vertical screen picture;
the acquiring of the face position in the video picture, determining the face image, and segmenting the face image to obtain dynamic information includes:
determining the face image according to the face position, generating a gray level image of the face image, and extracting a feature map of the gray level image;
predicting to obtain key points according to the gray level image, generating a similarity transformation matrix of the key points, and determining a key point hot spot diagram according to the similarity transformation matrix;
and iterating and determining the dynamic information of the human face through a feedforward neural network according to the gray level image, the feature map and the key point hot spot map.
2. The method for clipping horizontal and vertical screen transition picture according to claim 1, wherein the matching with the audio file according to the dynamic information and the determination of key characters according to the matching result comprise:
coding the dynamic information to obtain a first feature vector;
coding the audio file to obtain a second feature vector;
splicing the first feature vector and the second feature vector, and outputting a matching probability through a convolutional neural network; and determining the face position with the highest score in the matching probability as the key figure.
3. The method of claim 2, wherein the encoding according to the motion information to obtain a first feature vector comprises:
determining an average value of pixels in the dynamic information through principal component analysis;
calculating a covariance matrix of the average value of the pixels, and determining a pixel eigenvector according to the covariance matrix;
and projecting the dynamic information according to the pixel characteristic vector to obtain the first characteristic vector.
4. The method of claim 2, wherein the encoding the audio file to obtain the second feature vector comprises:
increasing the high-frequency part of the audio file to obtain a pre-emphasis signal, and framing the pre-emphasis signal to obtain a single-frame signal;
windowing the single-frame signal, and obtaining a frequency domain signal through fast Fourier transform;
and determining the energy of the frequency domain signal through a Mel filter bank, determining a Mel frequency cepstrum coefficient, and obtaining a second eigenvector through the Mel frequency cepstrum coefficient.
5. The method for cropping a horizontal/vertical screen converted picture according to any one of claims 1-4, wherein the cropping the video picture according to the key character to obtain a vertical screen picture comprises:
determining that the key character does not exist in the transverse screen video file, and keeping the picture height in the transverse screen video file unchanged;
determining the width of a picture according to a preset picture proportion, and determining a cutting area according to the height of the picture and the width of the picture;
and clipping according to the clipping area to obtain the vertical screen picture.
6. The method for cropping a horizontal-vertical screen converted picture according to claim 5, wherein the cropping the video picture according to the key character to obtain a vertical screen picture comprises:
determining that the key character exists in the horizontal screen video file, and determining the cropping area;
and controlling the cutting area to slide in the picture of the horizontal screen video file, determining that the face of the key figure is positioned in the cutting area, and cutting the picture to obtain the vertical screen picture.
7. A system of tailorring of horizontal vertical screen conversion picture, its characterized in that includes:
the system comprises a sound and picture separation module, a video and audio file acquisition module and a video and audio file acquisition module, wherein the sound and picture separation module is used for acquiring a horizontal screen video and audio file and separating the horizontal screen video and audio file from the horizontal screen video and audio file;
the characteristic cutting module is used for acquiring the face position in the video picture, determining a face image and segmenting the face image to obtain dynamic information;
the character matching module is used for matching the dynamic information with the audio file and determining key characters according to a matching result;
the picture cutting module is used for cutting the video picture according to the key character to obtain a vertical screen picture;
the acquiring the face position in the video picture, determining a face image, and segmenting the face image to obtain dynamic information includes:
determining the face image according to the face position, generating a gray level image of the face image, and extracting a feature map of the gray level image;
predicting to obtain key points according to the gray level image, generating a similarity transformation matrix of the key points, and determining a key point hot spot diagram according to the similarity transformation matrix;
and iterating and determining the dynamic information of the human face through a feedforward neural network according to the gray level image, the feature map and the key point hot spot map.
8. A cutting device for horizontal and vertical screen conversion pictures is characterized by comprising:
at least one processor;
at least one memory for storing at least one program;
when executed by the at least one processor, cause the at least one processor to perform a method of cropping a landscape or portrait screen conversion screen as defined in any one of claims 1 to 6.
9. A storage medium having stored therein a processor-executable program, wherein the processor-executable program is configured to execute a cropping method of a landscape screen and portrait screen conversion screen according to any one of claims 1 to 6 when executed by a processor.
CN202110324841.8A 2021-03-26 2021-03-26 Method, system, device and medium for cutting horizontal and vertical screen conversion picture Active CN113077470B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110324841.8A CN113077470B (en) 2021-03-26 2021-03-26 Method, system, device and medium for cutting horizontal and vertical screen conversion picture

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110324841.8A CN113077470B (en) 2021-03-26 2021-03-26 Method, system, device and medium for cutting horizontal and vertical screen conversion picture

Publications (2)

Publication Number Publication Date
CN113077470A CN113077470A (en) 2021-07-06
CN113077470B true CN113077470B (en) 2022-01-18

Family

ID=76610693

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110324841.8A Active CN113077470B (en) 2021-03-26 2021-03-26 Method, system, device and medium for cutting horizontal and vertical screen conversion picture

Country Status (1)

Country Link
CN (1) CN113077470B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113645483A (en) * 2021-08-20 2021-11-12 珠海九松科技有限公司 Cross-platform automatic video editing method
CN114257762B (en) * 2021-12-20 2024-09-13 咪咕音乐有限公司 Video conversion method, device, equipment and storage medium
CN114419716B (en) * 2022-01-26 2024-03-15 北方工业大学 Calibration method for face image face key point calibration
CN115018706A (en) * 2022-05-31 2022-09-06 新华智云科技有限公司 Method for converting horizontal screen video material into vertical screen video material

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103412647A (en) * 2013-08-13 2013-11-27 广东欧珀移动通信有限公司 Face recognition page display control method and mobile terminal
CN108234763A (en) * 2017-12-29 2018-06-29 广州优视网络科技有限公司 Horizontal/vertical screen switching method, device and terminal
CN109309808A (en) * 2017-07-26 2019-02-05 梭维智能科技(深圳)有限公司 A kind of monitoring system and method based on recognition of face
CN109376603A (en) * 2018-09-25 2019-02-22 北京周同科技有限公司 A kind of video frequency identifying method, device, computer equipment and storage medium

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI518675B (en) * 2013-08-15 2016-01-21 中華電信股份有限公司 A method for segmenting videos and audios into clips using speaker recognition
WO2018106213A1 (en) * 2016-12-05 2018-06-14 Google Llc Method for converting landscape video to portrait mobile layout
CN107483843B (en) * 2017-08-16 2019-11-15 成都品果科技有限公司 Audio-video matches clipping method and device
CN108650542B (en) * 2018-05-09 2022-02-01 腾讯科技(深圳)有限公司 Method for generating vertical screen video stream and processing image, electronic equipment and video system
CN110544481B (en) * 2019-08-27 2022-09-20 华中师范大学 S-T classification method and device based on voiceprint recognition and equipment terminal
CN110909613B (en) * 2019-10-28 2024-05-31 Oppo广东移动通信有限公司 Video character recognition method and device, storage medium and electronic equipment
CN111212245B (en) * 2020-01-15 2022-03-25 北京猿力未来科技有限公司 Method and device for synthesizing video
CN111556254B (en) * 2020-04-10 2021-04-02 早安科技(广州)有限公司 Method, system, medium and intelligent device for video cutting by using video content
CN111914811B (en) * 2020-08-20 2021-09-28 腾讯科技(深圳)有限公司 Image data processing method, image data processing device, computer equipment and storage medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103412647A (en) * 2013-08-13 2013-11-27 广东欧珀移动通信有限公司 Face recognition page display control method and mobile terminal
CN109309808A (en) * 2017-07-26 2019-02-05 梭维智能科技(深圳)有限公司 A kind of monitoring system and method based on recognition of face
CN108234763A (en) * 2017-12-29 2018-06-29 广州优视网络科技有限公司 Horizontal/vertical screen switching method, device and terminal
CN109376603A (en) * 2018-09-25 2019-02-22 北京周同科技有限公司 A kind of video frequency identifying method, device, computer equipment and storage medium

Also Published As

Publication number Publication date
CN113077470A (en) 2021-07-06

Similar Documents

Publication Publication Date Title
CN113077470B (en) Method, system, device and medium for cutting horizontal and vertical screen conversion picture
CN113192161A (en) Virtual human image video generation method, system, device and storage medium
US7123262B2 (en) Method of animating a synthesized model of a human face driven by an acoustic signal
CN116250036A (en) System and method for synthesizing photo-level realistic video of speech
JP2002507033A (en) Face synthesis device and face synthesis method
CN112001992A (en) Voice-driven 3D virtual human expression sound-picture synchronization method and system based on deep learning
CN114419702A (en) Digital human generation model, training method of model, and digital human generation method
EP4344199A1 (en) Speech and image synchronization measurement method and apparatus, and model training method and apparatus
CN113822968A (en) Method, system and storage medium for driving virtual human in real time by voice
KR102319753B1 (en) Method and apparatus for producing video contents based on deep learning
CN116580720A (en) Speaker vision activation interpretation method and system based on audio-visual voice separation
CN117409121A (en) Fine granularity emotion control speaker face video generation method, system, equipment and medium based on audio frequency and single image driving
CN115578512A (en) Method, device and equipment for training and using generation model of voice broadcast video
CN116013297A (en) Audio-visual voice noise reduction method based on multi-mode gating lifting model
CN115691539A (en) Two-stage voice separation method and system based on visual guidance
CN117593473B (en) Method, apparatus and storage medium for generating motion image and video
CN114255737A (en) Voice generation method and device and electronic equipment
CN110958417B (en) Method for removing compression noise of video call video based on voice clue
CN115438686A (en) Underwater sound target identification method based on data enhancement and residual CNN
Wang et al. Attention-based residual speech portrait model for speech to face generation
Zhou et al. Synthesizing a talking mouth
CN113722513B (en) Multimedia data processing method and equipment
CN118248147B (en) Audio-visual voice recognition method, equipment and storage medium based on self-supervision learning
CN117292437B (en) Lip language identification method, device, chip and terminal
CN114697687B (en) Data processing method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant