CN109862412B - Method and device for video co-shooting and storage medium - Google Patents

Method and device for video co-shooting and storage medium Download PDF

Info

Publication number
CN109862412B
CN109862412B CN201910193873.1A CN201910193873A CN109862412B CN 109862412 B CN109862412 B CN 109862412B CN 201910193873 A CN201910193873 A CN 201910193873A CN 109862412 B CN109862412 B CN 109862412B
Authority
CN
China
Prior art keywords
video
display window
video display
size
template
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910193873.1A
Other languages
Chinese (zh)
Other versions
CN109862412A (en
Inventor
吴娜
黄飞
张龄宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Kugou Computer Technology Co Ltd
Original Assignee
Guangzhou Kugou Computer Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Kugou Computer Technology Co Ltd filed Critical Guangzhou Kugou Computer Technology Co Ltd
Priority to CN201910193873.1A priority Critical patent/CN109862412B/en
Publication of CN109862412A publication Critical patent/CN109862412A/en
Application granted granted Critical
Publication of CN109862412B publication Critical patent/CN109862412B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • User Interface Of Digital Computer (AREA)

Abstract

The application discloses a method, a device and a storage medium for video co-shooting, and belongs to the field of computers. The method comprises the following steps: and acquiring a video close-shot template, wherein the video close-shot template comprises at least two video display windows. And displaying the video close-shot template in the screen display window according to the position description sequence of each video display window and the size of the screen display window, wherein the position description sequences of at least two video display windows are determined after the size of the video close-shot template is normalized. And carrying out video co-shooting through the currently displayed video co-shooting template and the recorded at least one video. By the method, the problem that at least two videos cannot be played simultaneously can be solved, and the video close-shot template can be simply and conveniently displayed in the screen display window because the position description sequence of each video display window is determined after the size of the video close-shot template is normalized.

Description

Method and device for video co-shooting and storage medium
Technical Field
The present application relates to the field of computers, and in particular, to a method, an apparatus, and a storage medium for video co-shooting.
Background
At present, videos are all recorded separately, and when the videos are played, the videos are also played separately. Therefore, the requirement of playing at least two videos at the same time cannot be met, and therefore, a method for taking a photo in time, that is, a method for taking a photo in time with a recorded video is urgently needed, so that the requirement of playing at least two videos at the same time can be met when the taken video is played subsequently.
Disclosure of Invention
The application provides a method, a device and a storage medium for video co-shooting, which can solve the problem that at least two videos cannot be played simultaneously. The technical scheme is as follows:
in a first aspect, a method for taking a video in time is provided, the method comprising:
the method comprises the steps of obtaining a video close-shot template, wherein the video close-shot template comprises at least two video display windows which are used for displaying at least two videos for close-shot;
displaying the video snapshot template in the screen display window according to the position description sequence of each of the at least two video display windows and the size of the screen display window, wherein the position description sequences of the at least two video display windows are determined by normalizing the size of the video snapshot template;
and carrying out video co-shooting through the currently displayed video co-shooting template and the recorded at least one video.
Optionally, the displaying the video close-shot template in the screen display window according to the position description sequence of each of the at least two video display windows and the size of the screen display window includes:
determining the position of each video display window in the screen display window according to the position description sequence of each video display window in the at least two video display windows and the size of the screen display window;
and displaying the video snapshot template in the screen display window according to the position of each video display window in the screen display window.
Optionally, the position description sequence includes at least three elements, which are respectively an abscissa and an ordinate of a position marker point in the video capture template after size normalization, and a size of a video display window relative to the video capture template after size normalization, where the position marker point is a point on the video display window;
determining the position of each video display window in the screen display window according to the position description sequence of each video display window in the at least two video display windows and the size of the screen display window, including:
for a first video display window in the at least two video display windows, determining the abscissa and the ordinate of a first position mark point in the screen display window according to the abscissa and the ordinate of the first position mark point in the video close-shot template after the size normalization and the size of the screen display window, wherein the first position mark point is a point on the first video display window, and the first video display window is any one of the at least two video display windows;
determining the size of the first video display window in the video display window according to the size of the first video display window relative to the video shooting template with the normalized size and the size of the screen display window;
and determining the position of the first video display window in the screen display window according to the abscissa and the ordinate of the first position marking point in the screen display window and the size of the first video display window in the screen display window.
Optionally, the video close-shooting is performed through the currently displayed video close-shooting template and at least one recorded video, and the method includes:
displaying the at least one video in at least one video display window included in the video close-shot template;
and in the process of displaying the at least one video through the at least one video display window, shooting the current video through the remaining video display window except the at least one video display window in the video close-shot template so as to realize close-shot of at least two videos.
Optionally, the method further comprises:
when a dragging operation aiming at a first video display window is detected, determining a dragging ending position, wherein the first video display window is any one of the at least two video display windows;
determining a second video display window according to the dragging ending position, wherein the second video display window is a video display window which is subjected to position exchange with the first video display window;
and exchanging the display position of a first video and the display position of a second video, wherein the first video is the video currently displayed by the first video display window, and the second video is the video currently displayed by the second video display window.
Optionally, exchanging the display position of the first video and the display position of the second video includes:
switching the first video from the first video presentation window to be presented in the second video presentation window, and switching the second video from the second video presentation window to be presented in the first video presentation window; alternatively, the first and second electrodes may be,
and exchanging the position of the first video display window and the position of the second video display window to realize the exchange of the display position of the first video and the display position of the second video.
Optionally, exchanging the position of the first video display window and the position of the second video display window includes:
adjusting the position of the second video display window in the screen display window according to the position description sequence of the first video display window and the size of the screen display window, and re-determining the position description sequence of the adjusted second video display window;
and adjusting the position of the first video display window in the screen display window according to the position description sequence of the second video display window and the size of the screen display window, and re-determining the position description sequence of the adjusted first video display window.
In a second aspect, there is provided an apparatus for co-shooting video, the apparatus comprising:
the system comprises an acquisition module, a storage module and a display module, wherein the acquisition module is used for acquiring a video co-shooting template, the video co-shooting template comprises at least two video display windows, and the at least two video display windows are used for displaying at least two videos for co-shooting;
the display module is used for displaying the video close-shot template in the screen display window according to the position description sequence of each of the at least two video display windows and the size of the screen display window, wherein the position description sequences of the at least two video display windows are determined after the size of the video close-shot template is normalized;
and the close-shooting module is used for carrying out close shooting on the video through the currently displayed video close-shooting template and the recorded at least one video.
Optionally, the display module comprises:
the determining submodule is used for determining the position of each video display window in the screen display window according to the position description sequence of each video display window in the at least two video display windows and the size of the screen display window;
and the first display sub-module is used for displaying the video snapshot template in the screen display window according to the position of each video display window in the screen display window.
Optionally, the position description sequence includes at least three elements, which are respectively an abscissa and an ordinate of a position marker point in the video capture template after size normalization, and a size of a video display window relative to the video capture template after size normalization, where the position marker point is a point on the video display window;
the determination sub-module includes:
a first determining unit, configured to determine, for a first video display window of the at least two video display windows, an abscissa and an ordinate of a first position mark point in the screen display window according to the abscissa and the ordinate of the first position mark point in the video capture template after size normalization and the size of the screen display window, where the first position mark point is a point on the first video display window, and the first video display window is any one of the at least two video display windows;
a second determining unit, configured to determine, according to the size of the first video display window relative to the video snapshot template after size normalization and the size of the screen display window, the size of the first video display window in the video display window;
and the third determining unit is used for determining the position of the first video display window in the screen display window according to the abscissa and the ordinate of the first position marking point in the screen display window and the size of the first video display window in the screen display window.
Optionally, the snap-in module comprises:
the second display submodule is used for displaying the at least one video in at least one video display window included in the video close-shot template;
and the shooting submodule is used for shooting the current video through the remaining video display window except the at least one video display window in the video close-shooting template in the process of displaying the at least one video through the at least one video display window so as to realize close shooting of at least two videos.
Optionally, the apparatus further comprises:
a first determining module, configured to determine a dragging ending position when a dragging operation for a first video display window is detected, where the first video display window is any one of the at least two video display windows;
a second determining module, configured to determine a second video display window according to the dragging ending position, where the second video display window is a video display window that performs position exchange with the first video display window;
the switching module is used for switching the display position of a first video and the display position of a second video, wherein the first video is a video currently displayed by the first video display window, and the second video is a video currently displayed by the second video display window.
Optionally, the switching module comprises:
the switching sub-module is used for switching and displaying the first video from the first video display window into the second video display window and switching and displaying the second video from the second video display window into the first video display window; alternatively, the first and second electrodes may be,
and the switching submodule is used for switching the position of the first video display window and the position of the second video display window so as to realize the switching of the display position of the first video and the display position of the second video.
Optionally, the switch submodule includes:
a fourth determining unit, configured to adjust a position of the second video display window in the screen display window according to the position description sequence of the first video display window and the size of the screen display window, and to re-determine the position description sequence of the adjusted second video display window;
and a fifth determining unit, configured to adjust the position of the first video display window in the screen display window according to the position description sequence of the second video display window and the size of the screen display window, and to re-determine the adjusted position description sequence of the first video display window.
In a third aspect, a computer-readable storage medium is provided, in which a computer program is stored, which, when being executed by a processor, carries out the steps of any of the methods provided in the first aspect.
In a fourth aspect, there is provided a computer program product containing instructions which, when run on a computer, cause the computer to perform the method of taking a video in time as provided in the first aspect.
The beneficial effects brought by the technical scheme provided by the embodiment of the application at least can comprise:
in the embodiment of the application, the video close-up template comprises at least two video display windows, and the position description sequence of each video display window is determined after the size of the video close-up template is normalized. Therefore, when the videos are subjected to close-shot, the close-shot of at least two videos can be realized, the problem that the at least two videos cannot be played simultaneously is solved, and compared with the position description sequence of the video display window which is determined after the size of the video close-shot template is normalized, the position description sequence of the at least two video display windows is simpler, and the subsequent coding is more convenient.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present application, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a flowchart of a first method for taking a video in time according to an embodiment of the present disclosure;
fig. 2 is a flowchart of a second method for taking a video in time according to an embodiment of the present disclosure;
FIG. 3 is a schematic diagram of a first video co-shooting template provided in an embodiment of the present application;
FIG. 4 is a schematic diagram of a second video co-shooting template provided in the embodiment of the present application;
fig. 5 is a schematic structural diagram of a first apparatus for taking a video in time according to an embodiment of the present disclosure;
fig. 6 is a schematic structural diagram of a second device for taking a video in time according to an embodiment of the present disclosure;
fig. 7 is a schematic structural diagram of a terminal according to an embodiment of the present application.
Detailed Description
To make the objects, technical solutions and advantages of the present application more clear, embodiments of the present application will be described in further detail below with reference to the accompanying drawings.
Fig. 1 is a flowchart of a method for snap-shooting a video according to an embodiment of the present application, where the method is applied to a device for snap-shooting a video, and the device for snap-shooting a video may be integrated in a terminal. Referring to fig. 1, the method includes:
step 101: and acquiring a video close-shooting template, wherein the video close-shooting template comprises at least two video display windows, and the at least two video display windows are used for displaying at least two videos for close shooting.
Step 102: and displaying the video close-shot template in the screen display window according to the position description sequence of each of the at least two video display windows and the size of the screen display window, wherein the position description sequences of the at least two video display windows are determined after the size of the video close-shot template is normalized.
Step 103: and carrying out video co-shooting through the currently displayed video co-shooting template and the recorded at least one video.
In the embodiment of the application, the video close-up template comprises at least two video display windows, and the position description sequence of each video display window is determined after the size of the video close-up template is normalized. Therefore, when the videos are subjected to close-shot, the close-shot of at least two videos can be realized, the problem that the at least two videos cannot be played simultaneously is solved, and compared with the position description sequence of the video display window which is determined after the size of the video close-shot template is normalized, the position description sequence of the at least two video display windows is simpler, and the subsequent coding is more convenient.
Optionally, displaying a video snapshot template in the screen presentation window according to the position description sequence of each of the at least two video presentation windows and the size of the screen presentation window, including:
determining the position of each video display window in the screen display window according to the position description sequence of each video display window in the at least two video display windows and the size of the screen display window;
and displaying the video snapshot template in the screen display window according to the position of each video display window in the screen display window.
Optionally, the position description sequence includes at least three elements, which are respectively an abscissa and an ordinate of the position mark point in the video capture template after the size normalization, and a size of the video display window relative to the video capture template after the size normalization, where the position mark point is a point on the video display window;
determining the position of each video display window in the screen display window according to the position description sequence of each video display window in the at least two video display windows and the size of the screen display window, wherein the determining comprises the following steps:
for a first video display window of the at least two video display windows, determining the abscissa and the ordinate of a first position mark point in a screen display window according to the abscissa and the ordinate of the first position mark point in a video close-shot template after size normalization and the size of the screen display window, wherein the first position mark point is a point on the first video display window, and the first video display window is any one of the at least two video display windows;
determining the size of the first video display window in the video display window according to the size of the first video display window relative to the video shooting template with the normalized size and the size of the screen display window;
and determining the position of the first video display window in the screen display window according to the abscissa and the ordinate of the first position marking point in the screen display window and the size of the first video display window in the screen display window.
Optionally, the video taking a close shot through the currently displayed video taking a close shot template and the recorded at least one video includes:
displaying at least one video in at least one video display window included in the video snapshot template;
in the process that at least one video is displayed in at least one video display window, shooting of the current video is carried out through the remaining video display window except the at least one video display window in the video snapshot template, so that the snapshot of at least two videos is realized.
Optionally, the method further comprises:
when the dragging operation aiming at a first video display window is detected, determining a dragging ending position, wherein the first video display window is any one of at least two video display windows;
determining a second video display window according to the dragging ending position, wherein the second video display window is a video display window which is subjected to position exchange with the first video display window;
and exchanging the display position of the first video and the display position of the second video, wherein the first video is the video currently displayed by the first video display window, and the second video is the video currently displayed by the second video display window.
Optionally, exchanging the display position of the first video and the display position of the second video includes:
switching and displaying the first video from the first video display window into the second video display window, and switching and displaying the second video from the second video display window into the first video display window; alternatively, the first and second electrodes may be,
and exchanging the position of the first video display window and the position of the second video display window to realize the exchange of the display position of the first video and the display position of the second video.
Optionally, exchanging the position of the first video presentation window and the position of the second video presentation window includes:
adjusting the position of a second video display window in the screen display window according to the position description sequence of the first video display window and the size of the screen display window, and re-determining the position description sequence of the adjusted second video display window;
and adjusting the position of the first video display window in the screen display window according to the position description sequence of the second video display window and the size of the screen display window, and re-determining the position description sequence of the adjusted first video display window.
All the above optional technical solutions can be combined arbitrarily to form an optional embodiment of the present application, and the present application embodiment is not described in detail again.
Fig. 2 is a flowchart of a method for snap-shooting a video according to an embodiment of the present application, where the method is applied to a device for snap-shooting a video, and the device for snap-shooting a video may be integrated in a terminal. Referring to fig. 2, the method includes:
step 201: and acquiring a video close-shooting template, wherein the video close-shooting template comprises at least two video display windows, and the at least two video display windows are used for displaying at least two videos for close shooting.
In this embodiment of the application, the terminal may obtain the video snapshot template when detecting the video snapshot instruction, and of course, the terminal may also obtain the video snapshot module at other occasions, for example, the terminal may determine the continuous playing time of the currently played video, and when the continuous playing time reaches the time threshold, it may be indicated that the user likes the currently played video very much, and at this time, the terminal may obtain the video snapshot template to indicate that the user may perform video snapshot with the currently played video.
The video snapshot command may be triggered by the user clicking a video snapshot button, or may be triggered by other operations, such as a shake-and-shake operation, a double-click operation, and the like on the terminal. The number threshold may be set in advance and fixed in the terminal, or may be set before the video is taken, or may be adjusted in different cases.
In some embodiments, the terminal may store one or more video-taken templates in advance. Because different video capture templates can include different numbers of video display windows, for example, some video capture templates can include 2 video display windows, some video capture templates can include 3 video display windows, and some video capture templates can include 4 video display windows. Therefore, when the terminal stores one video snapshot template, it can be determined that the terminal can only perform one form of video snapshot, for example, when the video snapshot template stored in the terminal includes 2 video display windows, it can be determined that the terminal can only perform 2 video snapshots. When the terminal stores a plurality of video close-shot templates, it can be determined that the terminal can perform video close-shot in various forms, such as close-shot of 2 videos, close-shot of 3 videos, close-shot of 4 videos, and the like.
Moreover, different video snap templates may include different numbers of video presentation windows, and the shapes of the included video presentation windows may also be different. For example, for two video capture templates that also include 2 video presentation windows, the video presentation window of one of the video capture templates may be a square, the video presentation window of the other video capture template may be a circle, and the shapes of different video presentation windows in the same video capture template may also be different.
Therefore, when the terminal acquires the video close-shot template, if the terminal only stores one video close-shot template, the terminal can directly acquire the stored video close-shot template. If the terminal stores a plurality of video close-shooting templates, the video close-shooting template with the highest use frequency can be obtained, the plurality of video close-shooting templates can also be displayed, a user selects one video close-shooting template from the plurality of video close-shooting templates, of course, the terminal can also obtain the number of videos which need to be closed-shot currently, and the corresponding video close-shooting template can be obtained according to the number of the obtained videos. For example, the number of videos which need to be currently captured by the terminal is 3, and at this time, the terminal may select a video capture template including 3 video display windows from the stored multiple video capture templates.
The video close-shooting instruction can carry the number of videos which need to be closed-shot currently, and certainly, the terminal can also display a window, and the user inputs the number of videos which need to be closed-shot currently in the window.
In other embodiments, the server may store one or more video capture templates in advance. When the terminal needs to acquire the video co-shooting template, a template acquisition request can be sent to the server, and the server acquires the video co-shooting template according to the template acquisition request and sends the video co-shooting template to the terminal. When the server stores a video snapshot template, the server can directly send the stored video snapshot template to the terminal when receiving a template acquisition request sent by the terminal. When the server stores a plurality of video snapshot templates, the server can acquire the video snapshot template with the highest use frequency and send the video snapshot template to the terminal. Certainly, the template acquisition request may carry the number of videos that need to be taken together currently, and the server acquires the corresponding video taking-together template according to the number of videos and sends the video taking-together template to the terminal.
Step 202: and displaying the video close-shot template in the screen display window according to the position description sequence of each of the at least two video display windows and the size of the screen display window, wherein the position description sequences of the at least two video display windows are determined after the size of the video close-shot template is normalized.
In this embodiment of the present application, according to the position description sequence of each of the at least two video presentation windows and the size of the screen presentation window, the operation of presenting the video snapshot template in the screen presentation window may be: and determining the position of each video display window in the screen display window according to the position description sequence of each video display window in at least two video display windows and the size of the screen display window. And displaying the video snapshot template in the screen display window according to the position of each video display window in the screen display window.
It should be noted that the position description sequence of the video presentation window is a set of information used to characterize the position of the video presentation window in the video snapshot template. In some embodiments, the location description sequence may include the abscissa and ordinate of each key location point in the video presentation window in the video snapshot template after size normalization. For example, when the video display window is a rectangular window, the position description sequence may include the abscissa and the ordinate of the four vertices in the video display window in the video snapshot template after the size normalization. In other embodiments, the position description sequence may include at least three elements, which are the abscissa and the ordinate of the position mark point in the video capture template after size normalization and the size of the video display window relative to the video capture template after size normalization, where the position mark point is a point on the video display window.
The position description sequence of the video display window is determined after the size of the video close-up template is normalized, namely, the size of the video close-up template is normalized, and the abscissa and the ordinate of each key position point in the video display window in the video close-up template after the size normalization form the position description sequence of the video display window. Or determining position marking points of the video display window, and forming a position description sequence of the video display window by using the abscissa and the ordinate of the position marking points in the video close-shot template with the normalized size and the size of the video display window relative to the video close-shot template with the normalized size. The key position point is a point capable of uniquely determining a video display window, and the position marking point can be one of the position key points. For example, when the video presentation window is a rectangle, the key position point may be four vertices of the video presentation window, and when the video presentation window is a circle, the key position point may be a center point of the video presentation window and a point on the circumference. For the position mark point, when the video display window is rectangular, the position mark point may be one of four vertices of the video display window, or may be a center point of the video display window. When the video display window is circular, the position mark point may be a center point of the video display window, or any one of two end points of the circumference of each quadrant after the video display window is divided into four parts according to horizontal lines and vertical lines.
For the convenience of understanding, the position description sequence is exemplified by including at least three elements, and the position description sequence of the video presentation window is explained in conjunction with two specific examples.
In a first example, referring to fig. 3, it is assumed that the video snapshot template includes two rectangular video display windows, which are a first video display window and a second video display window, respectively, where the first video display window corresponds to the first position mark point B, the second video display window corresponds to the second position mark point C, and the point B is used as a coordinate origin, the width of the video snapshot template is used as a horizontal axis, and the height of the video snapshot template is used as a vertical axis, so as to establish a coordinate system. Then, after normalizing the width of the video close-shot template and normalizing the height of the video close-shot template, the position description sequence of the first video display window is as follows: (0, 0, 0.5, 1), the position description sequence of the second video presentation window is: [ 0.5, 0, 0.5, 1 ].
In a second example, referring to fig. 4, it is assumed that the video close-shot template includes two circular video display windows, namely a first video display window and a second video display window, where the first video display window corresponds to the first position mark point E and the second video display window corresponds to the second position mark point F. And establishing a coordinate system by taking the circle center D of the video co-shooting template as a coordinate origin, taking the horizontal diameter of the video co-shooting template D as a horizontal axis and taking the vertical diameter of the video co-shooting template D as a vertical axis. Then, after normalizing the diameter of the video snapshot template D, the position description sequence of the first video presentation window is: -0.25, 0, 0.5 ] the position description sequence of the second video presentation window is: [ 0.25, 0, 0.5 ].
When the position description sequence includes at least three elements, which are respectively an abscissa and an ordinate of the position mark point in the video close-up template after the size normalization, and a size of the video display window relative to the video close-up template after the size normalization, the operation of determining the position of each video display window in the screen display window according to the position description sequence of each video display window in the at least two video display windows and the size of the screen display window may be: for a first video display window of the at least two video display windows, according to the abscissa and the ordinate of the first position mark point in the video close-shot template after the size normalization and the size of the screen display window, the abscissa and the ordinate of the first position mark point in the screen display window are determined, the first position mark point is a point on the first video display window, and the first video display window is any one of the at least two video display windows. And determining the size of the first video display window in the video display window according to the size of the first video display window relative to the video shooting template with the normalized size and the size of the screen display window. And determining the position of the first video display window in the screen display window according to the abscissa and the ordinate of the first position marking point in the screen display window and the size of the first video display window in the screen display window.
When the first video display window is a rectangular window, the position description sequence of the first video display window comprises four elements, namely an abscissa and an ordinate of the first position mark in the video close-shooting template after the size normalization, and a width and a height of the first video display window relative to the video close-shooting template after the size normalization. In this case, the abscissa of the first position mark point in the video capture template after size normalization and the width of the first video display window relative to the video capture template after size normalization may be multiplied by the width of the screen display window, respectively, to obtain the abscissa of the first position mark point in the screen display window and the width of the first video display window in the screen display window. And multiplying the vertical coordinate of the first position mark point in the video close-shot template after the size normalization and the height of the first video display window relative to the video close-shot template after the size normalization by the height of the screen display window respectively to obtain the vertical coordinate of the first position mark point in the screen display window and the height of the first video display window in the screen display window. And determining the position of the first video display window in the screen display window according to the abscissa and the ordinate of the first position mark point in the screen display window and the width and the height of the first video display window relative to the screen display window.
By way of example, taking fig. 3 as an example, referring to fig. 3, it can be known that the abscissa and the ordinate of the first position mark point B in the video capture template after size normalization are both 0, the width of the first video display window relative to the video capture template after size normalization is 0.5, the height of the first video display window relative to the video capture template after size normalization is 1, and at this time, assuming that the width of the screen display window is 4cm, it can be known that the abscissa of the first position mark point B in the screen display window is: 0 is 0 by 4cm, and the ordinate of the first position marking point B in the screen display window is: 0 x 6cm ═ 0, the width of the first video presentation window in the screen presentation window is: 0.5 x 4 cm-2 cm. The first video presentation window has a height in the screen presentation window of: 1 × 6cm ═ 6 cm. Therefore, the position of the first video display window in the screen display window can be determined according to the abscissa and the ordinate of the first position mark point B in the screen display window and the width and the height of the first video display window in the screen display window.
When the first video display window is a circular window, the position description sequence of the first video display window comprises three elements, namely an abscissa and an ordinate of the first position mark in the video close-shooting template after the size normalization and a diameter of the first video display window relative to the video close-shooting template after the size normalization. In this case, the abscissa of the first position mark point in the video snapshot template after the size normalization and the diameter of the first video display window relative to the video snapshot template after the size normalization may be multiplied by the diameter of the screen display window, respectively, to obtain the abscissa of the first position mark point in the screen display window and the diameter of the first video display window in the screen display window. And multiplying the vertical coordinate of the first position mark point in the video close-shot template with the normalized size by the diameter of the screen display window to obtain the vertical coordinate of the first position mark point in the screen display window. And determining the position of the first video display window in the screen display window according to the abscissa and the ordinate of the first position marking point in the screen display window and the diameter of the first video display window relative to the screen display window.
By way of example, taking fig. 4 as an example, referring to fig. 4, it can be seen that the abscissa and the ordinate of the first position mark point E in the video capture template after size normalization are-0.25 and 0, respectively, and the diameter of the first video display window relative to the video capture template after size normalization is 0.5, at this time, assuming that the diameter of the screen display window is 4cm, it can be seen that the abscissa of the first position mark point E in the screen display window is: 0.25 × 4 ═ 1, the ordinate of the first position marker point E in the screen presentation window is: 0 x 6 ═ 0, the diameter of the first video presentation window in the screen presentation window is: 0.5 x 4 cm-2 cm. Therefore, the position of the first video display window in the screen display window can be determined according to the abscissa and the ordinate of the first position mark point E in the screen display window and the diameter of the first video display window in the screen display window.
Step 203: and carrying out video co-shooting through the currently displayed video co-shooting template and the recorded at least one video.
In this embodiment of the present application, the operation of performing video snapshot through the currently displayed video snapshot template and the recorded at least one video may be: and displaying at least one video in at least one video display window included in the video close-shooting template, and shooting the current video through the remaining video display window except the at least one video display window in the video close-shooting template in the process that the at least one video display window displays the at least one video so as to realize close shooting of the at least two videos.
The terminal can perform video shooting when detecting a video shooting start instruction, wherein the video shooting start instruction can be triggered by a user clicking a video shooting start button, and certainly can also be triggered by the user through other operations. For example, in a possible implementation manner, the video close-shooting start instruction may be implemented by a determination button displayed on an operation interface of the terminal, that is, when the user has a satisfactory gesture and believes that the close-shooting of the video can be started, the determination button on the operation interface of the terminal may be clicked, and at this time, the terminal detects the video close-shooting start instruction and performs the close-shooting of the video.
Further, taking fig. 3 as an example, referring to fig. 3, the first video display window displays the first video, and the second video display window displays the second video, that is, in the process of displaying the first video by the first video display window, shooting the current video by using the remaining video display window except the first video display window, that is, the second video display window in the video co-shooting template, so as to implement co-shooting of the two videos. It should be noted that, when multiple videos are needed to be taken in time, the implementation method is similar to the implementation process of taking in time of two videos in this embodiment, and details are not repeated in this application.
Step 204: when the dragging operation aiming at the first video display window is detected, the dragging ending position is determined, and the first video display window is any one of the at least two video display windows.
Each video display window corresponds to one position description sequence, so that each video display window corresponds to one position area in the screen display window, when the terminal detects a dragging operation, the dragging start position can be determined, the position area where the dragging start position is located can be determined, and the video display window corresponding to the position area can be called as a first video display window. At this time, in order to implement the position exchange of the video presentation window, the drag end position may also be determined.
Further, when the terminal detects a drag operation with respect to the first video presentation window, a drag animation may also be displayed. For example, the dragging animation may be a gradient color ribbon or other cool and dazzling effects such as a fairy flower, and the effect of the dragging animation is not specifically limited in the embodiment of the present application.
Step 205: and determining a second video display window according to the dragging ending position, wherein the second video display window is a video display window which is subjected to position exchange with the first video display window.
Since the video snapshot template includes at least two video display windows, and each video display window corresponds to a location area in the screen display window as described above, after the drag end location is determined, the location area where the drag end location is located may be determined, and the video display window corresponding to the location area is referred to as a second video display window. I.e. the video presentation window being position exchanged with the first video presentation window.
Step 206: and exchanging the display position of the first video and the display position of the second video, wherein the first video is the video currently displayed by the first video display window, and the second video is the video currently displayed by the second video display window.
In the embodiment of the present application, the present step may be implemented in the following two different ways.
In a first implementation manner, the first video is switched and displayed from the first video display window to the second video display window, and the second video is switched and displayed from the second video display window to the first video display window.
In a second implementation manner, the position of the first video display window and the position of the second video display window are exchanged to realize the exchange of the display position of the first video and the display position of the second video.
For the two implementation manners, the first implementation manner is to actually exchange videos displayed in the two video display windows, and the second implementation manner is to actually exchange the positions of the two video display windows, and the video displayed in each video display window is not changed, so that the exchange of the display positions of the two videos can be realized.
In some embodiments, the second implementation is actually to swap the sizes of the two video presentation windows to achieve the position swap of the two video presentation windows. The operation process can be as follows: and adjusting the position of the second video display window in the screen display window according to the position description sequence of the first video display window and the size of the screen display window, and re-determining the position description sequence of the adjusted second video display window. And adjusting the position of the first video display window in the screen display window according to the position description sequence of the second video display window and the size of the screen display window, and re-determining the position description sequence of the adjusted first video display window.
Exemplarily, referring to fig. 3, the description sequence of the position of the known first video presentation window is: (0, 0, 0.5, 1), the position description sequence of the second video presentation window is: 0.5, 0, 0.5, 1, the ratio of the width to the height of the screen display window is as follows: 4cm and 6cm, then adjusting the position of the second video display window in the screen display window, and the position description sequence of the adjusted second video display window is: and [ 0, 0, 0.5, 1 ], wherein the position description sequence of the adjusted first video display window is [ 0.5, 0, 0.5, 1 ]. That is, the position of the first video presentation window and the position of the second video presentation window are swapped.
In the embodiment of the application, the video close-up template comprises at least two video display windows, and the position description sequence of each video display window is determined after the size of the video close-up template is normalized. Therefore, when the videos are subjected to close-shot, the close-shot of at least two videos can be realized, the problem that the at least two videos cannot be played simultaneously is solved, and compared with the position description sequence of the video display window which is determined after the size of the video close-shot template is normalized, the position description sequence of the at least two video display windows is simpler, and the subsequent coding is more convenient. In addition, this application embodiment can also exchange the show position of first video and the show position of second video when detecting the drag operation to first video show window, shows the dragging animation of dazzling simultaneously, therefore can improve user's viscosity, and the video function of taking a photo in good tune is also abundanter.
Referring to fig. 5, an apparatus for taking a video in time is provided in an embodiment of the present application, and the apparatus includes an obtaining module 501, a presentation module 502, and a taking module 503.
The acquiring module 501 is configured to acquire a video close-shooting template, where the video close-shooting template includes at least two video display windows, and the at least two video display windows are used for displaying at least two videos for close shooting;
the display module 502 is configured to display a video snapshot template in the screen display window according to the position description sequence of each of the at least two video display windows and the size of the screen display window, where the position description sequences of the at least two video display windows are determined by normalizing the size of the video snapshot template;
and a close-up module 503, configured to perform close-up shooting on a video through the currently displayed video close-up template and the recorded at least one video.
Optionally, the display module 502 comprises:
determining a submodule for determining the position of each video display window in the screen display window according to the position description sequence of each video display window in at least two video display windows and the size of the screen display window;
and the first display submodule is used for displaying the video snapshot template in the screen display window according to the position of each video display window in the screen display window.
Optionally, the position description sequence includes at least three elements, which are respectively an abscissa and an ordinate of the position mark point in the video capture template after the size normalization, and a size of the video display window relative to the video capture template after the size normalization, where the position mark point is a point on the video display window;
the determination submodule includes:
the first determining unit is used for determining the abscissa and the ordinate of the first position mark point in the screen display window according to the abscissa and the ordinate of the first position mark point in the video close-shot template after the size normalization and the size of the screen display window, wherein the first position mark point is a point on the first video display window, and the first video display window is any one of the at least two video display windows;
the second determining unit is used for determining the size of the first video display window in the video display window according to the size of the first video display window relative to the video snapshot template with the normalized size and the size of the screen display window;
and the third determining unit is used for determining the position of the first video display window in the screen display window according to the abscissa and the ordinate of the first position mark point in the screen display window and the size of the first video display window in the screen display window.
Optionally, the clap module 503 includes:
the second display submodule is used for displaying at least one video in at least one video display window included in the video snapshot template;
and the shooting submodule is used for shooting the current video through the remaining video display window except the at least one video display window in the video close-shooting template in the process of displaying the at least one video through the at least one video display window so as to realize the close shooting of the at least two videos.
Optionally, referring to fig. 6, the apparatus further comprises:
a first determining module 504, configured to determine a dragging ending position when a dragging operation for a first video display window is detected, where the first video display window is any one of the at least two video display windows;
a second determining module 505, configured to determine a second video display window according to the dragging ending position, where the second video display window is a video display window that performs position exchange with the first video display window;
the exchanging module 506 is configured to exchange a display position of a first video and a display position of a second video, where the first video is a video currently displayed in the first video display window, and the second video is a video currently displayed in the second video display window.
Optionally, the switching module 506 includes:
the switching submodule is used for switching and displaying the first video from the first video display window to the second video display window and switching and displaying the second video from the second video display window to the first video display window; alternatively, the first and second electrodes may be,
and the switching submodule is used for switching the position of the first video display window and the position of the second video display window so as to realize the switching of the display position of the first video and the display position of the second video.
Optionally, the switching submodule includes:
a fourth determining unit, configured to adjust a position of the second video display window in the screen display window according to the position description sequence of the first video display window and the size of the screen display window, and to re-determine the position description sequence of the adjusted second video display window;
and the fifth determining unit is used for adjusting the position of the first video display window in the screen display window according to the position description sequence of the second video display window and the size of the screen display window, and re-determining the position description sequence of the adjusted first video display window.
In summary, in the embodiment of the present application, the video capture template includes at least two video display windows, and the position description sequence of each video display window is determined after the size of the video capture template is normalized. Therefore, when the videos are subjected to close-shot, the close-shot of at least two videos can be realized, the problem that the at least two videos cannot be played simultaneously is solved, and compared with the position description sequence of the video display window which is determined after the size of the video close-shot template is normalized, the position description sequence of the at least two video display windows is simpler, and the subsequent coding is more convenient. In addition, the embodiment that this application provided can also exchange the show position of first video and the show position of second video when detecting the drag operation to first video show window, shows the drag animation of cool dazzling simultaneously, therefore can improve user's viscosity, and the video function of taking a photo in good tune is also abundanter.
It should be noted that: in the device for taking a video in time according to the foregoing embodiment, only the division of the functional modules is illustrated, and in practical applications, the functions may be distributed by different functional modules according to needs, that is, the internal structure of the device may be divided into different functional modules to complete all or part of the functions described above. In addition, the device for taking a video in tune provided by the above embodiment and the method embodiment for taking a video in tune belong to the same concept, and the specific implementation process thereof is detailed in the method embodiment and will not be described herein again.
Fig. 7 shows a block diagram of a terminal 700 according to an exemplary embodiment of the present application. The terminal 700 may be: a mobile phone, a tablet computer, a notebook computer or a desktop computer. Terminal 700 may also be referred to by other names such as user equipment, portable terminal, laptop terminal, desktop terminal, and so on.
In general, terminal 700 includes: a processor 701 and a memory 702.
The processor 701 may include one or more processing cores, such as a 4-core processor, an 8-core processor, and so on. The processor 701 may be implemented in at least one hardware form of a DSP (Digital Signal Processing), an FPGA (Field-Programmable Gate Array), and a PLA (Programmable Logic Array). The processor 701 may also include a main processor and a coprocessor, where the main processor is a processor for Processing data in an awake state, and is also called a Central Processing Unit (CPU); a coprocessor is a low power processor for processing data in a standby state. In some embodiments, the processor 701 may be integrated with a GPU (Graphics Processing Unit), which is responsible for rendering and drawing the content required to be displayed on the display screen. In some embodiments, the processor 701 may further include an AI (Artificial Intelligence) processor for processing computing operations related to machine learning.
Memory 702 may include one or more computer-readable storage media, which may be non-transitory. Memory 702 may also include high-speed random access memory, as well as non-volatile memory, such as one or more magnetic disk storage devices, flash memory storage devices. In some embodiments, a non-transitory computer readable storage medium in memory 702 is used to store at least one instruction for execution by processor 701 to implement a method of managing oil and gas pipeline maintenance repair information as provided by method embodiments herein.
In some embodiments, the terminal 700 may further optionally include: a peripheral interface 703 and at least one peripheral. The processor 701, the memory 702, and the peripheral interface 703 may be connected by buses or signal lines. Various peripheral devices may be connected to peripheral interface 703 via a bus, signal line, or circuit board. Specifically, the peripheral device includes: at least one of radio frequency circuitry 704, touch screen display 705, camera assembly 706, audio circuitry 707, positioning assembly 708, and power source 709.
The peripheral interface 703 may be used to connect at least one peripheral related to I/O (Input/Output) to the processor 701 and the memory 702. In some embodiments, processor 701, memory 702, and peripheral interface 703 are integrated on the same chip or circuit board; in some other embodiments, any one or two of the processor 701, the memory 702, and the peripheral interface 703 may be implemented on a separate chip or circuit board, which is not limited in this embodiment.
The Radio Frequency circuit 704 is used for receiving and transmitting RF (Radio Frequency) signals, also called electromagnetic signals. The radio frequency circuitry 704 communicates with communication networks and other communication devices via electromagnetic signals. The rf circuit 704 converts an electrical signal into an electromagnetic signal to transmit, or converts a received electromagnetic signal into an electrical signal. Optionally, the radio frequency circuit 704 includes: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a subscriber identity module card, and so forth. The radio frequency circuitry 704 may communicate with other terminals via at least one wireless communication protocol. The wireless communication protocols include, but are not limited to: the world wide web, metropolitan area networks, intranets, generations of mobile communication networks (2G, 3G, 4G, and 5G), Wireless local area networks, and/or WiFi (Wireless Fidelity) networks. In some embodiments, the radio frequency circuit 704 may also include NFC (Near Field Communication) related circuits, which are not limited in this application.
The display screen 705 is used to display a UI (User Interface). The UI may include graphics, text, icons, video, and any combination thereof. When the display screen 705 is a touch display screen, the display screen 705 also has the ability to capture touch signals on or over the surface of the display screen 705. The touch signal may be input to the processor 701 as a control signal for processing. At this point, the display 705 may also be used to provide virtual buttons and/or a virtual keyboard, also referred to as soft buttons and/or a soft keyboard. In some embodiments, the display 705 may be one, providing the front panel of the terminal 700; in other embodiments, the display 705 can be at least two, respectively disposed on different surfaces of the terminal 700 or in a folded design; in still other embodiments, the display 705 may be a flexible display disposed on a curved surface or on a folded surface of the terminal 700. Even more, the display 705 may be arranged in a non-rectangular irregular pattern, i.e. a shaped screen. The Display 705 may be made of LCD (Liquid Crystal Display), OLED (Organic Light-Emitting Diode), or the like.
The camera assembly 706 is used to capture images or video. Optionally, camera assembly 706 includes a front camera and a rear camera. Generally, a front camera is disposed at a front panel of the terminal, and a rear camera is disposed at a rear surface of the terminal. In some embodiments, the number of the rear cameras is at least two, and each rear camera is any one of a main camera, a depth-of-field camera, a wide-angle camera and a telephoto camera, so that the main camera and the depth-of-field camera are fused to realize a background blurring function, and the main camera and the wide-angle camera are fused to realize panoramic shooting and VR (Virtual Reality) shooting functions or other fusion shooting functions. In some embodiments, camera assembly 706 may also include a flash. The flash lamp can be a monochrome temperature flash lamp or a bicolor temperature flash lamp. The double-color-temperature flash lamp is a combination of a warm-light flash lamp and a cold-light flash lamp, and can be used for light compensation at different color temperatures.
The audio circuitry 707 may include a microphone and a speaker. The microphone is used for collecting sound waves of a user and the environment, converting the sound waves into electric signals, and inputting the electric signals to the processor 701 for processing or inputting the electric signals to the radio frequency circuit 704 to realize voice communication. For the purpose of stereo sound collection or noise reduction, a plurality of microphones may be provided at different portions of the terminal 700. The microphone may also be an array microphone or an omni-directional pick-up microphone. The speaker is used to convert electrical signals from the processor 701 or the radio frequency circuit 704 into sound waves. The loudspeaker can be a traditional film loudspeaker or a piezoelectric ceramic loudspeaker. When the speaker is a piezoelectric ceramic speaker, the speaker can be used for purposes such as converting an electric signal into a sound wave audible to a human being, or converting an electric signal into a sound wave inaudible to a human being to measure a distance. In some embodiments, the audio circuitry 707 may also include a headphone jack.
The positioning component 708 is used to locate the current geographic Location of the terminal 700 for navigation or LBS (Location Based Service). The Positioning component 708 can be a Positioning component based on the Global Positioning System (GPS) in the united states, the beidou System in china, or the galileo System in russia.
Power supply 709 is provided to supply power to various components of terminal 700. The power source 709 may be alternating current, direct current, disposable batteries, or rechargeable batteries. When the power source 709 includes a rechargeable battery, the rechargeable battery may be a wired rechargeable battery or a wireless rechargeable battery. The wired rechargeable battery is a battery charged through a wired line, and the wireless rechargeable battery is a battery charged through a wireless coil. The rechargeable battery may also be used to support fast charge technology.
In some embodiments, terminal 700 also includes one or more sensors 710. The one or more sensors 710 include, but are not limited to: acceleration sensor 711, gyro sensor 712, pressure sensor 713, fingerprint sensor 714, optical sensor 715, and proximity sensor 717.
The acceleration sensor 711 can detect the magnitude of acceleration in three coordinate axes of a coordinate system established with the terminal 700. For example, the acceleration sensor 711 may be used to detect components of the gravitational acceleration in three coordinate axes. The processor 701 may control the touch screen 705 to display the user interface in a landscape view or a portrait view according to the gravitational acceleration signal collected by the acceleration sensor 711. The acceleration sensor 711 may also be used for acquisition of motion data of a game or a user.
The gyro sensor 712 may detect a body direction and a rotation angle of the terminal 700, and the gyro sensor 712 may cooperate with the acceleration sensor 711 to acquire a 3D motion of the terminal 700 by the user. From the data collected by the gyro sensor 712, the processor 701 may implement the following functions: motion sensing (such as changing the UI according to a user's tilting operation), image stabilization at the time of photographing, game control, and inertial navigation.
Pressure sensors 713 may be disposed on a side bezel of terminal 700 and/or an underlying layer of touch display 705. When the pressure sensor 713 is disposed on a side frame of the terminal 700, a user's grip signal on the terminal 700 may be detected, and the processor 701 performs right-left hand recognition or shortcut operation according to the grip signal collected by the pressure sensor 713. When the pressure sensor 713 is disposed at a lower layer of the touch display 705, the processor 701 controls the operability control on the UI interface according to the pressure operation of the user on the touch display 705. The operability control comprises at least one of a button control, a scroll bar control, an icon control and a menu control.
The fingerprint sensor 714 is used for collecting a fingerprint of a user, and the processor 701 identifies the identity of the user according to the fingerprint collected by the fingerprint sensor 714, or the fingerprint sensor 714 identifies the identity of the user according to the collected fingerprint. When the user identity is identified as a trusted identity, the processor 701 authorizes the user to perform relevant sensitive operations, including unlocking a screen, viewing encrypted information, downloading software, paying, changing settings, and the like. The fingerprint sensor 714 may be disposed on the front, back, or side of the terminal 700. When a physical button or a vendor Logo is provided on the terminal 700, the fingerprint sensor 714 may be integrated with the physical button or the vendor Logo.
The optical sensor 715 is used to collect the ambient light intensity. In one embodiment, the processor 701 may control the display brightness of the touch display 705 based on the ambient light intensity collected by the optical sensor 715. Specifically, when the ambient light intensity is high, the display brightness of the touch display screen 705 is increased; when the ambient light intensity is low, the display brightness of the touch display 705 is turned down. In another embodiment, processor 701 may also dynamically adjust the shooting parameters of camera assembly 706 based on the ambient light intensity collected by optical sensor 715.
A proximity sensor 717, also called a distance sensor, is generally provided at a front panel of the terminal 700. The proximity sensor 717 serves to collect a distance between the user and the front surface of the terminal 700. In one embodiment, when the proximity sensor 717 detects that the distance between the user and the front surface of the terminal 700 gradually decreases, the processor 701 controls the touch display 705 to switch from the bright screen state to the dark screen state; when the proximity sensor 717 detects that the distance between the user and the front surface of the terminal 700 gradually increases, the processor 701 controls the touch display 705 to switch from the breath-screen state to the bright-screen state.
Those skilled in the art will appreciate that the configuration shown in fig. 7 is not intended to be limiting of terminal 700 and may include more or fewer components than those shown, or some components may be combined, or a different arrangement of components may be used.
The present application provides not only a terminal including a processor and a memory for storing processor-executable instructions, wherein the processor is configured to execute the method of co-shooting video shown in fig. 1 or 2, but also a computer-readable storage medium having a computer program stored therein, which when executed by the processor can implement the method of co-shooting video shown in fig. 1-2.
Embodiments of the present application further provide a computer program product containing instructions, which when run on a computer, cause the computer to execute the method for taking a video in time provided in the embodiments shown in fig. 1-2.
It will be understood by those skilled in the art that all or part of the steps for implementing the above embodiments may be implemented by hardware, or may be implemented by a program instructing relevant hardware, where the program may be stored in a computer-readable storage medium, and the above-mentioned storage medium may be a read-only memory, a magnetic disk or an optical disk, etc.
The above description is only exemplary of the present application and should not be taken as limiting, as any modification, equivalent replacement, or improvement made within the spirit and principle of the present application should be included in the protection scope of the present application.

Claims (6)

1. A method of co-shooting a video, the method comprising:
the method comprises the steps of obtaining a video close-shot template, wherein the video close-shot template comprises at least two video display windows which are used for displaying at least two videos for close-shot;
determining the position of each video display window in the screen display window according to the position description sequence of each video display window in the at least two video display windows and the size of the screen display window, and displaying the video close-shot template in the screen display window according to the position of each video display window in the screen display window, wherein the position description sequences of the at least two video display windows are determined after the size of the video close-shot template is normalized;
performing video co-shooting through the currently displayed video co-shooting template and at least one recorded video;
the method further comprises the following steps:
when the dragging operation aiming at a first video display window is detected, determining a dragging ending position and displaying a dragging animation, wherein the first video display window is any one of the at least two video display windows;
determining a second video display window according to the dragging ending position, wherein the second video display window is a video display window which is subjected to position exchange with the first video display window;
exchanging the display position of the first video and the display position of the second video; the first video is a video currently displayed by the first video display window, and the second video is a video currently displayed by the second video display window;
the position description sequence comprises at least three elements which are respectively an abscissa and an ordinate of a position marking point in the video close-shot template after the size normalization, and the size of a video display window relative to the video close-shot template after the size normalization, wherein the position marking point is one of key position points, and the key position points are points which are used for uniquely determining the video display window;
determining the position of each video display window in the screen display window according to the position description sequence of each video display window in the at least two video display windows and the size of the screen display window, including:
for a first video display window in the at least two video display windows, determining the abscissa and the ordinate of a first position mark point in the screen display window according to the abscissa and the ordinate of the first position mark point in the video close-shot template after the size normalization and the size of the screen display window, wherein the first position mark point is a point on the first video display window;
determining the size of the first video display window in the screen display window according to the size of the first video display window relative to the video shooting template with the normalized size and the size of the screen display window;
and determining the position of the first video display window in the screen display window according to the abscissa and the ordinate of the first position marking point in the screen display window and the size of the first video display window in the screen display window.
2. The method of claim 1, wherein said taking a video snap through the currently presented video snap template and at least one video that has been recorded comprises:
displaying the at least one video in at least one video display window included in the video close-shot template;
and in the process of displaying the at least one video through the at least one video display window, shooting the current video through the remaining video display window except the at least one video display window in the video close-shot template so as to realize close-shot of at least two videos.
3. The method of claim 1, wherein exchanging the presentation location of the first video and the presentation location of the second video comprises:
switching the first video from the first video presentation window to be presented in the second video presentation window, and switching the second video from the second video presentation window to be presented in the first video presentation window; alternatively, the first and second electrodes may be,
and exchanging the position of the first video display window and the position of the second video display window to realize the exchange of the display position of the first video and the display position of the second video.
4. The method of claim 3, wherein said swapping the position of the first video presentation window and the position of the second video presentation window comprises:
adjusting the position of the second video display window in the screen display window according to the position description sequence of the first video display window and the size of the screen display window, and re-determining the position description sequence of the adjusted second video display window;
and adjusting the position of the first video display window in the screen display window according to the position description sequence of the second video display window and the size of the screen display window, and re-determining the position description sequence of the adjusted first video display window.
5. An apparatus for co-shooting video, the apparatus comprising:
the system comprises an acquisition module, a storage module and a display module, wherein the acquisition module is used for acquiring a video co-shooting template, the video co-shooting template comprises at least two video display windows, and the at least two video display windows are used for displaying at least two videos for co-shooting;
the display module is used for determining the position of each video display window in the screen display window according to the position description sequence of each video display window in the at least two video display windows and the size of the screen display window, and displaying the video close-shot template in the screen display window according to the position of each video display window in the screen display window, wherein the position description sequences of the at least two video display windows are determined by normalizing the size of the video close-shot template;
the video-on-demand shooting module is used for carrying out video-on-demand shooting through the currently displayed video-on-demand shooting template and at least one recorded video;
the apparatus also includes means for:
when the dragging operation aiming at a first video display window is detected, determining a dragging ending position and displaying a dragging animation, wherein the first video display window is any one of the at least two video display windows; determining a second video display window according to the dragging ending position, wherein the second video display window is a video display window which is subjected to position exchange with the first video display window; exchanging a display position of a first video and a display position of a second video, wherein the first video is a video currently displayed in a first video display window, and the second video is a video currently displayed in a second video display window;
the position description sequence comprises at least three elements which are respectively an abscissa and an ordinate of a position marking point in the video close-shot template after the size normalization, and the size of a video display window relative to the video close-shot template after the size normalization, wherein the position marking point is one of key position points, and the key position points are points which are used for uniquely determining the video display window;
the display module is used for determining the abscissa and the ordinate of a first position mark point in a screen display window according to the abscissa and the ordinate of the first position mark point in the video close-shot template after the size normalization and the size of the screen display window for the first video display window in the at least two video display windows, wherein the first position mark point is a point on the first video display window; determining the size of the first video display window in the screen display window according to the size of the first video display window relative to the video shooting template with the normalized size and the size of the screen display window; and determining the position of the first video display window in the screen display window according to the abscissa and the ordinate of the first position marking point in the screen display window and the size of the first video display window in the screen display window.
6. A computer-readable storage medium, in which a computer program is stored which, when being executed by a processor, carries out the steps of the method according to any one of claims 1 to 4.
CN201910193873.1A 2019-03-14 2019-03-14 Method and device for video co-shooting and storage medium Active CN109862412B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910193873.1A CN109862412B (en) 2019-03-14 2019-03-14 Method and device for video co-shooting and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910193873.1A CN109862412B (en) 2019-03-14 2019-03-14 Method and device for video co-shooting and storage medium

Publications (2)

Publication Number Publication Date
CN109862412A CN109862412A (en) 2019-06-07
CN109862412B true CN109862412B (en) 2021-08-13

Family

ID=66900915

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910193873.1A Active CN109862412B (en) 2019-03-14 2019-03-14 Method and device for video co-shooting and storage medium

Country Status (1)

Country Link
CN (1) CN109862412B (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110121094A (en) * 2019-06-20 2019-08-13 广州酷狗计算机科技有限公司 Video is in step with display methods, device, equipment and the storage medium of template
CN110336968A (en) * 2019-07-17 2019-10-15 广州酷狗计算机科技有限公司 Video recording method, device, terminal device and storage medium
CN112004045A (en) * 2020-08-26 2020-11-27 Oppo(重庆)智能科技有限公司 Video processing method, device and storage medium
CN112004032B (en) * 2020-09-04 2022-02-18 北京字节跳动网络技术有限公司 Video processing method, terminal device and storage medium
CN115720292A (en) * 2021-08-23 2023-02-28 北京字跳网络技术有限公司 Video recording method, apparatus, storage medium, and program product
CN114401368B (en) * 2022-01-24 2024-05-03 杭州卡路里体育有限公司 Processing method and device for simultaneous video

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6452570B1 (en) * 1998-08-19 2002-09-17 Bayerische Motoren Werke Aktiengesellschaft Device for controlling the reproduction of an image displayed on a vehicle screen
JP2006185025A (en) * 2004-12-27 2006-07-13 Fuji Electric Systems Co Ltd Window display control method and program
CN105210051A (en) * 2013-03-18 2015-12-30 谷歌公司 Estimating visibility of content items
CN106406794A (en) * 2016-09-14 2017-02-15 乐视控股(北京)有限公司 Screen adapting method and device
CN108566519A (en) * 2018-04-28 2018-09-21 腾讯科技(深圳)有限公司 Video creating method, device, terminal and storage medium
CN108769561A (en) * 2018-06-22 2018-11-06 广州酷狗计算机科技有限公司 video recording method and device
CN108989692A (en) * 2018-10-19 2018-12-11 北京微播视界科技有限公司 Video capture method, apparatus, electronic equipment and computer readable storage medium
CN108989691A (en) * 2018-10-19 2018-12-11 北京微播视界科技有限公司 Video capture method, apparatus, electronic equipment and computer readable storage medium
CN109089059A (en) * 2018-10-19 2018-12-25 北京微播视界科技有限公司 Method, apparatus, electronic equipment and the computer storage medium that video generates
CN109151356A (en) * 2018-09-05 2019-01-04 传线网络科技(上海)有限公司 video recording method and device

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6452570B1 (en) * 1998-08-19 2002-09-17 Bayerische Motoren Werke Aktiengesellschaft Device for controlling the reproduction of an image displayed on a vehicle screen
JP2006185025A (en) * 2004-12-27 2006-07-13 Fuji Electric Systems Co Ltd Window display control method and program
CN105210051A (en) * 2013-03-18 2015-12-30 谷歌公司 Estimating visibility of content items
CN106406794A (en) * 2016-09-14 2017-02-15 乐视控股(北京)有限公司 Screen adapting method and device
CN108566519A (en) * 2018-04-28 2018-09-21 腾讯科技(深圳)有限公司 Video creating method, device, terminal and storage medium
CN108769561A (en) * 2018-06-22 2018-11-06 广州酷狗计算机科技有限公司 video recording method and device
CN109151356A (en) * 2018-09-05 2019-01-04 传线网络科技(上海)有限公司 video recording method and device
CN108989692A (en) * 2018-10-19 2018-12-11 北京微播视界科技有限公司 Video capture method, apparatus, electronic equipment and computer readable storage medium
CN108989691A (en) * 2018-10-19 2018-12-11 北京微播视界科技有限公司 Video capture method, apparatus, electronic equipment and computer readable storage medium
CN109089059A (en) * 2018-10-19 2018-12-25 北京微播视界科技有限公司 Method, apparatus, electronic equipment and the computer storage medium that video generates

Also Published As

Publication number Publication date
CN109862412A (en) 2019-06-07

Similar Documents

Publication Publication Date Title
CN109862412B (en) Method and device for video co-shooting and storage medium
CN108449641B (en) Method, device, computer equipment and storage medium for playing media stream
CN108965922B (en) Video cover generation method and device and storage medium
CN109922356B (en) Video recommendation method and device and computer-readable storage medium
WO2022134632A1 (en) Work processing method and apparatus
CN110288689B (en) Method and device for rendering electronic map
CN110941375A (en) Method and device for locally amplifying image and storage medium
CN110839174A (en) Image processing method and device, computer equipment and storage medium
CN111754386A (en) Image area shielding method, device, equipment and storage medium
CN111083526B (en) Video transition method and device, computer equipment and storage medium
CN112565806A (en) Virtual gift presenting method, device, computer equipment and medium
CN109783176B (en) Page switching method and device
CN111385525B (en) Video monitoring method, device, terminal and system
CN110769120A (en) Method, device, equipment and storage medium for message reminding
CN112612405B (en) Window display method, device, equipment and computer readable storage medium
CN110677713A (en) Video image processing method and device and storage medium
CN112396076A (en) License plate image generation method and device and computer storage medium
CN113160031A (en) Image processing method, image processing device, electronic equipment and storage medium
CN111158575B (en) Method, device and equipment for terminal to execute processing and storage medium
CN109032492B (en) Song cutting method and device
CN109033473B (en) Picture sharing method and device and computer readable storage medium
CN113535039B (en) Method and device for updating page, electronic equipment and computer readable storage medium
CN114594885A (en) Application icon management method, device and equipment and computer readable storage medium
CN109275015B (en) Method, device and storage medium for displaying virtual article
CN108881715B (en) Starting method and device of shooting mode, terminal and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant