CN111629269A - Method for automatically shooting and generating mobile terminal short video advertisement based on mechanical arm - Google Patents
Method for automatically shooting and generating mobile terminal short video advertisement based on mechanical arm Download PDFInfo
- Publication number
- CN111629269A CN111629269A CN202010452022.7A CN202010452022A CN111629269A CN 111629269 A CN111629269 A CN 111629269A CN 202010452022 A CN202010452022 A CN 202010452022A CN 111629269 A CN111629269 A CN 111629269A
- Authority
- CN
- China
- Prior art keywords
- lens
- script
- video
- mechanical arm
- motion
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 47
- 230000033001 locomotion Effects 0.000 claims abstract description 84
- 238000013515 script Methods 0.000 claims abstract description 80
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 17
- 238000010586 diagram Methods 0.000 claims abstract description 12
- 238000009877 rendering Methods 0.000 claims abstract description 7
- 230000003993 interaction Effects 0.000 claims abstract description 5
- 238000004519 manufacturing process Methods 0.000 claims abstract description 5
- 230000002194 synthesizing effect Effects 0.000 claims abstract description 5
- 230000000007 visual effect Effects 0.000 claims description 30
- 230000009471 action Effects 0.000 claims description 16
- 239000000463 material Substances 0.000 claims description 14
- 239000000203 mixture Substances 0.000 claims description 8
- 238000012545 processing Methods 0.000 claims description 7
- 238000004891 communication Methods 0.000 claims description 6
- 230000008569 process Effects 0.000 claims description 6
- 230000002452 interceptive effect Effects 0.000 claims description 4
- 230000033764 rhythmic process Effects 0.000 claims description 2
- 230000015572 biosynthetic process Effects 0.000 claims 1
- 238000003786 synthesis reaction Methods 0.000 claims 1
- 238000002372 labelling Methods 0.000 abstract description 5
- 238000013461 design Methods 0.000 description 4
- 230000001133 acceleration Effects 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 238000013499 data model Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 239000003381 stabilizer Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/472—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
- H04N21/47205—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/812—Monomedia components thereof involving advertisement data
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Databases & Information Systems (AREA)
- Human Computer Interaction (AREA)
- Business, Economics & Management (AREA)
- Marketing (AREA)
- Processing Or Creating Images (AREA)
- Studio Devices (AREA)
Abstract
The invention provides a method or a system for automatically shooting and generating a mobile-end short video advertisement based on a mechanical arm, wherein the method comprises the following steps: (1) determining the video style and commodity information through user interaction, and generating a video script by combining a video script model through the system; (2) the system plans the motion path of the mechanical arm to be planned and automatically shot, and interactively guides a user to arrange a scene through a line diagram; (3) and analyzing the video to be edited based on the saliency region algorithm, and labeling the video. And editing, synthesizing and rendering the video based on the label combination script. The invention develops and explores a new man-machine cooperation mode for generating the short video advertisements. The efficiency of short video advertisement production is obviously promoted, and the quality of short video advertisements is optimized.
Description
Technical Field
The invention belongs to the field of video generation, and particularly relates to a method and a system for automatically shooting and generating a mobile-end short video advertisement based on a mechanical arm.
Background
With the rise of short video applications, short video advertisements are in rapid development, and for small micro-enterprises, the short video advertisements put on social media have considerable conversion rate, so that the sales volume of commodities can be remarkably increased. But for small micro-businesses, making short video advertisements is an extremely time, effort, and cost consuming task. Among billions of good, the number of short video advertisements is less than 1%.
Short video generation can be roughly divided into three steps of script design, shooting, and post-clipping. In the aspect of script design, existing short video production tools, such as TopVid and Animoto, all rely on a fixed video template with a single structure, and these short video advertisement production tools often only perform general and limited classification on video templates, and lack the acquisition of user requirements for a shooting part.
In the aspect of shooting, although a series of exploration is available for the interaction mode of shooting guidance of non-professional users, most of the interaction modes only stay at the level of photo shooting. In contrast, video shooting has more dimensional features including lens movement direction, lens movement speed and dynamic composition, and has higher complexity, and interactive guidance in the video field is very rare. Professional camera shooting auxiliary equipment (such as a handheld stabilizer) including Xinjiang can reduce the shaking degree of a lens and help non-professional users to improve the quality of video shooting to a certain extent, but the equipment still needs the users to design the movement path and speed of the lens, and for the non-professional users, high learning cost exists and the users still have difficulty in shooting high-quality short videos.
Existing video clips are mostly done manually by hand. Related methods have explored video automatic editing methods. Video auto-editing is an important branch of the field of computer vision research. The current automatic clipping method mainly uses a visual method to detect human features (such as human faces) or automatically clips based on preset information (such as video scripts and human dialogue contents). However, these editing methods are limited to the scenes of the character, and cannot automatically edit the video with a large part of scenes containing no actor, such as short video advertisements.
Disclosure of Invention
The invention provides a method or a system for automatically shooting and generating a mobile terminal short video advertisement based on a mechanical arm. After shooting, the video is automatically edited through an algorithm based on the analysis of the salient regions. The invention explores a new man-machine cooperation mode for generating the short video advertisements, obviously improves the efficiency of making the short video advertisements and optimizes the quality of the short video advertisements. The automatic generation of the video mainly comprises the following steps:
1. the system generates a video script based on user requirements and a script generation model
The system provides the requirements for acquiring the front-end interactive interface on one hand, and generates the video script by combining the short video advertisement script model on the basis of the acquired user requirements on the back end on the other hand. And in a user requirement acquisition interface, the user selects the style of the video and the category of the commodity by clicking to be used as input parameters. The script model contains design rules for the visual effects of the advertising video. The video script finally generated by the system encodes the visual effect of the video and describes the specific contents of the video, including the video advertisement duration, the split mirror, the main tone and the like.
2. The system analyzes the video script, plans the motion path of the mechanical arm and guides the user to perform scene setting
The system calculates the movement speed and the movement curve of a main object (namely a commodity) in a two-dimensional camera plane according to the visual effect of each lens defined in the script; calculating the motion speed and the motion curve of the camera in the three-dimensional space according to the motion speed and the motion curve of the main object in the plane of the two-dimensional camera; and calculating and obtaining the position information of each degree of freedom of the mechanical arm in the mechanical arm coordinate based on a reverse dynamics algorithm according to the movement speed and the movement curve of the camera in the three-dimensional space, and planning the complete movement path of the mechanical arm for shooting all commodity lenses.
Meanwhile, the system reads the position of the article in the first commodity lens, calculates the position relation of the mechanical arm, the camera and the commodity in the three-dimensional space, and generates a corresponding wire frame diagram to guide the user to put the commodity and the susceptors in the shooting scene. The wire frame diagram includes the position of the commodity, the position of the susceptor, and the angle of the camera. The user sets under the guidance of the online block diagram and fixes the camera at the tail end of the mechanical arm. The robot arm takes this as an initial state of shooting.
3. Performing significance region analysis on video to be edited by utilizing significance algorithm and synthesizing rendering
The mechanical arm shoots and obtains a continuous section of video to be edited. The video to be edited comprises: a segment moving at a constant speed; shot-to-shot transition segments; a segment in which the robot arm accelerates after moving in the changing direction or a segment in which the robot arm decelerates to reach a stop state. The system needs to keep or discard video segments based on the video script. The saliency algorithm analyzes the video to be edited frame by frame to obtain the size and position coordinates of the range of the saliency area of each frame of shot. In order to improve the robustness of the algorithm, the average value processing is carried out on every five frames.
And selecting key frames every five frames, and performing labeling processing. For each key frame, the distance and angular offset of the key frame from several adjacent key frames are analyzed. And judging the motion state of the object in the picture according to the position offset, and labeling the key frame according to the motion state of the object in the picture. And after the labeling processing, combining the visual characteristic information defined by the script, editing and synthesizing the video to be edited, and finally obtaining the advertisement film.
Compared with the prior method, the method has the innovation points that:
1. the professional and abstract short video advertisement making method is converted into a data model in a computer system, so that the complicated video making process is replaced by simple operation steps, and the use threshold of a user is reduced in a man-machine cooperation system generated by the short video advertisement.
2. The shooting process of the short video advertisement is automatically completed by the aid of the mechanical arm, and a user only needs to select needed advertisement styles and product categories through a mobile phone interface and arrange shooting scenes under guidance. The problems of wrong composition, uneven speed, shaking and the like which possibly occur in manual shooting are avoided, the shooting efficiency is improved, and the video quality is optimized.
3. The video is analyzed and edited by using the salient region analysis algorithm, so that the time-consuming and labor-consuming steps in manual editing are replaced, the video editing efficiency is improved, and the video editing accuracy is optimized.
Drawings
FIG. 1 is a flow chart of the automatic video generation method of the present invention
FIG. 2 is a flow chart of a video script generation method of the present invention
FIG. 3 is a user request acquisition interface in accordance with the present invention
FIG. 4 is a flow chart of a robot arm shooting module of the present invention
FIG. 5 is a guide line diagram of the scenery according to the invention
FIG. 6 is an exemplary communication protocol format sent to a robotic arm in accordance with the present invention
FIG. 7 is a diagram illustrating an example of the relationship between the positions of the robot arm, the mobile phone and the object to be photographed
FIG. 8 is a flow chart of an automatic clipping method of the present invention
FIG. 9 is a flowchart of a method for automatically capturing and generating a short video advertisement in accordance with the present invention
Detailed Description
In order to make the objects, technical processes and technical innovation points of the present invention more clearly illustrated, the present invention is further described in detail with reference to the accompanying drawings and examples. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention. In addition, the technical features involved in the embodiments of the present invention described below may be combined with each other as long as they do not conflict with each other.
In order to achieve the aim, the invention provides a method for automatically shooting and generating a mobile-end short video advertisement based on a mechanical arm. The main process is shown in fig. 1, and the method comprises the following steps:
and step S01, the user interactively determines the video style and the commodity information to generate a video script. Referring to fig. 2, specifically, the following may be:
the user selects the video style and the commodity category as input on the demand acquisition interface. The interface is referred to in fig. 3. The video style comprises fresh wind and ancient wind; the commodity category comprises jewelry, tableware, food, daily necessities and stationery;
the script generator analyzes the user input by combining the formulated video advertisement script model;
and the script generator integrates the analysis results and writes the information of each shot into the script based on different video types. For each lens, if the lens is a commodity lens, writing the visual characteristics of the lens into a video script to be shot; and aiming at each lens, if the lens is an empty lens, entering a material database to search for an empty lens material. And traversing the material database by the system, and if a material matched with the visual characteristics of the lens is found, writing the index of the material into the script for rendering.
The script generator analyzes user input by combining with a formulated video advertisement script model, and specifically comprises the following steps:
the video advertisement script model comprises a production process of a professional short video advertisement video, and specifically comprises a series of rules; the rules contain visual characteristics that are appropriate for short video advertisements of different styles and different goods. The script generator defines the advertisement duration, the advertisement music, the lens quantity and the advertisement rhythm of the short video advertisement based on the video advertisement script model; further, the script generator defines the video type (empty lens or commodity lens), composition, mirror moving mode, color characteristics (saturation, contrast, etc.) of each lens of the short video advertisement based on the video advertisement script model and the advertisement overall information (duration, music … …).
And step S02, the system analyzes the video script, plans the motion path of the mechanical arm and guides the user to set scenes. Referring to fig. 4, specifically, the following may be:
the system reads the script and converts visual characteristics of the lens in the script into mechanical arm movement path planning aiming at the commodity lens in the script;
converting the motion path of the mechanical arm into a corresponding instruction format according to a communication protocol of the system;
and displaying a line diagram on a mobile phone interface according to the initial position of the mechanical arm for shooting to guide a user to arrange shooting scenes. The wire frame diagram contains the position of the commodity and the position of the susceptor as well as the placing angle of the camera. The user sets under the guidance of the online block diagram and fixes the camera at the tail end of the mechanical arm. The mechanical arm regards the final state as an initial state of the photographing process.
Wherein, the visual characteristic of camera lens in the script is converted into arm motion path planning to the system, specifically is:
aiming at each shot in the script, if the shot is a commodity shot, the system analyzes the visual characteristics of the shot; otherwise, skipping to enter the next lens until all the lenses are read;
the shooting analysis can be divided into two-dimensional camera screen plane analysis, three-dimensional shooting space analysis and mechanical arm motion path planning analysis:
and analyzing the visual effect of each lens defined in the script by the two-dimensional camera screen plane, and calculating the motion speed and the motion track of the main object (namely the commodity) in the two-dimensional camera screen plane. The system first reads the resolution of the camera view, as well as the visual characteristics of the current shot in the script. The visual features include: the moving path of the commodity in the picture and the moving speed of the commodity in the picture. Wherein, the commodity motion path in the picture comprises picture composition and commodity motion direction; establishing a two-dimensional camera plane coordinate system by taking the upper left corner of a camera plane as an origin and the horizontal direction and the vertical direction as an x axis and a y axis respectively; according to the visual characteristics of the lens in the script, the abstract visual characteristics are quantized into the commodity motion states in the two-dimensional camera screen plane in the initial state and the termination state of the lens, wherein the commodity motion states comprise the coordinates of the center of the commodity in a two-dimensional camera plane coordinate system and the radius of an circumscribed circle presented by the commodity in a picture; according to the visual characteristics of the lens in the script, the position relation of the main object (namely the commodity) and the set object is defined. Finally, transmitting the analysis result to a three-dimensional shooting space for analysis;
and the three-dimensional shooting space analysis calculates and obtains the motion speed and the motion trail of the camera in the three-dimensional space according to the motion speed and the motion trail of the main object in the plane of the two-dimensional camera screen. The system firstly reads the focal length of a camera and the motion information of a commodity in a current lens picture; and establishing a three-dimensional space coordinate system. Calculating motion information of the camera in the three-dimensional space of the initial state and the termination state of the lens by combining the focal length of the camera according to the sizes and the positions of the commodity and the susceptor and through a projection theorem; finally, the analysis result is transmitted to the mechanical arm movement path planning analysis;
and planning and analyzing the motion path of the mechanical arm, calculating and obtaining the position information of each degree of freedom of the mechanical arm in the mechanical arm coordinate based on a reverse dynamics algorithm according to the motion speed and the motion track of the camera in the three-dimensional space, and planning the global motion track of the mechanical arm. The system firstly reads the motion information of the current lens camera; based on the inverse dynamics algorithm, the positions of the mechanical arms with the starting points and the ending points are calculated according to the positions of the cameras (namely the tail ends of the mechanical arms) with the current lens starting points and the current lens ending points. Finally, according to the communication protocol of the system, the motion path of the mechanical arm is converted into a corresponding instruction format, and fig. 6 is a communication protocol example, specifically:
and calculating the sequence numbers of the initial point and the initial point of the current lens in the motion action sequence of the mechanical arm according to the lens sequence numbers. N001 in fig. 6 represents the first motion of the robot arm in the motion sequence. L01 is action ID; the delay waiting means that after the lens finishes a certain action, the lens stays for a plurality of times and then enters the next action. And a delay waiting is arranged between actions, so that the shaking phenomenon of the mechanical arm can be effectively relieved. X, Y, Z, A, B, C correspond to the robot arms with six degrees of freedom, all with an origin of 0000.000. And coding the mechanical arm state information of the current lens starting point and the current lens ending point according to the format, and writing the mechanical arm state information into a mechanical arm script.
And S03, analyzing the state information of the motion of the object in the video picture to be clipped based on the salient region analysis algorithm, and labeling the video. And based on the label and in combination with the script, performing clipping and synthesizing processing on the video. The method specifically comprises the following steps:
performing frame-by-frame significance region analysis on a video to be edited by a significance algorithm; carrying out mean processing on every five frames to obtain the final salient regions and the salient region centers of the five frames;
calculating the central distance and deflection angle of the salient region of the nth frame and the (n-5) th frame by taking the five frames as step lengths from the sixth frame, setting the lens serial number as k, and setting the initial value as 1;
if the central deflection angle in the salient region of the nth frame and the (n-5) th frame and the central deflection angle in the salient region of the (n-5) th frame and the (n-10) th frame are higher than a threshold value, recording a label 'lens switching' for the nth frame; if the previous 'shot switching' frame does not comprise the 'kth commodity shot' label, the 'kth commodity shot' label is recorded for the current frame. K plus 1;
if the distance between the center of the salient region of the nth frame and the (n-5) th frame and the distance between the center of the salient region of the (n-5) th frame and the center of the salient region of the (n-10) th frame are lower than the threshold, and the distance between the center of the salient region of the nth frame and the (n-5) th frame and the distance between the center of the salient region of the (n-5) th frame and the center of the salient region of the (n-10) th frame are higher than the threshold, recording a label' acceleration of the kth commodity lens;
if the central deflection angle of the salient region of the nth frame and the (n-5) th frame and the central distance of the salient region of the (n-5) th frame and the (n-10) th frame are higher than the threshold value, recording a label 'the deceleration start of the k-th commodity lens' for the nth frame;
and finally performing composite rendering on the video according to the script. If the shot is an empty lens, extracting materials from the database according to material retrieval in the script; and if the shot is a commodity shot, performing editing and composition according to the script information and the label. The simplified flow chart of the above steps is shown in fig. 9.
Can also use a plurality of arms to cooperate the shooting action of accomplishing high complexity jointly, include: marking a plurality of robotic arms, in order a1, a2,. and aN, wherein N is a natural number of at least 2;
planning a three-dimensional space position to which the mechanical arms a1, a2, aN aN should reach at each moment, and which mechanical arms start shooting actions at each moment, wherein the mechanical arms which do not start the shooting actions can be advanced to a position required by the next key moment point, so that zero-time difference switching is realized during shooting of the next group;
planning the movement speed, movement tracks and interactive actions of a plurality of mechanical arms in a three-dimensional space, and ensuring that the mechanical arms are mutually matched and do not interfere with each other;
the multiple mechanical arms are matched together to finish multi-angle shooting of the articles at the same moment or special track shooting of the articles at different moments, and shooting actions which cannot be finished by the single mechanical arm are carried out.
As another aspect, the present application further provides a system for automatically shooting and generating a mobile-end short video advertisement based on a mechanical arm, where the system includes: a generator and a memory for storing a mobile terminal program capable of running on the generator; the generator is configured to implement the method described in the embodiment of the present application when the generator runs the mobile terminal program.
In another aspect, the present application further provides a computer-readable storage medium, on which mobile end device executable instructions are stored, where the mobile end device executable instructions are configured to execute a method for implementing the descriptions of the embodiments of the present application.
It should be understood that portions of the present application may be implemented in hardware, software, firmware, or a combination thereof. In the above embodiments, the various steps or methods may be implemented in software or firmware stored in memory and executed by a suitable instruction execution system. If implemented in hardware, as in another embodiment, any one or combination of the following techniques, which are known in the art, may be used: a discrete logic circuit having a logic gate circuit for realizing a logic function for a data signal, an asic having an appropriate combinational logic gate circuit, a Programmable Gate Array (PGA), a Field Programmable Gate Array (FPGA), and the like.
It will be understood by those skilled in the art that all or part of the steps carried by the method for implementing the above embodiments may be implemented by hardware related to instructions of a program, which may be stored in a computer readable storage medium, and when the program is executed, the program includes one or a combination of the steps of the method embodiments.
In addition, functional units in the embodiments of the present application may be integrated into one processing module, or each unit may exist alone physically, or two or more units are integrated into one module. The integrated module can be realized in a hardware mode, and can also be realized in a software functional module mode. The integrated module, if implemented in the form of a software functional module and sold or used as a stand-alone product, may also be stored in a computer readable storage medium.
The storage medium mentioned above may be a read-only memory, a magnetic or optical disk, etc. Although embodiments of the present application have been shown and described above, it is understood that the above embodiments are exemplary and should not be construed as limiting the present application, and that variations, modifications, substitutions and alterations may be made to the above embodiments by those of ordinary skill in the art within the scope of the present application.
Claims (14)
1. A method for automatically shooting and generating a mobile-end short video advertisement based on a mechanical arm comprises the following steps:
determining the video style and commodity information through user interaction, and generating a video script by combining a script generation model based on user requirements through the system;
the system analyzes the video script, plans the motion path of the mechanical arm and guides the user to set a scene;
and carrying out significance region analysis on the video to be clipped by utilizing a significance algorithm and synthesizing and rendering.
2. The method of claim 1, wherein the user interaction determines a video style and commodity information, and the system generates a video script based on the user demand in combination with a script generation model, comprising:
selecting a video style and a commodity type on a demand acquisition interface by a user as input;
the script generator analyzes the user input by combining the formulated video advertisement script model;
and the script generator integrates the analysis results and writes the information of each shot into the script based on different video types.
3. The method of claim 1, wherein the system analyzes the video script, plans the path of the robot arm movement, and guides the user through the set, comprising:
the system reads the script and converts visual characteristics of the lens in the script into mechanical arm movement path planning aiming at the commodity lens in the script;
converting the motion path of the mechanical arm into a corresponding instruction format according to a communication protocol of the system;
according to the initial position of the mechanical arm in shooting, a line frame diagram is displayed on a mobile phone interface to guide a user to arrange a shooting scene, the line frame diagram comprises the positions of commodities, the positions of the susceptors and the placing angles of the cameras, the user sets scenes under the guidance of the line frame diagram and fixes the cameras at the tail end of the mechanical arm, and the mechanical arm regards the final state as the initial state of the shooting process.
4. The method of claim 1, wherein performing saliency region analysis and synthetic rendering on the video to be clipped by using a saliency algorithm comprises:
the saliency algorithm analyzes the saliency areas of the video to be edited frame by frame, and performs mean processing on every five frames to obtain the saliency areas and the saliency area centers of the five frames finally;
starting from the sixth frame, calculating the central distance and deflection angle of the salient region of the nth frame and the (n-5) th frame with the step length of five frames, setting the lens serial number as k, and setting the initial value as 1;
if the central deflection angle in the salient region of the nth frame and the (n-5) th frame and the central deflection angle in the salient region of the (n-5) th frame and the (n-10) th frame are higher than a threshold value, recording a label 'lens switching' for the nth frame; if the previous 'shot switching' frame does not comprise a 'kth commodity shot' label, a current frame recording label 'kth commodity shot' is marked, and 1 is added to K;
the distance between the center of the salient region of the nth frame and the (n-5) th frame and the distance between the center of the salient region of the (n-5) th frame and the center of the salient region of the (n-10) th frame are lower than a threshold value, and the distance between the center of the salient region of the nth frame and the (n-5) th frame and the distance between the center of the salient region of the (n-5) th frame and the center of the salient region of the (n-10) th frame are higher than the threshold value, then the nth frame is recorded with a label;
if the central deflection angle of the salient region of the nth frame and the (n-5) th frame and the central distance of the salient region of the (n-5) th frame and the (n-10) th frame are higher than the threshold value, recording a label 'the deceleration start of the k-th commodity lens' for the nth frame;
performing final synthesis rendering on the video according to the script, and if the video is an empty lens, extracting materials from a database according to material retrieval in the script; and if the shot is a commodity shot, performing editing and composition according to the script information and the label.
5. The method of claim 2, wherein the script generator analyzes the user input in conjunction with the formulated video advertisement script model, comprising:
the video advertisement script model comprises a production process of a professional short video advertisement video, and specifically comprises a series of rules; the rules contain visual characteristics suitable for short video advertisements of different styles and different commodities;
the script generator defines the advertisement duration, the advertisement music, the lens quantity and the advertisement rhythm of the short video advertisement based on the video advertisement script model;
the script generator defines the video type, composition, mirror moving mode and color characteristics of each lens of the short video advertisement based on the video advertisement script model and the advertisement overall information.
6. The method of claim 2, wherein the script generator integrates the analysis results, comprising:
for each lens, if the lens is a commodity lens, writing the visual characteristics of the lens into a video script to be shot;
and for each lens, if the lens is an empty lens, entering a material database to search for an empty lens material, traversing the material database by the system, and if a material matched with the visual characteristics of the lens is found, writing the index of the material into a script to be rendered for use.
7. The method of claim 3, wherein the system converts the visual characteristics of the shots in the script into the mechanical arm motion path plan, comprising:
aiming at each lens in the script, if the lens is a commodity lens, the system carries out shooting analysis on the visual characteristics of the lens; otherwise, skipping to enter the next lens until all the lenses are read;
the shooting analysis can be divided into two-dimensional camera screen plane analysis, three-dimensional shooting space analysis and mechanical arm motion path planning analysis:
analyzing the visual effect of each lens defined in the script by the two-dimensional camera screen plane, and calculating the motion speed and the motion track of the main object in the two-dimensional camera plane;
the three-dimensional shooting space analysis is used for calculating and obtaining the motion speed and the motion trail of the camera in the three-dimensional space according to the motion speed and the motion trail of the main object in the plane of the two-dimensional camera;
and the mechanical arm motion analysis is used for calculating and obtaining the position information of each degree of freedom of the mechanical arm in the mechanical arm coordinate based on a reverse dynamics algorithm according to the motion speed and the motion track of the camera in the three-dimensional space, and planning the overall motion track of the mechanical arm.
8. The method of claim 3, wherein converting the robot arm motion path into a corresponding command format according to a communication protocol of the system comprises:
calculating the serial numbers of the initial point and the initial point of the current lens in the motion sequence of the mechanical arm according to the lens serial numbers, wherein N001 represents the first motion of the mechanical arm in the motion sequence, and L01 represents a motion ID;
the time delay waiting means that after a certain action is finished by the lens, the lens stays for a plurality of times and then enters the next action, and the time delay waiting is set between the actions, so that the shaking phenomenon of the mechanical arm can be effectively relieved;
x, Y, Z, A, B, C respectively corresponding to the mechanical arm with six degrees of freedom, the original points are all 0000.000;
and coding the mechanical arm state information of the current lens starting point and the current lens ending point according to the format, and writing the mechanical arm state information into a mechanical arm script.
9. The method of claim 7, wherein the two-dimensional camera screen plane analyzing the visual effect for each lens defined in the script calculates motion information of objects in the two-dimensional camera plane, comprising:
reading the visual characteristics of the current lens in the script, wherein the visual characteristics comprise: the system comprises an intra-picture commodity motion path and an intra-picture commodity motion speed, wherein the intra-picture commodity motion path comprises a picture composition and a commodity motion direction;
reading the resolution of a camera picture;
establishing a two-dimensional camera plane coordinate system by taking the upper left corner of a camera plane as an origin and the horizontal direction and the vertical direction as an x axis and a y axis respectively;
according to the visual characteristics of the lens in the script, the abstract visual characteristics are quantized into the commodity motion states in the two-dimensional camera screen plane in the initial state and the termination state of the lens, wherein the commodity motion states comprise the coordinates of the center of the commodity in a two-dimensional camera plane coordinate system and the radius of an circumscribed circle presented by the commodity in a picture;
defining the position relation of the main object article and the dummy article according to the visual characteristics of the lens in the script;
and transmitting the analysis result to a three-dimensional shooting space for analysis.
10. The method of claim 7, wherein the three-dimensional shooting space analysis is used for calculating and obtaining the motion information of the camera in the three-dimensional space according to the motion information of the subject article in the two-dimensional camera plane, and comprises the following steps:
reading commodity motion information in a current lens picture;
reading a focal distance of a camera;
establishing a three-dimensional space coordinate system, and calculating motion information of the camera in the three-dimensional space of the initial state and the termination state of the lens through a projection theorem according to the sizes and the positions of the commodity and the susceptor and by combining the focal length of the camera;
and transmitting the analysis result to the mechanical arm motion path planning analysis.
11. The method of claim 7, wherein the mechanical arm motion path planning analysis is based on a reverse dynamics algorithm to calculate and plan a mechanical arm global motion track, and comprises the following steps:
reading the current lens camera motion information;
based on a reverse dynamics algorithm, the positions of the mechanical arms with the starting points and the ending points are calculated according to the positions of the cameras at the starting points and the ending points of the current lens, namely the tail ends of the mechanical arms.
12. The method of claim 7, wherein the high complexity photographing action is performed using a plurality of robotic arms cooperatively, comprising: marking a plurality of robotic arms, in order a1, a2,. and aN, wherein N is a natural number of at least 2;
planning a three-dimensional space position to which the mechanical arms a1, a2, aN aN should reach at each moment, and which mechanical arms start shooting actions at each moment, wherein the mechanical arms which do not start the shooting actions are advanced to a position required by the next key moment point, so that zero-time difference switching is realized during shooting of the next group;
planning the movement speed, movement tracks and interactive actions of a plurality of mechanical arms in a three-dimensional space, and ensuring that the mechanical arms are mutually matched and do not interfere with each other;
the multiple mechanical arms are matched together to finish multi-angle shooting of the articles at the same moment or special track shooting of the articles at different moments, and shooting actions which cannot be finished by the single mechanical arm are carried out.
13. A system for automatically shooting and generating a mobile terminal short video advertisement based on a mechanical arm is characterized by comprising: a generator and a memory for storing a mobile terminal program capable of running on the generator; wherein the generator is configured to execute the method according to any one of claims 1 to 12 when running the mobile terminal program.
14. A mobile end device-readable storage medium having mobile end device-executable instructions stored thereon, wherein the mobile end device-executable instructions are configured to perform the method according to any one of claims 1 to 12.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010452022.7A CN111629269B (en) | 2020-05-25 | 2020-05-25 | Method for automatically shooting and generating mobile terminal short video advertisement based on mechanical arm |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010452022.7A CN111629269B (en) | 2020-05-25 | 2020-05-25 | Method for automatically shooting and generating mobile terminal short video advertisement based on mechanical arm |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111629269A true CN111629269A (en) | 2020-09-04 |
CN111629269B CN111629269B (en) | 2021-05-25 |
Family
ID=72261198
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010452022.7A Expired - Fee Related CN111629269B (en) | 2020-05-25 | 2020-05-25 | Method for automatically shooting and generating mobile terminal short video advertisement based on mechanical arm |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111629269B (en) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112422831A (en) * | 2020-11-20 | 2021-02-26 | 广州太平洋电脑信息咨询有限公司 | Video generation method and device, computer equipment and storage medium |
CN112862861A (en) * | 2021-02-08 | 2021-05-28 | 广州富港生活智能科技有限公司 | Camera motion path determining method and device and shooting system |
CN113242470A (en) * | 2021-06-15 | 2021-08-10 | 广州聚焦网络技术有限公司 | Video publishing method and device applied to foreign trade marketing |
CN113590247A (en) * | 2021-07-21 | 2021-11-02 | 阿里巴巴达摩院(杭州)科技有限公司 | Text creation method and computer program product |
CN113596283A (en) * | 2021-07-28 | 2021-11-02 | 杭州更火数字科技有限公司 | Video customization method and system and electronic equipment |
CN113660526A (en) * | 2021-10-18 | 2021-11-16 | 阿里巴巴达摩院(杭州)科技有限公司 | Script generation method, system, computer storage medium and computer program product |
CN114363549A (en) * | 2022-01-12 | 2022-04-15 | 关晓辉 | Intelligent script walking show recording processing method, device and system |
CN114501061A (en) * | 2022-01-25 | 2022-05-13 | 上海影谱科技有限公司 | Video frame alignment method and system based on object detection |
CN114827456A (en) * | 2022-04-09 | 2022-07-29 | 王可 | Control system for intelligent studio |
CN117336539A (en) * | 2023-09-28 | 2024-01-02 | 北京风平智能科技有限公司 | Video script production method and system for short video IP (Internet protocol) construction |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN202433318U (en) * | 2012-01-06 | 2012-09-12 | 成都旭双太阳能科技有限公司 | Visual detection device for armorphous silicon solar thin film battery |
CN103660514A (en) * | 2013-12-06 | 2014-03-26 | 苏州逸美德自动化科技有限公司 | Glue punching, film pasting and glue adding integration machine |
CN104808590A (en) * | 2015-02-14 | 2015-07-29 | 浙江大学 | Mobile robot visual servo control method based on key frame strategy |
CN108833782A (en) * | 2018-06-20 | 2018-11-16 | 广州长鹏光电科技有限公司 | A kind of positioning device and method based on video auto-tracking shooting |
CN109002857A (en) * | 2018-07-23 | 2018-12-14 | 厦门大学 | A kind of transformation of video style and automatic generation method and system based on deep learning |
CN109631793A (en) * | 2018-12-12 | 2019-04-16 | 上海卫星装备研究所 | Type face digital photography automatic measurement method |
CN110414446A (en) * | 2019-07-31 | 2019-11-05 | 广东工业大学 | The operational order sequence generating method and device of robot |
CN110557568A (en) * | 2019-09-05 | 2019-12-10 | 昭世(北京)科技有限公司 | structure and method of module combined photographic equipment based on artificial intelligence |
-
2020
- 2020-05-25 CN CN202010452022.7A patent/CN111629269B/en not_active Expired - Fee Related
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN202433318U (en) * | 2012-01-06 | 2012-09-12 | 成都旭双太阳能科技有限公司 | Visual detection device for armorphous silicon solar thin film battery |
CN103660514A (en) * | 2013-12-06 | 2014-03-26 | 苏州逸美德自动化科技有限公司 | Glue punching, film pasting and glue adding integration machine |
CN104808590A (en) * | 2015-02-14 | 2015-07-29 | 浙江大学 | Mobile robot visual servo control method based on key frame strategy |
CN108833782A (en) * | 2018-06-20 | 2018-11-16 | 广州长鹏光电科技有限公司 | A kind of positioning device and method based on video auto-tracking shooting |
CN109002857A (en) * | 2018-07-23 | 2018-12-14 | 厦门大学 | A kind of transformation of video style and automatic generation method and system based on deep learning |
CN109631793A (en) * | 2018-12-12 | 2019-04-16 | 上海卫星装备研究所 | Type face digital photography automatic measurement method |
CN110414446A (en) * | 2019-07-31 | 2019-11-05 | 广东工业大学 | The operational order sequence generating method and device of robot |
CN110557568A (en) * | 2019-09-05 | 2019-12-10 | 昭世(北京)科技有限公司 | structure and method of module combined photographic equipment based on artificial intelligence |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112422831A (en) * | 2020-11-20 | 2021-02-26 | 广州太平洋电脑信息咨询有限公司 | Video generation method and device, computer equipment and storage medium |
CN112862861A (en) * | 2021-02-08 | 2021-05-28 | 广州富港生活智能科技有限公司 | Camera motion path determining method and device and shooting system |
CN112862861B (en) * | 2021-02-08 | 2024-05-07 | 广州富港生活智能科技有限公司 | Camera motion path determining method, determining device and shooting system |
CN113242470A (en) * | 2021-06-15 | 2021-08-10 | 广州聚焦网络技术有限公司 | Video publishing method and device applied to foreign trade marketing |
CN113590247B (en) * | 2021-07-21 | 2024-04-05 | 杭州阿里云飞天信息技术有限公司 | Text creation method and computer program product |
CN113590247A (en) * | 2021-07-21 | 2021-11-02 | 阿里巴巴达摩院(杭州)科技有限公司 | Text creation method and computer program product |
CN113596283A (en) * | 2021-07-28 | 2021-11-02 | 杭州更火数字科技有限公司 | Video customization method and system and electronic equipment |
CN113660526A (en) * | 2021-10-18 | 2021-11-16 | 阿里巴巴达摩院(杭州)科技有限公司 | Script generation method, system, computer storage medium and computer program product |
CN113660526B (en) * | 2021-10-18 | 2022-03-11 | 阿里巴巴达摩院(杭州)科技有限公司 | Script generation method, system, computer storage medium and computer program product |
CN114363549A (en) * | 2022-01-12 | 2022-04-15 | 关晓辉 | Intelligent script walking show recording processing method, device and system |
CN114501061A (en) * | 2022-01-25 | 2022-05-13 | 上海影谱科技有限公司 | Video frame alignment method and system based on object detection |
CN114501061B (en) * | 2022-01-25 | 2024-03-15 | 上海影谱科技有限公司 | Video frame alignment method and system based on object detection |
CN114827456B (en) * | 2022-04-09 | 2023-08-22 | 王可 | Control system for intelligent film studio |
CN114827456A (en) * | 2022-04-09 | 2022-07-29 | 王可 | Control system for intelligent studio |
CN117336539A (en) * | 2023-09-28 | 2024-01-02 | 北京风平智能科技有限公司 | Video script production method and system for short video IP (Internet protocol) construction |
CN117336539B (en) * | 2023-09-28 | 2024-05-14 | 北京风平智能科技有限公司 | Video script production method and system for short video IP (Internet protocol) construction |
Also Published As
Publication number | Publication date |
---|---|
CN111629269B (en) | 2021-05-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111629269B (en) | Method for automatically shooting and generating mobile terminal short video advertisement based on mechanical arm | |
CN108573527B (en) | Expression picture generation method and equipment and storage medium thereof | |
CN112700523B (en) | Virtual object face animation generation method and device, storage medium and terminal | |
KR101635730B1 (en) | Apparatus and method for generating montage, recording medium for performing the method | |
Cosatto et al. | Sample-based synthesis of photo-realistic talking heads | |
Cosatto et al. | Photo-realistic talking-heads from image samples | |
JP5355422B2 (en) | Method and system for video indexing and video synopsis | |
US6504546B1 (en) | Method of modeling objects to synthesize three-dimensional, photo-realistic animations | |
US8692940B2 (en) | Method for producing a blended video sequence | |
US6492990B1 (en) | Method for the automatic computerized audio visual dubbing of movies | |
JP2020071851A (en) | Method and apparatus for live broadcasting with avatar | |
US11184558B1 (en) | System for automatic video reframing | |
US20070165022A1 (en) | Method and system for the automatic computerized audio visual dubbing of movies | |
MXPA03001171A (en) | Image conversion and encoding techniques. | |
KR20030036747A (en) | Method and apparatus for superimposing a user image in an original image | |
US11282257B2 (en) | Pose selection and animation of characters using video data and training techniques | |
CN108320334B (en) | Method for establishing three-dimensional scene roaming system based on point cloud | |
US20030085901A1 (en) | Method and system for the automatic computerized audio visual dubbing of movies | |
CN110324728A (en) | The competitive sports whole audience based on deeply study looks back short video generation method | |
US20210304447A1 (en) | Customizing soundtracks and hairstyles in modifiable videos of multimedia messaging application | |
CA2741461C (en) | Automated cinematographic editing tool | |
KR20230130748A (en) | Image processing methods and apparatus, devices and media | |
KR101313285B1 (en) | Method and Device for Authoring Information File of Hyper Video and Computer-readable Recording Medium for the same | |
US11024071B2 (en) | Method of converting phoneme transcription data into lip sync animation data for 3D animation software | |
US6337916B1 (en) | Modifying image data |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20210525 |