CN113873080B - Multimedia file acquisition method and device - Google Patents

Multimedia file acquisition method and device Download PDF

Info

Publication number
CN113873080B
CN113873080B CN202111133380.2A CN202111133380A CN113873080B CN 113873080 B CN113873080 B CN 113873080B CN 202111133380 A CN202111133380 A CN 202111133380A CN 113873080 B CN113873080 B CN 113873080B
Authority
CN
China
Prior art keywords
multimedia file
objects
input
multimedia
information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111133380.2A
Other languages
Chinese (zh)
Other versions
CN113873080A (en
Inventor
杜享平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vivo Mobile Communication Co Ltd
Original Assignee
Vivo Mobile Communication Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vivo Mobile Communication Co Ltd filed Critical Vivo Mobile Communication Co Ltd
Priority to CN202111133380.2A priority Critical patent/CN113873080B/en
Publication of CN113873080A publication Critical patent/CN113873080A/en
Application granted granted Critical
Publication of CN113873080B publication Critical patent/CN113873080B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72439User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for image or video messaging

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Human Computer Interaction (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The application discloses a method and a device for acquiring a multimedia file, and belongs to the technical field of electronic equipment. The method for acquiring the multimedia file comprises the following steps: receiving first input of a user to the shooting preview picture, and determining at least two types of targets according to the first input; a second input is received and a plurality of multimedia files are generated in response to the second input, wherein each multimedia file contains a class of objects.

Description

Multimedia file acquisition method and device
Technical Field
The application belongs to the technical field of electronic equipment, and particularly relates to a method and a device for acquiring a multimedia file.
Background
At present, photographing (shooting) becomes a core function and a selling point of electronic equipment, and meanwhile, a user can take pictures (shooting) by using the electronic equipment and also becomes an important scene in life.
In the related art, a user often generates only one photo (video) each time the user takes a picture. If a user needs to change a different photo (video), the user needs to shoot again, the user can continuously and repeatedly operate the camera, and the operation steps are troublesome.
Disclosure of Invention
An object of the embodiments of the present application is to provide a method and an apparatus for acquiring a multimedia file of a multimedia file, which can solve the problem in the related art that only one multimedia file can be acquired by one operation.
In a first aspect, an embodiment of the present application provides a method for acquiring a multimedia file, including: receiving a first input of a user to the shooting preview picture, and determining at least two types of targets according to the first input; a second input is received and a plurality of multimedia files are generated in response to the second input, wherein each multimedia file contains a class of objects.
In a second aspect, an embodiment of the present application provides an apparatus for acquiring a multimedia file, including: the determining unit is used for receiving a first input of a user to the shooting preview picture and determining at least two types of targets according to the first input; and the processing unit is used for receiving a second input and responding to the second input to generate a plurality of multimedia files, wherein each multimedia file comprises a class of targets.
In a third aspect, an embodiment of the present application provides an electronic device, which includes a processor, a memory, and a program or instructions stored in the memory and executable on the processor, where the program or instructions, when executed by the processor, implement the steps of the method for acquiring a multimedia file according to the first aspect.
In a fourth aspect, the present application provides a readable storage medium, on which a program or instructions are stored, and when executed by a processor, the program or instructions implement the steps of the multimedia file acquiring method according to the first aspect.
In a fifth aspect, an embodiment of the present application provides a chip, where the chip includes a processor and a communication interface, where the communication interface is coupled to the processor, and the processor is configured to execute a program or instructions to implement the method for acquiring a multimedia file according to the first aspect.
In the embodiment of the application, a plurality of multimedia files can be obtained through one-time operation, and the convenience and the practicability of multimedia file acquisition are improved. In addition, different multimedia files correspond to different types of targets in the application, so that each multimedia file can have own style, and the visual effect of the multimedia files is further improved.
Drawings
Fig. 1 is a flowchart of a multimedia file acquiring method according to an embodiment of the present application;
fig. 2 is a schematic block diagram of an apparatus for acquiring a multimedia file according to an embodiment of the present application;
FIG. 3 is a schematic diagram of a screen display of an electronic device according to an embodiment of the present application;
FIG. 4 is a second schematic diagram of a screen display of an electronic device according to an embodiment of the present application;
FIG. 5 is a third schematic diagram of a screen display of an electronic device according to an embodiment of the present application;
FIG. 6 is a fourth schematic diagram of a screen display of an electronic device according to an embodiment of the present application;
FIG. 7 is a fifth schematic view of a screen display of an electronic device according to an embodiment of the present application;
FIG. 8 is a sixth exemplary screen display of an electronic device according to an embodiment of the present application;
FIG. 9 is a seventh schematic view of a screen display of an electronic device according to an embodiment of the present application;
FIG. 10 is an eighth schematic diagram of a screen display of an electronic device according to an embodiment of the present application;
FIG. 11 is a ninth illustration of a screen display of an electronic device according to an embodiment of the disclosure;
FIG. 12 is one of the schematic block diagrams of an electronic device of an embodiment of the present application;
FIG. 13 is a second schematic block diagram of an electronic device according to an embodiment of the present application;
Detailed Description
The technical solutions in the embodiments of the present application will be described clearly below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, embodiments of the present application. All other embodiments that can be derived by one of ordinary skill in the art from the embodiments given herein are intended to be within the scope of the present disclosure.
The terms first, second and the like in the description and in the claims of the present application are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It will be appreciated that the data so used may be interchanged under appropriate circumstances such that embodiments of the application are capable of operation in sequences other than those illustrated or described herein, and that the terms "first," "second," etc. are generally used in a generic sense and do not limit the number of terms, e.g., a first term can be one or more than one. In addition, "and/or" in the specification and claims means at least one of connected objects, a character "/" generally means that a preceding and succeeding related objects are in an "or" relationship.
The method for acquiring a multimedia file, the apparatus for acquiring a multimedia file, the electronic device, and the readable storage medium provided in the embodiments of the present application are described in detail below with reference to the accompanying drawings.
It should be noted that, in the method for acquiring a multimedia file provided in the embodiment of the present application, the execution main body may be an acquisition apparatus for a multimedia file, or a control module in the acquisition apparatus for a multimedia file, which is used for executing the method for acquiring a multimedia file. In the embodiment of the present application, a method for executing an acquisition of a multimedia file by an acquisition device of the multimedia file is taken as an example, and the acquisition device of the multimedia file provided in the embodiment of the present application is described.
An embodiment of a first aspect of the present application provides a method for acquiring a multimedia file, as shown in fig. 1, the method for acquiring a multimedia file includes:
102, receiving a first input of a user to a shooting preview picture, and determining at least two types of targets according to the first input;
step 104, receiving a second input, and generating a plurality of multimedia files in response to the second input, wherein each multimedia file comprises a class of objects.
The method for acquiring the multimedia file comprises the steps of firstly receiving first input of a user for shooting a preview picture, and determining at least two types of targets according to the first input; then, a second input is received and a plurality of multimedia files are generated in response to the second input. In particular, each multimedia file contains a class of objects. Specifically, the multi-type objects are respectively displayed on the multi-media files, and then the multi-media files are obtained. Like this, this application accessible one-time operation obtains a plurality of multimedia files, has promoted the convenience and the practicality that multimedia file acquireed. In addition, different multimedia files correspond to different types of targets in the application, so that each multimedia file can have own style, and the visual effect of the multimedia files is further improved.
For example, when the present application displays the photographed preview screen shown in fig. 4, the multimedia files shown in fig. 5 and 6 may be respectively generated in the course of one operation; when the present application displays the photographed preview screen as shown in fig. 7, multimedia files as shown in fig. 8 and 9 may be respectively generated in the course of one operation.
In one embodiment of the present application, the multimedia file comprises a photograph or video. When the multimedia photos are photos, a plurality of photos can be obtained through one-time photographing operation, and each photo is provided with a type of target, so that each photo can have the style of the photo, and the photographing effect is further improved. When the multimedia photo is a video, a plurality of videos can be obtained through one-time shooting operation, and each video is provided with a type of target, so that each video can have the style of the video, and the shooting effect is improved.
In one embodiment of the present application, in response to a first input, the present application may determine at least two objects specified by the first input in the photographing preview screen; then, the present application may display the mark information of the at least two objects, respectively. In particular, for any one tag information, all objects having the tag information are a kind of objects. Thus, a class of objects includes one or more objects.
Specifically, the method for acquiring the multimedia file provided by the application has a manual mode. In the manual mode, the user can freely mark an object in each of the photographed preview pictures. Specifically, during the operation, a plurality of objects are displayed on the shooting preview screen, and the user can select at least two objects in the shooting preview screen through a first input to the shooting preview screen and mark the at least two objects respectively to display mark information for the at least two objects respectively. In this way, for any piece of mark information, the object having the mark information in the shooting preview picture is the same type of object. Therefore, in the subsequent process of generating the multimedia file, one or more objects with the same marking information can be generated into one multimedia file.
Specifically, in this embodiment, the first input is a mark input. That is, the user directly marks at least two objects through the first input. Therefore, in this embodiment, the user can mark different objects according to his/her preference, and freely decide which objects in the shooting preview screen are generated in the same multimedia file.
For example, the marking information may adopt numbers of 1, 2, 3, 4, 5, etc., or may adopt different figures (for example, the marking information may adopt polygons, and the different marking information is distinguished by the number of sides of the polygons).
In one embodiment of the application, in response to a first input, the application automatically identifies the categories of at least two objects in the shooting preview picture, and displays mark information of the at least two objects according to the category of each object. In particular, for any one piece of tag information, all objects having the tag information are one type of object. Thus, a class of objects includes one or more objects. Specifically, the above categories may include people, animals, landscapes, and the like.
Specifically, the method for acquiring the multimedia file has an automatic mode. In the automatic mode, the method and the device can automatically identify the categories of at least two objects in a shooting preview picture; then, corresponding mark information is automatically displayed for each object according to the category of the object. In this way, for any piece of mark information, the object with the mark information in the shooting preview picture is the same type of object. Therefore, in the subsequent process of generating the multimedia file, one or more objects with the same marking information can be generated into one multimedia file. Specifically, as for identifying the category of the object, AI identification and other related technologies can be adopted, and the discussion is not provided herein, which can be understood by those skilled in the art.
Specifically, in this embodiment, the first input is a start input. That is, the user starts the automatic mode through the first input, and then the mark information is automatically displayed for at least two objects in the automatic mode. Therefore, in this embodiment, it is possible to automatically recognize at least two objects in the photographing preview screen, mark the at least two objects according to the recognition result, and display mark information of the objects. Therefore, the automation degree in the whole process is high, and the acquisition convenience of the multimedia file can be greatly improved.
For example, the marking information may adopt numbers of 1, 2, 3, 4, 5, etc., or may adopt different figures (for example, the marking information may adopt polygons, and the different marking information is distinguished by the number of sides of the polygons).
Further, in one embodiment of the present application, in the process of generating a plurality of multimedia files in response to the second input, for any tag information, the present application obtains all objects with the tag information; all objects with this tag information are then generated as one photo. Correspondingly, when the shooting preview picture has a plurality of mark information, the method can repeat the steps for a plurality of times, and then generate a plurality of photos.
Specifically, in the process of generating a photo, one piece of mark information in a shooting preview picture is selected as first target mark information, and all objects with the first target mark information are obtained; then, all objects having the first object tag information are generated in the first multimedia file. Correspondingly, after the first photo is generated, selecting next mark information as second target mark information, and acquiring all objects with the second target mark information; then, all objects having the second object tag information are generated in a second multimedia file. And repeating the steps until all the mark information in the shooting preview picture is selected as the target mark object and the corresponding multimedia file generation processing is finished.
Specifically, after all the photos are generated, the number of the multimedia files corresponds to the number of the categories of the mark information. That is, if there are two types of mark information in the photographed preview screen, two multimedia files are finally generated; correspondingly, if N types of mark information exist in the shooting preview picture, N multimedia files are finally generated.
For example, as shown in fig. 4, the shooting preview screen has mark information 1 and mark information 2 therein; the present application first acquires an object having tag information 1, acquires images of the two persons, and generates the images of the two persons in a multimedia file as shown in fig. 5. Then, the present application first acquires the object with the label information 2, and the present application acquires images of the sun and the butterfly, and generates the images of the sun and the butterfly in the multimedia file shown in fig. 6.
Further, in an embodiment of the present application, the present application may generate a silhouette area of an object as a multimedia file.
For example, as shown in fig. 4, the present application may mark silhouette areas of a person image (specifically, silhouette areas of a person) as mark information 1, and mark silhouette areas of an animal image and a scene image (specifically, silhouette areas of a butterfly and a sun) as mark information 2. Thus, as shown in fig. 5, the image (person) in the silhouette area marked 1 can generate a multimedia file, and as shown in fig. 6, the image (butterfly and sun) in the silhouette area marked 2 can generate a multimedia file.
Further, in an embodiment of the present application, the present application may generate the selected area where the object is located as a multimedia file. Moreover, the application allows the border of the selection area to be adjusted, so that the size and scale of the picture of the selection area presented on the final multimedia file can be adjusted.
For example, as shown in fig. 8 and 9, the present application directly obtains an image corresponding to a silhouette area of a photo-object in a selected area and a background image in the selected area, and generates a multimedia file from the image corresponding to the silhouette area and the background image in the selected area.
Further, in one embodiment of the present application, each tag information has a corresponding priority, and different tag information has different priority orders. In the process of generating the multimedia file, different targets with different marking information are sequenced according to the priority order of the different marking information, and the targets with different marking information are sequentially generated on different multimedia files.
For example, the tag information includes first tag information and second tag information, and the priority of the first tag information is higher than the priority of the second tag information. In the process of generating the multimedia file, firstly, generating all objects with first mark information in a first multimedia file; then, all objects with the second tag information are generated in the first multimedia file.
For example, the priority of tag information 1 is higher than that of tag information 2 in fig. 4, and the multimedia file shown in fig. 5 is generated first and then the multimedia file shown in fig. 6 is generated; in fig. 7, the priority of the tag information 1 is higher than that of the tag information 2, and the multimedia file shown in fig. 8 is generated first, and then the multimedia file shown in fig. 9 is generated.
Further, in an embodiment of the present application, as shown in fig. 10, after the multimedia file is generated, the present application may add a background to the generated multimedia file. Specifically, the method and the device can receive a third input, respond to the third input and obtain at least two background pictures, and then fuse the generated multimedia file with the selected background pictures to further obtain the multimedia file with the background.
Specifically, the user may directly select a corresponding number of background photos from the background, or may preview the alternative background photos first, then select a corresponding number of background photos from the previewed background images, and respectively fuse the corresponding number of photos with the corresponding number of background photos.
For example, as shown in fig. 10, the present application may be a multimedia file fusion background photo 1 generated corresponding to the marking information 1 in fig. 10, and a multimedia file fusion background photo 2 generated corresponding to the marking information 2 in fig. 10.
For example, as shown in fig. 11, the present application previews four background pictures (background picture 1, background picture 2, background picture 3, and background picture 4, respectively); and then, selecting one of the four background pictures to be fused with the multimedia file generated corresponding to the tag information 1, and selecting one of the four background pictures to be fused with the multimedia file generated corresponding to the tag information 2.
Further, in an embodiment of the present application, before fusing the generated multimedia file and the selected background photos, the present application first orders at least two background photos; specifically, the background photos may be sorted according to a selection order, with the selected background photo in front and the selected background photo in back. Then, sequencing the generated multimedia files; specifically, according to the priority order of the tag information corresponding to the object in the multimedia file, the multimedia file with high priority is before the multimedia file with low priority.
Furthermore, in the process of fusing the multimedia file and the selected background picture, one multimedia file and one background picture are selected each time according to the sequence, and the generated multimedia file is fused with the selected background picture to further obtain the multimedia file with the background. Correspondingly, repeating the selection and fusion process for multiple times until the fusion of all the multimedia files and the background photos is completed.
The following further explains the method for acquiring a multimedia file according to a specific embodiment:
according to the method for acquiring the multimedia file, different objects in the picture can be marked firstly when the preview picture is shot, and all objects with the marked information are targets of the same type aiming at any marked information; then, the same multimedia file is generated by the same type of object according to the marking information obtained by marking so as to obtain a plurality of multimedia files. In addition, the method and the device can support the generation of the sequence of the multimedia files according to the priority of the marking sequence and can also support the addition of different backgrounds to different multimedia files.
In the following, taking an example of generating a silhouette outline area of an object as a multimedia file, the method for acquiring a multimedia file provided by the present application is explained as follows:
in the using process, as shown in fig. 3, the user firstly opens the photographing function of the electronic device and previews the picture to be photographed. Then, as shown in fig. 4, one piece of tag information (e.g., tag information 1 and tag information 2 in fig. 4) is generated for each object according to the attributes of the object. Then, as shown in fig. 5 and 6, in the process of generating a multimedia file, for any tag information, all objects having tag information are acquired, and the same type of object having the tag information is generated as one multimedia file (that is, the object having tag information 1 in fig. 4 is generated as one multimedia file, and the object having tag information 2 in fig. 4 is generated as one multimedia file).
Further, as shown in FIG. 4, if the tag information is a number, the size of the number may represent the order in which the multimedia files are generated, such as the image of the "1" tag displaying the first multimedia file, and so on.
Further, as shown in fig. 10 and 11, the present application may add respective background photos to the separately generated multimedia files, so as to enrich the scenes and the practicability of the multimedia files, and each multimedia file has a separate background.
Further, as shown in fig. 5 and 6, when the user presses the photographing key, the terminal camera system generates a plurality of multimedia files according to the tag information set by the user.
Specifically, in the process of marking the object, the marking may be automatic or manual. The mark information may be a number, a figure, or the like.
In the manual marking mode, a user can mark images in the silhouette area of an object according to the attributes (objects, characters and the like) of the object in a shooting preview picture, and if the images are marked as the same group of numbers, the images can be considered as the same group, so that the same group of type marking information can be displayed in a multimedia file when the multimedia file is generated later (for example, a character with a mark 1 is generated in the multimedia file, and a butterfly with a mark information 2 and a sun are generated in the multimedia file); if there is only one type of information, the objects of the type of information are displayed in a multimedia file alone. In particular, for an object without tag information, it may or may not be generated in each multimedia file.
In the automatic marking mode, images in the silhouette outline area of the object can be automatically classified according to the attributes of the object in the shooting preview picture; for example, in fig. 5 and 6, a person is taken as a category, a scene is taken as a category, and a single category can generate a picture.
In the following, taking the example of generating the selection area where the object is located as a multimedia file, the method for acquiring the multimedia file proposed in the present application is explained as follows:
in the using process, as shown in fig. 3, the user firstly opens the photographing function of the electronic device and previews the picture to be photographed. Then, as shown in fig. 7, one piece of label information (for example, label information 1 and label information 2 in fig. 7) is generated for each selection area. Then, as shown in fig. 8 and 9, in the process of generating a photograph, for any piece of mark information, all the selection areas having the mark information are acquired, and images within all the selection areas having the mark information are generated into one multimedia file (that is, an image within a selection area having mark information 1 in fig. 7 is generated into one multimedia file, and an image within a selection area having mark information 2 in fig. 7 is generated into one multimedia file).
Specifically, as shown in fig. 7, the user divides different areas of the screen in the shooting preview screen, and each divided selection area can generate a single multimedia file. In addition, the user can drag the boundary of a selection area, and the enclosed part of each selection area is a photographing area.
In addition, the marking mode of the marking information can be an automatic mode or a manual mode, and the discussion is not repeated here. In addition, the multimedia file generated separately may be added with respective background photos, and the discussion is not repeated here. Furthermore, the generated multimedia files may also be stored or shared, and will not be discussed again.
As shown in fig. 2, a second embodiment of the present application provides an apparatus 200 for acquiring a multimedia file, including: a determining unit 202, configured to receive a first input of a user to capture a preview screen, and determine at least two types of targets according to the first input; the processing unit 204 is configured to receive a second input, and generate a plurality of multimedia files in response to the second input, wherein each multimedia file includes a type of object.
In the apparatus 200 for acquiring a multimedia file according to the embodiment of the second aspect of the present application, first, the determining unit 202 receives a first input of a user for shooting a preview image, and determines at least two types of targets according to the first input; the processing unit 204 then receives a second input and generates a plurality of multimedia files in response to the second input. In particular, each multimedia file contains a class of objects. Specifically, the multi-type objects are respectively displayed on the multi-media files, and then the multi-media files are obtained. Like this, this application accessible one-time operation obtains a plurality of multimedia files, has promoted the convenience and the practicality that multimedia file acquireed. In addition, different multimedia files correspond to different types of targets in the application, so that each multimedia file can have own style, and the visual effect of the multimedia files is further improved.
For example, when the present application displays a shooting preview screen as shown in fig. 4, multimedia files as shown in fig. 5 and 6 may be generated in the course of one operation, respectively; when the present application displays the photographing preview screen as shown in fig. 7, multimedia files as shown in fig. 8 and 9, respectively, can be generated in the course of one operation.
In one embodiment of the present application, the multimedia file comprises a photograph or video. When the multimedia photos are photos, a plurality of photos can be obtained through one-time photographing operation, and each photo is provided with a type of target, so that each photo can have the style of the photo, and the photographing effect is further improved. When the multimedia photo is a video, a plurality of videos can be obtained through one-time shooting operation, and each video is provided with a type of target, so that each video can have own style, and the shooting effect is further improved.
In one embodiment of the present application, the determination unit 202 is responsive to a first input and determines at least two objects specified by the first input in the photographing preview screen; then, the determination unit 202 may display the mark information of the at least two objects, respectively. In particular, for any one tag information, all objects having the tag information are a kind of objects. Thus, a class of objects includes one or more objects.
Specifically, the method for acquiring the multimedia file provided by the application has a manual mode. In the manual mode, the user is free to mark the object in each shot preview screen. Specifically, during the operation, a plurality of objects are displayed on the shooting preview screen, and the user can select at least two objects in the shooting preview screen through a first input to the shooting preview screen and mark the at least two objects respectively to display mark information for the at least two objects respectively. In this way, for any piece of mark information, the object having the mark information in the shooting preview picture is the same type of object. Therefore, in the subsequent process of generating the multimedia file, one or more objects with the same marking information can be generated into one multimedia file.
Specifically, in this embodiment, the first input is a mark input. That is, the user directly marks the at least two objects through the first input. Therefore, in this embodiment, the user can mark different objects according to his/her preference, and freely decide which objects in the shooting preview screen are generated in the same multimedia file.
For example, the marking information may be 1, 2, 3, 4, 5, etc., or may be different patterns (for example, the marking information may be a polygon, and the different marking information is distinguished by the number of sides of the polygon).
In one embodiment of the present application, in response to the first input, the determining unit 202 may automatically identify categories of at least two objects in the shooting preview screen, and display mark information of the at least two objects according to the category of each object. In particular, for any one piece of tag information, all objects having the tag information are one type of object. Thus, a class of objects includes one or more objects. Specifically, the above categories may include humans, animals, landscapes, and the like.
Specifically, the method for acquiring the multimedia file has an automatic mode. In the automatic mode, the method and the device can automatically identify the categories of at least two objects in the shooting preview picture; then, corresponding mark information is automatically displayed for each object according to the category of the object. In this way, for any piece of mark information, the object having the mark information in the shooting preview picture is the same type of object. Therefore, in the subsequent process of generating the multimedia file, one or more objects with the same marking information can be generated into one multimedia file. Specifically, as for identifying the category of the object, AI identification and other related technologies can be adopted, and the discussion is not provided herein, which can be understood by those skilled in the art.
Specifically, in this embodiment, the first input is a start input. That is, the user starts the automatic mode through the first input, and then the mark information is automatically displayed for at least two objects in the automatic mode. Therefore, in this embodiment, it is possible to automatically recognize at least two objects in the photographing preview screen, mark the at least two objects in the base recognition result, and display mark information of the objects. Therefore, the automation degree in the whole process is high, and the acquisition convenience of the multimedia file can be greatly improved.
For example, the marking information may adopt numbers of 1, 2, 3, 4, 5, etc., or may adopt different figures (for example, the marking information may adopt polygons, and the different marking information is distinguished by the number of sides of the polygons).
Further, in an embodiment of the present application, in response to the second input, the processing unit 204 may obtain, for any one of the tag information, all objects having the tag information during the process of generating the plurality of multimedia files; all objects with this tag information are then generated as one photo. Correspondingly, when there are a plurality of pieces of mark information on the shooting preview screen, the processing unit 204 repeats the above steps a plurality of times, and further generates a plurality of photos.
Specifically, in the process of generating a photo, the processing unit 204 selects one piece of mark information in the preview screen as first target mark information, and acquires all objects with the first target mark information; then, the processing unit 204 generates all objects with the first target mark information in the first multimedia file. Correspondingly, after the first photo is generated, the processing unit 204 selects the next mark information as the second target mark information, and obtains all objects with the second target mark information; the processing unit 204 then generates all objects with the second object tag information in a second multimedia file. And repeating the steps until all the mark information in the shooting preview picture is selected as the target mark object and the corresponding multimedia file generation processing is finished.
Specifically, after all the photos are generated, the number of the multimedia files corresponds to the number of the categories of the mark information. That is, if there are two types of mark information in the photographed preview screen, two multimedia files are finally generated; correspondingly, if N types of mark information exist in the shooting preview picture, N multimedia files are finally generated.
For example, as shown in fig. 4, the shooting preview screen has mark information 1 and mark information 2 therein; the processing unit 204 first acquires the object having the tag information 1, and the processing unit 204 acquires images of the two persons and generates the images of the two persons in the multimedia file shown in fig. 5. Then, the processing unit 204 first obtains the object with the tag information 2, and the processing unit 204 obtains the images of the sun and the butterfly and generates the images of the sun and the butterfly in the multimedia file shown in fig. 6.
Further, in an embodiment of the present application, the processing unit 204 may generate a silhouette area of the object as a multimedia file.
For example, as shown in fig. 4, the determination unit 202 may mark the silhouette area of the person image (specifically, the silhouette area of the person) as mark information 1, and mark the silhouette area of the animal image and the scenery image (specifically, the silhouette area of the butterfly and the sun) as mark information 2. Thus, as shown in fig. 5, the image (person) in the silhouette area marked 1 can generate a multimedia file, and as shown in fig. 6, the image (butterfly and sun) in the silhouette area marked 2 can generate a multimedia file.
Further, in an embodiment of the present application, the processing unit 204 may generate the selected area where the object is located as a multimedia file. Moreover, the application allows the border of the selection area to be adjusted, so that the size and scale of the picture of the selection area presented on the final multimedia file can be adjusted.
For example, as shown in fig. 8 and 9, the determining unit 202 directly obtains an image corresponding to a silhouette area of the photo-object in the selected area and a background image in the selected area, and generates a multimedia file from the image corresponding to the silhouette area and the background image in the selected area.
Further, in one embodiment of the present application, each tag information has a corresponding priority, and different tag information has different priority orders. In the process of generating a multimedia file, the processing unit 204 may sort different targets each having different tag information according to the priority order of the different tag information, and sequentially generate objects having different tag information on different multimedia files.
For example, the tag information includes first tag information and second tag information, and the priority of the first tag information is higher than the priority of the second tag information. In the process of generating a multimedia file, the processing unit 204 first generates all objects with first tag information in a first multimedia file; then, the processing unit 204 generates all objects with the second tag information in the first multimedia file.
For example, if the priority of tag information 1 is higher than that of tag information 2 in fig. 4, the processing unit 204 generates the multimedia file shown in fig. 5 first and then generates the multimedia file shown in fig. 6; in fig. 7, the priority of tag information 1 is higher than that of tag information 2, and the processing unit 204 generates the multimedia file shown in fig. 8 first and then generates the multimedia file shown in fig. 9.
Further, in an embodiment of the present application, as shown in fig. 10, after the multimedia file is generated, the present application may add a background to the generated multimedia file. Specifically, the determining unit 202 may receive a third input and obtain at least two background photos in response to the third input, and the processing unit 204 may fuse the generated multimedia file with the selected background photos to obtain the multimedia file with a background.
Specifically, the user may directly select a corresponding number of background photos from the background, or preview the alternative background photos, then select a corresponding number of background photos from the previewed background images, and respectively fuse the corresponding number of photos with the corresponding number of background photos.
For example, as shown in fig. 10, the present application may be a multimedia file fusion background photo 1 generated corresponding to the marking information 1 in fig. 10, and a multimedia file fusion background photo 2 generated corresponding to the marking information 2 in fig. 10.
For example, as shown in fig. 11, the present application previews four background pictures (background picture 1, background picture 2, background picture 3, and background picture 4, respectively); and then, selecting one of the four background pictures to fuse with the multimedia file generated correspondingly to the tag information 1, and selecting one of the four background pictures to fuse with the multimedia file generated correspondingly to the tag information 2.
Further, in an embodiment of the present application, before the generated multimedia file and the selected background photo are fused, the processing unit 204 first sorts at least two background photos; specifically, the background photos can be sorted according to the selection sequence of the background photos, wherein the background photos selected first are in front of the background photos selected later. Then, the processing unit 204 sequences the generated multimedia files; specifically, according to the priority order of the tag information corresponding to the object in the multimedia file, the multimedia file with high priority is before and the multimedia file with low priority is after.
Further, in the process of fusing the multimedia file and the selected background picture, the processing unit 204 selects one multimedia file and one background picture each time according to the above sorting, and fuses the generated multimedia file and the selected background picture, thereby obtaining a multimedia file with a background. Correspondingly, the processing unit 204 repeats the above selection and fusion process for a plurality of times until the fusion of all the multimedia files and the background photos is completed.
The apparatus 200 for acquiring a multimedia file in the embodiment of the present application may be an apparatus, and may also be a component, an integrated circuit, or a chip in a terminal. The device can be mobile electronic equipment or non-mobile electronic equipment. By way of example, the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a palm top computer, a vehicle-mounted electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook or a Personal Digital Assistant (PDA), and the like, and the non-mobile electronic device may be a server, a Network Attached Storage (NAS), a Personal Computer (PC), a Television (TV), a teller machine or a self-service machine, and the like, and the embodiments of the present application are not particularly limited.
The apparatus 200 for acquiring a multimedia file in the embodiment of the present application may be an apparatus having an operating system. The operating system may be an Android operating system, an ios operating system, or other possible operating systems, which is not specifically limited in the embodiment of the present application.
The apparatus 200 for acquiring a multimedia file provided in this embodiment of the present application can implement each process implemented in the method embodiment of fig. 1, and is not described here again to avoid repetition.
Optionally, as shown in fig. 12, an electronic device 1200 is further provided in an embodiment of the present application, and includes a processor 1201, a memory 1202, and a program or an instruction stored in the memory 1202 and executable on the processor 1201, where the program or the instruction is executed by the processor 1201 to implement each process of the above-mentioned embodiment of the method for acquiring a multimedia file, and can achieve the same technical effect, and no further description is provided here to avoid repetition.
It should be noted that the electronic devices in the embodiments of the present application include the mobile electronic device and the non-mobile electronic device described above.
Fig. 13 is a schematic hardware structure diagram of an electronic device implementing an embodiment of the present application.
The electronic device 1300 includes, but is not limited to: a radio frequency unit 1301, a network module 1302, an audio output unit 1303, an input unit 1304, a sensor 1305, a display unit 1306, a user input unit 1307, an interface unit 1308, a memory 1309, a processor 1310, and the like.
Those skilled in the art will appreciate that the electronic device 1300 may further comprise a power supply (e.g., a battery) for supplying power to the various components, and the power supply may be logically connected to the processor 1310 via a power management system, so as to manage charging, discharging, and power consumption management functions via the power management system. The electronic device structure shown in fig. 13 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than those shown, or combine some components, or arrange different components, and thus, the description is omitted here.
The display unit 1306 is configured to receive a first input of a user to take a preview picture, and determine at least two types of targets according to the first input; the processor 1310 is configured to receive a second input, and generate a plurality of multimedia files in response to the second input, wherein each multimedia file contains a class of objects. Therefore, the method and the device can obtain a plurality of multimedia files through one-time operation, and improve the convenience and the practicability of multimedia file acquisition; in addition, different multimedia files correspond to different types of targets in the application, so that each multimedia file can have own style, and the visual effect of the multimedia files is further improved.
Optionally, the display unit 1306 is specifically configured to, in response to the first input, determine at least two objects specified by the first input in the shooting preview screen; and respectively displaying the mark information of at least two objects, wherein all the objects with the mark information are targets of the same type aiming at any mark information.
Optionally, the display unit 1306 is specifically configured to, in response to the first input, identify categories of at least two objects in the shooting preview screen, where the categories include people, animals, and scenery; and displaying the mark information of at least two shot objects respectively according to the categories, wherein all the objects with the mark information are targets of the same category aiming at any mark information.
Optionally, the processor 1310 is specifically configured to, for any tag information, obtain all objects with tag information; and generating all the objects with the tag information as one multimedia file.
Optionally, each tag information has a corresponding priority, and the processor 1310 is specifically configured to generate different multimedia files sequentially according to the priorities of the tag information, where the different types of objects have different tag information.
Optionally, the processor 1310 is further configured to receive a third input, and determine at least two background pictures according to the third input; sequencing the at least two background pictures according to the determined sequence of the at least two background pictures; sequencing the plurality of multimedia files according to the priority order of the mark information corresponding to the objects in the multimedia files; and selecting one background picture and one multimedia file each time according to the sequence, and fusing the selected background picture and the multimedia file to generate a plurality of multimedia files with backgrounds.
It should be understood that in the embodiment of the present application, the input Unit 1304 may include a Graphics Processing Unit (GPU) 1041 and a microphone 13042, and the Graphics processor 13041 processes image data of still pictures or videos obtained by an image capturing apparatus (such as a camera) in a video capturing mode or an image capturing mode. The display unit 1306 may include a display panel 13061, and the display panel 13061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. The user input unit 1307 includes a touch panel 13071 and other input devices 13072. Touch panel 13071, also referred to as a touch screen. The touch panel 13071 may include two parts, a touch detection device and a touch controller. Other input devices 13072 may include, but are not limited to, a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, and a joystick, which are not described in detail herein. Memory 1309 may be used to store software programs as well as various data, including but not limited to application programs and operating systems. The processor 1310 may integrate an application processor, which mainly handles operating systems, user interfaces, application programs, etc., and a modem processor, which mainly handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into processor 1310.
The embodiment of the present application further provides a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or the instruction is executed by a processor, the program or the instruction implements each process of the above-mentioned method for acquiring a multimedia file, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
The processor is the processor in the electronic device in the above embodiment. Readable storage media, including computer-readable storage media, such as computer Read-Only Memory (ROM), random Access Memory (RAM), magnetic or optical disks, etc.
The embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to execute a program or an instruction to implement each process of the above-mentioned method for acquiring a multimedia file, and can achieve the same technical effect, and in order to avoid repetition, the description is omitted here.
It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-on-chip, system-on-chip or system-on-chip, etc.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrases "comprising a component of' 8230; \8230;" does not exclude the presence of another like element in a process, method, article, or apparatus that comprises the element. Further, it should be noted that the scope of the methods and apparatus of the embodiments of the present application is not limited to performing the functions in the order illustrated or discussed, but may include performing the functions in a substantially simultaneous manner or in a reverse order based on the functions involved, e.g., the methods described may be performed in an order different than that described, and various steps may be added, omitted, or combined. In addition, features described with reference to certain examples may be combined in other examples.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present application may be embodied in the form of a computer software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (such as a mobile phone, a computer, a server, or a network device) to execute the method of the embodiments of the present application.
While the present embodiments have been described with reference to the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiments described above, which are meant to be illustrative and not restrictive, and that various changes may be made therein by those skilled in the art without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (10)

1. A method for acquiring a multimedia file is characterized by comprising the following steps:
receiving first input of a user to a shooting preview picture, and determining at least two types of targets according to the first input;
receiving a second input, and generating a plurality of multimedia files in response to the second input, wherein each multimedia file comprises one type of the target, and different multimedia files correspond to different types of targets;
determining at least two types of targets according to the first input, including:
determining, in response to the first input, at least two objects specified by the first input in the photographing preview screen; and
and respectively displaying mark information of the at least two objects, wherein all the objects with the mark information are the same type of target aiming at any mark information.
2. The method for acquiring a multimedia file according to claim 1, wherein the determining at least two types of targets according to the first input comprises:
identifying categories of at least two objects in the shooting preview picture in response to the first input, wherein the categories comprise people, animals and scenery; and
and respectively displaying mark information of the at least two shooting objects according to the categories, wherein all the objects with the mark information are targets of the same category aiming at any mark information.
3. The method for acquiring multimedia files according to claim 1 or 2, wherein the generating a plurality of multimedia files in response to the second input comprises:
for any one of the label information, acquiring all the objects with the label information; and
and generating all the objects with the mark information into one multimedia file.
4. The method for acquiring a multimedia file according to claim 1 or 2,
each of the tag information has a corresponding priority; and
and generating different multimedia files for the different types of objects with different label information in sequence according to the priority of the label information.
5. The method for acquiring a multimedia file according to claim 4,
receiving a third input, and determining at least two background pictures according to the third input;
sequencing the at least two background pictures according to the determined sequence of the at least two background pictures;
sequencing the plurality of multimedia files according to the priority order of the mark information corresponding to the objects in the multimedia files;
and selecting one background picture and one multimedia file each time according to the sequence, and fusing the selected background picture and the multimedia file to generate a plurality of multimedia files with backgrounds.
6. An apparatus for acquiring a multimedia file, comprising:
the device comprises a determining unit, a judging unit and a judging unit, wherein the determining unit is used for receiving a first input of a user to a shooting preview picture and determining at least two types of targets according to the first input;
a processing unit, configured to receive a second input, and generate a plurality of multimedia files in response to the second input, where each of the multimedia files includes one type of the object, and different multimedia files correspond to different types of objects;
the determination unit is specifically configured to determine, in response to the first input, at least two objects specified by the first input in the shooting preview screen; and
and respectively displaying mark information of the at least two objects, wherein aiming at any mark information, all the objects with the mark information are the same type of target.
7. The apparatus for acquiring multimedia file according to claim 6,
the determining unit is specifically configured to, in response to the first input, identify categories of at least two objects in the shooting preview image, where the categories include people, animals, and scenery; and
and respectively displaying mark information of the at least two objects according to the categories, wherein all the objects with the mark information are the same type of target aiming at any mark information.
8. The apparatus for acquiring a multimedia file according to claim 6 or 7,
the processing unit is specifically configured to, for any one of the tag information, acquire all the objects having the tag information; and
and generating all the objects with the mark information into one multimedia file.
9. The apparatus for acquiring a multimedia file according to claim 6 or 7,
each of the tag information has a corresponding priority; and
the processing unit is specifically configured to sequentially generate different multimedia files for different types of objects having different tag information according to the priority of the tag information.
10. The apparatus for acquiring multimedia file according to claim 9,
the determining unit is further used for receiving a third input and determining at least two background pictures according to the third input;
the processing unit is further configured to sort the at least two background pictures according to the determined order of the at least two background pictures;
sequencing the plurality of multimedia files according to the priority order of the mark information corresponding to the object in the multimedia files; and
and selecting one background picture and one multimedia file each time according to the sequence, and fusing the selected background picture and the multimedia file to generate a plurality of multimedia files with backgrounds.
CN202111133380.2A 2021-09-27 2021-09-27 Multimedia file acquisition method and device Active CN113873080B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111133380.2A CN113873080B (en) 2021-09-27 2021-09-27 Multimedia file acquisition method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111133380.2A CN113873080B (en) 2021-09-27 2021-09-27 Multimedia file acquisition method and device

Publications (2)

Publication Number Publication Date
CN113873080A CN113873080A (en) 2021-12-31
CN113873080B true CN113873080B (en) 2022-12-13

Family

ID=78990902

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111133380.2A Active CN113873080B (en) 2021-09-27 2021-09-27 Multimedia file acquisition method and device

Country Status (1)

Country Link
CN (1) CN113873080B (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110198428A (en) * 2019-05-29 2019-09-03 维沃移动通信有限公司 A kind of multimedia file producting method and first terminal
CN111669503A (en) * 2020-06-29 2020-09-15 维沃移动通信有限公司 Photographing method and device, electronic equipment and medium
CN112135046A (en) * 2020-09-23 2020-12-25 维沃移动通信有限公司 Video shooting method, video shooting device and electronic equipment
CN112887609A (en) * 2021-01-27 2021-06-01 维沃移动通信有限公司 Shooting method, shooting device, electronic equipment and storage medium

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105869198A (en) * 2015-12-14 2016-08-17 乐视移动智能信息技术(北京)有限公司 Multimedia photograph generating method, apparatus and device, and mobile phone
KR20200009594A (en) * 2018-07-19 2020-01-30 (주)코믹스브이 Method for providing virtual-reality cartoon service using panoramic image
CN112241397B (en) * 2019-07-18 2023-08-22 腾讯科技(深圳)有限公司 Sharing method and device of multimedia files, electronic equipment and readable storage medium
CN111836112B (en) * 2020-06-28 2022-01-25 腾讯科技(深圳)有限公司 Multimedia file output method, device, medium and electronic equipment
CN112528049B (en) * 2020-12-17 2023-08-08 北京达佳互联信息技术有限公司 Video synthesis method, device, electronic equipment and computer readable storage medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110198428A (en) * 2019-05-29 2019-09-03 维沃移动通信有限公司 A kind of multimedia file producting method and first terminal
CN111669503A (en) * 2020-06-29 2020-09-15 维沃移动通信有限公司 Photographing method and device, electronic equipment and medium
CN112135046A (en) * 2020-09-23 2020-12-25 维沃移动通信有限公司 Video shooting method, video shooting device and electronic equipment
CN112887609A (en) * 2021-01-27 2021-06-01 维沃移动通信有限公司 Shooting method, shooting device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN113873080A (en) 2021-12-31

Similar Documents

Publication Publication Date Title
CN111601039B (en) Video shooting method and device and electronic equipment
CN106203286B (en) Augmented reality content acquisition method and device and mobile terminal
CN112714257B (en) Display control method, display control device, electronic device, and medium
CN113766296A (en) Live broadcast picture display method and device
CN111770386A (en) Video processing method, video processing device and electronic equipment
CN112306347B (en) Image editing method, image editing device and electronic equipment
CN109034242B (en) Grading method, device and system of image processing algorithm
CN113873166A (en) Video shooting method and device, electronic equipment and readable storage medium
CN113194256B (en) Shooting method, shooting device, electronic equipment and storage medium
CN112449110B (en) Image processing method and device and electronic equipment
CN113596574A (en) Video processing method, video processing apparatus, electronic device, and readable storage medium
CN111800574B (en) Imaging method and device and electronic equipment
CN117194697A (en) Label generation method and device and electronic equipment
CN113010738A (en) Video processing method and device, electronic equipment and readable storage medium
CN112330728A (en) Image processing method, image processing device, electronic equipment and readable storage medium
CN112287141A (en) Photo album processing method and device, electronic equipment and storage medium
CN113873080B (en) Multimedia file acquisition method and device
WO2022247766A1 (en) Image processing method and apparatus, and electronic device
CN112367487B (en) Video recording method and electronic equipment
CN113271378B (en) Image processing method and device and electronic equipment
CN114466140A (en) Image shooting method and device
CN112732961A (en) Image classification method and device
CN114025100A (en) Shooting method, shooting device, electronic equipment and readable storage medium
CN112492206B (en) Image processing method and device and electronic equipment
CN113873081B (en) Method and device for sending associated image and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant