WO2023116669A1 - 视频生成系统、方法及相关装置 - Google Patents

视频生成系统、方法及相关装置 Download PDF

Info

Publication number
WO2023116669A1
WO2023116669A1 PCT/CN2022/140279 CN2022140279W WO2023116669A1 WO 2023116669 A1 WO2023116669 A1 WO 2023116669A1 CN 2022140279 W CN2022140279 W CN 2022140279W WO 2023116669 A1 WO2023116669 A1 WO 2023116669A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
electronic device
image
geographic location
preset condition
Prior art date
Application number
PCT/CN2022/140279
Other languages
English (en)
French (fr)
Inventor
曾亦航
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Publication of WO2023116669A1 publication Critical patent/WO2023116669A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/268Signal distribution or switching

Definitions

  • the present application relates to the field of electronic technology, and in particular to a video generation system, method and related devices.
  • the mobile phone can edit and stitch pictures or videos under a certain period of time or under a theme, and automatically generate a new video (also called vlog).
  • the new video can be a period of time Highlights within a theme, such as highlights on weekends, highlights on holidays, etc.; it can also be highlights under a theme, such as highlights under the theme of sports, highlights under the theme of travel, highlights under the theme of graduation, etc. wait.
  • the generation process of the wonderful moments in the album needs to be completed when the electronic device is charging and the screen is turned off.
  • the analysis time is long. If there are a large number of photos/videos, the user will wait for a long time. Moreover, when traveling, there are often fleeting beautiful scenery, and the user wants to take pictures or record videos but it is too late. Therefore, how to improve the efficiency of video generation and reduce the waiting time of users is a problem being studied by those skilled in the art.
  • Embodiments of the present application provide a video generation system, method, and related devices, which can simultaneously analyze and edit real-time captured video, improve video generation efficiency, and reduce user waiting time.
  • the present application provides a method for generating a video, the method comprising: the first electronic device determines that its own geographic location information satisfies a first preset condition; the first electronic device obtains a video and/or image captured in real time; the second An electronic device selects a part of the video and/or image captured in real time; when the geographic location information of the first electronic device satisfies a second preset condition, the first electronic device generates a first video, and the first video is based on the Part of the screen spliced together.
  • the first electronic device determines the start and end time of the analysis of the video and/or image based on its own geographic location.
  • the first electronic device starts to analyze the The video and/or image data captured by other electronic devices (such as the second electronic device) in real time is analyzed and edited, so as to select a part of the video and/or image captured in real time. That is, while the other electronic devices are shooting the video, the first electronic device is synchronously selecting frames for the video.
  • the first electronic device When the geographic location of the first electronic device satisfies the second preset condition, the first electronic device generates the first video, and does not analyze and edit subsequent videos and/or images.
  • the first video is spliced based on the above-mentioned selected partial images. This way of synchronously analyzing video and/or images captured in real time can improve the efficiency of video generation and reduce the waiting time of users.
  • the first preset condition includes: the current geographic location of the first electronic device leaves the user's resident area; or the current geographic location of the first electronic device leaves the user's residence area.
  • the resident area and the duration of leaving exceeds the first threshold; or the shortest distance between the first electronic device's current geographic location and the preset location is greater than or equal to the second threshold, and the preset location includes the user's resident area; or An electronic device's current geographic location has entered a specific area.
  • the first preset condition determines the timing when the first electronic device starts to analyze the video and/or image. video and/or images for simultaneous analysis.
  • the first preset condition is related to the geographic location information of the first electronic device, which may be applicable to an application scenario where a user travels and generates a travel video.
  • the video and/or image is captured by the second electronic device in real time.
  • the video and/or image may be captured in real time by the first electronic device; optionally, the video and/or image may be captured in real time by multiple electronic devices.
  • the first electronic device selects a part of the video and/or image, including: the first electronic device segments the acquired video and/or image; the first The electronic device scores the segmented video and/or image based on the aesthetic scoring algorithm, and selects the highest-scoring frame in each segment of the segmented video and/or image, and some of the segments include each segment of the segmented video and/or image The highest rated picture in .
  • the scoring weight of the aesthetic scoring algorithm is associated with the scene type of the segmented video and/or image.
  • the scoring rules of the aesthetic scoring can be changed according to the change of the scene type of the video picture.
  • Each time the first electronic device divides a video segment it determines the scene type of the video segment based on the current geographic location of the first electronic device or image recognition technology, thereby determining the scoring rule corresponding to the scene type.
  • customized aesthetic scoring rules can be selected to select the most suitable picture fragments for the public's aesthetics and improve the user experience.
  • the second preset condition includes: the current geographic location of the first electronic device has entered a preset location; or the current geographic location of the first electronic device is far from the preset location The remaining time of is less than the fifth threshold.
  • the second preset condition determines the timing when the first electronic device finishes analyzing the video and/or image, and the first electronic device can generate the first video when the second preset condition is met, without waiting for a long time analysis in the background.
  • the second preset condition is related to the geographic location information of the first electronic device, which may be applicable to an application scenario where a user travels and generates a travel video.
  • the first electronic device generates the first video, and then further includes: the first electronic device outputs first prompt information, and the first prompt information is used to provide editing for the first video mode, the editing mode includes playing the first video, saving the first video, deleting the first video, adjusting the order of the display screens in the first video, deleting part of the display screens in the first video, and Insert transition effects in between, add text, stickers, watermarks on the display screen of the first video, and add one or more of the background music of the first video.
  • the method further includes: the first electronic device generates a video name of the first video based on geographic location information of the first electronic device or video content of the first video.
  • the problem of the single name of the album video in the current gallery is solved, and the ecology of the gallery to record life for users is enriched.
  • the method further includes: the first electronic device sends the first video to a third electronic device, where the third electronic device includes the second electronic device.
  • the present application provides a video generation system, including a first electronic device and a second electronic device, and the first electronic device establishes a connection with the second electronic device;
  • a second electronic device for taking video and/or images
  • the first electronic device is configured to determine that its own geographic location information satisfies a first preset condition
  • the first electronic device is further configured to acquire video and/or images captured by the second electronic device in real time;
  • the first electronic device is further configured to select a part of the video and/or image
  • the first electronic device is further configured to generate a first video when the geographic location information of the first electronic device satisfies a second preset condition, and the first video is spliced based on the partial images.
  • the first electronic device establishes a connection with the second electronic device, and the second electronic device can shoot videos/images at any time.
  • the first electronic device determines the start and end time of the analysis of the video and/or image based on its own geographic location.
  • the geographic location of the first electronic device satisfies the first preset condition, the first electronic device starts real-time analysis of the second electronic device.
  • the captured video and/or image data is analyzed and edited, so as to select a part of the real-time captured video and/or image. That is, while the second electronic device is shooting a video, the first electronic device simultaneously selects frames of the video.
  • the first electronic device When the geographic location of the first electronic device satisfies the second preset condition, the first electronic device generates the first video, and does not analyze and edit subsequent videos and/or images.
  • the first video is spliced based on the above-mentioned selected partial images. This way of synchronously analyzing video and/or images captured in real time can improve the efficiency of video generation and reduce the waiting time of users.
  • the first preset condition includes: the current geographic location of the first electronic device has left the user's resident area; or the current geographic location of the first electronic device has left the user The resident area and the duration of leaving exceeds the first threshold; or the shortest distance between the first electronic device's current geographic location and the preset location is greater than or equal to the second threshold, and the preset location includes the user's resident area; or An electronic device's current geographic location has entered a specific area.
  • the first preset condition determines the timing when the first electronic device starts to analyze the video and/or image. video and/or images for simultaneous analysis.
  • the first preset condition is related to the geographic location information of the first electronic device, which may be applicable to an application scenario where a user travels and generates a travel video.
  • the first electronic device is further configured to select a part of the video and/or images captured in real time, including: the first electronic device is also configured to process the acquired Segmenting the video and/or image; the first electronic device is also used to score the segmented video and/or image based on an aesthetic scoring algorithm, and select the highest-scoring frame in each video and/or image after segmentation , the part of frames includes the highest-scoring frame in each segment of video and/or image after segmenting.
  • the scoring weight of the aesthetic scoring algorithm is associated with the scene type of the segmented video and/or image.
  • the scoring rules of the aesthetic scoring can be changed according to the change of the scene type of the video picture.
  • Each time the first electronic device divides a video segment it determines the scene type of the video segment based on the current geographic location of the first electronic device or image recognition technology, thereby determining the scoring rule corresponding to the scene type.
  • customized aesthetic scoring rules can be selected to select the most suitable picture fragments for the public's aesthetics and improve the user experience.
  • the second preset condition includes: the current geographic location of the first electronic device has entered a preset location; or the current geographic location of the first electronic device is far from the preset location The remaining time of is less than the fifth threshold.
  • the second preset condition determines the timing when the first electronic device finishes analyzing the video and/or image, and the first electronic device can generate the first video when the second preset condition is met, without waiting for a long time analysis in the background.
  • the second preset condition is related to the geographic location information of the first electronic device, which may be applicable to an application scenario where a user travels and generates a travel video.
  • the first electronic device is further configured to output first prompt information after generating the first video, where the first prompt information is used to provide an editing method for the first video
  • Editing methods include playing the first video, saving the first video, deleting the first video, adjusting the order of the display screens in the first video, deleting part of the display screens in the first video, and inserting transitions between the display screens of the first video Dynamic effects, adding text, stickers, watermarks, and adding one or more of the background music of the first video to the display screen of the first video.
  • the first electronic device is further configured to generate a video name of the first video based on geographic location information of the first electronic device or video content of the first video.
  • the problem of the single name of the album video in the current gallery is solved, and the ecology of the gallery to record life for users is enriched.
  • the first electronic device is further configured to send the first video to the second electronic device.
  • the second electronic device is further configured to receive the first video; the second electronic device is further configured to output second prompt information, and the second prompt information is used to provide information about the first video.
  • a video editing method the editing method includes playing the first video, deleting the first video, adjusting the order of the display screens in the first video, deleting part of the display screens in the first video, inserting transitions between the display screens of the first video A dynamic effect, adding text, stickers, watermarks, and adding one or more of the background music of the first video to the display screen of the first video.
  • the present application provides an electronic device, including: one or more processors, one or more memories; the one or more stores are coupled with one or more processors; the one or more memories are used For storing computer program codes, the computer program codes include computer instructions; when the computer instructions run on the processor, the electronic device is made to execute the video generation method in any possible implementation of any aspect above.
  • the present application provides a server, including: one or more processors, one or more memories; the one or more stores are coupled with one or more processors; the one or more memories are used for Computer program code is stored, and the computer program code includes computer instructions; when the computer instructions are run on the processor, the electronic device is made to execute the video generation method in any possible implementation manner of any aspect above.
  • the embodiment of the present application provides a computer storage medium, including computer instructions, and when the computer instructions are run on the electronic device, the communication device executes the video generation method in any possible implementation of any of the above aspects .
  • an embodiment of the present application provides a computer program product, which, when the computer program product is run on a computer, causes the computer to execute the video generation method in any possible implementation manner of any one of the above aspects.
  • Fig. 1 is a flow chart of steps for automatically generating a picture stitching video provided by an embodiment of the present application
  • FIG. 2 is a schematic diagram of the architecture of a video generation system provided by an embodiment of the present application.
  • FIG. 3a is a schematic structural diagram of an electronic device provided in an embodiment of the present application.
  • FIG. 3b is a software structure diagram of an electronic device provided by an embodiment of the present application.
  • FIG. 4 is a flow chart of the steps of a method for generating a video provided in an embodiment of the present application
  • 5a to 5c are application scene diagrams of a video generation method provided by the embodiment of the present application.
  • FIG. 6 is an application scenario diagram of a video generation method provided by an embodiment of the present application.
  • FIG. 7 is an application interface diagram of a video generation method provided in the embodiment of the present application on the vehicle.
  • FIG. 8 is an application interface diagram on a mobile phone of a video generation method provided by an embodiment of the present application.
  • first and second are used for descriptive purposes only, and cannot be understood as implying or implying relative importance or implicitly specifying the quantity of indicated technical features. Therefore, the features defined as “first” and “second” may explicitly or implicitly include one or more of these features.
  • the “multiple” The meaning is two or more. The orientation or positional relationship indicated by the terms “middle”, “left”, “right”, “upper”, “lower” etc.
  • the electronic equipment involved in the embodiment of the present application can be a mobile phone, a tablet computer, a desktop, a laptop, a notebook computer, an ultra-mobile personal computer (Ultra-mobile Personal Computer, UMPC), a handheld computer, a netbook, a personal digital assistant (Personal Digital Assistant, PDA), virtual reality equipment, PDA (Personal Digital Assistant, personal digital assistant, also known as handheld computer), portable Internet devices, data storage devices, wearable devices (for example, wireless headsets, smart watches, smart bracelets , smart glasses, head-mounted display (HMD), electronic clothing, electronic bracelets, electronic necklaces, electronic accessories, electronic tattoos and smart mirrors), vehicles, car machines, vehicle terminals (such as smart screens, vehicle cameras , display screen), cameras and various electronic devices with cameras, etc.
  • wearable devices for example, wireless headsets, smart watches, smart bracelets , smart glasses, head-mounted display (HMD), electronic clothing, electronic bracelets, electronic necklaces, electronic accessories, electronic tattoos and smart mirrors
  • vehicles car machines, vehicle terminal
  • Image Aesthetic Scoring Score images based on an aesthetic scoring algorithm.
  • image aesthetic scoring algorithm some quantified image aesthetic rules can be used, such as the distribution of important features in the frame of the captured image. These aesthetic scoring rules can come from experience summarization. Commonly used aesthetic scoring rules include, for example, the rule of thirds, the principle of the main diagonal, visual balance, and determining the proportion of the subject in the picture. There can be different image aesthetic rules based on different requirements.
  • An aesthetic score Ei can be calculated according to each rule.
  • Video aesthetic score similar to image aesthetic score, image frames are extracted from a video, and the aesthetic score of a video can be the weighted average of the aesthetic scores of multiple image frames in the video.
  • the mobile phone can edit and stitch pictures or videos under a certain period of time or under a theme, and automatically generate a new video (also called vlog).
  • the new video can be a period of time Highlights within a theme, such as highlights on weekends, highlights on holidays, etc.; it can also be highlights under a theme, such as highlights under the theme of sports, highlights under the theme of travel, highlights under the theme of graduation, etc. wait.
  • the process of automatically generating a picture stitching video on the mobile phone is generated after time and place clustering, scene classification, and aesthetic scoring of the pictures in the user's album when the screen is turned off during charging.
  • the user needs to wait for a period of time after charging and turning off the screen.
  • the photos in the album are clustered by time and place, intelligently classified, and artistically generated cover is completed, a picture video composed of several photos can be obtained.
  • FIG. 1 shows a flow of steps for automatically generating a picture-stitched video.
  • the electronic device detects that it is currently in the state of charging and turning off the screen, the electronic device will classify the photos stored in the album.
  • the classification method can be based on time and/or location.
  • the saved photos are classified into one category, or the photos saved within a weekend are classified into one category.
  • the location-based classification can be, for example, the photos saved by electronic devices in the same location.
  • the time- and location-based classification can be, for example, Group photos of electronic devices that were saved over a weekend and in the same location.
  • the electronic device will analyze multiple photos classified into one category, deduplicate the multiple photos and score them based on an algorithm, filter out one of the two photos with a high repetition rate, and filter out the photos with a low score , to get the final multiple photos, sort the multiple photos, and arrange the photos of the same scene in the multiple photos together to generate the final picture stitching video.
  • the video can be named based on time or scene, for example named “ Hours on weekends", “Things on weekends” and so on.
  • the embodiment of the present application provides a video generation method.
  • the electronic device can synchronously analyze the video being shot in real time and edit the video. In this way, the problem of the long waiting time of the user in the above scene can be solved.
  • other users on the car can take pictures of the surrounding environment as records, and when the recorded video is over, the user can instantly obtain the edited new video.
  • FIG. 2 shows a system architecture of a video generation system according to an embodiment of the present application.
  • the video generation system includes an electronic device 100 and one or more electronic devices 101 .
  • the electronic device 100 and the electronic device 101 can be connected by wired or wireless communication, and the wireless communication can include, for example, Bluetooth (blue tooth, BT), near field communication (near field communication, NFC), wireless fidelity (wireless fidelity, WiFi) , or WiFi Direct.
  • one or more electronic devices 101 include cameras, cameras, tablets, iPads, mobile phones and other electronic devices with cameras.
  • the electronic device 100 is installed and runs an application program supporting video generation technology, and the application program may be a photography application program or a video processing application program.
  • the electronic device 101 shoots a video
  • the electronic device 100 can acquire video data in real time during the process of shooting a video by the electronic device 101 , and the electronic device 100 analyzes and edits the video data acquired in real time. That is, while the electronic device 101 is shooting a video, the electronic device 100 simultaneously analyzes and edits the video. When the electronic device 101 finishes shooting the video, the electronic device 100 can obtain a new video that has been edited. In this way, the efficiency of video generation can be improved, and the waiting time of the user can be reduced.
  • the electronic device 100 is a vehicle
  • the electronic device 101 may be a device with a camera such as a vehicle camera or a mobile phone.
  • the electronic device 101 can capture the roadside scenery during driving in real time, then the electronic device 100 can analyze and edit the video or picture captured in real time synchronously. 100 to generate a travel video about the driving process. And the electronic device 100 can share the generated travel video to other electronic devices (including the electronic device 101 ).
  • the electronic device 100 and the electronic device 101 may perform data communication through a server, and the server is used to provide background services for applications supporting video generation technology.
  • the server may be an application server, a cloud server, a background server, and the like.
  • the server can also be a cloud computing platform, a virtualization center, etc., including equipment for accessing a wide area network, such as a base station, a broadband remote access server (broadband access serve, BRAS), an optical network unit, and a broadband network gateway (broadband network gateway) , BNG), aggregation switch, etc.
  • the electronic device 100 and the electronic device 101 are devices logged into the same account, or the accounts logged in by the electronic device 100 and the electronic device 101 are associated accounts, wherein the associated account may be an account authorized by the same account.
  • the account here refers to an application account, a device account, an identity account, and the like.
  • the electronic device 100 involved in the embodiment of the present application is first introduced below.
  • FIG. 3a shows a schematic structural diagram of an exemplary electronic device 100 provided by an embodiment of the present application.
  • the electronic device 100 may include a processor 110, an external memory interface 120, an internal memory 121, a universal serial bus (universal serial bus, USB) interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, and an antenna 2 , mobile communication module 150, wireless communication module 160, audio module 170, speaker 170A, receiver 170B, microphone 170C, earphone jack 170D, sensor module 180, button 190, motor 191, indicator 192, camera 193, display screen 194, and A subscriber identification module (subscriber identification module, SIM) card interface 195 and the like.
  • SIM subscriber identification module
  • the sensor module 180 may include a pressure sensor 180A, a gyroscope sensor 180B, an air pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a distance sensor 180F, a proximity light sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, an ambient light sensor 180L, bone conduction sensor 180M, etc.
  • the structure illustrated in the embodiment of the present application does not constitute a specific limitation on the electronic device 100 .
  • the electronic device 100 may include more or fewer components than shown in the figure, or combine certain components, or separate certain components, or arrange different components.
  • the illustrated components can be realized in hardware, software or a combination of software and hardware.
  • the processor 110 may include one or more processing units, for example: the processor 110 may include an application processor (application processor, AP), a modem processor, a graphics processing unit (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), controller, memory, video codec, digital signal processor (digital signal processor, DSP), baseband processor, and/or neural network processor (neural-network processing unit, NPU) wait. Wherein, different processing units may be independent devices, or may be integrated in one or more processors.
  • application processor application processor, AP
  • modem processor graphics processing unit
  • GPU graphics processing unit
  • image signal processor image signal processor
  • ISP image signal processor
  • controller memory
  • video codec digital signal processor
  • DSP digital signal processor
  • baseband processor baseband processor
  • neural network processor neural-network processing unit, NPU
  • the controller may be the nerve center and command center of the electronic device 100 .
  • the controller can generate an operation control signal according to the instruction opcode and timing signal, and complete the control of fetching and executing the instruction.
  • a memory may also be provided in the processor 110 for storing instructions and data.
  • the memory in processor 110 is a cache memory.
  • the memory may hold instructions or data that the processor 110 has just used or recycled. If the processor 110 needs to use the instruction or data again, it can be called directly from the memory. Repeated access is avoided, and the waiting time of the processor 110 is reduced, thereby improving the efficiency of the system.
  • processor 110 may include one or more interfaces.
  • the interface may include an integrated circuit (inter-integrated circuit, I2C) interface, an integrated circuit built-in audio (inter-integrated circuit sound, I2S) interface, a pulse code modulation (pulse code modulation, PCM) interface, a universal asynchronous transmitter (universal asynchronous receiver/transmitter, UART) interface, mobile industry processor interface (mobile industry processor interface, MIPI), general-purpose input and output (general-purpose input/output, GPIO) interface, subscriber identity module (subscriber identity module, SIM) interface, and /or universal serial bus (universal serial bus, USB) interface, etc.
  • I2C integrated circuit
  • I2S integrated circuit built-in audio
  • PCM pulse code modulation
  • PCM pulse code modulation
  • UART universal asynchronous transmitter
  • MIPI mobile industry processor interface
  • GPIO general-purpose input and output
  • subscriber identity module subscriber identity module
  • SIM subscriber identity module
  • USB universal serial bus
  • the I2C interface is a bidirectional synchronous serial bus, including a serial data line (serial data line, SDA) and a serial clock line (derail clock line, SCL).
  • processor 110 may include multiple sets of I2C buses.
  • the processor 110 can be respectively coupled to the touch sensor 180K, the charger, the flashlight, the camera 193 and the like through different I2C bus interfaces.
  • the processor 110 may be coupled to the touch sensor 180K through the I2C interface, so that the processor 110 and the touch sensor 180K communicate through the I2C bus interface to realize the touch function of the electronic device 100 .
  • the I2S interface can be used for audio communication.
  • processor 110 may include multiple sets of I2S buses.
  • the processor 110 may be coupled to the audio module 170 through an I2S bus to implement communication between the processor 110 and the audio module 170 .
  • the audio module 170 can transmit audio signals to the wireless communication module 160 through the I2S interface, so as to realize the function of answering calls through the Bluetooth headset.
  • the PCM interface can also be used for audio communication, sampling, quantizing and encoding the analog signal.
  • the audio module 170 and the wireless communication module 160 may be coupled through a PCM bus interface.
  • the audio module 170 can also transmit audio signals to the wireless communication module 160 through the PCM interface, so as to realize the function of answering calls through the Bluetooth headset. Both the I2S interface and the PCM interface can be used for audio communication.
  • the UART interface is a universal serial data bus used for asynchronous communication.
  • the bus can be a bidirectional communication bus. It converts the data to be transmitted between serial communication and parallel communication.
  • a UART interface is generally used to connect the processor 110 and the wireless communication module 160 .
  • the processor 110 communicates with the Bluetooth module in the wireless communication module 160 through the UART interface to realize the Bluetooth function.
  • the audio module 170 can transmit audio signals to the wireless communication module 160 through the UART interface, so as to realize the function of playing music through the Bluetooth headset.
  • the MIPI interface can be used to connect the processor 110 with peripheral devices such as the display screen 194 and the camera 193 .
  • MIPI interface includes camera serial interface (camera serial interface, CSI), display serial interface (display serial interface, DSI), etc.
  • the processor 110 communicates with the camera 193 through the CSI interface to realize the shooting function of the electronic device 100 .
  • the processor 110 communicates with the display screen 194 through the DSI interface to realize the display function of the electronic device 100 .
  • the GPIO interface can be configured by software.
  • the GPIO interface can be configured as a control signal or as a data signal.
  • the GPIO interface can be used to connect the processor 110 with the camera 193 , the display screen 194 , the wireless communication module 160 , the audio module 170 , the sensor module 180 and so on.
  • the GPIO interface can also be configured as an I2C interface, I2S interface, UART interface, MIPI interface, etc.
  • the USB interface 130 is an interface conforming to the USB standard specification, specifically, it can be a Mini USB interface, a Micro USB interface, a USB Type C interface, and the like.
  • the USB interface 130 can be used to connect a charger to charge the electronic device 100 , and can also be used to transmit data between the electronic device 100 and peripheral devices. It can also be used to connect headphones and play audio through them. This interface can also be used to connect other electronic devices, such as AR devices.
  • the interface connection relationship between the modules shown in the embodiment of the present application is only a schematic illustration, and does not constitute a structural limitation of the electronic device 100 .
  • the electronic device 100 may also adopt different interface connection manners in the foregoing embodiments, or a combination of multiple interface connection manners.
  • the charging management module 140 is configured to receive a charging input from a charger.
  • the charger may be a wireless charger or a wired charger.
  • the charging management module 140 can receive charging input from the wired charger through the USB interface 130 .
  • the charging management module 140 may receive a wireless charging input through a wireless charging coil of the electronic device 100 . While the charging management module 140 is charging the battery 142 , it can also supply power to the electronic device through the power management module 141 .
  • the power management module 141 is used for connecting the battery 142 , the charging management module 140 and the processor 110 .
  • the power management module 141 receives the input from the battery 142 and/or the charging management module 140 to provide power for the processor 110 , the internal memory 121 , the external memory, the display screen 194 , the camera 193 , and the wireless communication module 160 .
  • the power management module 141 can also be used to monitor parameters such as battery capacity, battery cycle times, and battery health status (leakage, impedance).
  • the power management module 141 may also be disposed in the processor 110 .
  • the power management module 141 and the charging management module 140 may also be set in the same device.
  • the wireless communication function of the electronic device 100 can be realized by the antenna 1 , the antenna 2 , the mobile communication module 150 , the wireless communication module 160 , a modem processor, a baseband processor, and the like.
  • Antenna 1 and Antenna 2 are used to transmit and receive electromagnetic wave signals.
  • Each antenna in electronic device 100 may be used to cover single or multiple communication frequency bands. Different antennas can also be multiplexed to improve the utilization of the antennas.
  • Antenna 1 can be multiplexed as a diversity antenna of a wireless local area network.
  • the antenna may be used in conjunction with a tuning switch.
  • the mobile communication module 150 can provide wireless communication solutions including 2G/3G/4G/5G applied on the electronic device 100 .
  • the mobile communication module 150 may include at least one filter, switch, power amplifier, low noise amplifier (low noise amplifier, LNA) and the like.
  • the mobile communication module 150 can receive electromagnetic waves through the antenna 1, filter and amplify the received electromagnetic waves, and send them to the modem processor for demodulation.
  • the mobile communication module 150 can also amplify the signals modulated by the modem processor, and convert them into electromagnetic waves and radiate them through the antenna 1 .
  • at least part of the functional modules of the mobile communication module 150 may be set in the processor 110 .
  • at least part of the functional modules of the mobile communication module 150 and at least part of the modules of the processor 110 may be set in the same device.
  • a modem processor may include a modulator and a demodulator.
  • the modulator is used for modulating the low-frequency baseband signal to be transmitted into a medium-high frequency signal.
  • the demodulator is used to demodulate the received electromagnetic wave signal into a low frequency baseband signal. Then the demodulator sends the demodulated low-frequency baseband signal to the baseband processor for processing. After the low-frequency baseband signal is processed by the baseband processor, it is passed to the application processor.
  • the application processor outputs sound signals through audio equipment (not limited to speaker 170A, receiver 170B, etc.), or displays images or videos through display screen 194 .
  • the modem processor may be a stand-alone device. In some other embodiments, the modem processor may be independent from the processor 110, and be set in the same device as the mobile communication module 150 or other functional modules.
  • the wireless communication module 160 can provide applications on the electronic device 100 including UWB, wireless local area networks (wireless local area networks, WLAN) (such as wireless fidelity (wireless fidelity, WiFi) network), bluetooth (bluetooth, BT), global navigation satellite System (global navigation satellite system, GNSS), frequency modulation (frequency modulation, FM), near field communication technology (near field communication, NFC), infrared technology (infrared, IR) and other wireless communication solutions.
  • the wireless communication module 160 may be one or more devices integrating at least one communication processing module.
  • the wireless communication module 160 receives electromagnetic waves via the antenna 2 , frequency-modulates and filters the electromagnetic wave signals, and sends the processed signals to the processor 110 .
  • the wireless communication module 160 can also receive the signal to be sent from the processor 110 , frequency-modulate it, amplify it, and convert it into electromagnetic waves through the antenna 2 for radiation.
  • the antenna 1 of the electronic device 100 is coupled to the mobile communication module 150, and the antenna 2 is coupled to the wireless communication module 160, so that the electronic device 100 can communicate with the network and other devices through wireless communication technology.
  • the wireless communication technology may include global system for mobile communications (GSM), general packet radio service (general packet radio service, GPRS), code division multiple access (code division multiple access, CDMA), broadband Code division multiple access (wideband code division multiple access, WCDMA), time division code division multiple access (time-division code division multiple access, TD-SCDMA), long term evolution (long term evolution, LTE), BT, GNSS, WLAN, NFC , FM, and/or IR techniques, etc.
  • GSM global system for mobile communications
  • GPRS general packet radio service
  • code division multiple access code division multiple access
  • CDMA broadband Code division multiple access
  • WCDMA wideband code division multiple access
  • time division code division multiple access time-division code division multiple access
  • TD-SCDMA time-division code division multiple access
  • the GNSS may include a global positioning system (global positioning system, GPS), a global navigation satellite system (global navigation satellite system, GLONASS), a Beidou navigation satellite system (beidou navigation satellite system, BDS), a quasi-zenith satellite system (quasi -zenith satellite system (QZSS) and/or satellite based augmentation systems (SBAS).
  • GPS global positioning system
  • GLONASS global navigation satellite system
  • Beidou navigation satellite system beidou navigation satellite system
  • BDS Beidou navigation satellite system
  • QZSS quasi-zenith satellite system
  • SBAS satellite based augmentation systems
  • the electronic device 100 realizes the display function through the GPU, the display screen 194 , and the application processor.
  • the GPU is a microprocessor for image processing, and is connected to the display screen 194 and the application processor. GPUs are used to perform mathematical and geometric calculations for graphics rendering.
  • Processor 110 may include one or more GPUs that execute program instructions to generate or change display information.
  • the display screen 194 is used to display images, videos and the like.
  • the display screen 194 includes a display panel.
  • the display panel can be a liquid crystal display (LCD), an organic light-emitting diode (OLED), an active matrix organic light emitting diode or an active matrix organic light emitting diode (active-matrix organic light emitting diode, AMOLED), flexible light-emitting diode (flex light-emitting diode, FLED), Miniled, MicroLed, Micro-oLed, quantum dot light emitting diodes (quantum dot light emitting diodes, QLED), etc.
  • the electronic device 100 may include 1 or N display screens 194 , where N is a positive integer greater than 1.
  • the display screen 194 displays interface content currently output by the system.
  • the interface content is an interface provided by an instant messaging application.
  • the electronic device 100 can realize the shooting function through the ISP, the camera 193 , the video codec, the GPU, the display screen 194 and the application processor.
  • the ISP is used for processing the data fed back by the camera 193 .
  • the light is transmitted to the photosensitive element of the camera through the lens, and the light signal is converted into an electrical signal, and the photosensitive element of the camera transmits the electrical signal to the ISP for processing, and converts it into an image visible to the naked eye.
  • ISP can also optimize the algorithm for image noise and brightness.
  • ISP can also optimize the exposure, color temperature and other parameters of the shooting scene.
  • the ISP may be located in the camera 193 .
  • Camera 193 is used to capture still images or video.
  • the object generates an optical image through the lens and projects it to the photosensitive element.
  • the photosensitive element can be a charge coupled device (charge coupled device, CCD) or a complementary metal-oxide-semiconductor (complementary metal-oxide-semiconductor, CMOS) phototransistor.
  • CCD charge coupled device
  • CMOS complementary metal-oxide-semiconductor
  • the photosensitive element converts the light signal into an electrical signal, and then transmits the electrical signal to the ISP to convert it into a digital image signal.
  • the ISP outputs the digital image signal to the DSP for processing.
  • DSP converts digital image signals into standard RGB, YUV and other image signals.
  • the electronic device 100 may include 1 or N cameras 193 , where N is a positive integer greater than 1.
  • Digital signal processors are used to process digital signals. In addition to digital image signals, they can also process other digital signals. For example, when the electronic device 100 selects a frequency point, the digital signal processor is used to perform Fourier transform on the energy of the frequency point.
  • Video codecs are used to compress or decompress digital video.
  • the electronic device 100 may support one or more video codecs.
  • the electronic device 100 can play or record videos in various encoding formats, for example: moving picture experts group (moving picture experts group, MPEG) 1, MPEG2, MPEG3, MPEG4 and so on.
  • MPEG moving picture experts group
  • the NPU is a neural-network (NN) computing processor.
  • NN neural-network
  • Applications such as intelligent cognition of the electronic device 100 can be realized through the NPU, such as image recognition, face recognition, speech recognition, text understanding, and the like.
  • the external memory interface 120 can be used to connect an external memory card, such as a Micro SD card, so as to expand the storage capacity of the electronic device 100.
  • the external memory card communicates with the processor 110 through the external memory interface 120 to implement a data storage function. Such as saving music, video and other files in the external memory card.
  • the internal memory 121 may be used to store computer-executable program codes including instructions.
  • the processor 110 executes various functional applications and data processing of the electronic device 100 by executing instructions stored in the internal memory 121 .
  • the internal memory 121 may include an area for storing programs and an area for storing data.
  • the stored program area can store an operating system, at least one application program required by a function (such as a sound playing function, an image playing function, etc.) and the like.
  • the storage data area can store data created during the use of the electronic device 100 (such as audio data, phonebook, etc.) and the like.
  • the internal memory 121 may include a high-speed random access memory, and may also include a non-volatile memory, such as at least one magnetic disk storage device, flash memory device, universal flash storage (universal flash storage, UFS) and the like.
  • the electronic device 100 can implement audio functions through the audio module 170 , the speaker 170A, the receiver 170B, the microphone 170C, the earphone interface 170D, and the application processor. Such as music playback, recording, etc.
  • the audio module 170 is used to convert digital audio information into analog audio signal output, and is also used to convert analog audio input into digital audio signal.
  • the audio module 170 may also be used to encode and decode audio signals.
  • the audio module 170 may be set in the processor 110 , or some functional modules of the audio module 170 may be set in the processor 110 .
  • Speaker 170A also referred to as a "horn" is used to convert audio electrical signals into sound signals.
  • Electronic device 100 can listen to music through speaker 170A, or listen to hands-free calls.
  • Receiver 170B also called “earpiece” is used to convert audio electrical signals into sound signals.
  • the receiver 170B can be placed close to the human ear to receive the voice.
  • the microphone 170C also called “microphone” or “microphone” is used to convert sound signals into electrical signals. When making a phone call or sending a voice message, the user can put his mouth close to the microphone 170C to make a sound, and input the sound signal to the microphone 170C.
  • the electronic device 100 may be provided with at least one microphone 170C. In some other embodiments, the electronic device 100 may be provided with two microphones 170C, which may also implement a noise reduction function in addition to collecting sound signals. In some other embodiments, the electronic device 100 can also be provided with three, four or more microphones 170C to collect sound signals, reduce noise, identify sound sources, and realize directional recording functions, etc.
  • the earphone interface 170D is used for connecting wired earphones.
  • the earphone interface 170D may be a USB interface 130, or a 3.5mm open mobile terminal platform (OMTP) standard interface, or a cellular telecommunications industry association of the USA (CTIA) standard interface.
  • OMTP open mobile terminal platform
  • CTIA cellular telecommunications industry association of the USA
  • the pressure sensor 180A is used to sense the pressure signal and convert the pressure signal into an electrical signal.
  • pressure sensor 180A may be disposed on display screen 194 .
  • the pressure sensor 180A can be used to capture the pressure value generated when the user's finger part touches the display screen, and transmit the pressure value to the processor, so that the processor can identify which finger part the user inputs user action.
  • a capacitive pressure sensor may be comprised of at least two parallel plates with conductive material.
  • the electronic device 100 determines the intensity of pressure according to the change in capacitance.
  • the electronic device 100 detects the intensity of the touch operation according to the pressure sensor 180A.
  • the electronic device 100 may also calculate the touched position according to the detection signal of the pressure sensor 180A. In some embodiments, different touch positions may correspond to different operation instructions.
  • the pressure sensor 180A can also calculate the number of touch points according to the detected signal, and transmit the calculated value to the processor, so that the processor can recognize that the user inputs a user operation with a single finger or multiple fingers.
  • the gyro sensor 180B can be used to determine the motion posture of the electronic device 100 .
  • the angular velocity of the electronic device 100 around three axes can be determined by the gyro sensor 180B.
  • the gyro sensor 180B can be used for image stabilization. Exemplarily, when the shutter is pressed, the gyro sensor 180B detects the shaking angle of the electronic device 100, calculates the distance that the lens module needs to compensate according to the angle, and allows the lens to counteract the shaking of the electronic device 100 through reverse movement to achieve anti-shake.
  • the gyro sensor 180B can also be used for navigation and somatosensory game scenes.
  • the air pressure sensor 180C is used to measure air pressure.
  • the electronic device 100 calculates the altitude based on the air pressure value measured by the air pressure sensor 180C to assist positioning and navigation.
  • the magnetic sensor 180D includes a Hall sensor.
  • the electronic device 100 may use the magnetic sensor 180D to detect the opening and closing of the flip leather case.
  • the electronic device 100 when the electronic device 100 is a clamshell machine, the electronic device 100 can detect opening and closing of the clamshell according to the magnetic sensor 180D.
  • features such as automatic unlocking of the flip cover are set.
  • the acceleration sensor 180E can detect the acceleration of the electronic device 100 in various directions (generally three axes). The magnitude and direction of gravity can be detected when the electronic device 100 is stationary. It can also be used to identify the posture of electronic devices, and can be used in applications such as horizontal and vertical screen switching, pedometers, etc. In some optional embodiments of the present application, the acceleration sensor 180E may be used to capture the acceleration value generated when the user's finger touches the display screen (or the user's finger taps the rear side frame of the electronic device 100), and store the acceleration value The data is transmitted to the processor, so that the processor identifies which finger part the user uses to input the user operation.
  • the distance sensor 180F is used to measure the distance.
  • the electronic device 100 may measure the distance by infrared or laser. In some embodiments, when shooting a scene, the electronic device 100 may use the distance sensor 180F for distance measurement to achieve fast focusing.
  • Proximity light sensor 180G may include, for example, light emitting diodes (LEDs) and light detectors, such as photodiodes.
  • the light emitting diodes may be infrared light emitting diodes.
  • the electronic device 100 emits infrared light through the light emitting diode.
  • Electronic device 100 uses photodiodes to detect infrared reflected light from nearby objects. When sufficient reflected light is detected, it may be determined that there is an object near the electronic device 100 . When insufficient reflected light is detected, the electronic device 100 may determine that there is no object near the electronic device 100 .
  • the electronic device 100 can use the proximity light sensor 180G to detect that the user holds the electronic device 100 close to the ear to make a call, so as to automatically turn off the display screen to save power.
  • the proximity light sensor 180G can also be used in leather case mode, automatic unlock and lock screen in pocket mode.
  • the ambient light sensor 180L is used for sensing ambient light brightness.
  • the electronic device 100 can adaptively adjust the brightness of the display screen 194 according to the perceived ambient light brightness.
  • the ambient light sensor 180L can also be used to automatically adjust the white balance when taking pictures.
  • the ambient light sensor 180L can also cooperate with the proximity light sensor 180G to detect whether the electronic device 100 is in the pocket, so as to prevent accidental touch.
  • the fingerprint sensor 180H is used to collect fingerprints.
  • the electronic device 100 can use the collected fingerprint characteristics to implement fingerprint unlocking, access to application locks, take pictures with fingerprints, answer incoming calls with fingerprints, and the like.
  • the temperature sensor 180J is used to detect temperature.
  • the electronic device 100 uses the temperature detected by the temperature sensor 180J to implement a temperature treatment strategy. For example, when the temperature reported by the temperature sensor 180J exceeds the threshold, the electronic device 100 may reduce the performance of the processor located near the temperature sensor 180J, so as to reduce power consumption and implement thermal protection.
  • the electronic device 100 when the temperature is lower than another threshold, the electronic device 100 heats the battery 142 to avoid abnormal shutdown of the electronic device 100 caused by the low temperature.
  • the electronic device 100 boosts the output voltage of the battery 142 to avoid abnormal shutdown caused by low temperature.
  • Touch sensor 180K also known as "touch panel”.
  • the touch sensor 180K can be disposed on the display screen 194, and the touch sensor 180K and the display screen 194 form a touch screen, also called a “touch screen”.
  • the touch sensor 180K is used to detect a touch operation on or near it, and the touch operation refers to an operation in which the user's hand, elbow, stylus, etc. touch the display screen 194 .
  • the touch sensor can pass the detected touch operation to the application processor to determine the type of touch event.
  • Visual output related to the touch operation can be provided through the display screen 194 .
  • the touch sensor 180K may also be disposed on the surface of the electronic device 100 , which is different from the position of the display screen 194 .
  • the bone conduction sensor 180M can acquire vibration signals. In some embodiments, the bone conduction sensor 180M can acquire the vibration signal of the vibrating bone mass of the human voice. The bone conduction sensor 180M can also contact the human pulse and receive the blood pressure beating signal. In some embodiments, the bone conduction sensor 180M can also be disposed in the earphone, combined into a bone conduction earphone.
  • the audio module 170 can analyze the voice signal based on the vibration signal of the vibrating bone mass of the vocal part acquired by the bone conduction sensor 180M, so as to realize the voice function.
  • the application processor can analyze the heart rate information based on the blood pressure beating signal acquired by the bone conduction sensor 180M, so as to realize the heart rate detection function.
  • the keys 190 include a power key, a volume key and the like.
  • the key 190 may be a mechanical key. It can also be a touch button.
  • the electronic device 100 may receive key input and generate key signal input related to user settings and function control of the electronic device 100 .
  • the motor 191 can generate a vibrating reminder.
  • the motor 191 can be used for incoming call vibration prompts, and can also be used for touch vibration feedback.
  • touch operations applied to different applications may correspond to different vibration feedback effects.
  • the motor 191 may also correspond to different vibration feedback effects for touch operations acting on different areas of the display screen 194 .
  • Different application scenarios for example: time reminder, receiving information, alarm clock, games, etc.
  • the touch vibration feedback effect can also support customization.
  • the indicator 192 can be an indicator light, and can be used to indicate charging status, power change, and can also be used to indicate messages, missed calls, notifications, and the like.
  • the SIM card interface 195 is used for connecting a SIM card.
  • the SIM card can be connected and separated from the electronic device 100 by inserting it into the SIM card interface 195 or pulling it out from the SIM card interface 195 .
  • the software systems of the electronic device 100 and the electronic device 101 may adopt a layered architecture, an event-driven architecture, a micro-kernel architecture, a micro-service architecture, or a cloud architecture.
  • the embodiment of the present application takes the Android system as an example to illustrate the software architecture of the electronic device 100 .
  • the Android system is only a system example of the electronic device 100 or the electronic device 101 in the embodiment of the present application, and the present application can also be applied to other types of operating systems, such as ios, Windows, Hongmeng, etc. limit.
  • the following only uses the Android system as an example of the operating system of the electronic device 100 .
  • Fig. 3b is a block diagram of the software structure of the electronic device 100 according to the embodiment of the present invention.
  • the layered architecture divides the software into several layers, and each layer has a clear role and division of labor. Layers communicate through software interfaces.
  • the Android system is divided into four layers, which are respectively the application program layer, the application program framework layer, Android runtime (Android runtime) and system libraries, and the kernel layer from top to bottom.
  • the application layer can consist of a series of application packages.
  • the application package may include applications such as camera, gallery, calendar, call, map, navigation, WLAN, Bluetooth, music, video, and short message.
  • the application framework layer provides an application programming interface (application programming interface, API) and a programming framework for applications in the application layer.
  • the application framework layer includes some predefined functions.
  • the application framework layer can include window manager, content provider, view system, phone manager, resource manager, notification manager, etc.
  • a window manager is used to manage window programs.
  • the window manager can get the size of the display screen, determine whether there is a status bar, lock the screen, capture the screen, etc.
  • Content providers are used to store and retrieve data and make it accessible to applications.
  • Said data may include video, images, audio, calls made and received, browsing history and bookmarks, phonebook, etc.
  • the view system includes visual controls, such as controls for displaying text, controls for displaying pictures, and so on.
  • the view system can be used to build applications.
  • a display interface can consist of one or more views.
  • a display interface including a text message notification icon may include a view for displaying text and a view for displaying pictures.
  • the phone manager is used to provide communication functions of the electronic device 100 . For example, the management of call status (including connected, hung up, etc.).
  • the resource manager provides various resources for the application, such as localized strings, icons, pictures, layout files, video files, and so on.
  • the notification manager enables the application to display notification information in the status bar, which can be used to convey notification-type messages, and can automatically disappear after a short stay without user interaction.
  • the notification manager is used to notify the download completion, message reminder, etc.
  • the notification manager can also be a notification that appears on the top status bar of the system in the form of a chart or scroll bar text, such as a notification of an application running in the background, or a notification that appears on the screen in the form of a dialog window.
  • prompting text information in the status bar issuing a prompt sound, vibrating the electronic device, and flashing the indicator light, etc.
  • the Android Runtime includes core library and virtual machine. The Android runtime is responsible for the scheduling and management of the Android system.
  • the core library consists of two parts: one part is the function function that the java language needs to call, and the other part is the core library of Android.
  • the application layer and the application framework layer run in virtual machines.
  • the virtual machine executes the java files of the application program layer and the application program framework layer as binary files.
  • the virtual machine is used to perform functions such as object life cycle management, stack management, thread management, security and exception management, and garbage collection.
  • a system library can include multiple function modules. For example: surface manager (surface manager), media library (Media Libraries), 3D graphics processing library (eg: OpenGL ES), 2D graphics engine (eg: SGL), etc.
  • the surface manager is used to manage the display subsystem and provides the fusion of 2D and 3D layers for multiple applications.
  • the media library supports playback and recording of various commonly used audio and video formats, as well as still image files, etc.
  • the media library can support a variety of audio and video encoding formats, such as: MPEG4, H.264, MP3, AAC, AMR, JPG, PNG, etc.
  • the 3D graphics processing library is used to implement 3D graphics drawing, image rendering, compositing, and layer processing, etc.
  • 2D graphics engine is a drawing engine for 2D drawing.
  • the kernel layer is the layer between hardware and software.
  • the kernel layer includes at least a display driver, a camera driver, an audio driver, and a sensor driver.
  • the structure illustrated in the embodiment of the present application does not constitute a specific limitation on the electronic device 100 .
  • the electronic device 100 may include more or fewer components than shown in the figure, or combine certain components, or separate certain components, or arrange different components.
  • the illustrated components can be realized in hardware, software or a combination of software and hardware.
  • S101 The electronic device 100 establishes a connection with the electronic device 101.
  • the electronic device 100 and the electronic device 101 may be connected by wired or wireless communication, and the wireless communication may include, for example, Bluetooth, near field communication (near field communication, NFC), wireless fidelity (wireless fidelity, WiFi), or WiFi direct connection.
  • the electronic device 101 includes a camera, a camera, a tablet, an iPad, a mobile phone and other electronic devices with a camera.
  • the electronic device 100 may also be called a first electronic device, and the electronic device 101 may also be called a second electronic device.
  • the electronic device 100 and the electronic device 101 may perform data communication through a server, and the server may be an application server, a cloud server, a background server, and the like.
  • the electronic device 101 can be a mobile phone, a camera, etc.; or if the electronic device 100 is a mobile phone, the electronic device 101 can be a vehicle camera, other mobile phones, etc.
  • the electronic device 101 can be mounted on the electronic device 100, for example, the electronic device 100 is a vehicle, and the electronic device 101 is a vehicle camera; and for example, the electronic device 100 is a mobile phone, and the electronic device 101 is a camera on the mobile phone.
  • the electronic device 100 and the electronic device 101 are started, the electronic device 101 can take pictures and/or videos, and the electronic device 100 can synchronously acquire the pictures and/or video data taken by the electronic device 101 .
  • the electronic device 100 establishes a connection with multiple electronic devices 101, and the electronic device 100 can acquire the picture or video data taken by the multiple electronic devices 101 synchronously.
  • the electronic device 100 is a vehicle
  • the electronic device 101 is a vehicle camera and a camera on a mobile phone.
  • the electronic device 100 determines that the geographic location information of the electronic device 100 satisfies a first preset condition.
  • the electronic device 100 acquires its own geographic location information, where the geographic location information may include absolute geographic location information determined by the GPS system, such as latitude and longitude; it may also include a geographic location area located by a navigation system, such as a certain street in a certain area, etc. ; may also include the relative position of the electronic device 100 from the initial starting position; may also include the relative position of the electronic device 100 from a preset position.
  • the geographic location information may include absolute geographic location information determined by the GPS system, such as latitude and longitude; it may also include a geographic location area located by a navigation system, such as a certain street in a certain area, etc. ; may also include the relative position of the electronic device 100 from the initial starting position; may also include the relative position of the electronic device 100 from a preset position.
  • the first preset condition is that the current geographic location of the electronic device 100 leaves the user's resident area.
  • the resident area of the user may include one or more areas, and the resident area of the user may be a city as the smallest area, a region as the smallest area, a street as the smallest area, and so on.
  • the user's resident area is preset by the electronic device 100, and the user can add or delete the user's resident area.
  • the user's resident area is calculated by the electronic device 100 based on its historical movement trajectory, and the electronic device 100 may update (add or delete) the user's resident area.
  • the first preset condition is that the current geographic location of the electronic device 100 leaves the user's resident area and the duration of leaving exceeds the first threshold.
  • the first preset condition is that the shortest distance between the current geographic location of the electronic device 100 and the preset location of the electronic device 100 is greater than or equal to the second threshold.
  • the second threshold be 30 kilometers
  • the preset position of the electronic device 100 is point A
  • the distance between point B and point A is 30 kilometers or greater than 30 kilometers
  • the preset location may be a resident area of the user.
  • the first preset condition is that the current geographic location of the electronic device 100 enters a specific area, such as a scenic area, a tourist attraction, and the like.
  • the specific area is updated in real time, for example, new scenic spots are added based on the network; optionally, the specific area is preset by the user.
  • the first preset condition is that the distance between the current geographic location of the electronic device 100 and the initial starting position of the electronic device 100 is greater than or equal to a third threshold.
  • a third threshold be 30 kilometers
  • the electronic device 100 starts to move from point A, when the electronic device 100 moves to point B, and the distance between point B and point A is 30 kilometers or more, then the electronic device 100 at this time
  • the current geographic location (point B) of the device itself satisfies the first preset condition.
  • the first preset condition is that the electronic device 100 detects that the electronic device 100 enters a highway through a highway toll booth.
  • the electronic device 100 detects that the current geographical location of the electronic device 100 passes a high-speed toll station.
  • the electronic device 100 detects that the electronic device 100 passes a high-speed toll station based on the sensor detection.
  • the first preset condition is that the electronic device 100 receives a user confirmation instruction.
  • the electronic device 100 receives the user confirmation instruction, the electronic device 100 satisfies the first preset condition.
  • S103 The electronic device 100 analyzes the video captured by the electronic device 101 in real time.
  • the electronic device 100 determines that its own geographic location information satisfies the first preset condition, the electronic device 100 analyzes the video captured by the electronic device 101 in real time.
  • the electronic device 101 is a vehicle camera connected to the vehicle.
  • the vehicle When the vehicle is started, the vehicle power supplies power to the vehicle and the vehicle camera, and the vehicle camera can shoot video in real time. Make a note.
  • the car-machine determines that its current geographical location information satisfies the first preset condition, at this time, the car-machine obtains the video currently captured by the car camera in real time for analysis.
  • the electronic device 100 determines that its geographic location information satisfies the first preset condition, the electronic device 101 starts to shoot a video, and the electronic device 100 analyzes the video captured by the electronic device 101 in real time.
  • the electronic device 101 is shooting a video, and when the electronic device 100 determines that its geographic location information satisfies a first preset condition, the electronic device 100 analyzes the video captured by the electronic device 101 in real time. If the electronic device 101 detects that its own geographic location information satisfies the first preset condition after the electronic device 101 shoots for 5 minutes, then the electronic device 100 analyzes the video currently shot by the electronic device 101, that is, after 5 minutes video for analysis.
  • the electronic device 100 is a vehicle
  • the electronic device 101 is a vehicle camera. When the vehicle is started, the car machine and the on-board camera are powered, and the on-board camera starts to shoot video.
  • the car-machine determines that its own geographical location information meets the first preset condition, the car-machine then synchronously analyzes the video captured by the car camera in real time.
  • the following describes in detail the method and principle for the electronic device 100 to analyze the video captured by the electronic device 101 in real time.
  • the electronic device 100 synchronously acquires the video captured by the electronic device 101 in real time, selects a picture of the acquired video, and uses the selected picture as a piece of content in the finally generated new video.
  • the electronic device 100 acquires the video images captured by the electronic device 101 synchronously, and the electronic device 100 performs frame extraction on the acquired video images and scores them based on the aesthetic scoring algorithm.
  • the electronic device 100 determines the video segment with the highest score in the last scene. Then continue to score frames for the next scene until the end.
  • the electronic device 100 selects the video picture with the highest score in real time through the sliding window, so that the electronic device 100 can generate the first video when the second preset condition is met, without waiting for a long time analysis in the background.
  • the electronic device 100 is a car
  • the car has lower requirements on power consumption, and the analysis consumes an average of 800-1000ms per frame, saving resources.
  • the electronic device 100 segments the acquired video data, and then selects a part of the frame in each segmented video according to a preset method.
  • the electronic device 100 may segment the acquired video data by segmenting based on the scene shots, and the scene shots include character shots, landscape shots, transition shots, etc., as shown in FIG. 5b
  • FIG. 5b exemplarily shows the video captured by the electronic device 101 in real time
  • the electronic device 100 acquires the video captured by the electronic device 101 synchronously
  • the electronic device 100 recognizes the image frames in the video based on the image recognition technology, thereby classifying different categories
  • Figure 5b shows three types of scene shots, each type of scene shot corresponds to a section of video, the image frames in the video before time point 1 are classified as scene shot 1; the video frames between time point 1 and time point 2 The image frames in the video after time point 2 are classified as scene shot 2; the image frames in the video after time point 2 are classified as scene shot 3.
  • the electronic device 100 Each time the electronic device 100 divides a segment of video, it selects a frame of the segment of video, and takes the selected frame as a segment of content in the finally generated new video. In some embodiments, the electronic device 100 may generate a corresponding name or a brief description for the selected video frame based on geographic location information or video content, so as to improve the user's viewing experience of the generated new video.
  • the mode of picture selection can be to divide this section of video into several sections of videos, and then perform aesthetic scoring on each section of video in the several sections of videos, and select a section of video with the highest score as the final video in the generated new video. a piece of content.
  • the way to select the picture can be to refine the scene of this section of video, classify the scene again, perform an aesthetic score on each scene in this section of video, and select the scene with the highest score as the final generated new video. a piece of content.
  • the method of picture selection may be to extract frames at equal intervals for the segment of video, as shown in FIG. 5c .
  • scene shot 1 is taken as an example to exemplarily show a way of picture selection for a scene shot.
  • the electronic device 100 extracts 9 image frames from the scene shot 1, and the 9 image frames divide the scene shot 1 into 8 video frames.
  • the electronic device 100 performs an aesthetic score on the nine image frames, and determines the score corresponding to each image frame, thereby determining the score of each video frame.
  • the score of each video frame can be the sum of two image frames. Average rating.
  • the score of video frame 1 is the average score of image frame 1 and image frame 2
  • the score of video frame 2 is the average score of image frame 2 and image frame 3, and so on.
  • the electronic device 100 selects a segment of video frame with the highest score as a segment of content in the finally generated new video.
  • the electronic device 100 selects video frames with scores greater than the fourth threshold as content in the finally generated new video. In this way, there may be a situation where no frame is selected in a scene shot, or a situation where a large number of frames are selected in a scene shot.
  • the scoring rules of the aesthetic scoring can be changed according to the scene type of the video picture.
  • the electronic device 100 determines the scene type of the video segment based on the current geographic location of the electronic device 100 or image recognition technology, thereby determining the scoring rule corresponding to the scene type.
  • the landform features Danxia landform
  • the aesthetic scoring rules can increase or add the weight of the degree of color and terrain level; on the Qiandao Lake Ring Road, most of them are erosion and denudation low mountains and hills
  • the weight of water, island, and sky can be increased or added to the aesthetic scoring rules; when driving inside the city, such as Shanghai Yan’an Road Elevated Road, the brightness of the lights, sky and tall buildings can be increased or added to the aesthetic scoring rules The weight of the proportion of ; and so on.
  • customized aesthetic scoring rules can be selected to select the most suitable picture fragments for the public's aesthetics and improve the user experience.
  • the process of analyzing the video captured by the electronic device 101 in real time by the above-mentioned electronic device 100 is carried out synchronously with the video frame captured by the electronic device 101, and the electronic device 100 synchronously receives the video frame captured by the electronic device 101,
  • the video screen is segmented, and a part of the screen is intercepted as the content of the finally generated new video for each segment, until the geographic location information of the electronic device 100 satisfies the second preset condition, and the electronic device 100 generates the first video.
  • S104 Determine that the geographic location information of the electronic device 100 satisfies a second preset condition.
  • the second preset condition is that the current geographic location of the electronic device 100 reaches the destination.
  • the destination determination may be based on navigation software.
  • the destination may also be referred to as a preset location.
  • the second preset condition is that the remaining time for the electronic device 100 to reach the destination at its current geographic location is less than the fifth threshold.
  • the fifth threshold may be determined by navigation software.
  • the second preset condition is that the electronic device 100 detects that its own geographic location does not change or changes within a small range within a preset time.
  • the second preset condition is that the electronic device 100 detects that the electronic device 100 leaves the highway through a highway toll booth.
  • the electronic device 100 detects that the current geographic location of the electronic device 100 leaves the high-speed toll station.
  • the electronic device 100 detects that the electronic device 100 leaves the high-speed toll station based on the sensor detection.
  • the second preset condition is that the electronic device 100 receives a user confirmation instruction.
  • the electronic device 100 receives the user confirmation instruction, the electronic device 100 satisfies the second preset condition.
  • the electronic device 100 generates a first video, where the first video is obtained by the electronic device 100 through analysis and clipping of the video captured by the electronic device 101 .
  • the electronic device 100 determines that the geographic location information of the electronic device 100 satisfies the second preset condition, the electronic device 100 no longer analyzes the video content captured by the electronic device 101 later, and the electronic device 100 generates the first video based on the video content acquired before. (new video).
  • the first video is obtained by analyzing and clipping the video content acquired by the electronic device 100 .
  • the electronic device 100 synchronously receives the video picture taken by the electronic device 101, segments the received video picture, and intercepts a part of the picture for each segment as the content in the finally generated new video, until the electronic device 100 The geographic location information of 100 satisfies the second preset condition, and the electronic device 100 generates a new video.
  • the electronic device 100 synchronously receives the video screen shot by the electronic device 101, divides the scene shot 1, and selects The video frame 1 with the highest score in the scene shot 1; the electronic device 100 continues to receive the video frames shot by the electronic device 100, divides the scene shot 12, and selects the video frame 24 with the highest score in the scene shot 2.
  • the video frame can be 1 and video picture 2 are spliced together; ...; the electronic device 100 continues to receive the video pictures taken by the electronic device 100, divides the scene shot n-1, and selects the video picture a with the highest score in the scene shot n-1.
  • the electronic device 100 When the video frame a and the video frame selected in the scene shot n-2 can be spliced together; the electronic device 100 continues to receive the video frames captured by the electronic device 100, and the electronic device 100 determines that the geographic location information of the electronic device 100 satisfies the second preset After setting the conditions, the electronic device 100 divides the last scene shot n, and selects the video frame m with the highest score in the scene shot n, at this time, the video frame m and the video frame a can be spliced together; at this time, the electronic device 100 generates A first video, where the first video is formed by splicing n video frames.
  • a transition effect or a transition picture may be inserted between each video picture in the first video.
  • more than one video frame may be selected in each scene shot, or no video frame may be selected.
  • the electronic device 100 generates the first video and saves the first video. The user can view and play the first video on the electronic device 100 .
  • the electronic device 100 generates the first video, and based on the geographic location information of the electronic device 100 or the video content of the first video, generates a video name for the first video, such as Yunnan Erhai Tour, Thousand Islands
  • a video name for the first video such as Yunnan Erhai Tour, Thousand Islands
  • the rich and colorful customized names such as the trip around the lake solve the problem of the single name of the photo album and video in the current gallery, and enrich the ecology of the gallery to record life for users.
  • the electronic device 100 after the electronic device 100 generates the first video, the electronic device 100 outputs prompt information, which is used to remind the user that the first video has been generated and provide a processing method for the first video, including saving the first video, Delete first video, edit first video, share to app, send to other devices and more.
  • FIG. 7 takes the electronic device 100 as an example of a vehicle, and shows an application interface 701 in a navigation application after the electronic device 100 generates the first video.
  • the application interface 701 includes navigation The map 702, the tool bar 703 and the prompt information 711 generated by the first video, wherein,
  • the toolbar 703 includes one or more functional controls, providing functions such as making a phone call and playing music.
  • Prompt information 711 prompts the user that the first video has been generated (shown in FIG. 7 as "the vlog on this drive has been generated"), and the prompt information 711 includes a thumbnail 712 of the first video, a play control 713, and a save control 714, delete control 715, edit control 716, share control 717, continue clip control 718 and restart control 719.
  • Play control 713 is used to play the first video.
  • the save control 714 is used to save the first video to the electronic device 100 .
  • Delete control 715 is used to delete the first video.
  • the edit control 716 is used to edit the first video, including, for example, changing the order of the video frames in the first video, deleting some frames in the first video, inserting transition effects between two video frames, and adding text, stickers, and watermarks etc., add background music, edit the video name of the first video, etc.
  • the sharing control 717 is used to send the first video to other devices, or share it to an application program.
  • the continue clipping control 718 is used to instruct the electronic device 100 to continue to receive and analyze the video images captured by the electronic device 101 in real time, and regenerate the first video.
  • the continuation editing control can be applied to the situation that the electronic device 100 wrongly judges that the electronic device 100 satisfies the second preset condition. It is desired to edit the video content of the two drives into one vlog, and the user can select the electronic device 100 to continue analyzing and editing the video images captured by the electronic device 101 in real time.
  • the restart control 719 is used to instruct the electronic device 100 to delete the first video, and to re-receive and analyze the video images captured by the electronic device 101 in real time.
  • the electronic device 100 sends the first video to other electronic devices.
  • the electronic device 100 After the electronic device 100 determines that the geographic location information of the electronic device 100 satisfies the second preset condition, the electronic device 100 generates a first video and sends the first video to other electronic devices.
  • the other electronic devices may include the electronic device 101 , or other electronic devices that have established a connection with the electronic device 100 .
  • other electronic devices receive the first video
  • the other electronic devices output prompt information, which is used to prompt the user to receive the first video and provide a processing method for the first video, including confirming receipt of the first video.
  • video save first video
  • delete first video edit first video, share to app and more.
  • FIG. 8 exemplarily shows a lock screen interface 801 .
  • the lock screen interface 801 shown in FIG. 8 includes a status bar 802 and a notification 803 (prompt information), wherein,
  • Status bar 802 may include: one or more signal strength indicators 804 for mobile communication signals (also referred to as cellular signals), one or more signal strength indicators 805 for wireless fidelity (Wi-Fi) signals , Bluetooth indicator 806, battery status indicator 807.
  • signal strength indicators 804 for mobile communication signals also referred to as cellular signals
  • signal strength indicators 805 for wireless fidelity (Wi-Fi) signals Bluetooth indicator 806, battery status indicator 807.
  • Wi-Fi wireless fidelity
  • Bluetooth indicator 806 battery status indicator 807.
  • Notification 803 prompts that the user has received the first video (shown in FIG. 8 as "received the vlog on the way of driving, click to view"), and the prompt information notification 803 includes a play control 808, a delete control 809, an edit Control 810, share control 811.
  • the play control 808 is used to play the received first video.
  • Delete control 809 is used to delete the first video.
  • the edit control 810 is used to edit the first video, including, for example, changing the order of the video frames in the first video, deleting some frames in the first video, inserting transition effects between two video frames, and adding text, stickers, and watermarks etc., add background music, etc.
  • the sharing control 811 is used to send the first video to other electronic devices, or share it to an application program.
  • the electronic device 100 determines the start and end time of the analysis of the video image based on its own geographic location.
  • the video data is analyzed and edited. That is, while the electronic device 101 is shooting a video, the electronic device 100 simultaneously analyzes and edits the video.
  • the electronic device 100 When the geographic location of the electronic device 100 satisfies the second preset condition, the electronic device 100 generates the first video, and does not analyze and edit the video data captured by the electronic device 101 afterwards.
  • the first video is obtained by analyzing and editing video content obtained by the electronic device 100 . In this way, the manner in which the electronic device 100 synchronously analyzes the video content captured in real time can improve the efficiency of video generation and reduce the waiting time of the user.
  • the electronic device 100 is a vehicle, and the electronic device 101 may be a device with a camera such as a vehicle camera or a mobile phone.
  • the electronic device 101 can capture the roadside scenery during driving in real time, then the electronic device 100 can analyze and edit the video data or pictures captured in real time synchronously.
  • the electronic device 100 can then generate a travel video about the driving process.
  • the electronic device 100 can send the video to other electronic devices.
  • the electronic device 100 is a vehicle, and the electronic device 101 can be a camera on a mobile phone.
  • the mobile phone can capture the scenery on the roadside during driving in real time. video.
  • the electronic device 100 can send the video to other electronic devices, such as mobile phones.
  • the electronic device 100 is a mobile phone
  • the electronic device 101 can be a car camera, which can capture roadside scenery in real time during driving, and the mobile phone can generate a travel video about the driving process after the driving is over.
  • the electronic device 101 includes a plurality of electronic devices.
  • the electronic device 100 is a vehicle, and the electronic device 101 may be a vehicle camera or a camera on a mobile phone, that is, the electronic device 100 may receive video data captured by multiple electronic devices 101 in real time.
  • the electronic device 100 may perform analysis based on multiple pieces of video data acquired during the same period of time, and select multiple pieces of video content from the multiple pieces of video and stitch them together.
  • the video captured by the electronic device 101 is taken as an example.
  • the electronic device 101 can also capture images, and the electronic device 100 can synchronously acquire the image data captured by the electronic device 101 in real time.
  • the electronic device 100 acquires image data captured by the electronic device 101 in real time.
  • the electronic device 100 starts to analyze and edit the image data captured by the electronic device 101 in real time. That is, while the electronic device 101 captures images, the electronic device 100 synchronously analyzes and edits the received images.
  • the electronic device 100 performs scene classification on a series of captured images, and selects a part of the images in each scene as part of the content in the final generated video. taller photos. Furthermore, before aesthetic scoring, the electronic device 100 screens out photos with high repetition rate based on image recognition technology, and selects one or more of them (for example, the one with the highest aesthetic score) to keep.
  • the electronic device 100 first determines the number n of images needed, the electronic device 100 performs an aesthetic score on each received image, and the top n images are used as the content of the finally generated video.
  • the electronic device 100 When the geographic location of the electronic device 100 satisfies the second preset condition, the electronic device 100 generates the first video, and does not analyze and edit the image data captured by the electronic device 101 afterwards.
  • the first video is obtained by analyzing and editing based on the images acquired by the electronic device 100 . In this way, the manner in which the electronic device 100 synchronously analyzes the images captured in real time can improve the efficiency of video generation and reduce the waiting time of the user.
  • the electronic device 100 may acquire image data and video data sent by the electronic device 101 .
  • the electronic device 100 starts to analyze and edit the image data and video data captured by the electronic device 101 in real time. That is, while the electronic device 101 captures images and videos, the electronic device 100 synchronously analyzes and edits the received image data and video data.
  • the electronic device 100 generates the first video, and does not analyze and edit the image data and video data captured by the electronic device 101 afterwards.
  • all or part of them may be implemented by software, hardware, firmware or any combination thereof.
  • software When implemented using software, it may be implemented in whole or in part in the form of a computer program product.
  • the computer program product includes one or more computer instructions. When the computer program instructions are loaded and executed on the computer, the processes or functions according to the embodiments of the present application will be generated in whole or in part.
  • the computer can be a general purpose computer, a special purpose computer, a computer network, or other programmable devices.
  • the computer instructions may be stored in or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer instructions may be transmitted from a website, computer, server or data center Transmission to another website site, computer, server, or data center by wired (eg, coaxial cable, optical fiber, DSL) or wireless (eg, infrared, wireless, microwave, etc.) means.
  • the computer-readable storage medium may be any available medium that can be accessed by a computer, or a data storage device such as a server or a data center integrated with one or more available media.
  • the available media may be magnetic media (eg, floppy disk, hard disk, magnetic tape), optical media (eg, DVD), or semiconductor media (eg, solid state hard disk), etc.
  • the processes can be completed by computer programs to instruct related hardware.
  • the programs can be stored in computer-readable storage media.
  • When the programs are executed may include the processes of the foregoing method embodiments.
  • the aforementioned storage medium includes: ROM or random access memory RAM, magnetic disk or optical disk, and other various media that can store program codes.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

公开了一种视频生成方法及相关装置,其特征在于,该方法包括:当第一电子设备的地理位置满足第一预设条件,第一电子设备开始对其他电子设备(例如第二电子设备)实时拍摄到的视频和/或图像数据进行分析和剪辑,从而选取出该实时拍摄的视频和/或图像中的部分画面。即其他电子设备一边拍摄视频和/或图像,第一电子设备一边同步对该视频和/或图像进行画面选取。当第一电子设备的地理位置满足第二预设条件,第一电子设备生成第一视频,并不再对之后拍摄到的视频和/或图像进行分析和剪辑。该第一视频即为基于上述选取的部分画面拼接得到的。这种同步分析实时拍摄到的视频和/或图像的方式,可以提高视频生成的效率,减少用户等待的时间。

Description

视频生成系统、方法及相关装置
本申请要求于2021年12月22日提交中国专利局、申请号为202111580204.3、申请名称为“视频生成系统、方法及相关装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及电子技术领域,尤其涉及一种视频生成系统、方法及相关装置。
背景技术
手机可以基于相册中保存的图片或视频,对一段时间内或者一个主题下的图片或视频进行剪辑、拼接,自动生成一段新的视频(也可以称为vlog),该新的视频可以是一段时间内的精彩回顾,例如周末的精彩回顾、节假日的精彩回顾等等;也可以是一个主题下的精彩瞬间,例如运动主题下的精彩瞬间、旅游主题下的精彩瞬间、毕业主题下的精彩瞬间等等。
目前,相册中的精彩瞬间视频的生成过程需要在电子设备在充电灭屏的过程中完成,数据量大时,分析时间长,若有大量照片/视频的情况下,用户等待时间久。并且,在旅行途中,经常会有转瞬即逝的美景,用户想要拍照或者录像记录但来不及。因此,如何提高视频生成的效率,减少用户等待的时间是本领域技术人员正在研究的问题。
发明内容
本申请实施例提供了一种视频生成系统、方法及相关装置,可以同步对实时拍摄的视频进行分析和剪辑,提高视频生成的效率,减少用户等待的时间。
第一方面,本申请提供了一种视频生成方法,该方法包括:第一电子设备确定自身的地理位置信息满足第一预设条件;第一电子设备获取实时拍摄的视频和/或图像;第一电子设备选取该实时拍摄的视频和/或图像中的部分画面;当第一电子设备的地理位置信息满足第二预设条件,第一电子设备生成第一视频,第一视频是基于所述部分画面拼接而成的。
本申请实施例,第一电子设备基于自身的地理位置确定对视频和/或图像的分析开始和分析结束时间,当第一电子设备的地理位置满足第一预设条件,第一电子设备开始对其他电子设备(例如第二电子设备)实时拍摄到的视频和/或图像数据进行分析和剪辑,从而选取出该实时拍摄的视频和/或图像中的部分画面。即其他电子设备一边拍摄视频,第一电子设备一边同步对该视频进行画面选取。当第一电子设备的地理位置满足第二预设条件,第一电子设备生成第一视频,并不再对之后拍摄到的视频和/或图像进行分析和剪辑。该第一视频即为基于上述选取的部分画面拼接得到的。这种同步分析实时拍摄到的视频和/或图像的方式,可以提高视频生成的效率,减少用户等待的时间。
基于第一方面,在一种可能的实现方式中,第一预设条件包括:第一电子设备当前自身的地理位置离开了用户常驻区域;或者第一电子设备当前自身的地理位置离开了用户常驻区域并且离开的持续时间超过第一阈值;或者第一电子设备当前自身的地理位置和预设位置之间的最短距离大于或等于第二阈值,预设位置包括用户常驻区域;或者第一电子设备当前自身的地理位置进入了特定区域。其中,第一预设条件确定了第一电子设备开始分析视频和/或 图像的时机,当第一电子设备确定自身的地理位置信息满足第一预设条件,则第一电子设备开始对实时拍摄的视频和/或图像进行同步分析。本申请实施例,第一预设条件和第一电子设备的地理位置信息相关,可以适用于用户出门旅行生成旅行视频的应用场景。
基于第一方面,在一种可能的实现方式中,视频和/或图像是通过第二电子设备实时拍摄的。可选的,视频和/或图像可以是通过第一电子设备实时拍摄的;可选的,视频和/或图像可以是通过多个电子设备实时拍摄的。
基于第一方面,在一种可能的实现方式中,第一电子设备选取视频和/或图像中的部分画面,包括:第一电子设备对获取到的视频和/或图像进行分段;第一电子设备基于美学评分算法对分段后的视频和/或图像进行评分,选取分段后每一段视频和/或图像中评分最高的画面,部分画面包括该分段后每一段视频和/或图像中评分最高的画面。
在一种可能的实现方式中,美学评分算法的评分权重与分段后的视频和/或图像的场景类型相关联。美学评分的评分规则可以根据视频画面的场景类型的改变而改变。第一电子设备每划分出一个视频段,则基于当前第一电子设备的地理位置或者图像识别技术,确定该视频段的场景类型,从而确定该场景类型对应的评分规则。这样,根据精确的场景分类,进行定制化的美学评分规则,能够选出最贴近大众审美的画面片段,提升用户体验。
基于第一方面,在一种可能的实现方式中,第二预设条件包括:第一电子设备当前自身的地理位置进入了预设位置;或者第一电子设备当前自身的地理位置距离预设位置的剩余时间小于第五阈值。其中,第二预设条件确定了第一电子设备结束分析视频和/或图像的时机,第一电子设备可以在满足第二预设条件时即生成第一视频,无需等待后台的长时间分析。本申请实施例,第二预设条件和第一电子设备的地理位置信息相关,可以适用于用户出门旅行生成旅行视频的应用场景。
基于第一方面,在一种可能的实现方式中,第一电子设备生成第一视频,之后还包括:第一电子设备输出第一提示信息,第一提示信息用于提供针对第一视频的编辑方式,编辑方式包括播放所述第一视频,保存第一视频,删除第一视频,调整第一视频中显示画面的排序,删除第一视频中的部分显示画面,在第一视频的显示画面之间插入转场动效,在第一视频的显示画面上添加文字、贴纸、水印,添加第一视频的背景音乐中的一项或多项。
基于第一方面,在一种可能的实现方式中,方法还包括:第一电子设备基于第一电子设备的地理位置信息或者第一视频的视频内容,生成第一视频的视频名称。这里,解决了目前图库中相册视频的名称单一的问题,充实了图库为用户记录生活的生态。
基于第一方面,在一种可能的实现方式中,方法还包括:第一电子设备将第一视频发送给第三电子设备,第三电子设备包括第二电子设备。
第二方面,本申请提供了一种视频生成系统,包括第一电子设备和第二电子设备,第一电子设备和第二电子设备建立连接;
第二电子设备,用于拍摄视频和/或图像;
第一电子设备,用于确定自身的地理位置信息满足第一预设条件;
所述第一电子设备,还用于获取所述第二电子设备实时拍摄的视频和/或图像;
所述第一电子设备,还用于选取所述视频和/或图像中的部分画面;
第一电子设备,还用于当第一电子设备的地理位置信息满足第二预设条件,生成第一视频,第一视频是基于所述部分画面拼接而成的。
本申请实施例,第一电子设备和第二电子设备建立连接,第二电子设备可以随时进行视频/图像的拍摄。第一电子设备基于自身的地理位置确定对视频和/或图像的分析开始和分析结 束时间,当第一电子设备的地理位置满足第一预设条件,第一电子设备开始对第二电子设备实时拍摄到的视频和/或图像数据进行分析和剪辑,从而选取出该实时拍摄的视频和/或图像中的部分画面。即第二电子设备一边拍摄视频,第一电子设备一边同步对该视频进行画面选取。当第一电子设备的地理位置满足第二预设条件,第一电子设备生成第一视频,并不再对之后拍摄到的视频和/或图像进行分析和剪辑。该第一视频即为基于上述选取的部分画面拼接得到的。这种同步分析实时拍摄到的视频和/或图像的方式,可以提高视频生成的效率,减少用户等待的时间。
基于第二方面,在一种可能的实现方式中,第一预设条件包括:第一电子设备当前自身的地理位置离开了用户常驻区域;或者第一电子设备当前自身的地理位置离开了用户常驻区域并且离开的持续时间超过第一阈值;或者第一电子设备当前自身的地理位置和预设位置之间的最短距离大于或等于第二阈值,预设位置包括用户常驻区域;或者第一电子设备当前自身的地理位置进入了特定区域。其中,第一预设条件确定了第一电子设备开始分析视频和/或图像的时机,当第一电子设备确定自身的地理位置信息满足第一预设条件,则第一电子设备开始对实时拍摄的视频和/或图像进行同步分析。本申请实施例,第一预设条件和第一电子设备的地理位置信息相关,可以适用于用户出门旅行生成旅行视频的应用场景。
基于第二方面,在一种可能的实现方式中,第一电子设备,还用于选取实时拍摄的视频和/或图像中的部分画面,包括:第一电子设备,还用于对获取到的视频和/或图像进行分段;第一电子设备,还用于基于美学评分算法对分段后的视频和/或图像进行评分,选取分段后每一段视频和/或图像中评分最高的画面,该部分画面包括分段后每一段视频和/或图像中评分最高的画面。
在一种可能的实现方式中,美学评分算法的评分权重与分段后的视频和/或图像的场景类型相关联。美学评分的评分规则可以根据视频画面的场景类型的改变而改变。第一电子设备每划分出一个视频段,则基于当前第一电子设备的地理位置或者图像识别技术,确定该视频段的场景类型,从而确定该场景类型对应的评分规则。这样,根据精确的场景分类,进行定制化的美学评分规则,能够选出最贴近大众审美的画面片段,提升用户体验。
基于第二方面,在一种可能的实现方式中,第二预设条件包括:第一电子设备当前自身的地理位置进入了预设位置;或者第一电子设备当前自身的地理位置距离预设位置的剩余时间小于第五阈值。其中,第二预设条件确定了第一电子设备结束分析视频和/或图像的时机,第一电子设备可以在满足第二预设条件时即生成第一视频,无需等待后台的长时间分析。本申请实施例,第二预设条件和第一电子设备的地理位置信息相关,可以适用于用户出门旅行生成旅行视频的应用场景。
基于第二方面,在一种可能的实现方式中,第一电子设备,还用于在生成第一视频之后,输出第一提示信息,第一提示信息用于提供针对第一视频的编辑方式,编辑方式包括播放第一视频,保存第一视频,删除第一视频,调整第一视频中显示画面的排序,删除第一视频中的部分显示画面,在第一视频的显示画面之间插入转场动效,在第一视频的显示画面上添加文字、贴纸、水印,添加第一视频的背景音乐中的一项或多项。
基于第二方面,在一种可能的实现方式中,第一电子设备,还用于基于第一电子设备的地理位置信息或者第一视频的视频内容,生成第一视频的视频名称。这里,解决了目前图库中相册视频的名称单一的问题,充实了图库为用户记录生活的生态。
基于第二方面,在一种可能的实现方式中,第一电子设备,还用于将第一视频发送给第二电子设备。
基于第二方面,在一种可能的实现方式中,第二电子设备,还用于接收第一视频;第二电子设备,还用于输出第二提示信息,第二提示信息用于提供针对第一视频的编辑方式,编辑方式包括播放第一视频,删除第一视频,调整第一视频中显示画面的排序,删除第一视频中的部分显示画面,在第一视频的显示画面之间插入转场动效,在第一视频的显示画面上添加文字、贴纸、水印,添加第一视频的背景音乐中的一项或多项。
第三方面,本申请提供了一种电子设备,包括:一个或多个处理器、一个或多个存储器;该一个或多个存储与一个或多个处理器耦合;该一个或多个存储器用于存储计算机程序代码,该计算机程序代码包括计算机指令;当该计算机指令在该处理器上运行时,使得该电子设备执行上述任一方面任一种可能的实现方式中的视频生成方法。
第四方面,本申请提供了一种服务器,包括:一个或多个处理器、一个或多个存储器;该一个或多个存储与一个或多个处理器耦合;该一个或多个存储器用于存储计算机程序代码,该计算机程序代码包括计算机指令;当该计算机指令在该处理器上运行时,使得该电子设备执行上述任一方面任一种可能的实现方式中的视频生成方法。
第五方面,本申请实施例提供了一种计算机存储介质,包括计算机指令,当计算机指令在电子设备上运行时,使得通信装置执行上述任一方面任一项可能的实现方式中的视频生成方法。
第六方面,本申请实施例提供了一种计算机程序产品,当计算机程序产品在计算机上运行时,使得计算机执行上述任一方面任一项可能的实现方式中的视频生成方法。
附图说明
图1为本申请实施例提供的一种自动生成图片拼接视频的步骤流程图;
图2为本申请实施例提供的一种视频生成系统的架构示意图;
图3a为本申请实施例提供的一种电子设备的结构示意图;
图3b为本申请实施例提供的一种电子设备的软件结构图;
图4为本申请实施例提供的一种视频生成方法的步骤流程图;
图5a~图5c为本申请实施例提供的一种视频生成方法的应用场景图;
图6为本申请实施例提供的一种视频生成方法的应用场景图;
图7为本申请实施例提供的一种视频生成方法在车机上的应用界面图;
图8为本申请实施例提供的一种视频生成方法在手机上的应用界面图。
具体实施方式
下面将结合附图对本申请实施例中的技术方案进行地描述。其中,在本申请实施例的描述中,除非另有说明,“/”表示或的意思,例如,A/B可以表示A或B;文本中的“和/或”仅仅是一种描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况,另外,在本申请实施例的描述中,“多个”是指两个或多于两个。
以下,术语“第一”、“第二”仅用于描述目的,而不能理解为暗示或暗示相对重要性或者隐含指明所指示的技术特征的数量。由此,限定有“第一”、“第二”的特征可以明示或者隐含地包括一个或者更多个该特征,在本申请实施例的描述中,除非另有说明,“多个”的含义是两个或两个以上。术语“中间”、“左”、“右”、“上”、“下”等指示的方位或位置关系为基于附图所示的方位或位置关系,仅是为了便于描述本申请和简化描述,而不是指示或暗示所指的装置或元件必须具有特定的方位、以特定的方位构造和操作,因此不能理解为对本申请的 限制。
本申请实施例中涉及的电子设备可以是手机、平板电脑、桌面型、膝上型、笔记本电脑、超级移动个人计算机(Ultra-mobile Personal Computer,UMPC)、手持计算机、上网本、个人数字助理(Personal Digital Assistant,PDA)、虚拟现实设备、PDA(Personal Digital Assistant,个人数字助手,又称为掌上电脑)、便携式互联网设备、数据存储设备、可穿戴设备(例如,无线耳机、智能手表、智能手环、智能眼镜、头戴式设备(Head-mounted display,HMD)、电子衣物、电子手镯、电子项链、电子配件、电子纹身和智能镜子)、车辆、车机、车载终端(例如智慧屏、车载摄像头、显示屏)、相机以及带有摄像头的各类电子设备等等。
下面对本申请实施例中涉及到的技术关键词进行介绍。
图像美学评分:基于美学评分算法对图像进行评分。对于图像的美学评分算法,可以使用一些量化的图像美学规则,例如重要特征在所拍摄的图像的画面中的分布。这些美学评分规则可以来自于经验总结,常用的美学评分规则例如包括三分法、主对角线原则、视觉平衡、以及确定被拍摄主体占画面的比例等。基于不同的需求可以有不同的图像美学规则。根据每一种规则可以计算一个美学评分Ei。图像的最终的美学评分可以是根据任意一种规则所计算的美学评分,也可以是根据多个规则所计算的美学评分的加权平均,即E=∑wiEi。其中wi表示权重。具体地,每种规则下的美学评分可以表示为Ei=g(Si,Fi),其中Si描述被拍摄主体的大小和位置,Fi描述主要特征在图像画面中的分布,g为自定义函数(例如高斯函数)。
视频美学评分:同理于图像美学评分,在视频中抽取图像帧,一个视频的美学评分可以是该视频中多个图像帧的美学评分的加权平均。
下面对本申请实施例涉及的应用场景进行举例说明。
手机可以基于相册中保存的图片或视频,对一段时间内或者一个主题下的图片或视频进行剪辑、拼接,自动生成一段新的视频(也可以称为vlog),该新的视频可以是一段时间内的精彩回顾,例如周末的精彩回顾、节假日的精彩回顾等等;也可以是一个主题下的精彩瞬间,例如运动主题下的精彩瞬间、旅游主题下的精彩瞬间、毕业主题下的精彩瞬间等等。
在一些应用场景中,在用户驾驶汽车的过程中,汽车上的其他用户可以拍摄周边环境景色作为记录,若用户想要基于拍摄的图片或视频得到一个处理后的新的视频,需要手动进行剪辑、拼接;或者手机端可以自动生成。目前,手机端自动生成图片拼接视频的过程,是在充电灭屏时对用户相册的图片进行时间地点聚类、场景分类、美学评分后生成的。用户使用手机拍摄照片后,需要充电灭屏等待一段时间,等到对相册中的照片进行时间地点聚类、智慧化分类、生成艺术化封面完成后,可以获得一个由若干照片组成的图片视频。
具体的,如图1所示,图1示出了一种自动生成图片拼接视频的步骤流程。当电子设备检测到自身当前处于充电灭屏状态,电子设备将保存在相册中的照片进行分类,该分类的方式可以是基于时间和/或地点,基于时间分类例如可以是将电子设备在一周内保存的照片划为一类,或者将一个周末内保存的照片划为一类,基于地点分类例如可以是将电子设备在同一个地点保存的照片划为一类,基于时间和地点分类例如可以是将电子设备在一个周末内并且处于同一地点保存的照片划为一类。电子设备将划为一类的多个照片进行图片分析,对该多张照片进行去重以及基于算法进行打分,将重复率高的两张照片筛选掉一张,并且将打分低的照片筛选掉,得到最终的多张照片,将该多张照片进行排序,将多张照片中相同场景的照片排在一起,从而生成最终的图片拼接视频,该视频可以基于时间或场景命名,例如命名为“周末小时光”“周末那些事”等等。
其中,由于特殊场景或特殊时刻强依赖于智慧化能力,并且用户照片多为周末出游,生 成的相册视频多为无场景时刻视频。大量无场景时刻视频存在命名重复、单一的问题,并且上述方式需要在电子设备在充电灭屏的过程中完成,数据量大时,分析时间长,若有大量照片的情况下,用户等待时间久。
本申请实施例提供了一种视频生成方式,电子设备能够同步分析正在实时拍摄的视频,并对该视频进行剪辑,这样,能够解决上述场景中用户等待时间久的问题,在用户驾驶汽车的过程中,汽车上的其他用户可以拍摄周边环境景色作为记录,当记录的视频结束后,用户可以即时获取到剪接后的新的视频。
图2示出了本申请实施例的一种视频生成系统的系统架构,如图2所示,该视频生成系统包括电子设备100、一个或多个电子设备101。电子设备100和电子设备101可以通过有线或无线通信方式连接,无线通信方式例如可以包括蓝牙(blue tooth,BT)、近场通信(near field communication,NFC)、无线保真(wireless fidelity,WiFi)、或WiFi直连。其中,一个或多个电子设备101中包括摄像头、相机、平板、iPad、手机等具有摄像头的电子设备。可选的,电子设备100安装和运行有支持视频生成技术的应用程序,该应用程序可以是摄影类应用程序、视频处理类应用程序。
本申请实施例,电子设备101拍摄视频,电子设备100可以在电子设备101拍摄视频的过程中实时获取到视频数据,电子设备100对实时获取到的视频数据进行分析和剪辑。即电子设备101一边拍摄视频,电子设备100一边同步对该视频进行分析和剪辑,当电子设备101拍摄视频结束,电子设备100可得到剪辑完成的一个新的视频。这样,可以提高视频生成的效率,减少用户等待的时间。示例性的,电子设备100为车机,电子设备101可以为车载摄像头或手机等具有摄像头的设备。当用户驾驶着连接有电子设备101的车辆时,电子设备101可以实时拍摄驾驶过程中路边的风景,那么电子设备100可以对实时拍摄到的视频或图片同步进行分析和剪辑,驾驶结束后电子设备100即可生成一个关于驾驶过程中的旅行视频。并且电子设备100可以将生成的旅行视频分享到其他电子设备(包括电子设备101)。
在一些实施例中,电子设备100和电子设备101可以通过服务器进行数据通信,服务器用于为支持视频生成技术的应用程序提供后台服务。其中,服务器可以是应用服务器、云服务器、后台服务器等等。服务器还可以是云计算平台、虚拟化中心等,包括用于接入广域网的设备,例如基站,宽带远程接入服务器(broadband access serve,BRAS)、光网络单元、宽带网络业务网关(broadband network gateway,BNG),汇聚交换机等。
在一些实施例中,电子设备100和电子设备101为登录了同一账号的设备,或者电子设备100和电子设备101登陆的账户互为关联账户,其中,关联账户可以为被同一账户授权的账户。可选的,这里的账户指的可以是应用账户、设备账户、身份账户等等。
下面首先介绍本申请实施例中涉及的电子设备100。
参见图3a,图3a示出了本申请实施例提供的示例性电子设备100的结构示意图。
电子设备100可以包括处理器110,外部存储器接口120,内部存储器121,通用串行总线(universal serial bus,USB)接口130,充电管理模块140,电源管理模块141,电池142,天线1,天线2,移动通信模块150,无线通信模块160,音频模块170,扬声器170A,受话器170B,麦克风170C,耳机接口170D,传感器模块180,按键190,马达191,指示器192,摄像头193,显示屏194,以及用户标识模块(subscriber identification module,SIM)卡接口195等。其中传感器模块180可以包括压力传感器180A,陀螺仪传感器180B,气压传感器180C,磁传感器180D,加速度传感器180E,距离传感器180F,接近光传感器180G,指纹传感器 180H,温度传感器180J,触摸传感器180K,环境光传感器180L,骨传导传感器180M等。
可以理解的是,本申请实施例示意的结构并不构成对电子设备100的具体限定。在本申请另一些实施例中,电子设备100可以包括比图示更多或更少的部件,或者组合某些部件,或者拆分某些部件,或者不同的部件布置。图示的部件可以以硬件,软件或软件和硬件的组合实现。
处理器110可以包括一个或多个处理单元,例如:处理器110可以包括应用处理器(application processor,AP),调制解调处理器,图形处理器(graphics processing unit,GPU),图像信号处理器(image signal processor,ISP),控制器,存储器,视频编解码器,数字信号处理器(digital signal processor,DSP),基带处理器,和/或神经网络处理器(neural-network processing unit,NPU)等。其中,不同的处理单元可以是独立的器件,也可以集成在一个或多个处理器中。
其中,控制器可以是电子设备100的神经中枢和指挥中心。控制器可以根据指令操作码和时序信号,产生操作控制信号,完成取指令和执行指令的控制。
处理器110中还可以设置存储器,用于存储指令和数据。在一些实施例中,处理器110中的存储器为高速缓冲存储器。该存储器可以保存处理器110刚用过或循环使用的指令或数据。如果处理器110需要再次使用该指令或数据,可从所述存储器中直接调用。避免了重复存取,减少了处理器110的等待时间,因而提高了系统的效率。
在一些实施例中,处理器110可以包括一个或多个接口。接口可以包括集成电路(inter-integrated circuit,I2C)接口,集成电路内置音频(inter-integrated circuit sound,I2S)接口,脉冲编码调制(pulse code modulation,PCM)接口,通用异步收发传输器(universal asynchronous receiver/transmitter,UART)接口,移动产业处理器接口(mobile industry processor interface,MIPI),通用输入输出(general-purpose input/output,GPIO)接口,用户标识模块(subscriber identity module,SIM)接口,和/或通用串行总线(universal serial bus,USB)接口等。
I2C接口是一种双向同步串行总线,包括一根串行数据线(serial data line,SDA)和一根串行时钟线(derail clock line,SCL)。在一些实施例中,处理器110可以包含多组I2C总线。处理器110可以通过不同的I2C总线接口分别耦合触摸传感器180K,充电器,闪光灯,摄像头193等。例如:处理器110可以通过I2C接口耦合触摸传感器180K,使处理器110与触摸传感器180K通过I2C总线接口通信,实现电子设备100的触摸功能。
I2S接口可以用于音频通信。在一些实施例中,处理器110可以包含多组I2S总线。处理器110可以通过I2S总线与音频模块170耦合,实现处理器110与音频模块170之间的通信。在一些实施例中,音频模块170可以通过I2S接口向无线通信模块160传递音频信号,实现通过蓝牙耳机接听电话的功能。
PCM接口也可以用于音频通信,将模拟信号抽样,量化和编码。在一些实施例中,音频模块170与无线通信模块160可以通过PCM总线接口耦合。在一些实施例中,音频模块170也可以通过PCM接口向无线通信模块160传递音频信号,实现通过蓝牙耳机接听电话的功能。所述I2S接口和所述PCM接口都可以用于音频通信。
UART接口是一种通用串行数据总线,用于异步通信。该总线可以为双向通信总线。它将要传输的数据在串行通信与并行通信之间转换。在一些实施例中,UART接口通常被用于连接处理器110与无线通信模块160。例如:处理器110通过UART接口与无线通信模块160中的蓝牙模块通信,实现蓝牙功能。在一些实施例中,音频模块170可以通过UART接口向无线通信模块160传递音频信号,实现通过蓝牙耳机播放音乐的功能。
MIPI接口可以被用于连接处理器110与显示屏194,摄像头193等外围器件。MIPI接口包括摄像头串行接口(camera serial interface,CSI),显示屏串行接口(display serial interface,DSI)等。在一些实施例中,处理器110和摄像头193通过CSI接口通信,实现电子设备100的拍摄功能。处理器110和显示屏194通过DSI接口通信,实现电子设备100的显示功能。
GPIO接口可以通过软件配置。GPIO接口可以被配置为控制信号,也可被配置为数据信号。在一些实施例中,GPIO接口可以用于连接处理器110与摄像头193,显示屏194,无线通信模块160,音频模块170,传感器模块180等。GPIO接口还可以被配置为I2C接口,I2S接口,UART接口,MIPI接口等。
USB接口130是符合USB标准规范的接口,具体可以是Mini USB接口,Micro USB接口,USB Type C接口等。USB接口130可以用于连接充电器为电子设备100充电,也可以用于电子设备100与外围设备之间传输数据。也可以用于连接耳机,通过耳机播放音频。该接口还可以用于连接其他电子设备,例如AR设备等。
可以理解的是,本申请实施例示意的各模块间的接口连接关系,只是示意性说明,并不构成对电子设备100的结构限定。在本申请另一些实施例中,电子设备100也可以采用上述实施例中不同的接口连接方式,或多种接口连接方式的组合。
充电管理模块140用于从充电器接收充电输入。其中,充电器可以是无线充电器,也可以是有线充电器。在一些有线充电的实施例中,充电管理模块140可以通过USB接口130接收有线充电器的充电输入。在一些无线充电的实施例中,充电管理模块140可以通过电子设备100的无线充电线圈接收无线充电输入。充电管理模块140为电池142充电的同时,还可以通过电源管理模块141为电子设备供电。
电源管理模块141用于连接电池142,充电管理模块140与处理器110。电源管理模块141接收电池142和/或充电管理模块140的输入,为处理器110,内部存储器121,外部存储器,显示屏194,摄像头193,和无线通信模块160等供电。电源管理模块141还可以用于监测电池容量,电池循环次数,电池健康状态(漏电,阻抗)等参数。在其他一些实施例中,电源管理模块141也可以设置于处理器110中。在另一些实施例中,电源管理模块141和充电管理模块140也可以设置于同一个器件中。
电子设备100的无线通信功能可以通过天线1,天线2,移动通信模块150,无线通信模块160,调制解调处理器以及基带处理器等实现。
天线1和天线2用于发射和接收电磁波信号。电子设备100中的每个天线可用于覆盖单个或多个通信频带。不同的天线还可以复用,以提高天线的利用率。例如:可以将天线1复用为无线局域网的分集天线。在另外一些实施例中,天线可以和调谐开关结合使用。
移动通信模块150可以提供应用在电子设备100上的包括2G/3G/4G/5G等无线通信的解决方案。移动通信模块150可以包括至少一个滤波器,开关,功率放大器,低噪声放大器(low noise amplifier,LNA)等。移动通信模块150可以由天线1接收电磁波,并对接收的电磁波进行滤波,放大等处理,传送至调制解调处理器进行解调。移动通信模块150还可以对经调制解调处理器调制后的信号放大,经天线1转为电磁波辐射出去。在一些实施例中,移动通信模块150的至少部分功能模块可以被设置于处理器110中。在一些实施例中,移动通信模块150的至少部分功能模块可以与处理器110的至少部分模块被设置在同一个器件中。
调制解调处理器可以包括调制器和解调器。其中,调制器用于将待发送的低频基带信号调制成中高频信号。解调器用于将接收的电磁波信号解调为低频基带信号。随后解调器将解调得到的低频基带信号传送至基带处理器处理。低频基带信号经基带处理器处理后,被传递 给应用处理器。应用处理器通过音频设备(不限于扬声器170A,受话器170B等)输出声音信号,或通过显示屏194显示图像或视频。在一些实施例中,调制解调处理器可以是独立的器件。在另一些实施例中,调制解调处理器可以独立于处理器110,与移动通信模块150或其他功能模块设置在同一个器件中。
无线通信模块160可以提供应用在电子设备100上的包括UWB,无线局域网(wireless local area networks,WLAN)(如无线保真(wireless fidelity,WiFi)网络),蓝牙(bluetooth,BT),全球导航卫星系统(global navigation satellite system,GNSS),调频(frequency modulation,FM),近距离无线通信技术(near field communication,NFC),红外技术(infrared,IR)等无线通信的解决方案。无线通信模块160可以是集成至少一个通信处理模块的一个或多个器件。无线通信模块160经由天线2接收电磁波,将电磁波信号调频以及滤波处理,将处理后的信号发送到处理器110。无线通信模块160还可以从处理器110接收待发送的信号,对其进行调频,放大,经天线2转为电磁波辐射出去。
在一些实施例中,电子设备100的天线1和移动通信模块150耦合,天线2和无线通信模块160耦合,使得电子设备100可以通过无线通信技术与网络以及其他设备通信。所述无线通信技术可以包括全球移动通讯系统(global system for mobile communications,GSM),通用分组无线服务(general packet radio service,GPRS),码分多址接入(code division multiple access,CDMA),宽带码分多址(wideband code division multiple access,WCDMA),时分码分多址(time-division code division multiple access,TD-SCDMA),长期演进(long term evolution,LTE),BT,GNSS,WLAN,NFC,FM,和/或IR技术等。所述GNSS可以包括全球卫星定位系统(global positioning system,GPS),全球导航卫星系统(global navigation satellite system,GLONASS),北斗卫星导航系统(beidou navigation satellite system,BDS),准天顶卫星系统(quasi-zenith satellite system,QZSS)和/或星基增强系统(satellite based augmentation systems,SBAS)。
电子设备100通过GPU,显示屏194,以及应用处理器等实现显示功能。GPU为图像处理的微处理器,连接显示屏194和应用处理器。GPU用于执行数学和几何计算,用于图形渲染。处理器110可包括一个或多个GPU,其执行程序指令以生成或改变显示信息。
显示屏194用于显示图像,视频等。显示屏194包括显示面板。显示面板可以采用液晶显示屏(liquid crystal display,LCD),有机发光二极管(organic light-emitting diode,OLED),有源矩阵有机发光二极体或主动矩阵有机发光二极体(active-matrix organic light emitting diode的,AMOLED),柔性发光二极管(flex light-emitting diode,FLED),Miniled,MicroLed,Micro-oLed,量子点发光二极管(quantum dot light emitting diodes,QLED)等。在一些实施例中,电子设备100可以包括1个或N个显示屏194,N为大于1的正整数。
在本申请的一些实施例中,显示屏194中显示有系统当前输出的界面内容。例如,界面内容为即时通讯应用提供的界面。
电子设备100可以通过ISP,摄像头193,视频编解码器,GPU,显示屏194以及应用处理器等实现拍摄功能。
ISP用于处理摄像头193反馈的数据。例如,拍照时,打开快门,光线通过镜头被传递到摄像头感光元件上,光信号转换为电信号,摄像头感光元件将所述电信号传递给ISP处理,转化为肉眼可见的图像。ISP还可以对图像的噪点,亮度进行算法优化。ISP还可以对拍摄场景的曝光,色温等参数优化。在一些实施例中,ISP可以设置在摄像头193中。
摄像头193用于捕获静态图像或视频。物体通过镜头生成光学图像投射到感光元件。感 光元件可以是电荷耦合器件(charge coupled device,CCD)或互补金属氧化物半导体(complementary metal-oxide-semiconductor,CMOS)光电晶体管。感光元件把光信号转换成电信号,之后将电信号传递给ISP转换成数字图像信号。ISP将数字图像信号输出到DSP加工处理。DSP将数字图像信号转换成标准的RGB,YUV等格式的图像信号。在一些实施例中,电子设备100可以包括1个或N个摄像头193,N为大于1的正整数。
数字信号处理器用于处理数字信号,除了可以处理数字图像信号,还可以处理其他数字信号。例如,当电子设备100在频点选择时,数字信号处理器用于对频点能量进行傅里叶变换等。
视频编解码器用于对数字视频压缩或解压缩。电子设备100可以支持一种或多种视频编解码器。这样,电子设备100可以播放或录制多种编码格式的视频,例如:动态图像专家组(moving picture experts group,MPEG)1,MPEG2,MPEG3,MPEG4等。
NPU为神经网络(neural-network,NN)计算处理器,通过借鉴生物神经网络结构,例如借鉴人脑神经元之间传递模式,对输入信息快速处理,还可以不断的自学习。通过NPU可以实现电子设备100的智能认知等应用,例如:图像识别,人脸识别,语音识别,文本理解等。
外部存储器接口120可以用于连接外部存储卡,例如Micro SD卡,实现扩展电子设备100的存储能力。外部存储卡通过外部存储器接口120与处理器110通信,实现数据存储功能。例如将音乐,视频等文件保存在外部存储卡中。
内部存储器121可以用于存储计算机可执行程序代码,所述可执行程序代码包括指令。处理器110通过运行存储在内部存储器121的指令,从而执行电子设备100的各种功能应用以及数据处理。内部存储器121可以包括存储程序区和存储数据区。其中,存储程序区可存储操作系统,至少一个功能所需的应用程序(比如声音播放功能,图像播放功能等)等。存储数据区可存储电子设备100使用过程中所创建的数据(比如音频数据,电话本等)等。此外,内部存储器121可以包括高速随机存取存储器,还可以包括非易失性存储器,例如至少一个磁盘存储器件,闪存器件,通用闪存存储器(universal flash storage,UFS)等。
电子设备100可以通过音频模块170,扬声器170A,受话器170B,麦克风170C,耳机接口170D,以及应用处理器等实现音频功能。例如音乐播放,录音等。
音频模块170用于将数字音频信息转换成模拟音频信号输出,也用于将模拟音频输入转换为数字音频信号。音频模块170还可以用于对音频信号编码和解码。在一些实施例中,音频模块170可以设置于处理器110中,或将音频模块170的部分功能模块设置于处理器110中。
扬声器170A,也称“喇叭”,用于将音频电信号转换为声音信号。电子设备100可以通过扬声器170A收听音乐,或收听免提通话。
受话器170B,也称“听筒”,用于将音频电信号转换成声音信号。当电子设备100接听电话或语音信息时,可以通过将受话器170B靠近人耳接听语音。
麦克风170C,也称“话筒”,“传声器”,用于将声音信号转换为电信号。当拨打电话或发送语音信息时,用户可以通过人嘴靠近麦克风170C发声,将声音信号输入到麦克风170C。电子设备100可以设置至少一个麦克风170C。在另一些实施例中,电子设备100可以设置两个麦克风170C,除了采集声音信号,还可以实现降噪功能。在另一些实施例中,电子设备100还可以设置三个,四个或更多麦克风170C,实现采集声音信号,降噪,还可以识别声音来源,实现定向录音功能等。
耳机接口170D用于连接有线耳机。耳机接口170D可以是USB接口130,也可以是3.5mm 的开放移动电子设备平台(open mobile terminal platform,OMTP)标准接口,美国蜂窝电信工业协会(cellular telecommunications industry association of the USA,CTIA)标准接口。
压力传感器180A用于感受压力信号,可以将压力信号转换成电信号。在一些实施例中,压力传感器180A可以设置于显示屏194。在本申请一些可选的实施例中,压力传感器180A可用于捕获用户手指部位接触显示屏时生成的压力值,并将该压力值传输给处理器,以使得处理器识别用户通过哪个手指部位输入用户操作。
压力传感器180A的种类很多,如电阻式压力传感器,电感式压力传感器,电容式压力传感器等。电容式压力传感器可以是包括至少两个具有导电材料的平行板。当有力作用于压力传感器180A,电极之间的电容改变。电子设备100根据电容的变化确定压力的强度。当有触摸操作作用于显示屏194,电子设备100根据压力传感器180A检测所述触摸操作强度。电子设备100也可以根据压力传感器180A的检测信号计算触摸的位置。在一些实施例中,作用于不同触摸位置,可以对应不同的操作指令。在一些可选的实施例中,压力传感器180A还可根据检测到的信号计算触摸点的数量,并将计算值传输给处理器,以使得处理器识别用户通过单指或多指输入用户操作。
陀螺仪传感器180B可以用于确定电子设备100的运动姿态。在一些实施例中,可以通过陀螺仪传感器180B确定电子设备100围绕三个轴(电子设备的X轴、Y轴和Z轴)的角速度。陀螺仪传感器180B可以用于拍摄防抖。示例性的,当按下快门,陀螺仪传感器180B检测电子设备100抖动的角度,根据角度计算出镜头模组需要补偿的距离,让镜头通过反向运动抵消电子设备100的抖动,实现防抖。陀螺仪传感器180B还可以用于导航,体感游戏场景。
气压传感器180C用于测量气压。在一些实施例中,电子设备100通过气压传感器180C测得的气压值计算海拔高度,辅助定位和导航。
磁传感器180D包括霍尔传感器。电子设备100可以利用磁传感器180D检测翻盖皮套的开合。在一些实施例中,当电子设备100是翻盖机时,电子设备100可以根据磁传感器180D检测翻盖的开合。进而根据检测到的皮套的开合状态或翻盖的开合状态,设置翻盖自动解锁等特性。
加速度传感器180E可检测电子设备100在各个方向上(一般为三轴)加速度的大小。当电子设备100静止时可检测出重力的大小及方向。还可以用于识别电子设备姿态,应用于横竖屏切换,计步器等应用。在本申请一些可选的实施例中,加速度传感器180E可用于捕获用户手指部位接触显示屏(或者用户手指敲击电子设备100的后壳后侧边框)时生成的加速度值,并将该加速度值传输给处理器,以使得处理器识别用户通过哪个手指部位输入用户操作。
距离传感器180F,用于测量距离。电子设备100可以通过红外或激光测量距离。在一些实施例中,拍摄场景,电子设备100可以利用距离传感器180F测距以实现快速对焦。
接近光传感器180G可以包括例如发光二极管(LED)和光检测器,例如光电二极管。发光二极管可以是红外发光二极管。电子设备100通过发光二极管向外发射红外光。电子设备100使用光电二极管检测来自附近物体的红外反射光。当检测到充分的反射光时,可以确定电子设备100附近有物体。当检测到不充分的反射光时,电子设备100可以确定电子设备100附近没有物体。电子设备100可以利用接近光传感器180G检测用户手持电子设备100贴近耳朵通话,以便自动熄灭显示屏达到省电的目的。接近光传感器180G也可用于皮套模式,口袋模式自动解锁与锁屏。
环境光传感器180L用于感知环境光亮度。电子设备100可以根据感知的环境光亮度自适应调节显示屏194亮度。环境光传感器180L也可用于拍照时自动调节白平衡。环境光传感器 180L还可以与接近光传感器180G配合,检测电子设备100是否在口袋里,以防误触。
指纹传感器180H用于采集指纹。电子设备100可以利用采集的指纹特性实现指纹解锁,访问应用锁,指纹拍照,指纹接听来电等。
温度传感器180J用于检测温度。在一些实施例中,电子设备100利用温度传感器180J检测的温度,执行温度处理策略。例如,当温度传感器180J上报的温度超过阈值,电子设备100执行降低位于温度传感器180J附近的处理器的性能,以便降低功耗实施热保护。在另一些实施例中,当温度低于另一阈值时,电子设备100对电池142加热,以避免低温导致电子设备100异常关机。在其他一些实施例中,当温度低于又一阈值时,电子设备100对电池142的输出电压执行升压,以避免低温导致的异常关机。
触摸传感器180K,也称“触控面板”。触摸传感器180K可以设置于显示屏194,由触摸传感器180K与显示屏194组成触摸屏,也称“触控屏”。触摸传感器180K用于检测作用于其上或附近的触摸操作,该触摸触控操作是指用户手部、手肘、触控笔等接触显示屏194的操作。触摸传感器可以将检测到的触摸操作传递给应用处理器,以确定触摸事件类型。可以通过显示屏194提供与触摸操作相关的视觉输出。在另一些实施例中,触摸传感器180K也可以设置于电子设备100的表面,与显示屏194所处的位置不同。
骨传导传感器180M可以获取振动信号。在一些实施例中,骨传导传感器180M可以获取人体声部振动骨块的振动信号。骨传导传感器180M也可以接触人体脉搏,接收血压跳动信号。在一些实施例中,骨传导传感器180M也可以设置于耳机中,结合成骨传导耳机。音频模块170可以基于所述骨传导传感器180M获取的声部振动骨块的振动信号,解析出语音信号,实现语音功能。应用处理器可以基于所述骨传导传感器180M获取的血压跳动信号解析心率信息,实现心率检测功能。
按键190包括开机键,音量键等。按键190可以是机械按键。也可以是触摸式按键。电子设备100可以接收按键输入,产生与电子设备100的用户设置以及功能控制有关的键信号输入。
马达191可以产生振动提示。马达191可以用于来电振动提示,也可以用于触摸振动反馈。例如,作用于不同应用(例如拍照,音频播放等)的触摸操作,可以对应不同的振动反馈效果。作用于显示屏194不同区域的触摸操作,马达191也可对应不同的振动反馈效果。不同的应用场景(例如:时间提醒,接收信息,闹钟,游戏等)也可以对应不同的振动反馈效果。触摸振动反馈效果还可以支持自定义。
指示器192可以是指示灯,可以用于指示充电状态,电量变化,也可以用于指示消息,未接来电,通知等。
SIM卡接口195用于连接SIM卡。SIM卡可以通过插入SIM卡接口195,或从SIM卡接口195拔出,实现和电子设备100的接触和分离。
电子设备100和电子设备101的软件系统可以采用分层架构,事件驱动架构,微核架构,微服务架构,或云架构。本申请实施例以Android系统为例,示例性说明电子设备100的软件架构。其中,Android系统仅为本申请实施例中电子设备100或电子设备101的一种系统实例,本申请还可以适用于其他类型的操作系统,比如ios、Windows、鸿蒙等,本申请对此不加以限制。下述仅将Android系统作为电子设备100的操作系统的示例。
图3b是本发明实施例的电子设备100的软件结构框图。
分层架构将软件分成若干个层,每一层都有清晰的角色和分工。层与层之间通过软件接口通信。在一些实施例中,将Android系统分为四层,从上至下分别为应用程序层,应用程 序框架层,安卓运行时(Android runtime)和系统库,以及内核层。
应用程序层可以包括一系列应用程序包。
如图3b所示,应用程序包可以包括相机,图库,日历,通话,地图,导航,WLAN,蓝牙,音乐,视频,短信息等应用程序。
应用程序框架层为应用程序层的应用程序提供应用编程接口(application programming interface,API)和编程框架。应用程序框架层包括一些预先定义的函数。
如图3b所示,应用程序框架层可以包括窗口管理器,内容提供器,视图系统,电话管理器,资源管理器,通知管理器等。
窗口管理器用于管理窗口程序。窗口管理器可以获取显示屏大小,判断是否有状态栏,锁定屏幕,截取屏幕等。
内容提供器用来存放和获取数据,并使这些数据可以被应用程序访问。所述数据可以包括视频,图像,音频,拨打和接听的电话,浏览历史和书签,电话簿等。
视图系统包括可视控件,例如显示文字的控件,显示图片的控件等。视图系统可用于构建应用程序。显示界面可以由一个或多个视图组成的。例如,包括短信通知图标的显示界面,可以包括显示文字的视图以及显示图片的视图。
电话管理器用于提供电子设备100的通信功能。例如通话状态的管理(包括接通,挂断等)。
资源管理器为应用程序提供各种资源,比如本地化字符串,图标,图片,布局文件,视频文件等等。
通知管理器使应用程序可以在状态栏中显示通知信息,可以用于传达告知类型的消息,可以短暂停留后自动消失,无需用户交互。比如通知管理器被用于告知下载完成,消息提醒等。通知管理器还可以是以图表或者滚动条文本形式出现在系统顶部状态栏的通知,例如后台运行的应用程序的通知,还可以是以对话窗口形式出现在屏幕上的通知。例如在状态栏提示文本信息,发出提示音,电子设备振动,指示灯闪烁等。
Android Runtime包括核心库和虚拟机。Android runtime负责安卓系统的调度和管理。
核心库包含两部分:一部分是java语言需要调用的功能函数,另一部分是安卓的核心库。
应用程序层和应用程序框架层运行在虚拟机中。虚拟机将应用程序层和应用程序框架层的java文件执行为二进制文件。虚拟机用于执行对象生命周期的管理,堆栈管理,线程管理,安全和异常的管理,以及垃圾回收等功能。
系统库可以包括多个功能模块。例如:表面管理器(surface manager),媒体库(Media Libraries),三维图形处理库(例如:OpenGL ES),2D图形引擎(例如:SGL)等。
表面管理器用于对显示子系统进行管理,并且为多个应用程序提供了2D和3D图层的融合。
媒体库支持多种常用的音频,视频格式回放和录制,以及静态图像文件等。媒体库可以支持多种音视频编码格式,例如:MPEG4,H.264,MP3,AAC,AMR,JPG,PNG等。
三维图形处理库用于实现三维图形绘图,图像渲染,合成,和图层处理等。
2D图形引擎是2D绘图的绘图引擎。
内核层是硬件和软件之间的层。内核层至少包含显示驱动,摄像头驱动,音频驱动,传感器驱动。
可以理解的是,本申请实施例示意的结构并不构成对电子设备100的具体限定。在本申请另一些实施例中,电子设备100可以包括比图示更多或更少的部件,或者组合某些部件,或者拆分某些部件,或者不同的部件布置。图示的部件可以以硬件,软件或软件和硬件的组 合实现。
基于上述系统架构和软硬件结构,下面详细介绍本申请提供的一种视频生成方法的步骤流程,如图4所示,图4示例性的示出了一种视频生成方法的步骤流程图,可包括:
S101:电子设备100和电子设备101建立连接。
电子设备100和电子设备101可以通过有线或无线通信方式连接,无线通信方式例如可以包括蓝牙、近场通信(near field communication,NFC)、无线保真(wireless fidelity,WiFi)、或WiFi直连。其中,电子设备101包括摄像头、相机、平板、iPad、手机等具有摄像头的电子设备。本申请实施例,电子设备100也可以称为第一电子设备,电子设备101也可以称为第二电子设备。
在一些实施例中,电子设备100和电子设备101可以通过服务器进行数据通信,服务器可以是应用服务器、云服务器、后台服务器等等。
在一些实施例中,电子设备100为车辆,则电子设备101可以为手机、相机等;或者电子设备100为手机,则电子设备101可以为车载摄像头、其他手机等。可选的,电子设备101可以搭载在电子设备100上,例如电子设备100为车辆,电子设备101为车载摄像头;又例如电子设备100为手机,电子设备101为手机上的摄像头。
在一些实施例中,电子设备100和电子设备101启动,电子设备101可以拍摄图片和/或视频,电子设备100可以同步获取电子设备101拍摄到的图片和/或视频数据。
在一些实施例中,电子设备100和多个电子设备101建立连接,电子设备100可以同步获取到多个电子设备101拍摄的图片或视频数据。例如电子设备100为车辆,电子设备101为车载摄像头和手机上的摄像头。
S102:电子设备100确定电子设备100的地理位置信息满足第一预设条件。
电子设备100获取到自身的地理位置信息,其中地理位置信息可以包括通过GPS系统确定的绝对地理位置信息,例如经纬度;也可以包括通过导航系统定位出的地理位置区域,例如某区某街道等位置;也可以包括电子设备100距离初始启动位置的相对位置;也可以包括电子设备100距离某一预设位置的相对位置。
在一些实施例中,第一预设条件为电子设备100当前自身的地理位置离开了用户常驻区域。其中用户常驻区域可以包括一个或多个区域,用户常驻区域可以是以城市为最小区域,可以是以地区为最小区域,可以是以街道为最小区域等等。可选的,用户常驻区域是电子设备100预设的,用户可以对用户常驻地区进行添加或删除。可选的,用户常驻区域是电子设备100基于自身的历史移动轨迹统计出来的,电子设备100可以对用户常驻区域进行更新(添加或删除)。
进一步的,第一预设条件为电子设备100当前自身的地理位置离开了用户常驻区域并且离开的持续时间超过第一阈值。
在一些实施例中,第一预设条件为电子设备100当前自身的地理位置和电子设备100的预设位置之间的最短距离大于或等于第二阈值。示例性的,令第二阈值为30公里,电子设备100的预设位置为A点,当电子设备100当前的地理位置为B点,且B点和A点的距离为30公里或大于30公里,则此时电子设备当前自身的地理位置(B点)满足第一预设条件。可选的,该预设位置可以是用户常驻区域。
在一些实施例中,第一预设条件为电子设备100当前自身的地理位置进入了特定区域,例如风景区、旅游景点等等。可选的,该特定区域是实时更新的,例如基于网络添加新的风景区;可选的,该特定区域是用户预设的。
在一些实施例中,第一预设条件为电子设备100当前自身的地理位置和电子设备100的初始启动位置的距离大于或等于第三阈值。示例性的,令第三阈值为30公里,电子设备100从A点开始移动,当电子设备100移动到了B点,且B点和A点的距离为30公里或大于30公里,则此时电子设备当前自身的地理位置(B点)满足第一预设条件。
在一些实施例中,第一预设条件为电子设备100检测到通过高速收费站进入了高速。可选的,电子设备100基于导航软件检测到电子设备100当前自身的地理位置通过高速收费站。可选的,电子设备100基于传感器检测到电子设备100通过高速收费站。
在一些实施例中,第一预设条件为电子设备100接收到了用户确认指令。当电子设备100接收到用户确认指令,则电子设备100满足第一预设条件。
S103:电子设备100对电子设备101实时拍摄的视频进行分析。
当电子设备100确定自身的地理位置信息满足第一预设条件,电子设备100对电子设备101实时拍摄的视频进行分析。
在一些应用场景中,以电子设备100为车机为例,电子设备101为与车机连接的车载摄像头,当车辆启动时,车辆电源向车机和车载摄像头进行供电,车载摄像头可以实时拍摄视频进行记录。当车机确定当前自身的地理位置信息满足第一预设条件,此时,车机获取车载摄像头当前实时拍摄的视频进行分析。
在一些实施例中,当电子设备100确定自身的地理位置信息满足第一预设条件,电子设备101开始拍摄视频,电子设备100对电子设备101实时拍摄的视频进行分析。
在一些实施例中,电子设备101正在拍摄视频,当电子设备100确定自身的地理位置信息满足第一预设条件,电子设备100对电子设备101实时拍摄的视频进行分析。若电子设备101在拍摄了5分钟之后,电子设备100检测出此时自身的地理位置信息满足第一预设条件,则电子设备100对当下电子设备101拍摄的视频进行分析,即对5分钟之后的视频进行分析。示例性的,电子设备100为车机,电子设备101为车载摄像头。当车辆启动,车机和车载摄像头得到供电,车载摄像头开始进行拍摄视频。当车机确定自身的地理位置信息满足第一预设条件,车机再同步对车载摄像头实时拍摄的视频进行分析。
下面详细介绍电子设备100对电子设备101实时拍摄的视频进行分析的方法和原理。
电子设备100同步获取到电子设备101实时拍摄的视频,对获取到的视频进行画面选取,将选取到的画面作为最终生成的新视频中的一段内容。
如图5a所示,电子设备100同步获取到电子设备101拍摄到的视频画面,电子设备100对获取到的视频画面进行抽帧并基于美学评分算法进行评分,当检测到视频画面中的场景类型改变了,电子设备100确定上个场景中评分最高的视频段。然后继续对下个场景进行抽帧评分,直至结束。
本申请实施例,电子设备100通过滑动窗口实时选出评分最高的视频画面,这样电子设备100可以在满足第二预设条件时即生成第一视频,无需等待后台的长时间分析。并且,当电子设备100为车机的情况下,车机对功耗的要求更低,平均每帧分析消耗800-1000ms,节约资源。
在一些实施例中,电子设备100对获取到的视频数据进行分段,然后按照预设方式在分段后的每一段视频中选取部分画面。
进一步的,电子设备100对获取到的视频数据进行分段的分段方式可以是基于场景镜头进行分段,场景镜头包括人物镜头、风景镜头、转场镜头等等,如图5b所示,图5b示例性示出了电子设备101实时拍摄到的视频,电子设备100同步获取到电子设备101拍摄到的视 频,电子设备100基于图像识别技术对视频中的图像帧进行识别,从而划分出不同分类的镜头,图5b中示出了三类场景镜头,每类场景镜头对应一段视频,时间点1之前的视频中的图像帧分类为场景镜头1;时间点1和时间点2之间的视频中的图像帧分类为场景镜头2;时间点2之后的视频中的图像帧分类为场景镜头3。
电子设备100每划分出一段视频,则对该段视频进行画面选取,将选取到的画面作为最终生成的新视频中的一段内容。在一些实施例中,电子设备100可以基于地理位置信息或者视频内容,对选取出的视频画面生成对应的名称或简要描述,提升用户对于生成的新视频的观看体验。
可选的,画面选取的方式可以是将该段视频均分为若干段视频,然后对该若干段视频中每一段视频进行美学评分,选取出评分最高的一段视频作为最终生成的新视频中的一段内容。
可选的,画面选取的方式可以是将该段视频进行场景细化,再次进行场景分类,对该段视频中每个场景进行美学评分,选取出评分最高的一段场景作为最终生成的新视频中的一段内容。
可选的,画面选取的方式可以是将该段视频进行等间隔抽帧,如图5c所示,图5c中以场景镜头1为例,示例性示出了一个场景镜头进行画面选取的方式。图5c中,电子设备100在场景镜头1中抽出了9个图像帧,该9个图像帧把场景镜头1划分为了8段视频画面。电子设备100对该9个图像帧进行美学评分,确定出每个图像帧对应的评分,从而确定出每段视频画面的评分,可选的,每段视频画面的评分可以是两个图像帧的平均评分。例如视频画面1的评分为图像帧1和图像帧2的平均评分,视频画面2的评分为图像帧2和图像帧3的平均评分,以此类推。电子设备100选取出评分最高的一段视频画面作为最终生成的新视频中的一段内容。
可选的,电子设备100选取评分大于第四阈值的视频画面作为最终生成的新视频中的内容。这样可能会出现在一段场景镜头中没有选取到任何画面的情况,也可能会出现在一段场景镜头中选取了大量画面的情况。
在一些实施例中,美学评分的评分规则可以根据视频画面的场景类型的改变而改变。电子设备100每划分出一个视频段,则基于当前电子设备100的地理位置或者图像识别技术,确定该视频段的场景类型,从而确定该场景类型对应的评分规则。举例来说,用户在S101省道,该地地貌特征为丹霞地貌,美学评分规则中可以提高或加入色彩鲜艳程度和地形层次的权重;在千岛湖环湖公路,多为侵蚀剥蚀低山丘陵的库区地貌,美学评分规则中可以提高或加入水、岛、天空的占比的权重;在城市内部驾驶,如上海延安路高架,美学评分规则中可以提高或加入灯光鲜明程度,天空与高楼的占比的权重;等等。这样,根据精确的场景分类,进行定制化的美学评分规则,能够选出最贴近大众审美的画面片段,提升用户体验。
上述电子设备100对电子设备101实时拍摄的视频进行分析的过程,是随着电子设备101拍摄到的视频画面同步进行的,电子设备100同步接收到电子设备101拍摄的视频画面,对接收到的视频画面进行分段,每划分一段则截取一部分画面作为最终生成的新视频中的内容,直至电子设备100的地理位置信息满足第二预设条件,电子设备100生成第一视频。
S104:确定电子设备100的地理位置信息满足第二预设条件。
在一些实施例中,第二预设条件为电子设备100当前自身的地理位置到达了目的地。可选的,目的地确定可以是基于导航软件确定的。其中,目的地也可以称为预设位置。
可选的,第二预设条件为电子设备100当前自身的地理位置到达目的地的剩余时间小于第五阈值。可选的,第五阈值可以是导航软件确定的。
在一些实施例中,第二预设条件为电子设备100检测到在预设时间内自身的地理位置不变或者变化在小范围内。
在一些实施例中,第二预设条件为电子设备100检测到自身通过高速收费站离开了高速。可选的,电子设备100基于导航软件检测到电子设备100当前自身的地理位置离开高速收费站。可选的,电子设备100基于传感器检测到电子设备100离开高速收费站。
在一些实施例中,第二预设条件为电子设备100接收到了用户确认指令。当电子设备100接收到用户确认指令,则电子设备100满足第二预设条件。
S105:电子设备100生成第一视频,该第一视频为基于电子设备100获取到电子设备101拍摄的视频进行分析并剪辑得到的。
当电子设备100确定电子设备100的地理位置信息满足第二预设条件后,电子设备100不再分析电子设备101之后拍摄到的视频内容,基于之前获取到的视频内容电子设备100生成第一视频(新视频)。该第一视频为基于电子设备100获取到的视频内容进行分析并剪辑得到的。
在一些实施例中,电子设备100同步接收到电子设备101拍摄的视频画面,对接收到的视频画面进行分段,每划分一段则截取一部分画面作为最终生成的新视频中的内容,直至电子设备100的地理位置信息满足第二预设条件,电子设备100生成新视频。
示例性的,如图6所示,电子设备100确定电子设备100的地理位置信息满足第一预设条件后,电子设备100同步接收电子设备101拍摄的视频画面,划分出场景镜头1,选取出场景镜头1中评分最高的视频画面1;电子设备100持续接收到电子设备100拍摄的视频画面,划分出场景镜头12,选取出场景镜头2中评分最高的视频画面24,此时可以将视频画面1和视频画面2拼接到一起;……;电子设备100持续接收到电子设备100拍摄的视频画面,划分出场景镜头n-1,选取出场景镜头n-1中评分最高的视频画面a,此时可以将视频画面a和场景镜头n-2中选取的视频画面拼接到一起;电子设备100持续接收到电子设备100拍摄的视频画面,电子设备100确定电子设备100的地理位置信息满足第二预设条件后,电子设备100划分出最后一段场景镜头n,选取出场景镜头n中评分最高的视频画面m,此时可以将视频画面m和视频画面a拼接到一起;这时,电子设备100生成第一视频,该第一视频是由n个视频画面拼接而成的。可选的,第一视频中每个视频画面之间可以插入一个转场动效或转场画面。可选的,每个场景镜头中可以选取一个以上的视频画面,也可以不选取视频画面。
在一些实施例中,电子设备100生成第一视频,并保存该第一视频。用户可以在电子设备100上查看并播放该第一视频。
在一些实施例中,电子设备100生成第一视频,并基于电子设备100的地理位置信息或者该第一视频的视频内容,生成针对于该第一视频的视频名称,例如云南洱海之旅,千岛湖环湖之行等丰富多彩的定制名称,解决了目前图库中相册视频的名称单一的问题,充实了图库为用户记录生活的生态。
在一些实施例中,电子设备100生成第一视频后,电子设备100输出提示信息,该提示信息用于提示用户第一视频已经生成并且提供对第一视频的处理方式,包括保存第一视频、删除第一视频、编辑第一视频、分享到应用、发送到其他设备等等。示例性的,如图7所示,图7以电子设备100为车机为例,示出了电子设备100生成第一视频后的在一种导航应用中的应用界面701,应用界面701包括导航地图702,工具栏703以及第一视频生成的提示信息711,其中,
工具栏703包括一个或多个功能控件,提供了例如打电话、播放音乐等功能。
提示信息711提示了用户第一视频已经生成(图7中示出的为“本次驾驶途中的vlog已经生成”),在提示信息711中包括第一视频的缩略图712,播放控件713,保存控件714,删除控件715,编辑控件716,分享控件717,继续剪辑控件718和重新开始控件719。
播放控件713用于播放第一视频。
保存控件714用于将第一视频保存至电子设备100。
删除控件715用于删除第一视频。
编辑控件716用于编辑第一视频,包括例如改变第一视频中视频画面的顺序,删除第一视频中的部分画面,在两个视频画面之间插入转场动效,添加文字、贴纸、水印等,添加背景音乐,编辑第一视频的视频名称,等等。
分享控件717用于将第一视频发送到其他设备,或者分享到应用程序。
继续剪辑控件718用于指示电子设备100继续接收并分析电子设备101实时拍摄的视频画面,重新生成第一视频。其中,继续剪辑控件可以应用于电子设备100错误判断电子设备100满足第二预设条件的情况,此时用户可以选择电子设备100继续对电子设备101实时拍摄的视频画面进行分析和剪辑;或者用户想要将两次驾驶的视频内容剪辑到一个vlog中,用于可以选择电子设备100继续对电子设备101实时拍摄的视频画面进行分析和剪辑。
重新开始控件719用于指示电子设备100将第一视频删除,并重新接收并分析电子设备101实时拍摄的视频画面。
可选的,生成第一视频之后,电子设备100将该第一视频发送到其他电子设备。
当电子设备100确定电子设备100的地理位置信息满足第二预设条件后,电子设备100生成第一视频并将该第一视频发送给其他电子设备。该其他电子设备可以包括电子设备101,也可以是其他与电子设备100建立了连接的电子设备。
在一些实施例中,其他电子设备接收到该第一视频,其他电子设备输出提示信息,该提示信息用于提示用户接收到第一视频并且提供对第一视频的处理方式,包括确认接收第一视频、保存第一视频、删除第一视频、编辑第一视频、分享到应用等等。示例性的,如图8所示,以接收到第一视频的其他电子设备为手机为例,图8示例性的示出了一种锁屏界面801,手机在锁屏状态下接收到电子设备100发送的第一视频,手机输出提示信息,图8中示出的锁屏界面801包括状态栏802和通知803(提示信息),其中,
状态栏802可包括:移动通信信号(又可称为蜂窝信号)的一个或多个信号强度指示符804、无线高保真(wireless fidelity,Wi-Fi)信号的一个或多个信号强度指示符805,蓝牙指示符806,电池状态指示符807。当电子设备的蓝牙模块为开启状态(即电子设备为蓝牙模块进行供电)时,电子设备的显示界面上显示蓝牙指示符806。
通知803提示了用户已经接收到第一视频(图8中示出的为“接收到本次驾驶途中的vlog,点击查看”),在提示信息通知803中包括播放控件808,删除控件809,编辑控件810,分享控件811。
播放控件808用于播放接收到的第一视频。
删除控件809用于删除第一视频。
编辑控件810用于编辑第一视频,包括例如改变第一视频中视频画面的顺序,删除第一视频中的部分画面,在两个视频画面之间插入转场动效,添加文字、贴纸、水印等,添加背景音乐,等等。
分享控件811用于将第一视频发送到其他电子设备,或者分享到应用程序。
本申请实施例,电子设备100基于自身的地理位置确定对视频画面的分析开始和分析结束时间,当电子设备100的地理位置满足第一预设条件,电子设备100开始对电子设备101实时拍摄到的视频数据进行分析和剪辑。即电子设备101一边拍摄视频,电子设备100一边同步对该视频进行分析和剪辑。当电子设备100的地理位置满足第二预设条件,电子设备100生成第一视频,并不再对电子设备101之后拍摄到的视频数据进行分析和剪辑。该第一视频即为基于电子设备100获取到的视频内容进行分析并剪辑得到的。这样,电子设备100同步分析实时拍摄到的视频内容的方式,可以提高视频生成的效率,减少用户等待的时间。
示例性的,电子设备100为车机,电子设备101可以为车载摄像头或手机等具有摄像头的设备。当用户驾驶着搭载有电子设备100的车辆上时,电子设备101可以实时拍摄驾驶过程中路边的风景,那么电子设备100可以对实时拍摄到的视频数据或图片同步进行分析和剪辑,驾驶结束后电子设备100即可生成一个关于驾驶过程中的旅行视频。电子设备100可以将该视频发送到其他电子设备。
在一些应用场景中,电子设备100为车机,电子设备101可以为手机上的摄像头,手机可以实时拍摄驾驶过程中路边的风景,驾驶结束后电子设备100即可生成一个关于驾驶过程中的旅行视频。电子设备100可以将该视频发送到其他电子设备,例如手机上。
在一些应用场景中,电子设备100为手机,电子设备101可以为车载摄像头,车载摄像头可以实时拍摄驾驶过程中路边的风景,驾驶结束后手机即可生成一个关于驾驶过程中的旅行视频。
在一些实施例中,电子设备101包括多个电子设备。例如,电子设备100为车机,电子设备101可以为车载摄像头和手机上的摄像头,即电子设备100可以接收到多个电子设备101实时拍摄到的视频数据。可选的,电子设备100可以基于在同一段时间内获取到的多段视频数据进行分析,选取中多段视频中的多段视频内容拼接到一起。
上述实施例中均以电子设备101拍摄的是视频为例,不限于视频数据,电子设备101还可以拍摄图像,电子设备100可以同步获取到电子设备101实时拍摄到的图像数据。在一些实施例中,电子设备100获取电子设备101实时拍摄的图像数据。当电子设备100的地理位置满足第一预设条件,电子设备100开始对电子设备101实时拍摄到的图像数据进行分析和剪辑。即电子设备101一边拍摄图像,电子设备100一边同步对接收到的图像进行分析和剪辑。
示例性的,电子设备100对拍摄得到的一系列图像进行场景分类,在每个场景下的图像中选取部分图像作为最终生成的视频中的部分内容,选取方式包括例如基于美学评分算法筛选出评分较高的照片。进一步的,在进行美学评分之前,电子设备100基于图像识别技术筛选出重复率较高的照片,选取其中的一张或多张(例如为美学评分最高的一张)保留。
示例性的,电子设备100先确定需要的图像张数n张,电子设备100对接收到的每张图像进行美学评分,排名前n张的图像作为最终生成的视频中的内容。
当电子设备100的地理位置满足第二预设条件,电子设备100生成第一视频,并不再对电子设备101之后拍摄到的图像数据进行分析和剪辑。该第一视频即为基于电子设备100获取到的图像进行分析并剪辑得到的。这样,电子设备100同步分析实时拍摄到的图像的方式,可以提高视频生成的效率,减少用户等待的时间。
在一些实施例中,电子设备100可以获取电子设备101发送的图像数据和视频数据。当 电子设备100的地理位置满足第一预设条件,电子设备100开始对电子设备101实时拍摄到的图像数据和视频数据进行分析和剪辑。即电子设备101一边拍摄图像和视频,电子设备100一边同步对接收到的图像数据和视频数据进行分析和剪辑。当电子设备100的地理位置满足第二预设条件,电子设备100生成第一视频,并不再对电子设备101之后拍摄到的图像数据和视频数据进行分析和剪辑。
在上述实施例中,可以全部或部分地通过软件、硬件、固件或者其任意组合来实现。当使用软件实现时,可以全部或部分地以计算机程序产品的形式实现。所述计算机程序产品包括一个或多个计算机指令。在计算机上加载和执行所述计算机程序指令时,全部或部分地产生按照本申请实施例所述的流程或功能。所述计算机可以是通用计算机、专用计算机、计算机网络、或者其他可编程装置。所述计算机指令可以存储在计算机可读存储介质中,或者从一个计算机可读存储介质向另一个计算机可读存储介质传输,例如,所述计算机指令可以从一个网站站点、计算机、服务器或数据中心通过有线(例如同轴电缆、光纤、数字用户线)或无线(例如红外、无线、微波等)方式向另一个网站站点、计算机、服务器或数据中心进行传输。所述计算机可读存储介质可以是计算机能够存取的任何可用介质或者是包含一个或多个可用介质集成的服务器、数据中心等数据存储设备。所述可用介质可以是磁性介质,(例如,软盘、硬盘、磁带)、光介质(例如DVD)、或者半导体介质(例如固态硬盘)等。
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,该流程可以由计算机程序来指令相关的硬件完成,该程序可存储于计算机可读取存储介质中,该程序在执行时,可包括如上述各方法实施例的流程。而前述的存储介质包括:ROM或随机存储记忆体RAM、磁碟或者光盘等各种可存储程序代码的介质。

Claims (20)

  1. 一种视频生成方法,其特征在于,所述方法包括:
    第一电子设备确定自身的地理位置信息满足第一预设条件;
    所述第一电子设备获取实时拍摄的视频和/或图像;
    所述第一电子设备选取所述实时拍摄的视频和/或图像中的部分画面;
    当所述第一电子设备的地理位置信息满足第二预设条件,所述第一电子设备生成第一视频,所述第一视频是基于所述部分画面拼接而成的。
  2. 根据权利要求1所述的方法,其特征在于,所述第一预设条件包括:
    所述第一电子设备当前自身的地理位置离开了用户常驻区域;或者
    所述第一电子设备当前自身的地理位置离开了用户常驻区域并且离开的持续时间超过第一阈值;或者
    所述第一电子设备当前自身的地理位置和预设位置之间的最短距离大于或等于第二阈值,所述预设位置包括所述用户常驻区域;或者
    所述第一电子设备当前自身的地理位置进入了特定区域。
  3. 根据权利要求1所述的方法,其特征在于,所述视频和/或图像是通过第二电子设备实时拍摄的。
  4. 根据权利要求1所述的方法,其特征在于,所述第一电子设备选取所述视频和/或图像中的部分画面,包括:
    所述第一电子设备对获取到的视频和/或图像进行分段;
    所述第一电子设备基于美学评分算法对所述分段后的视频和/或图像进行评分,选取所述分段后每一段视频和/或图像中评分最高的画面,所述部分画面包括所述分段后每一段视频和/或图像中评分最高的画面。
  5. 根据权利要求4所述的方法,其特征在于,所述美学评分算法的评分权重与所述分段后的视频和/或图像的场景类型相关联。
  6. 根据权利要求1所述的方法,其特征在于,所述第二预设条件包括:
    所述第一电子设备当前自身的地理位置进入了预设位置;或者
    所述第一电子设备当前自身的地理位置距离预设位置的剩余时间小于第五阈值。
  7. 根据权利要求1所述的方法,其特征在于,所述第一电子设备生成第一视频,之后还包括:
    所述第一电子设备输出第一提示信息,所述第一提示信息用于提供针对所述第一视频的编辑方式,所述编辑方式包括播放所述第一视频,保存所述第一视频,删除所述第一视频,调整所述第一视频中显示画面的排序,删除所述第一视频中的部分显示画面,在所述第一视频的显示画面之间插入转场动效,在所述第一视频的显示画面上添加文字、贴纸、水印,添加所述第一视频的背景音乐中的一项或多项。
  8. 根据权利要求1所述的方法,其特征在于,所述方法还包括:
    所述第一电子设备基于所述第一电子设备的地理位置信息或者所述第一视频的视频内容,生成所述第一视频的视频名称。
  9. 根据权利要求1-8任一项所述的方法,其特征在于,所述方法还包括:
    所述第一电子设备将所述第一视频发送给第三电子设备,所述第三电子设备包括所述第二电子设备。
  10. 一种视频生成系统,包括第一电子设备和第二电子设备,所述第一电子设备和所述第二电子设备建立连接;
    所述第二电子设备,用于拍摄视频和/或图像;
    所述第一电子设备,用于确定自身的地理位置信息满足第一预设条件;
    所述第一电子设备,还用于获取所述第二电子设备实时拍摄的视频和/或图像;
    所述第一电子设备,还用于选取所述实时拍摄的视频和/或图像中的部分画面;
    所述第一电子设备,还用于当所述第一电子设备的地理位置信息满足第二预设条件,生成第一视频,所述第一视频是基于所述部分画面拼接而成的。
  11. 根据权利要求10所述的系统,其特征在于,所述第一预设条件包括:
    所述第一电子设备当前自身的地理位置离开了用户常驻区域;或者
    所述第一电子设备当前自身的地理位置离开了用户常驻区域并且离开的持续时间超过第一阈值;或者
    所述第一电子设备当前自身的地理位置和预设位置之间的最短距离大于或等于第二阈值,所述预设位置包括所述用户常驻区域;或者
    所述第一电子设备当前自身的地理位置进入了特定区域。
  12. 根据权利要求10所述的系统,其特征在于,所述第一电子设备,还用于选取所述实时拍摄的视频和/或图像中的部分画面,包括:
    所述第一电子设备,还用于对获取到的视频和/或图像进行分段;
    所述第一电子设备,还用于基于美学评分算法对所述分段后的视频和/或图像进行评分,选取所述分段后每一段视频和/或图像中评分最高的画面,所述部分画面包括所述分段后每一段视频和/或图像中评分最高的画面。
  13. 根据权利要求12所述的系统,其特征在于,所述美学评分算法的评分权重与所述分段后的视频和/或图像的场景类型相关联。
  14. 根据权利要求10所述的系统,其特征在于,所述第二预设条件包括:
    所述第一电子设备当前自身的地理位置进入了预设位置;或者
    所述第一电子设备当前自身的地理位置距离预设位置的剩余时间小于第五阈值。
  15. 根据权利要求10所述的系统,其特征在于,所述第一电子设备,还用于在生成第一 视频之后,输出第一提示信息,所述第一提示信息用于提供针对所述第一视频的编辑方式,所述编辑方式包括播放所述第一视频,保存所述第一视频,删除所述第一视频,调整所述第一视频中显示画面的排序,删除所述第一视频中的部分显示画面,在所述第一视频的显示画面之间插入转场动效,在所述第一视频的显示画面上添加文字、贴纸、水印,添加所述第一视频的背景音乐中的一项或多项。
  16. 根据权利要求10所述的系统,其特征在于,所述第一电子设备,还用于基于所述第一电子设备的地理位置信息或者所述第一视频的视频内容,生成所述第一视频的视频名称。
  17. 根据权利要求1-16任一项所述的系统,其特征在于,所述第一电子设备,还用于将所述第一视频发送给所述第二电子设备。
  18. 根据权利要求17任一项所述的系统,其特征在于,所述第二电子设备,还用于接收所述第一视频;
    所述第二电子设备,还用于输出第二提示信息,所述第二提示信息用于提供针对所述第一视频的编辑方式,所述编辑方式包括播放所述第一视频,删除所述第一视频,调整所述第一视频中显示画面的排序,删除所述第一视频中的部分显示画面,在所述第一视频的显示画面之间插入转场动效,在所述第一视频的显示画面上添加文字、贴纸、水印,添加所述第一视频的背景音乐中的一项或多项。
  19. 一种电子设备,其特征在于,包括:一个或多个处理器、一个或多个存储器;所述一个或多个存储器分别与所述一个或多个处理器耦合;所述一个或多个存储器用于存储计算机程序代码,所述计算机程序代码包括计算机指令;当所述计算机指令在所述处理器上运行时,使得所述电子设备执行如权利要求1-9所述的方法。
  20. 一种计算机可读介质,用于存储一个或多个程序,其中所述一个或多个程序被配置为被所述一个或多个处理器执行,所述一个或多个程序包括指令,所述指令用于执行如权利要求1-9所述的方法。
PCT/CN2022/140279 2021-12-22 2022-12-20 视频生成系统、方法及相关装置 WO2023116669A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202111580204.3 2021-12-22
CN202111580204.3A CN116366785A (zh) 2021-12-22 2021-12-22 视频生成系统、方法及相关装置

Publications (1)

Publication Number Publication Date
WO2023116669A1 true WO2023116669A1 (zh) 2023-06-29

Family

ID=86901350

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/140279 WO2023116669A1 (zh) 2021-12-22 2022-12-20 视频生成系统、方法及相关装置

Country Status (2)

Country Link
CN (1) CN116366785A (zh)
WO (1) WO2023116669A1 (zh)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004135208A (ja) * 2002-10-15 2004-04-30 Hitachi Ltd 映像撮影装置及び方法
CN104735353A (zh) * 2015-03-11 2015-06-24 广东欧珀移动通信有限公司 一种拍摄全景照片的方法及装置
US20150243326A1 (en) * 2014-02-24 2015-08-27 Lyve Minds, Inc. Automatic generation of compilation videos
CN112446363A (zh) * 2021-01-29 2021-03-05 广州市玄武无线科技股份有限公司 一种基于视频抽帧的图像拼接与去重方法及装置
CN112770061A (zh) * 2020-12-16 2021-05-07 影石创新科技股份有限公司 视频剪辑方法、系统、电子设备及存储介质
WO2021139728A1 (zh) * 2020-01-07 2021-07-15 影石创新科技股份有限公司 全景视频处理方法、装置、设备及存储介质

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004135208A (ja) * 2002-10-15 2004-04-30 Hitachi Ltd 映像撮影装置及び方法
US20150243326A1 (en) * 2014-02-24 2015-08-27 Lyve Minds, Inc. Automatic generation of compilation videos
CN104735353A (zh) * 2015-03-11 2015-06-24 广东欧珀移动通信有限公司 一种拍摄全景照片的方法及装置
WO2021139728A1 (zh) * 2020-01-07 2021-07-15 影石创新科技股份有限公司 全景视频处理方法、装置、设备及存储介质
CN112770061A (zh) * 2020-12-16 2021-05-07 影石创新科技股份有限公司 视频剪辑方法、系统、电子设备及存储介质
CN112446363A (zh) * 2021-01-29 2021-03-05 广州市玄武无线科技股份有限公司 一种基于视频抽帧的图像拼接与去重方法及装置

Also Published As

Publication number Publication date
CN116366785A (zh) 2023-06-30

Similar Documents

Publication Publication Date Title
WO2021129688A1 (zh) 显示方法及相关产品
CN110134316B (zh) 模型训练方法、情绪识别方法及相关装置和设备
CN110138959B (zh) 显示人机交互指令的提示的方法及电子设备
CN115866121B (zh) 应用界面交互方法、电子设备和计算机可读存储介质
WO2020078299A1 (zh) 一种处理视频文件的方法及电子设备
WO2021104485A1 (zh) 一种拍摄方法及电子设备
CN114390139B (zh) 一种电子设备在来电时呈现视频的方法、电子设备和存储介质
WO2020029306A1 (zh) 一种图像拍摄方法及电子设备
CN114173000B (zh) 一种回复消息的方法、电子设备和系统、存储介质
CN113170037B (zh) 一种拍摄长曝光图像的方法和电子设备
CN113691842A (zh) 一种跨设备的内容投射方法及电子设备
CN114756785A (zh) 页面显示的方法、装置、电子设备以及可读存储介质
CN112532508B (zh) 一种视频通信方法及视频通信装置
CN114064160A (zh) 应用图标布局方法及相关装置
WO2023071940A1 (zh) 跨设备的导航任务的同步方法、装置、设备及存储介质
CN116055856B (zh) 相机界面显示方法、电子设备和计算机可读存储介质
WO2021204103A1 (zh) 照片预览方法、电子设备和存储介质
WO2021147483A1 (zh) 数据分享的方法和装置
CN116561085A (zh) 图片分享方法和电子设备
WO2023116669A1 (zh) 视频生成系统、方法及相关装置
CN114205318B (zh) 头像显示方法及电子设备
WO2024114785A1 (zh) 一种图像处理方法、电子设备及系统
CN114115772B (zh) 灭屏显示的方法及装置
WO2023098467A1 (zh) 语音解析方法、电子设备、可读存储介质及芯片系统
CN116095223B (zh) 通知显示方法和终端设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22909981

Country of ref document: EP

Kind code of ref document: A1