WO2022166263A1 - 一种车载直播方法及装置 - Google Patents

一种车载直播方法及装置 Download PDF

Info

Publication number
WO2022166263A1
WO2022166263A1 PCT/CN2021/126289 CN2021126289W WO2022166263A1 WO 2022166263 A1 WO2022166263 A1 WO 2022166263A1 CN 2021126289 W CN2021126289 W CN 2021126289W WO 2022166263 A1 WO2022166263 A1 WO 2022166263A1
Authority
WO
WIPO (PCT)
Prior art keywords
video stream
server
viewing angle
terminal device
video
Prior art date
Application number
PCT/CN2021/126289
Other languages
English (en)
French (fr)
Inventor
段少华
金鑫垚
谢坪慧
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to EP21924260.9A priority Critical patent/EP4287623A4/en
Publication of WO2022166263A1 publication Critical patent/WO2022166263A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/21805Source of audio or video content, e.g. local disk arrays enabling multiple viewpoints, e.g. using a plurality of cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/233Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/23418Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/239Interfacing the upstream path of the transmission network, e.g. prioritizing client content requests
    • H04N21/2393Interfacing the upstream path of the transmission network, e.g. prioritizing client content requests involving handling client requests
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/251Learning process for intelligent management, e.g. learning user preferences for recommending movies
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/258Client or end-user data management, e.g. managing client capabilities, user preferences or demographics, processing of multiple end-users preferences to derive collaborative data
    • H04N21/25808Management of client data
    • H04N21/25841Management of client data involving the geographical location of the client
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/258Client or end-user data management, e.g. managing client capabilities, user preferences or demographics, processing of multiple end-users preferences to derive collaborative data
    • H04N21/25866Management of end-user data
    • H04N21/25891Management of end-user data being end-user preferences
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/27Server based end-user applications
    • H04N21/274Storing end-user multimedia data in response to end-user request, e.g. network recorder
    • H04N21/2743Video hosting of uploaded data from client
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/414Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance
    • H04N21/41422Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance located in transportation means, e.g. personal vehicle
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/4223Cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • H04N21/4316Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations for displaying supplemental content in a region of the screen, e.g. an advertisement in a separate window
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/4728End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for selecting a Region Of Interest [ROI], e.g. for requesting a higher resolution version of a selected region
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/65Transmission of management data between client and server
    • H04N21/658Transmission by the client directed to the server
    • H04N21/6587Control parameters, e.g. trick play commands, viewpoint selection
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/816Monomedia components thereof involving special video data, e.g 3D video

Definitions

  • the present application relates to the technical field of smart cars, and in particular, to a method and device for in-vehicle live broadcasting.
  • the mobile live broadcast technology is mainly realized through the mobile phone client and depends on the specific video live broadcast platform.
  • smart cars can enrich the mobile live broadcast technology by providing users with live video functions.
  • video live broadcast services have been provided for users through the dash cam or mobile phone installed inside the vehicle.
  • the way the device performs live broadcast is limited by the fixed position of the camera, and the camera faces in a fixed direction. For example, if the camera faces directly in front of the vehicle, the user can only see the live video directly in front.
  • the live viewing angle provided by this live broadcast method It is inconvenient to switch.
  • interesting scenery may appear in other directions than the fixed direction. If the anchor wants to switch the live viewing angle to show the user interesting scenery, the anchor can only manually control the camera to switch the angle of view. user experience.
  • the present application provides a method and device for in-vehicle live broadcasting, which are used to realize intelligent switching of live viewing angles.
  • an embodiment of the present application provides a method for in-vehicle live broadcasting.
  • the method includes: the server receives at least one first video stream from a first terminal device, and the first video stream is collected by collecting any direction outside the first vehicle.
  • the live video is obtained, and the server processes at least one first video stream to obtain a panoramic video stream.
  • a panoramic video stream is obtained, so that the second terminal device on the viewer side can be provided with a video stream of any viewing angle, and the viewer side The user of the terminal device can see the live video in any direction outside the first vehicle, thereby helping to realize intelligent switching of the live viewing angle.
  • the method further includes: the server receives first request information from the second terminal device, where the first request information is used to request a live video outside the first vehicle; the server obtains, according to the first request information, The video stream corresponding to the target viewing angle in the panoramic video stream; the server sends the video stream corresponding to the target viewing angle to the second terminal device. Therefore, the server may recommend a video stream corresponding to the target viewing angle for the second terminal device on the viewer side according to the first request information of the second terminal device on the viewer side.
  • the first request information includes first indication information
  • the first indication information is used to instruct the server to determine the video stream corresponding to the first viewing angle according to the content of the live video.
  • the second terminal device on the viewer side can instruct the server to intelligently recommend a viewing angle for the viewer side by means of user instructions, and provide a corresponding video stream.
  • the server obtains, according to the first request information, the video stream corresponding to the target viewing angle of view in the panoramic video stream, including: The video stream is framed to obtain at least one frame of picture corresponding to each preset viewing angle; the server matches the content of the at least one frame of picture with the content of the reference picture stored in the server, and determines the content of the reference picture.
  • the server can intelligently recommend a video stream of a viewing angle with a high user acceptance degree to the viewer in combination with the content whose user acceptance degree is greater than the first threshold.
  • the method further includes: the server obtains user preference data, where the user preference data includes the content of the video streams whose user acceptance degree is greater than the second threshold and the category to which the content of each video stream belongs.
  • the server obtains, according to the first request information, a video stream corresponding to a target viewing angle of view in the panoramic video stream, including: The video stream is classified to obtain the category to which the content of the video stream of each preset perspective belongs; the server determines the user of the video stream of each preset perspective according to the user preference data and the category to which the content of the video stream of each preset perspective belongs. Recognition degree; the server determines the first viewing angle according to the user acceptance degree of the video stream of each preset angle of view; the server obtains the video stream corresponding to the first viewing angle in the panoramic video stream. In this way, the server can intelligently recommend a video stream with a viewing angle that is highly recognized by the user for the viewer in combination with the user preference data.
  • the method further includes: the server receiving a second video stream from the first terminal device, where the second video stream is obtained by collecting the live video of the first host.
  • the method includes: the server determines, according to the second video stream, a gesture feature of the first anchor and a feature keyword in the audio; the gesture feature includes expressions and/or body movements; the server determines, according to the first request information to obtain the video stream corresponding to the target viewing angle in the panoramic video stream, including: the server determines, according to the first request information, the viewing angle corresponding to the posture feature of the first anchor and the feature keyword in the audio from the panoramic video stream , as the first viewing angle; the server obtains the video stream corresponding to the first viewing angle in the panoramic video stream.
  • the server can intelligently recommend a video stream from a viewing angle of interest to the viewer by combining the host's posture feature and the feature keyword in the audio.
  • the method further includes: the server acquires current environment information; wherein the current environment information includes at least one of the following: current location information of the first vehicle, current weather information at the location of the first vehicle, POI information on the map point of interest whose distance from the location of the first vehicle is within a preset range; the server obtains, according to the first request information, the video stream corresponding to the target viewing angle of view in the panoramic video stream, including: the server obtains, according to the first request information, the video stream corresponding to the target viewing angle of view , and determine the viewing angle related to the current environment information from the panoramic video stream as the first viewing angle; the server obtains the video stream corresponding to the first viewing angle in the panoramic video stream.
  • the server can intelligently recommend video streams with viewing angles that the public likes, such as scenic spots, natural landscapes, etc., for the viewer in combination with the current environment information.
  • the first request information includes second indication information
  • the second indication information is used to indicate a second viewing angle
  • the second viewing angle is used to instruct the server to deliver the video stream according to the second viewing angle.
  • the server obtains, according to the first request information, the video stream corresponding to the target viewing angle in the panoramic video stream, including: obtaining, according to the first request information, the server corresponding to the second viewing angle in the panoramic video stream
  • the server sends the video stream corresponding to the target viewing angle to the second terminal device, including: the server sending the video stream corresponding to the second viewing angle to the second terminal device. Therefore, the server can deliver the video stream of the specified viewing angle selected by the user according to the instruction of the viewer.
  • the method further includes: the server receiving second request information sent from the second terminal device; the second request information is used to request to switch the current viewing angle to the third viewing angle; The two terminal devices send the video stream corresponding to the third viewing angle.
  • the current viewing angle may be the first viewing angle or the second viewing angle.
  • the server may also provide the viewer with a video stream of the switched viewing angle.
  • the method further includes: the server sending the second video stream to the second terminal device.
  • the server sending the second video stream to the second terminal device.
  • the method server receives a third video stream from the first terminal device, where the third video stream is obtained by collecting the live video of the second host.
  • the first terminal device on the live broadcast side can replace the host, and accordingly, the server can receive the live video of the replaced host, so as to provide the viewer side with the live video of the replaced second host.
  • an embodiment of the present application provides a method for in-vehicle live broadcasting, the method includes: a first terminal device receives a first instruction; the first instruction is used to instruct to start a live video; the first terminal device starts a video according to the first instruction Live broadcast function; the first terminal device obtains live videos in all directions outside the first vehicle, and obtains at least one first video stream; the first terminal device sends at least one first video stream to the server.
  • the terminal device on the host side can obtain the live video outside the first vehicle and upload it to the server, so that the server can provide the terminal device on the viewer side with a video stream of a target viewing angle.
  • the method further includes: the first terminal device obtains the live video of the first host, and obtains the second video stream; The server sends the second video stream.
  • the terminal device on the anchor side can not only upload the live video outside the vehicle to the server, but also upload the live video of the first anchor to the server, so that the server can not only provide the viewer side with the live video outside the vehicle, but also provide The audience side provides the live video of the first anchor.
  • the first terminal device collects the live video of the first host through the second video capture device, and after obtaining the second video stream, the method further includes: the first terminal device receives the second instruction; the second instruction uses to instruct the live footage to switch from the first host to the second host; the first terminal device adjusts the live shot to the second host, collects the live video of the second host, and obtains a third video stream; the first terminal device sends the third video to the server flow.
  • the live broadcast lens can be switched to change the anchor, and the live video of the changed anchor can be uploaded to the server.
  • an embodiment of the present application provides a method for in-vehicle live broadcasting, including: a second terminal device receives a third instruction, where the third instruction is used to instruct to obtain a live video of the first vehicle; the second terminal device sends a first Request information, where the first request information is used to request the live video of the first vehicle; the second terminal device receives the video stream corresponding to the target viewing angle from the server.
  • the second terminal device on the viewer side can obtain the video stream corresponding to the target viewing angle delivered by the server by sending a request to the server, so that the viewer can see the live video of the first vehicle in the target viewing angle.
  • the first request information includes first indication information
  • the first indication information is used to instruct the server to determine the video stream corresponding to the first viewing angle according to the content of the live video
  • the second terminal device receives the video stream from the server.
  • the video stream corresponding to the target viewing angle includes: the second terminal device receives the video stream corresponding to the first viewing angle from the server. Therefore, the viewer side can instruct the server to determine a video stream corresponding to a viewing angle, so as to obtain a video stream of a viewing angle recommended by the server.
  • the first request information includes second indication information
  • the second indication information is used to indicate a second viewing angle
  • the second viewing angle is used to instruct the server to deliver the video stream according to the second viewing angle
  • the first The second terminal device receiving the video stream corresponding to the target viewing angle from the server includes: the second terminal device receiving the video stream corresponding to the second viewing angle from the server.
  • the method further includes: the second terminal device sends user preference data to the server, where the user preference data includes data of the video stream watched by the user of the second terminal device.
  • the method further includes: the second terminal device sends second request information to the server, where the second request information is used to request to switch the current viewing angle to the third viewing angle; the second terminal device receives The video stream corresponding to the target viewing angle from the server includes: the second terminal device receives the video stream corresponding to the third viewing angle sent by the server.
  • the current viewing angle may be the above-mentioned first viewing angle or the second viewing angle. In this way, the viewer side can freely switch the viewing angle in the process of watching the live video outside the vehicle, so that the video stream of the switched viewing angle instructed by the viewer can be obtained from the server.
  • the method further includes: the second terminal device receives a second video stream sent by the server, where the second video stream is obtained by collecting the live video of the first host; the second terminal device displays a floating window, The floating window is used to display the second video stream.
  • the user on the viewer side can not only watch the live video outside the vehicle on the second terminal device, but also watch the live video of the host.
  • the present application further provides a communication device.
  • the communication apparatus may be any device on the sending end or the device on the receiving end that performs data transmission in a wireless manner.
  • the device on the sending end and the device on the receiving end are relative.
  • the communication device can be used as the above-mentioned first terminal device or a communication chip that can be used in the first terminal device; in some communication processes, the communication device can be used as the above-mentioned second terminal device or can be used in the second terminal device.
  • the communication chip in some communication processes, the communication device can be used as the above-mentioned server or a communication chip that can be used for the server.
  • a communication device including a communication unit and a processing unit, so as to execute any one of the implementations of any one of the above-mentioned methods for in-vehicle live broadcasting of the first to third aspects.
  • the communication unit is used to perform functions related to transmission and reception.
  • the communication unit includes a receiving unit and a sending unit.
  • the communication device is a communication chip, and the communication unit may be an input/output circuit or port of the communication chip.
  • the communication unit may be a transmitter and receiver, or the communication unit may be a transmitter and receiver.
  • the communication device further includes various modules that can be used to execute any one of the implementation manners of any one of the above-mentioned in-vehicle live broadcast methods of the first aspect to the third aspect.
  • a communication device is provided, where the communication device is the above server, including a processor and a memory.
  • the memory is used to store a computer program or instruction
  • the processor is used to call and run the computer program or instruction from the memory, and when the processor executes the computer program or instruction in the memory, make the computer program or instruction in the memory.
  • the communication device executes any one of the implementations of any one of the above-mentioned in-vehicle live broadcast methods of the first aspect to the third aspect.
  • processors there are one or more processors and one or more memories.
  • the memory may be integrated with the processor, or the memory may be provided separately from the processor.
  • the transceiver may include a transmitter (transmitter) and a receiver (receiver).
  • a communication apparatus including a processor.
  • the processor coupled to the memory, is operable to perform the method of any one of the first to third aspects, and any one of the possible implementations of the first to third aspects.
  • the communication device further includes a memory.
  • the communication device further includes a communication interface, and the processor is coupled to the communication interface.
  • the communication apparatus is a terminal device end.
  • the communication interface may be a transceiver, or an input/output interface.
  • the transceiver may be a transceiver circuit.
  • the input/output interface may be an input/output circuit.
  • the communication device is a server.
  • the communication interface may be a transceiver, or an input/output interface.
  • the transceiver may be a transceiver circuit.
  • the input/output interface may be an input/output circuit.
  • the communication device is a chip or a system of chips.
  • the communication interface may be an input/output interface, an interface circuit, an output circuit, an input circuit, a pin or a related circuit, etc. on the chip or a chip system.
  • a processor may also be embodied as processing circuitry or logic circuitry.
  • a communication system in a seventh aspect, includes at least two of the above-mentioned first terminal device, second terminal and server.
  • a computer program product includes: a computer program (also referred to as code, or instruction), which, when the computer program is executed, enables the computer to execute any one of the above-mentioned first aspects.
  • the method in the manner, or causing the computer to execute the method in any one of the implementation manners of the first aspect to the third aspect.
  • a computer-readable storage medium stores a computer program (also referred to as code, or instruction), when it runs on a computer, so that the computer executes any one of the above-mentioned first aspects.
  • the method in one possible implementation manner, or causing the computer to execute the method in any one of the implementation manners of the first aspect to the third aspect.
  • a system-on-chip may include a processor.
  • the processor coupled to the memory, is operable to perform the method of any one of the first to third aspects, and any one of the possible implementations of any of the first to third aspects.
  • the chip system further includes a memory.
  • Memory used to store computer programs (also referred to as code, or instructions).
  • a processor for invoking and running a computer program from a memory, so that a device on which the chip system is installed performs any one of the first to third aspects, and any one of the first to third aspects is possible method in the implementation.
  • the above-mentioned processing device may be a chip
  • the input circuit may be an input pin
  • the output circuit may be an output pin
  • the processing circuit may be a transistor, a gate circuit, a flip-flop, and various logic circuits.
  • the input signal received by the input circuit may be received and input by, for example, but not limited to, a receiver
  • the signal output by the output circuit may be, for example, but not limited to, output to and transmitted by a transmitter
  • the circuit can be the same circuit that acts as an input circuit and an output circuit at different times.
  • the embodiments of the present application do not limit the specific implementation manners of the processor and various circuits.
  • FIG. 1 is a schematic diagram of the architecture of a possible communication system to which an embodiment of the present application is applicable;
  • FIG. 2 is a schematic structural diagram of a vehicle to which the embodiment of the application is applied;
  • FIG. 3 is a schematic diagram of a three-dimensional model of a panoramic video live broadcast provided by an embodiment of the present application.
  • FIG. 4 is a schematic flowchart of a method for in-vehicle live broadcasting provided by an embodiment of the present application
  • FIG. 5 is a schematic diagram of a communication device provided by an embodiment of the present application.
  • FIG. 6 is a schematic diagram of a communication device provided by an embodiment of the present application.
  • FIG. 7 is a schematic diagram of a communication device provided by an embodiment of the present application.
  • FIG. 1 is a schematic structural diagram of a possible communication system to which an embodiment of the present application is applied.
  • the communication system may include a terminal device and an intelligent car cloud service platform.
  • the terminal device includes at least one live broadcast side terminal device. 1 schematically shows a vehicle 110 .
  • the terminal device may further include at least one viewer-side terminal device.
  • the viewer-side terminal device is a mobile phone as an example for illustration, and FIG. 1 schematically shows the mobile phone 120 .
  • the smart car cloud service platform includes at least one server 130 .
  • the live broadcast side terminal device 110 is used for collecting real-time original live video, and uploading the collected original live video to the server in the smart car cloud service platform.
  • the original live video collected by the terminal device on the live broadcast side may include not only the live video outside the vehicle collected by the camera device, but also the host video collected by the camera device and including the host.
  • the smart car cloud service platform includes at least one server 130 .
  • the server 130 may be configured to receive the original live video sent by the terminal device on the live broadcast side, and use an artificial intelligence (artificial intelligence, AI) algorithm to determine the target viewing angle in the original live video video, and then send it to the terminal device on the viewer side through the live broadcast service interface. Delivers the video stream corresponding to the target viewing angle.
  • AI artificial intelligence
  • the viewer-side terminal device 120 is configured to receive and play the target live video.
  • the terminal device on the live broadcast side is taken as an example of a vehicle for illustration.
  • Any vehicle in the embodiments of the present application may be a smart car or a non-smart car, and the comparison of the embodiments of the present application is not limited.
  • Each vehicle is provided with a sensor, and the sensor is used to collect images near the vehicle, wherein the sensor may be a camera.
  • each vehicle may be provided with one or more sensors, and the number of each sensor may be one or more.
  • the sensor can be installed on the top of the vehicle (for example, it can be arranged in the middle position of the top of the vehicle), the front end of the vehicle, the rear end of the vehicle, the left side of the vehicle, the right side of the vehicle, etc. Not limited.
  • the components inside the vehicle involved in the embodiment of the present application may include an automatic driving system (auto driving system, ADS) system, a mobile data center (mobile data center, MDC), a cockpit domain controller (cockpit domain controller) , CDC), in-vehicle communication box (telematics box, TBOX), external camera, internal camera, microphone (MIC).
  • ADS automatic driving system
  • MDC mobile data center
  • cockpit domain controller cockpit domain controller
  • TBOX in-vehicle communication box
  • TBOX in-vehicle communication box
  • the external camera can be set with one camera for the four directions of front, rear, left, and right
  • the internal camera can be set in the driver's seat or in the passenger seat in the car. This embodiment of the present application does not limit this.
  • the vehicle can use the ADS system to send the audio and video outside the car (hereinafter referred to as the first video stream) collected by the external camera to the MDC, and can also use the CDC to send the audio and video inside the car collected by the internal camera in the car ( Hereinafter referred to as the second video stream) transmits the MDC.
  • MDC uploads the collected first video stream and second video stream to the smart car cloud service platform through the network channel provided by TBOX.
  • the smart car cloud service provides video splicing function and provides panoramic live video.
  • the cloud's AI algorithm provides an intelligent viewing angle of live video and pushes it to the live client.
  • the vehicle may communicate with other objects based on a wireless communication technology between the vehicle and the outside world (for example, vehicle to everything (V2X)).
  • V2X vehicle to everything
  • the communication between the vehicle and the terminal device on the audience side may be implemented based on an inter-vehicle wireless communication technology (eg, vehicle to vehicle (V2V)).
  • Communication between vehicles and other objects may be based on wireless high-fidelity (eg, wireless fidelity (Wi-Fi)), fifth-generation (5th generation, 5G) mobile communication technologies, and the like.
  • Wi-Fi wireless fidelity
  • 5G fifth-generation
  • the server may receive at least one first video stream sent from the terminal device on the live broadcast side, and the first video stream is obtained by collecting live video from any direction outside the first vehicle, and then the server can perform the processing on the at least one first video stream.
  • the first video stream is processed to obtain a panoramic video stream, and then the video stream corresponding to the target viewing angle in the panoramic video stream can be obtained and sent to the terminal device on the audience side, so as to intelligently provide the terminal device on the audience side with the video corresponding to the target viewing angle. flow, which can effectively improve the user experience.
  • the terminal device on the live broadcast side is referred to as the first terminal device, and the terminal device on the viewer side is referred to as the second terminal device.
  • an on-vehicle live broadcast method provided by an embodiment of the present application will be introduced below. As shown in FIG. 3 , the method includes:
  • Step 301 The first terminal device receives a first instruction, where the first instruction is used to instruct to start live video.
  • the first instruction may have multiple implementation manners.
  • the first instruction may be a voice instruction.
  • the first terminal device may receive the voice instruction through the microphone inside the first vehicle, and start the live video broadcast.
  • the first instruction may be implemented by the host triggering a button inside the first vehicle.
  • the first terminal device receives a first operation and generates a first instruction, and the first operation is for the first operation. Click operation of buttons inside the vehicle.
  • Step 302 the first terminal device starts the live video function according to the first instruction.
  • the user account of the first host when the first host uses the live video function for the first time, based on the operation of the first host, the user account of the first host can be registered, and the live client can be logged in to create a live room. Afterwards, when the first anchor activates the video live broadcast function, the user account of the first anchor can log in to the live broadcast client to live broadcast the live video around the outside of the first vehicle where the first anchor is located to the audience.
  • Step 303 the first terminal device obtains live video in all directions outside the first vehicle, and obtains at least one first video stream.
  • the first terminal device may enable at least one first video capture device to capture live video from various directions outside the first vehicle when the live video function is activated according to the first instruction.
  • a first video capture device may be set outside the first vehicle.
  • a panoramic camera is set at the top central position outside the first vehicle. The camera collects video frame data within a 360-degree field of view in the horizontal direction outside the first vehicle and video frame data within a 180-degree field of view in the vertical direction to obtain one channel of video frame data, that is, a first video stream.
  • a plurality of first video capture devices may be set outside the first vehicle.
  • the first video capture device as a camera as an example
  • the front, rear, left, and right directions outside the first vehicle The cameras are respectively set, and the multi-frame images collected by each camera are encoded to obtain a video stream, so that a total of four first video streams in the front, rear, left and right directions can be obtained.
  • Step 304 the first terminal device sends at least one first video stream to the server.
  • the server receives at least one first video stream from the first terminal device.
  • Step 305 The server processes at least one first video stream to obtain a panoramic video stream.
  • the server decodes the video frame data in the first video stream to generate panoramic video frames, and then uses 3D modeling to obtain a panoramic video stream, As shown in FIG. 4 , a 3D coordinate system xyz is established with the first video acquisition device on the first as the origin, a 3D sphere model is created under the 3D coordinate system xyz, and the panoramic video frame is set at a preset time (for example, millisecond level) The sequence is continuously projected to the 3D sphere model, and the 3D model of the panoramic video live broadcast is generated.
  • a preset time for example, millisecond level
  • the user of the second terminal device is hereinafter referred to as the first user for convenience of description.
  • the first user can watch the panoramic video on the live 3D model of the panoramic video from any viewing angle.
  • the positive half-axis direction of the x-axis corresponds to the first vehicle.
  • the positive half-axis of the z-axis corresponds to the right side of the first vehicle
  • the positive half-axis of the y-axis corresponds to the direction of the roof of the first vehicle.
  • the first user can view point M and its surrounding area on the three-dimensional sphere at the viewing angle of rotation degree ⁇ and elevation angle ⁇ at the origin O, wherein the rotation degree ⁇ takes the xOy plane as the starting point and rotates in the direction of the positive semi-axis of the z-axis.
  • the value of ⁇ can be [0, 360°]
  • the elevation angle ⁇ rotates in the direction of the positive semi-axis of the y-axis from the xOz plane as the starting point
  • the value of ⁇ can be [0, 180°]
  • the degree of rotation is ⁇
  • the elevation angle is ⁇ , which can uniquely determine a viewing angle.
  • the size of the surrounding area of point M depends on the size of the window on the second terminal device for displaying the live video.
  • the server decodes and splices the multiple first video streams to generate panoramic video frames, and uses 3D modeling to obtain a panoramic video stream.
  • Step 306 the second terminal device receives a third instruction, where the third instruction is used to instruct to acquire the live video of the first vehicle.
  • the second terminal device as the viewer-side terminal device, can log in to the live client through the user account of the first user based on a series of operations of the user of the second terminal device (hereinafter referred to as the first user), and Find the live broadcast room corresponding to the first vehicle in the live broadcast client, and join the live broadcast room.
  • the third instruction may be an instruction triggered by the first user entering the live broadcast room corresponding to the first vehicle.
  • the third instruction may be triggered by a click operation of the first user on the live broadcast control displayed on the second terminal device.
  • the live broadcast control is a control used to enter the live broadcast room corresponding to the first vehicle, and the third instruction may also be a voice command.
  • the third instruction may include a live video mode selected by the user, wherein the live video mode may be an intelligent viewing angle mode or a fixed viewing angle mode.
  • the live video mode may be an intelligent viewing angle mode or a fixed viewing angle mode.
  • the user-selected live video mode included in the third instruction is an intelligent viewing angle mode
  • the second terminal device generates first indication information
  • the first indication information is used to instruct the server to recommend a viewing angle of view, and deliver the recommendation.
  • the live video mode selected by the user included in the third instruction is a fixed viewing angle mode
  • the second terminal device generates second indication information
  • the second indication information includes the first viewing angle
  • the second indication information is used to indicate the server Delivers the video stream corresponding to the first viewing angle.
  • Step 307 The second terminal device sends first request information to the server, where the first request information is used to request a live video outside the first vehicle.
  • the server receives the first request information from the second terminal device.
  • the first request information may include first indication information for instructing the server to determine the video stream corresponding to the first viewing angle according to the content of the live video.
  • the server receives the first request information, and recommends a viewing angle, that is, the first viewing angle, for the user of the second terminal device according to the first indication information carried in the first request information.
  • the server can intelligently recommend a viewing angle for the first user, and can also intelligently switch the viewing angle for the first user during the live broadcast process.
  • the first request information includes second indication information
  • the second indication information is used to indicate a second viewing angle of view
  • the second viewing angle of view is used to instruct the server according to the second
  • the video stream from the viewing angle is delivered. That is, the second terminal device may instruct the server to deliver the video stream corresponding to the fixed viewing angle, and the fixed viewing angle is the second viewing angle.
  • the server receives the first request information, and delivers the video stream of the second viewing angle to the second terminal device according to the second indication information carried in the first request information. In this way, the first user can choose the viewing angle according to his own preference.
  • Step 308 The server obtains a video stream corresponding to the target viewing angle in the panoramic video stream according to the first request information.
  • the server obtains the video stream corresponding to the target viewing angle in the panoramic video stream according to the first request information, including: Frames are extracted from video streams of preset viewing angles, at least one frame of pictures corresponding to each preset viewing angle is obtained, and then, the content of the at least one frame of the picture is matched with the content of the reference picture stored in the server, and the content of the reference picture is determined to be the same as the reference picture.
  • the user recognition degree corresponding to the preset viewing angle that matches the content of the reference picture wherein the content included in the reference picture is the content whose user recognition degree is greater than the first threshold
  • the server determines the user recognition degree corresponding to the preset viewing angle that matches the content of the reference picture according to the user recognition degree. , determine the first viewing angle, and then acquire the video stream corresponding to the first viewing angle in the panoramic video stream.
  • the threshold can be set according to actual needs, for example, it is set to 90%, and for example, it is set to 98%, which is not limited here.
  • each reference picture stored in the server, and the content of each reference picture also corresponds to a user recognition degree.
  • at least one frame of pictures corresponding to the viewing angle may be the same as a reference.
  • the picture matches, and may also match with multiple reference pictures. The greater the number of matching reference pictures, the higher the user recognition corresponding to the viewing angle.
  • the first viewing angle may be the viewing angle with the highest user recognition degree, or may be one of the n viewing angles with the highest user recognition degree.
  • the server may acquire user preference data, wherein the user preference data includes the content of the video streams whose user approval degree is greater than the second threshold and the category to which the content of each video stream belongs.
  • the server obtains, according to the first request information, the video stream corresponding to the target viewing angle of view in the panoramic video stream, which specifically includes: according to the first request information, the server classifies the video streams of each preset angle of view in the panoramic video stream to obtain each preset angle of view.
  • the server determines the user approval degree of the video stream of each preset viewing angle according to the user preference data and the category to which the content of the video stream of each preset viewing angle belongs.
  • the user recognition degree of the video stream of the viewing angle is set, the first viewing angle is determined, and then the video stream corresponding to the first viewing angle in the panoramic video stream is obtained.
  • the second terminal device sends user preference data to the server, where the user preference data may include the first user's user recognition of the category to which the content of each video stream viewed in the past belongs to.
  • other viewer-side terminal devices other than the second terminal device send user preference data to the server, and the user preference data may also be the historical viewing of users of other viewer-side terminal devices (hereinafter referred to as the second user) User recognition of the category to which the content of each video stream that has been viewed.
  • the user preference data may also be user approval degrees of the first user and the at least one second user to the categories to which the content of each video stream that has been viewed historically belongs.
  • the server receives the second video stream from the first terminal device, the second video stream is obtained by collecting the live video of the first anchor, and the server determines the posture feature and audio of the first anchor according to the second video stream.
  • Feature keywords in , and gesture features include expressions and/or body movements.
  • the server obtains the video stream corresponding to the target viewing angle in the panoramic video stream, which specifically includes: the server determines, according to the first request information, the posture feature and the audio feature of the first anchor from the panoramic video stream.
  • the viewing angle corresponding to the keyword is used as the first viewing angle, and then the server obtains the video stream corresponding to the first viewing angle in the panoramic video stream.
  • the server can identify the feature keyword "look to the right” in the audio of the first anchor, and the server can combine the feature keyword "look to the right” in the audio ” to determine the viewing angle indicated by the first anchor, referring to the xyz coordinate system shown in Figure 4, the viewing angle corresponding to the right side of the first vehicle is: rotation 90°, elevation angle 0°, the server can determine the first anchor instructed
  • the viewing angle of view is a rotation degree of 90° and an elevation angle of 0°, that is, the first viewing angle of view is determined to be a rotation degree of 90° and an elevation angle of 0°.
  • the server can identify The feature keyword "that tree" in the audio of the first anchor, the posture feature of the first anchor is: the finger points to a direction with an angle of 30 degrees in front of the first vehicle, see the xyz coordinate system shown in Figure 4 , the viewing angle corresponding to the direction in which the angle directly in front of the first vehicle is 30 degrees is the rotation degree of 30° and the elevation angle is 0.
  • the server can combine the body movements of the first anchor and the characteristic keywords in the audio to determine the viewing angle as the rotation.
  • the first viewing angle is determined to be 90° of rotation and 0° of elevation.
  • the first host in the first vehicle can also switch to another host for live broadcast.
  • the first terminal device can receive a second instruction, and the second instruction is used to instruct the live broadcast shot Switching from the first anchor to the second anchor, the first terminal device adjusts the live lens to the second anchor, collects the live video of the second anchor, obtains a third video stream, and sends the third video stream to the server.
  • the server receives the third video stream, and can send the third video stream to the second terminal device.
  • the second instruction may be triggered by the first host, or may be triggered by the second host, and the specific triggering method may be a button trigger or a voice trigger.
  • the server may acquire current environment information, and the server acquires the video stream corresponding to the target viewing angle in the panoramic video stream according to the first request information, including: the server acquires the video stream corresponding to the target viewing angle in the panoramic video stream according to the first request information , the viewing angle related to the current environment information is determined from the panoramic video stream, and as the first viewing angle, the server obtains the video stream corresponding to the first viewing angle in the panoramic video stream.
  • the current environment information may include any one or more of the following:
  • the current location information of the first vehicle the current weather information at the location of the first vehicle, and the map point of interest (POI) information whose distance from the location of the first vehicle is within a preset range.
  • POI map point of interest
  • the first terminal device may obtain the current weather information at the location of the first vehicle, for example, the current temperature is -5°C, and there is an ice surface not far from the included angle of 45° between the right side of the vehicle and the front, which is related to the ice surface.
  • the viewing angle of view is 45° of rotation and 0° of elevation, and the server determines that the first viewing angle is 45° of rotation and 0° of elevation.
  • the first terminal device may obtain POI information around the first current location from the high-precision map cloud service, such as Internet celebrity attractions, and the name of a landmark building. For example, if there is an internet celebrity attraction in the direction of the right side of the vehicle and the front of the vehicle at an angle of 90°, the server determines that the first viewing angle is 90° of rotation and 0° of elevation.
  • the high-precision map cloud service such as Internet celebrity attractions, and the name of a landmark building. For example, if there is an internet celebrity attraction in the direction of the right side of the vehicle and the front of the vehicle at an angle of 90°, the server determines that the first viewing angle is 90° of rotation and 0° of elevation.
  • the server obtains the video stream corresponding to the target viewing angle in the panoramic video stream according to the first request information, which specifically includes: the server counts the number of viewers corresponding to each viewing angle in the panoramic video stream according to the first request information, The viewing angle with the largest number of viewers is determined from the panoramic video stream, and as the first viewing angle, the server obtains the video stream corresponding to the first viewing angle in the panoramic video stream.
  • Possible implementation b6 is to determine the first viewing angle based on the combination of any of the foregoing implementations b1 to b5.
  • the server may combine the content of the reference picture in implementation b1 and the user in implementation b2 Any of preference data, the posture feature of the first anchor in Embodiment b3 and the viewing angle corresponding to the feature keyword in the audio, the current environment information in Embodiment b4, and the number of viewers corresponding to each viewing angle in Embodiment b5.
  • the first viewing angle is determined by various factors, and the specific manner may refer to the above-mentioned possible implementation manners b1 to b5, which will not be repeated here.
  • each viewing angle is scored according to the video stream of the viewing angle and the following data content items, and the first viewing angle is determined according to the final result of each viewing angle score. See Table 1 below for a recommended scoring table for perspectives.
  • the data content items mainly include the content of video frames, the current number of online viewers, the directionality of the in-car anchor, and environmental factors.
  • the specific implementation can refer to the above-mentioned possible implementation b1, if the judgment result is yes, it can be scored according to the user's favorite degree. , the score can be any value from 1 to 10 points, if the judgment result is no, it is 0 points, the final score of this item is counted as S1, and the weight of this item is set as P1.
  • the final score of the current number of online viewers is counted as S2, and the weight of this item is set as P2.
  • the final score of the directivity of the in-car anchor is counted as S3, and the weight of this item is set as P3.
  • the final score of this item of environmental factors is counted as S4, and the weight of this item is set as P4. where the weights are tuned according to an internal algorithm.
  • the server obtains, according to the first request information, the video stream corresponding to the target viewing angle in the panoramic video stream, which specifically includes: the server obtains, according to the first request information, the video stream in the panoramic video stream.
  • the video stream corresponding to the second viewing angle of view is possible Embodiment b7, based on the above-mentioned possible Embodiment a2, the server obtains, according to the first request information, the video stream corresponding to the target viewing angle in the panoramic video stream, which specifically includes: the server obtains, according to the first request information, the video stream in the panoramic video stream.
  • the video stream corresponding to the second viewing angle of view is possible Embodiment b7, based on the above-mentioned possible Embodiment a2
  • Step 309 the server sends the video stream corresponding to the target viewing angle to the second terminal device.
  • the server sends the video stream corresponding to the target viewing angle to the second terminal device, which specifically includes: the server sends the video stream corresponding to the first viewing angle to the second terminal device.
  • the second terminal device receives the video stream corresponding to the first viewing angle from the server.
  • the server sends the video stream corresponding to the target viewing angle to the second terminal device, which specifically includes: the server sending the video stream corresponding to the second viewing angle to the second terminal device.
  • the second terminal device receives the video stream corresponding to the second viewing angle from the server.
  • Step 310 the second terminal device displays the video content of the video stream corresponding to the target viewing angle.
  • step 310 the second terminal device displays the video content of the video stream corresponding to the first viewing angle of view.
  • step 310 the second terminal device displays the video content of the video stream corresponding to the second viewing angle of view.
  • the server when the user of the second terminal device selects the intelligent viewing angle mode, the server may automatically calculate the target viewing angle and recommend it to the user.
  • the server may send the video stream of the viewing angle specified by the user to the second terminal device.
  • the second video stream may also be sent to the second terminal device, and the second terminal device may receive the second video stream from the server, and may display the video of the second video stream through a floating window
  • the video content of the second video stream may not be displayed, so that it will not affect the first user's viewing of the video content of the first video stream, and the user experience can be improved.
  • the user of the second terminal device can switch the viewing angle in the process of watching the live broadcast, so as to see the video content that the user wants to watch. For example, by triggering the fourth instruction, the user can indicate that the current viewing angle is switched to the second viewing angle, and the current viewing angle may be the above-mentioned first viewing angle or the second viewing angle.
  • the second terminal device may receive the fourth instruction, and the second terminal device sends second request information to the server in response to the fourth instruction, where the second request information is used to request to switch the current viewing angle to the third viewing angle; the server sends the second request information to the server.
  • the second terminal device sends the video stream corresponding to the third viewing angle, and correspondingly, the second terminal device receives the video stream corresponding to the third viewing angle sent by the server.
  • system and “network” in the embodiments of the present application may be used interchangeably.
  • At least one means one or more, and “plurality” means two or more.
  • And/or which describes the association relationship of the associated objects, indicates that there can be three kinds of relationships, for example, A and/or B, which can indicate: the existence of A alone, the existence of A and B at the same time, and the existence of B alone, where A, B can be singular or plural.
  • the character “/” generally indicates that the associated objects are an “or” relationship.
  • At least one item(s) below” or similar expressions thereof refer to any combination of these items, including any combination of single item(s) or plural items(s).
  • At least one item (a) of a, b, or c can represent: a, b, c, a-b, a-c, b-c, or a-b-c, where a, b, c may be single or multiple .
  • ordinal numbers such as “first” and “second” mentioned in the embodiments of the present application are used to distinguish multiple objects, and are not used to limit the order, sequence, priority or importance of multiple objects degree.
  • first vehicle and the second vehicle are only used to distinguish different vehicles, and do not indicate differences in the priority or importance of the two vehicles.
  • each network element in the above-mentioned implementation includes corresponding hardware structures and/or software modules for executing each function.
  • the present invention can be implemented in hardware or a combination of hardware and computer software in conjunction with the units and algorithm steps of each example described in the embodiments disclosed herein. Whether a function is performed by hardware or computer software driving hardware depends on the specific application and design constraints of the technical solution. Skilled artisans may implement the described functionality using different methods for each particular application, but such implementations should not be considered beyond the scope of the present invention.
  • FIG. 5 is a schematic structural diagram of a communication apparatus provided by an embodiment of the present application.
  • the communication apparatus may be a first terminal device, a second terminal device, or a server. It can also be a chip or a circuit, such as a chip or circuit that can be set in the map updating device at the terminal device, or a chip or a circuit that can be set in the second map updating device.
  • the communication device 1301 may further include a bus system, wherein the processor 1302, the memory 1304, and the transceiver 1303 may be connected through the bus system.
  • the above-mentioned processor 1302 may be a chip.
  • the processor 1302 may be a field programmable gate array (FPGA), an application specific integrated circuit (ASIC), a system on chip (SoC), or a system on chip (SoC). It can be a central processing unit (CPU), a network processor (NP), a digital signal processing circuit (DSP), or a microcontroller (microcontroller). unit, MCU), it can also be a programmable logic device (PLD) or other integrated chips.
  • FPGA field programmable gate array
  • ASIC application specific integrated circuit
  • SoC system on chip
  • SoC system on chip
  • SoC system on chip
  • MCU microcontroller
  • MCU programmable logic device
  • PLD programmable logic device
  • each step of the above-mentioned method can be completed by an integrated logic circuit of hardware in the processor 1302 or an instruction in the form of software.
  • the steps of the methods disclosed in conjunction with the embodiments of the present application may be directly embodied as being executed by a hardware processor, or executed by a combination of hardware and software modules in the processor 1302 .
  • the software modules may be located in random access memory, flash memory, read-only memory, programmable read-only memory or electrically erasable programmable memory, registers and other storage media mature in the art.
  • the storage medium is located in the memory 1304, and the processor 1302 reads the information in the memory 1304, and completes the steps of the above method in combination with its hardware.
  • processor 1302 in this embodiment of the present application may be an integrated circuit chip, which has a signal processing capability.
  • steps of the above method embodiments may be completed by hardware integrated logic circuits in the processor or instructions in the form of software.
  • the aforementioned processors may be general purpose processors, digital signal processors (DSPs), application specific integrated circuits (ASICs), field programmable gate arrays (FPGAs) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components .
  • DSPs digital signal processors
  • ASICs application specific integrated circuits
  • FPGAs field programmable gate arrays
  • the methods, steps, and logic block diagrams disclosed in the embodiments of this application can be implemented or executed.
  • a general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
  • the steps of the method disclosed in conjunction with the embodiments of the present application may be directly embodied as executed by a hardware decoding processor, or executed by a combination of hardware and software modules in the decoding processor.
  • the software modules may be located in random access memory, flash memory, read-only memory, programmable read-only memory or electrically erasable programmable memory, registers and other storage media mature in the art.
  • the storage medium is located in the memory, and the processor reads the information in the memory, and completes the steps of the above method in combination with its hardware.
  • the memory 1304 in this embodiment of the present application may be a volatile memory or a non-volatile memory, or may include both volatile and non-volatile memory.
  • the non-volatile memory may be read-only memory (ROM), programmable read-only memory (PROM), erasable programmable read-only memory (EPROM), electrically programmable Erase programmable read-only memory (electrically EPROM, EEPROM) or flash memory.
  • Volatile memory may be random access memory (RAM), which acts as an external cache.
  • RAM random access memory
  • DRAM dynamic random access memory
  • SDRAM synchronous DRAM
  • SDRAM double data rate synchronous dynamic random access memory
  • ESDRAM enhanced synchronous dynamic random access memory
  • SLDRAM synchronous link dynamic random access memory
  • direct rambus RAM direct rambus RAM
  • the communication device may include a processor 1302 , a transceiver 1303 and a memory 1304 .
  • the memory 1304 is used for storing instructions
  • the processor 1302 is used for executing the instructions stored in the memory 1304, so as to implement the relevant solution of the server in any one or more of the corresponding methods shown in FIG. 3 above.
  • the transceiver 1303 is configured to receive at least one first video stream from the first terminal device, where the first video stream is obtained by collecting live video from any direction outside the first vehicle;
  • the processor 1302 is configured to process the at least one first video stream to obtain a panoramic video stream.
  • the transceiver 1303 is further configured to receive first request information from a second terminal device, where the first request information is used to request a live video outside the first vehicle; the processor 1302 , and is further configured to obtain the video stream corresponding to the target viewing angle in the panoramic video stream according to the first request information; the transceiver 1303 is further configured to send the second terminal device the corresponding video stream of the target viewing angle video stream.
  • the first request information includes first indication information, where the first indication information is used to instruct the server to determine the video stream corresponding to the first viewing angle according to the content of the live video.
  • the processor 1302 is configured to, according to the first indication information, perform frame extraction on the video stream of each preset viewing angle in the panoramic video stream, and obtain each preset viewing angle corresponding at least one frame of picture; the content of the at least one frame of picture is matched with the content of the reference picture stored in the server, and the user recognition degree corresponding to the preset viewing angle that matches the content of the reference picture is determined ;
  • the content included in the reference picture is the content whose user approval degree is greater than the first threshold value; According to the user approval degree corresponding to the preset viewing angle that matches the content of the reference picture, determine the first viewing angle of view;
  • Transceiver 1303 is further used to obtain a video stream corresponding to the first viewing angle of view in the panoramic video stream.
  • the transceiver 1303 is further configured for the server to obtain user preference data, where the user preference data includes the content of the video streams whose user approval degree is greater than the second threshold and the content of each video stream to which the content belongs. category.
  • the processor 1302 is configured to, according to the first request information, classify the video streams of each preset viewing angle in the panoramic video stream to obtain the video stream of each preset viewing angle.
  • the transceiver 1303 is configured to receive a second video stream from the first terminal device, where the second video stream is obtained by collecting the live video of the first host.
  • the processor 1302 is further configured to determine, according to the second video stream, a gesture feature of the first anchor and a feature keyword in the audio; the gesture feature includes expressions and/or Body movements; according to the first request information, determine the viewing angle corresponding to the posture feature of the first anchor and the feature keyword in the audio from the panoramic video stream, as the first viewing angle;
  • the server acquires a video stream corresponding to the first viewing angle in the panoramic video stream.
  • the processor 1302 is further configured to acquire current environment information; wherein the current environment information includes at least one of the following: current location information of the first vehicle, location of the first vehicle The current weather information at the location, the POI information of the map point of interest whose distance from the location of the first vehicle is within a preset range; according to the first request information, it is determined from the panoramic video stream The viewing angle related to the current environment information is used as the first viewing angle; the video stream corresponding to the first viewing angle in the panoramic video stream is acquired.
  • the first request information includes second indication information
  • the second indication information is used to indicate a second viewing angle of view
  • the second viewing angle of view is used to instruct the server according to the The second viewing angle delivers the video stream.
  • the processor 1302 is used for the server to obtain the video stream corresponding to the second viewing angle of view in the panoramic video stream according to the first request information; the transceiver 1303 is configured to use for sending the video stream corresponding to the second viewing angle to the second terminal device.
  • the transceiver 1303 is configured to receive second request information sent from the second terminal device; the second request information is used to request to switch the current viewing angle to a third viewing angle ; send the video stream corresponding to the third viewing angle to the second terminal device.
  • the transceiver 1303 is configured to send the second video stream to the second terminal device.
  • the transceiver 1303 is further configured to receive a third video stream from the first terminal device, where the third video stream is obtained by collecting the live video of the second host.
  • the transceiver 1303 is used for the first terminal device to receive the first instruction; the first instruction is used to instruct to start a live video; the processor 1302 is used to according to the first instruction , start the live video function; obtain live videos from all directions outside the first vehicle, and obtain at least one first video stream; the transceiver 1303 is further configured to send the at least one first video stream to the server.
  • the processor 1302 is further configured to acquire the live video of the first host to obtain a second video stream; the transceiver 1303 is further configured to send the second video stream to the server.
  • the transceiver 1303 is further configured to receive a second instruction; the second instruction is used to instruct the live footage to switch from the first anchor to the second anchor; the transceiver 1303 is further configured to Adjust the live lens to face the second anchor, collect the live video of the second anchor, and obtain a third video stream; the transceiver 1303 is further configured to send the third video stream to the server.
  • the processor 1302 is configured to obtain the first position information of the map element; and obtain the first error information, where the first error information includes the positioning system of the first vehicle positioning the first vehicle. Error information; the transceiver 1303 is configured to send the first position information and the first error information of the map element to the second map updating apparatus.
  • the transceiver 1303 is configured to receive a third instruction, where the third instruction is used to instruct to obtain the live video of the first vehicle; send first request information to the server, the first request information It is used to request the live video of the first vehicle; and receive the video stream corresponding to the target viewing angle from the server.
  • the first request information includes first indication information
  • the first indication information is used to instruct the server to determine the video stream corresponding to the first viewing angle according to the content of the live video; the transceiver 1303, for receiving a video stream corresponding to the first viewing angle of view from the server.
  • the first request information includes second indication information
  • the second indication information is used to indicate a second viewing angle of view
  • the second viewing angle of view is used to instruct the server according to the The second viewing angle delivers a video stream
  • the transceiver 1303 is configured to receive a video stream corresponding to the second viewing angle from the server.
  • the transceiver 1303 is used for the transceiver 1303, and is further configured to send user preference data to the server, where the user preference data includes data of the user of the second terminal device watching the video stream .
  • the transceiver 1303 is configured to send second request information to the server, where the second request information is used to request to switch the current viewing angle to a third viewing angle; receive the server The sent video stream corresponding to the third viewing angle of view.
  • the transceiver 1303 is configured to receive a second video stream sent by the server, where the second video stream is obtained by collecting the video broadcast live by the first host; the processor 1302 is configured to use A floating window is displayed on the second terminal device, and the floating window is used to display the second video stream.
  • FIG. 6 is a schematic structural diagram of a communication apparatus provided by an embodiment of the present application.
  • the communication apparatus 1401 may include a communication interface 1403 , a processor 1402 , and a memory 1404 .
  • the communication interface 1403 is used for inputting and/or outputting information;
  • the processor 1402 is used for executing computer programs or instructions, so that the communication device 1401 can implement the method of the first terminal device in the above-mentioned related scheme of FIG. 3, or the communication device 1401 can realize the above-mentioned method
  • the method of the second terminal device in the related scheme of FIG. 3 or the method of enabling the communication apparatus 1401 to implement the method of the server in the related scheme of FIG. 3 .
  • the communication interface 1403 can implement the solution implemented by the transceiver 1303 in FIG. 5
  • the processor 1402 can implement the solution implemented by the processor 1302 in FIG. 5
  • the memory 1404 can implement the memory 1304 in FIG. 5. The implemented solution will not be repeated here.
  • FIG. 7 is a schematic diagram of a communication apparatus provided by an embodiment of the present application.
  • the communication apparatus 1501 may be a first terminal device, a second terminal device, or a third terminal device, or It can be a chip or a circuit, such as a chip or circuit that can be provided in the first terminal device, the second terminal device, or the third terminal device.
  • the communication apparatus may correspond to the first terminal device in the above method.
  • the communication apparatus may implement the steps performed by the first terminal device in any one or more of the corresponding methods shown in FIG. 3 above.
  • the communication apparatus may include a processing unit 1502 and a communication unit 1503 .
  • the communication unit 1503 is used for the first terminal device to receive a first instruction; the first instruction is used to instruct to start a live video; the processing unit 1502 is used to start a live video function according to the first instruction; obtain all directions outside the first vehicle to obtain at least one first video stream; the transceiver 1303 is further configured to send the at least one first video stream to the server.
  • the processing unit 1502 is configured to acquire the first position information of the map element; acquire the first error information, where the first error information includes a pair of positioning systems of the first vehicle. Error information of the positioning of the first vehicle; the communication unit 1503 is configured to send the first position information and the first error information of the map element to the second map updating apparatus.
  • the communication unit 1503 is configured to receive at least one first video stream from the first terminal device, where the first video stream is a live broadcast from any direction outside the first vehicle. video get;
  • the processing unit 1502 processes the at least one first video stream to obtain a panoramic video stream.
  • each unit in the foregoing communication apparatus 1501 may refer to the implementation of the corresponding method embodiments, and details are not described herein again.
  • the division of the units of the above communication apparatus is only a division of logical functions, and may be fully or partially integrated into a physical entity in actual implementation, or may be physically separated.
  • the communication unit 1503 may be implemented by the transceiver 1303 shown in FIG. 5 above, and the processing unit 1502 may be implemented by the processor 1302 shown in FIG. 5 above.
  • the present application also provides a computer program product, the computer program product includes: computer program code or instructions, when the computer program code or instructions are run on a computer, the computer is made to execute FIG. 3 The method of any one of the illustrated embodiments.
  • the present application also provides a computer-readable storage medium, where the computer-readable medium stores program codes, and when the program codes are run on a computer, the computer is made to execute the implementation shown in FIG. 3 .
  • the present application further provides a chip system, where the chip system may include a processor.
  • the processor is coupled to the memory and can be used to perform the method of any one of the embodiments shown in FIG. 3 .
  • the chip system further includes a memory. Memory for storing computer programs (also called code, or instructions).
  • the processor is used to call and run the computer program from the memory, so that the device installed with the chip system executes the method of any one of the embodiments shown in FIG. 3 .
  • the present application further provides a system, which includes the aforementioned one or more vehicles and a map update device on the server side, and the map update device is provided in the vehicle.
  • a computer program product includes one or more computer instructions.
  • the computer may be a general purpose computer, a special purpose computer, a computer network, or other programmable device.
  • Computer instructions may be stored in or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer instructions may be transmitted from a website site, computer, server, or data center over a wire (e.g.
  • coaxial cable, optical fiber, digital subscriber line (DSL)) or wireless means to transmit to another website site, computer, server or data center.
  • a computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device such as a server, a data center, or the like that includes an integration of one or more available media.
  • Useful media may be magnetic media (eg, floppy disk, hard disk, magnetic tape), optical media (eg, high-density digital video disc (DVD)), or semiconductor media (eg, solid state disc (SSD)) )Wait.
  • the second map updating device in each of the above device embodiments corresponds to the map updating device on the terminal device side and the second map updating device or the map updating device on the terminal device side in the method embodiments, and corresponding steps are performed by corresponding modules or units, such as
  • the communication unit (transceiver) performs the step of receiving or sending in the method embodiment, and other steps except the sending and receiving may be performed by the processing unit (processor).
  • processor For functions of specific units, reference may be made to corresponding method embodiments.
  • the number of processors may be one or more.
  • a component may be, but is not limited to, a process running on a processor, a processor, an object, an executable, a thread of execution, a program, and/or a computer.
  • an application running on a computing device and the computing device may be components.
  • One or more components may reside within a process and/or thread of execution, and a component may be localized on one computer and/or distributed between two or more computers.
  • these components can execute from various computer readable media having various data structures stored thereon.
  • a component may, for example, be based on a signal having one or more data packets (eg, data from two components interacting with another component between a local system, a distributed system, and/or a network, such as the Internet interacting with other systems via signals) Communicate through local and/or remote processes.
  • data packets eg, data from two components interacting with another component between a local system, a distributed system, and/or a network, such as the Internet interacting with other systems via signals
  • Units described as separate components may or may not be physically separated, and components shown as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution in this embodiment.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
  • the functions, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored in a computer-readable storage medium.
  • the technical solution of the present application can be embodied in the form of a software product in essence, or the part that contributes to the prior art or the part of the technical solution.
  • the computer software product is stored in a storage medium, including Several instructions are used to cause a computer device (which may be a personal computer, a server, etc.) to execute all or part of the steps of the methods of various embodiments of the present application.
  • the aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (ROM), random access memory (RAM), magnetic disk or optical disk and other media that can store program codes .

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Computer Graphics (AREA)
  • Computing Systems (AREA)
  • Human Computer Interaction (AREA)
  • Business, Economics & Management (AREA)
  • Marketing (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

一种车载直播方法及装置,用于实现智能切换直播视角。本申请中服务器接收来自第一终端设备的至少一条第一视频流,第一视频流为通过采集第一车辆外部的任一方向的直播视频得到,服务器对至少一条第一视频流进行处理,得到全景视频流。通过该方法,服务器对来自直播侧的第一终端设备的至少一条第一视频流进行处理之后,得到全景视频流,从而可以为观众侧的第二终端设备提供任意视角的视频流,进而观众侧终端设备的用户可以看到第一车辆外部的任意方向的直播视频,从而有助于实现智能切换直播视角。

Description

一种车载直播方法及装置
相关申请的交叉引用
本申请要求在2021年02月04日提交中国专利局、申请号为202110156289.6、申请名称为“一种车载直播方法及装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及智能汽车技术领域,尤其涉及一种车载直播方法及装置。
背景技术
近年来,随着移动互联网中直播技术的流行,目前每个人都可以作为信息生产者对外发布音视频作品,通过丰富的真实场景和内容呈现到用户眼前,用来吸引流量和粉丝以达成自己的用户资源和经济利益需求。
移动直播技术主要是通过手机客户端实现,依赖具体的视频直播平台。智能汽车作为互联网的移动终端,为用户提供视频直播功能可以丰富移动直播技术,目前已有通过在车辆内部加装的行车记录仪或者手机,为用户提供视频直播服务,但是,这种需要额外增加设备进行直播的方式,受限于摄像头的固定位置,摄像头朝向固定方向,例如,摄像头朝向车辆的正前方,用户就只能看到正前方的直播视频,而且,这种直播方式提供的直播视角不方便切换,在车辆行驶过程中,可能在固定方向之外的其它方向出现有趣的风景,主播想要切换直播视角给用户看有趣的风景,也只能通过主播手动控制摄像头进行视角切换,影响用户体验。
发明内容
本申请提供一种车载直播方法及装置,用于实现智能切换直播视角。
第一方面,本申请实施例提供了一种车载直播方法,该方法包括:服务器接收来自第一终端设备的至少一条第一视频流,第一视频流为通过采集第一车辆外部的任一方向的直播视频得到,服务器对至少一条第一视频流进行处理,得到全景视频流。
通过该方法,服务器对来自直播侧的第一终端设备的至少一条第一视频流进行处理之后,得到全景视频流,从而可以为观众侧的第二终端设备提供任意视角的视频流,进而观众侧终端设备的用户可以看到第一车辆外部的任意方向的直播视频,从而有助于实现智能切换直播视角。
在一种可能地实施方式中,该方法还包括:服务器接收来自第二终端设备的第一请求信息,第一请求信息用于请求第一车辆外部的直播视频;服务器根据第一请求信息,获取全景视频流中的目标观看视角对应的视频流;服务器向第二终端设备发送目标观看视角对应的视频流。因此,服务器可以根据观众侧的第二终端设备的第一请求信息,为观众侧的第二终端设备推荐目标观看视角对应的视频流。
在一种可能地实施方式中,第一请求信息包括第一指示信息,第一指示信息用于指示 服务器根据直播视频的内容确定第一观看视角对应的视频流。如此,观众侧的第二终端设备可以通过用户指示的方式,指示服务器智能的为观众侧推荐观看视角,并提供相应的视频流。
在一种可能地实施方式中,服务器根据第一请求信息,获取全景视频流中的目标观看视角对应的视频流,包括:服务器根据第一指示信息,对全景视频流中每个预设视角的视频流进行抽帧,获取每个预设视角对应的至少一帧图片;服务器将至少一帧图片的内容与服务器中存储的参考图片的内容进行匹配,确定出与参考图片的内容相匹配的预设视角对应的用户认可度;参考图片包括的内容为用户认可度大于第一阈值的内容;服务器根据与参考图片的内容相匹配的预设视角对应的用户认可度,确定第一观看视角;服务器获取全景视频流中的第一观看视角对应的视频流。如此,服务器可以结合用户认可度大于第一阈值的内容,智能的为观众侧推荐用户认可度高的观看视角的视频流。
在一种可能地实施方式中,该方法还包括:服务器获取用户偏好数据,用户偏好数据包括用户认可度大于第二阈值的视频流的内容以及每个视频流的内容所属的类别。
在一种可能地实施方式中,服务器根据第一请求信息,获取全景视频流中的目标观看视角对应的视频流,包括:服务器根据第一请求信息,对全景视频流中每个预设视角的视频流进行分类得到每个预设视角的视频流的内容所属的类别;服务器根据用户偏好数据和每个预设视角的视频流的内容所属的类别,确定每个预设视角的视频流的用户认可度;服务器根据每个预设视角的视频流的用户认可度,确定第一观看视角;服务器获取全景视频流中的第一观看视角对应的视频流。如此,服务器可以结合用户偏好数据,智能的为观众侧推荐用户认可度高的观看视角的视频流。
在一种可能地实施方式中,该方法还包括:服务器接收来自第一终端设备的第二视频流,第二视频流为通过采集第一主播的直播视频得到。
在一种可能地实施方式中,该方法包括:服务器根据第二视频流,确定第一主播的姿态特征和音频中的特征关键字;姿态特征包括表情和/或肢体动作;服务器根据第一请求信息,获取全景视频流中的目标观看视角对应的视频流,包括:服务器根据第一请求信息,从全景视频流中确定出与第一主播的姿态特征和音频中的特征关键字对应的观看视角,作为第一观看视角;服务器获取全景视频流中的第一观看视角对应的视频流。如此,服务器可以结合主播的姿态特征和音频中的特征关键字,智能的为观众侧推荐主播感兴趣的观看视角的视频流。
在一种可能地实施方式中,该方法还包括:服务器获取当前环境信息;其中,当前环境信息包括以下至少一项:第一车辆的当前位置信息、第一车辆所在位置处的当前天气信息、与第一车辆所在位置处的距离在预设范围内的地图兴趣点POI信息;服务器根据第一请求信息,获取全景视频流中的目标观看视角对应的视频流,包括:服务器根据第一请求信息,从全景视频流中确定出与当前环境信息相关的观看视角,作为第一观看视角;服务器获取全景视频流中的第一观看视角对应的视频流。如此,服务器可以结合当前环境信息,智能的为观众侧推荐例如景点、自然景观等大众喜欢的观看视角的视频流。
在一种可能地实施方式中,第一请求信息包括第二指示信息,第二指示信息用于指示第二观看视角,第二观看视角用于指示服务器根据第二观看视角下发视频流。如此,观众侧可以根据自己的喜好自己选择某个指定的观看视角,从而指示服务器下发指定的观看视角对应的视频流。
在一种可能地实施方式中,服务器根据第一请求信息,获取全景视频流中的目标观看视角对应的视频流,包括:服务器根据第一请求信息,获取全景视频流中的第二观看视角对应的视频流;服务器向第二终端设备发送目标观看视角对应的视频流,包括:服务器向第二终端设备发送第二观看视角对应的视频流。因此,服务器可以根据观众侧的指示,下发用户选择的指定的观看视角的视频流。
在一种可能地实施方式中,该方法还包括:服务器接收来自第二终端设备发送的第二请求信息;第二请求信息用于请求将当前的观看视角切换为第三观看视角;服务器向第二终端设备发送第三观看视角对应的视频流。该当前的观看视角可以为上述第一观看视角或者第二观看视角。如此,服务器还可以在接收到观众侧指示切换观看视角时,为观众侧提供切换后的观看视角的视频流。
在一种可能地实施方式中,该方法还包括:服务器向第二终端设备发送第二视频流。如此,可以使得观众侧在观看车辆外部的直播视频之外,还可以看到主播的直播视频。
在一种可能地实施方式中,该方法服务器接收来自第一终端设备的第三视频流,第三视频流为通过采集第二主播的直播视频得到。如此,直播侧的第一终端设备可以更换主播,相应的,服务器可以接收更换后的主播的直播视频,从而为观众侧提供更换后的第二主播的直播视频。
第二方面,本申请实施例提供了一种车载直播方法,该方法包括:第一终端设备接收第一指令;第一指令用于指示启动视频直播;第一终端设备根据第一指令,启动视频直播功能;第一终端设备获取第一车辆外部各个方向的直播视频,得到至少一条第一视频流;第一终端设备向服务器发送至少一条第一视频流。
通过该方法,主播侧的终端设备可以获取第一车辆外部的直播视频,并上传至服务器,以便于服务器可以为观众侧的终端设备提供目标观看视角的视频流。
在一种可能地实施方式中,第一终端设备根据第一指令,启动视频直播功能之后,还包括:第一终端设备获取第一主播的直播视频,得到第二视频流;第一终端设备向服务器发送第二视频流。如此,主播侧的终端设备不仅可以将车辆外部的直播视频上传至服务器,还可以将第一主播的直播视频上传至服务器,从而服务器可以不仅可以为观众侧提供车辆外部的直播视频,还可以为观众侧提供第一主播的直播视频。
在一种可能地实施方式中,第一终端设备通过第二视频采集装置采集第一主播直播的视频,得到第二视频流之后,还包括:第一终端设备接收第二指令;第二指令用于指示直播镜头从第一主播切换朝向第二主播;第一终端设备调整直播镜头朝向第二主播,采集第二主播的直播视频,得到第三视频流;第一终端设备向服务器发送第三视频流。如此,在第一主播直播的过程中,可以切换直播镜头用以更换主播,并上传更换后主播的直播视频上传至服务器。
第三方面,本申请实施例提供了一种车载直播方法,包括:第二终端设备接收第三指令,第三指令用于指示获取第一车辆的直播视频;第二终端设备向服务器发送第一请求信息,第一请求信息用于请求第一车辆的直播视频;第二终端设备接收来自服务器的目标观看视角对应的视频流。
通过该方法,观众侧的第二终端设备可以通过向服务器发送请求,从而获取到服务器下发的目标观看视角对应的视频流,使得观众可以看到目标观看视角的第一车辆的直播视频。
在一种可能地实施方式中,第一请求信息包括第一指示信息,第一指示信息用于指示服务器根据直播视频的内容确定第一观看视角对应的视频流;第二终端设备接收来自服务器的目标观看视角对应的视频流,包括:第二终端设备接收来自服务器的第一观看视角对应的视频流。因此,观众侧可以指示服务器确定一个观看视角对应的视频流,从而获取到服务器推荐的观看视角的视频流。
在一种可能地实施方式中,第一请求信息包括第二指示信息,第二指示信息用于指示第二观看视角,第二观看视角用于指示服务器根据第二观看视角下发视频流;第二终端设备接收来自服务器的目标观看视角对应的视频流,包括:第二终端设备接收来自服务器的第二观看视角对应的视频流。如此,观众侧可以根据自己的喜好自己指定一个第二观看视角,从而获取到服务器下发观众指定的第二观看视角对应的视频流。
在一种可能地实施方式中,该方法还包括:第二终端设备向服务器发送用户偏好数据,用户偏好数据包括第二终端设备的用户观看视频流的数据。
在一种可能地实施方式中,该方法还包括:第二终端设备向服务器发送第二请求信息,第二请求信息用于请求将当前的观看视角切换为第三观看视角;第二终端设备接收来自服务器的目标观看视角对应的视频流,包括:第二终端设备接收服务器发送的第三观看视角对应的视频流。该当前的观看视角可以为上述的第一观看视角或者第二观看视角。如此,观众侧可以在观看车辆外部的直播视频的过程中,自由的切换观看视角,从而可以从服务器获取到观众指示的切换后的观看视角的视频流。
在一种可能地实施方式中,该方法还包括:第二终端设备接收服务器发送的第二视频流,第二视频流为采集第一主播直播的视频得到的;第二终端设备显示悬浮窗口,悬浮窗口用于显示第二视频流。如此,观众侧的用户不仅可以在第二终端设备观看车辆外部的直播视频,还可以观看主播的直播视频。
相应于第一方面至第三方面任一种车载直播方法,本申请还提供了一种通信装置。通信装置可以是以无线方式进行数据传输的任意一种发送端的设备或接收端的设备。例如,第一终端设备、第二终端设备、服务器。在通信过程中,发送端的设备和接收端的设备是相对的。在某些通信过程中,通信装置可以作为上述第一终端设备或可用于第一终端设备的通信芯片;在某些通信过程中,通信装置可以作为上述第二终端设备或可用于第二终端设备的通信芯片;在某些通信过程中,通信装置可以作为上述服务器或可用于服务器的通信芯片。
第四方面,提供了一种通信装置,包括通信单元和处理单元,以执行上述第一方面至第三方面任一种车载直播方法中的任一种实施方式。通信单元用于执行与发送和接收相关的功能。可选地,通信单元包括接收单元和发送单元。在一种设计中,通信装置为通信芯片,通信单元可以为通信芯片的输入输出电路或者端口。
在另一种设计中,通信单元可以为发射器和接收器,或者通信单元为发射机和接收机。
可选的,通信装置还包括可用于执行上述第一方面至第三方面任一种车载直播方法中的任一种实施方式的各个模块。
第五方面,提供了一种通信装置,该通信装置为上述服务器,包括处理器和存储器。可选的,还包括收发器,该存储器用于存储计算机程序或指令,该处理器用于从存储器中调用并运行该计算机程序或指令,当处理器执行存储器中的计算机程序或指令时,使得该 通信装置执行上述第一方面至第三方面任一种车载直播方法中的任一种实施方式。
可选的,处理器为一个或多个,存储器为一个或多个。
可选的,存储器可以与处理器集成在一起,或者存储器与处理器分离设置。
可选的,收发器中可以包括,发射机(发射器)和接收机(接收器)。
第六方面,提供了一种通信装置,包括处理器。该处理器与存储器耦合,可用于执行第一方面至第三方面任一方面,以及第一方面至第三方面中任一种可能实现方式中的方法。可选地,该通信装置还包括存储器。可选地,该通信装置还包括通信接口,处理器与通信接口耦合。
在一种实现方式中,该通信装置为终端设备端。当该通信装置为第一终端设备时,通信接口可以是收发器,或,输入/输出接口。可选地,收发器可以为收发电路。可选地,输入/输出接口可以为输入/输出电路。
在另一种实现方式中,该通信装置为服务器。当该通信装置为服务器时,通信接口可以是收发器,或,输入/输出接口。可选地,收发器可以为收发电路。可选地,输入/输出接口可以为输入/输出电路。
在又一种实现方式中,该通信装置为芯片或芯片系统。当该通信装置为芯片或芯片系统时,通信接口可以是该芯片或芯片系统上的输入/输出接口、接口电路、输出电路、输入电路、管脚或相关电路等。处理器也可以体现为处理电路或逻辑电路。
第七方面,提供了一种通信系统,系统包括上述第一终端设备端、第二终端设备和服务器中的至少两项。
第八方面,提供了一种计算机程序产品,计算机程序产品包括:计算机程序(也可以称为代码,或指令),当计算机程序被运行时,使得计算机执行上述第一方面中任一种可能实现方式中的方法,或者使得计算机执行上述第一方面至第三方面任一种实现方式中的方法。
第九方面,提供了一种计算机可读存储介质,计算机可读介质存储有计算机程序(也可以称为代码,或指令)当其在计算机上运行时,使得计算机执行上述第一方面中任一种可能实现方式中的方法,或者使得计算机执行上述第一方面至第三方面任一种实现方式中的方法。
第十方面,提供了一种芯片系统,该芯片系统可以包括处理器。该处理器与存储器耦合,可用于执行第一方面至第三方面中任一方面,以及第一方面至第三方面中任一方面中任一种可能实现方式中的方法。可选地,该芯片系统还包括存储器。存储器,用于存储计算机程序(也可以称为代码,或指令)。处理器,用于从存储器调用并运行计算机程序,使得安装有芯片系统的设备执行第一方面至第三方面中任一方面,以及第一方面至第三方面中任一方面中任一种可能实现方式中的方法。
在具体实现过程中,上述处理装置可以为芯片,输入电路可以为输入管脚,输出电路可以为输出管脚,处理电路可以为晶体管、门电路、触发器和各种逻辑电路等。输入电路所接收的输入的信号可以是由例如但不限于接收器接收并输入的,输出电路所输出的信号可以是例如但不限于输出给发射器并由发射器发射的,且输入电路和输出电路可以是同一电路,该电路在不同的时刻分别用作输入电路和输出电路。本申请实施例对处理器及各种电路的具体实现方式不做限定。
附图说明
图1为本申请实施例适用的一种可能的通信系统的架构示意图;
图2为本申请实施例适用的车辆的结构示意图;
图3为本申请实施例提供的一种全景视频直播三维模型示意图;
图4本申请实施例提供的一种车载直播方法的流程示意图;
图5本申请实施例提供的一种通信装置示意图;
图6本申请实施例提供的一种通信装置示意图;
图7本申请实施例提供的一种通信装置示意图。
具体实施方式
下面将结合附图对本申请作进一步地详细描述。显然,所描述的实施例是本申请一部分实施例,而不是全部的实施例。方法实施例中的具体操作方法也可以应用于装置实施例中。其中,在本申请实施例的描述中,本领域普通技术人员可以理解:本申请中涉及的第一、第二等各种数字编号仅为描述方便进行的区分,并不用来限制本申请实施例的范围,也不用来表示先后顺序。“多个”的含义是两个或两个以上。“和/或”,描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。字符“/”一般表示前后关联对象是一种“或”的关系。“至少一个”是指一个或者多个。至少两个是指两个或者多个。“至少一个”、“任意一个”或其类似表达,是指的这些项中的任意组合,包括单项(个)或复数项(个)的任意组合。
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述。
图1为本申请实施例适用的一种可能的通信系统的架构示意图。在该通信系统中可以包括终端设备和智能汽车云服务平台,如图1所示,终端设备包括至少一个直播侧终端设备,在图1中是以直播侧终端设备为车辆为例进行示意,图1示意性示出了车辆110。终端设备还可以包括至少一个观众侧终端设备,在图1中是以观众侧终端设备为手机为例进行示意,图1示意性示出了手机120。智能汽车云服务平台中包含至少一个服务器130。
所述直播侧终端设备110,用于采集实时的原始直播视频,并将采集的原始直播视频上传到智能汽车云服务平台中的服务器。示例性的,所述直播侧终端设备采集的原始直播视频不仅可以包含通过摄像装置采集的车辆外部的直播视频,还可以包含通过摄像装置采集的包含主播的主播视频。
智能汽车云服务平台中包含至少一个服务器130。所述服务器130可以用于执行接收直播侧终端设备发送的原始直播视频,并使用人工智能(artificial intelligence,AI)算法确定原始直播视频中的目标观看视角,然后通过直播服务接口向观众侧终端设备下发目标观看视角对应的视频流。
所述观众侧终端设备120,用于接收并播放目标直播视频。
本申请实施例中以直播侧终端设备为车辆为例进行示意。本申请实施例中的任意一个车辆可以是智能车或非智能车,本申请实施例对比不做限定。每个车辆上设置有传感器,该传感器用于采集车辆附近的图像,其中,传感器可以为摄像头。另外,每辆车可以设置一种或多种传感器,每种传感器的个数可以为一个或多个。传感器可以安装在车辆的顶部 (例如可以设置在车辆顶部的中间位置)、车辆前端、车辆后端、车辆左侧、车辆右侧等位置,本申请实施例对每个车辆中传感器安装位置和数量并不做限定。
如图2所示,本申请实施例涉及到的车辆内部的部件可以包括自动驾驶系统(auto driving system,ADS)系统、移动数据中心(mobile data center,MDC)、座舱域控制器(cockpit domain controller,CDC)、车载通信盒子(telematics box,TBOX)、外部摄像头、内部摄像头、麦克风(MIC)。其中,外部摄像头可以为前后左右四个方向分别设置一个摄像头,内部摄像头可以设置在驾驶位,也可以设置在车内的乘客座位。本申请实施例对此不作限定。
其中,车辆可以利用ADS系统可以将外部摄像头采集的车外的音视频(下文称为第一视频流)发送至MDC,还可以利用CDC将车内的内部摄像头采集到的车内的音视频(下文称为第二视频流)发送MDC。然后,MDC将采集到的第一视频流和第二视频流通过TBOX提供的网络通道上传至智能汽车云服务平台,智能汽车云服务提供视频的拼接功能,并提供全景直播视频,同时使用智能汽车云的AI算法,提供智能的直播视频观看视角,推送到直播客户端。
本申请实施例中车辆可以基于车辆与外界无线通信技术(例如,vehicle to everything(V2X))与其它物体进行通信。例如,可以基于车辆间无线通信技术(例如,vehicle to vehicle(V2V))实现车辆与观众侧终端设备之间的通信。车辆与其它物体之间进行通信可以基于无线高保真(例如,wireless fidelity(Wi-Fi))、第五代(5th generation,5G)移动通信技术等进行通信。
在本申请实施例中,服务器可以接收来自直播侧终端设备发送的至少一条第一视频流,第一视频流通过采集第一车辆外部的任一方向的直播视频得到,然后服务器对所述至少一条第一视频流进行处理,得到全景视频流,然后可以获取全景视频流中的目标观看视角对应的视频流发送至观众侧终端设备,从而实现智能的为观众侧终端设备提供目标观看视角对应的视频流,能够有效的提升用户体验。
为了便于描述,将直播侧终端设备称为第一终端设备,将观众侧终端设备称为第二终端设备。
下面基于如图1所示的通信系统,对本申请实施例提供的一种车载直播方法进行介绍,如图3所示,方法包括:
步骤301,第一终端设备接收第一指令,第一指令用于指示启动视频直播。
可选地,该第一指令可以有多种实现方式,一种可能的实现方式中,该第一指令可以为语音指令,例如可以为一条语音“打开视频直播”,也可以为其它语音,此处不作限制。第一终端设备可以通过第一车辆内部的麦克风接收该语音指令,启动视频直播。另一种可能的实现方式中,该第一指令可以为主播触发第一车辆内部的按键实现的,例如,第一终端设备接收第一操作,生成第一指令,该第一操作为针对第一车辆内部的按键的点击操作。
步骤302,第一终端设备根据第一指令,启动视频直播功能。
本申请实施例中,在第一主播第一次使用视频直播功能时,可以基于第一主播的操作,注册第一主播的用户账号,并登陆直播客户端,创建直播间。之后,在第一主播启动视频直播功能时,可以通过第一主播的用户账号登陆直播客户端,向观众直播第一主播所在的第一车辆外部周围的直播视频。
步骤303,第一终端设备获取第一车辆外部各个方向的直播视频,得到至少一条第一 视频流。
本申请实施例中,第一终端设备可以在根据第一指令启动视频直播功能时,打开通过至少一个第一视频采集装置,以便采集第一车辆外部各个方向的直播视频。
一种可能的实现方式中,第一车辆外部可以设置一个第一视频采集装置,以第一视频采集装置为摄像头为例,在第一车辆外部的顶端中央位置设置一个全景摄像头,可以通过该全景摄像头采集第一车辆外部水平方向上360度范围视野内的视频帧数据以及垂直方向上180度范围视野内的视频帧数据,得到一路视频帧数据,即一条第一视频流。
另一种可能的实现方式中,第一车辆外部可以设置多个第一视频采集装置,以第一视频采集装置为摄像头为例,在第一车辆外部的前、后、左、右四个方向分别设置摄像头,每个摄像头采集的多帧图像经过编码得到一条视频流,这样可以得到前、后、左、右四个方向的总共四条第一视频流。
步骤304,第一终端设备向服务器发送至少一条第一视频流。相应的,服务器接收来自第一终端设备的至少一条第一视频流。
步骤305,服务器对至少一条第一视频流进行处理,得到全景视频流。
在一个示例中,以服务器接收一条第一视频流为例,服务器对第一视频流中的视频帧数据进行解码处理后生成全景视频帧,然后采用3D建模的方式处理得到一条全景视频流,如图4所示,以第一上的第一视频采集装置为原点建立3D坐标系xyz,在3D坐标系xyz下创建一个三维球体模型,将全景视频帧以预设时间(例如,毫秒级)顺序不断投射到三维球体模型,生成全景视频直播三维模型。
以通过全景视频直播三维模型模拟第二终端设备的用户置身于第一车辆内部观看第一车辆外部周围的风景为例,为便于描述,下文中将第二终端设备的用户称为第一用户。第一用户在原点O处,可以以任一个观看视角观看全景视频直播三维模型上的全景视频,例如,如图4所示,xyz坐标系中,x轴的正半轴方向对应指向第一车辆的正前方,z轴的正半轴对应指向第一车辆的右侧,y轴的正半轴方向对应指向第一车辆的车顶方向。第一用户在原点O处以旋转度为α、仰角为β的观看视角可以观看到三维球体上M点及其周围区域,其中,旋转度α以xOy面为起点向z轴的正半轴方向旋转,α的取值可以为[0,360°],仰角β以xOz面为起点向y轴的正半轴方向旋转,为β的取值可以为[0,180°],旋转度为α和仰角为β可以唯一确定一个观看角度,此处,M点的周围区域的小大视第二终端设备上的显示直播视频的窗口大小而定。
在另一个示例中,以服务器接收多条第一视频流为例,服务器对多条第一视频流进行解码、拼接处理,生成全景视频帧,采用3D建模的方式处理得到一条全景视频流。
步骤306,第二终端设备接收第三指令,该第三指令用于指示获取第一车辆的直播视频。
本申请实施例中,第二终端设备作为观众侧终端设备,可以基于第二终端设备的用户(下文称为第一用户)的一系列操作,通过第一用户的用户账号登录直播客户端,并在该直播客户端中查找到该第一车辆对应的直播间,并加入该直播间。第三指令可以为第一用户进入第一车辆对应的直播间而触发的指令。第三指令可以为第一用户针对第二终端设备上显示的直播控件的点击操作触发的,该直播控件为用于进入第一车辆对应的直播间的控件,第三指令也可以为语音指令。
在步骤306中,第三指令可以包括用户选择的视频直播模式,其中,视频直播模式可 以是智能视角模式,也可以是固定视角模式。下面通过下述可能的实施方式a1和可能的实施方式a2来进行具体的描述。
可能的实施方式a1,第三指令包括的用户选择的视频直播模式为智能视角模式,第二终端设备生成第一指示信息,该第一指示信息用于指示服务器推荐一个观看视角,并下发推荐的观看视角对应的视频流。
可能的实施方式a2,第三指令包括的用户选择的视频直播模式为固定视角模式,第二终端设备生成第二指示信息,第二指示信息包括第一观看视角,第二指示信息用于指示服务器下发第一观看视角对应的视频流。
步骤307,第二终端设备向服务器发送第一请求信息,第一请求信息用于请求第一车辆外部的直播视频。相应的,服务器接收来自第二终端设备的第一请求信息。
基于上述可能的实施方式a1中,步骤307中,第一请求信息可以包括第一指示信息,用于指示服务器根据直播视频的内容确定第一观看视角对应的视频流。相应的,服务器接收第一请求信息,根据第一请求信息携带的第一指示信息,为第二终端设备的用户推荐一个观看视角,即第一观看视角。如此,服务器可以为第一用户智能的推荐观看视角,在直播过程中,也可以智能的为第一用户切换观看视角。
基于上述可能的实施方式a2中,步骤307中,第一请求信息包括第二指示信息,第二指示信息用于指示第二观看视角,第二观看视角用于指示所述服务器根据所述第二观看视角下发视频流。也就是说,第二终端设备可以指示服务器下发固定观看视角对应的视频流,固定观看视角为第二观看视角。相应的,服务器接收第一请求信息,根据第一请求信息携带的第二指示信息,向第二终端设备的下发第二观看视角的视频流。如此,第一用户可以根据自己的喜好自己选择观看视角。
步骤308,服务器根据第一请求信息,获取全景视频流中的目标观看视角对应的视频流。
本申请实施例中,可实现上述步骤308的方式有多种,下面通过下述可能的实施方式b1、可能的实施方式b2、可能的实施方式b3、可能的实施方式b4和可能的实施方式b5来进行描述。
可能的实施方式b1,基于上述可能的实施方式a1,服务器根据第一请求信息,获取全景视频流中的目标观看视角对应的视频流,包括:服务器根据第一请求信息,对全景视频流中每个预设视角的视频流进行抽帧,获取每个预设视角对应的至少一帧图片,然后,将至少一帧图片的内容与服务器中存储的参考图片的内容进行匹配,确定出与参考图片的内容相匹配的预设视角对应的用户认可度,其中,参考图片包括的内容为用户认可度大于第一阈值的内容,服务器根据与参考图片的内容相匹配的预设视角对应的用户认可度,确定第一观看视角,然后获取全景视频流中的第一观看视角对应的视频流。
具体实施中,判断每个预设视角对应的至少一帧图片与参考图片的内容是否相匹配,可以通过判断每个预设视角对应的至少一帧图片与参考图片的内容之间的匹配度是否大于一定的阈值,这里的阈值可以根据实际需要进行设置,例如设置为90%,又例如设置为98%,此处不作限制。
示例性的,服务器中存储的参考图片有很多个,每个参考图片的内容也对应一个用户认可度,针对一个观看视角的视频流来说,该观看视角对应的至少一帧图片可能与一个参考图片相匹配,也可能与多个参考图片相匹配,相匹配的参考图片的数量越多,该观看视 角对应的用户认可度越高。
在实施方式b1中,第一观看视角可以为用户认可度最高的观看视角,也可以为用户认可度最高的n个观看视角中的一个。
可能的实施方式b2,基于上述可能的实施方式a1,服务器可以获取用户偏好数据,其中,用户偏好数据包括用户认可度大于第二阈值的视频流的内容以及每个视频流的内容所属的类别。服务器根据第一请求信息,获取全景视频流中的目标观看视角对应的视频流,具体包括:服务器根据第一请求信息,对全景视频流中每个预设视角的视频流进行分类得到每个预设视角的视频流的内容所属的类别,服务器根据用户偏好数据和每个预设视角的视频流的内容所属的类别,确定每个预设视角的视频流的用户认可度,服务器根据每个预设视角的视频流的用户认可度,确定第一观看视角,然后,获取全景视频流中的第一观看视角对应的视频流。
在一个示例中,第二终端设备向服务器发送用户偏好数据,用户偏好数据可以包括第一用户对历史观看过的各个视频流的内容所属的类别的用户认可度。
在另一个示例中,除第二终端设备之外的其它观众侧终端设备向服务器发送用户偏好数据,用户偏好数据也可以是其它观众侧终端设备的用户(下文称为第二用户)对历史观看过的各个视频流的内容所属的类别的用户认可度。
在又一个示例中,用户偏好数据也可以是第一用户和至少一个第二用户对历史观看过的各个视频流的内容所属的类别的用户认可度。
可能的实施方式b3,服务器接收来自第一终端设备的第二视频流,第二视频流为通过采集第一主播的直播视频得到,服务器根据第二视频流,确定第一主播的姿态特征和音频中的特征关键字,姿态特征包括表情和/或肢体动作。服务器根据第一请求信息,获取全景视频流中的目标观看视角对应的视频流,具体包括:服务器根据第一请求信息,从全景视频流中确定出与第一主播的姿态特征和音频中的特征关键字对应的观看视角,作为第一观看视角,然后服务器获取全景视频流中的第一观看视角对应的视频流。
示例的,第一主播在直播过程中,说道:请看右边的山,服务器可以识别出第一主播的音频中的特征关键字“看右边”,服务器可以结合音频中的特征关键字“看右边”,确定第一主播指示的观看角度,参见图4所示的xyz坐标系,第一车辆的右边对应的观看视角为:旋转度90°,仰角0°,服务器可以确定出第一主播指示的观看视角为旋转度90°,仰角0°,即确定第一观看视角为旋转度90°,仰角0°。
示例的,第一主播在直播过程中,第一主播的手指指向与第一车辆正前方的夹角为30度的方向,说道:“哇,那颗树的树叶真漂亮”,服务器可以识别出第一主播的音频中的特征关键字“那颗树”,第一主播的姿态特征为:手指指向与第一车辆正前方的夹角为30度的方向,参见图4所示的xyz坐标系,与第一车辆正前方的夹角为30度的方向对应的观看视角为旋转度30°,仰角0,服务器可以结合第一主播的肢体动作和音频中的特征关键字,确定观看视角为旋转度30°,仰角0°,即确定第一观看视角为旋转度90°,仰角0°。
在其它一些实施例中,第一车辆内部的第一主播在直播过程中,还可以切换为其他主播进行直播,示例的,第一终端设备可以接收第二指令,第二指令用于指示直播镜头从第一主播切换朝向第二主播,第一终端设备调整直播镜头朝向第二主播,采集第二主播直播的视频,得到第三视频流,并向服务器发送第三视频流。服务器接收第三视频流,可以将该第三视频流发送至第二终端设备。其中,第二指令可以为第一主播触发的,也可以为第 二主播触发,具体的触发方式可以为按键触发,也可以为语音触发。
可能的实施方式b4,基于上述可能的实施方式a1,服务器可以获取当前环境信息,服务器根据第一请求信息,获取全景视频流中的目标观看视角对应的视频流,包括:服务器根据第一请求信息,从全景视频流中确定出与当前环境信息相关的观看视角,作为第一观看视角,服务器获取全景视频流中的第一观看视角对应的视频流。
其中,当前环境信息可以包括以下内容中的任一项或多项:
第一车辆的当前位置信息、第一车辆所在位置处当前的天气信息、与第一车辆所在位置处的距离在预设范围内的地图兴趣点(point of interest,POI)信息。
示例的,第一终端设备可以获取第一车辆所在位置处当前的天气信息,例如当前气温-5℃,车辆右侧与正前方夹角45°的不远处有冰面,与该冰面相关的观看视角为旋转度45°,仰角0°,服务器确定第一观看视角为旋转度45°,仰角0°。
示例的,第一终端设备可以从高精地图云服务获取第一当前所在位置处周围的POI信息,例如网红景点,又例如标志性建筑名称等。例如,车辆右侧与正前方夹角90°的方向有网红景点,服务器确定第一观看视角为旋转度90°,仰角0°。
可能的实施方式b5,服务器根据第一请求信息,获取全景视频流中的目标观看视角对应的视频流,具体包括:服务器根据第一请求信息,统计全景视频流中各个观看视角对应的观看人数,从全景视频流中确定出观看人数最多的观看视角,作为第一观看视角,服务器获取全景视频流中的第一观看视角对应的视频流。
可能的实施方式b6,基于上述实施方式b1至b5中的任多种实施方式的组合,确定第一观看视角,例如,服务器可以结合实施方式b1中的参考图片的内容、实施方式b2中的用户偏好数据、实施方式b3中的第一主播的姿态特征和音频中的特征关键字对应的观看视角、实施方式b4中的当前环境信息、实施方式b5中的各个观看视角对应的观看人数中的任多种因素确定第一观看视角,具体方式参考上述可能的实施方式b1至b5,此处不再赘述。
下面结合具体的示例进行说明。
以全景视频流中的一个观看视角为例,根据该观看视角的视频流以及如下数据内容项,为每个观看视角进行评分,根据每个观看视角评分的最终结果确定第一观看视角。请参见下表1,为视角推荐评分表。
表1
Figure PCTCN2021126289-appb-000001
Figure PCTCN2021126289-appb-000002
如上表1,数据内容项主要包括视频抽帧的内容,目前在线观看人数,车内主播的指向性,环境因素等。以视频抽帧的内容为例进行说明,首先根据视频抽帧的内容判断是否为用户喜欢的内容,具体实施方式可参见上述可能的实施方式b1,若判断结果为是,可以根据用户喜爱程度评分,分值可为1~10分任一值,若判断结果为否,则为0分,这一项的最终得分计为S1,这一项所占的权重设为P1。依此类推,目前在线观看人数这一项的最终得分计为S2,这一项所占的权重设为P2。车内主播的指向性这一项的最终得分计为S3,这一项所占的权重设为P3。环境因素这一项的最终得分计为S4,这一项所占的权重设为P4。其中权重根据内部算法进行调优。
根据上述各项评分结果,计算每个观看视角的优先级分数S=S1*P1+S2*P2+S3*P3+S4*P4。根据各个观看视角的优先级分数,从高到低进行排序,向用户推荐优先级分数值最高的观看视角,对于优先级分数相同的观看视角,继续判断所占权重较高那一项的分数值,继续相同则按照内部的排序进行推荐。
可能的实施方式b7,基于上述可能的实施方式a2,服务器根据第一请求信息,获取全景视频流中的目标观看视角对应的视频流,具体包括:服务器根据第一请求信息,获取全景视频流中的第二观看视角对应的视频流。
步骤309,服务器向第二终端设备发送目标观看视角对应的视频流。
基于上述可能的实施方式b1至b6,步骤309,服务器向第二终端设备发送目标观看视角对应的视频流,具体包括:服务器向第二终端设备发送第一观看视角对应的视频流。相应的,第二终端设备接收来自服务器的第一观看视角对应的视频流。
基于上述可能的实施方式b7,步骤309,服务器向第二终端设备发送目标观看视角对应的视频流,具体包括:服务器向第二终端设备发送第二观看视角对应的视频流。相应的,第二终端设备接收来自服务器的第二观看视角对应的视频流。
步骤310,第二终端设备显示目标观看视角对应的视频流的视频内容。
基于上述可能的实施方式b1至b6,步骤310中,第二终端设备显示第一观看视角对应的视频流的视频内容。
基于上述可能的实施方式b7,步骤310中,第二终端设备显示第二观看视角对应的视频流的视频内容。
本申请实施例中,第二终端设备的用户选择智能视角模式时,服务器可以自动计算出目标观看视角,并推荐给用户。第二终端设备的用户选择固定视角模式时,服务器可以给第二终端设备发用户指定的观看视角的视频流。
在其它一些实施例中,所述还可以向所述第二终端设备发送所述第二视频流,第二终端设备接收来自服务器的第二视频流,可以通过悬浮窗口显示第二视频流的视频内容,也可以不显示第二视频流的视频内容,这样不会影响到第一用户观看第一视频流的视频内容, 可以提升用户体验。
在其它一些实施例中,第二终端设备的用户可以在观看直播的过程中切换观看视角,从而可以看到用户自己想看的视频内容。例如,用户通过触发第四指令,实现指示当前的观看视角切换为第二观看视角,该当前的观看视角可以为上述第一观看视角或者第二观看视角。第二终端设备可以接收第四指令,第二终端设备响应于第四指令,向服务器发送的第二请求信息,第二请求信息用于请求将当前的观看视角切换为第三观看视角;服务器向第二终端设备发送第三观看视角对应的视频流,相应的,第二终端设备接收服务器发送的第三观看视角对应的视频流。
本申请实施例中的术语“系统”和“网络”可被互换使用。“至少一个”是指一个或者多个,“多个”是指两个或两个以上。“和/或”,描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B的情况,其中A,B可以是单数或者复数。字符“/”一般表示前后关联对象是一种“或”的关系。“以下至少一项(个)”或其类似表达,是指的这些项中的任意组合,包括单项(个)或复数项(个)的任意组合。例如,a,b,或c中的至少一项(个),可以表示:a,b,c,a-b,a-c,b-c,或a-b-c,其中a,b,c可以是单个,也可以是多个。
以及,除非有特别说明,本申请实施例提及“第一”、“第二”等序数词是用于对多个对象进行区分,不用于限定多个对象的顺序、时序、优先级或者重要程度。例如,第一车辆和第二车辆,只是为了区分不同的车辆,而并不是表示这两个车辆的优先级或者重要程度等的不同。
需要说明的是,上述各个消息的名称仅仅是作为示例,随着通信技术的演变,上述任意消息均可能改变其名称,但不管其名称如何发生变化,只要其含义与本申请上述消息的含义相同,则均落入本申请的保护范围之内。
上述主要从各个网元之间交互的角度对本申请提供的方案进行了介绍。可以理解的是,上述实现各网元为了实现上述功能,其包含了执行各个功能相应的硬件结构和/或软件模块。本领域技术人员应该很容易意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,本发明能够以硬件或硬件和计算机软件的结合形式来实现。某个功能究竟以硬件还是计算机软件驱动硬件的方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本发明的范围。
根据前述方法,图5为本申请实施例提供的通信装置的结构示意图,如图5所示,该通信装置可以为第一终端设备、第二终端设备或服务器。也可以为芯片或电路,比如可设置于终端设备端的地图更新装置的芯片或电路,再比如可设置于第二地图更新装置内的芯片或电路。
进一步的,该通信装置1301还可以进一步包括总线系统,其中,处理器1302、存储器1304、收发器1303可以通过总线系统相连。
应理解,上述处理器1302可以是一个芯片。例如,该处理器1302可以是现场可编程门阵列(field programmable gate array,FPGA),可以是专用集成芯片(application specific integrated circuit,ASIC),还可以是系统芯片(system on chip,SoC),还可以是中央处理器(central processor unit,CPU),还可以是网络处理器(network processor,NP),还可以是数字信号处理电路(digital signal processor,DSP),还可以是微控制器(micro controller  unit,MCU),还可以是可编程控制器(programmable logic device,PLD)或其他集成芯片。
在实现过程中,上述方法的各步骤可以通过处理器1302中的硬件的集成逻辑电路或者软件形式的指令完成。结合本申请实施例所公开的方法的步骤可以直接体现为硬件处理器执行完成,或者用处理器1302中的硬件及软件模块组合执行完成。软件模块可以位于随机存储器,闪存、只读存储器,可编程只读存储器或者电可擦写可编程存储器、寄存器等本领域成熟的存储介质中。该存储介质位于存储器1304,处理器1302读取存储器1304中的信息,结合其硬件完成上述方法的步骤。
应注意,本申请实施例中的处理器1302可以是一种集成电路芯片,具有信号的处理能力。在实现过程中,上述方法实施例的各步骤可以通过处理器中的硬件的集成逻辑电路或者软件形式的指令完成。上述的处理器可以是通用处理器、数字信号处理器(DSP)、专用集成电路(ASIC)、现场可编程门阵列(FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件。可以实现或者执行本申请实施例中的公开的各方法、步骤及逻辑框图。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。结合本申请实施例所公开的方法的步骤可以直接体现为硬件译码处理器执行完成,或者用译码处理器中的硬件及软件模块组合执行完成。软件模块可以位于随机存储器,闪存、只读存储器,可编程只读存储器或者电可擦写可编程存储器、寄存器等本领域成熟的存储介质中。该存储介质位于存储器,处理器读取存储器中的信息,结合其硬件完成上述方法的步骤。
可以理解,本申请实施例中的存储器1304可以是易失性存储器或非易失性存储器,或可包括易失性和非易失性存储器两者。其中,非易失性存储器可以是只读存储器(read-only memory,ROM)、可编程只读存储器(programmable ROM,PROM)、可擦除可编程只读存储器(erasable PROM,EPROM)、电可擦除可编程只读存储器(electrically EPROM,EEPROM)或闪存。易失性存储器可以是随机存取存储器(random access memory,RAM),其用作外部高速缓存。通过示例性但不是限制性说明,许多形式的RAM可用,例如静态随机存取存储器(static RAM,SRAM)、动态随机存取存储器(dynamic RAM,DRAM)、同步动态随机存取存储器(synchronous DRAM,SDRAM)、双倍数据速率同步动态随机存取存储器(double data rate SDRAM,DDR SDRAM)、增强型同步动态随机存取存储器(enhanced SDRAM,ESDRAM)、同步连接动态随机存取存储器(synchlink DRAM,SLDRAM)和直接内存总线随机存取存储器(direct rambus RAM,DR RAM)。应注意,本文描述的系统和方法的存储器旨在包括但不限于这些和任意其它适合类型的存储器。
该通信装置1301对应上述方法中的服务器的情况下,该通信装置可以包括处理器1302、收发器1303和存储器1304。该存储器1304用于存储指令,该处理器1302用于执行该存储器1304存储的指令,以实现如上图3中所示的任一项或任多项对应的方法中服务器的相关方案。
收发器1303,用于接收来自第一终端设备的至少一条第一视频流,所述第一视频流为通过采集第一车辆外部的任一方向的直播视频得到;
处理器1302,用于对所述至少一条第一视频流进行处理,得到全景视频流。
在一种可能地实施方式中,收发器1303,还用于接收来自第二终端设备的第一请求信息,所述第一请求信息用于请求所述第一车辆外部的直播视频;处理器1302,还用于根据所述第一请求信息,获取所述全景视频流中的目标观看视角对应的视频流;收发器1303, 还用于向所述第二终端设备发送所述目标观看视角对应的视频流。
在一种可能地实施方式中,所述第一请求信息包括第一指示信息,所述第一指示信息用于指示所述服务器根据直播视频的内容确定第一观看视角对应的视频流。
在一种可能地实施方式中,处理器1302,用于根据所述第一指示信息,对所述全景视频流中每个预设视角的视频流进行抽帧,获取每个所述预设视角对应的至少一帧图片;将所述至少一帧图片的内容与所述服务器中存储的参考图片的内容进行匹配,确定出与所述参考图片的内容相匹配的预设视角对应的用户认可度;所述参考图片包括的内容为用户认可度大于第一阈值的内容;根据与所述参考图片的内容相匹配的预设视角对应的用户认可度,确定所述第一观看视角;收发器1303,还用于获取所述全景视频流中的所述第一观看视角对应的视频流。
在一种可能地实施方式中,收发器1303,还用于所述服务器获取用户偏好数据,所述用户偏好数据包括用户认可度大于第二阈值的视频流的内容以及每个视频流的内容所属的类别。
在一种可能地实施方式中,处理器1302,用于根据所述第一请求信息,对所述全景视频流中每个预设视角的视频流进行分类得到每个预设视角的视频流的内容所属的类别;根据所述用户偏好数据和每个预设视角的视频流的内容所属的类别,确定每个预设视角的视频流的用户认可度;根据每个预设视角的视频流的用户认可度,确定所述第一观看视角;获取所述全景视频流中的所述第一观看视角对应的视频流。
在一种可能地实施方式中,收发器1303,用于接收来自所述第一终端设备的第二视频流,所述第二视频流为通过采集第一主播的直播视频得到。
在一种可能地实施方式中,处理器1302,还用于根据所述第二视频流,确定所述第一主播的姿态特征和音频中的特征关键字;所述姿态特征包括表情和/或肢体动作;根据所述第一请求信息,从所述全景视频流中确定出与所述第一主播的姿态特征和音频中的特征关键字对应的观看视角,作为所述第一观看视角;
所述服务器获取所述全景视频流中的所述第一观看视角对应的视频流。
在一种可能地实施方式中,处理器1302,还用于获取当前环境信息;其中,所述当前环境信息包括以下至少一项:所述第一车辆的当前位置信息、所述第一车辆所在位置处的当前天气信息、与所述第一车辆所在位置处的距离在预设范围内的地图兴趣点POI信息;根据所述第一请求信息,从所述全景视频流中确定出与所述当前环境信息相关的观看视角,作为第一观看视角;获取所述全景视频流中的所述第一观看视角对应的视频流。
在一种可能地实施方式中,所述第一请求信息包括第二指示信息,所述第二指示信息用于指示第二观看视角,所述第二观看视角用于指示所述服务器根据所述第二观看视角下发视频流。
在一种可能地实施方式中,处理器1302,用于所述服务器根据所述第一请求信息,获取所述全景视频流中的所述第二观看视角对应的视频流;收发器1303,用于向所述第二终端设备发送所述第二观看视角对应的视频流。
在一种可能地实施方式中,收发器1303,用于接收来自所述第二终端设备发送的第二请求信息;所述第二请求信息用于请求将当前的观看视角切换为第三观看视角;向所述第二终端设备发送所述第三观看视角对应的视频流。
在一种可能地实施方式中,收发器1303,用于向所述第二终端设备发送所述第二视频 流。
在一种可能地实施方式中,收发器1303,还用于接收来自所述第一终端设备的第三视频流,所述第三视频流为通过采集第二主播的直播视频得到。
当通信装置1301为上述第一终端设备,收发器1303,用于第一终端设备接收第一指令;所述第一指令用于指示启动视频直播;处理器1302,用于根据所述第一指令,启动视频直播功能;获取第一车辆外部各个方向的直播视频,得到至少一条第一视频流;收发器1303,还用于向服务器发送所述至少一条第一视频流。
在一种可能地实施方式中,处理器1302,还用于获取第一主播的直播视频,得到第二视频流;收发器1303,还用于向所述服务器发送所述第二视频流。
在一种可能地实施方式中,收发器1303,还用于接收第二指令;所述第二指令用于指示直播镜头从所述第一主播切换朝向第二主播;收发器1303,还用于调整所述直播镜头朝向所述第二主播,采集所述第二主播的直播视频,得到第三视频流;收发器1303,还用于向所述服务器发送所述第三视频流。
当通信装置1301为上述第二终端设备,处理器1302,用于获取地图元素的第一位置信息;获取第一误差信息,第一误差信息包括第一车辆的定位系统对第一车辆进行定位的误差信息;收发器1303,用于向第二地图更新装置发送地图元素的第一位置信息和第一误差信息。
在一种可能地实施方式中,收发器1303,用于接收第三指令,所述第三指令用于指示获取第一车辆的直播视频;向服务器发送第一请求信息,所述第一请求信息用于请求所述第一车辆的直播视频;接收来自所述服务器的目标观看视角对应的视频流。
在一种可能地实施方式中,所述第一请求信息包括第一指示信息,所述第一指示信息用于指示所述服务器根据直播视频的内容确定第一观看视角对应的视频流;收发器1303,用于接收来自所述服务器的第一观看视角对应的视频流。
在一种可能地实施方式中,所述第一请求信息包括第二指示信息,所述第二指示信息用于指示第二观看视角,所述第二观看视角用于指示所述服务器根据所述第二观看视角下发视频流;收发器1303,用于接收来自所述服务器的所述第二观看视角对应的视频流。
在一种可能地实施方式中,收发器1303,用于收发器1303,还用于向所述服务器发送用户偏好数据,所述用户偏好数据包括所述第二终端设备的用户观看视频流的数据。
在一种可能地实施方式中,收发器1303,用于向所述服务器发送第二请求信息,所述第二请求信息用于请求将当前的观看视角切换为第三观看视角;接收所述服务器发送的所述第三观看视角对应的视频流。
在一种可能地实施方式中,收发器1303,用于接收所述服务器发送的第二视频流,所述第二视频流为采集第一主播直播的视频得到的;所述处理器1302,用于所述第二终端设备显示悬浮窗口,所述悬浮窗口用于显示所述第二视频流。
该通信装置所涉及的与本申请实施例提供的技术方案相关的概念,解释和详细说明及其他步骤请参见前述方法或其他实施例中关于这些内容的描述,此处不做赘述。
根据前述方法,图6为本申请实施例提供的通信装置的结构示意图,如图6所示,通信装置1401可以包括通信接口1403、处理器1402和存储器1404。通信接口1403,用于 输入和/或输出信息;处理器1402,用于执行计算机程序或指令,使得通信装置1401实现上述图3相关方案中第一终端设备的方法,或使得通信装置1401实现上述图3的相关方案中第二终端设备的方法,或使得通信装置1401实现上述图3的相关方案中服务器的方法。本申请实施例中,通信接口1403可以实现上述图5的收发器1303所实现的方案,处理器1402可以实现上述图5的处理器1302所实现的方案,存储器1404可以实现上述图5的存储器1304所实现的方案,在此不再赘述。
基于以上实施例以及相同构思,图7为本申请实施例提供的通信装置的示意图,如图7所示,该通信装置1501可以为第一终端设备或第二终端设备或第三终端设备,也可以为芯片或电路,比如可设置于第一终端设备或第二终端设备或第三终端设备的芯片或电路。
该通信装置可以对应上述方法中的第一终端设备。该通信装置可以实现如上图3中所示的任一项或任多项对应的方法中第一终端设备所执行的步骤。该通信装置可以包括处理单元1502和通信单元1503。通信单元1503用于第一终端设备接收第一指令;所述第一指令用于指示启动视频直播;处理单元1502用于根据所述第一指令,启动视频直播功能;获取第一车辆外部各个方向的直播视频,得到至少一条第一视频流;收发器1303,还用于向服务器发送所述至少一条第一视频流。
该通信装置1501对应上述方法中的第二终端设备的情况下,处理单元1502,用于获取地图元素的第一位置信息;获取第一误差信息,第一误差信息包括第一车辆的定位系统对第一车辆进行定位的误差信息;通信单元1503,用于向第二地图更新装置发送地图元素的第一位置信息和第一误差信息。
当通信装置1501对应上述服务器的情况下,通信单元1503,用于接收来自第一终端设备的至少一条第一视频流,所述第一视频流为通过采集第一车辆外部的任一方向的直播视频得到;
处理单元1502,对所述至少一条第一视频流进行处理,得到全景视频流。
该通信装置所涉及的与本申请实施例提供的技术方案相关的概念,解释和详细说明及其他步骤请参见前述方法或其他实施例中关于这些内容的描述,此处不做赘述。
可以理解的是,上述通信装置1501中各个单元的功能可以参考相应方法实施例的实现,此处不再赘述。
应理解,以上通信装置的单元的划分仅仅是一种逻辑功能的划分,实际实现时可以全部或部分集成到一个物理实体上,也可以物理上分开。本申请实施例中,通信单元1503可以由上述图5的收发器1303实现,处理单元1502可以由上述图5的处理器1302实现。
根据本申请实施例提供的方法,本申请还提供一种计算机程序产品,该计算机程序产品包括:计算机程序代码或指令,当该计算机程序代码或指令在计算机上运行时,使得该计算机执行图3所示实施例中任意一个实施例的方法。
根据本申请实施例提供的方法,本申请还提供一种计算机可读存储介质,该计算机可读介质存储有程序代码,当该程序代码在计算机上运行时,使得该计算机执行图3所示实施例中任意一个实施例的方法。
根据本申请实施例提供的方法,本申请还提供一种芯片系统,该芯片系统可以包括处理器。该处理器与存储器耦合,可用于执行图3所示实施例中任意一个实施例的方法。可选地,该芯片系统还包括存储器。存储器,用于存储计算机程序(也可以称为代码,或指 令)。处理器,用于从存储器调用并运行计算机程序,使得安装有芯片系统的设备执行图3所示实施例中任意一个实施例的方法。
根据本申请实施例提供的方法,本申请还提供一种系统,其包括前述的一个或多个车辆以及服务器端的地图更新装置,车辆中设置有地图更新装置。
在上述实施例中,可以全部或部分地通过软件、硬件、固件或者其任意组合来实现。当使用软件实现时,可以全部或部分地以计算机程序产品的形式实现。计算机程序产品包括一个或多个计算机指令。在计算机上加载和执行计算机指令时,全部或部分地产生按照本申请实施例的流程或功能。计算机可以是通用计算机、专用计算机、计算机网络、或者其他可编程装置。计算机指令可以存储在计算机可读存储介质中,或者从一个计算机可读存储介质向另一个计算机可读存储介质传输,例如,计算机指令可以从一个网站站点、计算机、服务器或数据中心通过有线(例如同轴电缆、光纤、数字用户线(digital subscriber line,DSL))或无线(例如红外、无线、微波等)方式向另一个网站站点、计算机、服务器或数据中心进行传输。计算机可读存储介质可以是计算机能够存取的任何可用介质或者是包含一个或多个可用介质集成的服务器、数据中心等数据存储设备。可用介质可以是磁性介质(例如,软盘、硬盘、磁带)、光介质(例如,高密度数字视频光盘(digital video disc,DVD))、或者半导体介质(例如,固态硬盘(solid state disc,SSD))等。
需要指出的是,本专利申请文件的一部分包含受著作权保护的内容。除了对专利局的专利文件或记录的专利文档内容制作副本以外,著作权人保留著作权。
上述各个装置实施例中第二地图更新装置与终端设备端的地图更新装置和方法实施例中的第二地图更新装置或终端设备端的地图更新装置对应,由相应的模块或单元执行相应的步骤,例如通信单元(收发器)执行方法实施例中接收或发送的步骤,除发送、接收外的其它步骤可以由处理单元(处理器)执行。具体单元的功能可以参考相应的方法实施例。其中,处理器可以为一个或多个。
在本说明书中使用的术语“部件”、“模块”、“系统”等用于表示计算机相关的实体、硬件、固件、硬件和软件的组合、软件、或执行中的软件。例如,部件可以是但不限于,在处理器上运行的进程、处理器、对象、可执行文件、执行线程、程序和/或计算机。通过图示,在计算设备上运行的应用和计算设备都可以是部件。一个或多个部件可驻留在进程和/或执行线程中,部件可位于一个计算机上和/或分布在两个或更多个计算机之间。此外,这些部件可从在上面存储有各种数据结构的各种计算机可读介质执行。部件可例如根据具有一个或多个数据分组(例如来自与本地系统、分布式系统和/或网络间的另一部件交互的二个部件的数据,例如通过信号与其它系统交互的互联网)的信号通过本地和/或远程进程来通信。
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各种说明性逻辑块(illustrative logical block)和步骤(step),能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统、装置和单元的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。
在本申请所提供的几个实施例中,应该理解到,所揭露的系统、装置和方法,可以通 过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。
作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。
另外,在本申请各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。
功能如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器等)执行本申请各个实施例方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(read-only memory,ROM)、随机存取存储器(random access memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。
以上,仅为本申请的具体实施方式,但本申请的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本申请揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本申请的保护范围之内。因此,本申请的保护范围应以所述权利要求的保护范围为准。

Claims (27)

  1. 一种车载直播方法,其特征在于,包括:
    服务器接收来自第一终端设备的至少一条第一视频流,所述第一视频流为通过采集第一车辆外部的任一方向的直播视频得到;
    所述服务器对所述至少一条第一视频流进行处理,得到全景视频流。
  2. 如权利要求1所述的方法,其特征在于,所述方法还包括:
    所述服务器接收来自第二终端设备的第一请求信息,所述第一请求信息用于请求所述第一车辆外部的直播视频;
    所述服务器根据所述第一请求信息,获取所述全景视频流中的目标观看视角对应的视频流;
    所述服务器向所述第二终端设备发送所述目标观看视角对应的视频流。
  3. 如权利要求2所述的方法,其特征在于,所述第一请求信息包括第一指示信息,所述第一指示信息用于指示所述服务器根据直播视频的内容确定第一观看视角对应的视频流。
  4. 如权利要求3所述的方法,其特征在于,所述服务器根据所述第一请求信息,获取所述全景视频流中的目标观看视角对应的视频流,包括:
    所述服务器根据所述第一指示信息,对所述全景视频流中每个预设视角的视频流进行抽帧,获取每个所述预设视角对应的至少一帧图片;
    所述服务器将所述至少一帧图片的内容与所述服务器中存储的参考图片的内容进行匹配,确定出与所述参考图片的内容相匹配的预设视角对应的用户认可度;所述参考图片包括的内容为用户认可度大于第一阈值的内容;
    所述服务器根据与所述参考图片的内容相匹配的预设视角对应的用户认可度,确定所述第一观看视角;
    所述服务器获取所述全景视频流中的所述第一观看视角对应的视频流。
  5. 如权利要求3所述的方法,其特征在于,所述方法还包括:
    所述服务器获取用户偏好数据,所述用户偏好数据包括用户认可度大于第二阈值的视频流的内容以及每个视频流的内容所属的类别。
  6. 如权利要求5所述的方法,其特征在于,所述服务器根据所述第一请求信息,获取所述全景视频流中的目标观看视角对应的视频流,包括:
    所述服务器根据所述第一请求信息,对所述全景视频流中每个预设视角的视频流进行分类得到每个预设视角的视频流的内容所属的类别;
    所述服务器根据所述用户偏好数据和每个预设视角的视频流的内容所属的类别,确定每个预设视角的视频流的用户认可度;
    所述服务器根据每个预设视角的视频流的用户认可度,确定所述第一观看视角;
    所述服务器获取所述全景视频流中的所述第一观看视角对应的视频流。
  7. 如权利要求3所述的方法,其特征在于,所述方法还包括:
    所述服务器接收来自所述第一终端设备的第二视频流,所述第二视频流为通过采集第一主播的直播视频得到。
  8. 如权利要求7所述的方法,其特征在于,所述方法包括:
    所述服务器根据所述第二视频流,确定所述第一主播的姿态特征和音频中的特征关键字;所述姿态特征包括表情和/或肢体动作;
    所述服务器根据所述第一请求信息,获取所述全景视频流中的目标观看视角对应的视频流,包括:
    所述服务器根据所述第一请求信息,从所述全景视频流中确定出与所述第一主播的姿态特征和音频中的特征关键字对应的观看视角,作为所述第一观看视角;
    所述服务器获取所述全景视频流中的所述第一观看视角对应的视频流。
  9. 如权利要求3所述的方法,其特征在于,所述方法还包括:
    所述服务器获取当前环境信息;其中,所述当前环境信息包括以下至少一项:
    所述第一车辆的当前位置信息、所述第一车辆所在位置处的当前天气信息、与所述第一车辆所在位置处的距离在预设范围内的地图兴趣点POI信息;
    所述服务器根据所述第一请求信息,获取所述全景视频流中的目标观看视角对应的视频流,包括:
    所述服务器根据所述第一请求信息,从所述全景视频流中确定出与所述当前环境信息相关的观看视角,作为第一观看视角;
    所述服务器获取所述全景视频流中的所述第一观看视角对应的视频流。
  10. 如权利要求2所述的方法,其特征在于,所述第一请求信息包括第二指示信息,所述第二指示信息用于指示第二观看视角,所述第二观看视角用于指示所述服务器根据所述第二观看视角下发视频流。
  11. 如权利要求10所述的方法,其特征在于,所述服务器根据所述第一请求信息,获取所述全景视频流中的目标观看视角对应的视频流,包括:
    所述服务器根据所述第一请求信息,获取所述全景视频流中的所述第二观看视角对应的视频流;
    所述服务器向所述第二终端设备发送所述目标观看视角对应的视频流,包括:
    所述服务器向所述第二终端设备发送所述第二观看视角对应的视频流。
  12. 如权利要求1-11任一项所述的方法,其特征在于,还包括:
    所述服务器接收来自所述第二终端设备发送的第二请求信息;所述第二请求信息用于请求将当前的观看视角切换为第三观看视角;
    所述服务器向所述第二终端设备发送所述第三观看视角对应的视频流。
  13. 如权利要求7或8所述的方法,其特征在于,还包括:
    所述服务器向所述第二终端设备发送所述第二视频流。
  14. 如权利要求1-13任一项所述的方法,其特征在于,还包括:
    所述服务器接收来自所述第一终端设备的第三视频流,所述第三视频流为通过采集第二主播的直播视频得到。
  15. 一种车载直播方法,其特征在于,包括:
    第一终端设备接收第一指令;所述第一指令用于指示启动视频直播;
    所述第一终端设备根据所述第一指令,启动视频直播功能;
    所述第一终端设备获取第一车辆外部各个方向的直播视频,得到至少一条第一视频流;
    所述第一终端设备向服务器发送所述至少一条第一视频流。
  16. 如权利要求15所述的方法,其特征在于,所述第一终端设备根据所述第一指令, 启动视频直播功能之后,还包括:
    所述第一终端设备获取第一主播的直播视频,得到第二视频流;
    所述第一终端设备向所述服务器发送所述第二视频流。
  17. 如权利要求16所述的方法,其特征在于,所述第一终端设备通过第二视频采集装置采集第一主播直播的视频,得到第二视频流之后,还包括:
    所述第一终端设备接收第二指令;所述第二指令用于指示直播镜头从所述第一主播切换朝向第二主播;
    所述第一终端设备调整所述直播镜头朝向所述第二主播,采集所述第二主播的直播视频,得到第三视频流;
    所述第一终端设备向所述服务器发送所述第三视频流。
  18. 一种车载直播方法,其特征在于,包括:
    第二终端设备接收第三指令,所述第三指令用于指示获取第一车辆的直播视频;
    所述第二终端设备向服务器发送第一请求信息,所述第一请求信息用于请求所述第一车辆的直播视频;
    所述第二终端设备接收来自所述服务器的目标观看视角对应的视频流。
  19. 如权利要求18所述的方法,其特征在于,所述第一请求信息包括第一指示信息,所述第一指示信息用于指示所述服务器根据直播视频的内容确定第一观看视角对应的视频流;
    所述第二终端设备接收来自所述服务器的目标观看视角对应的视频流,包括:
    所述第二终端设备接收来自所述服务器的第一观看视角对应的视频流。
  20. 如权利要求18所述的方法,其特征在于,所述第一请求信息包括第二指示信息,所述第二指示信息用于指示第二观看视角,所述第二观看视角用于指示所述服务器根据所述第二观看视角下发视频流;
    所述第二终端设备接收来自所述服务器的目标观看视角对应的视频流,包括:
    所述第二终端设备接收来自所述服务器的所述第二观看视角对应的视频流。
  21. 如权利要求18-20任一项所述的方法,其特征在于,所述方法还包括:
    所述第二终端设备向所述服务器发送用户偏好数据,所述用户偏好数据包括所述第二终端设备的用户观看视频流的数据。
  22. 如权利要求18-21任一项所述的方法,其特征在于,所述方法还包括:
    所述第二终端设备向所述服务器发送第二请求信息,所述第二请求信息用于请求将当前的观看视角切换为第三观看视角;
    所述第二终端设备接收来自所述服务器的目标观看视角对应的视频流,包括:
    所述第二终端设备接收所述服务器发送的所述第三观看视角对应的视频流。
  23. 如权利要求18-22任一项所述的方法,其特征在于,所述方法还包括:
    所述第二终端设备接收所述服务器发送的第二视频流,所述第二视频流为采集第一主播直播的视频得到的;
    所述第二终端设备显示悬浮窗口,所述悬浮窗口用于显示所述第二视频流。
  24. 一种服务器,其特征在于,包括处理器和存储器,包括处理器和存储器,所述存储器用于存储计算机执行指令,所述服务器运行时,所述处理器执行所述存储器中的计算机执行指令以利用所述服务器中的硬件资源执行权利要求1至14中任一所述方法的操作步 骤。
  25. 一种终端设备,其特征在于,包括处理器和存储器,包括处理器和存储器,所述存储器用于存储计算机执行指令,所述服务器运行时,所述处理器执行所述存储器中的计算机执行指令以利用所述服务器中的硬件资源执行权利要求15至17中任一所述方法的操作步骤,或者执行权利要求18至23中任一所述方法的操作步骤。
  26. 一种服务器,其特征在于,
    所述通信接口,用于输入和/或输出信息;
    所述处理器,用于执行计算机程序,使得权利要求1至14中任一项所述的方法被执行。
  27. 一种终端设备,其特征在于,
    所述通信接口,用于输入和/或输出信息;
    所述处理器,用于执行计算机程序,使得权利要求15至17中任一项所述的方法被执行,或者使得权利要求18至23中任一项所述的方法被执行。
PCT/CN2021/126289 2021-02-04 2021-10-26 一种车载直播方法及装置 WO2022166263A1 (zh)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP21924260.9A EP4287623A4 (en) 2021-02-04 2021-10-26 VEHICLE MOUNTED LIVE STREAMING METHOD AND APPARATUS

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110156289.6 2021-02-04
CN202110156289.6A CN114866789A (zh) 2021-02-04 2021-02-04 一种车载直播方法及装置

Publications (1)

Publication Number Publication Date
WO2022166263A1 true WO2022166263A1 (zh) 2022-08-11

Family

ID=82623378

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/126289 WO2022166263A1 (zh) 2021-02-04 2021-10-26 一种车载直播方法及装置

Country Status (3)

Country Link
EP (1) EP4287623A4 (zh)
CN (1) CN114866789A (zh)
WO (1) WO2022166263A1 (zh)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117156258B (zh) * 2023-10-24 2024-01-23 北京优贝在线网络科技有限公司 一种基于全景直播的多视角自切换系统

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130278732A1 (en) * 2012-04-24 2013-10-24 Mobitv, Inc. Control of perspective in multi-dimensional media
CN106657733A (zh) * 2016-11-25 2017-05-10 深圳市元征科技股份有限公司 一种基于无人机的全景直播方法及终端
CN106803966A (zh) * 2016-12-31 2017-06-06 北京星辰美豆文化传播有限公司 一种多人网络直播方法、装置及其电子设备
CN107835435A (zh) * 2017-06-09 2018-03-23 越野族(北京)传媒科技有限公司 一种赛事宽视角直播设备和相关联的直播系统和方法
US20180146216A1 (en) * 2016-11-18 2018-05-24 Twitter, Inc. Live interactive video streaming using one or more camera devices
CN110798692A (zh) * 2019-09-27 2020-02-14 咪咕视讯科技有限公司 一种视频直播方法、服务器及存储介质
CN111277904A (zh) * 2020-02-17 2020-06-12 腾讯科技(深圳)有限公司 一种视频的播放控制方法、装置及计算设备
CN111355966A (zh) * 2020-03-05 2020-06-30 上海乐杉信息技术有限公司 环绕自由视角直播方法及系统
CN111726649A (zh) * 2020-06-28 2020-09-29 百度在线网络技术(北京)有限公司 视频流处理方法、装置、计算机设备和介质

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10015551B2 (en) * 2014-12-25 2018-07-03 Panasonic Intellectual Property Management Co., Ltd. Video delivery method for delivering videos captured from a plurality of viewpoints, video reception method, server, and terminal device
US10531048B2 (en) * 2016-12-15 2020-01-07 Motorola Solutions, Inc. System and method for identifying a person, object, or entity (POE) of interest outside of a moving vehicle

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130278732A1 (en) * 2012-04-24 2013-10-24 Mobitv, Inc. Control of perspective in multi-dimensional media
US20180146216A1 (en) * 2016-11-18 2018-05-24 Twitter, Inc. Live interactive video streaming using one or more camera devices
CN106657733A (zh) * 2016-11-25 2017-05-10 深圳市元征科技股份有限公司 一种基于无人机的全景直播方法及终端
CN106803966A (zh) * 2016-12-31 2017-06-06 北京星辰美豆文化传播有限公司 一种多人网络直播方法、装置及其电子设备
CN107835435A (zh) * 2017-06-09 2018-03-23 越野族(北京)传媒科技有限公司 一种赛事宽视角直播设备和相关联的直播系统和方法
CN110798692A (zh) * 2019-09-27 2020-02-14 咪咕视讯科技有限公司 一种视频直播方法、服务器及存储介质
CN111277904A (zh) * 2020-02-17 2020-06-12 腾讯科技(深圳)有限公司 一种视频的播放控制方法、装置及计算设备
CN111355966A (zh) * 2020-03-05 2020-06-30 上海乐杉信息技术有限公司 环绕自由视角直播方法及系统
CN111726649A (zh) * 2020-06-28 2020-09-29 百度在线网络技术(北京)有限公司 视频流处理方法、装置、计算机设备和介质

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP4287623A4 *

Also Published As

Publication number Publication date
CN114866789A (zh) 2022-08-05
EP4287623A1 (en) 2023-12-06
EP4287623A4 (en) 2024-05-01

Similar Documents

Publication Publication Date Title
US10820038B2 (en) Method and apparatus for augmenting media services
CN111083364B (zh) 一种控制方法、电子设备、计算机可读存储介质、芯片
US9576394B1 (en) Leveraging a multitude of dynamic camera footage to enable a user positional virtual camera
CN106576184B (zh) 信息处理装置、显示装置、信息处理方法、程序和信息处理系统
US10659685B2 (en) Control of viewing angles for 360-degree video playback
WO2017134706A1 (ja) 映像表示方法及び映像表示装置
US11039109B2 (en) System and method for adjusting an image for a vehicle mounted camera
JP2021524626A (ja) 車両の搭乗者及び遠隔ユーザのための共有環境
US20120213212A1 (en) Life streaming
CA3087256A1 (en) Camera enhanced ride sharing
JPWO2018030206A1 (ja) カメラワーク生成方法及び映像処理装置
US20170264822A1 (en) Mounting Device for Portable Multi-Stream Video Recording Device
JP6944138B2 (ja) 画像処理装置および画像処理方法
JP2017139725A (ja) 映像表示方法及び映像表示装置
US20180103197A1 (en) Automatic Generation of Video Using Location-Based Metadata Generated from Wireless Beacons
WO2012122046A1 (en) Eyeglasses with integrated camera for video streaming
US10156898B2 (en) Multi vantage point player with wearable display
US20160073013A1 (en) Handheld multi vantage point player
KR20150080003A (ko) 모션 패럴랙스를 이용한 2d 이미지로부터의 3d 지각 생성
WO2022166263A1 (zh) 一种车载直播方法及装置
CN106231152A (zh) 传输图片的方法及装置
CN112333458B (zh) 直播房间显示方法、装置、设备及存储介质
AU2019271924A1 (en) System and method for adjusting an image for a vehicle mounted camera
CN107317952B (zh) 一种视频图像处理方法及基于电子地图的拼图方法
US11394882B2 (en) Display control device, display control method, and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21924260

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2021924260

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 2021924260

Country of ref document: EP

Effective date: 20230831

NENP Non-entry into the national phase

Ref country code: DE