CN117528125A - Video playing method, device, computer equipment and storage medium - Google Patents

Video playing method, device, computer equipment and storage medium Download PDF

Info

Publication number
CN117528125A
CN117528125A CN202311491230.8A CN202311491230A CN117528125A CN 117528125 A CN117528125 A CN 117528125A CN 202311491230 A CN202311491230 A CN 202311491230A CN 117528125 A CN117528125 A CN 117528125A
Authority
CN
China
Prior art keywords
camera
target object
video data
playing
cameras
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311491230.8A
Other languages
Chinese (zh)
Inventor
罗绮琪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Priority to CN202311491230.8A priority Critical patent/CN117528125A/en
Publication of CN117528125A publication Critical patent/CN117528125A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/21805Source of audio or video content, e.g. local disk arrays enabling multiple viewpoints, e.g. using a plurality of cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/262Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/437Interfacing the upstream path of the transmission network, e.g. for transmitting client requests to a VOD server
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Human Computer Interaction (AREA)
  • Studio Devices (AREA)

Abstract

The application provides a video playing method, a video playing device, computer equipment and a storage medium, and belongs to the technical field of computers. The method comprises the following steps: acquiring a plurality of video data shot by a plurality of cameras; determining a correlation coefficient between each of the plurality of cameras and the target object based on the plurality of video data and information of the target object; sequencing the cameras according to the sequence of the correlation coefficient from big to small to obtain a camera sequence; and sending the camera sequence to a terminal. According to the technical scheme, the user can know the condition of the target object in multiple directions, and can watch the content shot by the camera for shooting the target object more clearly, so that the target object can be watched more clearly, manual operation of the user is not needed, the man-machine interaction efficiency is improved, and further the watching experience of the user is improved.

Description

Video playing method, device, computer equipment and storage medium
Technical Field
The present disclosure relates to the field of computer technologies, and in particular, to a video playing method, a video playing device, a computer device, and a storage medium.
Background
With the development of computer technology, the personalized appeal of the user's viewing is getting stronger. The current users hope to watch the scene in multiple directions and angles, and have extremely high viewing appeal for a certain person object.
Currently, live broadcast programs provided by a video platform, such as a variety, a show-choosing total jeopardy and the like, are provided on site with a plurality of cameras for shooting at the same time, and shot contents are fed back to a 'center console' on site and are uniformly managed and controlled by a program director. Thus, only one camera can be used to capture the content of the user terminal, regardless of the main camera or the current multi-camera switching function. That is, the user passively accepts the program picture transmitted by the central console, and cannot select the object of interest on site by himself or herself, and cannot closely watch a certain target object in real time, so that the man-machine interaction efficiency is low, and the watching experience of the user is affected.
Disclosure of Invention
The embodiment of the application provides a video playing method, a video playing device, computer equipment and a storage medium, so that a user can know the condition of a target object in multiple directions, can watch the target object more clearly, does not need manual operation of the user, improves the man-machine interaction efficiency, and further improves the watching experience of the user. The technical scheme is as follows:
In one aspect, a video playing method is provided, and the method includes:
acquiring a plurality of video data shot by a plurality of cameras, wherein the plurality of cameras are in one-to-one correspondence with the plurality of video data;
determining a correlation coefficient between each camera in the plurality of cameras and the target object based on the plurality of video data and the information of the target object, wherein the correlation coefficient is used for representing the correlation degree between the content shot by the corresponding camera and the target object;
sequencing the cameras according to the sequence of the correlation coefficient from big to small to obtain a camera sequence;
and sending the camera sequence to a terminal, wherein the camera sequence is used for indicating the sizes of playing areas corresponding to the cameras in the terminal, each playing area is used for playing video data shot by the corresponding camera, and the sizes of the playing areas are positively correlated with the front and back of the camera sequencing in the camera sequence.
In another aspect, a video playing method is provided, the method including:
acquiring information of a target object based on a video playing interface, wherein the video playing interface is used for playing video data shot by a plurality of cameras, and the target object is a locked object;
Transmitting information of the target object to a server;
receiving a camera sequence returned by the server, wherein the cameras in the camera sequence are arranged according to the sequence of the association coefficients from large to small, and the association coefficients are used for representing the correlation degree of the content shot by the corresponding camera and the target object;
and displaying playing areas corresponding to the cameras in the video playing interface based on the camera sequence, wherein each playing area is used for playing video data shot by the corresponding camera, and the size of the playing area is positively correlated with the front and back of the camera sequencing in the camera sequence.
In another aspect, there is provided a video playing device, the device comprising:
the first acquisition module is used for acquiring a plurality of video data shot by a plurality of cameras, and the plurality of cameras are in one-to-one correspondence with the plurality of video data;
a determining module, configured to determine, based on the plurality of video data and information of a target object, a correlation coefficient between each camera of the plurality of cameras and the target object, where the correlation coefficient is used to represent a degree of correlation between content captured by a corresponding camera and the target object;
The sequencing module is used for sequencing the cameras according to the sequence of the correlation coefficients from large to small to obtain a camera sequence;
the sending module is used for sending the camera sequence to the terminal, the camera sequence is used for indicating the sizes of playing areas corresponding to the cameras in the terminal, each playing area is used for playing video data shot by the corresponding camera, and the sizes of the playing areas are positively correlated with the front and rear of the camera sequencing in the camera sequence.
In some embodiments, the determining module includes:
a screening unit, configured to screen at least one target video data from the plurality of video data based on the information of the target object, where the target video data includes the target object;
and the determining unit is used for determining the association coefficient between the camera corresponding to the target video data and the target object based on the data amount occupied by the target object in the target video data for any one of the at least one target video data.
In some embodiments, the determining unit includes:
an obtaining subunit, configured to obtain, for any one target video data of the at least one target video data, a number of pixels of the target object in the target video data;
And the determining subunit is used for determining the association coefficient between the camera corresponding to the target video data and the target object based on the ratio between the pixel number of the target object and the total pixel number of the target video data.
In some embodiments, the determining subunit is configured to determine an intermediate coefficient based on a ratio between a number of pixels of the target object and a total number of pixels of the target video data; determining a front scale of the target object based on a ratio between a face presented by the target object and a front of the target object in the target video data, wherein the front scale is used for representing the degree of shooting the target object by the front of the camera; and determining a correlation coefficient between a camera corresponding to the target video data and the target object based on the intermediate coefficient and the front scale.
In some embodiments, the screening unit is configured to screen the at least one target video data from the plurality of video data based on text information of the target object;
the screening unit is further configured to screen the at least one target video data from the plurality of video data based on the image information of the target object.
In some embodiments, the apparatus further comprises:
the second acquisition module is used for responding to a filter adding request of the terminal for any play area, and acquiring a filter text, wherein the filter text is used for describing the style of a filter to be added;
the generating module is used for determining a filter to be added based on the filter text;
the sending module is further configured to send the filter to be added to the terminal, where the terminal is configured to update the style of the playing area based on the filter to be added.
In another aspect, there is provided a video playing device, the device comprising:
the acquisition module is used for acquiring information of a target object based on a video playing interface, wherein the video playing interface is used for playing video data shot by a plurality of cameras, and the target object is a locked object;
the sending module is used for sending the information of the target object to a server;
the receiving module is used for receiving a camera sequence returned by the server, the cameras in the camera sequence are arranged according to the sequence of the correlation coefficient from big to small, and the correlation coefficient is used for representing the correlation degree of the content shot by the corresponding camera and the target object;
The display module is used for displaying playing areas corresponding to the cameras in the video playing interface based on the camera sequence, each playing area is used for playing video data shot by the corresponding camera, and the size of the playing area is positively correlated with the front and back of the camera sequencing in the camera sequence.
In some embodiments, the display module is configured to determine, for a play area corresponding to any one camera, a size weight of the play area based on a position of the camera in the camera sequence; weighting the preset size of the playing area based on the size weight to obtain the playing size; and displaying the playing area in the playing size in the video playing interface.
In some embodiments, the receiving module is configured to receive the camera sequence returned by the server every a preset period of time;
the apparatus further comprises:
the first adjusting module is used for adjusting playing areas corresponding to the cameras in the video playing interface based on the latest camera sequence under the condition that the camera sequence is changed;
And the second adjusting module is used for adjusting the volume of the playing area based on the size change condition of the playing area under the condition that the size of the playing area is changed for the playing area corresponding to any video camera, and the volume of the playing area is positively related to the size of the playing area.
In another aspect, a computer device is provided, the computer device including a processor and a memory for storing at least one segment of a computer program, the at least one segment of the computer program being loaded and executed by the processor to implement a video playing method in an embodiment of the present application.
In another aspect, a computer readable storage medium is provided, in which at least one segment of a computer program is stored, the at least one segment of the computer program being loaded and executed by a processor to implement a video playing method as in an embodiment of the present application.
In another aspect, a computer program product is provided, comprising a computer program stored in a computer readable storage medium, the computer program being read from the computer readable storage medium by a processor of a computer device, the computer program being executed by the processor to cause the computer device to perform the video playback method provided in each of the above aspects or in various alternative implementations of each of the aspects.
In the video playing method, in the process of playing videos shot by a plurality of cameras, for any one of the plurality of cameras, calculating the correlation degree between video data of the camera and a target object through the video data shot by the camera and the information of the target object to obtain a correlation coefficient, and then sorting the plurality of cameras according to the order of the correlation coefficient from large to small, namely sorting the plurality of cameras according to the correlation degree between the content shot by the camera and the target object, so that the camera sequence obtained by sorting can accurately reflect the situation of shooting the target object by the camera, and compared with cameras with the ranks at the front, the cameras with the ranks more focus on shooting the target object; then, the size of each playing area of the plurality of cameras in the terminal is indicated through the camera sequence, compared with the cameras with the rear sequence, the size of the playing area of the cameras with the front sequence is larger, so that a user can watch target objects shot by the plurality of cameras at a plurality of angles at the same time, so that the situation of the target objects can be known in multiple directions, the content shot by the cameras for shooting the target objects can be watched more clearly, manual operation of the user is not needed, the man-machine interaction efficiency is improved, and the watching experience of the user is further improved.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings that are needed in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present application, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
Fig. 1 is a schematic diagram of an implementation environment of a video playing method according to an embodiment of the present application;
fig. 2 is a flowchart of a video playing method according to an embodiment of the present application;
fig. 3 is a flowchart of another video playing method according to an embodiment of the present application;
fig. 4 is a flowchart of still another video playing method according to an embodiment of the present application;
FIG. 5 is a schematic diagram of calculating correlation coefficients according to an embodiment of the present application;
FIG. 6 is a schematic diagram of dynamic change of a playing area according to an embodiment of the present application;
fig. 7 is a schematic diagram of a relationship between a playing area and a playing total area of a camera according to an embodiment of the present application;
FIG. 8 is a schematic diagram of an add filter provided according to an embodiment of the present application;
Fig. 9 is a schematic diagram of a video playing method according to an embodiment of the present application;
fig. 10 is a block diagram of a video playing device according to an embodiment of the present application;
fig. 11 is a block diagram of another video playback device provided according to an embodiment of the present application;
fig. 12 is a block diagram of a terminal according to an embodiment of the present application;
fig. 13 is a schematic structural diagram of a server according to an embodiment of the present application.
Detailed Description
For the purpose of making the objects, technical solutions and advantages of the present application more apparent, the embodiments of the present application will be described in further detail below with reference to the accompanying drawings.
The terms "first," "second," and the like in this application are used to distinguish between identical or similar items that have substantially the same function and function, and it should be understood that there is no logical or chronological dependency between the "first," "second," and "nth" terms, nor is it limited to the number or order of execution.
The term "at least one" in this application means one or more, and the meaning of "a plurality of" means two or more.
It should be noted that, information (including but not limited to user equipment information, user personal information, etc.), data (including but not limited to data for analysis, stored data, presented data, etc.), and signals referred to in this application are all authorized by the user or are fully authorized by the parties, and the collection, use, and processing of relevant data is required to comply with relevant laws and regulations and standards of relevant countries and regions. For example, information and video data of a target object referred to in the present application are acquired with sufficient authorization.
In order to facilitate understanding, terms related to the present application are explained below.
Video player: refers to a container used to play video data on a terminal. For example, a play area when the cell phone views a video.
Shooting sites: the live broadcast shooting site consists of site personnel and elements such as a camera, a director, a shot object and the like, and the shot content is presented to the terminal in real time. The shooting scene may also be referred to as a shooting scene.
Machine position: it means that the imaging of one camera is one camera position, and the camera shoots from its own position.
Play area: in a video player of the terminal, n areas are divided, each area is used for displaying contents shot by a corresponding camera, and the area is a playing area. The sizes of the n playing areas may be equal or unequal, which is not limited in the embodiment of the present application.
Center console: the device is used for collecting shooting contents of all the sites on a shooting scene and processing the shooting contents of all the sites. In the center console, a shooting director can perform operations such as machine position, lens switching, filter adding, subtitle adding, lens shielding and the like according to requirements and a table book.
The video playing method provided by the embodiment of the application can be executed by the computer equipment. In some embodiments, the computer device is a terminal or a server. In the following, taking a computer device as an example of a server, an implementation environment of a video playing method provided in an embodiment of the present application is introduced, and fig. 1 is a schematic diagram of an implementation environment of a video playing method provided in an embodiment of the present application. Referring to fig. 1, the implementation environment includes a terminal 101 and a server 102. The terminal 101 and the server 102 can be directly or indirectly connected through wired or wireless communication, and the present application is not limited herein.
In some embodiments, terminal 101 is, but is not limited to, a smart phone, tablet, notebook, desktop, smart speaker, smart watch, smart voice-interactive device, smart home appliance, vehicle-mounted terminal, etc. The terminal 101 runs an application program supporting video playback. The application may be a multimedia-type application, a social-type application, an information-type application, a browser-type application, or a meeting-type application, to which the embodiments of the present application are not limited. Illustratively, the terminal 101 may be a terminal used by a user. The terminal 101 can acquire video data photographed by a plurality of cameras from the server 102. Then, the terminal 101 displays video data photographed by a plurality of cameras in a video playback interface. The user may be a viewer object viewing video, which is not limited by the embodiments of the present application.
Those skilled in the art will recognize that the number of terminals may be greater or lesser. Such as the above-mentioned terminals may be only one, or the above-mentioned terminals may be several tens or hundreds, or more. The number of terminals and the device type are not limited in the embodiment of the present application.
In some embodiments, the server 102 is a stand-alone physical server, can be a server cluster or a distributed system formed by a plurality of physical servers, and can also be a cloud server providing cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communication, middleware services, domain name services, security services, CDNs (Content Delivery Network, content delivery networks), basic cloud computing services such as big data and artificial intelligence platforms, and the like. The server 102 is used to provide background services for applications supporting video playback. In some embodiments, the server 102 takes on primary computing work and the terminal 101 takes on secondary computing work; alternatively, the server 102 takes on secondary computing work and the terminal 101 takes on primary computing work; alternatively, a distributed computing architecture is used for collaborative computing between the server 102 and the terminal 101.
Fig. 2 is a flowchart of a video playing method according to an embodiment of the present application, and referring to fig. 2, in the embodiment of the present application, an example of the method is described by way of execution by a server. The video playing method comprises the following steps:
201. the server acquires a plurality of video data shot by a plurality of cameras, and the plurality of cameras are in one-to-one correspondence with the plurality of video data.
In the embodiment of the application, the plurality of cameras may be located at the same shooting site or may be located at different shooting sites. The plurality of cameras are located at different positions. Accordingly, the contents photographed by the plurality of cameras are different. Things such as people, animals or articles can be included in the shooting site, and the embodiment of the application is not limited to this. In the shooting process, for any camera, the camera uploads video data obtained by shooting to a server. That is, the server can acquire video data captured by each of the plurality of cameras from among the plurality of cameras. In the case where a plurality of cameras are in the same shooting site, the plurality of cameras can shoot things in the shooting site in multiple directions. That is, a plurality of cameras can take a picture of a target object in a shooting scene in multiple directions.
202. The server determines, based on the plurality of video data and the information of the target object, a correlation coefficient between each of the plurality of cameras and the target object, the correlation coefficient being used to represent a degree of correlation between the content captured by the corresponding camera and the target object.
In the embodiment of the application, the shooting scene comprises a target object. The target object may be any person, any animal, any article, or the like in the shooting scene, which is not limited by the embodiment of the present application. The server obtains information of the target object. The information of the target object has personalized characteristics, and can directly reflect the characteristics of the target object. For video data of any camera, the server calculates a correlation coefficient between the video data and information of the target object. The larger the correlation coefficient is, the greater the degree of correlation between the content photographed by the camera and the target object is indicated. That is, the more relevant the content captured by the camera is to the target object. The smaller the correlation coefficient is, the smaller the correlation degree between the content shot by the camera and the target object is. That is, the less relevant the content captured by the camera is to the target object.
In other words, the server searches for the presence of the target object from the plurality of video data based on the information of the target object. That is, the server determines that the plurality of cameras capture the target object based on the information of the target object. The case of photographing the target object may be that the target object is not photographed, may be that a portion of the target object is photographed, may be that the entire target object is photographed, or the like, and the embodiment of the present application does not limit this.
203. And the server sorts the plurality of cameras according to the sequence of the association coefficients from large to small to obtain a camera sequence.
In the embodiment of the application, the server compares the association coefficients corresponding to the cameras and determines the magnitude relation between the association coefficients corresponding to the cameras. Then, the server orders the identifications of the plurality of cameras according to the sequence of the association coefficients from large to small to obtain a camera sequence. The camera sequence can reflect the attention of a plurality of cameras to a shooting target object. That is, the camera ranked forward is more focused on photographing the target object than the camera ranked rearward. In other words, the content of the target object in the video data captured by the camera in the front order occupies a larger amount of data than the camera in the rear order.
Taking video data as an example of one frame of image captured by a camera, the size of the image captured by each camera is the same. Compared with the cameras with the rear-ranking camera, the images shot by the cameras with the front-ranking camera have larger image areas occupied by the target objects.
204. The server sends a camera sequence to the terminal, the camera sequence is used for indicating the sizes of playing areas corresponding to a plurality of cameras in the terminal, each playing area is used for playing video data shot by the corresponding camera, and the sizes of the playing areas are positively correlated with the front and back of the camera sequencing in the camera sequence.
In the embodiment of the present application, the terminal is a terminal for use of a viewer object. The terminal can play video data photographed by a plurality of cameras. Each camera corresponds to a playing area. Correspondingly, the terminal displays playing areas corresponding to the cameras. For any camera, the terminal plays the video data shot by the camera in the playing area corresponding to the camera. That is, the viewer object can view video data photographed by a plurality of cameras at the same time. The server sends the camera sequence to the terminal. The terminal can adjust playing areas corresponding to a plurality of cameras according to the camera sequences. Compared with cameras with the rear order, the cameras with the front order correspond to a larger playing area.
In the video playing method, in the process of playing videos shot by a plurality of cameras, for any one of the plurality of cameras, the correlation degree between the video data of the camera and a target object is calculated through the video data shot by the camera and the information of the target object to obtain a correlation coefficient, then the plurality of cameras are ranked according to the order of the correlation coefficient from large to small, namely, the plurality of cameras are ranked according to the correlation degree between the content shot by the camera and the target object, so that the camera sequence obtained by ranking can accurately reflect the situation of shooting the target object by the camera, and compared with the cameras with the ranks at the front, the cameras with the ranks more focus on shooting the target object; then, the size of each playing area of the plurality of cameras in the terminal is indicated through the camera sequence, compared with the cameras with the rear sequence, the size of the playing area of the cameras with the front sequence is larger, so that a user can watch target objects shot by the plurality of cameras at a plurality of angles at the same time, so that the situation of the target objects can be known in multiple directions, the content shot by the cameras for shooting the target objects can be watched more clearly, manual operation of the user is not needed, the man-machine interaction efficiency is improved, and the watching experience of the user is further improved.
Fig. 3 is a flowchart of another video playing method according to an embodiment of the present application, and referring to fig. 3, in the embodiment of the present application, a terminal is taken as an example for illustration. The video playing method comprises the following steps:
301. the terminal acquires information of a target object based on a video playing interface, wherein the video playing interface is used for playing video data shot by a plurality of cameras, and the target object is a locked object.
In the embodiment of the application, the terminal displays a video playing interface. The terminal can play video data shot by a plurality of cameras in the video play interface. A plurality of cameras are used to capture things in a certain capture site or capture scene. The shooting scene or shooting scene has a target object. Multiple cameras may capture a target object. Accordingly, the viewer object (user) can view the target object through the video play interface displayed by the terminal. The target object may be a shooting scene or any object in the shooting scene, and the embodiment of the present application does not limit the target object. In other words, the target object is an object locked by the viewer object. That is, the target object is an object of interest to the viewer object. The target object may be customized by the viewer object. Accordingly, the terminal can acquire the information of the target object according to the information input operation of the audience object in the video playing interface. The information of the target object may be a name, an image, sex, or shape information of the object, etc., which is not limited in the embodiment of the present application. The appearance information is used to describe the appearance of the target object. The profile information may contain descriptive information about at least one of a hairstyle, wearing, and five sense organs of the target object.
302. And the terminal sends the information of the target object to the server.
In the embodiment of the application, the terminal uploads the information of the target object to the server so that the server determines the object focused by the audience object. The server can determine the condition of shooting the target object by a plurality of cameras according to the information of the target object. That is, the server can determine the degree of correlation of the target object with the contents photographed by the respective cameras from the information of the target object and the video data photographed by the plurality of cameras. Then, the server can sort the plurality of cameras according to the order of the degree of correlation between the target object and the content shot by the cameras from large to small, and a camera sequence is obtained.
303. The terminal receives a camera sequence returned by the server, wherein a plurality of cameras in the camera sequence are arranged according to the sequence of the correlation coefficient from big to small, and the correlation coefficient is used for representing the correlation degree of the content shot by the corresponding camera and the target object.
In the embodiment of the application, the terminal acquires the camera sequence from the server. In the camera sequence, the identifications of the plurality of cameras are arranged in order of the association coefficient from large to small. The more relevant the content captured by the top ranked camera is to the target object. That is, the camera ranked forward is more focused on photographing the target object than the camera ranked backward.
304. The terminal displays playing areas corresponding to a plurality of cameras in a video playing interface based on the camera sequence, wherein each playing area is used for playing video data shot by the corresponding camera, and the size of the playing area is positively correlated with the front and back of the camera sequencing in the camera sequence.
In the embodiment of the application, the camera sequence can indicate the sizes of playing areas corresponding to a plurality of cameras in the terminal. In the camera sequence, the size of the playing area corresponding to the camera with the higher ranking is larger; the size of the playing area corresponding to the camera with the later sequencing is smaller. That is, the terminal displays the playing areas corresponding to the plurality of cameras on the video playing interface according to the situation that the plurality of cameras shoot the object focused by the audience object. The more the camera focuses on photographing the target object, the larger the size of the play area corresponding to the camera. "focusing on photographing of a target object" means that the amount of data occupied by the target object is large in video data photographed by a camera. That is, the larger the amount of data occupied by the target object in the video data captured by the camera, the larger the size of the play area corresponding to the camera.
The embodiment of the application provides a video playing method, wherein in the process of playing videos shot by a plurality of cameras, a user can autonomously select an interested target object, so that the method meets the intention of the user; then, providing information of the target object to a server, so that the server can sort the plurality of cameras according to the correlation degree between the video data of the plurality of cameras and the target object, and the camera sequence obtained by sorting can accurately reflect the condition that the cameras shoot the target object, and compared with cameras with the rear sequences, the cameras with the front sequences pay more attention to shooting the target object; then, the camera sequence is directly acquired from the server, so that the running consumption is saved; then, the sizes of the playing areas of the cameras are displayed according to the camera sequence, compared with the cameras with the rear sequence, the sizes of the playing areas of the cameras with the front sequence are larger, so that a user can watch target objects shot by the cameras at a plurality of angles at the same time, so that the situation of the target objects can be known in multiple directions, the content shot by the cameras for shooting the target objects can be watched more clearly, manual operation of the user is not needed, the man-machine interaction efficiency is improved, and the watching experience of the user is further improved.
Fig. 4 is a flowchart of still another video playing method according to an embodiment of the present application, referring to fig. 4, in the embodiment of the present application, an interaction between a terminal and a server is taken as an example. The video playing method comprises the following steps:
401. the terminal acquires information of a target object based on a video playing interface, wherein the video playing interface is used for playing video data shot by a plurality of cameras, and the target object is a locked object.
In the embodiment of the application, the terminal displays a video playing interface. The terminal can play video data shot by a plurality of cameras in the video play interface. The video data may be data obtained by shooting in real time by a camera, or may be data obtained after shooting is completed, which is not limited in the embodiment of the present application. For example, the terminal can play live content in a video play interface. Live broadcast content is formed by gathering video data shot by a plurality of cameras in real time.
The terminal may directly display video data photographed by a plurality of cameras in the video play interface. Correspondingly, the terminal displays playing areas corresponding to the cameras in the video playing interface. Each play area is used for playing video data shot by a corresponding camera. Or, the terminal may play only video data captured by any one of the plurality of cameras in the video play interface. The video data played in the video playing interface may be customized by the audience object, which is not limited in the embodiments of the present application.
The shooting sites where the plurality of cameras are located include a target object. Accordingly, a plurality of cameras may capture a target object. That is, the viewer object may view the target object through the video playback interface. The target object is the object of interest to the viewer object. The target object may be customized by the viewer object. Accordingly, the terminal can determine the object to be focused on by the audience object according to the information input by the audience object in the video playing interface.
In some embodiments, in the case where the object locking function is turned on, the terminal displays an information input window in the video playback interface. Then, in response to an information input operation to the information input window, the terminal acquires information of the target object. The object lock function is used to represent an object that is focused on being locked when playing the photographed content. Under the condition that the object locking function is not started, the terminal can only play video data shot by any one of the plurality of cameras in the video playing interface.
The information of the target object may be text information such as name, sex or shape information of the object, image information such as a photograph or portrait of the target object, or audio information such as video including the target object or voice of the target object, which is not limited in the embodiment of the present application.
402. And the terminal sends the information of the target object to the server.
In the embodiment of the application, the terminal encodes the information of the target object according to the type of the information of the target object. The terminal then uploads the encoded information to the server to cause the server to determine the object of interest to the viewer object.
403. The server acquires a plurality of video data shot by a plurality of cameras, and the plurality of cameras are in one-to-one correspondence with the plurality of video data.
In this embodiment of the present application, for any one camera, the server may acquire video data obtained by capturing by the camera. Alternatively, the camera may upload captured video data in real time. That is, the server can acquire video data photographed by the camera in real time. The plurality of cameras are located at different positions, and thus the photographed contents are different. The server can acquire video data photographed in a plurality of directions at the same time. The video data may contain a target object.
404. The server screens out at least one target video data from the plurality of video data based on the information of the target object, the target video data including the target object.
In the embodiment of the application, the server determines that the object to be focused on by the audience object is the target object according to the information of the target object. Then, the server screens out at least one video data containing the target object from the plurality of video data. The embodiment of the application does not limit the number of the target video data.
In some embodiments, the information of the target object may be at least one of text information, image information, video information, and audio information. Accordingly, the server may filter out at least one target video data by at least one of the following filtering methods. Four screening methods are described below by way of example, but are in no way limited thereto.
In the first way, the information of the target object is text information. Accordingly, the server screens out at least one target video data from the plurality of video data based on the text information of the target object. The server can perform semantic analysis on the text information, so that characteristic information of the target object is determined. Then, the server identifies the content in the plurality of video data based on the characteristic information of the target object, thereby screening out video data conforming to the characteristic information of the target object. The feature information of the target object is used to reflect the image of the target object. The image of the target object may contain information on the sex, appearance, etc. of the target object.
Alternatively, the server may create the object repository in advance. The object resource library contains characteristic information of each object in the shooting scene. The feature information includes text information and image information. After the text information is subjected to semantic analysis, the server can search the characteristic information of the target object corresponding to the analysis result from the object resource library according to the analysis result. For example, the server may search for image information corresponding to text information from the object repository based on the text information of the target object. Then, the server identifies the contents in the plurality of video data according to the image information of the target object, thereby screening out the video data containing the target object.
In the second mode, the information of the target object is image information. Accordingly, the server screens out at least one target video data from the plurality of video data based on the image information of the target object. The server performs image recognition on the image information, so that the appearance characteristics of the target object are determined. The server performs image recognition on the image in any video data, thereby determining the characteristics of the video data. The server then compares the appearance characteristics of the target object with the characteristics of the video data. Then, for any video data, in the case where there is the same feature between the feature of the video data and the appearance feature of the target object, the server regards the video data as target video data.
Optionally, an AI (Artificial Intelligence ) recognition system is run in the server. The server may perform image recognition on the image in the video data through the AI recognition system, thereby determining whether the target object exists in the video data.
In a third way, the information of the target object is video information. Accordingly, the server screens out at least one target video data from the plurality of video data based on the video information of the target object. The server performs image recognition on images in the video information of the target object, so that the appearance characteristics of the target object are determined. Then, for any video data, the server compares the appearance characteristics of the target object with the characteristics of the video data. Then, in the case where the similarity between the feature of the video data and the appearance feature of the subject reaches a first preset value, the server regards the video data as target video data.
In a fourth way, the information of the target object is audio information. Accordingly, the server screens out at least one target video data from the plurality of video data based on the audio information of the target object. The server extracts voiceprint of the audio information of the target object to obtain voiceprint characteristics of the target object. Then, for any video data, the server performs voiceprint extraction on the audio information of the video data to obtain audio characteristics. The server then compares the voiceprint characteristics of the target object with the audio characteristics of the video data. Then, in the case where the similarity between the audio feature of the video data and the voiceprint feature of the target object reaches a second preset value, the server regards the video data as target video data.
The server can label the video camera corresponding to the target video data, so that the correlation coefficient of the video data shot by the labeled video camera can be calculated according to the label. That is, the server continues to step 405. The annotation of the camera may be referred to as an "enhanced tag" for distinguishing from other cameras that do not capture the target object. For any one of the plurality of video data other than the target video data, the server may set an association coefficient between a camera corresponding to the video data and the target object to a default value. For example, the default value is 0, which is not limited in the embodiments of the present application.
405. For any one of the at least one target video data, the server determines a correlation coefficient between the camera corresponding to the target video data and the target object based on the data amount occupied by the target object in the target video data.
In the embodiment of the application, for any target video data, the server calculates the data amount occupied by the target object in the target video data. Then, the server determines a correlation coefficient between the camera corresponding to the target video data and the target object according to the data amount occupied by the target object in the target video data. The association coefficient is used for representing the degree of correlation between the content shot by the corresponding camera and the target object. The more the target object occupies the target video data, the greater the degree of correlation between the content captured by the corresponding camera and the target object, and thus the greater the correlation coefficient between the camera and the target object. The smaller the data amount occupied by the target object in the target video data, the smaller the correlation degree between the content shot by the corresponding camera and the target object, and thus the smaller the correlation coefficient between the camera and the target object.
In some embodiments, the amount of data that a target object occupies in the target video data refers to the number of pixels of the target object in the target video data. Accordingly, the process of determining the association coefficient between the camera corresponding to the target video data and the target object by the server comprises the following steps: for any one of the at least one target video data, the server obtains a number of pixels of the target object in the target video data. Then, the server determines a correlation coefficient between the camera corresponding to the target video data and the target object based on a ratio between the number of pixels of the target object and the total number of pixels of the target video data. According to the scheme provided by the embodiment of the application, the proportion of the pixel number of the target object in the total number of pixels of the target video data is calculated, so that the proportion of the target object in the content shot by the camera can be more accurately determined, the attention degree of the camera to the target object is more accurately determined, the follow-up more accurate display of the corresponding playing areas of the cameras is facilitated, that is, the display of the content shot by the camera which is more focused on the shooting target object in a larger size is facilitated, the target object can be clearly watched by a user, the playing area of the camera does not need to be manually enlarged by the user, the man-machine interaction efficiency is improved, and the watching experience of the user is further improved.
For example, fig. 5 is a schematic diagram of calculating a correlation coefficient according to an embodiment of the present application. Referring to fig. 5, a camera that captures target video data includes a camera 1, a camera 3, and a camera 6. The server tags the three cameras with "enhanced labels". The server then further analyzes the video data captured by the "enhanced tagged" camera. The server acquires video data shot by the camera in real time. For the acquired video data at any moment, the video data is a video frame, that is, an image. Accordingly, for any camera, the server determines a correlation coefficient between the camera and the target object based on a ratio between the number of pixels of the target object in the image captured by the camera and the total number of pixels of the image. Wherein the server determines the outline of the target object in the image captured by the camera. The server then counts the number of pixels in the outline of the target object. As shown in fig. 5, a target object in an image captured by the camera 1 includes X1 pixels; the target object in the image shot by the camera 3 comprises X3 pixels; the target object in the image taken by the camera 6 contains X6 pixels. The total number of pixels of the images captured by the plurality of cameras is the same. The server then determines a correlation coefficient between the camera and the target object based on a ratio between a number of pixels of the target object in an image captured by the camera and a total number of pixels of the image.
In some embodiments, the server may determine the association coefficient between the camera and the target object based on whether the camera captured the front of the target object. Accordingly, the process of determining the association coefficient between the camera corresponding to the target video data and the target object by the server based on the ratio between the number of pixels of the target object and the total number of pixels of the target video data includes: the server determines an intermediate coefficient based on a ratio between the number of pixels of the target object and the total number of pixels of the target video data. The server then determines a front scale of the target object based on a ratio between a face presented by the target object in the target video data and a front of the target object. The front scale is used to represent the extent to which the camera captures the target object in front. Then, the server determines a correlation coefficient between the camera corresponding to the target video data and the target object based on the intermediate coefficient and the front scale. The server can weight the intermediate coefficients of the camera according to the front proportion of the target object, so that the association coefficients between the camera and the target object are obtained. According to the scheme provided by the embodiment of the invention, as the user is generally used to knowing the interested object from the front, the user generally wants to watch the front of the target object, the proportion of the front of the target object is used to weight the duty ratio of the target object in the content shot by the camera, so that the attention degree of the camera to the target object is more accurately determined, the subsequent more accurate display of the respective corresponding playing areas of the cameras is facilitated, namely, the display of the content shot by the camera which is more focused on the shot target object in a larger size is facilitated, so that the user can watch the target object clearly, the user does not need to manually enlarge the playing area of the camera, the man-machine interaction efficiency is improved, and the watching experience of the user is further improved.
Alternatively, the server may acquire a frontal image of the target object. The front image is an image taken from the front of the target object. The server may determine a front scale of the target object based on a similarity between a region of the target object in the target video data and a front image of the target object.
406. And the server sorts the plurality of cameras according to the sequence of the association coefficients from large to small to obtain a camera sequence.
In the embodiment of the application, the server compares the association coefficients corresponding to the cameras and determines the magnitude relation between the association coefficients corresponding to the cameras. Then, the server orders the identifications of the plurality of cameras according to the sequence of the association coefficients from large to small to obtain a camera sequence. The camera sequence can reflect the attention of a plurality of cameras to a shooting target object. That is, the camera ranked forward is more focused on photographing the target object than the camera ranked rearward.
For example, the camera 1 captures a face close-up of the target object. The camera 2 captures an overall view of the shooting site, including the entirety of the target object. The camera 3 does not capture the target object. Accordingly, in the camera sequence, camera 1 is located before camera 2 and camera 2 is located before camera 3.
407. The server sends the camera sequence to the terminal.
In the embodiment of the application, the camera sequence is used for indicating the sizes of playing areas corresponding to a plurality of cameras in the terminal. Each play area is used for playing video data shot by a corresponding camera. The size of the play area is positively correlated to the front and back of the camera ordering in the camera sequence. The more forward the camera is sequenced, the larger the playing area corresponding to the camera is; the more the camera is ranked, the smaller the play area the camera corresponds to.
The camera sequence may include all of the plurality of cameras, or may include only the camera of the plurality of cameras that captures the target object, which is not limited in this embodiment of the present application. In the case where the camera sequence includes all of the plurality of cameras, the terminal may display play areas corresponding to the plurality of cameras in the video play interface according to the camera sequence. In the case where the camera sequence includes only a camera of the plurality of cameras that captured the target object, the association coefficient between the camera that does not exist in the terminal camera sequence and the target object is set to a default value. For example, the default value is 0, which is not limited in the embodiments of the present application. That is, the terminal automatically arranges the cameras that are not in the camera sequence after the camera sequence, thereby displaying the play areas corresponding to the plurality of cameras.
408. The terminal receives the camera sequence returned by the server.
In the embodiment of the application, a plurality of cameras in the camera sequence are arranged according to the order of the association coefficients from large to small. The association coefficient is used for representing the degree of correlation between the content shot by the corresponding camera and the target object.
In some embodiments, the terminal receives the camera sequence returned by the server at preset time intervals. The preset time period may be a duration of one video frame or a duration of a plurality of video frames, and the size of the preset time period is not limited in the embodiment of the present application. If the preset time period is equal to the duration of one video frame, the server acquires video data (one video frame) shot by a plurality of cameras in real time, and determines a camera sequence according to the video data shot by the plurality of cameras, so that the terminal can acquire the camera sequence in real time. According to the scheme provided by the embodiment of the application, the new camera sequence is acquired again every preset time period, namely, the condition that a plurality of cameras shoot a target object every preset time period is detected, so that the size of a playing area corresponding to the plurality of cameras is automatically adjusted according to the new camera sequence, the content shot by the camera for shooting the target object is watched more clearly, manual operation of a user is not needed, the man-machine interaction efficiency is improved, and the watching experience of the user is further improved.
409. The terminal displays playing areas corresponding to a plurality of cameras in a video playing interface based on the camera sequence, wherein each playing area is used for playing video data shot by the corresponding camera, and the size of the playing area is positively correlated with the front and back of the camera sequencing in the camera sequence.
In the embodiment of the application, in the camera sequence, the size of the playing area corresponding to the camera with the higher ranking is larger; the size of the playing area corresponding to the camera with the later sequencing is smaller. The terminal displays playing areas corresponding to a plurality of cameras in the video playing interface based on the camera sequence. That is, the terminal displays the playing areas corresponding to the plurality of cameras on the video playing interface according to the situation that the plurality of cameras shoot the object focused by the audience object.
In some embodiments, the play area of each camera has an original preset size. The terminal can adjust the size of the playing area of each camera according to the camera sequence. Correspondingly, the process of displaying playing areas corresponding to a plurality of cameras in the video playing interface by the terminal based on the camera sequence comprises the following steps: and for the playing area corresponding to any camera, the terminal determines the size weight of the playing area based on the position of the camera in the camera sequence. And then, the terminal weights the preset size of the playing area based on the size weight to obtain the playing size. Then, the terminal displays the play area in a play size in the video play interface. The more forward the camera is sequenced, the larger the size weight of the playing area corresponding to the camera is; the more the camera is ranked, the smaller the size weight of the playing area corresponding to the camera. That is, the large size weight of the play area is positively correlated with the front and back of the camera ordering in the camera sequence. According to the scheme provided by the embodiment of the application, the size weight of the playing area is determined according to the position of the camera in the camera sequence, so that the size of the playing area of the camera is adjusted, compared with the camera with the rear ordering, the size of the playing area of the camera with the front ordering is larger, so that a user can watch target objects shot by a plurality of cameras at a plurality of angles at the same time, so that the situation of the target objects can be known in multiple directions, the content shot by the camera for shooting the target objects can be watched more clearly, manual operation of the user is not needed, the man-machine interaction efficiency is improved, and the watching experience of the user is further improved.
In some embodiments, the length and width of the play area corresponding to each camera is dynamically varied. That is, the terminal may adjust the size of each play area according to the new camera sequence.
For example, fig. 6 is a schematic diagram of dynamic change of a playing area according to an embodiment of the present application. Referring to fig. 6, fig. 6 (a) illustrates a schematic view of a plurality of play areas before a change in size. Before the dimensional change, the target object in the video data photographed by the camera 1 is the front, and the target object occupies 70% of the data amount. The target object is a side in the video data shot by the camera 3, and the target object occupies 10% of the data amount. The target object in the video data shot by the camera 6 is the back surface, and the target object occupies 5% of the data volume. The camera 2, the camera 4, the camera 5, and the camera 7 do not capture the target object. Accordingly, as shown in fig. 6 (a), the size of the play area of the camera 1 is maximum, the size of the play area of the camera 3 is sub-sized, the size of the play area of the camera 6 is sub-sized, and the sizes of the camera 2, the camera 4, the camera 5, and the camera 7 are the same and minimum. Then, the server determines that the target object is the front side in the video data shot by the camera 5 by recognizing the content shot by each camera in real time, and the target object occupies 80% of the data amount. The target object is a side face in the video data shot by the camera 1, and the target object occupies 10% of the data amount. The target object is a side in the video data shot by the camera 3, and the target object occupies 10% of the data amount. The camera 2, the camera 4, the camera 6, and the camera 7 do not capture the target object. Accordingly, the terminal enlarges the play area of the camera 5 and reduces the play area of the camera 1. As shown in fig. 6 (b), the size of the play area of the camera 5 is largest, the sizes of the play areas of the camera 1 and the camera 3 are the same, and the sizes of the camera 2, the camera 4, the camera 6, and the camera 7 are smallest.
It should be noted that, the length a of each playing area does not exceed the length a of the playing total area; the width B of each play area does not exceed the width B of the total play area. The sum of the areas of the plurality of play areas does not exceed the area of the total play area. For example, fig. 7 is a schematic diagram of a relationship between a playing area and a playing total area of a video camera according to an embodiment of the present application. Referring to fig. 7, the area of the play area of the camera 1 is a1×b1; the area of the playing area of the video camera 2 is the sum of the areas of the playing areas of all video cameras of a2×b2 … …, and the sum of the areas of the playing areas of all video cameras is not more than the area of the playing total area (a total×b total) in the video playing interface. For each playing area, the terminal weights the preset size of the playing area based on the size weight of the playing area to obtain the playing size. Then, the terminal displays the play area in a play size in the video play interface. The area of the total play area= (A1 x B1) the size weight corresponding to camera 1 + (A2 x B2) the size weight corresponding to camera 2 + … … + (An x Bn) the size weight corresponding to camera n.
In some embodiments, the terminal may also adjust the volume of the play area while adjusting the size of the play area. Correspondingly, under the condition that the camera sequence is changed, the terminal adjusts playing areas corresponding to a plurality of cameras in the video playing interface based on the latest camera sequence. Then, for the playing area corresponding to any video camera, when the size of the playing area is changed, the terminal adjusts the volume of the playing area based on the size change of the playing area. The volume size of the play area is positively correlated with the size of the play area. According to the scheme provided by the embodiment of the application, the larger the size of the playing area is, the easier the user is attracted to watch the video of the playing area, and the volume of the playing area is increased while the playing area is enlarged, so that the user can watch the target object in the playing area better; the volume of the play area is reduced while the play area is reduced, the interference of the video of the play area to the user is reduced, manual operation of the user is not needed, the man-machine interaction efficiency is improved, and the watching experience of the user is further improved.
In some embodiments, the terminal may also add a filter to any play area. Correspondingly, the server acquires the filter text in response to a filter adding request of the terminal for any play area. The filter text is used to describe the style of filter to be added. Then, the server determines a filter to be added based on the filter text. Then, the server transmits the filter to be added to the terminal. The terminal is used for updating the style of the playing area based on the filter to be added. The server can generate a filter to be added according to the filter text; alternatively, the server may also select, according to the filter text, a filter to be added from a plurality of filter templates, which is not limited in the embodiment of the present application. According to the scheme provided by the embodiment of the application, the corresponding text can be added for the corresponding playing area according to the filter text, namely, the style of each area can be adjusted according to the user demand, the operation is simple, and the management and the control of the playing area of the lock are convenient.
For example, fig. 8 is a schematic diagram of an additive filter according to an embodiment of the present application. Referring to (a) of fig. 8, in the case where the play area of the camera 1 is selected, the terminal displays an editing panel of the play area. Then, in response to an input operation to the editing panel, the terminal acquires a filter text. The filter text is a secondary love cartoon filter, pink and warm colors are dominant, people are irrelevant and exquisite, eyes are bright and refreshing, the filter text is close to a beauty man character …, and then the terminal uploads the filter text to a server. Then, the server performs semantic analysis on the filter text, thereby generating a filter to be added. Then, the server transmits the filter to be added to the terminal. Then, the terminal updates the style of the play area based on the filter to be added. In the case where the size of the play area of the video camera 1 is changed, the terminal keeps the filter of the play area unchanged, see (b) in fig. 8.
In order to more clearly describe the video playing method provided by the embodiment of the present application, the video playing method is further described below with reference to the accompanying drawings. Fig. 9 is a schematic diagram of a video playing method according to an embodiment of the present application. Referring to fig. 9, in case that the object locking function is not turned on, the terminal displays video data photographed by a certain camera transmitted from a center console of a photographing site. And under the condition that the object locking function is started, the terminal displays an information input window in the video playing interface. In response to an information input operation to the information input window, the terminal acquires information of the target object. And the terminal sends the information of the target object to the server. The server identifies the information of the target object. Then, the server sorts the plurality of cameras based on the plurality of video data and the information identification result of the target object, to obtain a camera sequence. The server sends the camera sequence to the terminal. The terminal displays playing areas corresponding to a plurality of cameras in the video playing interface based on the camera sequence.
In the video playing method, in the process of playing videos shot by a plurality of cameras, for any one of the plurality of cameras, the correlation degree between the video data of the camera and a target object is calculated through the video data shot by the camera and the information of the target object to obtain a correlation coefficient, then the plurality of cameras are ranked according to the order of the correlation coefficient from large to small, namely, the plurality of cameras are ranked according to the correlation degree between the content shot by the camera and the target object, so that the camera sequence obtained by ranking can accurately reflect the situation of shooting the target object by the camera, and compared with the cameras with the ranks at the front, the cameras with the ranks more focus on shooting the target object; then, the size of each playing area of the plurality of cameras in the terminal is indicated through the camera sequence, compared with the cameras with the rear sequence, the size of the playing area of the cameras with the front sequence is larger, so that a user can watch target objects shot by the plurality of cameras at a plurality of angles at the same time, so that the situation of the target objects can be known in multiple directions, the content shot by the cameras for shooting the target objects can be watched more clearly, manual operation of the user is not needed, the man-machine interaction efficiency is improved, and the watching experience of the user is further improved.
Fig. 10 is a block diagram of a video playing device according to an embodiment of the present application. The video playing device is configured to execute the steps when the video playing method is executed, referring to fig. 10, where the video playing device includes: a first acquisition module 1001, a determination module 1002, a ranking module 1003, and a sending module 1004.
A first obtaining module 1001, configured to obtain a plurality of video data captured by a plurality of cameras, where the plurality of cameras are in one-to-one correspondence with the plurality of video data;
a determining module 1002, configured to determine, based on the plurality of video data and the information of the target object, a correlation coefficient between each of the plurality of cameras and the target object, where the correlation coefficient is used to represent a degree of correlation between the content captured by the corresponding camera and the target object;
a sorting module 1003, configured to sort the plurality of cameras according to the order of the association coefficients from the big to the small, so as to obtain a camera sequence;
the sending module 1004 is configured to send a camera sequence to the terminal, where the camera sequence is configured to indicate sizes of playing areas corresponding to a plurality of cameras in the terminal, each playing area is configured to play video data shot by a corresponding camera, and the sizes of the playing areas are positively correlated to the front and back of the camera ordering in the camera sequence.
In some embodiments, the determining module comprises:
a screening unit, configured to screen at least one target video data from a plurality of video data based on information of the target object, where the target video data includes the target object;
and the determining unit is used for determining the association coefficient between the camera corresponding to the target video data and the target object based on the data amount occupied by the target object in the target video data for any one of the at least one target video data.
In some embodiments, the determining unit comprises:
an obtaining subunit, configured to obtain, for any one target video data in the at least one target video data, a number of pixels of the target object in the target video data;
and the determining subunit is used for determining the association coefficient between the camera corresponding to the target video data and the target object based on the ratio between the pixel number of the target object and the total number of pixels of the target video data.
In some embodiments, the determining subunit is configured to determine the intermediate coefficient based on a ratio between a number of pixels of the target object and a total number of pixels of the target video data; determining the front proportion of the target object based on the ratio between the surface presented by the target object and the front of the target object in the target video data, wherein the front proportion is used for representing the degree of shooting the target object by the front of the camera; and determining the association coefficient between the camera corresponding to the target video data and the target object based on the intermediate coefficient and the front scale.
In some embodiments, the filtering unit is configured to filter at least one target video data from the plurality of video data based on the text information of the target object;
and the screening unit is also used for screening at least one target video data from the plurality of video data based on the image information of the target object.
In some embodiments, the apparatus further comprises:
the second acquisition module is used for responding to a filter adding request of the terminal for any play area, acquiring a filter text, wherein the filter text is used for describing the style of a filter to be added;
the generating module is used for determining a filter to be added based on the filter text;
and the sending module is also used for sending the filter to be added to the terminal, and the terminal is used for updating the style of the playing area based on the filter to be added.
In the video playing device, in the process of playing videos shot by a plurality of cameras, for any one of the plurality of cameras, calculating the correlation degree between video data of the camera and a target object through the video data shot by the camera and the information of the target object to obtain a correlation coefficient, and then sorting the plurality of cameras according to the order of the correlation coefficient from large to small, namely sorting the plurality of cameras according to the correlation degree between the content shot by the camera and the target object, so that the camera sequence obtained by sorting can accurately reflect the situation of shooting the target object by the camera, and compared with cameras with the ranks at the front, the cameras with the ranks more focus on shooting the target object; then, the size of each playing area of the plurality of cameras in the terminal is indicated through the camera sequence, compared with the cameras with the rear sequence, the size of the playing area of the cameras with the front sequence is larger, so that a user can watch target objects shot by the plurality of cameras at a plurality of angles at the same time, so that the situation of the target objects can be known in multiple directions, the content shot by the cameras for shooting the target objects can be watched more clearly, manual operation of the user is not needed, the man-machine interaction efficiency is improved, and the watching experience of the user is further improved.
Fig. 11 is a block diagram of another video playback device according to an embodiment of the present application. The video playing device is configured to execute the steps when the video playing method is executed, referring to fig. 11, where the video playing device includes: an acquisition module 1101, a transmission module 1102, a reception module 1103, and a display module 1104.
The obtaining module 1101 is configured to obtain information of a target object based on a video playing interface, where the video playing interface is configured to play video data captured by a plurality of cameras, and the target object is a locked object;
a sending module 1102, configured to send information of a target object to a server;
a receiving module 1103, configured to receive a camera sequence returned by the server, where a plurality of cameras in the camera sequence are arranged according to a sequence of from big to small association coefficients, where the association coefficients are used to represent a degree of correlation between content captured by the corresponding camera and a target object;
the display module 1104 is configured to display, on the basis of the camera sequence, play areas corresponding to a plurality of cameras in the video play interface, where each play area is used to play video data captured by a corresponding camera, and the size of the play area is positively correlated to the front and rear of the camera sequence in the camera sequence.
In some embodiments, the display module is configured to determine, for a play area corresponding to any one of the cameras, a size weight of the play area based on a position of the camera in the camera sequence; weighting the preset size of the playing area based on the size weight to obtain the playing size; in the video playback interface, the playback area is displayed in a playback size.
In some embodiments, the receiving module is configured to receive the camera sequence returned by the server at intervals of a preset time period;
the apparatus further comprises:
the first adjusting module is used for adjusting playing areas corresponding to a plurality of cameras in the video playing interface based on the latest camera sequence under the condition that the camera sequence is changed;
the second adjusting module is used for adjusting the volume of the playing area based on the size change condition of the playing area under the condition that the size of the playing area is changed for the playing area corresponding to any video camera, and the volume of the playing area is positively related to the size of the playing area.
The embodiment of the application provides a video playing device, wherein in the process of playing videos shot by a plurality of cameras, a user can autonomously select an interested target object, so that the user intention is met; then, providing information of the target object to a server, so that the server can sort the plurality of cameras according to the correlation degree between the video data of the plurality of cameras and the target object, and the camera sequence obtained by sorting can accurately reflect the condition that the cameras shoot the target object, and compared with cameras with the rear sequences, the cameras with the front sequences pay more attention to shooting the target object; then, the camera sequence is directly acquired from the server, so that the running consumption is saved; then, the sizes of the playing areas of the cameras are displayed according to the camera sequence, compared with the cameras with the rear sequence, the sizes of the playing areas of the cameras with the front sequence are larger, so that a user can watch target objects shot by the cameras at a plurality of angles at the same time, so that the situation of the target objects can be known in multiple directions, the content shot by the cameras for shooting the target objects can be watched more clearly, manual operation of the user is not needed, the man-machine interaction efficiency is improved, and the watching experience of the user is further improved.
It should be noted that, when the video playing device provided in the foregoing embodiment runs an application program, only the division of the foregoing functional modules is used as an example, in practical application, the foregoing functional allocation may be performed by different functional modules according to needs, that is, the internal structure of the device is divided into different functional modules, so as to complete all or part of the functions described above. In addition, the video playing device and the video playing method provided in the foregoing embodiments belong to the same concept, and specific implementation processes of the video playing device and the video playing method are detailed in the method embodiments and are not repeated herein.
In the embodiment of the present application, the computer device may be configured as a terminal or a server, and when the computer device is configured as a terminal, the technical solution provided in the embodiment of the present application may be implemented by the terminal as an execution body, and when the computer device is configured as a server, the technical solution provided in the embodiment of the present application may be implemented by the server as an execution body, and also the technical solution provided in the present application may be implemented by interaction between the terminal and the server, which is not limited in this embodiment of the present application.
Fig. 12 is a block diagram of a terminal 1200 according to an embodiment of the present application.
In general, the terminal 1200 includes: a processor 1201 and a memory 1202.
Processor 1201 may include one or more processing cores, such as a 4-core processor, an 8-core processor, or the like. The processor 1201 may be implemented in at least one hardware form of DSP (Digital Signal Processing ), FPGA (Field-Programmable Gate Array, field programmable gate array), PLA (Programmable Logic Array ). The processor 1201 may also include a main processor, which is a processor for processing data in an awake state, also called a CPU (Central Processing Unit ), and a coprocessor; a coprocessor is a low-power processor for processing data in a standby state. In some embodiments, the processor 1201 may be integrated with a GPU (Graphics Processing Unit, image processor) for taking care of rendering and rendering of content that the display screen is required to display. In some embodiments, the processor 1201 may also include an AI (Artificial Intelligence ) processor for processing computing operations related to machine learning.
Memory 1202 may include one or more computer-readable storage media, which may be non-transitory. Memory 1202 may also include high-speed random access memory as well as non-volatile memory, such as one or more magnetic disk storage devices, flash memory storage devices. In some embodiments, a non-transitory computer readable storage medium in memory 1202 is used to store at least one computer program for execution by processor 1201 to implement the video playback methods provided by the method embodiments herein.
In some embodiments, the terminal 1200 may further optionally include: a peripheral interface 1203, and at least one peripheral. The processor 1201, the memory 1202, and the peripheral interface 1203 may be connected by a bus or signal lines. The individual peripheral devices may be connected to the peripheral device interface 1203 via buses, signal lines, or a circuit board. Specifically, the peripheral device includes: at least one of radio frequency circuitry 1204, a display 1205, a camera assembly 1206, audio circuitry 1207, and a power supply 1208.
The peripheral interface 1203 may be used to connect at least one peripheral device associated with an I/O (Input/Output) to the processor 1201 and the memory 1202. In some embodiments, the processor 1201, the memory 1202, and the peripheral interface 1203 are integrated on the same chip or circuit board; in some other embodiments, either or both of the processor 1201, the memory 1202, and the peripheral interface 1203 may be implemented on separate chips or circuit boards, which is not limited in this embodiment.
The Radio Frequency circuit 1204 is used for receiving and transmitting RF (Radio Frequency) signals, also called electromagnetic signals. The radio frequency circuit 1204 communicates with a communication network and other communication devices via electromagnetic signals. The radio frequency circuit 1204 converts an electrical signal into an electromagnetic signal for transmission, or converts a received electromagnetic signal into an electrical signal. In some embodiments, the radio frequency circuit 1204 includes: antenna systems, RF transceivers, one or more amplifiers, tuners, oscillators, digital signal processors, codec chipsets, subscriber identity module cards, and so forth. The radio frequency circuit 1204 may communicate with other terminals via at least one wireless communication protocol. The wireless communication protocol includes, but is not limited to: the world wide web, metropolitan area networks, intranets, generation mobile communication networks (2G, 3G, 4G, and 5G), wireless local area networks, and/or WiFi (Wireless Fidelity ) networks. In some embodiments, the radio frequency circuit 1204 may also include NFC (Near Field Communication ) related circuits, which are not limited in this application.
The display 1205 is used to display a UI (User Interface). The UI may include graphics, text, icons, video, and any combination thereof. When the display 1205 is a touch display, the display 1205 also has the ability to collect touch signals at or above the surface of the display 1205. The touch signal may be input as a control signal to the processor 1201 for processing. At this time, the display 1205 may also be used to provide virtual buttons and/or a virtual keyboard, also referred to as soft buttons and/or a soft keyboard. In some embodiments, the display 1205 may be one and disposed on a front panel of the terminal 1200; in other embodiments, the display 1205 may be at least two, respectively disposed on different surfaces of the terminal 1200 or in a folded design; in other embodiments, the display 1205 may be a flexible display disposed on a curved surface or a folded surface of the terminal 1200. Even more, the display 1205 may be arranged in an irregular pattern that is not rectangular, i.e., a shaped screen. The display 1205 can be made of LCD (Liquid Crystal Display ), OLED (Organic Light-Emitting Diode) or other materials.
The camera assembly 1206 is used to capture images or video. In some embodiments, camera assembly 1206 includes a front camera and a rear camera. Typically, the front camera is disposed on the front panel of the terminal and the rear camera is disposed on the rear surface of the terminal. In some embodiments, the at least two rear cameras are any one of a main camera, a depth camera, a wide-angle camera and a tele camera, so as to realize that the main camera and the depth camera are fused to realize a background blurring function, and the main camera and the wide-angle camera are fused to realize a panoramic shooting and Virtual Reality (VR) shooting function or other fusion shooting functions. In some embodiments, camera assembly 1206 may also include a flash. The flash lamp can be a single-color temperature flash lamp or a double-color temperature flash lamp. The dual-color temperature flash lamp refers to a combination of a warm light flash lamp and a cold light flash lamp, and can be used for light compensation under different color temperatures.
The audio circuitry 1207 may include a microphone and a speaker. The microphone is used for collecting sound waves of a user and the environment, converting the sound waves into electric signals, and inputting the electric signals to the processor 1201 for processing, or inputting the electric signals to the radio frequency circuit 1204 for voice communication. For purposes of stereo acquisition or noise reduction, a plurality of microphones may be respectively disposed at different portions of the terminal 1200. The microphone may also be an array microphone or an omni-directional pickup microphone. The speaker is used to convert electrical signals from the processor 1201 or the radio frequency circuit 1204 into sound waves. The speaker may be a conventional thin film speaker or a piezoelectric ceramic speaker. When the speaker is a piezoelectric ceramic speaker, not only the electric signal can be converted into a sound wave audible to humans, but also the electric signal can be converted into a sound wave inaudible to humans for ranging and other purposes. In some embodiments, the audio circuitry 1207 may also include a headphone jack.
The power supply 1208 is used to power the various components in the terminal 1200. The power source 1208 may be alternating current, direct current, disposable battery, or rechargeable battery. When the power source 1208 includes a rechargeable battery, the rechargeable battery may be a wired rechargeable battery or a wireless rechargeable battery. The wired rechargeable battery is a battery charged through a wired line, and the wireless rechargeable battery is a battery charged through a wireless coil. The rechargeable battery may also be used to support fast charge technology.
In some embodiments, terminal 1200 also includes one or more sensors 1209. The one or more sensors 1209 include, but are not limited to: acceleration sensor 1210, gyro sensor 1211, pressure sensor 1212, optical sensor 1213, and proximity sensor 1214.
The acceleration sensor 1210 may detect the magnitudes of accelerations on three coordinate axes of a coordinate system established with the terminal 1200. For example, the acceleration sensor 1210 may be used to detect components of gravitational acceleration in three coordinate axes. The processor 1201 may control the display 1205 to display a user interface in a landscape view or a portrait view based on the gravitational acceleration signal acquired by the acceleration sensor 1210. The acceleration sensor 1210 may also be used for the acquisition of motion data of a game or a user.
The gyro sensor 1211 may detect a body direction and a rotation angle of the terminal 1200, and the gyro sensor 1211 may collect a 3D motion of the user to the terminal 1200 in cooperation with the acceleration sensor 1210. The processor 1201 can implement the following functions based on the data collected by the gyro sensor 1211: motion sensing (e.g., changing UI according to a tilting operation by a user), image stabilization at shooting, game control, and inertial navigation.
The pressure sensor 1212 may be disposed at a side frame of the terminal 1200 and/or at an underlying layer of the display 1205. When the pressure sensor 1212 is provided at a side frame of the terminal 1200, a grip signal of the terminal 1200 by a user may be detected, and the processor 1201 performs a left-right hand recognition or a shortcut operation according to the grip signal collected by the pressure sensor 1212. When the pressure sensor 1212 is provided at the lower layer of the display 1205, the processor 1201 realizes control of the operability control on the UI interface according to the pressure operation of the user on the display 1205. The operability controls include at least one of a button control, a scroll bar control, an icon control, and a menu control.
The optical sensor 1213 is used to collect the ambient light intensity. In one embodiment, processor 1201 may control the display brightness of display 1205 based on the intensity of ambient light collected by optical sensor 1213. Specifically, when the intensity of the ambient light is high, the display brightness of the display screen 1205 is turned up; when the ambient light intensity is low, the display brightness of the display screen 1205 is turned down. In another embodiment, processor 1201 may also dynamically adjust the shooting parameters of camera assembly 1206 based on the intensity of ambient light collected by optical sensor 1213.
A proximity sensor 1214, also referred to as a distance sensor, is typically provided on the front panel of the terminal 1200. The proximity sensor 1214 serves to collect a distance between the user and the front surface of the terminal 1200. In one embodiment, when the proximity sensor 1214 detects that the distance between the user and the front surface of the terminal 1200 gradually decreases, the processor 1201 controls the display 1205 to switch from the bright screen state to the off screen state; when the proximity sensor 1214 detects that the distance between the user and the front surface of the terminal 1200 gradually increases, the processor 1201 controls the display 1205 to switch from the off-screen state to the on-screen state.
It will be appreciated by those skilled in the art that the structure shown in fig. 12 is not limiting and that more or fewer components than shown may be included or certain components may be combined or a different arrangement of components may be employed.
Fig. 13 is a schematic structural diagram of a server according to an embodiment of the present application, where the server 1300 may have a relatively large difference due to different configurations or performances, and may include one or more processors (Central Processing Units, CPU) 1301 and one or more memories 1302, where at least one computer program is stored in the memories 1302, and the at least one computer program is loaded and executed by the processor 1301 to implement the video playing method provided in the above method embodiments. Of course, the server 1300 may also have a wired or wireless network interface, a keyboard, an input/output interface, etc. for performing input/output, and the server 1300 may also include other components for implementing the functions of the device, which are not described herein.
The present application also provides a computer readable storage medium, in which at least one section of a computer program is stored, where the at least one section of the computer program is loaded and executed by a processor of a computer device to implement the operations performed by the computer device in the video playing method of the above embodiment. For example, the computer readable storage medium may be Read-Only Memory (ROM), random-access Memory (Random Access Memory, RAM), compact disc Read-Only Memory (Compact Disc Read-Only Memory, CD-ROM), magnetic tape, floppy disk, optical data storage device, and the like.
Embodiments of the present application also provide a computer program product comprising a computer program stored in a computer readable storage medium. The processor of the computer device reads the computer program from the computer-readable storage medium, and the processor executes the computer program so that the computer device performs the video playback method provided in the above-described various alternative implementations.
It will be understood by those skilled in the art that all or part of the steps for implementing the above embodiments may be implemented by hardware, or may be implemented by a program for instructing relevant hardware, where the program may be stored in a computer readable storage medium, and the storage medium may be a read-only memory, a magnetic disk or an optical disk, etc.
The foregoing description of the preferred embodiments is merely exemplary in nature and is in no way intended to limit the invention, since it is intended that all modifications, equivalents, improvements, etc. that fall within the spirit and scope of the invention.

Claims (14)

1. A video playing method, the method comprising:
acquiring a plurality of video data shot by a plurality of cameras, wherein the plurality of cameras are in one-to-one correspondence with the plurality of video data;
determining a correlation coefficient between each camera in the plurality of cameras and the target object based on the plurality of video data and the information of the target object, wherein the correlation coefficient is used for representing the correlation degree between the content shot by the corresponding camera and the target object;
sequencing the cameras according to the sequence of the correlation coefficient from big to small to obtain a camera sequence;
and sending the camera sequence to a terminal, wherein the camera sequence is used for indicating the sizes of playing areas corresponding to the cameras in the terminal, each playing area is used for playing video data shot by the corresponding camera, and the sizes of the playing areas are positively correlated with the front and back of the camera sequencing in the camera sequence.
2. The method of claim 1, wherein the determining the association coefficients between each of the plurality of cameras and the target object based on the plurality of video data and the information of the target object comprises:
screening at least one target video data from the plurality of video data based on the information of the target object, wherein the target video data comprises the target object;
and for any one of the at least one target video data, determining a correlation coefficient between a camera corresponding to the target video data and the target object based on the data amount occupied by the target object in the target video data.
3. The method according to claim 2, wherein the determining, for any one of the at least one target video data, the association coefficient between the camera corresponding to the target video data and the target object based on the amount of data occupied by the target object in the target video data, includes:
for any one of the at least one target video data, acquiring the number of pixels of the target object in the target video data;
And determining a correlation coefficient between a camera corresponding to the target video data and the target object based on the ratio between the pixel number of the target object and the total pixel number of the target video data.
4. A method according to claim 3, wherein said determining an association coefficient between a camera corresponding to the target video data and the target object based on a ratio between the number of pixels of the target object and the total number of pixels of the target video data comprises:
determining an intermediate coefficient based on a ratio between a number of pixels of the target object and a total number of pixels of the target video data;
determining a front scale of the target object based on a ratio between a face presented by the target object and a front of the target object in the target video data, wherein the front scale is used for representing the degree of shooting the target object by the front of the camera;
and determining a correlation coefficient between a camera corresponding to the target video data and the target object based on the intermediate coefficient and the front scale.
5. The method of claim 2, wherein the screening at least one target video data from the plurality of video data based on the information of the target object comprises:
Screening the at least one target video data from the plurality of video data based on the text information of the target object; or,
and screening out the at least one target video data from the plurality of video data based on the image information of the target object.
6. The method according to claim 1, wherein the method further comprises:
responding to a filter adding request of the terminal for any playing area, and acquiring a filter text, wherein the filter text is used for describing the style of a filter to be added;
determining a filter to be added based on the filter text;
and sending the filter to be added to the terminal, wherein the terminal is used for updating the style of the playing area based on the filter to be added.
7. A video playing method, the method comprising:
acquiring information of a target object based on a video playing interface, wherein the video playing interface is used for playing video data shot by a plurality of cameras, and the target object is a locked object;
transmitting information of the target object to a server;
receiving a camera sequence returned by the server, wherein the cameras in the camera sequence are arranged according to the sequence of the association coefficients from large to small, and the association coefficients are used for representing the correlation degree of the content shot by the corresponding camera and the target object;
And displaying playing areas corresponding to the cameras in the video playing interface based on the camera sequence, wherein each playing area is used for playing video data shot by the corresponding camera, and the size of the playing area is positively correlated with the front and back of the camera sequencing in the camera sequence.
8. The method of claim 7, wherein displaying the play areas corresponding to the plurality of cameras in the video play interface based on the camera sequence comprises:
for a playing area corresponding to any camera, determining the size weight of the playing area based on the position of the camera in the camera sequence;
weighting the preset size of the playing area based on the size weight to obtain the playing size;
and displaying the playing area in the playing size in the video playing interface.
9. The method of claim 7, wherein the receiving the camera sequence returned by the server comprises:
receiving the camera sequence returned by the server every preset time period;
the method further comprises the steps of:
Under the condition that the camera sequence is changed, based on the latest camera sequence, adjusting playing areas corresponding to the cameras in the video playing interface;
and adjusting the volume of the playing area based on the size change condition of the playing area under the condition that the size of the playing area is changed for the playing area corresponding to any video camera, wherein the volume of the playing area is positively related to the size of the playing area.
10. A video playback device, the device comprising:
the first acquisition module is used for acquiring a plurality of video data shot by a plurality of cameras, and the plurality of cameras are in one-to-one correspondence with the plurality of video data;
a determining module, configured to determine, based on the plurality of video data and information of a target object, a correlation coefficient between each camera of the plurality of cameras and the target object, where the correlation coefficient is used to represent a degree of correlation between content captured by a corresponding camera and the target object;
the sequencing module is used for sequencing the cameras according to the sequence of the correlation coefficients from large to small to obtain a camera sequence;
The sending module is used for sending the camera sequence to the terminal, the camera sequence is used for indicating the sizes of playing areas corresponding to the cameras in the terminal, each playing area is used for playing video data shot by the corresponding camera, and the sizes of the playing areas are positively correlated with the front and rear of the camera sequencing in the camera sequence.
11. A video playback device, the device comprising:
the acquisition module is used for acquiring information of a target object based on a video playing interface, wherein the video playing interface is used for playing video data shot by a plurality of cameras, and the target object is a locked object;
the sending module is used for sending the information of the target object to a server;
the receiving module is used for receiving the camera sequence returned by the server, the cameras in the camera sequence are arranged according to the sequence from big to small of the association coefficient, and the association coefficient is used for representing the correlation degree of the content shot by the corresponding camera and the target object;
the display module is used for displaying playing areas corresponding to the cameras in the video playing interface based on the camera sequence, each playing area is used for playing video data shot by the corresponding camera, and the size of the playing area is positively correlated with the front and back of the camera sequencing in the camera sequence.
12. A computer device, characterized in that it comprises a processor and a memory for storing at least one piece of computer program, which is loaded by the processor and which performs the video playback method of any one of claims 1 to 6; alternatively, the at least one computer program is loaded by the processor and performs the video playing method of any one of claims 7 to 9.
13. A computer readable storage medium for storing at least one segment of a computer program for performing the video playback method of any one of claims 1 to 6; alternatively, the at least one computer program is configured to perform the video playing method of any one of claims 7 to 9.
14. A computer program product comprising a computer program which, when executed by a processor, implements the video playback method of any one of claims 1 to 6; alternatively, the computer program, when executed by a processor, implements a video playback method as claimed in any one of claims 7 to 9.
CN202311491230.8A 2023-11-09 2023-11-09 Video playing method, device, computer equipment and storage medium Pending CN117528125A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311491230.8A CN117528125A (en) 2023-11-09 2023-11-09 Video playing method, device, computer equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311491230.8A CN117528125A (en) 2023-11-09 2023-11-09 Video playing method, device, computer equipment and storage medium

Publications (1)

Publication Number Publication Date
CN117528125A true CN117528125A (en) 2024-02-06

Family

ID=89741293

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311491230.8A Pending CN117528125A (en) 2023-11-09 2023-11-09 Video playing method, device, computer equipment and storage medium

Country Status (1)

Country Link
CN (1) CN117528125A (en)

Similar Documents

Publication Publication Date Title
CN110865754B (en) Information display method and device and terminal
CN109040297B (en) User portrait generation method and device
CN111506758B (en) Method, device, computer equipment and storage medium for determining article name
CN111541907A (en) Article display method, apparatus, device and storage medium
WO2022048398A1 (en) Multimedia data photographing method and terminal
CN111432245B (en) Multimedia information playing control method, device, equipment and storage medium
CN112788359B (en) Live broadcast processing method and device, electronic equipment and storage medium
CN112363660B (en) Method and device for determining cover image, electronic equipment and storage medium
CN109618192B (en) Method, device, system and storage medium for playing video
CN113395566B (en) Video playing method and device, electronic equipment and computer readable storage medium
CN111836073B (en) Method, device and equipment for determining video definition and storage medium
CN110675473B (en) Method, device, electronic equipment and medium for generating GIF dynamic diagram
CN111327819A (en) Method, device, electronic equipment and medium for selecting image
CN114302160A (en) Information display method, information display device, computer equipment and medium
CN112131473B (en) Information recommendation method, device, equipment and storage medium
CN110891181B (en) Live broadcast picture display method and device, storage medium and terminal
CN110853124B (en) Method, device, electronic equipment and medium for generating GIF dynamic diagram
CN112419143A (en) Image processing method, special effect parameter setting method, device, equipment and medium
CN114415907B (en) Media resource display method, device, equipment and storage medium
CN112004134A (en) Multimedia data display method, device, equipment and storage medium
CN111369434B (en) Method, device, equipment and storage medium for generating spliced video covers
CN114297150A (en) Media file processing method, device, equipment and storage medium
CN117528125A (en) Video playing method, device, computer equipment and storage medium
CN113609387A (en) Playing content recommendation method and device, electronic equipment and storage medium
CN110662113B (en) Video playing method and device and computer readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication