CN110324672A - A kind of video data handling procedure, device, system and medium - Google Patents

A kind of video data handling procedure, device, system and medium Download PDF

Info

Publication number
CN110324672A
CN110324672A CN201910463414.0A CN201910463414A CN110324672A CN 110324672 A CN110324672 A CN 110324672A CN 201910463414 A CN201910463414 A CN 201910463414A CN 110324672 A CN110324672 A CN 110324672A
Authority
CN
China
Prior art keywords
video
location information
information
target
interactive
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910463414.0A
Other languages
Chinese (zh)
Inventor
刘小荣
谢年华
张富春
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Priority to CN201910463414.0A priority Critical patent/CN110324672A/en
Publication of CN110324672A publication Critical patent/CN110324672A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/239Interfacing the upstream path of the transmission network, e.g. prioritizing client content requests
    • H04N21/2393Interfacing the upstream path of the transmission network, e.g. prioritizing client content requests involving handling client requests
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/262Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists
    • H04N21/26208Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists the scheduling operation being performed under constraints
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47217End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for controlling playback functions for recorded or on-demand content, e.g. using progress bars, mode or play-point indicators or bookmarks

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Human Computer Interaction (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The invention discloses a kind of video data handling procedure, device, system and media, this method comprises: receiving the location information acquisition request that terminal is sent;Based on the location information acquisition request, the video object location information to match with the target video identification information is searched;Searched the video object location information is sent to terminal, so that the terminal exports the second video data based on the first video data of current video, the interactive information of interaction trigger action and interactive placement information.In this way, interaction effect can be enriched, interactive response duration is reduced, reduces the calculation amount and power consumption of terminal, reduces terminal hardware requirement, promotes whole interactive processing efficiency, user experience is more preferable.

Description

A kind of video data handling procedure, device, system and medium
Technical field
The present invention relates to field of computer technology more particularly to a kind of video data handling procedure, device, system and Jie Matter.
Background technique
With becoming increasingly abundant for internet video information, user watches video at the terminal becomes main leisure joy at present Happy mode.
When playing video, user, which can participate in, carries out video interactive in the content of video.Existing video traffic leads to Often only thumb up, make comments, interaction forms, the interaction forms such as object of giving gifts it is more single.If wanting to realize more flexible interaction effect Fruit, possible implementation are as follows: corresponding algorithm packet is installed in terminal, in video display process, calculates and ties to the algorithm packet Shu Hou carries out interaction effect displaying further according to calculated result.Obvious this implementation is deposited due to more demanding to terminal hardware The defects of old type can not execute the processing requirement, and treatment effeciency is slow, power consumption is fast in part, these are all to a certain extent Influence user interaction experience.
Summary of the invention
The present invention provides a kind of video data handling procedure, device, system and medium, with solve it is in the prior art extremely A kind of few technical problem.
On the one hand, the present invention provides a kind of video data handling procedures, comprising:
Receive the location information acquisition request that terminal is sent;The location information acquisition request is the terminal in response to right The interaction trigger action of target object is generated in current video frame, and the location information acquisition request works as forward sight including described The target video identification information of frequency;
Based on the location information acquisition request, the video object position to match with the target video identification information is searched Confidence breath;The video object location information is used to characterize each the video object in video in each frame picture of default frame sequence Coordinate position;
Searched the video object location information is sent to the terminal, so that the terminal is based on the video pair searched Interactive placement information is determined as the object information of location information and the target object, and based on the current video One video data, the interactive information of the interaction trigger action and the interactive placement information export the second video data.
On the other hand, the present invention also provides a kind of video data handling procedures, comprising:
In response to the interaction trigger action to target object in current video frame, location information acquisition request is generated;It is described Location information acquisition request includes the target video identification information of the current video;
The location information acquisition request is sent to server;The location information acquisition request is used to indicate the service Device searches the video object location information to match with the target video identification information;
Receive the video object location information searched that the server is sent;
Based on the object information of the video object location information and the target object searched, determine that interactive placement is believed Breath;
The interactive information of the first video data, the interaction trigger action based on the current video and the interaction position Confidence breath, exports the second video data.
On the other hand, the present invention also provides a kind of video data processing apparatus, comprising:
Request receiving module, for receiving the location information acquisition request of terminal transmission;The location information acquisition request Target video identification information including current video;The location information acquisition request is the terminal in response to described current The interaction trigger action of target object is generated in video frame;
Information searching module is searched and the target video identification information for being based on the location information acquisition request The video object location information to match;The video object location information is for characterizing in video each the video object in default frame Coordinate position in each frame picture of sequence;
Information sending module, for sending searched the video object location information to the terminal, so that the terminal Interactive placement information, Yi Jiji are determined based on the object information of the video object location information and the target object searched Come in the first video data of the current video, the interactive information of the interaction trigger action and the interactive placement information defeated Second video data out.
On the other hand, the present invention also provides a kind of video data processing apparatus, comprising:
Generation module is requested, for generating position in response to the interaction trigger action to target object in current video frame Information acquisition request;The location information acquisition request includes the target video identification information of the current video;
Request sending module, for sending the location information acquisition request to server;The location information acquisition is asked Seek the video object location information for being used to indicate that the server is searched and the target video identification information matches;
First location information receiving module, the video object position letter searched sent for receiving the server Breath;
Information receiving module, the video object location information searched sent for receiving the server;
Interactive placement determining module, for the object based on the video object location information and the target object searched Information determines interactive placement information.
On the other hand, the present invention also provides a kind of computer storage medium, at least one is stored in the storage medium Instruction, at least a Duan Chengxu, code set or instruction set, at least one instruction, at least a Duan Chengxu, code set or the instruction set It is loaded by processor and executes video data handling procedure as described above.
On the other hand, the present invention also provides a kind of video data processing systems, including terminal and server;
The terminal includes:
Memory, for storing program instruction;
Processor is executed for calling and executing the program instruction in the memory such as any one of preceding claim The video data handling procedure step;
The server includes:
Memory, for storing program instruction;
Processor executes view as described in any one of the above embodiments for calling and executing the program instruction in the memory Frequency data processing method steps.
A kind of video data handling procedure, device, system and medium provided by the invention at least have following technology effect Fruit:
The location information acquisition request that the present invention is sent by receiving terminal;The location information acquisition request is the end End is generated in response to the interaction trigger action to target object in current video frame, and the location information acquisition request includes The target video identification information of the current video;Based on the location information acquisition request, search and the target video mark Know the video object location information that information matches;Searched the video object location information is sent to the terminal, so that institute Terminal is stated based on the object information of the video object location information and the target object searched to determine interactive placement information, And the interactive information and interactive placement letter of the first video data, the interaction trigger action based on the current video Breath is to export the second video data.In this way, interaction effect can be enriched, interactive response duration is reduced, is not necessarily to terminal downloads algorithm packet And video processing is carried out, the calculation amount and power consumption of terminal are reduced, terminal hardware requirement is reduced, promotes whole interactive processing efficiency, User experience is more preferable.
Detailed description of the invention
It in order to more clearly explain the embodiment of the invention or the technical proposal in the existing technology and advantage, below will be to implementation Example or attached drawing needed to be used in the description of the prior art are briefly described, it should be apparent that, the accompanying drawings in the following description is only It is only some embodiments of the present invention, for those of ordinary skill in the art, without creative efforts, It can also be obtained according to these attached drawings other attached drawings.
Fig. 1 is the schematic diagram of implementation environment provided by one embodiment of the present invention.
Fig. 2 is the architecture diagram of background server provided by one embodiment of the present invention.
Fig. 3 is video data handling procedure flow chart provided by one embodiment of the present invention.
Fig. 4 is the video data handling procedure flow chart that another embodiment of the present invention provides.
Fig. 5 is the video data handling procedure flow chart that another embodiment of the invention provides.
Fig. 6 is the application scenarios schematic diagram of one embodiment of the invention.
Fig. 7 is server rack composition provided by one embodiment of the present invention.
Fig. 8 is the video data handling procedure flow chart that another embodiment of the invention provides.
Fig. 9 is video data processing apparatus block diagram provided by one embodiment of the present invention.
Figure 10 is the view data processing equipment block diagram that another embodiment of the present invention provides.
Figure 11 is a kind of hardware knot of equipment for realizing method provided by the embodiment of the present invention provided by the invention Structure schematic diagram.
Specific embodiment
In order to enable those skilled in the art to better understand the solution of the present invention, below in conjunction in the embodiment of the present invention Attached drawing, technical scheme in the embodiment of the invention is clearly and completely described, it is clear that described embodiment is only The embodiment of a part of the invention, instead of all the embodiments.Based on the embodiments of the present invention, ordinary skill people The model that the present invention protects all should belong in member's every other embodiment obtained without making creative work It encloses.
To make the object, technical solutions and advantages of the present invention clearer, below in conjunction with attached drawing to embodiment party of the present invention Formula is described in further detail.
Referring to FIG. 1, it illustrates the schematic diagrames of implementation environment provided by one embodiment of the present invention.The implementation environment packet It includes: terminal 120 and server 140.
Operation has sender's client in terminal 120.Terminal 120 can be mobile phone, tablet computer, E-book reader, MP3 (Moving Picture Experts Group Audio Layer III, dynamic image expert's compression standard audio level 3) player, MP4 (Moving Picture Experts Group Audio Layer IV, dynamic image expert's compression standard Audio level 4) player, pocket computer on knee and desktop computer etc..
In embodiments of the present invention, client can be any client for having video playback capability.For example, client Can be social category applications client, instant communication client, payment class applications client, game client, read client, It is exclusively used in the client etc. of video or audio and video playing.
Server 140 can be established with terminal 120 respectively by network and be communicated to connect.The network can be wireless network, It is also possible to cable network.Wireless mode include but is not limited to wireless network (such as bluetooth, Wi-Fi (Wireless Fidelity, Wireless Fidelity), ZigBee (purple honeybee) etc.), infrared ray etc.;Wired mode includes but is not limited to cable network, USB (Universal Serial Bus, universal serial bus), serial ports etc..
Server 140 can be a server, be also possible to the server cluster consisted of several servers, or It is a cloud computing service center.
In one example, Fig. 2 and 7 are please referred to, when server 140 is the aggregated structure that multiple servers are constituted, clothes Business device 140 may include: dispatch server 142, calculation server 144 and storage server 146 and access server 148.Its In, between each server can wirelessly or wired mode establish communication connection.Wireless mode includes but is not limited to nothing Gauze network (such as bluetooth, Wi-Fi (Wireless Fidelity, Wireless Fidelity), ZigBee (purple honeybee) etc.), infrared ray etc.;It is wired Mode includes but is not limited to cable network, USB (Universal Serial Bus, universal serial bus), serial ports etc..
Dispatch server 142 can be used for the video flowing in poll video database, generate video task list to be processed, And according to scheduled scheduling rule or allocation algorithm, corresponding junior's video processing is distributed to corresponding calculation server 144 and is appointed Business list, so that calculation server 144 executes corresponding video processing duties.
Calculation server 144 can be used for receiving junior's video task list of the distribution of dispatch server 142, regard to junior The corresponding video data of frequency task list carries out corresponding calculation processing and storage, obtains video processing data.The calculating service The quantity of device 144 can be one or more.If the quantity of calculation server 144 be it is multiple, it is exemplary, can be divided into node A, Node B ... node N.At this point it is possible to using the selected node B of dispatch server 142 as destination node, then other nodes It can be used as alternate node.
Storage server 146 can be used for receiving and storing the video file of user's upload.The storage server 146 may be used also Index record corresponding to the video processing data of 144 calculation processings of calculation server and corresponding video file for storage Video identifier index relative data.The quantity of the storage server 146 can be one or more.If storage server When 146 quantity is multiple, storage server 146 may include: cloud storage center for storing video file and for depositing Store up the index memory node of the index relative data.
Access server 148 can be used for receiving the location information acquisition request of terminal, which is accessed and is corresponded to Server execute corresponding operation so that the server accessed responds the data acquisition request.The server accessed It can be calculation server.If calculation server to be accessed has multiple server nodes, can by default access rules or Algorithm determines corresponding Target Acquisition node.
The quantity of above-mentioned each server, type and function are without being limited thereto, can also adjust accordingly according to actual needs It is whole.
Referring to FIG. 3, it illustrates a kind of video data handling procedures provided by one embodiment of the present invention.This method can Applied in implementation environment shown in FIG. 1.This method may include steps of:
S302, terminal generate location information and obtain in response to the interaction trigger action to target object in current video frame Request;The location information acquisition request includes the target video identification information of the current video.
In one embodiment, the current video frame is that the video that video reproduction program is currently played is (i.e. current Video) a certain frame frame picture or a certain section of frame frame picture.Terminal can play user by running video reproduction program The video chosen.The video reproduction program can be short-sighted sound interval sequence, long video program etc..The video reproduction program is in addition to tool Standby interaction function, can also have the other function such as video upload function, video sharing function, video download function.
In the playing process of current video, such as people, object, building, object, wind can be presented in each frame picture of video At least one object such as scape, region, point.Target object can be user's object of interest, or in video frame picture The object chosen.The target object can be an object in video frame, can also be multiple objects in video frame.One In embodiment, target object can be the region chosen in video or pixel collection.
It in one embodiment, can be to meet interaction condition to the interaction trigger action of target object in current video frame Under, to the selection operation for the target object that needs interact.Exemplary, which may include clicking, long-pressing, frame choosing, hooking Choosing, drop-down list is chosen, sliding select, certain gestures select or specific pattern select, it is specific input etc..The selection operation can be by Terminal is realized by input equipments such as mouse, keyboard, handle, touch screens.It is exemplary, after receiving corresponding interaction instruction, The selection interface for being directed to target object is generated, that is, is determined as meeting the interaction condition, if it is right to certain in video frame to receive user As progress frame selection operation interested, then this is used as target object by the region that frame selects.
The interaction trigger action to target object in current video frame is for generating location information acquisition request.Position Setting information acquisition request is the location information for the video object each in request current video in each frame picture.Position The target video identification information of the current video is carried in information acquisition request.Exemplary, target video identification information can be For characterizing the uniform resource locator of current video, such as video source identification information.
S304, terminal to server send the location information acquisition request.
The communication connection of terminal and server is pre-established, directly the location information can be sent to server later and obtain Request, can also be by intermediate equipment come the transfer location information acquisition request.The location information acquisition request is used to indicate institute It states server and searches the video object location information to match with the target video identification information.
S306, the location information acquisition request that server receiving terminal is sent.
After terminal and server establishes communication connection, server can be verified the enrollment status of terminal user, Respond the location information acquisition request again after authentication passes through.In one embodiment, after establishing communication connection, server It can start charging.Certainly, server can also carry out charging according to the content (video size, duration etc.) of request.
S308, server are based on the location information acquisition request, and lookup matches with the target video identification information The video object location information;The video object location information is used to characterize each the video object in video and is presetting frame sequence Coordinate position in each frame picture.
In one embodiment, server can identify letter by target video entrained in location information acquisition request Breath searches the video object location information to match with target video identification information in video location information aggregate.The view Frequency location information set can be used for the video object location information by the video of tracking processing and corresponding view processed The video identification information of frequency is associated storage.The video location information aggregate can store in local server, can also It is stored in other equipment or cloud.
The video object location information is used to characterize each the video object in video in each frame picture of default frame sequence In coordinate position.In an embodiment, which may include multiple the video objects mark and each described Coordinate data of the video object in corresponding each video frame picture.The video object location information can also include that video is drawn Multiple pixels or multiple pixel units and each pixel or each pixel unit in face are in each video frame picture Offset data.
The default frame sequence can be all frames in whole section of video, or each corresponding to certain in video is a bit of Frame corresponding to frame or stuffer fragment.The default frame sequence can be corresponding with location information acquisition request.It is exemplary, if position Setting in information acquisition request includes interaction start frame and interaction duration, then can be determined according to interaction start frame and interaction duration default Frame sequence can determine the co-ordinate position information for needing to obtain which frame sequence;Otherwise, default obtains each the video object at whole section Coordinate position in each frame picture of video.
In embodiments of the present invention, which may include:
S3081 is based on the location information acquisition request, inquires index note corresponding with the target video identification information Record.
In one embodiment, the index record is generated after carrying out tracking processing to video, if not passing through The video of tracking processing will not generate the index record.The index record is deposited for characterizing video identification information associated there In corresponding the video object location information.Index record generated is in addition to may include video identifier and right in one embodiment The index value (such as 1,2,3 ... ..) answered.Certainly, index record generated is in addition to may include video identifier and corresponding Index value can also include the storage location etc. that index record generates time and the video by tracking processing.
The index file comprising at least one index record is pre-established, searches whether to exist in indexed file and target The corresponding index record of video identifier.If finding corresponding index record, illustrate that the target video identification information is corresponding Video has already passed through tracking processing, and there is the video object location information to match with the target video identification information;If Corresponding index record is not found, then it is processed to illustrate that the corresponding video of target video identification information is not tracked, There is no the video object location informations to match.
It in another embodiment, can be one index record of each video configuration.At this point, index record may include view Frequency marking knows and associated tracking processing mark value (whether video is tracked processed).For example, place will can not be tracked The tracking processing mark value for the video managed is set as " 0 ", and the tracking processing mark value of tracked processed video is set For " 1 ", certainly, tracking processing mark value may also be configured to other values.Certainly, the index record can also include it is corresponding with Track marks storage location, label time of trace labelling etc..
S3082, according to search index as a result, being searched and the target video identification information in video location information aggregate The video object location information to match.
In one embodiment, if index record be it is generated after carrying out tracking processing to video, i.e., only carry out with The video of track processing can just have corresponding index record.At this point, if search index result is to inquire the index record, root According to the mapping relations of video identification information and the video object location information, searched and the target in video location information aggregate The video object location information that video identification information matches.If search index result is not inquire the index record, Then illustrating the target video identification information, there is no corresponding the video object location informations, then do not have to again in video location information collection It searches in conjunction, directly return prompt information (such as there is no request data, do not find matched result etc.), is used with prompt terminal This video of family is without request data.
In another embodiment, if each video corresponds to an index record.At this point, if search index result is the mesh Marking the processing mark value of tracking corresponding to video identification information is " 1 ", then is believed according to video identification information and the video object position The mapping relations of breath search the video object position to match with the target video identification information in video location information aggregate Confidence breath.If search index result is tracking corresponding to the target video identification information, processing mark value is " 0 ", illustrates this Corresponding the video object location information is not present in target video identification information, then does not have to look into video location information aggregate again It looks for, directly return prompt information (such as there is no request datas etc.), with this video of warning terminal user without request data.
S310, server send searched the video object location information to the terminal.
S312, terminal receive the video object location information searched that the server is sent.
S314, terminal are determined mutual based on the video object location information searched and the object information of the target object Dynamic location information.
It in one embodiment, include view corresponding to multiple the video objects in the searched Video object tracking information Frequency object location information.In order to realize the interaction for being directed to target object, the object for the target object that terminal is chosen according to user Information, determines the corresponding target position information of the target object from the video object location information of acquisition, and by the target pair The target position information of elephant is determined as interactive placement information.The target position that interactive element is followed by target object can be set Information realization video interactive.The object information of target object may include the object id information or target object of target object Image coordinate of the broadcast nodes and target object of place frame picture in the frame picture.
It is illustrative one, if above-mentioned the video object location information is calculated by optical flow tracking algorithm.It is logical It crosses optical flow tracking algorithm and line trace is clicked through to all pixels in video, include all pictures in obtained the video object location information The location information of vegetarian refreshments.It, can be according to the frame picture where target object when user has selected target object in current video In region or position, obtain pixel corresponding to the region or position or the region comprising multiple pixels, later basis Identified pixel determines the corresponding target position information of target object.
S316, the first video data based on the current video, the interactive information of the interaction trigger action and described Interactive placement information exports the second video data.
Interactive data is obtained according to the interactive information of interaction trigger action and interactive placement information, later according to interactive data With the first video data of current video, obtains after the second video data and export, to realize video interactive.
In one embodiment, if interactive information includes interactive element, can according to the interactive placement information of target object, Establish the binding relationship of interactive element and interactive placement information.Obtained interactive data includes interactive element and interaction member The binding relationship of element and interactive placement information.Later, the interactive data is rendered on current video, it is enterprising at video playing interface Row shows interactive element interaction effect relevant to target object.The interactive element can be system default, can also be user It is customized.It is exemplary, the interactive element include but is not limited to be interactive icons, barrage, picture, video, voice, animation or net Location etc..The interactive data include but is not limited to be animation data, video data etc..
Furthermore, it is possible to which interactive data is merged or is added in the first video data of current video, the second video is obtained Data carry out video playing displaying on video playing interface, can also be after playing display to other equipment or other users Second video data is sent, to further realize video interactive.
By precalculating and storing the location information of the video object in server, device to be serviced needs abundant mutual the present invention When dynamic effect, the location information acquisition request of the video, the video object that server feedback is found can be sent to server Location information, terminal can determine interactive placement information according to the video object location information, be based on the interactive placement information later Realize that more flexible interaction effect is shown.In this way, not only enriching interaction effect, interactive response duration is reduced, without under terminal It carries algorithm packet and carries out video processing, reduce the calculation amount and power consumption of terminal, reduce terminal hardware requirement, promoted at whole interaction Efficiency is managed, user experience is more preferable.
Referring to FIG. 4, it illustrates a kind of video data handling procedures that another embodiment of the present invention provides.This method Executing subject be server, may include steps of:
S402 receives the location information acquisition request that terminal is sent;The location information acquisition request is that the terminal is rung Ying Yu is generated to the interaction trigger action of target object in current video frame, and the location information acquisition request includes described The target video identification information of current video.
S404 is based on the location information acquisition request, searches the video to match with the target video identification information Object location information;The video object location information is used to characterize each the video object in video in each frame of default frame sequence Coordinate position in picture.
Step S404 may include:
S4041 is based on the location information acquisition request, inquires index note corresponding with the target video identification information Record.
S4042, according to search index as a result, being searched and the target video identification information in video location information aggregate The video object location information to match.
Optionally, the video location information aggregate can according to preset track algorithm, to video to be tracked carry out with Obtained by track processing.
S406, Xiang Suoshu terminal sends searched the video object location information, so that the terminal is based on being searched The object information of the video object location information and the target object works as forward sight to determine interactive placement information, and based on described First video data of frequency, the interactive information of the interaction trigger action and the interactive placement information export the second video counts According to.
It is consistent with previous embodiment in the specific implementation details and effect of the above-mentioned steps S402-S406 of the present embodiment, be It avoids repeating, details are not described herein.
In one embodiment, which can be obtained by server oneself foundation.At this point, the side Method can also include:
S401, server establish video location information aggregate.
Server carries out tracking calculating to the video object in video to be tracked, obtains comprising at least one the video object position The video location information aggregate of information.This step can specifically include:
S4012, obtains the tracing task list of video to be tracked, the tracing task list include at least one it is described to Track the video identification information of video.
According to predetermined period, server is polled the video data in video database, obtains at without tracking The video identification information of the video to be tracked of reason, generates corresponding tracing task list.The predetermined period interval can with minute, Time or number of days calculate.For example including but be not limited to be polled video database every 5~30 minutes, with obtain to Track video.The video database can be the database corresponding to certain video reproduction program and for storing video resource. In embodiments of the present invention, which is to place beyond the clouds, and the video data stored in video database is to use Family be uploaded to cloud and store.Specifically, the user of each department can pass through nearest access node uploaded videos file to cloud End storage, the relevant information of video file is imported in database, obtain video database later.
Tracing task list can be determined according to the sequencing of poll time;Or according to scheduled rule Obtained task list is ranked up to video to be tracked, the scheduled ordering rule include but is not limited to be to be arranged according to the time Sequence, the sequence of video temperature, clicking rate, forward rate, playback volume etc..It certainly, can be every if preset ordering rule is multiple A ordering rule configures weight, and corresponding alignment score is obtained according to the sum of weight, is determined according to alignment score result each The task number of video to be tracked generates corresponding tracing task list.
S4014 is based on the tracing task list, obtains corresponding video to be tracked.
Server is based on the tracing task list, can send downloading request to the cloud for storing video file, obtain Corresponding video to be tracked.
In one embodiment, it before obtaining corresponding video to be tracked, is also based in tracing task list and is remembered Whether the video identifier of the video to be tracked carried, inquire the video identifier of the video to be tracked in corresponding index file It is processed to carry out tracking.If it is determined that tracking processing has been carried out, then do not download at corresponding video to be tracked and subsequent tracking Manage step.If it is determined that carrying out tracking processing not yet, then corresponding video to be tracked is downloaded.
Although different users can respective uploaded videos, the video uploaded may be same video, such as have The video of standby same video information, the video information can be source video mark, current video mark etc..So in the task of acquisition column Before the corresponding video data of table, the video identifier of the video to be tracked can be inquired, can specifically inquire this to Tracking video has corresponding the video object location information or intermediate file with the presence or absence of no, and if it exists, then illustrates the view to be tracked Frequency has been performed for tracking processing task, without repeating downloading and subsequent tracking processing;This tracking can be skipped to appoint Business, into next tracing task.
Certainly, after server gets video task tracking list, list can be tracked directly in response to the video task, under Carry the corresponding video data of video task tracking list.It is of course also possible to forward pre- to sorting in video task tracking list The video to be tracked of fixed number amount carries out priority processing, determines goal task list;It is obtained later according to the goal task list Video data corresponding with the video identification information of video to be tracked.
S4016 is based on preset track algorithm, carries out tracking processing at least one described video to be tracked respectively, obtain To the video location information aggregate comprising at least one the video object location information.
The preset track algorithm can be common track algorithm, be including but not limited to particle filter (pf) tracking Method, average drifting (meanshift) tracing, KLT tracing or light stream (Lucas) method etc..
In embodiments of the present invention, tracking processing is carried out using optical flow tracking algorithm.Light stream is that space motion object exists Observe the instantaneous velocity of the pixel motion on imaging surface.Exemplary, according to optical flow tracking algorithm, at least one is to be tracked to described Video carries out tracking processing, may include:
S40161 obtains the characteristic information of target object and the target in the previous frame picture of the video to be tracked The previous frame Optical-flow Feature point of object.
S40162 determines the spy in present frame picture with the target object according to the characteristic information of the target object The corresponding current Optical-flow Feature point of reference manner of breathing.
S40163 is determined according to the location information of the location information of previous frame Optical-flow Feature point and current Optical-flow Feature point The positional shift information (such as offset) of the target object in the video frame, obtains corresponding the video object location information.
S4018 saves the video location information aggregate.
By optical flow tracking algorithm, tracking processing can be carried out to the pixel in frame picture each in video data, obtained To the positional shift information comprising all pixels point.
It in one embodiment, can be to the view for the video to be tracked for carrying out tracking processing before carrying out tracking processing Frequency carries out tracking processing to compressed the video object later according to compression processing is carried out.Specifically, optical flow tracking can be used Algorithm is to compressed the video object.For example, if video compress to 1/4 original size is tracked, i.e., with every 4 Original image pixel is a unit record offset, and the length and width of obtained video pictures will reduce twice, required at this time to deposit The file size of storage reduces 4 times.It should be noted that the compression ratio before tracking processing can be according to practical application request, network-like Condition and hardware performance adjust with carrying out suitability.
In one embodiment, video location information aggregate can be stored in the local disk of server.Specifically, can The video object location information in video location information aggregate is generated corresponding intermediate file, intermediate file is kept later In local disk.In addition, being stored in local disk after the intermediate file can be compressed according to certain compression ratio. The compression ratio can be 5~100 times of compression ratios.By way of being locally stored, loading velocity can be promoted, save machine at This;Storage occupied space can be reduced by compressing storage.
It should be noted that can store video location information corresponding with corresponding video identification information for the intermediate file Relationship.After server obtains the location information acquisition request, according to the corresponding relationship of the video object location information and video identifier, The video object location information to match with the target video identification information is searched from video location information aggregate, and will inquiry Terminal is returned to the video object location information to match.
In one embodiment, described to save the video location information aggregate, may include:
S40182 is grouped according to video information to obtained each the video object location information is handled by tracking.
Optionally, to the video information for handling video corresponding to obtained each the video object location information by tracking Analyzed, extract video features parameter, for example, big classification belonging to video (such as life kind, science and technology, entertain class, knowledge class, Information class etc.), small classification (such as life kind can be divided into mother and baby's class, makeups skin type, cuisines class etc. again) belonging to video, video wound Build the time (year, month, day etc.), video size etc..Later according to extracted video features parameter, handle tracking is had already passed through The corresponding information of video be grouped, obtain the size groups of different classifications.
S40184 stores each the video object location information according to group result respectively.
According to group result, obtained each the video object location information is stored respectively, such as according to video institute Belong to small classification belonging to big classification, video and the video creation time is grouped storage.Furthermore, it is possible to for each group of configuration group number and Storage location can recorde group number corresponding with video identifier and storage location in index record.
Correspondingly, it may include that the corresponding video of target video is special that terminal, which is sent in the location information acquisition request of server, Levy parameter.In this way, search efficiency can be improved, shorten interaction by being in advance that the video object location information is grouped storage Response time improves interaction effect.
In one embodiment, after carrying out tracking processing, the method can also include:
S502 generates index record associated with video identification information based on tracking processing result.
There are corresponding the video object positions to believe for characterizing video identification information associated there for the index record Breath.
The index record is sent to index memory node by S504, so that the index memory node establishes index text Part.
In embodiments of the present invention, index file can be K-V index file, Hash table.Index memory node can be set It sets except server.If server is the aggregated structure for including multiple servers, deposited by establishing index except server Node is stored up, the index record that index memory node is uploaded to store each server can before searching the video object location information The first search index record on index memory node, guides according to index record operate in next step later.
Exemplary, index memory node includes K-V index file, and wherein K can be with corresponding video identification information, and V can be with Corresponding index record.The corresponding intermediate file of video identification information or the video object location information if it exists, then illustrate to regard Frequency identification information is tracked processed, and corresponding index record result can be " true " (represent and exist), otherwise, rope Drawing result can be " faulse " (representative is not present).
By establishing index file in index memory node, so when receiving inquiry request, index can be first passed through in advance Memory node searches the video identification information with the presence or absence of corresponding index record, if so, then returning to existing result;If not yet Have, then return to the query result being not present, then avoid frequently executing look-up command in the server, improve overall treatment efficiency and Interaction effect.
In one embodiment, described to be based on the location information acquisition request, it searches and the target video identification information The video object location information to match, can specifically include:
S602 is based on the location information acquisition request, generates the search index for carrying the target video identification information Request.
S604 sends the search index request to the index memory node;The search index request is used to indicate The index memory node inquiry whether there is index record corresponding with the target video identification information.
S606, if the query result is to search in video location information aggregate and institute there are when the index record State the video object location information that target video identification information matches.
S608, if the query result is to return to prompt information to user there is no when the index record.
In embodiments of the present invention, it is that there is no matched video identification informations in query result, then illustrates the video mark Knowing information does not have a corresponding intermediate file, can so return and prompt to user, for example, can in the form of explanatory note or voice shape Formula returns to query result, such as " not finding matched result " etc..In one embodiment, if do not inquire matched result, It can terminate corresponding tracking trigger action.
In this way, data query speed so can be improved, reduce data by establishing index data in index memory node Delay improves user experience.
Referring to FIG. 5, it illustrates a kind of video data handling procedures that another embodiment of the present invention provides.This method Executing subject be terminal, may include steps of:
S702 generates location information acquisition request in response to the interaction trigger action to target object in current video frame; The location information acquisition request includes the target video identification information of the current video.
S704 sends the location information acquisition request to server;The location information acquisition request is used to indicate institute It states server and searches the video object location information to match with the target video identification information.
S706 receives the video object location information searched that the server is sent.
S708 determines interaction position based on the object information of the video object location information and the target object searched Confidence breath.
S710, the first video data based on the current video, the interactive information of the interaction trigger action and described Interactive placement information exports the second video data.
In one embodiment, the step S708 may include:
S7082, based on the object information of the target object, from the video object location information searched described in extraction The corresponding target position information of target object.
The object information of target object may include the object id information or target object place frame picture of target object Image coordinate in the frame picture of broadcast nodes and target object.It include multiple videos pair in the video object location information The location information of elephant can extract the target from the video object location information of lookup according to the object information of target object The corresponding target position information of object.The target position information may include target object mark and target object corresponding every Coordinate data in a video frame picture also may include the target object in video pictures in included pixel or target The offset numbers of multiple pixel units and the pixel included by object or the pixel unit in each video frame picture According to.
In addition, when determining target position information, can be uploaded according to the broadcasting attribute of current video and user by with The video attribute of track processing, obtains location information correction factor, according to the location information correction factor come the target position to extraction Confidence breath is adjusted correspondingly, the target position information updated.In this way, current video broadcast interface size, broadcast Put clarity and code rate and when tracked video that user is uploaded has differences, can by the location information correction factor come Extracted target position information is adjusted, to guarantee the accuracy of the target position information of finally determining target object, is avoided It is subsequent interactive placement offset occur and influence interaction effect.
S7084 is based on the target position information, determines interactive placement information.
In one embodiment, after the target position information of target object has been determined, which can be made For interactive placement information.It can be respectively that each object configures different interactive placements if the quantity of target object is multiple Information.
In one embodiment, the step S710 may include:
S7102 determines the interactive element and the target object based on the interactive information of the interaction trigger action Relative position.
In one embodiment, the interactive information may include interactive element, interactive configuration information and interaction content.This is mutually Dynamic element include but is not limited to be interactive icons, barrage, picture, video, voice, animation or network address etc..The interactive data includes But be not limited to animation data, video data etc..Interactive configuration information include but be not limited to the allocation position of interactive element, mutually Displaying time, the displaying size of interactive element, quantity, the time etc. of dynamic element.Interaction content includes associated with interactive element Content of text, web page contents etc..It is exemplary, if interactive element is barrage, the corresponding text in barrage frame of interaction content This content.
The relative position includes the interactive placement between interactive element and target object.It is exemplary, in interaction, interaction Element can show the positions such as the top in target object, left side, right side, lower section, middle part.
S7104 is based on the interactive placement information and the relative position, determines the interactive element in the frame of corresponding frame The target interactive placement information of picture.
According to the relative position of interactive element and mutual moving-target, position data in interactive placement information is adjusted, is obtained To interactive element the frame picture of corresponding frame target interactive placement information.The target interactive placement information is that interactive element will open up Show the interaction locus on video playing interface.
S7106 is based on the target interactive placement information and the interactive information, generates interactive data.
S7108, the first video data and the interactive data based on the current video export the second video data.
In one embodiment, interactive data can be merged or is added in the first video data of current video, obtained Second video data plays second video data on video playing interface, realizes interaction display.During displaying, base In target interactive placement information, interactive element can be opened up on video playing interface by modes such as floating layer or intercalations Show, and then interactive element is presented and follows the effect that target object is moved, increases interaction effect, promote user experience.
In addition, second video data can also be sent to other equipment or other users after playing display, with into one Step realizes video interactive.
It should be noted that the execution sequence of above-mentioned steps is without being limited thereto, it is adjusted correspondingly according to actual needs. The specific implementation details and effect of above-mentioned steps S702 to S710 are consistent with previous embodiment, in order to avoid repeating, herein no longer It repeats.
The embodiment of the present invention can promote the treatment effeciency of terminal, reduce the power consumption of terminal, greatly reduce terminal hardware and want It asks, while in advance that the information processing of needs is good, shortens period of reservation of number, user experience is more preferable.By by interactive element It is contacted with target following location information by target object, realizes and interactive element is realized according to target following location information Flexible exhibition method, so can flexible modulation interactive element display location, the association for enhancing interactive element and target object is mutual It is dynamic, improve interaction effect and user experience.
In order to make it easy to understand, now in conjunction with specific application scenarios, the present invention will be described.
As shown in fig. 6, under video interactive scene, if desired to terminal be playing in video target object (such as Kitten in figure) carry out tracking interaction.User's trigger control, pause play current video, click kitten at currently playing interface, it Location information acquisition request is generated afterwards, which is sent to server by terminal, and server responds the position Information acquisition request simultaneously feeds back searched the video object location information to terminal.Terminal obtains the video object location information Afterwards, it according to the broadcast nodes for the coordinate position and current video frame for clicking kitten, is extracted from the video object location information small The corresponding target position information of cat.If the interactive element is icon (Fishbone in figure), which is added to selected Click location.If the interactive element is barrage, the barrage text (" I will eat fish " in figure) of available user's input.Exist later In video display process, which is added on the crown with kitten.With the broadcasting of video, the position of cat changes, The position of icon and barrage is also followed by cat movement, realizes the interaction effect of dynamically track, increases the interest interacted between user Taste.
Fig. 7 and 8 are please referred to, it illustrates a kind of video data handling procedures provided by one embodiment of the present invention.The party Method can be applied to the server end in application environment shown in Fig. 2.Server 140 is the aggregated structure being made of multiple servers When, server 140 may include: dispatch server 142, calculation server 144, storage server 146 and access server 148.This method may include steps of:
S802, server establish Video object tracking information aggregate.
In embodiments of the present invention, Video object tracking information aggregate is established by multiple servers to execute, and specifically may be used To include:
S8021, dispatch server obtain the tracing task list of video to be tracked, and the tracing task list includes at least The video identification information of one video to be tracked.
S8022, dispatch server generate multiple junior's tracing task lists according to the information of the video to be tracked.
S8023 sends the tracking to the video to be processed to corresponding target computing nodes and handles request, the tracking Processing request includes junior's tracing task list.
Junior's tracing task list includes the video identification information of at least one video to be tracked.
The target computing nodes are video information and Hash consistency of the dispatch server according to the video to be tracked What algorithm determined.The video information can be video identification information.
What Hash consistency algorithm migrated clockwise when increasing and delete by annular hash space, dummy node and node Rule realizes and still is able to externally provide good service when the interstitial content of system changes, and becomes in nodal information When more, it can guarantee that original allocated content can be mapped in new buffer area, without being mapped to old delay Other buffer areas in punching set;The load balancing for effectively realizing node simultaneously, reduces the load of buffering as far as possible.
S8022, target computing nodes described in target computing nodes are based on junior's tracing task list, obtain corresponding Video to be tracked.
S8024, target computing nodes are based on preset track algorithm, carry out respectively at least one described video to be tracked Tracking processing, obtains the video location information aggregate comprising at least one the video object location information.
S8026, target computing nodes save the video location information aggregate.
S804, access server receive the location information acquisition request that terminal is sent.
Access server and terminal are established and communicated to connect, after a connection is established, access server 146 can obtain end Hold the location information acquisition request sent.Target video identification information is carried in the location information acquisition request.
It should be noted that access server and dispatch server here can be merged into one, or respectively The server being separately provided.
S806, access server determine Target Acquisition according to the target video identification information and Hash consistency algorithm Node.
Access server determines Target Acquisition node according to target video identification information and Hash consistency algorithm.The mesh The target computing nodes that mark obtains node and obtained by Hash consistency algorithm are corresponding.In embodiments of the present invention, for phase Same video information is the same node by the obtained target computing nodes of Hash consistency algorithm and Target Acquisition node.
S808, access server send the location information acquisition request to the Target Acquisition node.
Position letter acquisition request be used to indicate the Target Acquisition node is searched from video location information aggregate and The video object location information corresponding to the target video identification information.
S810, Target Acquisition node obtain the location information acquisition request that the access server is sent, and store to index Node sends the inquiry request for carrying the target video identification information.
S812, Target Acquisition node receive the query result of the index memory node feedback.
S814, if Target Acquisition node is there are when the index record, in the view being locally stored in the query result The video object location information to match with the target video identification information is searched in frequency location information set.
S816, Target Acquisition node send found the video object location information to the access server.
The determining the video object location information to match can be sent to access server by Target Acquisition node, by Access server is sent to terminal.Certainly, Target Acquisition node can also be established directly with terminal and be communicated to connect, and directly be sent out to terminal The Video object tracking information for giving this to match.
The present invention can so carry out large batch of video processing, pre-process by using the server of aggregated structure Video location information promotes loading velocity by way of being locally stored, and saves machine cost, compared to tradition such as terminal operatings Tracing algorithm deployment way, have the advantages that efficient, simple.
Following is apparatus of the present invention embodiment, can be used for executing embodiment of the present invention method.For apparatus of the present invention reality Undisclosed details in example is applied, embodiment of the present invention method is please referred to.
Referring to FIG. 9, it illustrates a kind of block diagrams of video data processing apparatus provided by one embodiment of the present invention.It should Device has the function of server side in realization above method example, and the function can be by hardware realization, can also be by hardware Execute corresponding software realization.The apparatus may include:
Request receiving module 910, for receiving the location information acquisition request of terminal transmission;The location information acquisition is asked Terminal described in Seeking Truth is generated in response to the interaction trigger action to target object in current video frame, and the location information obtains Taking request includes the target video identification information of the current video;
Information searching module 920 is searched and is believed with target video mark for being based on the location information acquisition request The matched the video object location information of manner of breathing;The video object location information is being preset for characterizing each the video object in video Coordinate position in each frame picture of frame sequence;
Information sending module 930, for sending searched the video object location information to the terminal, so that the end End group determines interactive placement information in the object information of the video object location information and the target object searched, and The first video data based on the current video, the interactive information of the interaction trigger action and the interactive placement information come Export the second video data.
In one embodiment, information searching module 920 may include:
Submodule is inquired, for being based on the location information acquisition request, inquiry and the target video identification information pair The index record answered;
Submodule is searched, for regarding as a result, searching in video location information aggregate with the target according to search index The video object location information that frequency identification information matches
In one embodiment, which can also include:
Information aggregate establishes module, for establishing video location information aggregate.
The information aggregate establishes module
Task acquiring unit, for obtaining the tracing task list of video to be tracked, the tracing task list includes extremely The video identification information of a few video to be tracked;
Video data acquiring unit obtains corresponding video to be tracked for being based on the tracing task list;
Information combination determination unit is wrapped for carrying out tracking processing at least one described video to be tracked respectively Video location information aggregate containing at least one the video object location information;
Information memory cell, for saving the video location information aggregate.
In one embodiment, described device can also include:
Task acquisition module, for obtaining the tracing task list of video to be tracked;
Subtask generation module generates multiple junior's tracing task lists according to the information of the video to be tracked;
Request sending module is handled, for sending to the tracking of the video to be processed to corresponding target computing nodes Reason request, the tracking processing request include junior's tracing task list, and junior's tracing task list includes at least The video identification information of one video to be tracked, so that the target computing nodes are arranged based on junior's tracing task Table obtains corresponding video to be tracked;And be based on preset track algorithm, respectively at least one described video to be tracked into Line trace processing, obtains the video location information aggregate comprising at least one the video object location information;And save the view Frequency location information set;
Information receiving module, the video location information aggregate sent for receiving the target computing nodes.
In one embodiment, described device can also include:
Generation module is recorded, for generating index record associated with video identification information based on tracking processing result, There are corresponding the video object location informations for characterizing video identification information associated there for the index record;
Sending module is recorded, for the index record to be sent to index memory node, so that index storage section Point establishes index file.
In one embodiment, the information searching module 920 may include:
Search index request unit generates for being based on the location information acquisition request and carries the target video mark Know the search index request of information;
Search index transmission unit, for sending the search index request to the index memory node;The index Inquiry request is used to indicate the index memory node inquiry with the presence or absence of index corresponding with the target video identification information Record;
Indexed results receiving unit, for receiving the query result of the index memory node feedback;
Information searching unit, if being there are when the index record, in video location information collection for the query result The video object location information to match with the target video identification information is searched in conjunction.
Referring to FIG. 10, it illustrates a kind of frames of video data processing apparatus device provided by one embodiment of the present invention Figure.The device has the function of terminal side in realization above method example, and the function can be by hardware realization, can also be by hard Part executes corresponding software realization.The apparatus may include:
Generation module 1002 is requested, for generating in response to the interaction trigger action to target object in current video frame Location information acquisition request;The location information acquisition request includes the target video identification information of the current video;
Request sending module 1004, for sending the location information acquisition request to server;The location information obtains It takes request to be used to indicate the server and searches the video object location information to match with the target video identification information;
Information receiving module 1006, the video object location information searched sent for receiving the server;
Interactive placement determining module 1008, for based on the video object location information and the target object searched Object information determines interactive placement information;
Output module 1010, for based on the current video the first video data, it is described interaction trigger action it is mutual Dynamic information and the interactive placement information export the second video data.
In one embodiment, the interactive placement determining module 1008 can also include:
Target position determination unit, for the object information based on the target object, from the video object position searched The corresponding target position information of the target object is extracted in confidence breath;
Interactive placement determination unit determines interactive placement information for being based on the target position information;
Wherein, the object information of the target object include the target object object identity and/or the target pair As the image coordinate in the frame picture of broadcast nodes and target object of place frame picture.
In one embodiment, the output module 1010 can also include:
Relative position determination unit determines the interactive element for the interactive information based on the interaction trigger action With the relative position of the target object;The interactive information may include interactive element, interactive configuration information and interaction content;
Target interactive placement determination unit, described in determining based on the interactive placement information and the relative position Target interactive placement information of the interactive element in the frame picture of corresponding frame;
Interactive data generation unit generates interaction for being based on the target interactive placement information and the interactive information Data;
Output unit, for the first video data and the interactive data based on the current video, the second view of output Frequency evidence.
The embodiment of the invention also provides a kind of computer storage medium, at least one finger is stored in the storage medium Enable, at least a Duan Chengxu, code set or instruction set, at least one instruction, at least a Duan Chengxu, code set or instruction set by Processor loads and executes described in the embodiment of the present invention that a kind of video data handling procedure step, specific implementation procedure can be square Method embodiment illustrates, herein without repeating.
The embodiment of the invention also provides a kind of video data processing systems, including terminal and server.
The terminal includes: memory, for storing program instruction;Processor, for calling and executing the memory In program instruction, execute video data handling procedure step as described in the claim embodiment of the present invention.
The server includes: memory, for storing program instruction;Processor, for calling and executing the storage Program instruction in device executes video data handling procedure step as described in the embodiments of the present invention.
Further, Figure 11 shows a kind of hardware knot of equipment for realizing method provided by the embodiment of the present invention Structure schematic diagram, the equipment can be terminal, mobile terminal or other equipment, the equipment may also participate in composition or Include device provided by the embodiment of the present invention.As shown in figure 11, terminal 10 may include one or more (adopts in figure With 102a, 102b ... ..., 102n is shown) processor 102 (processor 102 can include but is not limited to Micro-processor MCV or The processing unit of programmable logic device FPGA etc.), memory 104 for storing data and the biography for communication function Defeated device 106.It in addition to this, can also include: display, input/output interface (I/O interface), universal serial bus (USB) Port (a port that can be used as in the port of I/O interface is included), network interface, power supply and/or camera.This field is general Logical technical staff is appreciated that structure shown in Figure 11 is only to illustrate, and does not cause to limit to the structure of above-mentioned electronic device. For example, terminal 10 may also include than shown in Figure 11 more perhaps less component or have with shown in Figure 11 not Same configuration.
It is to be noted that said one or multiple processors 102 and/or other data processing circuits lead to herein Can often " data processing circuit " be referred to as.The data processing circuit all or part of can be presented as software, hardware, firmware Or any other combination.In addition, data processing circuit for single independent processing module or all or part of can be integrated to meter In any one in other elements in calculation machine terminal 10 (or mobile device).As involved in the embodiment of the present application, The data processing circuit controls (such as the selection for the variable resistance end path connecting with interface) as a kind of processor.
Memory 104 can be used for storing the software program and module of application software, as described in the embodiment of the present invention Corresponding program instruction/the data storage device of method, the software program that processor 102 is stored in memory 104 by operation And module realizes a kind of above-mentioned Processing with Neural Network method thereby executing various function application and data processing.It deposits Reservoir 104 may include high speed random access memory, may also include nonvolatile memory, as one or more magnetic storage fills It sets, flash memory or other non-volatile solid state memories.In some instances, memory 104 can further comprise relative to place The remotely located memory of device 102 is managed, these remote memories can pass through network connection to terminal 10.Above-mentioned network Example include but is not limited to internet, intranet, local area network, mobile radio communication and combinations thereof.
Transmitting device 106 is used to that data to be received or sent via a network.Above-mentioned network specific example may include The wireless network that the communication providers of terminal 10 provide.In an example, transmitting device 106 includes that a network is suitable Orchestration (Network Interface Controller, NIC), can be connected by base station with other network equipments so as to Internet is communicated.In an example, transmitting device 106 can be radio frequency (Radio Frequency, RF) module, For wirelessly being communicated with internet.
Display can such as touch-screen type liquid crystal display (LCD), the liquid crystal display aloow user with The user interface of terminal 10 (or mobile device) interacts.
It should be understood that embodiments of the present invention sequencing is for illustration only, do not represent the advantages or disadvantages of the embodiments. And above-mentioned this specification specific embodiment is described.Other embodiments are within the scope of the appended claims.One In a little situations, the movement recorded in detail in the claims or step can be executed according to the sequence being different from embodiment and Still desired result may be implemented.In addition, process depicted in the drawing not necessarily requires the particular order shown or company Continuous sequence is just able to achieve desired result.In some embodiments, multitasking and parallel processing it is also possible or It may be advantageous.
All the embodiments in this specification are described in a progressive manner, same and similar portion between each embodiment Dividing may refer to each other, and each embodiment focuses on the differences from other embodiments.Especially for device and For server example, since it is substantially similar to the method embodiment, so being described relatively simple, related place is referring to side The part of method embodiment illustrates.
Those of ordinary skill in the art will appreciate that realizing that all or part of the steps of above-described embodiment can pass through hardware It completes, relevant hardware can also be instructed to complete by program, the program can store in a kind of computer-readable In storage medium, storage medium mentioned above can be read-only memory, disk or CD etc..
The foregoing is merely presently preferred embodiments of the present invention, is not intended to limit the invention, it is all in spirit of the invention and Within principle, any modification, equivalent replacement, improvement and so on be should all be included in the protection scope of the present invention.

Claims (13)

1. a kind of video data handling procedure characterized by comprising
Receive the location information acquisition request that terminal is sent;The location information acquisition request is the terminal in response to current The interaction trigger action of target object is generated in video frame, and the location information acquisition request includes the current video Target video identification information;
Based on the location information acquisition request, searches the video object position to match with the target video identification information and believe Breath;The video object location information is used to characterize seat of each the video object in each frame picture of default frame sequence in video Cursor position;
Searched the video object location information is sent to the terminal, so that the terminal is based on the video object position searched Confidence, which is ceased with the object information of the target object, determines interactive placement information, and the based on the current video first view Frequency exports the second video data according to, the interactive information of the interaction trigger action and the interactive placement information.
2. the method according to claim 1, wherein it is described be based on the location information acquisition request, search with The video object location information that the target video identification information matches, comprising:
Based on the location information acquisition request, index record corresponding with the target video identification information is inquired;
According to search index as a result, searching the view to match with the target video identification information in video location information aggregate Frequency object location information.
3. according to the method described in claim 2, it is characterized in that, the method also includes:
The tracing task list of video to be tracked is obtained, the tracing task list includes at least one described video to be tracked Video identification information;
Based on the tracing task list, corresponding video to be tracked is obtained;
Based on preset track algorithm, tracking processing is carried out at least one described video to be tracked respectively, is obtained comprising at least The video location information aggregate of one the video object location information;
Save the video location information aggregate.
4. according to the method described in claim 2, it is characterized in that, the method also includes:
Obtain the tracing task list of video to be tracked;
According to the information of the video to be tracked, multiple junior's tracing task lists are generated;
The tracking to the video to be processed is sent to corresponding target computing nodes and handles request, and the tracking handles request packet Junior's tracing task list is included, junior's tracing task list includes the video mark of at least one video to be tracked Know information, so that the target computing nodes are based on junior's tracing task list, obtains corresponding video to be tracked;And Based on preset track algorithm, tracking processing is carried out at least one described video to be tracked respectively, is obtained comprising at least one The video location information aggregate of the video object location information;And save the video location information aggregate;
Receive the video location information aggregate that the target computing nodes are sent.
5. the method according to claim 3 or 4, which is characterized in that the method also includes:
Based on tracking processing result, index record associated with video identification information is generated, the index record is for characterizing There are corresponding the video object location informations for video identification information associated there;
The index record is sent to index memory node, so that the index memory node establishes index file;
Correspondingly, described be based on the location information acquisition request, the view to match with the target video identification information is searched Frequency object location information, comprising:
Based on the location information acquisition request, the search index request for carrying the target video identification information is generated;
The search index request is sent to the index memory node;The search index request is used to indicate the index and deposits Storing up querying node whether there is index record corresponding with the target video identification information;
Receive the query result of the index memory node feedback;
If the query result is to search and the target video in video location information aggregate there are when the index record The video object location information that identification information matches.
6. according to the method described in claim 5, it is characterized in that, it is described be based on the location information acquisition request, search with The video object location information that the target video identification information matches, comprising:
According to the target video identification information and Hash consistency algorithm, Target Acquisition node is determined;The Target Acquisition section Point is corresponding with target computing nodes, and the target computing nodes are the video information and Hash one according to the video to be tracked Determined by cause property algorithm;
The location information acquisition request is sent to the Target Acquisition node;The position letter acquisition request is used to indicate described Target Acquisition node is searched and the video object position corresponding to the target video identification information from video location information aggregate Confidence breath;
Receive the video object location information found that the Target Acquisition node is sent.
7. a kind of video data handling procedure characterized by comprising
In response to the interaction trigger action to target object in current video frame, location information acquisition request is generated;The position Information acquisition request includes the target video identification information of the current video;
The location information acquisition request is sent to server;The location information acquisition request is used to indicate the server and looks into Look for the video object location information to match with the target video identification information;
Receive the video object location information searched that the server is sent;
Based on the object information of the video object location information and the target object searched, interactive placement information is determined;
The interactive information and interactive placement letter of the first video data, the interaction trigger action based on the current video Breath exports the second video data.
8. the method according to the description of claim 7 is characterized in that described based on the video object location information searched and institute The object information for stating target object determines interactive placement information, comprising:
Based on the object information of the target object, the target object pair is extracted from the video object location information searched The target position information answered;
Based on the target position information, interactive placement information is determined;
Wherein, the object information of the target object includes object identity and/or the target object institute of the target object In the image coordinate in the frame picture of broadcast nodes and target object of frame picture.
9. the method according to the description of claim 7 is characterized in that first video data based on the current video, The interactive information of the interaction trigger action and the interactive placement information, export the second video data, comprising:
Based on the interactive information of the interaction trigger action, the relative position of the interactive element and the target object is determined; The interactive information may include interactive element, interactive configuration information and interaction content;
Based on the interactive placement information and the relative position, determine the interactive element in the target of the frame picture of corresponding frame Interactive placement information;
Based on the target interactive placement information and the interactive information, interactive data is generated;
The first video data and the interactive data based on the current video export the second video data.
10. a kind of video data processing apparatus characterized by comprising
Request receiving module, for receiving the location information acquisition request of terminal transmission;The location information acquisition request includes The target video identification information of current video;The location information acquisition request is the terminal in response to the current video The interaction trigger action of target object is generated in frame;
Information searching module is searched and the target video identification information phase for being based on the location information acquisition request The video object location information matched;The video object location information is for characterizing in video each the video object in default frame sequence Each frame picture in coordinate position;
Information sending module, for sending searched the video object location information to the terminal, so that the terminal is based on The object information of the video object location information and the target object searched determines interactive placement information, and is based on institute The first video data of current video, the interactive information of the interaction trigger action and the interactive placement information are stated to export the Two video datas.
11. a kind of video data processing apparatus characterized by comprising
Generation module is requested, for generating location information in response to the interaction trigger action to target object in current video frame Acquisition request;The location information acquisition request includes the target video identification information of the current video;
Request sending module, for sending the location information acquisition request to server;The location information acquisition request is used In the video object location information for indicating that the server is searched and the target video identification information matches;
First location information receiving module, the video object location information searched sent for receiving the server;
Information receiving module, the video object location information searched sent for receiving the server;
Interactive placement determining module, for the object letter based on the video object location information and the target object searched Breath, determines interactive placement information.
12. a kind of computer storage medium, which is characterized in that be stored at least one instruction, at least one in the storage medium Duan Chengxu, code set or instruction set, at least one instruction, an at least Duan Chengxu, code set or instruction set are added by processor Carry and execute the video data handling procedure as described in claim 1 to 6 is any or the view as described in claim 7 to 9 is any Frequency data processing method.
13. a kind of video data processing system, which is characterized in that including server and terminal;
The server includes:
Memory, for storing program instruction;
Processor is executed as described in any one of claim 1 to 6 for calling and executing the program instruction in the memory Video data handling procedure step;
The terminal includes:
Memory, for storing program instruction;
Processor executes the video as described in claim 7 to 9 for calling and executing the program instruction in the memory Data processing method steps.
CN201910463414.0A 2019-05-30 2019-05-30 A kind of video data handling procedure, device, system and medium Pending CN110324672A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910463414.0A CN110324672A (en) 2019-05-30 2019-05-30 A kind of video data handling procedure, device, system and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910463414.0A CN110324672A (en) 2019-05-30 2019-05-30 A kind of video data handling procedure, device, system and medium

Publications (1)

Publication Number Publication Date
CN110324672A true CN110324672A (en) 2019-10-11

Family

ID=68119193

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910463414.0A Pending CN110324672A (en) 2019-05-30 2019-05-30 A kind of video data handling procedure, device, system and medium

Country Status (1)

Country Link
CN (1) CN110324672A (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111263185A (en) * 2020-02-27 2020-06-09 中电福富信息科技有限公司 Video service cross-region access method based on GBT2260-2007 national standard coding specification
CN111314713A (en) * 2020-02-27 2020-06-19 中电福富信息科技有限公司 Nearby access method of video service based on GBT2260-2007 national standard coding specification
CN112637541A (en) * 2020-12-23 2021-04-09 平安银行股份有限公司 Audio and video labeling method and device, computer equipment and storage medium
CN113179445A (en) * 2021-04-15 2021-07-27 腾讯科技(深圳)有限公司 Video sharing method based on interactive article and interactive article
CN114339452A (en) * 2022-01-25 2022-04-12 福建中科多特健康科技有限公司 Video playing method and storage device
WO2022073409A1 (en) * 2020-10-10 2022-04-14 腾讯科技(深圳)有限公司 Video processing method and apparatus, computer device, and storage medium
CN114584824A (en) * 2020-12-01 2022-06-03 阿里巴巴集团控股有限公司 Data processing method and system, electronic equipment, server and client equipment
CN117201873A (en) * 2023-11-07 2023-12-08 湖南博远翔电子科技有限公司 Intelligent analysis method and device for video image

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105847998A (en) * 2016-03-28 2016-08-10 乐视控股(北京)有限公司 Video playing method, playing terminal, and media server
CN106096062A (en) * 2016-07-15 2016-11-09 乐视控股(北京)有限公司 video interactive method and device
CN106210808A (en) * 2016-08-08 2016-12-07 腾讯科技(深圳)有限公司 Media information put-on method, terminal, server and system

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105847998A (en) * 2016-03-28 2016-08-10 乐视控股(北京)有限公司 Video playing method, playing terminal, and media server
CN106096062A (en) * 2016-07-15 2016-11-09 乐视控股(北京)有限公司 video interactive method and device
CN106210808A (en) * 2016-08-08 2016-12-07 腾讯科技(深圳)有限公司 Media information put-on method, terminal, server and system

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111263185A (en) * 2020-02-27 2020-06-09 中电福富信息科技有限公司 Video service cross-region access method based on GBT2260-2007 national standard coding specification
CN111314713A (en) * 2020-02-27 2020-06-19 中电福富信息科技有限公司 Nearby access method of video service based on GBT2260-2007 national standard coding specification
WO2022073409A1 (en) * 2020-10-10 2022-04-14 腾讯科技(深圳)有限公司 Video processing method and apparatus, computer device, and storage medium
EP4106337A4 (en) * 2020-10-10 2023-10-18 Tencent Technology (Shenzhen) Company Limited Video processing method and apparatus, computer device, and storage medium
CN114584824A (en) * 2020-12-01 2022-06-03 阿里巴巴集团控股有限公司 Data processing method and system, electronic equipment, server and client equipment
CN112637541A (en) * 2020-12-23 2021-04-09 平安银行股份有限公司 Audio and video labeling method and device, computer equipment and storage medium
CN113179445A (en) * 2021-04-15 2021-07-27 腾讯科技(深圳)有限公司 Video sharing method based on interactive article and interactive article
CN114339452A (en) * 2022-01-25 2022-04-12 福建中科多特健康科技有限公司 Video playing method and storage device
CN117201873A (en) * 2023-11-07 2023-12-08 湖南博远翔电子科技有限公司 Intelligent analysis method and device for video image
CN117201873B (en) * 2023-11-07 2024-01-02 湖南博远翔电子科技有限公司 Intelligent analysis method and device for video image

Similar Documents

Publication Publication Date Title
CN110324672A (en) A kind of video data handling procedure, device, system and medium
US20210349604A1 (en) Shared Virtual Area Communication Environment Based Apparatus and Methods
KR102324987B1 (en) Method and apparatus for resource allocation
CN110809175B (en) Video recommendation method and device
US20090070673A1 (en) System and method for presenting multimedia content and application interface
CN106846040A (en) Virtual present display methods and system in a kind of direct broadcasting room
CN104243598B (en) A kind of information recommendation method and device
CN109348237A (en) A kind of direct broadcasting room management method, device, server and storage medium
US11197074B2 (en) Synchronized video annotation and control system for sexual stimulation devices
CN106062806A (en) Utilizing interactivity signals to generate relationships and promote content
US20180063064A1 (en) Modifying a mood through selective feeding of content
CN109600628A (en) Video creating method, device, computer equipment and storage medium
JP2023551476A (en) Graphic interchange format file identification for inclusion in video game content
US10798446B2 (en) Content narrowing of a live feed based on cognitive profiling
JP2021503111A (en) Recommendation method and system that uses content including location data
US9721321B1 (en) Automated interactive dynamic audio/visual performance with integrated data assembly system and methods
US20170017382A1 (en) System and method for interaction between touch points on a graphical display
CN109462777B (en) Video heat updating method, device, terminal and storage medium
EP2775703B1 (en) Method and apparatus for managing crowd sourced content creation
US11040278B2 (en) Server device distributing video data and replay data and storage medium used in same
CN112492328A (en) Virtual room creating method, device, terminal and storage medium
CN111683280A (en) Video processing method and device and electronic equipment
WO2023045912A1 (en) Selective content transfer for streaming content
CN115146087A (en) Resource recommendation method, device, equipment and storage medium
CN104737231A (en) Information processing device, information processing method and program, and information processing system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20191011