CN112866786A - Video data processing method and device, terminal equipment and storage medium - Google Patents

Video data processing method and device, terminal equipment and storage medium Download PDF

Info

Publication number
CN112866786A
CN112866786A CN202110051307.4A CN202110051307A CN112866786A CN 112866786 A CN112866786 A CN 112866786A CN 202110051307 A CN202110051307 A CN 202110051307A CN 112866786 A CN112866786 A CN 112866786A
Authority
CN
China
Prior art keywords
video
image frame
timestamp
target
position information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110051307.4A
Other languages
Chinese (zh)
Inventor
谢文龙
李云鹏
吕亚亚
杨春晖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Visionvera Information Technology Co Ltd
Original Assignee
Visionvera Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Visionvera Information Technology Co Ltd filed Critical Visionvera Information Technology Co Ltd
Priority to CN202110051307.4A priority Critical patent/CN112866786A/en
Publication of CN112866786A publication Critical patent/CN112866786A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7837Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/7867Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title and artist information, manually generated time, location and usage information, user ratings
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Library & Information Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The embodiment of the invention provides a method and a device for processing video data, terminal equipment and a storage medium, wherein the method comprises the following steps: receiving a first monitoring video sent by monitoring equipment; sending the first monitoring video to a target video networking terminal; the second monitoring video is sent to a video analysis server, the video analysis server determines a second time stamp of a second image frame and position information of a target object according to the second monitoring video, the first time stamp returned by the video analysis server and the position information of the target object corresponding to the first time stamp are sent to a target video network terminal, and the target video network terminal simultaneously displays the first image frame and a marking frame corresponding to the position information of the target object, so that the object in the first image frame corresponds to the position of the marking frame.

Description

Video data processing method and device, terminal equipment and storage medium
Technical Field
The present invention relates to the field of video networking technologies, and in particular, to a method and an apparatus for processing video data, a terminal device, and a storage medium.
Background
Along with the continuous development of the video networking technology, the field that involves is more and more, and video networking and internet also can carry out data transmission, can show the surveillance video of supervisory equipment on the video networking terminal, just need insert the supervisory equipment to the video networking through the control access server, acquire the surveillance video from the supervisory equipment again to show on sending this surveillance video to the video networking terminal.
When playing the surveillance video on the video networking terminal, if the surveillance video is formed by the multilayer data combination, then the video networking terminal is when playing the surveillance video, then needs a plurality of decoders to decode multilayer data, then adopts a plurality of encoders to encode the data after the multilayer is decoded again and plays, every increases one deck data, then needs to increase a set of encoder and decoder, like this, will increase the processing pressure at video networking terminal, and video broadcast efficiency is lower.
Disclosure of Invention
In view of the above problems, embodiments of the present invention are proposed to provide a video data processing method, apparatus, terminal device and storage medium that overcome or at least partially solve the above problems.
In a first aspect, an embodiment of the present invention provides a method for processing video data, where the method includes:
receiving a first monitoring video sent by monitoring equipment;
sending the first monitoring video to a target video network terminal so that the target video network terminal can acquire a first image frame in the first monitoring video and a first timestamp of the first image frame;
sending the second monitoring video to a video analysis server so that the video analysis server determines a second timestamp of a second image frame and position information of a target object in the second image frame according to the second monitoring video, wherein the first monitoring video is the same as the second monitoring video;
receiving a second timestamp returned by the video analysis server and position information of a target object corresponding to the second timestamp, and sending the second timestamp and the position information of the target object corresponding to the second timestamp to the target video network terminal, so that the target video network terminal displays the first image frame and a labeling frame corresponding to the position information of the target object at the same time in a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, and the object in the first image frame corresponds to the position of the labeling frame.
Optionally, the acquiring, by the target video network terminal, a first image frame in a first monitoring video and a first timestamp of the first image frame includes:
the target video network terminal decodes the first monitoring video to obtain the first image frame and obtains a first time stamp of the first image frame.
Optionally, the displaying, by the target internet of view terminal, the first image frame and the labeling frame corresponding to the position information of the target object simultaneously when the difference between the first timestamp and the second timestamp is smaller than a preset time range, so that the object in the first image frame corresponds to the position of the labeling frame, includes:
when the difference value of the first timestamp and the second timestamp is within a preset time range, the target video network terminal acquires a first image frame corresponding to the first timestamp;
the target video network terminal acquires the position information of the target object corresponding to the second timestamp;
the target video network terminal draws a marking frame corresponding to the position information on a preset canvas according to the position information of the target object;
and the target video network terminal simultaneously displays the first image frame and a marking frame corresponding to the position information of the target object, so that the object in the first image frame corresponds to the position of the marking frame.
Optionally, the method further comprises:
and after the target video network terminal finishes displaying the first image frame corresponding to the first timestamp, deleting the marking frame which is drawn on the preset canvas and corresponds to the second timestamp.
Optionally, the video analysis server determines, according to the second surveillance video, a second timestamp of the second image frame and location information of the target object in the second image frame, including:
the video analysis server decodes the second monitoring video to obtain a second image frame, and the second image frame is stored as an image in a YUV format;
acquiring a second timestamp of the second image frame;
and the video analysis server determines the position information of the target object in the second image frame according to the second image frame and a pre-established object recognition model.
Optionally, the determining, by the video analysis server, the position information of the target object in the second image frame according to the second image frame and a pre-established object recognition model includes:
the video analysis server inputs the second image frame into a pre-established object recognition model to obtain the image characteristics of a target object in the second image frame;
determining the similarity of the target object according to the image characteristics of the target object and preset image characteristic information in a pre-stored database, wherein the pre-stored database comprises object names and corresponding preset image characteristic information;
if the similarity of the target object is greater than or equal to a preset similarity, the video analysis server acquires position information of the target object in the second image frame, wherein the position information comprises coordinate information, and the coordinate information comprises preset vertex coordinate information, a preset length and a preset width.
In a second aspect, an embodiment of the present invention provides an apparatus for processing video data, where the apparatus includes:
the receiving module is used for receiving a first monitoring video sent by monitoring equipment;
the first sending module is used for sending the first monitoring video to a target video network terminal so that the target video network terminal can acquire a first image frame in the first monitoring video and a first timestamp of the first image frame;
the second sending module is used for sending the second monitoring video to the video analysis server so that the video analysis server can determine a second timestamp of a second image frame and position information of a target object in the second image frame according to the second monitoring video, wherein the first monitoring video is the same as the second monitoring video;
the display module is used for receiving a second timestamp returned by the video analysis server and position information of a target object corresponding to the second timestamp, and sending the second timestamp and the position information of the target object corresponding to the second timestamp to the target video network terminal, so that the target video network terminal displays the first image frame and a labeling frame corresponding to the position information of the target object at the same time within a preset time range according to the difference value between the first timestamp and the second timestamp, and the object in the first image frame corresponds to the position of the labeling frame.
Optionally, the target video networking terminal includes a decoding module, and the decoding module is configured to:
and decoding the first monitoring video to obtain the first image frame, and acquiring a first time stamp of the first image frame.
Optionally, the target video networking terminal includes a processing module, and the processing module is configured to:
acquiring a first image frame corresponding to the first timestamp when the difference value between the first timestamp and the second timestamp is within a preset time range;
acquiring position information of the target object corresponding to the second timestamp;
according to the position information of the target object, drawing a marking frame corresponding to the position information on a preset canvas;
and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object, so that the object in the first image frame corresponds to the position of the marking frame.
Optionally, the target video networking terminal further includes a deletion module, and the deletion module is configured to:
after the target video network terminal finishes displaying the first image frame corresponding to the first timestamp, deleting the marking frame corresponding to the second timestamp drawn on the preset canvas
Optionally, the video analysis server comprises an identification module, and the identification module is configured to:
decoding the second monitoring video to obtain a second image frame, and storing the second image frame as an image in a YUV format;
acquiring a second timestamp of the second image frame;
and determining the position information of the target object in the second image frame according to the second image frame and a pre-established object recognition model.
Optionally, the identification module is specifically configured to:
the video analysis server inputs the second image frame into a pre-established object recognition model to obtain the image characteristics of a target object in the second image frame;
determining the similarity of the target object according to the image characteristics of the target object and preset image characteristic information in a pre-stored database, wherein the pre-stored database comprises object names and corresponding preset image characteristic information;
if the similarity of the target object is greater than or equal to a preset similarity, the video analysis server acquires position information of the target object in the second image frame, wherein the position information comprises coordinate information, and the coordinate information comprises preset vertex coordinate information, a preset length and a preset width.
In a third aspect, an embodiment of the present invention provides a terminal device, including: at least one processor and memory;
the memory stores a computer program; the at least one processor executes the computer program stored by the memory to implement the method for processing video data provided by the first aspect.
In a fourth aspect, an embodiment of the present invention provides a computer-readable storage medium, where a computer program is stored, and when the computer program is executed, the computer program implements the video data processing method provided in the first aspect.
The embodiment of the invention has the following advantages:
according to the video data processing method, the video data processing device, the terminal equipment and the storage medium, the first monitoring video sent by the monitoring equipment is received; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
Drawings
FIG. 1 is a flow chart of the steps of an embodiment of a method for processing video data according to the present invention;
FIG. 2 is a flow chart of steps in another embodiment of a method of processing video data in accordance with the present invention;
fig. 3 is a block diagram of an embodiment of a video data processing apparatus according to the present invention;
fig. 4 is a schematic structural diagram of a terminal device of the present invention.
Detailed Description
In order to make the aforementioned objects, features and advantages of the present invention comprehensible, embodiments accompanied with figures are described in further detail below.
The video networking is an important milestone for network development, is a real-time network, can realize high-definition video real-time transmission, and pushes a plurality of internet applications to high-definition video, and high-definition faces each other.
The video networking adopts a real-time high-definition video exchange technology, can integrate required services such as dozens of services of video, voice, pictures, characters, communication, data and the like on a system platform on a network platform, such as high-definition video conference, video monitoring, intelligent monitoring analysis, emergency command, digital broadcast television, delayed television, network teaching, live broadcast, VOD on demand, television mail, Personal Video Recorder (PVR), intranet (self-office) channels, intelligent video broadcast control, information distribution and the like, and realizes high-definition quality video broadcast through a television or a computer.
Based on the characteristics of the video network, the invention provides one of the core concepts of the embodiment, and the first monitoring video sent by the monitoring equipment is received according to the protocol of the video network; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
An embodiment of the present invention provides a method for processing video data, which is used for displaying layered video data. The execution subject of this embodiment is a processing apparatus of video data, and is disposed in the monitoring access server.
Referring to fig. 1, a flowchart illustrating steps of an embodiment of a method for processing video data according to the present invention is shown, where the method specifically includes the following steps:
s101, receiving a first monitoring video sent by monitoring equipment;
specifically, the embodiment of the invention is applied to a video analysis system, which comprises a monitoring access server and a video analysis server, wherein the monitoring access server is connected with a plurality of monitoring devices and is connected with a video networking terminal. The monitoring access server obtains a first monitoring video from the monitoring device through an RTSP (Real Time Streaming Protocol) Protocol. The monitoring equipment is installed at each intersection of different road sections and used for collecting monitoring video data of each intersection.
S102, sending the first monitoring video to a target video network terminal so that the target video network terminal can acquire a first image frame in the first monitoring video and a first timestamp of the first image frame;
specifically, the monitoring access server copies the first monitoring video to obtain a second monitoring video, sends the first monitoring video to the target video network terminal, sends the second monitoring video to the video analysis server, and the target video network terminal decodes the first monitoring video to obtain a first image frame, wherein the first image frame is an I-frame image and records a first timestamp of the first image frame.
S103, sending the second monitoring video to a video analysis server so that the video analysis server determines a second timestamp of a second image frame and position information of a target object in the second image frame according to the second monitoring video, wherein the first monitoring video is the same as the second monitoring video;
specifically, the monitoring access server sends a second monitoring video to the video analysis server, and the video analysis server decodes the second monitoring video to obtain a second image frame, wherein the second image frame is an I-frame image, and records a second timestamp of the second image frame;
meanwhile, the video analysis server identifies the position information of all the target objects in the second image frame, and sends the obtained second time stamp and the position information of the target objects corresponding to the second time stamp to the monitoring access server, wherein the target objects can be people or vehicles.
S104, receiving a second time stamp returned by the video analysis server and position information of a target object corresponding to the second time stamp, and sending the second time stamp and the position information of the target object corresponding to the second time stamp to the target video network terminal, so that the target video network terminal displays the first image frame and a labeling frame corresponding to the position information of the target object at the same time in a preset time range when a difference value between the first time stamp and the second time stamp is within the preset time range, and the object in the first image frame corresponds to the position of the labeling frame.
Specifically, the monitoring access server receives a second timestamp returned by the video analysis server and the position information of the target object corresponding to the second timestamp, and sends the second timestamp and the position information of the target object corresponding to the second timestamp to the target video network terminal;
and the target video network terminal draws a marking frame corresponding to the target object on a preset canvas according to the position information of the target object corresponding to the second timestamp within a preset time range at the difference value of the first timestamp and the second timestamp, and then displays the first image frame and the marking frame corresponding to the position information of the target object at the same time, so that the object in the first image frame corresponds to the position of the marking frame.
According to the video data processing method provided by the embodiment of the invention, a first monitoring video sent by monitoring equipment is received; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
The present invention further provides a supplementary description of the video data processing method provided in the above embodiment.
As shown in fig. 2, a flow chart of steps of another embodiment of a method for processing video data according to the present invention is shown, and the method for processing video data includes:
s201, receiving a first monitoring video sent by monitoring equipment;
specifically, the embodiment of the invention is applied to a video analysis system, which comprises a monitoring access server and a video analysis server, wherein the monitoring access server is connected with a plurality of monitoring devices and is connected with front-end equipment. The monitoring access server obtains a first monitoring video from the monitoring device through an RTSP (Real Time Streaming Protocol) Protocol. The monitoring equipment is installed at each intersection of different road sections, and acquires monitoring video data of each intersection, wherein the first monitoring video can be media data in various forms such as H264, H265 and the like, and contains character characteristic information or object characteristic information.
S202, sending the first monitoring video to a target video network terminal so that the target video network terminal can acquire a first image frame in the first monitoring video and a first timestamp of the first image frame;
specifically, the monitoring access server copies the first monitoring video to obtain a second monitoring video, sends the first monitoring video to the target video network terminal, sends the second monitoring video to the video analysis server, and the target video network terminal decodes the first monitoring video to obtain a first image frame, wherein the first image frame is an I-frame image and records a first timestamp of the first image frame.
S203, sending the second monitoring video to a video analysis server so that the video analysis server determines a second timestamp of a second image frame and position information of a target object in the second image frame according to the second monitoring video, wherein the first monitoring video is the same as the second monitoring video;
specifically, the monitoring access server sends a second monitoring video to the video analysis server, and the video analysis server decodes the second monitoring video to obtain a second image frame, wherein the second image frame is an I-frame image, and records a second timestamp of the second image frame;
the video analysis server identifies the position information of all the target objects in the second image frame, and the identification comprises the following steps:
step A1, the video analysis server decodes the second monitoring video to obtain a second image frame, and stores the second image frame as an image in YUV format;
specifically, the video analysis server decodes the second monitoring video by using H264 to obtain a second image frame, and stores the second image frame in a YUV format.
YUV, a color coding method, can be used in various video processing components. YUV allows for reduced bandwidth of chrominance in view of human perception when encoding photos or videos.
YUV is a kind of compiled true-color space (color space), and the proper terms such as Y' UV, YUV, YCbCr, YPbPr, etc. may be called YUV, overlapping with each other. "Y" represents brightness (Luma) or gray scale value, and "U" and "V" represent Chroma (Chroma or Chroma) and are used to describe the color and saturation of the image for specifying the color of the pixel.
Step A2, acquiring a second time stamp of the second image frame;
step A3, the video analysis server determines the position information of the target object in the second image frame according to the second image frame and the pre-established object recognition model.
The method specifically comprises the following steps:
step A31, the video analysis server inputs the second image frame into a pre-established object recognition model to obtain the image characteristics of the target object in the second image frame;
the pre-established object recognition model can be obtained by calculation through a deep learning algorithm and the like, and is not repeated in the embodiment of the invention, and information such as the name of the target object can be obtained through the object recognition model.
Step A32, determining the similarity of the target object according to the image characteristics of the target object and preset image characteristic information in a pre-stored database, wherein the pre-stored database comprises object names and corresponding preset image characteristic information;
specifically, the names of all objects and the corresponding image feature information are stored in the video analysis server in advance, so that the similarity calculation is facilitated.
The video analysis server searches preset image characteristic information corresponding to the target name in a pre-stored database according to the name of the target object, and similarity calculation is carried out on the obtained image characteristic information of the target object and the preset image characteristic information in the database to obtain the similarity of the target object.
Step A33, if the similarity of the target object is greater than or equal to a preset similarity, the video analysis server obtains position information of the target object in the second image frame, wherein the position information includes coordinate information, and the coordinate information includes preset vertex coordinate information, a preset length and a preset width.
S204, receiving a second time stamp returned by the video analysis server and position information of the target object corresponding to the second time stamp, and sending the second time stamp and the position information of the target object corresponding to the second time stamp to the target video network terminal;
s205, the target video network terminal displays the first image frame and a marking frame corresponding to the position information of the target object at the same time when the difference value of the first time stamp and the second time stamp is within a preset time range, so that the object in the first image frame corresponds to the position of the marking frame;
in particular, the amount of the solvent to be used,
step B1, when the difference value between the first timestamp and the second timestamp is within a preset time range, the target video network terminal acquires a first image frame corresponding to the first timestamp;
step B2, the target video network terminal acquires the position information of the target object corresponding to the second timestamp;
step B3, the target video network terminal draws a marking frame corresponding to the position information on a preset canvas according to the position information of the target object;
specifically, the target internet-of-view terminal draws a labeling frame corresponding to the position information on a preset canvas according to the position information, namely coordinate information, of the target object, where the labeling frame may be a square frame, a round frame, a red label, or a blue label, and a specific labeling form is not specifically limited in the embodiment of the present invention. Preferably, the embodiment of the present invention employs a square frame, and the size of the specific square frame is set according to the actual target object.
In the embodiment of the invention, the transparency of the preset canvas can be set, and as a preferred implementation mode, the preset canvas is set as the transparent canvas, so that the first monitoring video and the marking frame can be clearly displayed.
And step B4, the target video network terminal displays the first image frame and a marking frame corresponding to the position information of the target object at the same time, so that the object in the first image frame corresponds to the position of the marking frame.
Specifically, a player is installed on the target video network terminal, a preset canvas is arranged on the player, and the marking frames corresponding to the position information of the target object in the first image frame and the second image frame are displayed simultaneously according to the first time stamp and the second time stamp.
S206, after the target video network terminal finishes displaying the first image frame corresponding to the first timestamp, deleting the marking frame which is drawn on the preset canvas and corresponds to the second timestamp.
Illustratively, if the target video network terminal displays the marking frames corresponding to the position information of the target object in the first image frame and the second image frame at the same time in the 1 st second, and after the 1 st second display is completed, the image frame in the 2 nd second needs to be played, the marking frame of the target object in the 1 st second on the preset canvas needs to be deleted, and then the marking frame corresponding to the position information of the target object in the 2 nd second is drawn on the preset canvas and displayed together with the second image frame.
It should be noted that, for simplicity of description, the method embodiments are described as a series of acts or combination of acts, but those skilled in the art will recognize that the present invention is not limited by the illustrated order of acts, as some steps may occur in other orders or concurrently in accordance with the embodiments of the present invention. Further, those skilled in the art will appreciate that the embodiments described in the specification are presently preferred and that no particular act is required to implement the invention.
According to the video data processing method provided by the embodiment of the invention, a first monitoring video sent by monitoring equipment is received; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
Another embodiment of the present invention provides a video data processing apparatus, configured to execute the video data processing method provided in the foregoing embodiment.
Referring to fig. 3, a block diagram of an embodiment of a video data processing apparatus according to the present invention is shown, where the apparatus may be applied in a video network, and specifically may include the following modules: a receiving module 301, a first sending module 302, a second sending module 303 and a display module 304, wherein:
the receiving module 301 is configured to receive a first monitoring video sent by a monitoring device;
the first sending module 302 is configured to send the first monitoring video to a target video network terminal, so that the target video network terminal obtains a first image frame in the first monitoring video and a first timestamp of the first image frame;
the second sending module 303 is configured to send the second surveillance video to the video analysis server, so that the video analysis server determines, according to the second surveillance video, a second timestamp of a second image frame and position information of a target object in the second image frame, where the first surveillance video is the same as the second surveillance video;
the display module 304 is configured to receive a second timestamp returned by the video analysis server and position information of a target object corresponding to the second timestamp, and send the second timestamp and the position information of the target object corresponding to the second timestamp to the target video network terminal, so that the target video network terminal displays the first image frame and a labeling frame corresponding to the position information of the target object simultaneously within a preset time range when a difference between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the labeling frame
The processing device of the video data provided by the embodiment of the invention receives the first monitoring video sent by the monitoring equipment; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
The present invention further provides a supplementary description of the video data processing apparatus provided in the above embodiments.
Optionally, the target video networking terminal includes a decoding module, and the decoding module is configured to:
and decoding the first monitoring video to obtain the first image frame, and acquiring a first time stamp of the first image frame.
Optionally, the target video networking terminal includes a processing module, and the processing module is configured to:
acquiring a first image frame corresponding to the first timestamp when the difference value between the first timestamp and the second timestamp is within a preset time range;
acquiring position information of the target object corresponding to the second timestamp;
according to the position information of the target object, drawing a marking frame corresponding to the position information on a preset canvas;
and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object, so that the object in the first image frame corresponds to the position of the marking frame.
Optionally, the target video networking terminal further includes a deletion module, and the deletion module is configured to:
after the target video network terminal finishes displaying the first image frame corresponding to the first timestamp, deleting the marking frame corresponding to the second timestamp drawn on the preset canvas
Optionally, the video analysis server comprises an identification module, and the identification module is configured to:
decoding the second monitoring video to obtain a second image frame, and storing the second image frame as an image in a YUV format;
acquiring a second timestamp of the second image frame;
and determining the position information of the target object in the second image frame according to the second image frame and a pre-established object recognition model.
Optionally, the identification module is specifically configured to:
the video analysis server inputs the second image frame into a pre-established object recognition model to obtain the image characteristics of a target object in the second image frame;
determining the similarity of the target object according to the image characteristics of the target object and preset image characteristic information in a pre-stored database, wherein the pre-stored database comprises object names and corresponding preset image characteristic information;
if the similarity of the target object is greater than or equal to a preset similarity, the video analysis server acquires position information of the target object in the second image frame, wherein the position information comprises coordinate information, and the coordinate information comprises preset vertex coordinate information, a preset length and a preset width.
It should be noted that the respective implementable modes in the present embodiment may be implemented individually, or may be implemented in combination in any combination without conflict, and the present application is not limited thereto.
For the device embodiment, since it is basically similar to the method embodiment, the description is simple, and for the relevant points, refer to the partial description of the method embodiment.
The processing device of the video data provided by the embodiment of the invention receives the first monitoring video sent by the monitoring equipment; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
Still another embodiment of the present invention provides a terminal device, configured to execute the video data processing method provided in the foregoing embodiment.
Fig. 4 is a schematic structural diagram of a terminal device of the present invention, and as shown in fig. 4, the terminal device includes: at least one processor 401 and memory 402;
the memory stores a computer program; the at least one processor executes the computer program stored in the memory to implement the video data processing method provided by the above-mentioned embodiments.
The terminal device provided by this embodiment receives a first monitoring video sent by a monitoring device; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
Yet another embodiment of the present application provides a computer-readable storage medium, in which a computer program is stored, and when the computer program is executed, the computer program implements the video data processing method provided in any of the above embodiments.
According to the computer-readable storage medium of the embodiment, a first monitoring video sent by a monitoring device is received; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
The embodiments in the present specification are described in a progressive manner, each embodiment focuses on differences from other embodiments, and the same and similar parts among the embodiments are referred to each other.
As will be appreciated by one skilled in the art, embodiments of the present invention may be provided as a method, apparatus, or computer program product. Accordingly, embodiments of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, embodiments of the present invention may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
Embodiments of the present invention are described with reference to flowchart illustrations and/or block diagrams of methods, electronic devices (systems), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing electronic device to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing electronic device, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing electronic devices to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing electronic device to cause a series of operational steps to be performed on the computer or other programmable electronic device to produce a computer implemented process such that the instructions which execute on the computer or other programmable electronic device provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
While preferred embodiments of the present invention have been described, additional variations and modifications of these embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. Therefore, it is intended that the appended claims be interpreted as including preferred embodiments and all such alterations and modifications as fall within the scope of the embodiments of the invention.
Finally, it should also be noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or electronic device that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or electronic device. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or electronic device that comprises the element.
The present invention provides a method and a device for processing video data, which are described in detail above, and the principles and embodiments of the present invention are explained herein by using specific examples, and the descriptions of the above examples are only used to help understand the method and the core ideas of the present invention; meanwhile, for a person skilled in the art, according to the idea of the present invention, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present invention.

Claims (10)

1. A method for processing video data, the method comprising:
receiving a first monitoring video sent by monitoring equipment;
sending the first monitoring video to a target video network terminal so that the target video network terminal can acquire a first image frame in the first monitoring video and a first timestamp of the first image frame;
sending the second monitoring video to a video analysis server so that the video analysis server determines a second timestamp of a second image frame and position information of a target object in the second image frame according to the second monitoring video, wherein the first monitoring video is the same as the second monitoring video;
receiving a second timestamp returned by the video analysis server and position information of a target object corresponding to the second timestamp, and sending the second timestamp and the position information of the target object corresponding to the second timestamp to the target video network terminal, so that the target video network terminal displays the first image frame and a labeling frame corresponding to the position information of the target object at the same time in a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, and the object in the first image frame corresponds to the position of the labeling frame.
2. The method of claim 1, wherein the target video network terminal acquiring a first image frame in a first surveillance video and a first timestamp of the first image frame comprises:
the target video network terminal decodes the first monitoring video to obtain the first image frame and obtains a first time stamp of the first image frame.
3. The method according to claim 1, wherein the target video network terminal simultaneously displays the first image frame and a labeling frame corresponding to the position information of the target object when the difference value between the first timestamp and the second timestamp is less than a preset time range, so that the object in the first image frame corresponds to the position of the labeling frame, and the method comprises the following steps:
when the difference value of the first timestamp and the second timestamp is within a preset time range, the target video network terminal acquires a first image frame corresponding to the first timestamp;
the target video network terminal acquires the position information of the target object corresponding to the second timestamp;
the target video network terminal draws a marking frame corresponding to the position information on a preset canvas according to the position information of the target object;
and the target video network terminal simultaneously displays the first image frame and a marking frame corresponding to the position information of the target object, so that the object in the first image frame corresponds to the position of the marking frame.
4. The method of claim 3, further comprising:
and after the target video network terminal finishes displaying the first image frame corresponding to the first timestamp, deleting the marking frame which is drawn on the preset canvas and corresponds to the second timestamp.
5. The method of claim 1, wherein the video analysis server determines a second timestamp for a second image frame and location information for a target object in the second image frame from a second surveillance video, comprising:
the video analysis server decodes the second monitoring video to obtain a second image frame, and the second image frame is stored as an image in a YUV format;
acquiring a second timestamp of the second image frame;
and the video analysis server determines the position information of the target object in the second image frame according to the second image frame and a pre-established object recognition model.
6. The method of claim 5, wherein the video analysis server determines position information of a target object in the second image frame from the second image frame and a pre-established object recognition model, comprising:
the video analysis server inputs the second image frame into a pre-established object recognition model to obtain the image characteristics of a target object in the second image frame;
determining the similarity of the target object according to the image characteristics of the target object and preset image characteristic information in a pre-stored database, wherein the pre-stored database comprises object names and corresponding preset image characteristic information;
if the similarity of the target object is greater than or equal to a preset similarity, the video analysis server acquires position information of the target object in the second image frame, wherein the position information comprises coordinate information, and the coordinate information comprises preset vertex coordinate information, a preset length and a preset width.
7. An apparatus for processing video data, the apparatus comprising:
the receiving module is used for receiving a first monitoring video sent by monitoring equipment;
the first sending module is used for sending the first monitoring video to a target video network terminal so that the target video network terminal can acquire a first image frame in the first monitoring video and a first timestamp of the first image frame;
the second sending module is used for sending the second monitoring video to the video analysis server so that the video analysis server can determine a second timestamp of a second image frame and position information of a target object in the second image frame according to the second monitoring video, wherein the first monitoring video is the same as the second monitoring video;
the display module is used for receiving a second timestamp returned by the video analysis server and position information of a target object corresponding to the second timestamp, and sending the second timestamp and the position information of the target object corresponding to the second timestamp to the target video network terminal, so that the target video network terminal displays the first image frame and a labeling frame corresponding to the position information of the target object at the same time within a preset time range according to the difference value between the first timestamp and the second timestamp, and the object in the first image frame corresponds to the position of the labeling frame.
8. The apparatus of claim 7, wherein the target video networking terminal comprises a processing module configured to:
acquiring a first image frame corresponding to the first timestamp when the difference value between the first timestamp and the second timestamp is within a preset time range;
acquiring position information of the target object corresponding to the second timestamp;
according to the position information of the target object, drawing a marking frame corresponding to the position information on a preset canvas;
and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object, so that the object in the first image frame corresponds to the position of the marking frame.
9. A terminal device, comprising: at least one processor and memory;
the memory stores a computer program; the at least one processor executes the computer program stored by the memory to implement the method of processing video data of any of claims 1-6.
10. A computer-readable storage medium, characterized in that a computer program is stored therein, which when executed implements the method of processing video data of any of claims 1-6.
CN202110051307.4A 2021-01-14 2021-01-14 Video data processing method and device, terminal equipment and storage medium Pending CN112866786A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110051307.4A CN112866786A (en) 2021-01-14 2021-01-14 Video data processing method and device, terminal equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110051307.4A CN112866786A (en) 2021-01-14 2021-01-14 Video data processing method and device, terminal equipment and storage medium

Publications (1)

Publication Number Publication Date
CN112866786A true CN112866786A (en) 2021-05-28

Family

ID=76006202

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110051307.4A Pending CN112866786A (en) 2021-01-14 2021-01-14 Video data processing method and device, terminal equipment and storage medium

Country Status (1)

Country Link
CN (1) CN112866786A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113453071A (en) * 2021-06-28 2021-09-28 杭州海康威视数字技术股份有限公司 Data superposition method and device and electronic equipment
CN115396741A (en) * 2022-07-29 2022-11-25 北京势也网络技术有限公司 Panoramic video playing method and device, electronic equipment and readable storage medium
CN116761019A (en) * 2023-08-24 2023-09-15 瀚博半导体(上海)有限公司 Video processing method, system, computer device and computer readable storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102752574A (en) * 2011-04-18 2012-10-24 中兴通讯股份有限公司 Video monitoring system and method
CN107332830A (en) * 2017-06-19 2017-11-07 腾讯科技(深圳)有限公司 Video code conversion, video broadcasting method and device, computer equipment, storage medium
CN108040221A (en) * 2017-11-30 2018-05-15 江西洪都航空工业集团有限责任公司 A kind of intelligent video analysis and monitoring system
CN111246237A (en) * 2020-01-22 2020-06-05 视联动力信息技术股份有限公司 Panoramic video live broadcast method and device
CN111953932A (en) * 2020-06-30 2020-11-17 视联动力信息技术股份有限公司 Data processing method and device, terminal equipment and storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102752574A (en) * 2011-04-18 2012-10-24 中兴通讯股份有限公司 Video monitoring system and method
US20140043480A1 (en) * 2011-04-18 2014-02-13 Zte Corporation Video monitoring system and method
CN107332830A (en) * 2017-06-19 2017-11-07 腾讯科技(深圳)有限公司 Video code conversion, video broadcasting method and device, computer equipment, storage medium
CN108040221A (en) * 2017-11-30 2018-05-15 江西洪都航空工业集团有限责任公司 A kind of intelligent video analysis and monitoring system
CN111246237A (en) * 2020-01-22 2020-06-05 视联动力信息技术股份有限公司 Panoramic video live broadcast method and device
CN111953932A (en) * 2020-06-30 2020-11-17 视联动力信息技术股份有限公司 Data processing method and device, terminal equipment and storage medium

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113453071A (en) * 2021-06-28 2021-09-28 杭州海康威视数字技术股份有限公司 Data superposition method and device and electronic equipment
CN115396741A (en) * 2022-07-29 2022-11-25 北京势也网络技术有限公司 Panoramic video playing method and device, electronic equipment and readable storage medium
CN116761019A (en) * 2023-08-24 2023-09-15 瀚博半导体(上海)有限公司 Video processing method, system, computer device and computer readable storage medium

Similar Documents

Publication Publication Date Title
CN112866786A (en) Video data processing method and device, terminal equipment and storage medium
US11070833B2 (en) Method and system for encoding video with overlay
JP6510404B2 (en) Transmission apparatus, transmission method, reception apparatus and reception method
CN108063976B (en) Video processing method and device
CN108271024B (en) Image coding and decoding method and device
WO2016038775A1 (en) Image processing apparatus and image processing method
EP1871098A2 (en) Processing of removable media that stores full frame video & sub-frame metadata
EP1871101A2 (en) Adaptive video processing circuitry & player using sub-frame metadata
CN104754366A (en) Audio and video file live broadcasting method, device and system
EP1998570A2 (en) Method and apparatus for replacing a portion of a first digital signal with a second digital signal
CN103108160B (en) Monitor video data capture method, server and terminal
CN111147860B (en) Video data decoding method and device
CN112532923A (en) Video stream processing method and device, terminal equipment and storage medium
RU2718118C2 (en) Information processing device and information processing method
CN111901630A (en) Data transmission method, device, terminal equipment and storage medium
JP2019512985A (en) Method and apparatus for encoding video into a bitstream carrying post-processing parameters for each region into SEI nesting messages
CN112584093A (en) Video data processing method and device, terminal equipment and storage medium
US20190068987A1 (en) Systems and Methods for Embedding Metadata into Video Contents
CN111105521A (en) Data reading method and device
CN111510717B (en) Image splicing method and device
CN111968031B (en) Image stitching method and device, storage medium and electronic device
CN102246532B (en) Method and apparatus for avoiding quality deterioration of transmitted media content
CN109429070A (en) A kind of mobile terminal video coding method, device and mobile terminal
CN114827620A (en) Image processing method, apparatus, device and medium
CN111343475B (en) Data processing method and device, live broadcast server and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20210528

RJ01 Rejection of invention patent application after publication