CN112866786A - Video data processing method and device, terminal equipment and storage medium - Google Patents
Video data processing method and device, terminal equipment and storage medium Download PDFInfo
- Publication number
- CN112866786A CN112866786A CN202110051307.4A CN202110051307A CN112866786A CN 112866786 A CN112866786 A CN 112866786A CN 202110051307 A CN202110051307 A CN 202110051307A CN 112866786 A CN112866786 A CN 112866786A
- Authority
- CN
- China
- Prior art keywords
- video
- image frame
- timestamp
- target
- position information
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000003860 storage Methods 0.000 title claims abstract description 13
- 238000003672 processing method Methods 0.000 title description 12
- 238000012544 monitoring process Methods 0.000 claims abstract description 150
- 238000004458 analytical method Methods 0.000 claims abstract description 87
- 238000012545 processing Methods 0.000 claims abstract description 43
- 238000000034 method Methods 0.000 claims abstract description 40
- 230000006855 networking Effects 0.000 claims abstract description 20
- 238000004590 computer program Methods 0.000 claims description 20
- 238000002372 labelling Methods 0.000 claims description 19
- 238000010586 diagram Methods 0.000 description 12
- 238000012806 monitoring device Methods 0.000 description 7
- 238000012217 deletion Methods 0.000 description 4
- 230000037430 deletion Effects 0.000 description 4
- 230000006870 function Effects 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 230000009471 action Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 241000023320 Luma <angiosperm> Species 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 238000004422 calculation algorithm Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 238000009826 distribution Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- OSWPMRLSEDHDFF-UHFFFAOYSA-N methyl salicylate Chemical compound COC(=O)C1=CC=CC=C1O OSWPMRLSEDHDFF-UHFFFAOYSA-N 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 239000002904 solvent Substances 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/7837—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/7867—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title and artist information, manually generated time, location and usage information, user ratings
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Library & Information Science (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
The embodiment of the invention provides a method and a device for processing video data, terminal equipment and a storage medium, wherein the method comprises the following steps: receiving a first monitoring video sent by monitoring equipment; sending the first monitoring video to a target video networking terminal; the second monitoring video is sent to a video analysis server, the video analysis server determines a second time stamp of a second image frame and position information of a target object according to the second monitoring video, the first time stamp returned by the video analysis server and the position information of the target object corresponding to the first time stamp are sent to a target video network terminal, and the target video network terminal simultaneously displays the first image frame and a marking frame corresponding to the position information of the target object, so that the object in the first image frame corresponds to the position of the marking frame.
Description
Technical Field
The present invention relates to the field of video networking technologies, and in particular, to a method and an apparatus for processing video data, a terminal device, and a storage medium.
Background
Along with the continuous development of the video networking technology, the field that involves is more and more, and video networking and internet also can carry out data transmission, can show the surveillance video of supervisory equipment on the video networking terminal, just need insert the supervisory equipment to the video networking through the control access server, acquire the surveillance video from the supervisory equipment again to show on sending this surveillance video to the video networking terminal.
When playing the surveillance video on the video networking terminal, if the surveillance video is formed by the multilayer data combination, then the video networking terminal is when playing the surveillance video, then needs a plurality of decoders to decode multilayer data, then adopts a plurality of encoders to encode the data after the multilayer is decoded again and plays, every increases one deck data, then needs to increase a set of encoder and decoder, like this, will increase the processing pressure at video networking terminal, and video broadcast efficiency is lower.
Disclosure of Invention
In view of the above problems, embodiments of the present invention are proposed to provide a video data processing method, apparatus, terminal device and storage medium that overcome or at least partially solve the above problems.
In a first aspect, an embodiment of the present invention provides a method for processing video data, where the method includes:
receiving a first monitoring video sent by monitoring equipment;
sending the first monitoring video to a target video network terminal so that the target video network terminal can acquire a first image frame in the first monitoring video and a first timestamp of the first image frame;
sending the second monitoring video to a video analysis server so that the video analysis server determines a second timestamp of a second image frame and position information of a target object in the second image frame according to the second monitoring video, wherein the first monitoring video is the same as the second monitoring video;
receiving a second timestamp returned by the video analysis server and position information of a target object corresponding to the second timestamp, and sending the second timestamp and the position information of the target object corresponding to the second timestamp to the target video network terminal, so that the target video network terminal displays the first image frame and a labeling frame corresponding to the position information of the target object at the same time in a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, and the object in the first image frame corresponds to the position of the labeling frame.
Optionally, the acquiring, by the target video network terminal, a first image frame in a first monitoring video and a first timestamp of the first image frame includes:
the target video network terminal decodes the first monitoring video to obtain the first image frame and obtains a first time stamp of the first image frame.
Optionally, the displaying, by the target internet of view terminal, the first image frame and the labeling frame corresponding to the position information of the target object simultaneously when the difference between the first timestamp and the second timestamp is smaller than a preset time range, so that the object in the first image frame corresponds to the position of the labeling frame, includes:
when the difference value of the first timestamp and the second timestamp is within a preset time range, the target video network terminal acquires a first image frame corresponding to the first timestamp;
the target video network terminal acquires the position information of the target object corresponding to the second timestamp;
the target video network terminal draws a marking frame corresponding to the position information on a preset canvas according to the position information of the target object;
and the target video network terminal simultaneously displays the first image frame and a marking frame corresponding to the position information of the target object, so that the object in the first image frame corresponds to the position of the marking frame.
Optionally, the method further comprises:
and after the target video network terminal finishes displaying the first image frame corresponding to the first timestamp, deleting the marking frame which is drawn on the preset canvas and corresponds to the second timestamp.
Optionally, the video analysis server determines, according to the second surveillance video, a second timestamp of the second image frame and location information of the target object in the second image frame, including:
the video analysis server decodes the second monitoring video to obtain a second image frame, and the second image frame is stored as an image in a YUV format;
acquiring a second timestamp of the second image frame;
and the video analysis server determines the position information of the target object in the second image frame according to the second image frame and a pre-established object recognition model.
Optionally, the determining, by the video analysis server, the position information of the target object in the second image frame according to the second image frame and a pre-established object recognition model includes:
the video analysis server inputs the second image frame into a pre-established object recognition model to obtain the image characteristics of a target object in the second image frame;
determining the similarity of the target object according to the image characteristics of the target object and preset image characteristic information in a pre-stored database, wherein the pre-stored database comprises object names and corresponding preset image characteristic information;
if the similarity of the target object is greater than or equal to a preset similarity, the video analysis server acquires position information of the target object in the second image frame, wherein the position information comprises coordinate information, and the coordinate information comprises preset vertex coordinate information, a preset length and a preset width.
In a second aspect, an embodiment of the present invention provides an apparatus for processing video data, where the apparatus includes:
the receiving module is used for receiving a first monitoring video sent by monitoring equipment;
the first sending module is used for sending the first monitoring video to a target video network terminal so that the target video network terminal can acquire a first image frame in the first monitoring video and a first timestamp of the first image frame;
the second sending module is used for sending the second monitoring video to the video analysis server so that the video analysis server can determine a second timestamp of a second image frame and position information of a target object in the second image frame according to the second monitoring video, wherein the first monitoring video is the same as the second monitoring video;
the display module is used for receiving a second timestamp returned by the video analysis server and position information of a target object corresponding to the second timestamp, and sending the second timestamp and the position information of the target object corresponding to the second timestamp to the target video network terminal, so that the target video network terminal displays the first image frame and a labeling frame corresponding to the position information of the target object at the same time within a preset time range according to the difference value between the first timestamp and the second timestamp, and the object in the first image frame corresponds to the position of the labeling frame.
Optionally, the target video networking terminal includes a decoding module, and the decoding module is configured to:
and decoding the first monitoring video to obtain the first image frame, and acquiring a first time stamp of the first image frame.
Optionally, the target video networking terminal includes a processing module, and the processing module is configured to:
acquiring a first image frame corresponding to the first timestamp when the difference value between the first timestamp and the second timestamp is within a preset time range;
acquiring position information of the target object corresponding to the second timestamp;
according to the position information of the target object, drawing a marking frame corresponding to the position information on a preset canvas;
and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object, so that the object in the first image frame corresponds to the position of the marking frame.
Optionally, the target video networking terminal further includes a deletion module, and the deletion module is configured to:
after the target video network terminal finishes displaying the first image frame corresponding to the first timestamp, deleting the marking frame corresponding to the second timestamp drawn on the preset canvas
Optionally, the video analysis server comprises an identification module, and the identification module is configured to:
decoding the second monitoring video to obtain a second image frame, and storing the second image frame as an image in a YUV format;
acquiring a second timestamp of the second image frame;
and determining the position information of the target object in the second image frame according to the second image frame and a pre-established object recognition model.
Optionally, the identification module is specifically configured to:
the video analysis server inputs the second image frame into a pre-established object recognition model to obtain the image characteristics of a target object in the second image frame;
determining the similarity of the target object according to the image characteristics of the target object and preset image characteristic information in a pre-stored database, wherein the pre-stored database comprises object names and corresponding preset image characteristic information;
if the similarity of the target object is greater than or equal to a preset similarity, the video analysis server acquires position information of the target object in the second image frame, wherein the position information comprises coordinate information, and the coordinate information comprises preset vertex coordinate information, a preset length and a preset width.
In a third aspect, an embodiment of the present invention provides a terminal device, including: at least one processor and memory;
the memory stores a computer program; the at least one processor executes the computer program stored by the memory to implement the method for processing video data provided by the first aspect.
In a fourth aspect, an embodiment of the present invention provides a computer-readable storage medium, where a computer program is stored, and when the computer program is executed, the computer program implements the video data processing method provided in the first aspect.
The embodiment of the invention has the following advantages:
according to the video data processing method, the video data processing device, the terminal equipment and the storage medium, the first monitoring video sent by the monitoring equipment is received; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
Drawings
FIG. 1 is a flow chart of the steps of an embodiment of a method for processing video data according to the present invention;
FIG. 2 is a flow chart of steps in another embodiment of a method of processing video data in accordance with the present invention;
fig. 3 is a block diagram of an embodiment of a video data processing apparatus according to the present invention;
fig. 4 is a schematic structural diagram of a terminal device of the present invention.
Detailed Description
In order to make the aforementioned objects, features and advantages of the present invention comprehensible, embodiments accompanied with figures are described in further detail below.
The video networking is an important milestone for network development, is a real-time network, can realize high-definition video real-time transmission, and pushes a plurality of internet applications to high-definition video, and high-definition faces each other.
The video networking adopts a real-time high-definition video exchange technology, can integrate required services such as dozens of services of video, voice, pictures, characters, communication, data and the like on a system platform on a network platform, such as high-definition video conference, video monitoring, intelligent monitoring analysis, emergency command, digital broadcast television, delayed television, network teaching, live broadcast, VOD on demand, television mail, Personal Video Recorder (PVR), intranet (self-office) channels, intelligent video broadcast control, information distribution and the like, and realizes high-definition quality video broadcast through a television or a computer.
Based on the characteristics of the video network, the invention provides one of the core concepts of the embodiment, and the first monitoring video sent by the monitoring equipment is received according to the protocol of the video network; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
An embodiment of the present invention provides a method for processing video data, which is used for displaying layered video data. The execution subject of this embodiment is a processing apparatus of video data, and is disposed in the monitoring access server.
Referring to fig. 1, a flowchart illustrating steps of an embodiment of a method for processing video data according to the present invention is shown, where the method specifically includes the following steps:
s101, receiving a first monitoring video sent by monitoring equipment;
specifically, the embodiment of the invention is applied to a video analysis system, which comprises a monitoring access server and a video analysis server, wherein the monitoring access server is connected with a plurality of monitoring devices and is connected with a video networking terminal. The monitoring access server obtains a first monitoring video from the monitoring device through an RTSP (Real Time Streaming Protocol) Protocol. The monitoring equipment is installed at each intersection of different road sections and used for collecting monitoring video data of each intersection.
S102, sending the first monitoring video to a target video network terminal so that the target video network terminal can acquire a first image frame in the first monitoring video and a first timestamp of the first image frame;
specifically, the monitoring access server copies the first monitoring video to obtain a second monitoring video, sends the first monitoring video to the target video network terminal, sends the second monitoring video to the video analysis server, and the target video network terminal decodes the first monitoring video to obtain a first image frame, wherein the first image frame is an I-frame image and records a first timestamp of the first image frame.
S103, sending the second monitoring video to a video analysis server so that the video analysis server determines a second timestamp of a second image frame and position information of a target object in the second image frame according to the second monitoring video, wherein the first monitoring video is the same as the second monitoring video;
specifically, the monitoring access server sends a second monitoring video to the video analysis server, and the video analysis server decodes the second monitoring video to obtain a second image frame, wherein the second image frame is an I-frame image, and records a second timestamp of the second image frame;
meanwhile, the video analysis server identifies the position information of all the target objects in the second image frame, and sends the obtained second time stamp and the position information of the target objects corresponding to the second time stamp to the monitoring access server, wherein the target objects can be people or vehicles.
S104, receiving a second time stamp returned by the video analysis server and position information of a target object corresponding to the second time stamp, and sending the second time stamp and the position information of the target object corresponding to the second time stamp to the target video network terminal, so that the target video network terminal displays the first image frame and a labeling frame corresponding to the position information of the target object at the same time in a preset time range when a difference value between the first time stamp and the second time stamp is within the preset time range, and the object in the first image frame corresponds to the position of the labeling frame.
Specifically, the monitoring access server receives a second timestamp returned by the video analysis server and the position information of the target object corresponding to the second timestamp, and sends the second timestamp and the position information of the target object corresponding to the second timestamp to the target video network terminal;
and the target video network terminal draws a marking frame corresponding to the target object on a preset canvas according to the position information of the target object corresponding to the second timestamp within a preset time range at the difference value of the first timestamp and the second timestamp, and then displays the first image frame and the marking frame corresponding to the position information of the target object at the same time, so that the object in the first image frame corresponds to the position of the marking frame.
According to the video data processing method provided by the embodiment of the invention, a first monitoring video sent by monitoring equipment is received; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
The present invention further provides a supplementary description of the video data processing method provided in the above embodiment.
As shown in fig. 2, a flow chart of steps of another embodiment of a method for processing video data according to the present invention is shown, and the method for processing video data includes:
s201, receiving a first monitoring video sent by monitoring equipment;
specifically, the embodiment of the invention is applied to a video analysis system, which comprises a monitoring access server and a video analysis server, wherein the monitoring access server is connected with a plurality of monitoring devices and is connected with front-end equipment. The monitoring access server obtains a first monitoring video from the monitoring device through an RTSP (Real Time Streaming Protocol) Protocol. The monitoring equipment is installed at each intersection of different road sections, and acquires monitoring video data of each intersection, wherein the first monitoring video can be media data in various forms such as H264, H265 and the like, and contains character characteristic information or object characteristic information.
S202, sending the first monitoring video to a target video network terminal so that the target video network terminal can acquire a first image frame in the first monitoring video and a first timestamp of the first image frame;
specifically, the monitoring access server copies the first monitoring video to obtain a second monitoring video, sends the first monitoring video to the target video network terminal, sends the second monitoring video to the video analysis server, and the target video network terminal decodes the first monitoring video to obtain a first image frame, wherein the first image frame is an I-frame image and records a first timestamp of the first image frame.
S203, sending the second monitoring video to a video analysis server so that the video analysis server determines a second timestamp of a second image frame and position information of a target object in the second image frame according to the second monitoring video, wherein the first monitoring video is the same as the second monitoring video;
specifically, the monitoring access server sends a second monitoring video to the video analysis server, and the video analysis server decodes the second monitoring video to obtain a second image frame, wherein the second image frame is an I-frame image, and records a second timestamp of the second image frame;
the video analysis server identifies the position information of all the target objects in the second image frame, and the identification comprises the following steps:
step A1, the video analysis server decodes the second monitoring video to obtain a second image frame, and stores the second image frame as an image in YUV format;
specifically, the video analysis server decodes the second monitoring video by using H264 to obtain a second image frame, and stores the second image frame in a YUV format.
YUV, a color coding method, can be used in various video processing components. YUV allows for reduced bandwidth of chrominance in view of human perception when encoding photos or videos.
YUV is a kind of compiled true-color space (color space), and the proper terms such as Y' UV, YUV, YCbCr, YPbPr, etc. may be called YUV, overlapping with each other. "Y" represents brightness (Luma) or gray scale value, and "U" and "V" represent Chroma (Chroma or Chroma) and are used to describe the color and saturation of the image for specifying the color of the pixel.
Step A2, acquiring a second time stamp of the second image frame;
step A3, the video analysis server determines the position information of the target object in the second image frame according to the second image frame and the pre-established object recognition model.
The method specifically comprises the following steps:
step A31, the video analysis server inputs the second image frame into a pre-established object recognition model to obtain the image characteristics of the target object in the second image frame;
the pre-established object recognition model can be obtained by calculation through a deep learning algorithm and the like, and is not repeated in the embodiment of the invention, and information such as the name of the target object can be obtained through the object recognition model.
Step A32, determining the similarity of the target object according to the image characteristics of the target object and preset image characteristic information in a pre-stored database, wherein the pre-stored database comprises object names and corresponding preset image characteristic information;
specifically, the names of all objects and the corresponding image feature information are stored in the video analysis server in advance, so that the similarity calculation is facilitated.
The video analysis server searches preset image characteristic information corresponding to the target name in a pre-stored database according to the name of the target object, and similarity calculation is carried out on the obtained image characteristic information of the target object and the preset image characteristic information in the database to obtain the similarity of the target object.
Step A33, if the similarity of the target object is greater than or equal to a preset similarity, the video analysis server obtains position information of the target object in the second image frame, wherein the position information includes coordinate information, and the coordinate information includes preset vertex coordinate information, a preset length and a preset width.
S204, receiving a second time stamp returned by the video analysis server and position information of the target object corresponding to the second time stamp, and sending the second time stamp and the position information of the target object corresponding to the second time stamp to the target video network terminal;
s205, the target video network terminal displays the first image frame and a marking frame corresponding to the position information of the target object at the same time when the difference value of the first time stamp and the second time stamp is within a preset time range, so that the object in the first image frame corresponds to the position of the marking frame;
in particular, the amount of the solvent to be used,
step B1, when the difference value between the first timestamp and the second timestamp is within a preset time range, the target video network terminal acquires a first image frame corresponding to the first timestamp;
step B2, the target video network terminal acquires the position information of the target object corresponding to the second timestamp;
step B3, the target video network terminal draws a marking frame corresponding to the position information on a preset canvas according to the position information of the target object;
specifically, the target internet-of-view terminal draws a labeling frame corresponding to the position information on a preset canvas according to the position information, namely coordinate information, of the target object, where the labeling frame may be a square frame, a round frame, a red label, or a blue label, and a specific labeling form is not specifically limited in the embodiment of the present invention. Preferably, the embodiment of the present invention employs a square frame, and the size of the specific square frame is set according to the actual target object.
In the embodiment of the invention, the transparency of the preset canvas can be set, and as a preferred implementation mode, the preset canvas is set as the transparent canvas, so that the first monitoring video and the marking frame can be clearly displayed.
And step B4, the target video network terminal displays the first image frame and a marking frame corresponding to the position information of the target object at the same time, so that the object in the first image frame corresponds to the position of the marking frame.
Specifically, a player is installed on the target video network terminal, a preset canvas is arranged on the player, and the marking frames corresponding to the position information of the target object in the first image frame and the second image frame are displayed simultaneously according to the first time stamp and the second time stamp.
S206, after the target video network terminal finishes displaying the first image frame corresponding to the first timestamp, deleting the marking frame which is drawn on the preset canvas and corresponds to the second timestamp.
Illustratively, if the target video network terminal displays the marking frames corresponding to the position information of the target object in the first image frame and the second image frame at the same time in the 1 st second, and after the 1 st second display is completed, the image frame in the 2 nd second needs to be played, the marking frame of the target object in the 1 st second on the preset canvas needs to be deleted, and then the marking frame corresponding to the position information of the target object in the 2 nd second is drawn on the preset canvas and displayed together with the second image frame.
It should be noted that, for simplicity of description, the method embodiments are described as a series of acts or combination of acts, but those skilled in the art will recognize that the present invention is not limited by the illustrated order of acts, as some steps may occur in other orders or concurrently in accordance with the embodiments of the present invention. Further, those skilled in the art will appreciate that the embodiments described in the specification are presently preferred and that no particular act is required to implement the invention.
According to the video data processing method provided by the embodiment of the invention, a first monitoring video sent by monitoring equipment is received; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
Another embodiment of the present invention provides a video data processing apparatus, configured to execute the video data processing method provided in the foregoing embodiment.
Referring to fig. 3, a block diagram of an embodiment of a video data processing apparatus according to the present invention is shown, where the apparatus may be applied in a video network, and specifically may include the following modules: a receiving module 301, a first sending module 302, a second sending module 303 and a display module 304, wherein:
the receiving module 301 is configured to receive a first monitoring video sent by a monitoring device;
the first sending module 302 is configured to send the first monitoring video to a target video network terminal, so that the target video network terminal obtains a first image frame in the first monitoring video and a first timestamp of the first image frame;
the second sending module 303 is configured to send the second surveillance video to the video analysis server, so that the video analysis server determines, according to the second surveillance video, a second timestamp of a second image frame and position information of a target object in the second image frame, where the first surveillance video is the same as the second surveillance video;
the display module 304 is configured to receive a second timestamp returned by the video analysis server and position information of a target object corresponding to the second timestamp, and send the second timestamp and the position information of the target object corresponding to the second timestamp to the target video network terminal, so that the target video network terminal displays the first image frame and a labeling frame corresponding to the position information of the target object simultaneously within a preset time range when a difference between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the labeling frame
The processing device of the video data provided by the embodiment of the invention receives the first monitoring video sent by the monitoring equipment; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
The present invention further provides a supplementary description of the video data processing apparatus provided in the above embodiments.
Optionally, the target video networking terminal includes a decoding module, and the decoding module is configured to:
and decoding the first monitoring video to obtain the first image frame, and acquiring a first time stamp of the first image frame.
Optionally, the target video networking terminal includes a processing module, and the processing module is configured to:
acquiring a first image frame corresponding to the first timestamp when the difference value between the first timestamp and the second timestamp is within a preset time range;
acquiring position information of the target object corresponding to the second timestamp;
according to the position information of the target object, drawing a marking frame corresponding to the position information on a preset canvas;
and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object, so that the object in the first image frame corresponds to the position of the marking frame.
Optionally, the target video networking terminal further includes a deletion module, and the deletion module is configured to:
after the target video network terminal finishes displaying the first image frame corresponding to the first timestamp, deleting the marking frame corresponding to the second timestamp drawn on the preset canvas
Optionally, the video analysis server comprises an identification module, and the identification module is configured to:
decoding the second monitoring video to obtain a second image frame, and storing the second image frame as an image in a YUV format;
acquiring a second timestamp of the second image frame;
and determining the position information of the target object in the second image frame according to the second image frame and a pre-established object recognition model.
Optionally, the identification module is specifically configured to:
the video analysis server inputs the second image frame into a pre-established object recognition model to obtain the image characteristics of a target object in the second image frame;
determining the similarity of the target object according to the image characteristics of the target object and preset image characteristic information in a pre-stored database, wherein the pre-stored database comprises object names and corresponding preset image characteristic information;
if the similarity of the target object is greater than or equal to a preset similarity, the video analysis server acquires position information of the target object in the second image frame, wherein the position information comprises coordinate information, and the coordinate information comprises preset vertex coordinate information, a preset length and a preset width.
It should be noted that the respective implementable modes in the present embodiment may be implemented individually, or may be implemented in combination in any combination without conflict, and the present application is not limited thereto.
For the device embodiment, since it is basically similar to the method embodiment, the description is simple, and for the relevant points, refer to the partial description of the method embodiment.
The processing device of the video data provided by the embodiment of the invention receives the first monitoring video sent by the monitoring equipment; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
Still another embodiment of the present invention provides a terminal device, configured to execute the video data processing method provided in the foregoing embodiment.
Fig. 4 is a schematic structural diagram of a terminal device of the present invention, and as shown in fig. 4, the terminal device includes: at least one processor 401 and memory 402;
the memory stores a computer program; the at least one processor executes the computer program stored in the memory to implement the video data processing method provided by the above-mentioned embodiments.
The terminal device provided by this embodiment receives a first monitoring video sent by a monitoring device; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
Yet another embodiment of the present application provides a computer-readable storage medium, in which a computer program is stored, and when the computer program is executed, the computer program implements the video data processing method provided in any of the above embodiments.
According to the computer-readable storage medium of the embodiment, a first monitoring video sent by a monitoring device is received; the first monitoring video is sent to a target video network terminal, and the target video network terminal acquires a first image frame in the first monitoring video and a first timestamp of the first image frame; sending the second monitoring video to a video analysis server, determining a second timestamp of a second image frame and position information of a target object in the second image frame by the video analysis server according to the second monitoring video, receiving a first timestamp returned by the video analysis server and the position information of the target object corresponding to the first timestamp, sending the first timestamp and the position information of the target object corresponding to the first timestamp to a target video network terminal, and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object by the target video network terminal within a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, so that the object in the first image frame corresponds to the position of the marking frame, and by adopting the embodiment of the invention, the settings of an encoder and a decoder are reduced, and the processing pressure of the video network terminal is reduced, and the video playing efficiency is improved.
The embodiments in the present specification are described in a progressive manner, each embodiment focuses on differences from other embodiments, and the same and similar parts among the embodiments are referred to each other.
As will be appreciated by one skilled in the art, embodiments of the present invention may be provided as a method, apparatus, or computer program product. Accordingly, embodiments of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, embodiments of the present invention may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
Embodiments of the present invention are described with reference to flowchart illustrations and/or block diagrams of methods, electronic devices (systems), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing electronic device to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing electronic device, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing electronic devices to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing electronic device to cause a series of operational steps to be performed on the computer or other programmable electronic device to produce a computer implemented process such that the instructions which execute on the computer or other programmable electronic device provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
While preferred embodiments of the present invention have been described, additional variations and modifications of these embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. Therefore, it is intended that the appended claims be interpreted as including preferred embodiments and all such alterations and modifications as fall within the scope of the embodiments of the invention.
Finally, it should also be noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or electronic device that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or electronic device. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or electronic device that comprises the element.
The present invention provides a method and a device for processing video data, which are described in detail above, and the principles and embodiments of the present invention are explained herein by using specific examples, and the descriptions of the above examples are only used to help understand the method and the core ideas of the present invention; meanwhile, for a person skilled in the art, according to the idea of the present invention, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present invention.
Claims (10)
1. A method for processing video data, the method comprising:
receiving a first monitoring video sent by monitoring equipment;
sending the first monitoring video to a target video network terminal so that the target video network terminal can acquire a first image frame in the first monitoring video and a first timestamp of the first image frame;
sending the second monitoring video to a video analysis server so that the video analysis server determines a second timestamp of a second image frame and position information of a target object in the second image frame according to the second monitoring video, wherein the first monitoring video is the same as the second monitoring video;
receiving a second timestamp returned by the video analysis server and position information of a target object corresponding to the second timestamp, and sending the second timestamp and the position information of the target object corresponding to the second timestamp to the target video network terminal, so that the target video network terminal displays the first image frame and a labeling frame corresponding to the position information of the target object at the same time in a preset time range when a difference value between the first timestamp and the second timestamp is within the preset time range, and the object in the first image frame corresponds to the position of the labeling frame.
2. The method of claim 1, wherein the target video network terminal acquiring a first image frame in a first surveillance video and a first timestamp of the first image frame comprises:
the target video network terminal decodes the first monitoring video to obtain the first image frame and obtains a first time stamp of the first image frame.
3. The method according to claim 1, wherein the target video network terminal simultaneously displays the first image frame and a labeling frame corresponding to the position information of the target object when the difference value between the first timestamp and the second timestamp is less than a preset time range, so that the object in the first image frame corresponds to the position of the labeling frame, and the method comprises the following steps:
when the difference value of the first timestamp and the second timestamp is within a preset time range, the target video network terminal acquires a first image frame corresponding to the first timestamp;
the target video network terminal acquires the position information of the target object corresponding to the second timestamp;
the target video network terminal draws a marking frame corresponding to the position information on a preset canvas according to the position information of the target object;
and the target video network terminal simultaneously displays the first image frame and a marking frame corresponding to the position information of the target object, so that the object in the first image frame corresponds to the position of the marking frame.
4. The method of claim 3, further comprising:
and after the target video network terminal finishes displaying the first image frame corresponding to the first timestamp, deleting the marking frame which is drawn on the preset canvas and corresponds to the second timestamp.
5. The method of claim 1, wherein the video analysis server determines a second timestamp for a second image frame and location information for a target object in the second image frame from a second surveillance video, comprising:
the video analysis server decodes the second monitoring video to obtain a second image frame, and the second image frame is stored as an image in a YUV format;
acquiring a second timestamp of the second image frame;
and the video analysis server determines the position information of the target object in the second image frame according to the second image frame and a pre-established object recognition model.
6. The method of claim 5, wherein the video analysis server determines position information of a target object in the second image frame from the second image frame and a pre-established object recognition model, comprising:
the video analysis server inputs the second image frame into a pre-established object recognition model to obtain the image characteristics of a target object in the second image frame;
determining the similarity of the target object according to the image characteristics of the target object and preset image characteristic information in a pre-stored database, wherein the pre-stored database comprises object names and corresponding preset image characteristic information;
if the similarity of the target object is greater than or equal to a preset similarity, the video analysis server acquires position information of the target object in the second image frame, wherein the position information comprises coordinate information, and the coordinate information comprises preset vertex coordinate information, a preset length and a preset width.
7. An apparatus for processing video data, the apparatus comprising:
the receiving module is used for receiving a first monitoring video sent by monitoring equipment;
the first sending module is used for sending the first monitoring video to a target video network terminal so that the target video network terminal can acquire a first image frame in the first monitoring video and a first timestamp of the first image frame;
the second sending module is used for sending the second monitoring video to the video analysis server so that the video analysis server can determine a second timestamp of a second image frame and position information of a target object in the second image frame according to the second monitoring video, wherein the first monitoring video is the same as the second monitoring video;
the display module is used for receiving a second timestamp returned by the video analysis server and position information of a target object corresponding to the second timestamp, and sending the second timestamp and the position information of the target object corresponding to the second timestamp to the target video network terminal, so that the target video network terminal displays the first image frame and a labeling frame corresponding to the position information of the target object at the same time within a preset time range according to the difference value between the first timestamp and the second timestamp, and the object in the first image frame corresponds to the position of the labeling frame.
8. The apparatus of claim 7, wherein the target video networking terminal comprises a processing module configured to:
acquiring a first image frame corresponding to the first timestamp when the difference value between the first timestamp and the second timestamp is within a preset time range;
acquiring position information of the target object corresponding to the second timestamp;
according to the position information of the target object, drawing a marking frame corresponding to the position information on a preset canvas;
and simultaneously displaying the first image frame and a marking frame corresponding to the position information of the target object, so that the object in the first image frame corresponds to the position of the marking frame.
9. A terminal device, comprising: at least one processor and memory;
the memory stores a computer program; the at least one processor executes the computer program stored by the memory to implement the method of processing video data of any of claims 1-6.
10. A computer-readable storage medium, characterized in that a computer program is stored therein, which when executed implements the method of processing video data of any of claims 1-6.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110051307.4A CN112866786A (en) | 2021-01-14 | 2021-01-14 | Video data processing method and device, terminal equipment and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110051307.4A CN112866786A (en) | 2021-01-14 | 2021-01-14 | Video data processing method and device, terminal equipment and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112866786A true CN112866786A (en) | 2021-05-28 |
Family
ID=76006202
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110051307.4A Pending CN112866786A (en) | 2021-01-14 | 2021-01-14 | Video data processing method and device, terminal equipment and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112866786A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113453071A (en) * | 2021-06-28 | 2021-09-28 | 杭州海康威视数字技术股份有限公司 | Data superposition method and device and electronic equipment |
CN115396741A (en) * | 2022-07-29 | 2022-11-25 | 北京势也网络技术有限公司 | Panoramic video playing method and device, electronic equipment and readable storage medium |
CN116761019A (en) * | 2023-08-24 | 2023-09-15 | 瀚博半导体(上海)有限公司 | Video processing method, system, computer device and computer readable storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102752574A (en) * | 2011-04-18 | 2012-10-24 | 中兴通讯股份有限公司 | Video monitoring system and method |
CN107332830A (en) * | 2017-06-19 | 2017-11-07 | 腾讯科技(深圳)有限公司 | Video code conversion, video broadcasting method and device, computer equipment, storage medium |
CN108040221A (en) * | 2017-11-30 | 2018-05-15 | 江西洪都航空工业集团有限责任公司 | A kind of intelligent video analysis and monitoring system |
CN111246237A (en) * | 2020-01-22 | 2020-06-05 | 视联动力信息技术股份有限公司 | Panoramic video live broadcast method and device |
CN111953932A (en) * | 2020-06-30 | 2020-11-17 | 视联动力信息技术股份有限公司 | Data processing method and device, terminal equipment and storage medium |
-
2021
- 2021-01-14 CN CN202110051307.4A patent/CN112866786A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102752574A (en) * | 2011-04-18 | 2012-10-24 | 中兴通讯股份有限公司 | Video monitoring system and method |
US20140043480A1 (en) * | 2011-04-18 | 2014-02-13 | Zte Corporation | Video monitoring system and method |
CN107332830A (en) * | 2017-06-19 | 2017-11-07 | 腾讯科技(深圳)有限公司 | Video code conversion, video broadcasting method and device, computer equipment, storage medium |
CN108040221A (en) * | 2017-11-30 | 2018-05-15 | 江西洪都航空工业集团有限责任公司 | A kind of intelligent video analysis and monitoring system |
CN111246237A (en) * | 2020-01-22 | 2020-06-05 | 视联动力信息技术股份有限公司 | Panoramic video live broadcast method and device |
CN111953932A (en) * | 2020-06-30 | 2020-11-17 | 视联动力信息技术股份有限公司 | Data processing method and device, terminal equipment and storage medium |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113453071A (en) * | 2021-06-28 | 2021-09-28 | 杭州海康威视数字技术股份有限公司 | Data superposition method and device and electronic equipment |
CN115396741A (en) * | 2022-07-29 | 2022-11-25 | 北京势也网络技术有限公司 | Panoramic video playing method and device, electronic equipment and readable storage medium |
CN116761019A (en) * | 2023-08-24 | 2023-09-15 | 瀚博半导体(上海)有限公司 | Video processing method, system, computer device and computer readable storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112866786A (en) | Video data processing method and device, terminal equipment and storage medium | |
US11070833B2 (en) | Method and system for encoding video with overlay | |
JP6510404B2 (en) | Transmission apparatus, transmission method, reception apparatus and reception method | |
CN108063976B (en) | Video processing method and device | |
CN108271024B (en) | Image coding and decoding method and device | |
WO2016038775A1 (en) | Image processing apparatus and image processing method | |
EP1871098A2 (en) | Processing of removable media that stores full frame video & sub-frame metadata | |
EP1871101A2 (en) | Adaptive video processing circuitry & player using sub-frame metadata | |
CN104754366A (en) | Audio and video file live broadcasting method, device and system | |
EP1998570A2 (en) | Method and apparatus for replacing a portion of a first digital signal with a second digital signal | |
CN103108160B (en) | Monitor video data capture method, server and terminal | |
CN111147860B (en) | Video data decoding method and device | |
CN112532923A (en) | Video stream processing method and device, terminal equipment and storage medium | |
RU2718118C2 (en) | Information processing device and information processing method | |
CN111901630A (en) | Data transmission method, device, terminal equipment and storage medium | |
JP2019512985A (en) | Method and apparatus for encoding video into a bitstream carrying post-processing parameters for each region into SEI nesting messages | |
CN112584093A (en) | Video data processing method and device, terminal equipment and storage medium | |
US20190068987A1 (en) | Systems and Methods for Embedding Metadata into Video Contents | |
CN111105521A (en) | Data reading method and device | |
CN111510717B (en) | Image splicing method and device | |
CN111968031B (en) | Image stitching method and device, storage medium and electronic device | |
CN102246532B (en) | Method and apparatus for avoiding quality deterioration of transmitted media content | |
CN109429070A (en) | A kind of mobile terminal video coding method, device and mobile terminal | |
CN114827620A (en) | Image processing method, apparatus, device and medium | |
CN111343475B (en) | Data processing method and device, live broadcast server and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20210528 |
|
RJ01 | Rejection of invention patent application after publication |