CN116668806B - Method and device for adding target tracking mark at playing end - Google Patents
Method and device for adding target tracking mark at playing end Download PDFInfo
- Publication number
- CN116668806B CN116668806B CN202310916067.9A CN202310916067A CN116668806B CN 116668806 B CN116668806 B CN 116668806B CN 202310916067 A CN202310916067 A CN 202310916067A CN 116668806 B CN116668806 B CN 116668806B
- Authority
- CN
- China
- Prior art keywords
- target tracking
- information
- video
- frame
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 34
- 238000004891 communication Methods 0.000 claims abstract description 6
- 230000007246 mechanism Effects 0.000 claims abstract description 6
- 230000008569 process Effects 0.000 claims description 8
- 230000000153 supplemental effect Effects 0.000 claims description 4
- 238000013507 mapping Methods 0.000 claims description 3
- 238000012545 processing Methods 0.000 description 4
- 238000013473 artificial intelligence Methods 0.000 description 3
- 238000011160 research Methods 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 239000000969 carrier Substances 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000000149 penetrating effect Effects 0.000 description 1
- 230000008092 positive effect Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/845—Structuring of content, e.g. decomposing content into time segments
- H04N21/8456—Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/46—Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/02—Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L69/00—Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
- H04L69/16—Implementation or adaptation of Internet protocol [IP], of transmission control protocol [TCP] or of user datagram protocol [UDP]
- H04L69/161—Implementation details of TCP/IP or UDP/IP stack architecture; Specification of modified or new header fields
- H04L69/162—Implementation details of TCP/IP or UDP/IP stack architecture; Specification of modified or new header fields involving adaptations of sockets based mechanisms
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/4302—Content synchronisation processes, e.g. decoder synchronisation
Abstract
The invention provides a method and a device for adding a target tracking mark at a playing end, comprising the following steps: acquiring video information and characteristic image information of a vehicle; determining a video key frame according to video information of a vehicle, and calculating target tracking information data of characteristic image information in the video key frame, wherein the target tracking information data comprises position information and time stamp information; marking target tracking information data, and sending the target tracking information data to the front end through a WebSocket communication protocol; the front end separates video stream data and target tracking information data according to a predefined protocol; and playing the video stream data, obtaining frame data of target tracking by adopting a positioning strategy, and realizing a synchronization mechanism according to the time stamp. The method solves the problems that the targets are added with marks on the original video after the video is processed by the current target tracking, the marked video is required to be backed up independently, and multiple video backups are required when different targets are tracked and marked by the same video.
Description
Technical Field
The invention relates to the technical field of video tracking and display, in particular to a method and a device for adding a target tracking mark by a playing end.
Background
In recent years, with the development and interactive integration of big data, cloud computing, artificial intelligence and other fields, concepts such as smart electronic commerce, smart transportation, smart city and the like are attracting attention. Along with the heading of people for more intelligent, more convenient and higher-quality life, and with great academic value and wide commercial prospect, a great deal of manpower, material resources and financial resources are invested in related industries by a plurality of universities, scientific research institutions and the like. Artificial intelligence is silently penetrating into various industries and changing our lifestyle. Computer vision is an important branch of the field of artificial intelligence, aiming at studying how to let a computer perceive, analyze, process the real world as intelligently as the human visual system. Various computer vision algorithms using images and videos as information carriers have been penetrated into the daily life of the public, such as face recognition, man-machine interaction, commodity retrieval, intelligent monitoring, visual navigation and the like. Video object tracking technology, one of the fundamental, important research directions in the field of computer vision, has been a focus of attention for researchers.
Video object tracking requires continuous localization and scale estimation of the object in subsequent video frames given the position and scale information of the object of interest in the first frame. However, the current target tracking adds marks to the targets on the original video after processing the video, so that not only is the marked video required to be backed up independently, but also multiple copies of video are required to be backed up when the same video tracks and marks different targets. There is now a lack of a method to display tracking objects while video is being played and presented without changing and adding video files.
Disclosure of Invention
In view of the above, the present invention provides a method and apparatus for adding a target tracking mark at a playing end.
In order to solve the technical problems, the invention adopts the following technical scheme: a method for adding a target tracking mark at a playing end comprises the following steps:
s1: acquiring video information and characteristic image information of a vehicle;
s2: determining a video key frame according to video information of a vehicle, and calculating target tracking information data of characteristic image information in the video key frame, wherein the target tracking information data comprises position information and time stamp information;
s3: marking target tracking information data, and sending the target tracking information data to the front end through a WebSocket communication protocol;
s4: the front end separates video stream data and target tracking information data according to a predefined protocol;
s5: and playing the video stream data, obtaining frame data of target tracking by adopting a positioning strategy, and realizing a synchronization mechanism according to the time stamp.
In the present invention, preferably, the specific steps of determining video key frames according to the video information of the vehicle are as follows:
the video information comprises a plurality of image frames which are arranged according to time sequence, each image frame corresponds to a plurality of characteristic points in the characteristic image information, and the image frame of the first frame is used as a first key frame;
and traversing the image frames after the first key frame, judging that the difference value between the characteristic points in one image frame and the characteristic points of the first key frame is smaller than a preset value, and determining that the image frame is a second key frame.
In the present invention, preferably, the first keyframe includes position information of an object of interest and a region of interest, the position information of the region of interest is tracked by the image acquisition device, and the first keyframe is generated according to the acquired region of interest including the object of interest, where the region of interest includes coordinates of key points in the region of interest and a length and a width of the region of interest.
In the invention, preferably, calculating the target tracking information data of the characteristic image information in the video key frame is specifically to input the characteristic points corresponding to the first key frame and the second key frame into a preset object judgment model of interest for processing, and establish a mapping relation among the image frame, the characteristic points and the target tracking information, wherein the object judgment model of interest comprises a first judgment model and a second judgment model, the characteristic points and the position information are matched through the first judgment model to obtain the position information of the video key frame, and the second judgment model is called to match the characteristic points and the timestamp information after the matching is completed to obtain the timestamp information of the video key frame.
In the present invention, preferably, the position information is set to a frame number of the image frame.
In the present invention, preferably, the specific process of marking the target tracking information data is: and marking the calculated target tracking information data as additional information, and adding the additional information into a supplementary enhancement information frame of the video.
In the present invention, preferably, the supplemental enhancement information frame supports h.264 standard characteristics.
In the present invention, preferably, the positioning strategy specifically adopts a cascading style sheet method to obtain frame data of target tracking.
The invention has the advantages and positive effects that: according to the method, a video key frame is determined according to video information of a vehicle, target tracking information data of characteristic image information in the video key frame is calculated, and the target tracking information data comprises position information and time stamp information; marking target tracking information data, and sending the target tracking information data to the front end through a WebSocket communication protocol; the front end separates video stream data and target tracking information data according to a predefined protocol; the video stream data is played, frame data for target tracking is obtained by adopting a positioning strategy, and a synchronization mechanism is realized according to a time stamp, so that the problem that the targets are marked on the original video after the current target tracking is processed is solved, the marked video is required to be backed up independently, and multiple video backups are required when different targets are tracked and marked by the same video.
Drawings
The accompanying drawings are included to provide a further understanding of the invention and are incorporated in and constitute a part of this specification, illustrate the invention and together with the embodiments of the invention, serve to explain the invention. In the drawings:
fig. 1 is a flow chart of a method for adding a target tracking mark at a playing end according to the present invention.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
It will be understood that when an element is referred to as being "fixed to" another element, it can be directly on the other element or intervening elements may also be present. When a component is considered to be "connected" to another component, it can be directly connected to the other component or intervening components may also be present. When an element is referred to as being "disposed on" another element, it can be directly on the other element or intervening elements may also be present. The terms "vertical," "horizontal," "left," "right," and the like are used herein for illustrative purposes only.
Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. The terminology used herein in the description of the invention is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. The term "and/or" as used herein includes any and all combinations of one or more of the associated listed items.
As shown in fig. 1, the present invention provides a method for adding a target tracking mark at a playing end, which includes the following steps:
s1: acquiring video information and characteristic image information of a vehicle;
s2: determining a video key frame according to video information of a vehicle, and calculating target tracking information data of characteristic image information in the video key frame, wherein the target tracking information data comprises position information and time stamp information;
s3: marking target tracking information data, and sending the target tracking information data to the front end through a WebSocket communication protocol;
s4: the front end separates video stream data and target tracking information data according to a predefined protocol;
s5: and playing the video stream data, obtaining frame data of target tracking by adopting a positioning strategy, and realizing a synchronization mechanism according to the time stamp. The identity of the vehicles is uniquely determined by a reader unit that collects the vehicle's electronic tag radio frequency signals, each corresponding to several video messages.
In this embodiment, further, the specific steps of determining the video key frame according to the video information of the vehicle are as follows:
the video information comprises a plurality of image frames which are arranged according to time sequence, each image frame corresponds to a plurality of characteristic points in the characteristic image information, and the image frame of the first frame is used as a first key frame;
and traversing the image frames after the first key frame, judging that the difference value between the characteristic points in one image frame and the characteristic points of the first key frame is smaller than a preset value, and determining that the image frame is a second key frame.
In this embodiment, further, the first keyframe includes position information of the object of interest and the region of interest, the position information of the region of interest is tracked by the image acquisition device, and the first keyframe is generated according to the acquired region of interest including the object of interest, where the region of interest includes coordinates of key points in the region of interest and a length and a width of the region of interest. The method is a process of extracting the region of interest from the video information so as to be more targeted for tracking the vehicle target, and further reduces the problem of large occupied space of hardware equipment in the process of extracting the region of interest and reduces the load of image processing.
In this embodiment, further, calculating the target tracking information data of the feature image information in the video key frame specifically includes inputting feature points corresponding to the first key frame and the second key frame into a preset object judgment model of interest for processing, and establishing a mapping relationship between the image frame, the feature points and the target tracking information, where the object judgment model of interest includes a first judgment model and a second judgment model, the first judgment model is used to match the feature points with the position information to obtain the position information of the video key frame, and after matching is completed, the second judgment model is called to match the feature points with the timestamp information to obtain the timestamp information of the video key frame.
In this embodiment, further, the position information is set to the frame number of the image frame.
In this embodiment, further, the specific process of marking the target tracking information data is: and marking the calculated target tracking information data as additional information, and adding the additional information into a supplementary enhancement information frame of the video.
In this embodiment, further, the supplemental enhancement information frame supports h.264 standard characteristics.
In this embodiment, further, the positioning policy specifically obtains the frame data of the target tracking by using a cascading style sheet method. Specifically, the initial position of the image frame tracked by the target is defined by using the cascading style sheet, the relative positioning of the div label is defined, the absolute positioning of the dynamic picture in the img label is defined, and the initial position of the dynamic picture in the img label is defined.
The working principle and working process of the invention are as follows: the image acquisition device acquires video and tracking target information, performs video AI tracking according to the tracking target information through the target tracking unit, separates the position information of the tracking target according to the AI tracking result, and adds the information into an SEI (supplemental enhancement information) frame of the video. SEI frames, which belong to the category of code streams, provide a method of adding additional information to a video code stream, which is one of the characteristics of the h.264 standard. The method is not an essential option in the video decoding process, and can be integrated into a video stream for playing and displaying, and when playing, the additional information is extracted and marked on the picture of each frame for displaying.
The device comprises equipment which is installed on the equipment for recording video and collecting characteristic target information, and can be specifically divided into a reader unit for collecting radio frequency signals of an automobile electronic tag and a special camera for collecting target image information.
According to the method, a video key frame is determined according to video information of a vehicle, target tracking information data of characteristic image information in the video key frame is calculated, and the target tracking information data comprises position information and time stamp information; marking target tracking information data, and sending the target tracking information data to the front end through a WebSocket communication protocol; the front end separates video stream data and target tracking information data according to a predefined protocol; the video stream data is played, frame data for target tracking is obtained by adopting a positioning strategy, and a synchronization mechanism is realized according to a time stamp, so that the problem that the targets are marked on the original video after the current target tracking is processed is solved, the marked video is required to be backed up independently, and multiple video backups are required when different targets are tracked and marked by the same video.
The foregoing describes the embodiments of the present invention in detail, but the description is only a preferred embodiment of the present invention and should not be construed as limiting the scope of the invention. All equivalent changes and modifications within the scope of the present invention are intended to be covered by this patent.
Claims (7)
1. The method for adding the target tracking mark at the playing end is characterized by comprising the following steps:
s1: acquiring video information and characteristic image information of a vehicle;
s2: determining a video key frame according to video information of a vehicle, and calculating target tracking information data of characteristic image information in the video key frame, wherein the target tracking information data comprises position information and time stamp information;
s3: marking target tracking information data, and sending the target tracking information data to the front end through a WebSocket communication protocol;
s4: the front end separates video stream data and target tracking information data according to a predefined protocol;
s5: playing video stream data, obtaining frame data of target tracking by adopting a positioning strategy, and realizing a synchronization mechanism according to a time stamp;
the specific steps of determining the video key frames according to the video information of the vehicle are as follows:
the video information comprises a plurality of image frames which are arranged according to time sequence, each image frame corresponds to a plurality of characteristic points in the characteristic image information, and the image frame of the first frame is used as a first key frame;
traversing the image frames after the first key frames, judging that the difference value between the characteristic points in one image frame and the characteristic points of the first key frames is smaller than a preset value, and determining that the image frame is a second key frame; the method comprises the steps of calculating target tracking information data of characteristic image information in video key frames, specifically inputting characteristic points corresponding to a first key frame and a second key frame into a preset object judgment model of interest to be processed, establishing a mapping relation among the image frames, the characteristic points and the target tracking information, wherein the object judgment model of interest comprises a first judgment model and a second judgment model, matching the characteristic points with position information through the first judgment model to obtain the position information of the video key frames, and calling the second judgment model to match the characteristic points with time stamp information after matching is completed to obtain the time stamp information of the video key frames.
2. The method for adding a target tracking mark to a playing end according to claim 1, wherein the first key frame includes an object of interest and position information of a region of interest, the position information of the region of interest is used for carrying out target tracking on the first key frame through the image acquisition device, and the target tracking mark is generated according to the acquired region of interest including the object of interest, and the region of interest includes coordinates of key points in the region of interest and length and width of the region of interest.
3. The method for adding a target tracking mark at a playback end according to claim 1, wherein said position information is set to a frame number of an image frame.
4. The method for adding the target tracking mark to the playing end according to claim 1, wherein the specific process of marking the target tracking information data is as follows: and marking the calculated target tracking information data as additional information, and adding the additional information into a supplementary enhancement information frame of the video.
5. The method for adding a target tracking mark at a playback end according to claim 4, wherein said supplemental enhancement information frame supports h.264 standard features.
6. The method for adding a target tracking mark to a playing end according to claim 1, wherein the positioning strategy specifically adopts a cascading style sheet method to obtain frame data of target tracking.
7. A device for adding a target tracking mark at a playing end, wherein a method for adding a target tracking mark at a playing end according to any one of claims 1 to 6 is adopted.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310916067.9A CN116668806B (en) | 2023-07-25 | 2023-07-25 | Method and device for adding target tracking mark at playing end |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310916067.9A CN116668806B (en) | 2023-07-25 | 2023-07-25 | Method and device for adding target tracking mark at playing end |
Publications (2)
Publication Number | Publication Date |
---|---|
CN116668806A CN116668806A (en) | 2023-08-29 |
CN116668806B true CN116668806B (en) | 2023-10-27 |
Family
ID=87724375
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310916067.9A Active CN116668806B (en) | 2023-07-25 | 2023-07-25 | Method and device for adding target tracking mark at playing end |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116668806B (en) |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004046647A (en) * | 2002-07-12 | 2004-02-12 | Univ Waseda | Method and device for tracking moving object based on dynamic image data |
FR3070766A1 (en) * | 2017-09-01 | 2019-03-08 | Thales | IMPROVED RADAR SYSTEM FOR ENHANCED TRACKING |
CN110868600A (en) * | 2019-11-11 | 2020-03-06 | 腾讯云计算(北京)有限责任公司 | Target tracking video plug-flow method, display method, device and storage medium |
CN112905824A (en) * | 2021-02-08 | 2021-06-04 | 智慧眼科技股份有限公司 | Target vehicle tracking method and device, computer equipment and storage medium |
CN113760091A (en) * | 2021-08-13 | 2021-12-07 | 江苏仁哆多智能化科技有限公司 | Mobile terminal perception computing technology and Internet of things technology application system |
CN115623264A (en) * | 2022-10-20 | 2023-01-17 | 上海哔哩哔哩科技有限公司 | Live stream subtitle processing method and device and live stream playing method and device |
CN116320506A (en) * | 2023-03-13 | 2023-06-23 | 苏州青林文化传播有限公司 | Stereoscopic interaction service management method for film and television videos |
-
2023
- 2023-07-25 CN CN202310916067.9A patent/CN116668806B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004046647A (en) * | 2002-07-12 | 2004-02-12 | Univ Waseda | Method and device for tracking moving object based on dynamic image data |
FR3070766A1 (en) * | 2017-09-01 | 2019-03-08 | Thales | IMPROVED RADAR SYSTEM FOR ENHANCED TRACKING |
CN110868600A (en) * | 2019-11-11 | 2020-03-06 | 腾讯云计算(北京)有限责任公司 | Target tracking video plug-flow method, display method, device and storage medium |
CN112905824A (en) * | 2021-02-08 | 2021-06-04 | 智慧眼科技股份有限公司 | Target vehicle tracking method and device, computer equipment and storage medium |
CN113760091A (en) * | 2021-08-13 | 2021-12-07 | 江苏仁哆多智能化科技有限公司 | Mobile terminal perception computing technology and Internet of things technology application system |
CN115623264A (en) * | 2022-10-20 | 2023-01-17 | 上海哔哩哔哩科技有限公司 | Live stream subtitle processing method and device and live stream playing method and device |
CN116320506A (en) * | 2023-03-13 | 2023-06-23 | 苏州青林文化传播有限公司 | Stereoscopic interaction service management method for film and television videos |
Also Published As
Publication number | Publication date |
---|---|
CN116668806A (en) | 2023-08-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Ma et al. | A saliency prior context model for real-time object tracking | |
CN109961051B (en) | Pedestrian re-identification method based on clustering and block feature extraction | |
CN107004271B (en) | Display method, display apparatus, electronic device, computer program product, and storage medium | |
CN102077580B (en) | Display control device, display control method | |
US8620026B2 (en) | Video-based detection of multiple object types under varying poses | |
CN110189333B (en) | Semi-automatic marking method and device for semantic segmentation of picture | |
CN112669349A (en) | Passenger flow statistical method, electronic equipment and storage medium | |
Zhang et al. | Coarse-to-fine object detection in unmanned aerial vehicle imagery using lightweight convolutional neural network and deep motion saliency | |
CN110458115B (en) | Multi-frame integrated target detection algorithm based on time sequence | |
CN112132103A (en) | Video face detection and recognition method and system | |
CN108876672A (en) | A kind of long-distance education teacher automatic identification image optimization tracking and system | |
Song et al. | A novel deep learning network for accurate lane detection in low-light environments | |
Zhang et al. | Fine-grained-based multi-feature fusion for occluded person re-identification | |
CN113572981B (en) | Video dubbing method and device, electronic equipment and storage medium | |
CN116668806B (en) | Method and device for adding target tracking mark at playing end | |
CN111310595B (en) | Method and device for generating information | |
CN109886996B (en) | Visual tracking optimization method | |
CN103974074A (en) | Education video and lantern slide synchronization method | |
CN112215205B (en) | Target identification method and device, computer equipment and storage medium | |
CN114067356B (en) | Pedestrian re-recognition method based on combined local guidance and attribute clustering | |
Chen et al. | Surveillance video summarisation by jointly applying moving object detection and tracking | |
Liu et al. | Fast tracking via spatio-temporal context learning based on multi-color attributes and pca | |
CN115019241A (en) | Pedestrian identification and tracking method and device, readable storage medium and equipment | |
CN114004854A (en) | System and method for processing and displaying slice image under microscope in real time | |
CN113378765A (en) | Intelligent statistical method and device for advertisement attention crowd and computer readable storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |