CN115601678A - Remote video conference safety guarantee method - Google Patents
Remote video conference safety guarantee method Download PDFInfo
- Publication number
- CN115601678A CN115601678A CN202211306722.0A CN202211306722A CN115601678A CN 115601678 A CN115601678 A CN 115601678A CN 202211306722 A CN202211306722 A CN 202211306722A CN 115601678 A CN115601678 A CN 115601678A
- Authority
- CN
- China
- Prior art keywords
- conference
- video
- probe
- frame
- target
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
Abstract
The invention relates to a remote video conference safety guarantee method, which solves the technical problem that the existing remote video conference is lack of abnormal behavior control, and comprises the following steps: establishing a conference, adding a probe and starting abnormal behavior detection in a system; the probe collects original videos and audios and uploads the videos and audios to a conference video analysis server; the conference video analysis server detects the video and feeds back the result to the probe; uploading, by the probe, the abnormal behavior record to the system; and for the detected abnormal behaviors, the video conference security probe performs masking control on the video in the conference to ensure that the candid shooting effect is poor and sensitive contents are not leaked. The invention can be widely applied to the security guarantee of the remote video conference.
Description
Technical Field
The invention relates to the field of video conferences, in particular to a remote video conference safety guarantee method.
Background
With the continuous improvement of the informatization level, the working modes of video conferences, remote consultation, remote diagnosis and the like greatly improve the office efficiency, and the utilization rate of the video conferences is higher and higher. However, in the process of holding a sensitive indoor conference, various security risks relating to sensitive words, candid shooting of a conference screen, leakage of a conference video and the like are faced, so that great challenges are brought to the use of remote videos, and the use of the remote videos is also limited. At present, a private network is needed for a video conference in a secret-related video conference, and a private network can be established on a public network by a VPN private network to carry out encryption communication. The method is widely applied to the secret-related enterprise network. The VPN gateway realizes remote access through encryption of a data packet and conversion of a data packet target address, but abnormal behaviors such as candid shooting in a conference cannot be managed and controlled. Therefore, it is necessary to manage abnormal behaviors such as candid shooting in a remote video conference.
Disclosure of Invention
The invention provides a remote video conference safety guarantee method for effectively controlling candid behaviors in order to solve the technical problem that abnormal behaviors are not controlled in the existing remote video conference.
The invention provides a remote video conference safety guarantee method, which comprises the following steps:
step 1, establishing a conference, adding a probe and starting abnormal behavior detection in a system;
step 2, the probe acquires an original video and uploads the original video to a conference video analysis server;
step 3, the conference video analysis server detects the video and feeds the result back to the probe;
step 4, uploading the abnormal behavior record to a system by the probe;
and 5, for the detected abnormal behaviors, the video conference security probe executes masking control on video contents in the conference to ensure that the candid shooting effect is poor and sensitive contents are not leaked.
Preferably, the specific steps of step 1 are:
step 11, establishing a conference and setting a conference watermark;
and 12, adding a probe in equipment management and carrying out equipment entry, wherein the equipment number is consistent with the communication number of the equipment.
Preferably, the specific steps of step 3 are:
step 31, making conference behavior data set pictures, selecting 60% -70% of the conference behavior data set pictures as training data sets, marking the specific targets in the pictures, wherein the data set pictures are character targets with candid behaviors;
step 32, performing model training by using the conference behavior data set, and performing parameter training of the model by using an improved Yolov4 algorithm to obtain an optimal training parameter of the model; the improved YOLOV4 algorithm changes a main network into a lighter-weight network mobileNet, and uses a channel and space attention mechanism CBAM to enhance the extraction of features;
step 33, reading the teleconferencing video, loading a detection model and training parameters, carrying out frame-by-frame detection on images of a video sequence, and if abnormal candid behavior is detected, storing a detection result and returning the detection result to a probe interface for subsequent processing;
step 34, the conference video analysis server and the probe are linked through a folder: when a conference is initiated, the probe stores a photo file in a directory specified by the NFS client, and the file naming format is as follows: IP., serial No. jpg; the component continuously scans the directory shared by the NFS server, performs abnormal behavior analysis if the directory has a file, and after the analysis is completed, if the directory is abnormal, tags and probability values of the pictures are printed on the abnormal picture folder, and the original pictures are deleted; if the image is normal, the original image is directly deleted; acquiring the equipment identification of the probe according to the folder where the picture is located, and feeding back a detection result to the probe;
and step 35, the conference video analysis server obtains the ip address reported by the file by analyzing the file name and splitting, and feeds back the analysis result to the probe.
Preferably, the specific steps of step 31 are:
step 311, acquiring a picture of a person candid behavior comprising different camera angles and light intensities, wherein the picture requires the behaviors of lifting a mobile phone and aligning a video camera by a person target;
and step 312, labeling the picture by using a labeling image labeling tool, wherein the labeling comprises labeling a target category of a target in the image and a target position in the image, the labeled picture is used as a conference behavior data set, the labeled picture is scaled to a size of 960 × 540, and model training is performed, so that the weight of the model training is obtained.
Preferably, the specific steps of step 32 are:
step 321, training and learning the conference behavior data set obtained in step 31 by using a YOLOV4 algorithm model, performing iterative computation on the input sample set for tens of thousands of times, and then adjusting the pre-trained model parameters by using a designed loss function, so as to obtain an optimal model weight parameter file;
step 322, loading the video of the conference scene as input into a detection model to obtain a detection result of an abnormal behavior target in each frame of the video sequence;
for video sequence Frame t Carrying out cyclic reading, wherein t is the serial number of the video frame and the value range is 1-n;
I t pixel information representing an image of a t-th frame, wherein I t The image feature extraction method comprises the steps of including the Width, height and Size of a picture and the pixel value of each pixel point, and providing a data basis for feature extraction of a target;
using DB t Representing the target detection result of the t-th frame image, DB t ={BB i I =1,2, …, n }, where BB i Information indicating an ith detection target in the t-th frame;
with PB i Denotes BB i In the Frame t+1 A predicted target location;
BB i the specific target information contained is: midpoint coordinates Cent = (x, y) of the object detection envelope box, width and Height of the object detection envelope box, and Size, the object is at the current I t Pixel information Roi, confidence P, class, frame number t, and Frame in middle envelope box t+1 A predicted target position PB, where PB has an alignment with BB i The same attribute.
The invention has the beneficial effects that:
the invention can detect abnormal behaviors such as candid photograph of the meeting place in real time only according to the video image and the audio of the meeting place under the environment without other sensors. And reading the video of the conference site through a probe interface, loading the trained model hyper-parameters and managing the system, and immediately cutting off the real-time input of the video if abnormal candid shooting exists, thereby playing a role in keeping the conference site secret.
Drawings
FIG. 1 is an exemplary image of a teleconference of the present invention;
FIG. 2 is an image of a model training process of the present invention;
FIG. 3 is an image of the test results of the present invention;
fig. 4 is a schematic diagram of an improved Yolov4 model.
Detailed Description
The present invention is further described below in conjunction with the drawings and examples to enable those skilled in the art to practice the present invention.
Example 1: as shown in fig. 1-3, the steps of the present invention are:
(1) Conference basic data such as a conference name, a watermark, a conference start and end time and the like are maintained, and the correspondence between the conference and the probe device is maintained.
(2) The embodiment is specifically a conference video of a certain conference room, the video comprises specific target detection behaviors, namely behaviors that participants take candid shots of a reporter by using camera equipment such as a mobile phone and the like, and the video is loaded into a model as input for detection. 5000 pictures of the behavior of candid shooting of people with different camera angles and light intensities are collected.
(3) And selecting 60-70% of sample images as training samples, wherein the number of the training samples is about 3000, performing model training, and storing the hyper-parameters after 100 rounds of iterative training.
(4) And testing 30% -40% of picture data of the rest samples, loading the trained hyper-parameters, and obtaining a detection result shown in figure 3 through model detection in the scheme, wherein the detection result comprises candid behavior and a detection probability value corresponding to the candid behavior. The detection precision of the model in the scheme under a specific data set can be calculated to reach 95% by using the formula (1).
Where TP is an image where the classifier detects a positive sample and is actually a positive sample, and FP is an image where the classifier detects a positive sample but is not actually a positive sample.
(5) And analyzing the video conference content by loading the trained model hyper-parameters.
(6) And immediately cutting off the real-time input of the video for abnormal candid behavior and carrying out masking control on sensitive content.
The method utilizes a conference behavior data set to train the model, and adopts an improved YOLOV4 algorithm to train the parameters of the model, so as to obtain the optimal training parameters of the model; an improved strategy is to change the backbone network into a more lightweight network mobileNet and use a channel and spatial attention mechanism CBAM to enhance feature extraction, as shown in fig. 4.
According to the implementation of the technical scheme, the control on abnormal behaviors is added in the remote video conference through a set of remote video conference safety guarantee system, so that the risk of sensitive information leakage in the conference can be greatly reduced, and the invention can also ensure that the video can be traced from the leaked video to the video outlet after a leakage event occurs.
The above description is only for the purpose of illustrating preferred embodiments of the present invention and is not to be construed as limiting the present invention, and it is apparent to those skilled in the art that various modifications and variations can be made in the present invention. All modifications, equivalents, improvements and the like which come within the scope of the invention as defined by the claims should be understood as falling within the scope of the invention.
Claims (5)
1. A remote video conference safety guarantee method is characterized by comprising the following steps:
step 1, establishing a conference, adding a probe and starting abnormal behavior detection in a system;
step 2, the probe acquires an original video and uploads the original video to a conference video analysis server;
step 3, the conference video analysis server detects the video and feeds back the result to the probe;
step 4, uploading the abnormal behavior record to a system by the probe;
and 5, for the detected abnormal behaviors, the video conference security probe executes masking control on video contents in the conference to ensure that the candid effect is poor and sensitive contents are not leaked.
2. The remote video conference security assurance method according to claim 1, wherein the specific steps of the step 1 are as follows:
step 11, establishing a conference and setting a conference watermark;
and 12, adding a probe in equipment management and performing equipment entry, wherein the equipment number is consistent with the communication number of the equipment.
3. The remote video conference security assurance method according to claim 1, wherein the specific steps of the step 3 are as follows:
step 31, making conference behavior data set pictures, selecting 60% -70% of the conference behavior data set pictures as a training data set, marking specific targets in the pictures, wherein the data set pictures are character targets with candid shooting behaviors;
step 32, performing model training by using the conference behavior data set, performing parameter training of the model by using an improved Yolov4 algorithm, and acquiring an optimal training parameter of the model; the improved YOLOV4 algorithm changes a main network into a lighter-weight network mobileNet, and uses a channel and space attention mechanism CBAM to enhance the extraction of features;
step 33, reading the teleconferencing video, loading a detection model and training parameters, carrying out frame-by-frame detection on images of a video sequence, and if abnormal candid behavior is detected, storing a detection result and returning the detection result to a probe interface for subsequent processing;
step 34, the conference video analysis server and the probe are linked through a folder: when a conference is initiated, the probe stores a photo file in a directory specified by the NFS client, and the file naming format is as follows: IP., serial No. jpg; the component continuously scans the directory shared by the NFS server, performs abnormal behavior analysis if the directory has a file, and after the analysis is completed, if the directory is abnormal, tags and probability values of the pictures are printed on the abnormal picture folder, and the original pictures are deleted; if the image is normal, the original image is directly deleted; acquiring the equipment identification of the probe according to the folder where the picture is located, and feeding back a detection result to the probe;
and step 35, the conference video analysis server obtains the ip address reported by the file by analyzing the file name and splitting, and feeds back the analysis result to the probe.
4. The remote video conference security assurance method according to claim 3, wherein the specific steps of the step 31 are as follows:
step 311, acquiring a picture of a person candid behavior comprising different camera angles and light intensities, wherein the picture requires the behaviors of lifting a mobile phone and aligning a video camera by a person target;
and step 312, labeling the picture by using a labeling image labeling tool, wherein the labeling comprises labeling a target category of a target in the image and a target position in the image, the labeled picture is used as a conference behavior data set, the labeled picture is scaled to a size of 960 × 540, and model training is performed, so that the weight of the model training is obtained.
5. The remote video conference security assurance method according to claim 3, wherein the specific steps of the step 32 are as follows:
step 321, training and learning the conference behavior data set obtained in step 31 by using a YOLOV4 algorithm model, performing iterative computation on the input sample set for tens of thousands of times, and then adjusting the pre-trained model parameters by using a designed loss function, so as to obtain an optimal model weight parameter file;
step 322, loading the video of the conference scene as input into a detection model to obtain a detection result of the abnormal behavior target in each frame of the video sequence;
for video sequence Frame t Carrying out cyclic reading, wherein t is the serial number of the video frame and the value range is 1-n;
I t pixel information representing an image of a t-th frame, wherein I t The image feature extraction method comprises the steps of including the Width, height and Size of a picture and the pixel value of each pixel point, and providing a data basis for feature extraction of a target;
using DB t Representing the target detection result of the t-th frame image, DB t ={BB i I =1,2, …, n }, where BB i Information indicating an ith detection target in the t-th frame;
with PB i Denotes BB i In the Frame t+1 A predicted target location;
BB i the specific target information contained is: midpoint coordinates Cent = (x, y) of the object detection envelope box, width and Height of the object detection envelope box, and Size, the object is at the current I t Pixel information Roi, confidence P, class, frame number t, and Frame in middle envelope box t+1 A predicted target position PB, where PB has an alignment with BB i The same attribute.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211306722.0A CN115601678A (en) | 2022-10-25 | 2022-10-25 | Remote video conference safety guarantee method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211306722.0A CN115601678A (en) | 2022-10-25 | 2022-10-25 | Remote video conference safety guarantee method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN115601678A true CN115601678A (en) | 2023-01-13 |
Family
ID=84848925
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202211306722.0A Pending CN115601678A (en) | 2022-10-25 | 2022-10-25 | Remote video conference safety guarantee method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115601678A (en) |
-
2022
- 2022-10-25 CN CN202211306722.0A patent/CN115601678A/en active Pending
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5612310B2 (en) | User interface for face recognition | |
EP3063731B1 (en) | Image cache for replacing portions of images | |
KR101810578B1 (en) | Automatic media sharing via shutter click | |
WO2021029648A1 (en) | Image capturing apparatus and auxiliary photographing method therefor | |
US20150356345A1 (en) | Systems and methods for detecting, identifying and tracking objects and events over time | |
US8406571B2 (en) | Automatic super-resolution transformation for images | |
US20120086792A1 (en) | Image identification and sharing on mobile devices | |
CN110543811B (en) | Deep learning-based non-cooperative examination personnel management method and system | |
US20210258584A1 (en) | Static video recognition | |
MX2011009714A (en) | Method and apparatus for video authentication of user. | |
CN110765134A (en) | File establishing method, equipment and storage medium | |
CN112601022B (en) | On-site monitoring system and method based on network camera | |
US20130343618A1 (en) | Searching for Events by Attendants | |
CN112004046A (en) | Image processing method and device based on video conference | |
CN114128255A (en) | Video conferencing using hybrid edge/cloud reasoning in conjunction with machine learning systems | |
CN105072478A (en) | Life recording system and method based on wearable equipment | |
US8300256B2 (en) | Methods, systems, and computer program products for associating an image with a communication characteristic | |
CN115601678A (en) | Remote video conference safety guarantee method | |
US20230231973A1 (en) | Streaming data processing for hybrid online meetings | |
US20060257003A1 (en) | Method for the automatic identification of entities in a digital image | |
CN114143429B (en) | Image shooting method, device, electronic equipment and computer readable storage medium | |
CN113329137B (en) | Picture transmission method, device, computer equipment and computer readable storage medium | |
CN112199547A (en) | Image processing method and device, storage medium and electronic equipment | |
KR102460292B1 (en) | COLLECTIVE BUILDING VoC(Voice of Customer) PROCESSING SYSTEM | |
CN111858992B (en) | Hydraulic engineering photo management method and system based on GPS and tag information |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |