CN111327939A - Distributed teaching video processing system - Google Patents
Distributed teaching video processing system Download PDFInfo
- Publication number
- CN111327939A CN111327939A CN202010114831.7A CN202010114831A CN111327939A CN 111327939 A CN111327939 A CN 111327939A CN 202010114831 A CN202010114831 A CN 202010114831A CN 111327939 A CN111327939 A CN 111327939A
- Authority
- CN
- China
- Prior art keywords
- video
- module
- human eye
- eye feature
- human
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000003860 storage Methods 0.000 claims abstract description 20
- 238000005520 cutting process Methods 0.000 claims abstract description 11
- 230000001133 acceleration Effects 0.000 claims abstract description 10
- 230000005540 biological transmission Effects 0.000 claims abstract description 8
- 238000007639 printing Methods 0.000 claims description 19
- 238000013528 artificial neural network Methods 0.000 claims description 18
- 238000001514 detection method Methods 0.000 claims description 16
- 235000014510 cooky Nutrition 0.000 claims description 12
- 238000000034 method Methods 0.000 claims description 12
- 238000003062 neural network model Methods 0.000 claims description 11
- 238000013135 deep learning Methods 0.000 claims description 10
- 210000000887 face Anatomy 0.000 claims description 7
- 230000015572 biosynthetic process Effects 0.000 claims description 6
- 238000005065 mining Methods 0.000 claims description 6
- 238000003786 synthesis reaction Methods 0.000 claims description 6
- 238000006243 chemical reaction Methods 0.000 claims description 3
- 230000006870 function Effects 0.000 abstract description 6
- 238000005516 engineering process Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000003252 repetitive effect Effects 0.000 description 1
- 230000002194 synthesizing effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
- H04N21/4312—Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/165—Detection; Localisation; Normalisation using facial parts and geometric relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/19—Sensors therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/193—Preprocessing; Feature extraction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/21—Server components or server architectures
- H04N21/218—Source of audio or video content, e.g. local disk arrays
- H04N21/2181—Source of audio or video content, e.g. local disk arrays comprising remotely distributed storage units, e.g. when movies are replicated over a plurality of video servers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4402—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
- H04N21/440236—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by media transcoding, e.g. video is transformed into a slideshow of still pictures, audio is converted into text
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- General Health & Medical Sciences (AREA)
- Theoretical Computer Science (AREA)
- Oral & Maxillofacial Surgery (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Signal Processing (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Ophthalmology & Optometry (AREA)
- Databases & Information Systems (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Geometry (AREA)
- Evolutionary Computation (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a distributed teaching video processing system which comprises a teaching video file transmission module, a teaching video GPU acceleration processing module and a teaching video automatic eye mosaic making module, wherein each module can realize a corresponding software module by C + + language, so that the teaching video processing system capable of running on a server is realized. The teaching video file transmission module is used for downloading an original video file and uploading a processed video file from the video storage server. The teaching video GPU acceleration processing module can conduct GPU acceleration processing on video cutting, video transcoding and video resolution adjustment. The automatic human eye mosaic playing module for the teaching video can realize automatic mosaic playing on human eyes in the video. The invention transfers the local video processing to the server for processing, accelerates the video processing speed, provides the video automatic human eye mosaic-making function which is not provided by the video processing software, and simplifies and facilitates the video processing flow.
Description
Technical Field
The invention relates to the technical field of video processing, in particular to a distributed teaching video processing system.
Background
At present, in a learning management system, a teaching video of a teacher's lectures is generally required to be taken and then made into a network course for students to learn. However, since the size and resolution of the original captured teaching video are usually very large, in order to save bandwidth resources in the learning management system, some processing is usually performed on the videos, such as a series of operations of cropping the original video, or adjusting the resolution to reduce the size of the video, or transcoding the video. In addition, for some teaching videos of traditional Chinese medicine diagnosis, some high-order processing needs to be performed on the videos, such as playing mosaic on human eyes to further protect the privacy of patients.
In the face of these video processing tasks, we usually use some third-party video processing software, such as adobe playback, etc., to perform processing. However, these video processing tasks are generally very tedious and repetitive for teachers, and these video processing software are usually proprietary, and have more basic functions, lack high-level functions, and are not friendly to general users.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provides a distributed teaching video processing system, which fully exploits the parallel processing capability of a GPU to process teaching videos and solves the problems that the local teaching video processing speed of a user is low, the local processing operation of the user is complicated, third-party video processing software is required to be relied on and some video processing functions which are not possessed by the third-party video processing software are provided. The system serves as a role of a teaching video processing engine, on one hand, GPU acceleration can be carried out on traditional teaching video processing such as video transcoding, resolution and clipping, and on the other hand, the engine provides some high-order teaching video processing functions such as automatic mosaic printing by human eyes in teaching videos.
In order to achieve the purpose, the technical scheme provided by the invention is as follows: a distributed teaching video processing system comprising:
the teaching video file transmission module is used for downloading an original video file from the video storage server and uploading the video file processed by the video processing system;
the teaching video GPU acceleration processing module is used for adjusting the resolution of a video according to the video name and the video resolution input by a user, transcoding the video according to the video name and the video format input by the user, and cutting the video according to the video name, the starting time and the ending time input by the user;
the automatic human eye mosaic printing module for teaching video is used for decomposing teaching video into image frames, detecting human faces in the images, detecting human eye feature points in the human faces, printing mosaic on human eye feature point areas, and making the image frames and audio after printing into video.
Further, the teaching video file transmission module comprises a user login module, a Cookie processing module, a video file downloading module and a video file uploading module, wherein:
the user login module logs in the video storage server according to a user name and a password input by a user;
the Cookie processing module is responsible for storing Cookies returned by the video storage server to the local and loading the Cookies from the local;
the video file downloading module is responsible for downloading video files from the video storage server;
and the video file uploading module is responsible for uploading the processed video to the video storage server.
Further, teaching video GPU accelerates processing module includes video resolution ratio adjustment module, video transcoding module, video cutting module, wherein:
the video resolution adjusting module adjusts the video resolution: analyzing request parameters in the HTTP POST request to obtain a target resolution of a video, then using a video processing tool FFMPEG to adjust the resolution of a specified video file, fully mining the parallel computing capability of a GPU in the adjusting process, and accelerating video coding and decoding by using NVENC and NVDEC provided by NVIDIA to realize GPU accelerated video processing and improve the processing efficiency;
the video transcoding module converts the video format: analyzing request parameters in the HTTP POST request to obtain a target video format, then converting the video format of a specified video file by using a video processing tool FFMPEG, fully mining the parallel computing capability of a GPU in the conversion process, and accelerating video coding and decoding by using NVENC and NVDEC provided by NVIDIA to realize GPU accelerated video processing and improve the processing efficiency;
the video clipping module clips videos: the starting time and the ending time of the video needing to be cut are obtained by analyzing the request parameters in the HTTP POST request, and then the video processing tool FFMPEG is used for cutting the specified video file.
Further, the automatic human eye mosaic printing module for teaching video comprises a human face detection module in an image, a human eye feature point detection module in a human face, a continuous inter-frame human eye feature point tracking module, a human eye feature point region mosaic printing module and an image frame and audio synthesis video module, wherein:
the in-image face detection module automatically detects coordinate points of a face area in one image: pre-training a neural network model for detecting a human face based on deep learning, storing a weight value of the neural network to a local area, loading the weight value of the neural network from the local area when detecting the human face, and automatically detecting a coordinate area of the human face in an image by the model when inputting an image frame into the model;
the module for detecting the characteristic points of the human eyes in the human face automatically detects the coordinate points of the characteristic points of the human eyes in the human face: pre-training a neural network model based on deep learning for detecting human eye feature points from a human face, storing the weight value of the neural network to the local, loading the weight value of the neural network from the local during human eye feature point detection, and automatically detecting the human eye feature points of the human face by the model when a human face region in an image frame is input into the model;
the continuous inter-frame human eye feature point tracking module tracks the human eye feature points of two continuous frames of images: pre-training a neural network model tracked by human eye feature points and based on deep learning, storing the weight value of the neural network to the local, loading the weight value of the neural network from the local during the tracking of the human eye feature points, when inputting the face area of the next frame into the model, tracking the human eye feature points by a tracker, if the tracked human eye feature point result is in line with expectation, continuing to process the image of the next frame, otherwise, handing the result to a human eye feature point detection module to detect the human eye feature points again;
the human eye characteristic point region mosaic printing module performs mosaic printing on the human eye characteristic point region: after human eye feature points are detected in the image, calculating a corresponding mosaic area according to the human eye feature points, and then generating a corresponding human eye mosaic;
the image frame and audio synthesis video module synthesizes the coded image frame and audio into a video file: after the coded image frames are obtained, the image frames and audio are combined into the final coded video using the video processing tool FFMPEG in combination with the previously separated audio in the video.
Compared with the prior art, the invention has the following advantages and beneficial effects:
1. the teaching video processing task is transferred to the server from the local area to be processed, the server fully excavates the parallel processing capacity of the GPU, and combines a video processing tool FFMPEG to carry out acceleration processing on video transcoding, video resolution adjustment and video cutting, so that the video processing speed is greatly improved, the video processing flow is simplified, and a user does not need to learn and use third-party video processing software.
2. Based on the latest deep learning technology, the neural network technology and the powerful GPU computing power, the system provides a high-order automatic mosaic playing function for human eyes, which is not provided by common video processing software, by deploying a human face detection model, a human eye feature point detection model and a human eye feature point tracking model in the system.
3. The video processing system is distributed, a teaching video storage server is separated from a teaching video processing server, the video processing system is composed of a plurality of video processing servers, and load balancing is achieved by distributing processing requests through a Web server Nginx.
Drawings
FIG. 1 is a schematic diagram of the overall architecture of the system of the present invention.
Fig. 2 is a flowchart of the entire video processing process.
Detailed Description
The present invention will be further described with reference to the following specific examples.
The distributed teaching video processing system provided in this embodiment is a teaching video processing system developed using Visual Studio Code software and C + + and Python languages and running on a server, and as shown in fig. 1, the system forwards a request and performs load balancing by a Web server Nginx, and forwards a video processing request to a Fast CGI server for processing. The teaching video processing system comprises:
the teaching video file transmission module is used for downloading an original video file from the video storage server and uploading the video file processed by the video processing system;
the teaching video GPU acceleration processing module is used for adjusting the resolution of a video according to the video name and the video resolution input by a user, transcoding the video according to the video name and the video format input by the user, and cutting the video according to the video name and the start and end time input by the user;
the automatic human eye mosaic printing module for teaching video is used for decomposing teaching video into image frames, detecting human faces in the images, detecting human eye feature points in the human faces, printing mosaic on human eye feature point areas and manufacturing the coded image frames into video.
The teaching video file transmission module comprises a user login module, a Cookie processing module, a video file downloading module and a video file uploading module, wherein:
the user login module logs in the video storage server according to a user name and a password input by a user;
the Cookie processing module is responsible for storing Cookies returned by the video storage server to the local and loading the Cookies from the local;
the video file downloading module is responsible for downloading video files from the video storage server;
and the video file uploading module is responsible for uploading the processed video to the video storage server.
The teaching video GPU acceleration processing module comprises a video resolution adjusting module, a video transcoding module and a video cutting module, wherein:
the video resolution adjusting module adjusts the video resolution: analyzing request parameters in the HTTP POST request to obtain a target resolution of a video, then using an FFMPEG tool to adjust the resolution of a specified video file, fully mining the parallel computing capability of a GPU in the adjustment process, and accelerating video coding and decoding by using NVENC and NVDEC provided by NVIDIA to realize GPU accelerated video processing and improve the processing efficiency;
the video transcoding module converts the video format: analyzing request parameters in an HTTP POST request to obtain a target video format, then converting the video format of a specified video file by using an FFMPEG tool, fully mining the parallel computing capability of a GPU in the conversion process, and accelerating video coding and decoding by using NVENC and NVDEC provided by NVIDIA to realize GPU accelerated video processing and improve the processing efficiency;
the video clipping module clips videos: analyzing request parameters in the HTTP POST request to obtain the starting time and the ending time of the video to be cut, and then cutting the specified video file by using an FFMPEG tool;
the automatic human eye mosaic printing module for teaching video comprises a human face detection module in an image, a human eye feature point detection module in a human face, a human eye feature point region mosaic printing module and an image frame and audio synthesis video module, wherein:
the in-image face detection module automatically detects coordinate points of a face area in one image: a deep learning-based neural network model for detecting human faces is trained in advance, the model mainly improves a faceBox face detector, improves the network structure of the faceBox face detector, increases the depth and the width of the network, and then eliminates unnecessary indistinct samples in a system for training. After training the model, saving the weight value of the neural network model to the local. When the human face is detected, the weight value of the neural network is loaded from the local, and when an image frame is input into the neural network model, the model can automatically detect the coordinate area of the position of the human face in the image, wherein the coordinate area comprises the coordinate value of the upper left corner of the human face area, the coordinate value of the lower right corner of the human face area and the coordinate value of the center of the human face area.
The module for detecting the characteristic points of the human eyes in the human face automatically detects the coordinate points of the characteristic points of the human eyes in the human face: a deep learning-based neural network model for detecting human eye feature points from a human face is trained in advance, the model improves O-Net in an MTCNN model, and a data set is adjusted for training. After the model is trained, the weight value of the neural network is stored to the local. When detecting the human eye feature points, the weight value of the neural network is loaded from the local, and when a human face area in an image frame is input into the model, the model can automatically detect the human eye feature points of the human face;
the human eye feature point tracking module tracks the human eye feature points of two continuous frames of images: pre-training a deep learning-based neural network model tracked by human eye feature points, and storing the weight value of the neural network to the local. When tracking the human eye feature points, loading the weight value of the neural network from the local part, when inputting the human face area of the next frame into the model, the tracker firstly tracks the human eye feature points, if the tracked human eye feature point result is in line with the expectation, continuing to process the next frame of image, otherwise, handing the result to the human eye feature point detection module to detect the human eye feature points again;
the human eye characteristic point region mosaic printing module performs mosaic printing on the human eye characteristic point region: after human eye feature points are detected in the image, calculating a mosaic area according to the human eye feature points, and then generating a corresponding mosaic;
the image frame and audio synthesis video module synthesizes the coded image frame and audio into a video file: and after the coded image frame is obtained, combining the audio separated from the video, and synthesizing the image frame and the audio into a final video by using an FFMPEG tool.
As shown in fig. 2, the whole processing flow of the distributed teaching video processing system in this embodiment is that, first, a video is uploaded from a browser to a video storage server, after the video is uploaded successfully, the browser sends a video processing request to a video processing system, the video processing system downloads and processes the video from the video storage server after receiving the request, and after the processing is completed, the processed video is uploaded to the video storage server.
The above-mentioned embodiments are merely preferred embodiments of the present invention, and the scope of the present invention is not limited thereto, so that the changes in the shape and principle of the present invention should be covered within the protection scope of the present invention.
Claims (4)
1. A distributed teaching video processing system, comprising:
the teaching video file transmission module is used for downloading an original video file from the video storage server and uploading the video file processed by the video processing system;
the teaching video GPU acceleration processing module is used for adjusting the resolution of a video according to the video name and the video resolution input by a user, transcoding the video according to the video name and the video format input by the user, and cutting the video according to the video name, the starting time and the ending time input by the user;
the automatic human eye mosaic printing module for teaching video is used for decomposing teaching video into image frames, detecting human faces in the images, detecting human eye feature points in the human faces, printing mosaic on human eye feature point areas, and making the image frames and audio after printing into video.
2. A distributed instructional video processing system according to claim 1, wherein: the teaching video file transmission module comprises a user login module, a Cookie processing module, a video file downloading module and a video file uploading module, wherein:
the user login module logs in the video storage server according to a user name and a password input by a user;
the Cookie processing module is responsible for storing Cookies returned by the video storage server to the local and loading the Cookies from the local;
the video file downloading module is responsible for downloading video files from the video storage server;
and the video file uploading module is responsible for uploading the processed video to the video storage server.
3. A distributed instructional video processing system according to claim 1, wherein: the teaching video GPU acceleration processing module comprises a video resolution adjusting module, a video transcoding module and a video cutting module, wherein:
the video resolution adjusting module adjusts the video resolution: analyzing request parameters in the HTTP POST request to obtain a target resolution of a video, then using a video processing tool FFMPEG to adjust the resolution of a specified video file, fully mining the parallel computing capability of a GPU in the adjusting process, and accelerating video coding and decoding by using NVENC and NVDEC provided by NVIDIA to realize GPU accelerated video processing and improve the processing efficiency;
the video transcoding module converts the video format: analyzing request parameters in the HTTP POST request to obtain a target video format, then converting the video format of a specified video file by using a video processing tool FFMPEG, fully mining the parallel computing capability of a GPU in the conversion process, and accelerating video coding and decoding by using NVENC and NVDEC provided by NVIDIA to realize GPU accelerated video processing and improve the processing efficiency;
the video clipping module clips videos: the starting time and the ending time of the video needing to be cut are obtained by analyzing the request parameters in the HTTP POST request, and then the video processing tool FFMPEG is used for cutting the specified video file.
4. A distributed instructional video processing system according to claim 1, wherein: automatic people's eye of teaching video beats mosaic module includes in the image people's eye detection module, people's eye feature point detection module in the people's face, continuous interframe people's eye feature point tracking module, people's eye feature point region beat mosaic module and image frame and audio synthesis video module, wherein:
the in-image face detection module automatically detects coordinate points of a face area in one image: pre-training a neural network model for detecting a human face based on deep learning, storing a weight value of the neural network to a local area, loading the weight value of the neural network from the local area when detecting the human face, and automatically detecting a coordinate area of the human face in an image by the model when inputting an image frame into the model;
the module for detecting the characteristic points of the human eyes in the human face automatically detects the coordinate points of the characteristic points of the human eyes in the human face: pre-training a neural network model based on deep learning for detecting human eye feature points from a human face, storing the weight value of the neural network to the local, loading the weight value of the neural network from the local during human eye feature point detection, and automatically detecting the human eye feature points of the human face by the model when a human face region in an image frame is input into the model;
the continuous inter-frame human eye feature point tracking module tracks the human eye feature points of two continuous frames of images: pre-training a neural network model tracked by human eye feature points and based on deep learning, storing the weight value of the neural network to the local, loading the weight value of the neural network from the local during the tracking of the human eye feature points, when inputting the face area of the next frame into the model, tracking the human eye feature points by a tracker, if the tracked human eye feature point result is in line with expectation, continuing to process the image of the next frame, otherwise, handing the result to a human eye feature point detection module to detect the human eye feature points again;
the human eye characteristic point region mosaic printing module performs mosaic printing on the human eye characteristic point region: after human eye feature points are detected in the image, calculating a corresponding mosaic area according to the human eye feature points, and then generating a corresponding human eye mosaic;
the image frame and audio synthesis video module synthesizes the coded image frame and audio into a video file: after the coded image frames are obtained, the image frames and audio are combined into the final coded video using the video processing tool FFMPEG in combination with the previously separated audio in the video.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010114831.7A CN111327939A (en) | 2020-02-25 | 2020-02-25 | Distributed teaching video processing system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010114831.7A CN111327939A (en) | 2020-02-25 | 2020-02-25 | Distributed teaching video processing system |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111327939A true CN111327939A (en) | 2020-06-23 |
Family
ID=71171165
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010114831.7A Pending CN111327939A (en) | 2020-02-25 | 2020-02-25 | Distributed teaching video processing system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111327939A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111738769A (en) * | 2020-06-24 | 2020-10-02 | 湖南快乐阳光互动娱乐传媒有限公司 | Video processing method and device |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060104366A1 (en) * | 2004-11-16 | 2006-05-18 | Ming-Yen Huang | MPEG-4 streaming system with adaptive error concealment |
CN101420452A (en) * | 2008-12-05 | 2009-04-29 | 深圳市迅雷网络技术有限公司 | Video file publishing method and device |
CN109743579A (en) * | 2018-12-24 | 2019-05-10 | 秒针信息技术有限公司 | A kind of method for processing video frequency and device, storage medium and processor |
CN110418144A (en) * | 2019-08-28 | 2019-11-05 | 成都索贝数码科技股份有限公司 | A method of realizing that one enters to have more transcoding multi code Rate of Chinese character video file based on NVIDIA GPU |
-
2020
- 2020-02-25 CN CN202010114831.7A patent/CN111327939A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060104366A1 (en) * | 2004-11-16 | 2006-05-18 | Ming-Yen Huang | MPEG-4 streaming system with adaptive error concealment |
CN101420452A (en) * | 2008-12-05 | 2009-04-29 | 深圳市迅雷网络技术有限公司 | Video file publishing method and device |
CN109743579A (en) * | 2018-12-24 | 2019-05-10 | 秒针信息技术有限公司 | A kind of method for processing video frequency and device, storage medium and processor |
CN110418144A (en) * | 2019-08-28 | 2019-11-05 | 成都索贝数码科技股份有限公司 | A method of realizing that one enters to have more transcoding multi code Rate of Chinese character video file based on NVIDIA GPU |
Non-Patent Citations (1)
Title |
---|
吴家贤: "基于卷积神经网络的自动人眼马赛克系统", 《中国优秀硕士学位论文全文数据库 信息科技辑》 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111738769A (en) * | 2020-06-24 | 2020-10-02 | 湖南快乐阳光互动娱乐传媒有限公司 | Video processing method and device |
CN111738769B (en) * | 2020-06-24 | 2024-02-20 | 湖南快乐阳光互动娱乐传媒有限公司 | Video processing method and device |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11941883B2 (en) | Video classification method, model training method, device, and storage medium | |
WO2021114881A1 (en) | Intelligent commentary generation method, apparatus and device, intelligent commentary playback method, apparatus and device, and computer storage medium | |
CN111242962A (en) | Method, device and equipment for generating remote training video and storage medium | |
CN114419300A (en) | Stylized image generation method and device, electronic equipment and storage medium | |
CN110652726B (en) | Game auxiliary system based on image recognition and audio recognition | |
WO2023138549A1 (en) | Image processing method and apparatus, and electronic device and storage medium | |
CN113452944B (en) | Picture display method of cloud mobile phone | |
US20210117723A1 (en) | Method and system for auto multiple image captioning | |
WO2023125181A1 (en) | Image processing method and apparatus, electronic device, and storage medium | |
CN112383807A (en) | Online education method and device and electronic equipment | |
CN111327939A (en) | Distributed teaching video processing system | |
CN113689440A (en) | Video processing method and device, computer equipment and storage medium | |
CN114463470A (en) | Virtual space browsing method and device, electronic equipment and readable storage medium | |
US20210195134A1 (en) | Method and device for generating dynamic image, mobile platform, and storage medium | |
KR20170127354A (en) | Apparatus and method for providing video conversation using face conversion based on facial motion capture | |
CN111274925A (en) | Method and device for generating recommended video, electronic equipment and computer storage medium | |
CN115526772B (en) | Video processing method, device, equipment and storage medium | |
CN115052197B (en) | Virtual portrait video generation method and device | |
CN116229311A (en) | Video processing method, device and storage medium | |
CN115862172A (en) | Attendance system with expression discernment | |
KR102526263B1 (en) | Method and System for Auto Multiple Image Captioning | |
CN109255996A (en) | A kind of the broadcasting optimization method and system of Online class | |
CN113327306B (en) | Exclusive animation generation method and system based on hand shadow realization | |
US11665215B1 (en) | Content delivery system | |
KR102541599B1 (en) | Digital sheet music management system and method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20200623 |