CN112911195A - Online live television teleconference intelligent management system based on cloud computing and artificial intelligence - Google Patents

Online live television teleconference intelligent management system based on cloud computing and artificial intelligence Download PDF

Info

Publication number
CN112911195A
CN112911195A CN202110052425.7A CN202110052425A CN112911195A CN 112911195 A CN112911195 A CN 112911195A CN 202110052425 A CN202110052425 A CN 202110052425A CN 112911195 A CN112911195 A CN 112911195A
Authority
CN
China
Prior art keywords
speaking
person
module
speaker
current
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110052425.7A
Other languages
Chinese (zh)
Other versions
CN112911195B (en
Inventor
尹晓东
赵洋梵
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Weimai Technology Co.,Ltd.
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN202110052425.7A priority Critical patent/CN112911195B/en
Publication of CN112911195A publication Critical patent/CN112911195A/en
Application granted granted Critical
Publication of CN112911195B publication Critical patent/CN112911195B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems
    • H04N7/152Multipoint control units therefor
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems
    • H04N7/155Conference systems involving storage of or access to video conference sessions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems
    • H04N7/157Conference systems defining a virtual conference space and using avatars or agents
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The invention discloses an intelligent management system for online live television telephone conference based on cloud computing and artificial intelligence, the system comprises a participant information acquisition module, a conference flow importing module, a conference flow processing module, an information verification module, a time correction module, a speech tracking module, a warning module, a conference processing module and a database, wherein the conference processing module and the time judgment module are connected through the speech tracking module, the speaking time of each speaking person is tracked, the speaking time point of each speaking person is obtained, and then each speaking person is reminded in time, the intelligent management of the on-line live television conference is realized, the stability of the conference flow is ensured, meanwhile, the next speaker can be adjusted according to the current speaker's speaking time, so that the effect of the live video teleconference is effectively improved, and the management efficiency of the live video teleconference is improved.

Description

Online live television teleconference intelligent management system based on cloud computing and artificial intelligence
Technical Field
The invention belongs to the technical field of conference management, and relates to an intelligent management system for an online live television teleconference based on cloud computing and artificial intelligence.
Background
Under the promotion of big data development strategy, people can acquire information more conveniently, and the national government actively promotes the information disclosure of a plurality of public fields all the time, so that an online live television conference is generated, real-time information of the conference can be shared in a live broadcast mode, but later adjustment cannot be performed, and the online live television conference management is particularly important.
In traditional online live telecommand conference management mode, manage network and the equipment operation state that corresponds in this meeting room, therefore traditional online live telecommand conference management mode has still existed many drawbacks of violence, on the one hand can't trace each speech person's speech time, can't accomplish timely warning to each speech person, on the one hand can't adjust the speech time of next speech person according to current speech person's speech time, and then can't improve the effect of this live telecommand conference, on the other hand lacks the management to the meeting flow, can't the managerial efficiency of effectual guarantee online live telecommand conference.
Disclosure of Invention
In view of this, in order to solve the problems existing in the background art, an intelligent management system for an online live tv teleconference based on cloud computing and artificial intelligence is proposed, so as to realize intelligent management of the online live tv teleconference.
The purpose of the invention can be realized by the following technical scheme:
an online live television teleconference intelligent management system based on cloud computing and artificial intelligence comprises a participant information acquisition module, a conference flow importing module, a conference flow processing module, a speech tracking module, an information verification module, a time correction module, a warning module and a database;
the speech tracking module is respectively connected with the conference flow importing module, the information verification module, the database, the time judging module and the warning module, the conference flow importing module is respectively connected with the participant information acquisition module and the conference flow processing module, and the warning module is connected with the information verification module and the speech tracking module;
the conference personnel information acquisition module is used for acquiring basic information of the conference personnel, counting the number of the conference personnel, numbering the counted conference personnel according to a preset sequence, sequentially marking the number as 1,2, aSetting the corresponding seat position of each participant, and acquiring the corresponding seat number of each participant, wherein the basic information of each participant comprises the name of each participant, the face image of each participant and the seat number of each participant, thereby constructing a participant information set Fw(Fw1,Fw2,...,Fwj,...Fwm),Fwj represents the w-th basic information corresponding to the jth participant, w represents the basic information, and w is a1, a2, a3, a1, a2 and a3 which are respectively represented as the participant name, the participant image and the participant seat number;
the conference flow importing module is used for importing flow information of the online live television teleconference, the flow information of the online live television teleconference comprises conference starting time and conference ending time, speaking subjects corresponding to all speaking persons, speaking time points and speaking durations corresponding to all speaking persons, and the imported conference starting time points and the imported conference ending time points, the speaking subjects corresponding to all speaking persons, the speaking time points corresponding to all speaking persons and the speaking durations are sent to the conference flow processing module;
the conference flow processing module is used for receiving the conference starting time point and the conference ending time point sent by the conference flow importing module, the speaking subjects corresponding to the speakers, the speaking time points and the speaking durations corresponding to the speakers, sequencing the received flow information of the online live video teleconference according to the speaking time points of the speakers, further acquiring the conference flow corresponding to the conference, and sending the conference flow to the speaking tracking module;
the speech tracking module is used for receiving the conference flow sent by the conference flow processing module and tracking the time of the current speech person according to the flow, and comprises a primary speech tracking module and a secondary speech tracking module;
the first-stage speaking tracking module is used for recording the initial speaking time point of the current speaking person when the current speaking person starts speaking, timing the current speaking person at the initial speaking time point, tracking the current speaking person in real time, counting the timing speaking duration of the current speaking person, acquiring the time point corresponding to the timing speaking duration according to the corresponding speaking time point and the timing speaking duration of the current speaking person in the conference flow, recording the time point as the speaking process time point, comparing the speaking process time point corresponding to the current speaking person with the speaking reminding time point corresponding to the current speaking person, if the speaking process time point corresponding to the current speaking person is at the speaking reminding time point corresponding to the current speaking person, extracting the number of the current speaking person to be reminded, sending the number of the speaking person to the warning module, and according to the speaking time point and the speaking duration of the current speaking person corresponding to the conference flow, acquiring a predicted speech stop time point of a current speaker, comparing the predicted speech stop time point of the current speaker with a speech reminding time point corresponding to the current speaker, further acquiring a difference value between the predicted speech stop time point of the current speaker and the speech reminding time point corresponding to the current speaker, recording the difference value as the predicted speech stop time of the current speaker, further extracting a name corresponding to a next speaker from a conference flow, extracting a face image and a seat number of the next speaker from a participant information set according to the name of the speaker, recording the face image as an original face image, and sending the extracted original face image, the seat number and the predicted speech stop time of the current speaker to an information verification module;
the information verification module is used for verifying the information of the next speaking person, wherein the specific verification process of the information verification module comprises the following steps:
b1, receiving an original face image, a seat number and an estimated stop duration of the current speaker of the next speaker sent by the primary speaker tracking module, and recording the next speaker as a to-be-uttered speaker;
b2, according to the seat number corresponding to the person to be announced, calling the camera of the seat to collect the face image of the person with the seat number, and recording the collected face image as an actual face image;
b3, extracting the features of the actual face image corresponding to the seat number to obtain the actual face feature image of the person waiting to speak, and extracting the features of the original face image corresponding to the seat number to obtain the original face feature image of the person waiting to speak;
b4, comparing the actual face feature image of the person to be pronounced with the original face feature image of the person to be pronounced, and further obtaining the matching degree of the actual face feature image of the person to be pronounced and the original face feature image of the person to be pronounced;
b5, if the image matching degree difference is smaller than zero, judging that the face image corresponding to the seat number does not accord with the face image input by the person to be spoken, and further starting a camera of the seat where each participant is located to acquire the face image of each participant;
b6, further carrying out face feature extraction on the collected face images of the participants to obtain face feature images corresponding to the participants;
b7, comparing the face characteristic image corresponding to each participant with the original face characteristic image of the person to be spoken, further screening out the face characteristic image with the highest matching degree with the face image input by the person to be spoken, further extracting the seat number corresponding to the face characteristic image, recording the seat number as the actual seat number of the person to be spoken, and further sending the actual seat number of the person to be spoken and the estimated stop time of the current speaker to the warning module;
the warning module comprises a primary warning module and a secondary warning module and is used for reminding the current person and the person to be called;
the first-stage warning module is used for receiving the number of the current speaker to be reminded, which is sent by the conference processing module, and further reminding the current speaker;
the second-level warning module is used for receiving the actual seat number of the person to be uttered and the estimated stop time of the current speaker sent by the information verification module, calling the position of the person to be uttered according to the actual seat number of the person to be uttered, calling a display corresponding to the position of the speaker, and displaying the estimated stop time and the estimated signal to be uttered corresponding to the current speaker through the display;
the secondary speaking tracking module is used for acquiring the actual speaking duration of the current speaking person when the current speaking person stops speaking, and sending the actual speaking duration of the current speaking person to the time judging module;
the time judging module is used for receiving the actual speaking duration of the current speaking person sent by the secondary speaking tracking module, and compares the actual speaking duration of the current speaking person with the speaking duration of the current speaking person corresponding to the conference flow, further obtaining the difference value between the actual speaking time length of the current speaking person and the speaking time length of the current speaking person corresponding to the conference flow, if the difference value is larger than zero, judging the speaking timeout of the speaking person, counting the timeout time of the current speaking person, if the difference is less than zero, judging that the speaker finishes speaking in advance, counting the advance time of the current speaker, if the difference between the actual speaking time of the current speaker and the speaking time of the current speaker corresponding to the conference flow is equal to zero, judging that the speaking person stops speaking according to the speaking duration corresponding to the conference flow, and sending the overtime time of the current speaking person and the advance time of the current speaking person to a time correction module;
the time correction module is used for correcting the speaking time point of the next speaking person when the speaking time of the current speaking person is not equal to the speaking time of the current speaking person corresponding to the conference flow, wherein the time correction module comprises an overtime correction module and a forward correction module;
the overtime correction module is used for receiving the overtime of the current speaker sent by the time judgment module, delaying the starting time corresponding to the next speaker backwards according to the overtime of the current speaker and the speaker time point corresponding to the next speaker corresponding to the conference flow, counting the speaker time point corresponding to the next speaker after correction, recording the speaker time point as an overtime correction speaker time point, and sending the overtime correction speaker time point corresponding to the next speaker to the speaker tracking module;
the advance correction module is used for carrying out time advance on the starting time corresponding to the next speaker according to the advance time of the current speaker and the speaker time point corresponding to the next speaker corresponding to the conference flow, counting the speaker time point corresponding to the next speaker after correction, recording the corrected speaker time point as an advance corrected speaker time point, and sending the advance corrected speaker time point corresponding to the next speaker to the speaker tracking module;
the database is used for storing speaking reminding time points corresponding to all speaking persons.
Furthermore, the speaking time tracking module tracks speaking through a timing device and a sound detection device, wherein the timing device is a timer and is used for timing speaking time when a speaking person starts speaking, and the sound detection device is a sound sensor and is used for detecting the sound of the speaking person so as to judge whether the speaking person is in the speaking process.
Furthermore, the primary warning module reminds through a warning device, the warning device is a wearable warning terminal, when the speaking process time point corresponding to the current speaking person is located at the speaking reminding time point corresponding to the current speaking person, the current speaking person number needing to be reminded is extracted, and then the warning terminal corresponding to the number is called for vibration reminding.
Furthermore, the secondary level warning module reminds the current speaking person of estimating the stop duration and the signal to be spoken in a text reminding mode, wherein the text reminding mode is the increase of text fonts and the deepening of text colors, and therefore reminding of the person to be spoken is achieved.
Further, the system also comprises a device for installing the corresponding seats of the participants, wherein the device comprises a plurality of cameras and a plurality of displays, the seats of the participants correspond to one camera and one display respectively, the cameras are used for collecting images of the participants at the seats, and the displays are used for displaying video images corresponding to the current speakers and conference flows corresponding to the participants to be spoken.
Furthermore, the camera is high-definition and can automatically focus, and is used for collecting high-definition face images of each participant.
The invention has the beneficial effects that:
(1) according to the intelligent management system for the online live television teleconference based on the cloud computing and the artificial intelligence, the current speakers are tracked in real time through the speech tracking module and the warning module, the initial speech time point and the speech duration of each speaker are obtained, and then each speaker is reminded in time, so that the intelligent management of the online live television teleconference is realized, the stability of the conference flow is guaranteed, meanwhile, the next speaker can be adjusted according to the speech time of the current speaker, the effect of the live television teleconference is effectively improved, and the management efficiency of the live television teleconference is improved.
(2) According to the invention, the speaking time of the next speaking person is rapidly adjusted by correcting the speaking time point of the next speaking person in the time correction module, so that the progress of the live video teleconference is effectively promoted, and the order of the live video teleconference is effectively maintained.
(3) According to the invention, the warning module reminds the current speaker and the speakers to be sent, so that the current speaker and the speakers to be sent are reminded in time, and the smoothness of the progress of the television/telephone conference is effectively guaranteed.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings used in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art that other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a schematic diagram of the system module connections of the present invention;
fig. 2 is a schematic diagram of the connection of the utterance tracking module in the present invention;
FIG. 3 is a schematic diagram of an alarm module according to the present invention.
Detailed Description
While the foregoing is directed to embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow.
Referring to fig. 1, an online live tv conference intelligent management system based on cloud computing and artificial intelligence includes a participant information acquisition module, a conference flow import module, a conference flow processing module, a speech tracking module, an information verification module, a time correction module, a warning module, and a database;
the speech tracking module is respectively connected with the conference flow importing module, the information verification module, the database, the time judging module and the warning module, the conference flow importing module is respectively connected with the participant information acquisition module and the conference flow processing module, and the warning module is connected with the information verification module and the speech tracking module;
the participant information acquisition module is used for acquiring basic information of participants, counting the number of the participants, numbering the counted participants according to a preset sequence, sequentially marking the counted participants as 1,2, a, j, a, m, counting the number of seats in the conference room, numbering the seats in the conference room according to the preset sequence, sequentially marking the seats as 1,2, a, i, a, n, and acquiring the seat numbers corresponding to the participants according to the preset seat positions corresponding to the participants, wherein the basic information of the participants comprises names of the participants, face images of the participants and the seat numbers of the participants, and further establishing a participant information set Fw(Fw1,Fw2,...,Fwj,...Fwm),Fwj represents the jth participant correspondenceThe w-th basic information of the present invention, w represents basic information, w is a1, a2, a3, a1, a2 and a3 respectively represent names of participants, images of the participants and seat numbers of the participants, wherein each seat corresponds to a camera and a display, the camera is used for collecting images of the participants at the position, and the display is used for displaying video images corresponding to the current speaker and conference flows corresponding to the participants to be spoken;
the conference flow importing module is used for importing flow information of the online live television teleconference, the flow information of the online live television teleconference comprises conference starting time and conference ending time, speaking subjects corresponding to all speaking persons, speaking time points and speaking durations corresponding to all speaking persons, and the imported conference starting time points and the imported conference ending time points, the speaking subjects corresponding to all speaking persons, the speaking time points corresponding to all speaking persons and the speaking durations are sent to the conference flow processing module;
the conference flow processing module is used for receiving the conference starting time point and the conference ending time point sent by the conference flow importing module, the speaking subjects corresponding to the speaking persons, the speaking time points and the speaking durations corresponding to the speaking persons, and sequencing the received flow information of the online live television conference according to the speaking time points of the speaking persons, so that the conference flow corresponding to the conference is obtained, and the conference flow is sent to the speaking tracking module.
Referring to fig. 2, the utterance tracking module is configured to receive a conference flow sent by the conference flow processing module, where the utterance tracking module includes a primary utterance tracking module and a secondary utterance tracking module;
the embodiment of the invention tracks the current speakers in real time through the speaker tracking module, acquires the initial speaking time point and the speaking duration of each speaker, and further timely reminds each speaker, thereby realizing intelligent management of the on-line live television conference and ensuring the stability of the conference flow;
the first-stage speaking tracking module tracks the time of the current speaking person according to the flow through a timing device and a sound detection device, wherein the timing device is a timer and is used for timing the speaking time when the speaking person starts speaking, the sound detection device is a sound sensor and is used for detecting the sound of the speaking person so as to judge whether the current speaking person is in the speaking process, when the sound sensor detects the sound, the initial speaking time point of the current speaking person is recorded and timed, the speaking time of the current speaking person is counted, the initial speaking time point of the current speaking person is recorded, the current speaking person is tracked in real time, the timed speaking time of the current speaking person is counted, and the time point corresponding to the timed speaking time is obtained according to the speaking time point and the timed speaking time corresponding to the current speaking person in the conference flow, recording the current speech process time point as a speech process time point, comparing the speech process time point corresponding to the current speaker with a speech reminding time point corresponding to the current speaker, if the speech process time point corresponding to the current speaker is at the speech reminding time point corresponding to the current speaker, extracting the number of the current speaker to be reminded, sending the number of the speaker to a warning module, acquiring the estimated speech stop time point of the current speaker according to the speech time point and the speech duration of the current speaker corresponding to the conference flow, comparing the estimated speech stop time point of the current speaker with the speech reminding time point corresponding to the current speaker, further acquiring the difference value between the estimated speech stop time point of the current speaker and the speech reminding time point corresponding to the current speaker, and recording the difference value as the estimated speech stop time of the current speaker, further extracting a name corresponding to the next speaker from the conference flow, extracting a face image and a seat number of the next speaker from the participant information set according to the name of the speaker, recording the face image as an original face image, and sending the extracted original face image, the seat number and the estimated stop time of the current speaker to the information verification module;
the information verification module verifies information of the next speaking person through a camera, the camera is high-definition and can automatically focus, and the specific verification process of the information verification module for collecting high-definition face images of all participants comprises the following steps:
b1, receiving an original face image, a seat number and an estimated stop duration of the current speaker of the next speaker sent by the primary speaker tracking module, and recording the next speaker as a to-be-uttered speaker;
b2, according to the seat number corresponding to the person to be announced, calling the camera of the seat to collect the face image of the person with the seat number, and recording the collected face image as an actual face image;
b3, extracting the features of the actual face image corresponding to the seat number to obtain the actual face feature image of the person waiting to speak, and extracting the features of the original face image corresponding to the seat number to obtain the original face feature image of the person waiting to speak;
b4, comparing the actual face feature image of the person to be pronounced with the original face feature image of the person to be pronounced, and further obtaining the matching degree of the actual face feature image of the person to be pronounced and the original face feature image of the person to be pronounced;
b5, if the image matching degree difference is smaller than zero, judging that the face image corresponding to the seat number does not accord with the face image input by the person to be spoken, and further starting a camera of the seat where each participant is located to acquire the face image of each participant;
b6, further carrying out face feature extraction on the collected face images of the participants to obtain face feature images corresponding to the participants;
b7, comparing the face characteristic image corresponding to each participant with the original face characteristic image of the person to be spoken, further screening out the face characteristic image with the highest matching degree with the face image input by the person to be spoken, further extracting the seat number corresponding to the face characteristic image, recording the seat number as the actual seat number of the person to be spoken, and further sending the actual seat number of the person to be spoken and the estimated stop time of the current speaker to the warning module;
according to the embodiment of the invention, the information verification module verifies the information of the next speaker through the camera, so that the warning accuracy of the warning module is effectively guaranteed, and the stability of the live video teleconference is effectively maintained;
referring to fig. 3, the warning module includes a primary warning module and a secondary warning module, and is configured to remind a current person and a person to be called;
according to the embodiment of the invention, the warning module reminds the current speaker and the speakers to be announced, so that the current speaker and the speakers to be announced can be reminded in time, and the smoothness of the progress of the television teleconference can be effectively guaranteed.
The first-stage warning module is used for receiving a current speaker number which needs to be reminded and is sent by the first-stage speech tracking module, the first-stage warning module reminds through a warning device, the warning device is a wearable warning terminal, and when a speech process time point corresponding to a current speaker is located at a speech reminding time point corresponding to the current speaker, the current speaker number which needs to be reminded is extracted, and then the reminding terminal corresponding to the number is called for vibration reminding;
the second-level warning module is used for receiving the actual seat number of the person to be uttered and the estimated stop time of the current person to be uttered, which are sent by the information verification module, calling the position of the person to be uttered according to the actual seat number of the person to be uttered, calling a display corresponding to the position of the person to be uttered, and displaying and reminding the estimated stop time of the current person to be uttered and a signal to be uttered in a text reminding mode through the display, wherein the text reminding mode is that the font of the text is enlarged and the color of the text is deepened, so that the reminding of the person to be uttered is realized;
the secondary speaking tracking module is used for acquiring the actual speaking duration of the current speaking person when the current speaking person stops speaking, and sending the actual speaking duration of the current speaking person to the time judging module;
the time judging module is used for receiving the actual speaking time of the current speaking person sent by the second-stage speaking tracking module, comparing the actual speaking time of the current speaking person with the speaking time of the current speaking person corresponding to the conference flow, further acquiring a difference value between the actual speaking time of the current speaking person and the speaking time of the current speaking person corresponding to the conference flow, if the difference value is greater than zero, judging that the speaking of the speaking person is overtime, counting the overtime time of the current speaking person, if the difference value is less than zero, judging that the speaking person finishes speaking in advance, counting the advance time of the current speaking person, if the difference value between the actual speaking time of the current speaking person and the speaking time of the current speaking person corresponding to the conference flow is equal to zero, judging that the speaking person stops speaking according to the speaking time corresponding to the conference flow, and sending the overtime of the current speaking person and the advance time of the current speaking person to the time correcting module, according to the embodiment of the invention, the actual speaking duration of the current speaking person is compared with the speaking duration of the current speaking person corresponding to the conference flow through the time judgment module, so that the actual speaking duration of the current speaking person is quickly judged, and a data basis is provided for the processing of the time of the next speaking person;
the time correction module is used for correcting the speaking time point of the next speaking person when the speaking time of the current speaking person is not equal to the speaking time of the current speaking person corresponding to the conference flow, wherein the time correction module comprises an overtime correction module and an advance correction module;
according to the embodiment of the invention, through the time correction module, when the speaking duration of the current speaking person is not equal to the speaking duration of the current speaking person corresponding to the conference flow, the speaking time point of the next speaking person is corrected, and through the adjustment of the speaking time point of the next speaking person, the effect of the live telecommand is effectively improved, and the management efficiency of the live telecommand is improved;
the overtime correction module is used for receiving the overtime of the current speaker sent by the time judgment module, delaying the starting time corresponding to the next speaker backwards according to the overtime of the current speaker and the speaker time point corresponding to the next speaker corresponding to the conference flow, counting the speaker time point corresponding to the next speaker after correction, recording the speaker time point as an overtime correction speaker time point, and sending the overtime correction speaker time point corresponding to the next speaker to the speaker tracking module;
the advance correction module is used for carrying out time advance on the starting time corresponding to the next speaker according to the advance time of the current speaker and the speaker time point corresponding to the next speaker corresponding to the conference flow, counting the speaker time point corresponding to the next speaker after correction, recording the corrected speaker time point as an advance corrected speaker time point, and sending the advance corrected speaker time point corresponding to the next speaker to the speaker tracking module;
the database is used for storing speaking reminding time points corresponding to all speaking persons.
The foregoing is merely exemplary and illustrative of the principles of the present invention and various modifications, additions and substitutions of the specific embodiments described herein may be made by those skilled in the art without departing from the principles of the present invention or exceeding the scope of the claims set forth herein.

Claims (6)

1. The utility model provides a live TV teleconference intelligent management system on line based on cloud calculates and artificial intelligence which characterized in that: the system comprises a participant information acquisition module, a conference flow importing module, a conference flow processing module, a speaking tracking module, an information verification module, a time correction module, a warning module and a database;
the speech tracking module is respectively connected with the conference flow importing module, the information verification module, the database, the time judging module and the warning module, the conference flow importing module is respectively connected with the participant information acquisition module and the conference flow processing module, and the warning module is connected with the information verification module and the speech tracking module;
the participant information acquisition module is used for acquiring basic information of participants, counting the number of the participants, numbering the counted participants according to a preset sequence, sequentially marking the counted participants as 1,2, a, j, a, m, counting the number of seats in the conference room, numbering the seats in the conference room according to the preset sequence, sequentially marking the seats as 1,2, a, i, a, n, and acquiring the seat numbers corresponding to the participants according to the preset seat positions corresponding to the participants, wherein the basic information of the participants comprises names of the participants, face images of the participants and the seat numbers of the participants, and further establishing a participant information set Fw(Fw1,Fw2,...,Fwj,...Fwm),Fwj represents the w-th basic information corresponding to the jth participant, w represents the basic information, and w is a1, a2, a3, a1, a2 and a3 which are respectively represented as the participant name, the participant image and the participant seat number;
the conference flow importing module is used for importing flow information of the online live television teleconference, the flow information of the online live television teleconference comprises conference starting time and conference ending time, speaking subjects corresponding to all speaking persons, speaking time points and speaking durations corresponding to all speaking persons, and the imported conference starting time points and the imported conference ending time points, the speaking subjects corresponding to all speaking persons, the speaking time points corresponding to all speaking persons and the speaking durations are sent to the conference flow processing module;
the conference flow processing module is used for receiving the conference starting time point and the conference ending time point sent by the conference flow importing module, the speaking subjects corresponding to the speakers, the speaking time points and the speaking durations corresponding to the speakers, sequencing the received flow information of the online live video teleconference according to the speaking time points of the speakers, further acquiring the conference flow corresponding to the conference, and sending the conference flow to the speaking tracking module;
the speech tracking module is used for receiving the conference flow sent by the conference flow processing module and tracking the time of the current speech person according to the flow, and comprises a primary speech tracking module and a secondary speech tracking module;
the first-stage speaking tracking module is used for recording the initial speaking time point of the current speaking person when the current speaking person starts speaking, timing the current speaking person at the initial speaking time point, tracking the current speaking person in real time, counting the timing speaking duration of the current speaking person, acquiring the time point corresponding to the timing speaking duration according to the corresponding speaking time point and the timing speaking duration of the current speaking person in the conference flow, recording the time point as the speaking process time point, comparing the speaking process time point corresponding to the current speaking person with the speaking reminding time point corresponding to the current speaking person, if the speaking process time point corresponding to the current speaking person is at the speaking reminding time point corresponding to the current speaking person, extracting the number of the current speaking person to be reminded, sending the number of the speaking person to the warning module, and according to the speaking time point and the speaking duration of the current speaking person corresponding to the conference flow, acquiring a predicted speech stop time point of a current speaker, comparing the predicted speech stop time point of the current speaker with a speech reminding time point corresponding to the current speaker, further acquiring a difference value between the predicted speech stop time point of the current speaker and the speech reminding time point corresponding to the current speaker, recording the difference value as the predicted speech stop time of the current speaker, further extracting a name corresponding to a next speaker from a conference flow, extracting a face image and a seat number of the next speaker from a participant information set according to the name of the speaker, recording the face image as an original face image, and sending the extracted original face image, the seat number and the predicted speech stop time of the current speaker to an information verification module;
the information verification module is used for verifying the information of the next speaking person, wherein the specific verification process of the information verification module comprises the following steps:
b1, receiving an original face image, a seat number and an estimated stop duration of the current speaker of the next speaker sent by the primary speaker tracking module, and recording the next speaker as a to-be-uttered speaker;
b2, according to the seat number corresponding to the person to be announced, calling the camera of the seat to collect the face image of the person with the seat number, and recording the collected face image as an actual face image;
b3, extracting the features of the actual face image corresponding to the seat number to obtain the actual face feature image of the person waiting to speak, and extracting the features of the original face image corresponding to the seat number to obtain the original face feature image of the person waiting to speak;
b4, comparing the actual face feature image of the person to be pronounced with the original face feature image of the person to be pronounced, and further obtaining the matching degree of the actual face feature image of the person to be pronounced and the original face feature image of the person to be pronounced;
b5, if the image matching degree difference is smaller than zero, judging that the face image corresponding to the seat number does not accord with the face image input by the person to be spoken, and further starting a camera of the seat where each participant is located to acquire the face image of each participant;
b6, further carrying out face feature extraction on the collected face images of the participants to obtain face feature images corresponding to the participants;
b7, comparing the face characteristic image corresponding to each participant with the original face characteristic image of the person to be spoken, further screening out the face characteristic image with the highest matching degree with the face image input by the person to be spoken, further extracting the seat number corresponding to the face characteristic image, recording the seat number as the actual seat number of the person to be spoken, and further sending the actual seat number of the person to be spoken and the estimated stop time of the current speaker to the warning module;
the warning module comprises a primary warning module and a secondary warning module and is used for reminding the current person and the person to be called;
the first-stage warning module is used for receiving the number of the current speaker to be reminded, which is sent by the conference processing module, and further reminding the current speaker;
the second-level warning module is used for receiving the actual seat number of the person to be uttered and the estimated stop time of the current speaker sent by the information verification module, calling the position of the person to be uttered according to the actual seat number of the person to be uttered, calling a display corresponding to the position of the speaker, and displaying the estimated stop time and the estimated stop time signal corresponding to the current speaker through the display;
the secondary speaking tracking module is used for acquiring the actual speaking duration of the current speaking person when the current speaking person stops speaking, and sending the actual speaking duration of the current speaking person to the time judging module;
the time judging module is used for receiving the actual speaking duration of the current speaking person sent by the secondary speaking tracking module, and compares the actual speaking duration of the current speaking person with the speaking duration of the current speaking person corresponding to the conference flow, further obtaining the difference value between the actual speaking time length of the current speaking person and the speaking time length of the current speaking person corresponding to the conference flow, if the difference value is larger than zero, judging the speaking timeout of the speaking person, counting the timeout time of the current speaking person, if the difference is less than zero, judging that the speaker finishes speaking in advance, counting the advance time of the current speaker, if the difference between the actual speaking time of the current speaker and the speaking time of the current speaker corresponding to the conference flow is equal to zero, judging that the speaking person stops speaking according to the speaking duration corresponding to the conference flow, and sending the overtime time of the current speaking person and the advance time of the current speaking person to a time correction module;
the time correction module is used for correcting the speaking time point of the next speaking person when the speaking time of the current speaking person is not equal to the speaking time of the current speaking person corresponding to the conference flow, wherein the time correction module comprises an overtime correction module and a forward correction module;
the overtime correction module is used for receiving the overtime of the current speaker sent by the time judgment module, delaying the starting time corresponding to the next speaker backwards according to the overtime of the current speaker and the speaker time point corresponding to the next speaker corresponding to the conference flow, counting the speaker time point corresponding to the next speaker after correction, recording the speaker time point as an overtime correction speaker time point, and sending the overtime correction speaker time point corresponding to the next speaker to the speaker tracking module;
the advance correction module is used for carrying out time advance on the starting time corresponding to the next speaker according to the advance time of the current speaker and the speaker time point corresponding to the next speaker corresponding to the conference flow, counting the speaker time point corresponding to the next speaker after correction, recording the corrected speaker time point as an advance corrected speaker time point, and sending the advance corrected speaker time point corresponding to the next speaker to the speaker tracking module;
the database is used for storing speaking reminding time points corresponding to all speaking persons.
2. The intelligent management system for the online live video teleconference based on the cloud computing and the artificial intelligence, as claimed in claim 1, wherein: the speaking time tracking module tracks speaking through a timing device and a sound detection device, wherein the timing device is a timer and is used for timing speaking time when a speaking person starts speaking, and the sound detection device is a sound sensor and is used for detecting the sound of the speaking person so as to judge whether the speaking person is in the speaking process.
3. The intelligent management system for the online live video teleconference based on the cloud computing and the artificial intelligence, as claimed in claim 1, wherein: the one-level warning module reminds through a warning device, the warning device is a wearable warning terminal, when the speaking process time point corresponding to the current speaking person is at the speaking reminding time point corresponding to the current speaking person, the number of the current speaking person to be reminded is extracted, and then the warning terminal corresponding to the number is called for vibration reminding.
4. The intelligent management system for the online live video teleconference based on the cloud computing and the artificial intelligence, as claimed in claim 1, wherein: the secondary warning module reminds the current speaking person of estimating the stop duration and the signal to be spoken in a text reminding mode, wherein the text reminding mode is the increase of text fonts and the deepening of text colors, and therefore reminding of the person to be spoken is achieved.
5. The intelligent management system for the online live video teleconference based on the cloud computing and the artificial intelligence, as claimed in claim 1, wherein: the system further comprises a seat carrying device which corresponds to each participant, wherein the device comprises a plurality of cameras and a plurality of displays, the positions of the participants correspond to one camera and one display respectively, the cameras are used for carrying out image acquisition on the participants at the positions, and the displays are used for displaying video images corresponding to the current speakers and conference flows corresponding to the participants to be spoken.
6. The intelligent management system for the online live video teleconference based on the cloud computing and the artificial intelligence, as claimed in claim 1, wherein: the camera is high-definition and can automatically focus and is used for collecting high-definition face images of all participants.
CN202110052425.7A 2021-01-15 2021-01-15 Online live television teleconference intelligent management system based on cloud computing and artificial intelligence Active CN112911195B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110052425.7A CN112911195B (en) 2021-01-15 2021-01-15 Online live television teleconference intelligent management system based on cloud computing and artificial intelligence

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110052425.7A CN112911195B (en) 2021-01-15 2021-01-15 Online live television teleconference intelligent management system based on cloud computing and artificial intelligence

Publications (2)

Publication Number Publication Date
CN112911195A true CN112911195A (en) 2021-06-04
CN112911195B CN112911195B (en) 2022-08-23

Family

ID=76113296

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110052425.7A Active CN112911195B (en) 2021-01-15 2021-01-15 Online live television teleconference intelligent management system based on cloud computing and artificial intelligence

Country Status (1)

Country Link
CN (1) CN112911195B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115268820A (en) * 2022-06-22 2022-11-01 晨雨初听(武汉)文化艺术传播有限公司 Multimedia interaction analysis method for digital scene recognition and cloud platform
CN116866509A (en) * 2023-07-10 2023-10-10 深圳市创载网络科技有限公司 Conference scene picture tracking method, device and storage medium

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020120788A1 (en) * 2001-02-23 2002-08-29 Wang Douglas W. Method and system of activating computer peripherals by processed messages in real time
JP2003304337A (en) * 2002-04-08 2003-10-24 Nec Corp Speech control system in remote conference system
CN104468137A (en) * 2013-09-12 2015-03-25 华为技术有限公司 Web conference control method and device
CN106796692A (en) * 2014-09-29 2017-05-31 惠普发展公司,有限责任合伙企业 Technical support is provided a user with via wearable computing devices
CN108764807A (en) * 2018-03-29 2018-11-06 联想(北京)有限公司 A kind of information processing method and intelligent meeting system
US20190057698A1 (en) * 2015-06-01 2019-02-21 AffectLayer, Inc. In-call virtual assistant
CN110008361A (en) * 2019-03-21 2019-07-12 深圳创维-Rgb电子有限公司 Automatic identification shows electronics famous brand processing method, system and the storage medium of information
CN110401811A (en) * 2019-08-07 2019-11-01 单正建 A kind of video conferencing system towards hierarchical organization
CN110557596A (en) * 2018-06-04 2019-12-10 杭州海康威视数字技术股份有限公司 conference system
CN111091646A (en) * 2018-10-22 2020-05-01 大连艾米移动科技有限公司 Intelligent conference security check system based on face recognition
CN111861048A (en) * 2019-04-24 2020-10-30 阿里巴巴集团控股有限公司 Conference data processing method, device and system and electronic equipment

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020120788A1 (en) * 2001-02-23 2002-08-29 Wang Douglas W. Method and system of activating computer peripherals by processed messages in real time
JP2003304337A (en) * 2002-04-08 2003-10-24 Nec Corp Speech control system in remote conference system
CN104468137A (en) * 2013-09-12 2015-03-25 华为技术有限公司 Web conference control method and device
CN106796692A (en) * 2014-09-29 2017-05-31 惠普发展公司,有限责任合伙企业 Technical support is provided a user with via wearable computing devices
US20190057698A1 (en) * 2015-06-01 2019-02-21 AffectLayer, Inc. In-call virtual assistant
CN108764807A (en) * 2018-03-29 2018-11-06 联想(北京)有限公司 A kind of information processing method and intelligent meeting system
CN110557596A (en) * 2018-06-04 2019-12-10 杭州海康威视数字技术股份有限公司 conference system
CN111091646A (en) * 2018-10-22 2020-05-01 大连艾米移动科技有限公司 Intelligent conference security check system based on face recognition
CN110008361A (en) * 2019-03-21 2019-07-12 深圳创维-Rgb电子有限公司 Automatic identification shows electronics famous brand processing method, system and the storage medium of information
CN111861048A (en) * 2019-04-24 2020-10-30 阿里巴巴集团控股有限公司 Conference data processing method, device and system and electronic equipment
CN110401811A (en) * 2019-08-07 2019-11-01 单正建 A kind of video conferencing system towards hierarchical organization

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115268820A (en) * 2022-06-22 2022-11-01 晨雨初听(武汉)文化艺术传播有限公司 Multimedia interaction analysis method for digital scene recognition and cloud platform
CN115268820B (en) * 2022-06-22 2023-10-24 广东三鼎智慧信息科技有限公司 Digital scene recognition multimedia interaction analysis method and cloud platform
CN116866509A (en) * 2023-07-10 2023-10-10 深圳市创载网络科技有限公司 Conference scene picture tracking method, device and storage medium
CN116866509B (en) * 2023-07-10 2024-02-23 深圳市创载网络科技有限公司 Conference scene picture tracking method, device and storage medium

Also Published As

Publication number Publication date
CN112911195B (en) 2022-08-23

Similar Documents

Publication Publication Date Title
CN112911195B (en) Online live television teleconference intelligent management system based on cloud computing and artificial intelligence
CN110991381B (en) Real-time classroom student status analysis and indication reminding system and method based on behavior and voice intelligent recognition
US20090123035A1 (en) Automated Video Presence Detection
CN112801608A (en) Remote video conference intelligent management system based on big data and cloud computing and cloud conference management platform
CN112037788B (en) Voice correction fusion method
CN110033403A (en) A kind of unaware check class attendance method and system
CN105205646A (en) Automatic roll call system and realization method thereof
CN112183238B (en) Remote education attention detection method and system
CN110648667B (en) Multi-person scene human voice matching method
CN108229441A (en) A kind of classroom instruction automatic feedback system and feedback method based on image and speech analysis
CN108109446B (en) Teaching class condition monitoring system
CN108109445B (en) Teaching course condition monitoring method
CN113052085A (en) Video clipping method, video clipping device, electronic equipment and storage medium
CN110827432B (en) Class attendance checking method and system based on face recognition
CN112395461A (en) Business meeting intelligent management system based on big data analysis
CN115052126B (en) Ultra-high definition video conference analysis management system based on artificial intelligence
CN109858809A (en) Learning quality appraisal procedure and system based on the analysis of classroom students ' behavior
CN109560941A (en) Minutes method, apparatus, intelligent terminal and storage medium
WO2022151639A1 (en) Method and apparatus for extracting picture to be identified, and device and storage medium
CN112200088A (en) Sitting posture monitoring method, device, equipment and system
CN104135638A (en) Optimized video snapshot
JP7204337B2 (en) CONFERENCE SUPPORT DEVICE, CONFERENCE SUPPORT SYSTEM, CONFERENCE SUPPORT METHOD AND PROGRAM
CN111915755A (en) Rapid classroom attendance checking method and system based on artificial intelligence image prediction
CN114971975B (en) Learning abnormity prompting method and system for online education platform
CN115294630A (en) Conference system and conference management method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20220802

Address after: 200000 room 2201-1, building 1, No. 2167, Zhenbei Road, Putuo District, Shanghai

Applicant after: Weimaikejian Group Co.,Ltd.

Address before: 430071 No.2 Bayi Road, Wuchang District, Wuhan City, Hubei Province

Applicant before: Yin Xiaodong

GR01 Patent grant
GR01 Patent grant
CP01 Change in the name or title of a patent holder
CP01 Change in the name or title of a patent holder

Address after: 200000 room 2201-1, building 1, No. 2167, Zhenbei Road, Putuo District, Shanghai

Patentee after: Weimai Technology Co.,Ltd.

Address before: 200000 room 2201-1, building 1, No. 2167, Zhenbei Road, Putuo District, Shanghai

Patentee before: Weimaikejian Group Co.,Ltd.