CN112911195B - Online live television teleconference intelligent management system based on cloud computing and artificial intelligence - Google Patents

Online live television teleconference intelligent management system based on cloud computing and artificial intelligence Download PDF

Info

Publication number
CN112911195B
CN112911195B CN202110052425.7A CN202110052425A CN112911195B CN 112911195 B CN112911195 B CN 112911195B CN 202110052425 A CN202110052425 A CN 202110052425A CN 112911195 B CN112911195 B CN 112911195B
Authority
CN
China
Prior art keywords
speaking
person
speaker
module
current
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110052425.7A
Other languages
Chinese (zh)
Other versions
CN112911195A (en
Inventor
尹晓东
赵洋梵
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Weimai Technology Co.,Ltd.
Original Assignee
Weimaikejian Group Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Weimaikejian Group Co ltd filed Critical Weimaikejian Group Co ltd
Priority to CN202110052425.7A priority Critical patent/CN112911195B/en
Publication of CN112911195A publication Critical patent/CN112911195A/en
Application granted granted Critical
Publication of CN112911195B publication Critical patent/CN112911195B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems
    • H04N7/152Multipoint control units therefor
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems
    • H04N7/155Conference systems involving storage of or access to video conference sessions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems
    • H04N7/157Conference systems defining a virtual conference space and using avatars or agents
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The invention discloses an intelligent management system for online live video teleconference based on cloud computing and artificial intelligence, the system comprises a participant information acquisition module, a conference flow importing module, a conference flow processing module, an information verification module, a time correction module, a speech tracking module, a warning module, a conference processing module and a database, wherein the conference processing module and the time judgment module are connected through the speech tracking module, the speaking time of each speaking person is tracked, the speaking time point of each speaking person is obtained, and then each speaking person is reminded in time, the intelligent management of the on-line live television conference is realized, the stability of the conference flow is ensured, meanwhile, the next speaker can be adjusted according to the current speaker's speaking time, so that the effect of the live video teleconference is effectively improved, and the management efficiency of the live video teleconference is improved.

Description

Online live television teleconference intelligent management system based on cloud computing and artificial intelligence
Technical Field
The invention belongs to the technical field of conference management, and relates to an intelligent management system for an online live video teleconference based on cloud computing and artificial intelligence.
Background
Under the promotion of big data development strategy, people can acquire information more conveniently, and the national government actively promotes the information disclosure of a plurality of public fields all the time, so that an online live television conference is generated, real-time information of the conference can be shared in a live broadcast mode, but later adjustment cannot be performed, and the online live television conference management is particularly important.
In traditional online live telecommand conference management mode, manage network and the equipment operation state that corresponds in this meeting room, therefore traditional online live telecommand conference management mode has still existed many drawbacks of violence, on the one hand can't trace each speech person's speech time, can't accomplish timely warning to each speech person, on the one hand can't adjust the speech time of next speech person according to current speech person's speech time, and then can't improve the effect of this live telecommand conference, on the other hand lacks the management to the meeting flow, can't the managerial efficiency of effectual guarantee online live telecommand conference.
Disclosure of Invention
In view of this, in order to solve the problems existing in the background art, an intelligent management system for an online live tv teleconference based on cloud computing and artificial intelligence is proposed, so as to realize intelligent management of the online live tv teleconference.
The purpose of the invention can be realized by the following technical scheme:
an online live television teleconference intelligent management system based on cloud computing and artificial intelligence comprises a participant information acquisition module, a conference flow importing module, a conference flow processing module, a speech tracking module, an information verification module, a time correction module, a warning module and a database;
the conference flow importing module is respectively connected with the participant information acquiring module and the conference flow processing module, and the warning module is connected with the information verifying module and the speech tracking module;
the participant information acquisition module is used for acquiring basic information of participants, counting the number of the participants, numbering the counted participants according to a preset sequence, sequentially marking the counted participants as 1,2, a, j, a, m, counting the number of seats in the conference room, numbering the seats in the conference room according to the preset sequence, sequentially marking the seats as 1,2, a, i, a, n, and acquiring the seat numbers corresponding to the participants according to the preset seat positions corresponding to the participants, wherein the basic information of the participants comprises names of the participants, face images of the participants and the seat numbers of the participants, and further establishing a participant information set F w (F w 1,F w 2,...,F w j,...F w m),F w j represents the w-th basic information corresponding to the jth participant, w represents the basic information, and w is a1, a2, a3, a1, a2 and a3 which are respectively represented as the participant name, the participant image and the participant seat number;
the conference flow importing module is used for importing flow information of the online live television teleconference, the flow information of the online live television teleconference comprises conference starting time and conference ending time, speaking subjects corresponding to all speaking persons, speaking time points and speaking durations corresponding to all speaking persons, and the imported conference starting time points and the imported conference ending time points, the speaking subjects corresponding to all speaking persons, the speaking time points corresponding to all speaking persons and the speaking durations are sent to the conference flow processing module;
the conference flow processing module is used for receiving the conference starting time point and the conference ending time point sent by the conference flow importing module, the speaking subjects corresponding to the speakers, the speaking time points and the speaking durations corresponding to the speakers, sequencing the received flow information of the online live video teleconference according to the speaking time points of the speakers, further acquiring the conference flow corresponding to the conference, and sending the conference flow to the speaking tracking module;
the speech tracking module is used for receiving the conference flow sent by the conference flow processing module and tracking the time of the current speech person according to the flow, and comprises a primary speech tracking module and a secondary speech tracking module;
the first-stage speaking tracking module is used for recording the initial speaking time point of the current speaking person when the current speaking person starts speaking, timing the current speaking person at the initial speaking time point, tracking the current speaking person in real time, counting the timing speaking duration of the current speaking person, acquiring the time point corresponding to the timing speaking duration according to the corresponding speaking time point and the timing speaking duration of the current speaking person in the conference flow, recording the time point as the speaking process time point, comparing the speaking process time point corresponding to the current speaking person with the speaking reminding time point corresponding to the current speaking person, if the speaking process time point corresponding to the current speaking person is at the speaking reminding time point corresponding to the current speaking person, extracting the number of the current speaking person to be reminded, sending the number of the speaking person to the warning module, and according to the speaking time point and the speaking duration of the current speaking person corresponding to the conference flow, acquiring a predicted speech stop time point of a current speaker, comparing the predicted speech stop time point of the current speaker with a speech reminding time point corresponding to the current speaker, further acquiring a difference value between the predicted speech stop time point of the current speaker and the speech reminding time point corresponding to the current speaker, recording the difference value as the predicted speech stop time of the current speaker, further extracting a name corresponding to a next speaker from a conference flow, extracting a face image and a seat number of the next speaker from a participant information set according to the name of the speaker, recording the face image as an original face image, and sending the extracted original face image, the seat number and the predicted speech stop time of the current speaker to an information verification module;
the information verification module is used for verifying the information of the next speaking person, wherein the specific verification process of the information verification module comprises the following steps:
b1, receiving an original face image, a seat number and an estimated stop duration of the current speaker of the next speaker sent by the primary speaker tracking module, and recording the next speaker as a to-be-uttered speaker;
b2, according to the seat number corresponding to the person to be announced, calling the camera of the seat to collect the face image of the person with the seat number, and recording the collected face image as an actual face image;
b3, extracting the features of the actual face image corresponding to the seat number to obtain the actual face feature image of the person waiting to speak, and extracting the features of the original face image corresponding to the seat number to obtain the original face feature image of the person waiting to speak;
b4, comparing the actual face feature image of the person to be pronounced with the original face feature image of the person to be pronounced, and further obtaining the matching degree of the actual face feature image of the person to be pronounced and the original face feature image of the person to be pronounced;
b5, if the image matching degree difference is smaller than zero, judging that the face image corresponding to the seat number does not accord with the face image input by the person to be announced, and starting a camera of the seat where each participant is located to acquire the face image of each participant;
b6, further carrying out face feature extraction on the collected face images of the participants to obtain face feature images corresponding to the participants;
b7, comparing the face characteristic image corresponding to each participant with the original face characteristic image of the person to be spoken, further screening out the face characteristic image with the highest matching degree with the face image input by the person to be spoken, further extracting the seat number corresponding to the face characteristic image, recording the seat number as the actual seat number of the person to be spoken, and further sending the actual seat number of the person to be spoken and the estimated stop time of the current speaker to the warning module;
the warning module comprises a primary warning module and a secondary warning module and is used for reminding a current person and a person to be uttered;
the first-stage warning module is used for receiving the number of the current speaker to be reminded, which is sent by the conference processing module, and further reminding the current speaker;
the second-level warning module is used for receiving the actual seat number of the person to be uttered and the estimated stop time of the current speaker sent by the information verification module, calling the position of the person to be uttered according to the actual seat number of the person to be uttered, calling a display corresponding to the position of the speaker, and displaying the estimated stop time and the estimated signal to be uttered corresponding to the current speaker through the display;
the secondary speaking tracking module is used for acquiring the actual speaking duration of the current speaking person when the current speaking person stops speaking, and sending the actual speaking duration of the current speaking person to the time judging module;
the time judging module is used for receiving the actual speaking duration of the current speaking person sent by the second-level speaking tracking module, and compares the actual speaking duration of the current speaking person with the speaking duration of the current speaking person corresponding to the conference flow, and further obtaining the difference value between the actual speaking time length of the current speaking person and the speaking time length of the current speaking person corresponding to the conference flow, if the difference value is greater than zero, judging the speaking timeout of the speaking person, counting the timeout time of the current speaking person, if the difference is less than zero, judging that the speaker finishes speaking in advance, counting the advance time of the current speaker, if the difference between the actual speaking time of the current speaker and the speaking time of the current speaker corresponding to the conference flow is equal to zero, judging that the speaking person stops speaking according to the speaking duration corresponding to the conference flow, and sending the overtime time of the current speaking person and the advance time of the current speaking person to a time correction module;
the time correction module is used for correcting the speaking time point of the next speaking person when the speaking time of the current speaking person is not equal to the speaking time of the current speaking person corresponding to the conference flow, wherein the time correction module comprises an overtime correction module and a forward correction module;
the overtime correction module is used for receiving the overtime of the current speaker sent by the time judgment module, delaying the starting time corresponding to the next speaker backwards according to the overtime of the current speaker and the speaker time point corresponding to the next speaker corresponding to the conference flow, counting the speaker time point corresponding to the next speaker after correction, recording the speaker time point as an overtime correction speaker time point, and sending the overtime correction speaker time point corresponding to the next speaker to the speaker tracking module;
the advance correction module is used for carrying out time advance on the starting time corresponding to the next speaker according to the advance time of the current speaker and the speaker time point corresponding to the next speaker corresponding to the conference flow, counting the speaker time point corresponding to the next speaker after correction, recording the corrected speaker time point as an advance corrected speaker time point, and sending the advance corrected speaker time point corresponding to the next speaker to the speaker tracking module;
the database is used for storing speaking reminding time points corresponding to all speaking persons.
Furthermore, the speaking time tracking module tracks speaking through a timing device and a sound detection device, wherein the timing device is a timer and is used for timing the speaking duration when the speaking person starts speaking, and the sound detection device is a sound sensor and is used for detecting the sound of the speaking person so as to judge whether the speaking person is in the speaking process.
Furthermore, the primary warning module reminds through a warning device, the warning device is a wearable warning terminal, when the speaking process time point corresponding to the current speaking person is located at the speaking reminding time point corresponding to the current speaking person, the current speaking person number needing to be reminded is extracted, and then the warning terminal corresponding to the number is called for vibration reminding.
Furthermore, the secondary level warning module reminds the current speaking person of estimating the stop duration and the signal to be spoken in a text reminding mode, wherein the text reminding mode is the increase of text fonts and the deepening of text colors, and therefore reminding of the person to be spoken is achieved.
Further, the system also comprises a device for installing the corresponding seats of the participants, wherein the device comprises a plurality of cameras and a plurality of displays, the seats of the participants correspond to one camera and one display respectively, the cameras are used for collecting images of the participants at the seats, and the displays are used for displaying video images corresponding to the current speakers and conference flows corresponding to the participants to be spoken.
Furthermore, the camera is high-definition and can automatically focus, and is used for collecting high-definition face images of each participant.
The invention has the beneficial effects that:
(1) according to the intelligent management system for the online live television teleconference based on the cloud computing and the artificial intelligence, the current speakers are tracked in real time through the speech tracking module and the warning module, the initial speech time point and the speech duration of each speaker are obtained, and then each speaker is reminded in time, so that the intelligent management of the online live television teleconference is realized, the stability of the conference flow is guaranteed, meanwhile, the next speaker can be adjusted according to the speech time of the current speaker, the effect of the live television teleconference is effectively improved, and the management efficiency of the live television teleconference is improved.
(2) The invention quickly adjusts the speaking time of the next speaking person by correcting the speaking time point of the next speaking person in the time correction module, thereby effectively promoting the process of the live video teleconference and effectively maintaining the order of the live video teleconference.
(3) According to the invention, the warning module reminds the current speaker and the speakers to be announced, so that the current speaker and the speakers to be announced can be reminded in time, and the smoothness of the progress of the video teleconference can be effectively guaranteed.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the description below are only some embodiments of the present invention, and it is obvious for those skilled in the art that other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a schematic diagram of the system module connections of the present invention;
fig. 2 is a schematic diagram of the connection of the utterance tracking module in the present invention;
FIG. 3 is a schematic diagram of an alarm module according to the present invention.
Detailed Description
While the foregoing is directed to embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow.
Referring to fig. 1, an online live tv conference intelligent management system based on cloud computing and artificial intelligence includes a participant information acquisition module, a conference flow import module, a conference flow processing module, a speech tracking module, an information verification module, a time correction module, a warning module, and a database;
the speech tracking module is respectively connected with the conference flow importing module, the information verification module, the database, the time judging module and the warning module, the conference flow importing module is respectively connected with the participant information acquisition module and the conference flow processing module, and the warning module is connected with the information verification module and the speech tracking module;
the participant information acquisition module is used for acquiring basic information of participants, counting the number of the participants, numbering the counted participants according to a preset sequence, sequentially marking the counted participants as 1,2, a, j, a, m, counting the number of seats in the conference room, numbering the seats in the conference room according to the preset sequence, sequentially marking the seats as 1,2, a, i, a, n, and acquiring the seat numbers corresponding to the participants according to the preset seat positions corresponding to the participants, wherein the basic information of the participants comprises names of the participants, face images of the participants and the seat numbers of the participants, and further establishing a participant information set F w (F w 1,F w 2,...,F w j,...F w m),F w j represents the w-th basic information corresponding to the jth participant, w represents the basic information, w is a1, a2, a3, a1, a2 and a3 respectively represent names of the participants, images of the participants and seat numbers of the participants, wherein each seat corresponds to a camera and a display, the camera is used for collecting images of the participants at the position, and the display is used for displaying video images corresponding to the current speaking personnel and conference processes corresponding to the participants to be spoken;
the conference flow importing module is used for importing flow information of the online live television teleconference, the flow information of the online live television teleconference comprises conference starting time and conference ending time, speaking subjects corresponding to all speaking persons, speaking time points and speaking durations corresponding to all speaking persons, and the imported conference starting time points and the imported conference ending time points, the speaking subjects corresponding to all speaking persons, the speaking time points corresponding to all speaking persons and the speaking durations are sent to the conference flow processing module;
the conference flow processing module is used for receiving the conference starting time point and the conference ending time point sent by the conference flow importing module, the speaking subjects corresponding to the speaking persons, the speaking time points and the speaking durations corresponding to the speaking persons, and sequencing the received flow information of the online live television conference according to the speaking time points of the speaking persons, so that the conference flow corresponding to the conference is obtained, and the conference flow is sent to the speaking tracking module.
Referring to fig. 2, the utterance tracking module is configured to receive a conference flow sent by the conference flow processing module, where the utterance tracking module includes a primary utterance tracking module and a secondary utterance tracking module;
according to the embodiment of the invention, the current speakers are tracked in real time through the speaker tracking module, and the initial speaking time point and the speaking duration of each speaker are obtained, so that each speaker can be reminded in time, the intelligent management of the on-line live television conference and telephone conference is realized, and the stability of the conference process is ensured;
the first-stage speaking tracking module tracks the time of the current speaking person according to the flow through a timing device and a sound detection device, wherein the timing device is a timer and is used for timing the speaking time when the speaking person starts speaking, the sound detection device is a sound sensor and is used for detecting the sound of the speaking person so as to judge whether the current speaking person is in the speaking process, when the sound sensor detects the sound, the initial speaking time point of the current speaking person is recorded and timed, the speaking time of the current speaking person is counted, the initial speaking time point of the current speaking person is recorded, the current speaking person is tracked in real time, the timed speaking time of the current speaking person is counted, and the time point corresponding to the timed speaking time is obtained according to the speaking time point and the timed speaking time corresponding to the current speaking person in the conference flow, recording the current speech process time point as a speech process time point, comparing the speech process time point corresponding to the current speaker with a speech reminding time point corresponding to the current speaker, if the speech process time point corresponding to the current speaker is at the speech reminding time point corresponding to the current speaker, extracting the number of the current speaker to be reminded, sending the number of the speaker to a warning module, acquiring the estimated speech stop time point of the current speaker according to the speech time point and the speech duration of the current speaker corresponding to the conference flow, comparing the estimated speech stop time point of the current speaker with the speech reminding time point corresponding to the current speaker, further acquiring the difference value between the estimated speech stop time point of the current speaker and the speech reminding time point corresponding to the current speaker, and recording the difference value as the estimated speech stop time of the current speaker, further extracting a name corresponding to a next speaker from the conference flow, extracting a face image and a seat number of the next speaker from the participant information set according to the name of the speaker, marking the face image as an original face image, and sending the extracted original face image, the seat number and the estimated stop time of the current speaker of the next speaker to the information verification module;
the information verification module verifies information of the next speaking person through a camera, the camera is high-definition and can automatically focus, and the information verification module is used for collecting high-definition face images of all participants, wherein the specific verification process of the information verification module comprises the following steps:
b1, receiving an original face image, a seat number and a predicted stop duration of the current speaker of the next speaker sent by the primary speaker tracking module, and recording the next speaker as a to-be-uttered person;
b2, according to the seat number corresponding to the person to be announced, calling the camera of the seat to collect the face image of the person with the seat number, and recording the collected face image as an actual face image;
b3, extracting the features of the actual face image corresponding to the seat number to obtain the actual face feature image of the person waiting to speak, and extracting the features of the original face image corresponding to the seat number to obtain the original face feature image of the person waiting to speak;
b4, comparing the actual face feature image of the person to be pronounced with the original face feature image of the person to be pronounced, and further obtaining the matching degree of the actual face feature image of the person to be pronounced and the original face feature image of the person to be pronounced;
b5, if the image matching degree difference is smaller than zero, judging that the face image corresponding to the seat number does not accord with the face image input by the person to be announced, and starting a camera of the seat where each participant is located to acquire the face image of each participant;
b6, further carrying out face feature extraction on the collected face images of the participants to obtain face feature images corresponding to the participants;
b7, comparing the facial feature image corresponding to each participant with the original facial feature image of the person to be announced, further screening out the facial feature image with the highest matching degree with the facial image input by the person to be announced, further extracting the seat number corresponding to the facial feature image, marking the seat number as the actual seat number of the person to be announced, and further sending the actual seat number of the person to be announced and the estimated stop duration of the current speaker to the warning module;
according to the embodiment of the invention, the information verification module verifies the information of the next speaker through the camera, so that the warning accuracy of the warning module is effectively guaranteed, and the stability of the live video teleconference is effectively maintained;
referring to fig. 3, the warning module includes a primary warning module and a secondary warning module, and is configured to remind a current person and a person to be announced;
according to the embodiment of the invention, the warning module reminds the current speaker and the speakers to be announced, so that the current speaker and the speakers to be announced can be reminded in time, and the smoothness of the progress of the television teleconference can be effectively guaranteed.
The first-stage warning module is used for receiving a current speaker number which needs to be reminded and is sent by the first-stage speech tracking module, the first-stage warning module reminds through a warning device, the warning device is a wearable warning terminal, and when a speech process time point corresponding to a current speaker is located at a speech reminding time point corresponding to the current speaker, the current speaker number which needs to be reminded is extracted, and then the reminding terminal corresponding to the number is called for vibration reminding;
the second-level warning module is used for receiving the actual seat number of the person to be uttered and the estimated stop time of the current person to be uttered, which are sent by the information verification module, calling the position of the person to be uttered according to the actual seat number of the person to be uttered, calling a display corresponding to the position of the person to be uttered, and displaying and reminding the estimated stop time of the current person to be uttered and a signal to be uttered in a text reminding mode through the display, wherein the text reminding mode is that the font of the text is enlarged and the color of the text is deepened, so that the reminding of the person to be uttered is realized;
the secondary speaking tracking module is used for acquiring the actual speaking duration of the current speaking person when the current speaking person stops speaking, and sending the actual speaking duration of the current speaking person to the time judging module;
the time judging module is used for receiving the actual speaking time of the current speaking person sent by the second-stage speaking tracking module, comparing the actual speaking time of the current speaking person with the speaking time of the current speaking person corresponding to the conference flow, further acquiring a difference value between the actual speaking time of the current speaking person and the speaking time of the current speaking person corresponding to the conference flow, if the difference value is greater than zero, judging that the speaking of the speaking person is overtime, counting the overtime time of the current speaking person, if the difference value is less than zero, judging that the speaking person finishes speaking in advance, counting the advance time of the current speaking person, if the difference value between the actual speaking time of the current speaking person and the speaking time of the current speaking person corresponding to the conference flow is equal to zero, judging that the speaking person stops speaking according to the speaking time corresponding to the conference flow, and sending the overtime of the current speaking person and the advance time of the current speaking person to the time correcting module, according to the embodiment of the invention, the actual speaking duration of the current speaking person is compared with the speaking duration of the current speaking person corresponding to the conference flow through the time judgment module, so that the actual speaking duration of the current speaking person is quickly judged, and a data basis is provided for the processing of the time of the next speaking person;
the time correction module is used for correcting the speaking time point of the next speaking person when the speaking time of the current speaking person is not equal to the speaking time of the current speaking person corresponding to the conference flow, wherein the time correction module comprises an overtime correction module and a correction module in advance;
according to the embodiment of the invention, through the time correction module, when the speaking duration of the current speaking person is not equal to the speaking duration of the current speaking person corresponding to the conference flow, the speaking time point of the next speaking person is corrected, and through the adjustment of the speaking time point of the next speaking person, the effect of the live telecommand is effectively improved, and the management efficiency of the live telecommand is improved;
the overtime correction module is used for receiving the overtime of the current speaker sent by the time judgment module, delaying the starting time corresponding to the next speaker backwards according to the overtime of the current speaker and the speaker time point corresponding to the next speaker corresponding to the conference flow, counting the speaker time point corresponding to the next speaker after correction, recording the speaker time point as an overtime correction speaker time point, and sending the overtime correction speaker time point corresponding to the next speaker to the speaker tracking module;
the advance correction module is used for carrying out time advance on the starting time corresponding to the next speaker according to the advance time of the current speaker and the speaker time point corresponding to the next speaker corresponding to the conference flow, counting the speaker time point corresponding to the next speaker after correction, recording the corrected speaker time point as an advance corrected speaker time point, and sending the advance corrected speaker time point corresponding to the next speaker to the speaker tracking module;
the database is used for storing speaking reminding time points corresponding to all speaking persons.
The foregoing is merely exemplary and illustrative of the principles of the present invention and various modifications, additions and substitutions of the specific embodiments described herein may be made by those skilled in the art without departing from the principles of the present invention or exceeding the scope of the claims set forth herein.

Claims (4)

1. The utility model provides a live TV teleconference intelligent management system on line based on cloud calculates and artificial intelligence which characterized in that: the system comprises a participant information acquisition module, a conference flow importing module, a conference flow processing module, a speaking tracking module, an information verification module, a time correction module, a warning module and a database;
the speech tracking module is respectively connected with the conference flow importing module, the information verification module, the database, the time judging module and the warning module, the conference flow importing module is respectively connected with the participant information acquisition module and the conference flow processing module, and the warning module is connected with the information verification module and the speech tracking module;
the participant information acquisition module is used for acquiring basic information of participants, counting the number of the participants, numbering the counted participants according to a preset sequence, sequentially marking the numbered participants as 1,2, j, m, counting the number of seats in a conference room, and carrying out entry to the seats in the conference room according to the preset sequenceThe lines are numbered and are sequentially marked as 1,2, 1, i, n, and then the seat numbers corresponding to all the participants are obtained according to the preset seat positions corresponding to all the participants, wherein the basic information of all the participants comprises names of all the participants, face images of all the participants and seat numbers of all the participants, and then a participant information set F is constructed w (F w 1,F w 2,...,F w j,...F w m),F w j represents the w-th basic information corresponding to the jth participant, w represents the basic information, and w is a1, a2, a3, a1, a2 and a3 which are respectively represented as the participant name, the participant image and the participant seat number;
the conference flow importing module is used for importing flow information of the online live television teleconference, the flow information of the online live television teleconference comprises conference starting time and conference ending time, speaking subjects corresponding to all speaking persons, speaking time points and speaking durations corresponding to all speaking persons, and the imported conference starting time points and the imported conference ending time points, the speaking subjects corresponding to all speaking persons, the speaking time points corresponding to all speaking persons and the speaking durations are sent to the conference flow processing module;
the conference flow processing module is used for receiving the conference starting time point and the conference ending time point sent by the conference flow importing module, the speaking subjects corresponding to the speakers, the speaking time points and the speaking durations corresponding to the speakers, sequencing the received flow information of the online live video teleconference according to the speaking time points of the speakers, further acquiring the conference flow corresponding to the conference, and sending the conference flow to the speaking tracking module;
the speech tracking module is used for receiving the conference flow sent by the conference flow processing module and tracking the time of the current speech person according to the flow, and comprises a primary speech tracking module and a secondary speech tracking module;
the speech tracking module tracks speech through a timing device and a sound detection device, wherein the timing device is a timer and is used for timing the speech duration when a speaker starts to speak, and the sound detection device is a sound sensor and is used for detecting the sound of the speaker so as to judge whether the speaker is in the speech process;
the first-stage speaking tracking module is used for recording the initial speaking time point of the current speaking person when the current speaking person starts speaking, timing the current speaking person at the initial speaking time point, tracking the current speaking person in real time, counting the timing speaking time of the current speaking person, acquiring the time point corresponding to the timing speaking time according to the corresponding speaking time point and the timing speaking time of the current speaking person in the conference flow, recording the time point as the speaking process time point, comparing the speaking process time point corresponding to the current speaking person with the reminding time point corresponding to the current speaking person, if the speaking process time point corresponding to the current speaking person is at the speaking reminding time point corresponding to the current speaking person, extracting the number of the current speaking person to be reminded, sending the number of the speaking person to the warning module, and according to the speaking time point and the speaking time length of the current speaking person corresponding to the conference flow, acquiring a predicted speech stop time point of a current speaker, comparing the predicted speech stop time point of the current speaker with a speech reminding time point corresponding to the current speaker, further acquiring a difference value between the predicted speech stop time point of the current speaker and the speech reminding time point corresponding to the current speaker, recording the difference value as the predicted speech stop time of the current speaker, further extracting a name corresponding to a next speaker from a conference flow, extracting a face image and a seat number of the next speaker from an attendee information set according to the name of the speaker, recording the face image as an original face image, and sending the extracted original face image, the seat number and the predicted speech stop time of the current speaker to an information verification module;
the information verification module is used for verifying the information of the next speaking person, wherein the specific verification process of the information verification module comprises the following steps:
b1, receiving an original face image, a seat number and an estimated stop duration of the current speaker of the next speaker sent by the primary speaker tracking module, and recording the next speaker as a to-be-uttered speaker;
b2, according to the seat number corresponding to the person to be announced, the camera of the seat is further called to collect the face image of the person with the seat number, and the collected face image is recorded as an actual face image;
b3, extracting the features of the actual face image corresponding to the seat number to obtain the actual face feature image of the person waiting to speak, and extracting the features of the original face image corresponding to the seat number to obtain the original face feature image of the person waiting to speak;
b4, comparing the actual face feature image of the person to be pronounced with the original face feature image of the person to be pronounced, and further obtaining the matching degree of the actual face feature image of the person to be pronounced and the original face feature image of the person to be pronounced;
b5, if the image matching degree difference is smaller than zero, judging that the face image corresponding to the seat number does not accord with the face image input by the person to be announced, and starting a camera of the seat where each participant is located to acquire the face image of each participant;
b6, further carrying out face feature extraction on the collected face images of the participants to obtain face feature images corresponding to the participants;
b7, comparing the facial feature image corresponding to each participant with the original facial feature image of the person to be announced, further screening out the facial feature image with the highest matching degree with the facial image input by the person to be announced, further extracting the seat number corresponding to the facial feature image, marking the seat number as the actual seat number of the person to be announced, and further sending the actual seat number of the person to be announced and the estimated stop duration of the current speaker to the warning module;
the warning module comprises a primary warning module and a secondary warning module and is used for reminding the current person and the person to be called;
the first-level warning module is used for receiving the number of the current speaker needing to be reminded and sent by the conference processing module so as to remind the current speaker;
the first-stage warning module reminds through a warning device, the warning device is a wearable warning terminal, and when the speaking process time point corresponding to the current speaking person is at the speaking reminding time point corresponding to the current speaking person, the number of the current speaking person to be reminded is extracted, and then the warning terminal corresponding to the number is called for vibration reminding;
the second-level warning module is used for receiving the actual seat number of the person to be uttered and the estimated stop time of the current speaker sent by the information verification module, calling the position of the person to be uttered according to the actual seat number of the person to be uttered, calling a display corresponding to the position of the speaker, and displaying the estimated stop time and the estimated stop time signal corresponding to the current speaker through the display;
the secondary speaking tracking module is used for acquiring the actual speaking duration of the current speaking person when the current speaking person stops speaking, and sending the actual speaking duration of the current speaking person to the time judging module;
the time judging module is used for receiving the actual speaking duration of the current speaking person sent by the secondary speaking tracking module, and compares the actual speaking duration of the current speaking person with the speaking duration of the current speaking person corresponding to the conference flow, and further obtaining the difference value between the actual speaking time length of the current speaking person and the speaking time length of the current speaking person corresponding to the conference flow, if the difference value is greater than zero, judging the speaking timeout of the speaking person, counting the timeout time of the current speaking person, if the difference is less than zero, judging that the speaker finishes speaking in advance, counting the advance time of the current speaker, if the difference between the actual speaking time of the current speaker and the speaking time of the current speaker corresponding to the conference flow is equal to zero, judging that the speaking person stops speaking according to the speaking duration corresponding to the conference flow, and sending the overtime time of the current speaking person and the advance time of the current speaking person to a time correction module;
the time correction module is used for correcting the speaking time point of the next speaking person when the speaking time of the current speaking person is not equal to the speaking time of the current speaking person corresponding to the conference flow, wherein the time correction module comprises an overtime correction module and a forward correction module;
the overtime correction module is used for receiving the overtime of the current speaker sent by the time judgment module, delaying the starting time corresponding to the next speaker backwards according to the overtime of the current speaker and the speaker time point corresponding to the next speaker corresponding to the conference flow, counting the speaker time point corresponding to the next speaker after correction, recording the speaker time point as an overtime correction speaker time point, and sending the overtime correction speaker time point corresponding to the next speaker to the speaker tracking module;
the advance correction module is used for carrying out time advance on the starting time corresponding to the next speaker according to the advance time of the current speaker and the speaker time point corresponding to the next speaker corresponding to the conference flow, counting the speaker time point corresponding to the next speaker after correction, recording the corrected speaker time point as an advance corrected speaker time point, and sending the advance corrected speaker time point corresponding to the next speaker to the speaker tracking module;
the database is used for storing speaking reminding time points corresponding to all speaking persons.
2. The intelligent management system for the online live video teleconference based on the cloud computing and the artificial intelligence, as claimed in claim 1, wherein: the secondary warning module reminds the current speaking person of estimating the stop duration and the signal to be spoken in a text reminding mode, wherein the text reminding mode is the increase of text fonts and the deepening of text colors, and therefore reminding of the person to be spoken is achieved.
3. The intelligent management system for the online live video teleconference based on the cloud computing and the artificial intelligence, as claimed in claim 1, wherein: the system further comprises a seat carrying device which corresponds to each participant, wherein the device comprises a plurality of cameras and a plurality of displays, the positions of the participants correspond to one camera and one display respectively, the cameras are used for carrying out image acquisition on the participants at the positions, and the displays are used for displaying video images corresponding to the current speakers and conference flows corresponding to the participants to be spoken.
4. The intelligent management system for the online live video teleconference based on the cloud computing and the artificial intelligence, as claimed in claim 1, wherein: the camera is high-definition and can automatically focus and is used for collecting high-definition face images of all participants.
CN202110052425.7A 2021-01-15 2021-01-15 Online live television teleconference intelligent management system based on cloud computing and artificial intelligence Active CN112911195B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110052425.7A CN112911195B (en) 2021-01-15 2021-01-15 Online live television teleconference intelligent management system based on cloud computing and artificial intelligence

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110052425.7A CN112911195B (en) 2021-01-15 2021-01-15 Online live television teleconference intelligent management system based on cloud computing and artificial intelligence

Publications (2)

Publication Number Publication Date
CN112911195A CN112911195A (en) 2021-06-04
CN112911195B true CN112911195B (en) 2022-08-23

Family

ID=76113296

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110052425.7A Active CN112911195B (en) 2021-01-15 2021-01-15 Online live television teleconference intelligent management system based on cloud computing and artificial intelligence

Country Status (1)

Country Link
CN (1) CN112911195B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115268820B (en) * 2022-06-22 2023-10-24 广东三鼎智慧信息科技有限公司 Digital scene recognition multimedia interaction analysis method and cloud platform
CN116866509B (en) * 2023-07-10 2024-02-23 深圳市创载网络科技有限公司 Conference scene picture tracking method, device and storage medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003304337A (en) * 2002-04-08 2003-10-24 Nec Corp Speech control system in remote conference system
CN104468137A (en) * 2013-09-12 2015-03-25 华为技术有限公司 Web conference control method and device
CN108764807A (en) * 2018-03-29 2018-11-06 联想(北京)有限公司 A kind of information processing method and intelligent meeting system
CN110008361A (en) * 2019-03-21 2019-07-12 深圳创维-Rgb电子有限公司 Automatic identification shows electronics famous brand processing method, system and the storage medium of information
CN110557596A (en) * 2018-06-04 2019-12-10 杭州海康威视数字技术股份有限公司 conference system
CN111091646A (en) * 2018-10-22 2020-05-01 大连艾米移动科技有限公司 Intelligent conference security check system based on face recognition
CN111861048A (en) * 2019-04-24 2020-10-30 阿里巴巴集团控股有限公司 Conference data processing method, device and system and electronic equipment

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020120788A1 (en) * 2001-02-23 2002-08-29 Wang Douglas W. Method and system of activating computer peripherals by processed messages in real time
CN106796692A (en) * 2014-09-29 2017-05-31 惠普发展公司,有限责任合伙企业 Technical support is provided a user with via wearable computing devices
US10586539B2 (en) * 2015-06-01 2020-03-10 AffectLayer, Inc. In-call virtual assistant
CN110401811A (en) * 2019-08-07 2019-11-01 单正建 A kind of video conferencing system towards hierarchical organization

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003304337A (en) * 2002-04-08 2003-10-24 Nec Corp Speech control system in remote conference system
CN104468137A (en) * 2013-09-12 2015-03-25 华为技术有限公司 Web conference control method and device
CN108764807A (en) * 2018-03-29 2018-11-06 联想(北京)有限公司 A kind of information processing method and intelligent meeting system
CN110557596A (en) * 2018-06-04 2019-12-10 杭州海康威视数字技术股份有限公司 conference system
CN111091646A (en) * 2018-10-22 2020-05-01 大连艾米移动科技有限公司 Intelligent conference security check system based on face recognition
CN110008361A (en) * 2019-03-21 2019-07-12 深圳创维-Rgb电子有限公司 Automatic identification shows electronics famous brand processing method, system and the storage medium of information
CN111861048A (en) * 2019-04-24 2020-10-30 阿里巴巴集团控股有限公司 Conference data processing method, device and system and electronic equipment

Also Published As

Publication number Publication date
CN112911195A (en) 2021-06-04

Similar Documents

Publication Publication Date Title
CN112911195B (en) Online live television teleconference intelligent management system based on cloud computing and artificial intelligence
CN110991381B (en) Real-time classroom student status analysis and indication reminding system and method based on behavior and voice intelligent recognition
CN108229441B (en) Classroom teaching automatic feedback system and feedback method based on image and voice analysis
US20090123035A1 (en) Automated Video Presence Detection
CN112037788B (en) Voice correction fusion method
CN112801608A (en) Remote video conference intelligent management system based on big data and cloud computing and cloud conference management platform
CN112183238B (en) Remote education attention detection method and system
CN110033403A (en) A kind of unaware check class attendance method and system
CN111833876A (en) Conference speech control method, system, electronic device and storage medium
CN110648667B (en) Multi-person scene human voice matching method
CN111970471B (en) Conference participant scoring method, device, equipment and medium based on video conference
CN108109446B (en) Teaching class condition monitoring system
CN109858809A (en) Learning quality appraisal procedure and system based on the analysis of classroom students ' behavior
CN108109445B (en) Teaching course condition monitoring method
CN115052126B (en) Ultra-high definition video conference analysis management system based on artificial intelligence
CN110827432A (en) Class attendance checking method and system based on face recognition
CN104135638A (en) Optimized video snapshot
JP2004118314A (en) Utterer detection system and video conference system using same
JP7204337B2 (en) CONFERENCE SUPPORT DEVICE, CONFERENCE SUPPORT SYSTEM, CONFERENCE SUPPORT METHOD AND PROGRAM
US8452599B2 (en) Method and system for extracting messages
CN111915755A (en) Rapid classroom attendance checking method and system based on artificial intelligence image prediction
CN110265038A (en) A kind of processing method and electronic equipment
CN113611308B (en) Voice recognition method, device, system, server and storage medium
CN110516562A (en) A kind of colleges and universities' personnel positioning method and device based on recognition of face
CN112348493A (en) Intelligent conference recording system and method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right

Effective date of registration: 20220802

Address after: 200000 room 2201-1, building 1, No. 2167, Zhenbei Road, Putuo District, Shanghai

Applicant after: Weimaikejian Group Co.,Ltd.

Address before: 430071 No.2 Bayi Road, Wuchang District, Wuhan City, Hubei Province

Applicant before: Yin Xiaodong

TA01 Transfer of patent application right
GR01 Patent grant
GR01 Patent grant
CP01 Change in the name or title of a patent holder

Address after: 200000 room 2201-1, building 1, No. 2167, Zhenbei Road, Putuo District, Shanghai

Patentee after: Weimai Technology Co.,Ltd.

Address before: 200000 room 2201-1, building 1, No. 2167, Zhenbei Road, Putuo District, Shanghai

Patentee before: Weimaikejian Group Co.,Ltd.

CP01 Change in the name or title of a patent holder