KR20170074015A - Method for editing video conference image and apparatus for executing the method - Google Patents
Method for editing video conference image and apparatus for executing the method Download PDFInfo
- Publication number
- KR20170074015A KR20170074015A KR1020150182982A KR20150182982A KR20170074015A KR 20170074015 A KR20170074015 A KR 20170074015A KR 1020150182982 A KR1020150182982 A KR 1020150182982A KR 20150182982 A KR20150182982 A KR 20150182982A KR 20170074015 A KR20170074015 A KR 20170074015A
- Authority
- KR
- South Korea
- Prior art keywords
- video conference
- editing
- video
- image
- interest
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/15—Conference systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
- G06Q50/10—Services
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/488—Data services, e.g. news ticker
- H04N21/4886—Data services, e.g. news ticker for displaying a ticker, e.g. scrolling banner for news, stock exchange, weather data
-
- H04N5/225—
Abstract
A video conference video editing method and apparatus for performing the same are disclosed. A method for editing a video conference image according to an exemplary embodiment includes the steps of receiving a video conference video, receiving a request for editing a content of interest from a video conference terminal in a video conference, A step of recognizing a voice in a part of the image corresponding to the request for editing a part of interest and converting the recognized voice into a text, And creating an edited image.
Description
Embodiments of the invention relate to video conferencing technology.
Conferences are held for the purpose of conveying information, collecting opinions, and drawing out results among people. In recent years, with the development of IT technology, there has been an increasing tendency to collect information and receive opinions through video conferences (video conferences or teleconferences).
In these videoconferences, it is difficult for the presenter to know when the content of the meeting is being delivered to the participants of the conference, as intended. It is also difficult for participants to understand the content of the presentation if they are interested in the presentation and the terminology or abbreviation is used in the presentation. Finally, when editing the meeting results is necessary, it takes a lot of time and effort to understand and organize important contents. Particularly, in the case of a person who can not attend the meeting, it is necessary to watch the whole meeting video.
Korean Patent Registration No. 10-0575634 (Nov. 10, 2006)
An embodiment of the present invention is to provide a video conference video editing method and an apparatus for performing the video conference video editing method in which a video conference participant can easily edit an image of a necessary part in a video conference.
An embodiment of the present invention is to provide a video conference video editing method capable of generating a highlight editing video for a video conference video by grasping the interest of video conference participants and an apparatus for performing the video conference video editing method.
A video conference video editing method according to an exemplary embodiment includes: receiving a video conference video at a video conference editing server; Extracting an image of a portion corresponding to the interest portion edit request from the video conference image when a request for editing the interest portion is received from the video conference terminal in the video conference by the video conference editing server; Recognizing a voice in an image of a portion corresponding to the request for editing the interest portion in the video conference editing server and converting the recognized voice into text; And generating, in the video conference editing server, an edited image by inserting the converted text as a subtitle into an image of a portion corresponding to the interest portion editing request.
Wherein the step of extracting the image corresponding to the request for editing the interest portion comprises the steps of: receiving, at the video conference editing server, Respectively; And extracting, at the video conference editing server, a section from the start of speech to the end of speech in the video conference video.
The method may further include the step of providing the edited video to the video conference terminal in the video conference at the video conference editing server after the editing video is generated.
Confirming at the video conference editing server whether or not an annotation required term is used in the video conference video; Retrieving annotation information describing the meaning of the annotation required term in the video conference editing server when the annotation required term is used; And providing the retrieved annotation information to each video conference terminal in the video conference at the video conference editing server.
Further comprising the step of statistically processing the interest of the participants of the video conference by analyzing the interest part editing requests received from each video conference terminal in the video conference at the video conference editing server after the step of generating the edited video can do.
The method may further include, at the video conference editing server, generating a highlight editing image for the video conference image based on the interest of the participants of the video conference after the statistical processing step.
The step of generating the highlight editing video may further include the step of providing the highlight editing video to each video conference terminal participating in the video conference at the video conference editing server.
A video conference video editing method according to another exemplary embodiment includes: receiving a video conference video at a video conference editing server; Receiving, at the video conference editing server, a request for editing a point of interest from each video conference terminal in a video conference; Analyzing the interest part editing requests received from each video conference terminal and statistically processing the participants' interest in the video conference at the video conference editing server; And generating, in the video conference editing server, a highlight editing video for the video conference video based on the interest of the participants of the video conference.
The highlight editing video may be generated for a portion of the video conference image having the highest degree of interest of the participants.
The step of generating the highlight editing image may include extracting an image of a portion of the video conference image having the highest degree of interest by the participant in the video conference editing server, Recognizing a voice in an image of a part having the highest degree of interest of the participants in the video conference editing server and converting the recognized voice into text; And inserting the converted text as a subtitle into an image of a portion of the participant having the highest interest in the video conference editing server.
Confirming at the video conference editing server whether or not an annotation required term is used in the video conference video; Retrieving annotation information describing the meaning of the annotation required term in the video conference editing server when the annotation required term is used; And providing the retrieved annotation information to each video conference terminal in the video conference at the video conference editing server.
The step of generating the highlight editing video may further include the step of providing the highlight editing video to each video conference terminal participating in the video conference at the video conference editing server.
An apparatus according to one exemplary embodiment includes one or more processors; Memory; And one or more programs configured to be stored in the memory and configured to be executed by the one or more processors, the program comprising the steps of: determining whether a point of interest editing request is received from a video conferencing terminal in a video conference; ; Extracting an image of a portion corresponding to the interest portion editing request from the video conference video when the interested portion editing request is received; Recognizing speech in a portion of the image corresponding to the interest portion edit request, and converting the recognized speech into text; And inserting the converted text as a subtitle into an image of a portion corresponding to the interest portion editing request to generate an edited image.
The program may further include a step of extracting an image of a portion corresponding to the request for editing the interest portion and detecting a start portion of the speaker's utterance and an end portion of the utterance in the video conference image, ; And extracting an interval from the start of speech to the end of speech in the video conference video.
The program may further include instructions for performing the step of providing the edited video to the corresponding video conference terminal in the video conference after the step of generating the edited video.
The program comprising the steps of: determining whether an annotation required term is used in the video conference video; Searching for annotation information describing the meaning of the annotation required term if the annotation required term is used; And providing the retrieved annotation information to each videoconference terminal in the videoconference.
The program further comprising the step of analyzing the interest portion editing requests received from each video conference terminal in the video conference after the step of generating the edited video to further statistically process the participants' Lt; / RTI >
The program may include instructions for further performing the step of generating a highlighted edit image for the video conference video based on the interest of the participants of the video conference after the statistical processing step.
The program may further comprise instructions for performing the step of providing the highlight edit image to each videoconference terminal participating in the videoconference, after the step of generating the highlight edit video.
An apparatus according to another exemplary embodiment includes one or more processors; Memory; And one or more programs, wherein the one or more programs are stored in the memory and are configured to be executed by the one or more processors, the program comprising the steps of: receiving a request for editing a point of interest from each videoconference terminal in a videoconference step; Analyzing the interest portion editing requests received from each videoconference terminal to statistically process the participants' interest in the videoconference; And generating a highlight edit image for the videoconference image based on the interest of the participants of the videoconference.
The program may generate the highlight editing image for a portion of the video conference image having the highest degree of interest of the participants.
The program comprising the steps of: extracting an image of a portion of the video conference image having the highest degree of interest of the participant in the step of generating the highlight edit video; Recognizing speech in a portion of the image of the highest degree of interest of the participants and converting the recognized speech into text; And inserting the transformed text into a caption of an image of a part having the highest degree of interest of the participants.
The program comprising the steps of: determining whether an annotation required term is used in the video conference video; Searching for annotation information describing the meaning of the annotation required term if the annotation required term is used; And providing the retrieved annotation information to each videoconference terminal in the videoconference.
The program may further comprise instructions for performing the step of providing the highlight edit image to each videoconference terminal participating in the videoconference, after the step of generating the highlight edit video.
According to the embodiment of the present invention, a participant of a video conference during a video conference can edit and manage a part of his / her interest in real time. In other words, by providing editing video reflecting the interest of the participant in each video conference attendee, the presenter can grasp the interest information and the understanding level of the attendees of each conference, and attendees can edit and summarize the video conference contents in a simple way . In addition, by providing the highlighted portion of the video conference as the most interesting part of the video conference attendees, those who can not attend the meeting can quickly grasp the contents of the video conference.
1 is a block diagram illustrating the configuration of a video conferencing system according to an exemplary embodiment;
2 is a block diagram showing a configuration of a video conference editing server according to an exemplary embodiment;
3 is a flowchart showing a video conference video editing method according to the exemplary embodiment
4 is a view showing a user interface screen provided in a video conference terminal according to an exemplary embodiment;
5 illustrates a computing environment including an exemplary computing device suitable for use in the exemplary embodiments
Hereinafter, specific embodiments of the present invention will be described with reference to the drawings. The following detailed description is provided to provide a comprehensive understanding of the methods, apparatus, and / or systems described herein. However, this is merely an example and the present invention is not limited thereto.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS Hereinafter, exemplary embodiments of the present invention will be described in detail with reference to the accompanying drawings. In the following description, well-known functions or constructions are not described in detail since they would obscure the invention in unnecessary detail. The following terms are defined in consideration of the functions of the present invention, and may be changed according to the intention or custom of the user, the operator, and the like. Therefore, the definition should be based on the contents throughout this specification. The terms used in the detailed description are intended only to describe embodiments of the invention and should in no way be limiting. Unless specifically stated otherwise, the singular form of a term includes plural forms of meaning. In this description, the expressions "comprising" or "comprising" are intended to indicate certain features, numbers, steps, operations, elements, parts or combinations thereof, Should not be construed to preclude the presence or possibility of other features, numbers, steps, operations, elements, portions or combinations thereof.
In the following description, terms such as " transmission ", "transmission "," transmission ", "reception ", and the like, of a signal or information refer not only to the direct transmission of signals or information from one component to another But also through other components. In particular, "transmitting" or "transmitting" a signal or information to an element is indicative of the final destination of the signal or information and not a direct destination. This is the same for "reception" of a signal or information. Also, in this specification, the fact that two or more pieces of data or information are "related" means that when one piece of data (or information) is acquired, at least a part of the other data (or information) can be obtained based thereon.
1 is a block diagram showing the configuration of a video conferencing system according to an exemplary embodiment.
Referring to FIG. 1, the video conferencing system 100 may include a
The
The
The video
The video
When a technical term or an abbreviation is used in a video conference, the video
The video
Here, it is described that the video
Although the video
According to the embodiment of the present invention, a participant of a video conference during a video conference can edit and manage a part of his / her interest in real time. In other words, by providing editing video reflecting the interest of the participant in each video conference attendee, the presenter can grasp the interest information and the understanding level of the attendees of each conference, and attendees can edit and summarize the video conference contents in a simple way . In addition, by providing the highlighted portion of the video conference as the most interesting part of the video conference attendees, those who can not attend the meeting can quickly grasp the contents of the video conference.
2 is a block diagram showing a configuration of a video conference editing server according to an exemplary embodiment.
2, the video
The
When the
In addition, the
The
When the annotation required term is used in the video conference, the
The
The
3 is a flowchart illustrating a video conference video editing method according to an exemplary embodiment. In the illustrated flow chart, the method is described as being divided into a plurality of steps, but at least some of the steps may be performed in reverse order, combined with other steps, performed together, omitted, divided into detailed steps, One or more steps may be added and performed. Also, one or more steps not shown in the method according to the embodiment may be performed with the method.
Referring to FIG. 3, the video
Next, the video
If it is determined in step S 103 that the interested part editing request is received, the video
Next, the video
Next, the video
Next, the video
Next, the video
Next, the video
4 is a diagram illustrating a user interface screen provided in a video conference terminal according to an exemplary embodiment.
Referring to FIG. 4, the
FIG. 5 illustrates a computing environment including an exemplary computing device suitable for use in the exemplary embodiments.
The
The
The
Computer
The
The
On the other hand, certain embodiments may include a computer readable storage medium comprising a program for performing the procedures described herein on a computer. Such computer-readable storage media may include program instructions, local data files, local data structures, etc., alone or in combination. The computer-readable storage medium may be those specially designed and constructed for the present invention. Examples of computer-readable storage media include magnetic media such as hard disks, floppy disks and magnetic tape, optical recording media such as CD-ROMs and DVDs, magneto-optical media such as floppy disks, and ROMs, And hardware devices specifically configured to store and execute the same program instructions. Examples of program instructions may include machine language code such as those generated by a compiler, as well as high-level language code that may be executed by a computer using an interpreter or the like.
While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it is to be understood that the invention is not limited to the disclosed exemplary embodiments, but, on the contrary, . Therefore, the scope of the present invention should not be limited to the above-described embodiments, but should be determined by equivalents to the appended claims, as well as the appended claims.
100: Video conferencing system
102: video conference terminal
104: Video conference management server
106: Video conference editing server
111:
113: Editor
115: Text conversion section
117:
119: Statistical Analysis Department
121:
Claims (24)
Extracting an image of a portion corresponding to the interest portion edit request from the video conference image when a request for editing the interest portion is received from the video conference terminal in the video conference by the video conference editing server;
Recognizing a voice in an image of a portion corresponding to the request for editing the interest portion in the video conference editing server and converting the recognized voice into text; And
And editing the video conference by the video conference editing server by inserting the converted text as a subtitle into an image of a portion corresponding to the interest portion editing request to generate an edited image.
Wherein the step of extracting an image of a portion corresponding to the interest-
Detecting, at the video conference editing server, a start point of a speaker and a end point of a speaker in the video conference video based on a point of time when the request for editing the point of interest is received; And
And extracting a section from the start of speech to the end of speech in the video conference video in the video conference editing server.
After the step of generating the edited image,
Further comprising the step of providing, in the video conference editing server, the edited video to the video conference terminal in the video conference.
Confirming at the video conference editing server whether or not an annotation required term is used in the video conference video;
Retrieving annotation information describing the meaning of the annotation required term in the video conference editing server when the annotation required term is used; And
Further comprising the step of providing, at the video conference editing server, the retrieved annotation information to each video conference terminal in the video conference.
After the step of generating the edited image,
Further comprising the step of statistically processing the interest of the participants of the video conference by analyzing the interest part editing requests received from each video conference terminal in the video conference at the video conference editing server.
After the statistical processing step,
Further comprising the step of creating, in the video conference editing server, a highlight editing video for the video conference video based on the interest of the participants of the video conference.
After the step of generating the highlight edit image,
Further comprising the step of, at the video conference editing server, providing the highlight editing video to each video conference terminal participating in the video conference.
Receiving, at the video conference editing server, a request for editing a point of interest from each video conference terminal in a video conference;
Analyzing the interest part editing requests received from each video conference terminal and statistically processing the participants' interest in the video conference at the video conference editing server; And
And a step of generating, in the video conference editing server, a highlight editing video for the video conference video based on the interest of the participants of the video conference.
The highlight editing image may include:
Wherein a portion of the video conference image having the highest degree of interest of the participants is generated.
Wherein the step of generating the highlight editing image comprises:
Extracting, at the video conference editing server, an image of a portion of the video conference image in which the participants are most interested;
Recognizing a voice in an image of a part having the highest degree of interest of the participants in the video conference editing server and converting the recognized voice into text; And
And inserting the converted text as a subtitle into an image of a portion of the participant having the highest degree of interest in the video conference editing server.
Confirming at the video conference editing server whether or not an annotation required term is used in the video conference video;
Retrieving annotation information describing the meaning of the annotation required term in the video conference editing server when the annotation required term is used; And
Further comprising the step of providing, at the video conference editing server, the retrieved annotation information to each video conference terminal in the video conference.
After the step of generating the highlight edit image,
Further comprising the step of, at the video conference editing server, providing the highlight editing video to each video conference terminal participating in the video conference.
Memory; And
An apparatus comprising one or more programs,
Wherein the one or more programs are stored in the memory and are configured to be executed by the one or more processors,
The program includes:
Confirming whether a request for editing a point of interest is received from a video conference terminal in a video conference;
Extracting an image of a portion corresponding to the interest portion editing request from the video conference image when the interested portion editing request is received;
Recognizing speech in a portion of the image corresponding to the interest portion edit request, and converting the recognized speech into text; And
And inserting the converted text as a subtitle into an image of a portion corresponding to the interest portion edit request to generate an edited image.
Wherein the program extracts an image of a portion corresponding to the interest portion edit request,
Detecting a speaker's utterance start point and a utterance end point in the video conference video based on a point of time when the request for editing the point of interest is received; And
And extracting an interval from the start of speech to the end of speech in the video conference image.
The program may further include, after the step of generating the edited video,
And providing the edited video to a corresponding videoconference terminal in the videoconference.
The program includes:
Confirming whether an annotation necessary term is used in the video conference video;
Searching for annotation information describing the meaning of the annotation required term if the annotation required term is used; And
And providing the retrieved annotation information to each videoconference terminal in the videoconference.
The program may further include, after the step of generating the edited video,
Further comprising analyzing interest portion editing requests received from each videoconference terminal in the videoconference to statistically process the participants' interest in the videoconference.
The program may further comprise, after the statistical processing step,
Further comprising generating a highlighted edit image for the video conference image based on an interest of the participants of the video conference.
The program may further include, after the step of generating the highlight edit image,
And providing the highlight edit image to each videoconference terminal participating in the videoconference.
Memory; And
An apparatus comprising one or more programs,
Wherein the one or more programs are stored in the memory and are configured to be executed by the one or more processors,
The program includes:
Receiving a request for editing a point of interest from each videoconference terminal in a videoconference;
Analyzing the interest portion editing requests received from each videoconference terminal to statistically process the participants' interest in the videoconference; And
And generating a highlighted edit image for the videoconference image based on the interest of the participants of the videoconference.
The program includes:
And generates the highlight edited image for a portion of the video conference image having the highest degree of interest of the participants.
The program may further include, in the step of generating the highlight editing image,
Extracting an image of a portion of the video conference image in which the participants are most interested;
Recognizing speech in a portion of the image of the highest degree of interest of the participants and converting the recognized speech into text; And
And inserting the transformed text into a caption in an image of a portion of the participants having the highest degree of interest.
The program includes:
Confirming whether an annotation necessary term is used in the video conference video;
Searching for annotation information describing the meaning of the annotation required term if the annotation required term is used; And
And providing the retrieved annotation information to each videoconference terminal in the videoconference.
The program may further include, after the step of generating the highlight edit image,
And providing the highlight edit image to each videoconference terminal participating in the videoconference.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020150182982A KR20170074015A (en) | 2015-12-21 | 2015-12-21 | Method for editing video conference image and apparatus for executing the method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020150182982A KR20170074015A (en) | 2015-12-21 | 2015-12-21 | Method for editing video conference image and apparatus for executing the method |
Publications (1)
Publication Number | Publication Date |
---|---|
KR20170074015A true KR20170074015A (en) | 2017-06-29 |
Family
ID=59280277
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020150182982A KR20170074015A (en) | 2015-12-21 | 2015-12-21 | Method for editing video conference image and apparatus for executing the method |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR20170074015A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20190009201A (en) * | 2017-07-18 | 2019-01-28 | 엘지전자 주식회사 | Mobile terminal and method for controlling the same |
CN113645431A (en) * | 2021-07-07 | 2021-11-12 | 四川腾云法智互联网科技有限公司 | Method and device for realizing bankruptcy case video conference, electronic equipment and storage medium |
-
2015
- 2015-12-21 KR KR1020150182982A patent/KR20170074015A/en unknown
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20190009201A (en) * | 2017-07-18 | 2019-01-28 | 엘지전자 주식회사 | Mobile terminal and method for controlling the same |
CN113645431A (en) * | 2021-07-07 | 2021-11-12 | 四川腾云法智互联网科技有限公司 | Method and device for realizing bankruptcy case video conference, electronic equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20190007469A1 (en) | Copy and paste for web conference content | |
US20190253474A1 (en) | Media production system with location-based feature | |
US8805929B2 (en) | Event-driven annotation techniques | |
Nagao et al. | Discussion mining: Annotation-based knowledge discovery from real world activities | |
US11315569B1 (en) | Transcription and analysis of meeting recordings | |
US9569428B2 (en) | Providing an electronic summary of source content | |
US20120233155A1 (en) | Method and System For Context Sensitive Content and Information in Unified Communication and Collaboration (UCC) Sessions | |
US20170371496A1 (en) | Rapidly skimmable presentations of web meeting recordings | |
CN112653902B (en) | Speaker recognition method and device and electronic equipment | |
KR20120102043A (en) | Automatic labeling of a video session | |
JP2008282397A (en) | Method for creating annotated transcript of presentation, information processing system, and computer program | |
KR20080037947A (en) | Method and apparatus of generating meta data of content | |
JP2005341015A (en) | Video conference system with minute creation support function | |
US10084829B2 (en) | Auto-generation of previews of web conferences | |
CN110211590B (en) | Conference hotspot processing method and device, terminal equipment and storage medium | |
US20140200888A1 (en) | System and Method for Generating a Script for a Web Conference | |
US9525896B2 (en) | Automatic summarizing of media content | |
US10841115B2 (en) | Systems and methods for identifying participants in multimedia data streams | |
US9361714B2 (en) | Enhanced video description | |
JP5030868B2 (en) | Conference audio recording system | |
KR101618084B1 (en) | Method and apparatus for managing minutes | |
KR20170074015A (en) | Method for editing video conference image and apparatus for executing the method | |
US20140222840A1 (en) | Insertion of non-realtime content to complete interaction record | |
US20220222449A1 (en) | Presentation transcripts | |
KR101783872B1 (en) | Video Search System and Method thereof |