KR101630614B1 - System and method for producing lecture contents - Google Patents
System and method for producing lecture contents Download PDFInfo
- Publication number
- KR101630614B1 KR101630614B1 KR1020150041692A KR20150041692A KR101630614B1 KR 101630614 B1 KR101630614 B1 KR 101630614B1 KR 1020150041692 A KR1020150041692 A KR 1020150041692A KR 20150041692 A KR20150041692 A KR 20150041692A KR 101630614 B1 KR101630614 B1 KR 101630614B1
- Authority
- KR
- South Korea
- Prior art keywords
- sound
- voice
- time difference
- transmitted
- mixer
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/8106—Monomedia components thereof involving special audio data, e.g. different tracks for different languages
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
- G06Q50/10—Services
- G06Q50/20—Education
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
- H04N21/854—Content authoring
Abstract
Description
BACKGROUND OF THE
Recently, the rate of using video lectures is increasing according to various purposes (for example, learning) for individuals. Video lectures are one of the ways to complement the shortcomings of offline real-time on-site lectures.
In the production of contents of this video lecture, it is produced in a studio equipped with a broadcasting equipment. However, video lecture contents produced in the studio are limited in the variety of lectures. That is, since only the content of the lecturer's lecture invited to the studio is produced as the content, the diversity of contents of the lecture can not be limited.
Therefore, in order to produce contents of various video lectures, a technique has been proposed in which video lecture contents can be easily produced at an arbitrary place. However, this technology also requires basic camera and sound equipment, and at the same time, it needs to solve at least the problem of automating camera and sound equipment in order to minimize the production staff (or one-person production system).
In other words, lectures on the off-line are often performed by lecturers for writing. In this case, it is possible to acquire various images through the movement and zoom function of the camera and the tracking method. However, there is a problem that it is difficult to acquire voice when the lecturer is moved away from the microphone formed in the camera or fixed separately.
As a method for acquiring voice, a lecturer uses a method of directly holding a microphone or attaching a wired / wireless microphone to the body. However, this has the problem of making lecturers uncomfortable.
Therefore, it is necessary to have a way to obtain voice effectively.
SUMMARY OF THE INVENTION Accordingly, the present invention has been made keeping in mind the above problems occurring in the prior art, and it is an object of the present invention to provide a portable sound recorder capable of automatically adjusting the position of a camera using a time difference of a voice signal transmitted from a plurality of sound collectors And a video lecture contents production system and method in which video lecture contents are produced by finally mixing audio and video compensated for time differences.
According to another aspect of the present invention, there is provided a moving picture lecture contents production system including: a plurality of sound collectors for collecting sounds; A voice controller for combining voice corresponding to a time difference of the voice signal transmitted from the plurality of sound collectors and generating location data of the sound source; A mobile camera for performing movement control corresponding to the position data; A mixer for mixing an image transmitted from the mobile camera and a voice transmitted from the voice controller; And a database for storing moving images output from the mixer.
Here, a fixed camera for photographing a certain space may be further included, and the mixer mixes the video transmitted from the mobile camera and the fixed camera with the voice transmitted from the audio controller.
It is also preferable that a speaker is connected to the audio controller.
A wired / wireless communication device for transmitting moving images to the mixer is preferably connected.
Wherein the sound controller comprises: a time difference measuring unit for measuring a time difference of a voice signal transmitted from each sound collector; A position determiner for determining a position of a lecturer based on a time difference of a voice signal of each sound collector to generate position data; A time difference correcting unit for correcting a time difference corresponding to a time difference of a sound signal of each sound collector; And a synthesizer for outputting the voice data synthesized with the time-lapse corrected voice.
The sound collector includes a first collecting part having a curvature in a first direction and collecting sound into the inside thereof; A second collecting part having a curvature in the first direction and guiding and collecting sound in a gap separated from the first collecting part; And an acoustic processor for performing acoustic processing on the sound collected using the plurality of microphones.
According to another aspect of the present invention, there is provided a method for producing a moving picture lecture content, comprising: generating each sound signal from a sound collected using a plurality of sound pickup units; Generating location data from a time difference of each voice signal in a voice controller; Moving the camera corresponding to the position data; And mixing the image data transmitted from the camera and the voice data transmitted from the voice controller using a hop sequence.
Preferably, the voice data is combined with voice corresponding to a time difference between voice signals transmitted from a plurality of sound pickup units.
Here, the audio controller can output a field sound through the connected speaker, and can transmit the moving picture in real time through the wired / wireless communication device connected to the mixer.
As described above, according to the video lecture contents production system and method of the present invention, since the loudspeaker is used to acquire the voice, it is possible not only to improve the convenience of the lecturer, So that control can be unified.
Further, according to the present invention, the entire system can be configured with a relatively low installation cost, and the ease of production of video lecture contents can be improved through automation using voice.
1 is a block diagram of a video lecture contents production system according to an embodiment of the present invention.
2 is a configuration diagram of a voice controller according to an embodiment of the present invention.
3 to 11 are various examples of the sound pickup machine applied to the present invention.
FIG. 12 is a flowchart of a method of producing a moving picture lecture content according to an embodiment of the present invention.
FIG. 13 and FIG. 14 are conceptual diagrams showing an extended contents production system of a moving picture lecture of the present invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS The present invention will be described in detail with reference to the accompanying drawings.
1 is a block diagram of a video lecture contents production system according to an embodiment of the present invention.
First, the same reference numerals are assigned to the components that perform the same function in the drawings.
1, a moving picture lecture contents production system according to the present invention includes a plurality of
Here, it may further include a
Also, the
The
On the other hand, in the present embodiment, the case where a plurality of
2 is a configuration diagram of a voice controller according to an embodiment of the present invention.
2, the
Here, the time
Then, the
3 to 11 are various examples of the sound pickup machine applied to the present invention.
3 to 11, the
The
Here, the
Meanwhile, the
The outer peripheral portion of the receiving
In addition, the receiving
In this case, the fixing
At this time, it is preferable that a substrate or the like is provided at the entrance of the inserting
On the other hand, the
On the other hand, the
The
On the other hand, the function of the
Hereinafter, a method for producing a moving picture lecture content of the present invention using the system configured as described above will be described.
FIG. 12 is a flowchart of a method of producing a moving picture lecture content according to an embodiment of the present invention.
Referring to FIG. 12, a sound signal is generated by each sound collector corresponding to a sound propagated from a sound source position (lecture position), and is transmitted to a sound controller (S1).
The voice controller measures the time difference of each voice signal transmitted (S2).
The voice controller generates position data from the time difference of the voice signal and outputs the generated position data (S3).
Then, the position data is transmitted to the mobile camera, and the mobile camera moves the mobile camera in accordance with the position data (S4). Thereafter, it is preferable that the lecturer can be photographed more accurately through a video tracking technique or the like. On the other hand, the fixed camera captures, for example, a part or all of the blackboard in a fixed manner.
On the other hand, the time difference is corrected corresponding to the time difference of the audio signal (S5). At this time, it is preferable to correct the reference of correction based on the strongest speech signal. Of course, it can be corrected on the basis of the speech signal which is inputted first among the speech signals. Subsequently, the voice controller outputs voice data obtained by synthesizing the time-difference-corrected voice (S6).
Then, the video data photographed from the mobile camera and the fixed camera are mixed with the audio data transmitted from the audio controller (S7), and the mixed video data is stored in the database (S8).
On the other hand, when a speaker is connected to the voice controller, the lecture voice can be output in real time through a speaker as a field sound.
When a wired / wireless communication device for transmitting a moving picture to the mixer is connected, the moving picture can be delivered to a field student in real time.
FIG. 13 and FIG. 14 are conceptual diagrams showing an extended contents production system of a moving picture lecture of the present invention.
13 and 14, as an example, it is possible to construct a server for managing a video-lecture contents production system on-line by combining network equipment with a mixer, and thereby, video data transmitted through a network can be stored in a database There will be.
In addition, the video lecture contents production system may be installed for each classroom, and video data transmitted from a plurality of video lecture contents production systems may be integrally managed using a server and a database. It will also be manageable by group (eg school, school).
While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it is to be understood that the invention is not limited to the disclosed exemplary embodiments, but, on the contrary, is intended to cover various modifications and equivalent arrangements included within the spirit and scope of the invention.
1: Sound collector
2: Audio controller
3: Moving camera
4: Mixer
5: Database
6: Fixed camera
7: Speaker
8: wired / wireless communication device
Claims (10)
A voice controller for generating position data of a lecturer by combining the plurality of voice signals in consideration of a time difference between each of a plurality of voice signals transmitted from each of the plurality of sound collectors and a receiving intensity of each of the plurality of voice signals;
A mobile camera for performing movement control corresponding to the position data;
A mixer for mixing an image transmitted from the mobile camera and a voice transmitted from the voice controller; And
And a database for storing moving images output from the mixer,
Wherein the voice controller comprises:
A time difference measurement unit for measuring a time difference of each of the plurality of audio signals transmitted from each of the plurality of sound collectors;
A position determination unit for determining the position of the lecturer from the time difference of each of the plurality of sound signals of each of the plurality of sound collector to generate the position data;
A time difference correction unit for correcting the time difference of each of the plurality of audio signals; And
And a synthesizer for outputting voice data obtained by synthesizing the plurality of voice signals corrected by the time difference corrector,
Wherein the time difference measuring unit determines whether or not the sound source intensity has a sound source strength higher than a set value and measures a first output energy including a delay function for each voice signal frame of each of the plurality of voice signals having a sound source strength higher than the set value, Determining a cross-correlation between the plurality of sound pickups based on the output energy,
Wherein the position determining unit selects the sound source direction candidate group based on the cross correlation and selects a predetermined number of candidate directions based on the sampling frequency and the interval between the plurality of sound trackers for the sound source direction candidate group, The second output energy is calculated using the cross correlation for each of the candidate directions, and the direction in which the second output energy is the largest among the candidate directions is determined as the position data.
And a fixed camera for photographing a predetermined space,
Wherein the mixer mixes the video transmitted from the mobile camera and the fixed camera with the voice transmitted from the audio controller.
And a speaker is connected to the audio controller.
And a wired / wireless communication device for transmitting a moving picture to the mixer is connected to the mixer.
Wherein each of the plurality of sound collectors comprises :
A first collecting part having a curvature in a first direction and collecting sound into the first collecting part;
A second collecting part having a curvature in the first direction and guiding and collecting sound in a gap separated from the first collecting part; And
And a sound processor for performing sound processing on the sound collected using a plurality of microphones.
Generating position data of a lecturer by combining the plurality of voice signals based on a time difference for each of the plurality of voice signals and a reception intensity of the plurality of voice signals in a voice controller;
Moving the camera corresponding to the position data; And
Mixing the video data transmitted from the camera with the plurality of audio signals transmitted from the audio controller using a mixer,
Wherein the voice controller comprises:
A time difference measurement unit for measuring a time difference of each of the plurality of audio signals transmitted from each of the plurality of sound collectors;
A position determination unit for determining the position of the lecturer from the time difference of each of the plurality of sound signals of each of the plurality of sound collector to generate the position data;
A time difference correction unit for correcting the time difference of each of the plurality of audio signals; And
And a synthesizer for outputting voice data obtained by synthesizing the plurality of voice signals corrected by the time difference corrector,
Wherein the time difference measuring unit determines whether or not the sound source intensity has a sound source strength higher than a set value and measures a first output energy including a delay function for each voice signal frame of each of the plurality of voice signals having a sound source strength higher than the set value, Determining a cross-correlation between the plurality of sound pickups based on the output energy,
Wherein the position determining unit selects the sound source direction candidate group based on the cross correlation and selects a predetermined number of candidate directions based on the sampling frequency and the interval between the plurality of sound trackers for the sound source direction candidate group, The second output energy is calculated using the cross correlation for each of the candidate directions, and the direction in which the second output energy is the largest among the candidate directions is determined as the position data.
The audio data is the time difference by way content generated video lecture to combine the speech corresponding to the speech signal for each transmitted from the sound collector can suit.
Wherein the audio controller outputs field sounds through connected speakers.
And transmitting the moving picture in real time through the wired / wireless communication device connected to the mixer.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020150041692A KR101630614B1 (en) | 2015-03-25 | 2015-03-25 | System and method for producing lecture contents |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020150041692A KR101630614B1 (en) | 2015-03-25 | 2015-03-25 | System and method for producing lecture contents |
Publications (1)
Publication Number | Publication Date |
---|---|
KR101630614B1 true KR101630614B1 (en) | 2016-06-24 |
Family
ID=56343498
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020150041692A KR101630614B1 (en) | 2015-03-25 | 2015-03-25 | System and method for producing lecture contents |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR101630614B1 (en) |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10227849A (en) * | 1997-02-14 | 1998-08-25 | Fuji Xerox Co Ltd | Sound source position measuring device, camera image pick-up controller, sound source position recording device, and sound source position measurement method |
JP2003230049A (en) * | 2002-02-06 | 2003-08-15 | Sharp Corp | Camera control method, camera controller and video conference system |
JP2007124140A (en) * | 2005-10-26 | 2007-05-17 | Yamaha Corp | Photographing device and communication conference system |
KR20080049431A (en) | 2006-11-30 | 2008-06-04 | 지창훈 | Production system real time lecture contents using projector and method thereof |
KR101188828B1 (en) * | 2011-04-08 | 2012-10-09 | 백민호 | The sound collector which uses on microphon of cassegrain method |
KR20140078043A (en) | 2012-12-14 | 2014-06-25 | 김용민 | A lecture contents manufacturing system and method which anyone can easily make |
-
2015
- 2015-03-25 KR KR1020150041692A patent/KR101630614B1/en active IP Right Grant
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10227849A (en) * | 1997-02-14 | 1998-08-25 | Fuji Xerox Co Ltd | Sound source position measuring device, camera image pick-up controller, sound source position recording device, and sound source position measurement method |
JP2003230049A (en) * | 2002-02-06 | 2003-08-15 | Sharp Corp | Camera control method, camera controller and video conference system |
JP2007124140A (en) * | 2005-10-26 | 2007-05-17 | Yamaha Corp | Photographing device and communication conference system |
KR20080049431A (en) | 2006-11-30 | 2008-06-04 | 지창훈 | Production system real time lecture contents using projector and method thereof |
KR101188828B1 (en) * | 2011-04-08 | 2012-10-09 | 백민호 | The sound collector which uses on microphon of cassegrain method |
KR20140078043A (en) | 2012-12-14 | 2014-06-25 | 김용민 | A lecture contents manufacturing system and method which anyone can easily make |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102186049B (en) | Conference terminal audio signal processing method, conference terminal and video conference system | |
US20180213345A1 (en) | Multi-Apparatus Distributed Media Capture for Playback Control | |
US10206030B2 (en) | Microphone array system and microphone array control method | |
US8989552B2 (en) | Multi device audio capture | |
CN108089152B (en) | Equipment control method, device and system | |
US20080219485A1 (en) | Apparatus, System and Method for Acoustic Signals | |
CN111918018B (en) | Video conference system, video conference apparatus, and video conference method | |
EP2352290B1 (en) | Method and apparatus for matching audio and video signals during a videoconference | |
CN103004238A (en) | Facilitating communications using a portable communication device and directed sound output | |
CN204539315U (en) | A kind of video conference machine of auditory localization | |
US11601731B1 (en) | Computer program product and method for auto-focusing a camera on an in-person attendee who is speaking into a microphone at a hybrid meeting that is being streamed via a videoconferencing system to remote attendees | |
CN111163281A (en) | Panoramic video recording method and device based on voice tracking | |
JP7428763B2 (en) | Information acquisition system | |
CN103414992A (en) | Audio file adjustment system | |
KR101976937B1 (en) | Apparatus for automatic conference notetaking using mems microphone array | |
KR101630614B1 (en) | System and method for producing lecture contents | |
US11665391B2 (en) | Signal processing device and signal processing system | |
WO2011108377A1 (en) | Coordinated operation apparatus, coordinated operation method, coordinated operation control program and apparatus coordination system | |
Meyer-Kahlen et al. | Design and measurement of first-order, horizontally beam-controlling loudspeaker cubes | |
WO2018173139A1 (en) | Imaging/sound acquisition device, sound acquisition control system, method for controlling imaging/sound acquisition device, and method for controlling sound acquisition control system | |
KR101687676B1 (en) | System and method for managing lecture contents | |
Ishigaki et al. | Zoom microphone | |
JP2021197658A (en) | Sound collecting device, sound collecting system, and sound collecting method | |
WO2013045533A1 (en) | Multimodal mobile video telephony | |
US11877058B1 (en) | Computer program product and automated method for auto-focusing a camera on a person in a venue who is wearing, or carrying, or holding, or speaking into a microphone at the venue |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
E701 | Decision to grant or registration of patent right | ||
GRNT | Written decision to grant | ||
FPAY | Annual fee payment |
Payment date: 20190124 Year of fee payment: 4 |