WO2005039175A1 - 映像音声記録再生装置、映像音声記録方法及び映像音声再生方法 - Google Patents
映像音声記録再生装置、映像音声記録方法及び映像音声再生方法 Download PDFInfo
- Publication number
- WO2005039175A1 WO2005039175A1 PCT/JP2004/009489 JP2004009489W WO2005039175A1 WO 2005039175 A1 WO2005039175 A1 WO 2005039175A1 JP 2004009489 W JP2004009489 W JP 2004009489W WO 2005039175 A1 WO2005039175 A1 WO 2005039175A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- audio
- information
- recording
- video
- additional
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/91—Television signal processing therefor
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
- G11B27/32—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on separate auxiliary tracks of the same or an auxiliary record carrier
- G11B27/322—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on separate auxiliary tracks of the same or an auxiliary record carrier used signal is digitally coded
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B20/00—Signal processing not specific to the method of recording or reproducing; Circuits therefor
- G11B20/10—Digital recording or reproducing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
Definitions
- Video / audio recording / reproducing apparatus video / audio recording method, and video / audio reproducing method
- the present invention relates to an apparatus and method for recording and reproducing video and audio, such as a memory recorder, and more particularly, to an apparatus and method for recording and reproducing audio and additional information in association with main video and audio information.
- Nonlinear editing devices that perform such editing work capture information such as video and audio recorded on tapes that have been collected as raw materials into a recording medium such as a hard disk that can be accessed randomly, and capture the information into this hard disk. Edit video and audio with random access.
- a video / audio recording / reproducing apparatus comprises: a video / audio input unit for inputting main information of video and audio; and an audio additional information input unit for inputting additional audio information added to the main information.
- a video / audio output unit for outputting main information and additional audio information; a recording medium for recording the main information and additional audio information; and a recording / reproduction for recording / reproducing the main information and additional audio information to / from the recording medium.
- control unit for controlling operations of the video / audio input unit, the audio / voice additional information input unit, the video / audio output unit, and the recording / reproducing unit.
- the control unit controls the audio additional information to be recorded on the recording medium in association with a specific frame position in the main information.
- a video / audio recording method includes: inputting audio additional information added to the main information by manually inputting video and audio main information; and converting the audio additional information to a specific frame in the main information. The information is recorded on the recording medium in association with the position.
- a method for reproducing a recording medium on which main information and audio additional information are recorded by the video / audio recording method wherein a thumbnail image of the main information is displayed and the same main information is displayed. Is associated with one or more pieces of additional audio information, and when any one of the one or more pieces of additional audio information is selected, the selected additional audio information is located at the associated frame position. The thumbnail image of the main information is displayed.
- the main information and the audio This is a method of reproducing a recording medium on which additional information is recorded, and reproduces the audio additional information without synchronization with the time axis of the main information.
- audio additional information for explaining the content of video and audio main information
- voice memo for explaining the content of video and audio main information
- Multiple voice memos can be recorded at one point.
- a specific frame position in the main information may be designated by the number of frames from the beginning of the main information, so that even when the time code of the material data is discontinuous, the voice additional data is stored in the unique position of the material data.
- audio additional information may be associated with each of the continuously recorded material data (clips), whereby the audio additional information can be used as a memo for each scene.
- audio additional data associated with the entire recording medium may be recorded.
- audio additional data may be associated with each recording medium, so that even if some recording media are removed, the remaining recording may be performed. Audio-added data associated with main information recorded on a medium can be recorded and reproduced.
- the recording of the audio additional information may be terminated when the recording of the main information is completed, thereby saving the user the trouble of ending the recording of the audio additional data when the recording of the material data is completed.
- the additional audio information may be recorded at a sampling rate or a bit rate different from that of the main information audio data. For example, by recording the additional audio information at a lower rate, the recordable time of the additional audio data is increased. be able to.
- the audio additional information may be recorded in a file format different from the audio data of the main information. By recording the data in the format for general-purpose PCs, it becomes possible to reproduce the audio additional data on the PC.
- an area for recording the audio additional information may be reserved in the recording medium in advance, and the recording of the audio additional data can be ensured even when the free space of the main information is exhausted. It is also possible to record additional audio information while recording main information, while recording is paused, when recording is stopped, during playback, when playback is paused, and when playback is stopped. This makes editing work easier.
- the additional audio information related to the deleted main information may be deleted at the same time, so that unnecessary audio additional data can be prevented from being forgotten to be erased. it can.
- the selected additional audio information is associated.
- the thumbnail image of the main information at the specified frame position may be displayed, which makes it easy to find the necessary audio additional information.
- a thumbnail of the main information associated with the additional audio information or the video information in the main information may be displayed, and the main information can be confirmed during the reproduction of the additional audio information. It becomes.
- the main information may be reproduced from a frame position of the main information associated with the selected additional audio information, and the additional audio information is used as a key. After searching, the associated main information can be checked immediately, which improves editing work efficiency.
- the main information may be reproducible from the frame position of the main information associated with the audio additional information being reproduced. Since information can be checked immediately, editing work efficiency is improved.
- management information on the audio additional information including information indicating a state at the start of recording the audio additional information, and reproducing the audio additional information by various methods with reference to this management information. Becomes possible.
- the audio additional information is recorded in association with the main information asynchronously with the time axis of the main information. This makes it easier to control the reproduction of the audio additional information.
- FIG. 1 is a diagram showing a configuration of a video / audio recording / playback apparatus according to Embodiment 1 of the present invention.
- FIG. 2 is a diagram for explaining association of a voice memo with a specific position in a clip.
- FIG. 3 is a diagram showing an example of management information (voice memo management table) indicating the relationship between a voice memo file and a clip.
- FIG. 4 is a diagram showing an example of management information (clip management table) indicating the relationship between a clip and material files (video and audio files) constituting the clip.
- FIG. 5 is a flowchart of a process for reproducing main data (clip) associated with voice memo data during voice memo reproduction.
- FIG. 6 is a flowchart of a process for reproducing a voice memo associated with a clip during clip reproduction.
- FIG. 7 is a diagram showing a configuration of a video / audio recording / reproducing apparatus having a plurality of recording media according to Embodiment 2 of the present invention.
- FIG. 8 is a diagram for explaining association of a voice memo with a specific position in a shot recorded over a plurality of recording media.
- FIG. 9 is a diagram showing an example of an operation unit in the recording / reproducing device.
- FIG. 10 is a flowchart of the recording operation of the voice memo.
- FIG. 11 is a diagram illustrating a display example of a clip list screen.
- FIG. 12 is a diagram showing a display example of a voice memo tapping list screen.
- FIG. 13 is a flowchart of a voice memo playback operation.
- FIG. 14 is a diagram illustrating a display example of a screen during voice memo playback.
- FIG. 15 is a diagram showing a directory structure of contents in a recording medium.
- FIG. 16 is a diagram for explaining tags for managing clip information.
- Figure 17 is a diagram showing an example of an XML description of a clip file.
- Embodiment 1
- FIG. 1 is a block diagram showing a schematic configuration of a memory recording power mera recorder according to the present invention. '
- the video / audio input unit 100 inputs audio information and video information as main information.
- Video information can be input via an image sensor or a playback device, and audio information can be input via a microphone, a playback device, or the like.
- any means can be used as long as video and audio information can be input.
- “main information” refers to video / audio information to which audio additional information can be added, and is also referred to as “main part”.
- the compression / expansion circuit 101 is capable of compressing the main information of video and audio input from the video / audio input unit 100 and outputting it to the recording / reproducing unit 140 as main data of video and audio data.
- the video / audio main data and the audio additional data reproduced from the recording / reproducing unit 140 are decompressed to the video / audio output unit 102 as video / audio main information and audio additional information. Output.
- the video / audio output unit 102 externally outputs video and audio main information and audio additional information from the compression / expansion circuit 101.
- the voice memo microphone 110 is voice additional information input means, and inputs a voice memo as voice additional information.
- the audio additional information input means instead of providing a microphone in the camera recorder, an audio input terminal may simply be provided, and an input means such as a microphone may be connected and used there.
- the voice memo processing circuit 111 converts the voice additional information input by the voice memo microphone 110, compresses the data, and outputs it to the recording / reproducing unit 140 as voice additional data.
- the control unit 120 controls the operation of each unit such as the recording / reproducing unit 140 and the display unit 121.
- the display unit 122 displays the voice memo number and the thumbnail (representative image) specified by the control unit 120.
- the operation unit 130 has a recording button, a reproduction button, a voice memo reproduction button, and the like, and receives a user's operation from the outside.
- Record The playback unit 140 records the video and audio main data from the compression / decompression circuit 101 and the voice additional data of the voice memo processing circuit 111 on the recording medium 150, and records the data.
- the video and audio main data and audio additional data reproduced from the medium 150 are output to the compression / decompression circuit 101.
- the recording medium 150 is a random-accessible recording medium for recording video and audio main data and audio additional data from the recording / reproducing unit 140.
- the recording medium 150 may be any type of recording medium as long as it is a randomly accessible recording medium, and may be plural without any limitation such as a built-in type, an external type, and a removable type.
- a hard disk, an optical disk, a magneto-optical disk, a semiconductor memory, and the like can be considered. In this embodiment, it is assumed that there is only one recording medium.
- the clip is composed of one material file, but the main video data and main audio data are different.
- a clip may be composed of a plurality of material files.
- the main video data and the main audio data are recorded on the recording medium 150 as separate files.
- the main video data is composed of one video file, and It is assumed that the main data is composed of audio files of multiple channels.
- the main video data is simply referred to as “video data”
- the main audio data is simply referred to as “audio data”.
- the voice information input from the voice memo microphone 110 is converted into voice additional data by the voice memo processing circuit 111 and output.
- This additional voice data is called "voice memo data”.
- the voice memo data is recorded in association with the time code in the clip.
- the time code to be associated may be the time code for the first frame in the clip, or May be a time code for an arbitrary frame.
- voice memo data By recording the voice memo data in association with the time code of the clip, it becomes possible to record a plurality of voice memos for one clip. In addition, it is possible to associate the material data with detailed positions in frame units. At the time of editing, listening to the voice memo has an effect that the position of desired material data can be easily found.
- the voice memo data may be associated with the clip frame offset (the number of frames from the beginning) instead of the clip time code.
- the time at the position of the frame offset associated with voice memo # 2 (412) may be earlier than the end time of voice memo # 1 (41 1).
- It is also possible to record another voice memo # 3 (413) in association with the frame (frame offset 8) to which voice memo # 2 (412) is associated.
- the recording time of the voice memo does not directly affect the recording time of the material clips constituting the main data.
- the voice memo is recorded at one point on the frame offset of the associated clip. Therefore, it is also possible to record a voice memo for a long time even with the material clip.
- the upper limit of the recording time of the voice memo is described later, but it must be within the set time.
- the voice memo data may be associated with a specific frame offset value of the clip, for example, the first frame of the clip.
- the voice memo may be defined as being associated with the entire clip.
- a management table (hereinafter referred to as a “voice memo management table”) showing a relationship between a clip and a voice memo file as shown in FIG. It is conceivable to use a management table (hereinafter referred to as “clip management table”) that indicates the relationship between clips and their material files (video and audio data files).
- FIGS. 3 and 4 show a management table in a case where voice memo data is associated with a frame offset of a clip.
- the clip name 200 indicates the ID of the clip. Note that all clips in the same recording medium have unique IDs.
- the frame offset 201 is the number of frames from the beginning of the clip.
- Memo ID 202 is a unique ID attached to multiple voice memos associated with the same tulip.
- Voice memo file name 203 is the file name of the voice memo file, and all voice memo file names within the same clip have unique filenames.
- the AV type 301 is information indicating whether the type of clip (material file) constituting the main data is video data or audio data.
- the channel number 302 specify the channel number for audio data, and do not need to specify it for video data.
- the material file name 303 is a unique file name of video data or audio data as a material file constituting a clip.
- the voice memo file name of the voice memo being played is unique within the same clip. Therefore, refer to the voice memo management tape file 20 and enter the voice memo file name.
- a clip name and a frame offset associated with the clip name are obtained (S11).
- the file names (material file names 303) of all material files constituting the clip of the obtained clip name are acquired (S 12). In other words, the material file names are obtained for the number of files that make up the clip. Playback is started from the position indicated by the previously obtained frame offset for each of the data files having the respective material file names obtained here (S13).
- the clip name of the clip currently being played is obtained (S 21). Then, by referring to the voice memo management table 20, a memo ID associated with the obtained clip name and a voice memo file name corresponding to the memo ID are obtained (S22). The voice memo data indicated by the acquired voice memo file name is reproduced (S23). The specific method of specifying the voice memo to be reproduced will be described later.
- the clip and the voice memo data can be associated with each other using the management information 20 and 30. Also, since the voice memo is associated with the time code / frame offset in the clip, a plurality of voice memo data can be associated with one clip. It is also possible to associate multiple voice memos on the same frame offset of a particular clip.
- the video and audio information is compressed by the compression / decompression circuit 101, and the video and audio data is decompressed.
- the non-compressed data is not compressed and decompressed. Certain video and audio information may be handled as is.
- the management table shown in FIGS. 3 and 4 is used as a means for associating the frame offset in the clip with the pois memo data, but other means may be used as long as the association can be realized. .
- the voice memo is calculated by dividing the frame offset of the clip by the time. If the voice memo is associated with the code, but the voice memo can be associated with a position on the specific time axis within the clip, that is, if the information can specify the frame position within the clip, the voice memo is associated with the clip frame offset. It does not have to be a time code.
- audio data may be additionally recorded by video recording with respect to video and audio data shot in advance, and may be regarded as audio data for video data and reproduced.
- the audio data additionally recorded by post-recording is recorded on the assumption that it is reproduced in synchronization with the video data recorded at the time of the first shooting. Therefore, when additionally recording audio data by after-recording, it is common to reproduce the video data and record the additional audio data while maintaining synchronization.
- the voice memo in the present invention is memo information for indicating what the clip (material data) is, and the synchronous reproduction with the main video and audio data is not essential. Therefore, there is no limitation on the state of the main data when recording the voice memo, and the voice memo is recorded even if the main data is in various states, such as stopped, playing, special playing (multiple search playback, reverse playback, etc.). can do.
- the voice memo is associated with a specific point on the time axis of the main data, and can be recorded without synchronization with the main data.
- the number of additional audio data is limited to the number of audio output channels of the device. For example, for a device that can output only up to four channels of audio, the audio can be recorded on up to four channels.
- the voice memo of the present invention a plurality of voice memos can be recorded in association with the same time axis position of the main data regardless of the number of audio output channels.
- the memory recording power mera recorder has only one recording medium 150, but in the present embodiment, the recording medium 150 is A case will be described in which the recording medium is composed of a plurality of removable recording media (recording medium # 1 (501), recording medium # 2 (502), recording medium # 3 (503)) as shown.
- the recording data unit when video and audio main data are recorded continuously over a plurality of recording media, the recording data unit is called a “shot”. For example, when the material of one shot is recorded on one recording medium, the shot becomes one clip. On the other hand, if the material of one shot is recorded over a plurality of recording media, a separate clip will be created for each recording medium. In this case, the voice memo data is associated with each divided clip.
- shot 600 starts recording from recording medium # 1 (501) and ends recording on recording medium # 3 (503) over recording medium # 2 (502).
- the shot 600 is composed of the clip # 1 (611) in the recording medium # 1 (501), the clip # 2 (612) in the recording medium # 2 (502), and the clip in the recording medium # 3 (503). # 3 (613) recorded.
- the voice memo data when voice memo data is recorded in association with a specific position in the shot 600, the voice memo data is recorded on the same recording medium as the recording medium on which the entity data to which the voice memo data is added is recorded. .
- the voice memo data (voice memo # 1 (621)) is recorded on recording medium # 1 (501).
- the voice memo data is recorded on recording medium # 2 (502) (voice memo # 2 (622)).
- the end time of voice memo # 2 (622) may be later than the end time of clip # 2 (612).
- voice memo # 2 (622) is transferred from recording medium # 2 (502) to recording medium # 3.
- the voice memo data (voice memo # 3 (6 2 3)) is recorded on the recording medium # 3 (503).
- the end time of voice memo # 3 (6 2 3) may be later than the end time of shot 600.
- the voice memo data is recorded on the recording medium in which the main part data of the frame offset of the clip to which the voice memo data is associated is recorded.
- the recording time of the voice memo data must be within the upper limit of the recording time of the voice memo, as in the first embodiment.
- voice memo data is associated with a clip including video or audio data.
- a clip hereinafter, referred to as “dummy clip”
- Voice memo data may be associated with the dummy clip.
- the voice memo data associated with the dummy clip may be associated with the entire recording medium.
- the recording medium can be distinguished from other recording media. This makes it easier to perform the task.
- FIG. 9 shows an example of the operation unit 130 operated by the user when recording and reproducing a voice memo.
- a voice memo recording button 1101 On the operation unit 130, a voice memo recording button 1101, a selection button 1102, and a decision button 111 are provided.
- the voice memo recording button 111 is used when recording a voice memo is started and when recording a voice memo is ended. When the voice memo recording button 111 is pressed while no voice memo is recorded, the voice memo recording operation starts. If the voice memo recording button 111 is pressed while a voice memo is being recorded, the recording operation of the voice memo ends. Note that the voice memo recording button 111 may be divided into a button for starting recording and a button for ending recording.
- the selection button 1102 is, for example, a button for moving a force in a thumbnail (representative image) list of clips and moving a force in various option items.
- the decision button 111 is a button for deciding the selection. For example, when the enter button 1103 is pressed while a certain voice memo is selected, the reproduction of the voice memo is started.
- the operation unit 130 may include various buttons (not shown).
- the voice memo recording operation starts.
- the clip name and the frame offset of the clip that associates the voice memo to be recorded are obtained, and the information is stored (S31).
- the specific method of determining the clip-to-frame offset for associating a voice memo will be described later (see Embodiment 5).
- the memo ID and the file name are determined and stored so as not to be duplicated in the same clip by referring to the management tables 20 and 30 (S32).
- the method of determining the file name will be described later (see Embodiment 7).
- recording of the voice memo is started (S33).
- the voice memo is memo information indicating what the material data is, and is rarely edited after recording. Also, unlike voice data of material, voice memos do not require high sound quality. Therefore, the sampling rate and bit rate of the voice memo are recorded at a lower rate than the audio data of the material. This reduces the file size of voice memos and makes them more efficient.
- recording is performed with the sampling rate of the audio data of the material set to 48 kHz and the sampling rate of the voice memo set to 8 kHz.
- bit rate of the audio data of the material is set to 16 bps (Bits Persample), and the bit rate of the voice memo is set to 8 bps.
- the voice memo can be recorded in the size of the audio data of the material in the size of 112, so that it is possible to record more video and audio data of the material on a limited-capacity recording medium. Become.
- the file format of the voice memo can be different from the file format of the audio data of the material.
- WAVE used on a general-purpose PC is used as the format of MXF (Material Exchange Format) s voice memo, which is a format for material exchange.
- MXF Media Exchange Format
- a material exchange format such as MX F facilitates editing on an editing machine, etc., and increases editing efficiency.
- a voice memo uses a format that can be handled by a general-purpose PC, such as WAVE, thumbnails (representative images), etc., described later, can be used to easily edit titles and other data on a PC without having to look at actual material data. Is obtained.
- the following is a specific example of recording material data and voice memos on a recording medium.
- the parameters of each data are as follows, assuming that the sound quality of the voice memo is set slightly higher.
- Frame G: 30 fps (Frames PerSecnd), Frame size of main video data: 120 kB,
- Bit rate of main audio data 16 bps
- the clip is composed of video data 1 ch and audio data 2 ch.
- the data size per second of the clip is
- an area dedicated to voice memo recording may be reserved in advance on the recording medium.
- the recording capacity (7.2 MB) required to record a voice memo for 5 minutes (300 seconds) is equivalent to the recording capacity for recording a clip for about 1.9 seconds (about 57 frames).
- voice memo # 2 can be recorded on recording medium # 2 (502).
- the maximum recordable time of the voice memo is set to 5 minutes.
- the force may not be set to 5 minutes, and may be set arbitrarily by the user.
- the capacity of the voice memo recording dedicated area is set by the voice memo recording time, the ratio of the voice memo recording area to the total capacity of the recording medium may be set. Further, the capacity to be secured may be directly set in units of bits.
- the voice memo starts recording while recording the main information of video and audio, recording pause, recording stop, playback, playback pause, playback stop. be able to. Since voice memos can be recorded from a plurality of states, the effect of increasing the convenience of voice memo recording is obtained. The method of recording voice memos in each state is described below.
- the method of recording a voice memo while recording the main information will be described.
- the voice signal input from the voice memo microphone 110 is processed by the voice memo processing circuit 111.
- the data is converted and recorded on the recording medium 150 as a voice memo.
- the voice memo is associated with the frame offset of the clip recorded when the voice memo button is pressed. This association is performed by registering and updating management information.
- the voice memo continues to be recorded on the recording medium at the start of recording of the voice memo.
- the voice memo can be reproduced.
- the voice memo button When the recording operation is stopped during the recording operation of the main information and the voice memo button is pressed, the audio signal input from the voice memo microphone 110 is subjected to data conversion and recorded as a voice memo. The voice memo is then linked to the last shot recorded. When a shot is recorded in multiple clips, The text is associated with the entire last recorded clip. As a result, the voice memo can be recorded after the main information recording (video shooting), so that the user can concentrate on recording the main information (video shooting) during the recording.
- the voice memo button is pressed while the recording of the main information is stopped, the voice memo may be recorded in association with the next shot to be taken. At that time, a dummy clip is created temporarily, and a voice memo is associated with the entire dummy clip. Then, when shooting is resumed, the recorded voice memo is re-associated with the clip being shot, and the dummy clip is deleted. If the next shot does not start, delete the recorded voice memo.
- a voice memo can be recorded before the main information is recorded, so that the user can concentrate on shooting video during recording.
- Whether to record the voice memo after recording the main information or whether to record the voice memo before recording the main information may be set so that the setting can be changed according to the use of the user.
- the voice memo button When the voice memo button is pressed during the reproduction of the main information, the voice input from the voice memo microphone 110 is converted into data and recorded on the recording medium as a voice memo. At this time, this voice memo is associated with the frame offset of the clip being played at the time when the voice memo button is pressed. According to this method, after recording the main information, the voice memo can be associated while confirming the video of the main information, so that it can be associated with a more accurate position of a predetermined scene.
- the voice memo button When the voice memo button is pressed while the main information is being reproduced, the voice input from the voice memo microphone 110 is converted into data and recorded as a voice memo. At this time, the voice memo is associated with the frame offset of the clip at the position in the clip where playback is paused. According to this method, as in the case where the voice memo button is pressed during playback, the voice memo can be associated while confirming the main information after recording, so that it can be associated with a more accurate position in a predetermined scene. can get.
- the stop position will be If it is in the middle of a voice memo, the voice signal input from the voice memo microphone 110 is data converted and recorded as a voice memo. The voice memo is then associated with the entire shot. If the shot is recorded in multiple clips, it is associated with the entire clip including the stop position. According to this method, a voice memo can be recorded in association with a shot or an entire clip during editing, and an effect that retrieval of a clip unit using a voice memo as a key is facilitated is obtained.
- the sampling rates of the audio data and the voice memo of the main information are set to 48 kHz and 12 kHz, respectively, but the respective values may be different values.
- the bit rates of the main information voice data and voice memo are both 16 bps, but each value may be different.
- a common sampling rate or common bit is used for voice data of main information and voice memo.
- a rate may be used, and the magnitude relation does not matter.
- MXF and WAVE are used as the format of the voice data and voice memo of the main information, other formats may be used.
- a common format may be used for the voice data of the main information and the voice memo for reasons such as simplifying the control.
- the clip is composed of one channel of video data and two channels of audio data, but the number of channels may be arbitrary. For example, a clip of only one channel of audio data may be used. Is also good.
- FIG. 11 shows an example of the clip list screen displayed on the display unit 12 1.
- the clip list screen displays a list of clips recorded in the recording medium 150. If all the clips do not fit on the screen, scroll the screen using the selection cursor 1102 and display them.
- thumbnails depictative images of clips of each recorded clip are displayed side by side.
- the thumbnail 1402 may be video data of the first frame of the clip or video data of another frame in the clip.
- an image such as a blue background is embedded in the thumbnail 1402. It should be noted that a different image other than the video data in the clip may be set as the thumbnail 1442 by the user.
- the clip number 1 4 0 3 of the clip is also displayed together with the thumbnail 1 4 0 2.
- the clip number 1403 can be determined independently of the clip name, and may be freely set as long as it is unique within the recording medium.
- a voice memo mark 144 is displayed for a clip to which a voice memo is associated.
- the voice memo is associated with the clips having the clip numbers “0 2”, “0 3 J”, and “0 5”.
- a display 144 indicating that the thumbnail is selected is added to the outer frame.
- FIG. 12 is a diagram showing an example of the voice memo clip list screen.
- the voice memo clip list screen is derived from the clip list screen.
- the voice memo clip list screen displays a list of only the clips associated with the voice memo among the clips recorded on the recording medium.
- the transition to the voice memo clip list screen can be performed by using an option button on the operation unit 130, and the like.
- the means does not matter.
- the voice memo clip list screen has a voice memo display area 1502 and a clip display area 1504.
- the clip display area 1504 is an area for displaying a thumbnail (hereinafter referred to as “clip 'thumbnail J”) 1402 of the clip to which the voice memo is associated.
- voice memo display area 1502 a list of thumbnails related to the voice memo associated with the currently selected clip (hereinafter referred to as “voice memo 'thumbnail') 1501 is displayed.
- the voice memo .Thumbnail 1501 is a reduced image of the still image at the position in the clip with which the voice memo is associated. If there is no video data in the associated clip, that is, if the clip is composed of only audio data, an image such as a blueback is embedded in the voice memo / thumbnail 1501.
- the voice memo number 1503 is displayed in the thumbnail 1501.
- the voice memo number 1503 can be determined independently of MemoID985 described later, and the voice memo number 1503 may be set freely as long as it is a uke in a clip.
- voice memo information 1505 displaying information of the currently selected voice memo is displayed. For example, the clip number 1443 associated with the currently selected voice memo and the voice memo number 1503 of the currently selected voice memo are displayed. In the example of FIG. 12, the voice memo information 1505 indicates that the voice memo with the clip number “0 2” and the voice memo number “0 2” is selected. In this example, it is shown that a total of three voice memos are associated with the clip having the clip number “02”.
- the information displayed in the voice memo information 1505 may not be displayed unless it is necessary, and other information may be displayed if necessary.
- the user moves to the voice memo clip list screen to play the voice memo, and selects and determines the clip associated with the voice memo to be played. Selection of the tap to be played is performed by the selection button 1 1 0 2 on the operation unit 1 3 0, and the decision button 1 The selection is determined by 103.
- a tap is selected / determined by a user operation on the voice memo clip list screen (S41).
- the cursor is moved to the voice memo display area 1502, and the voice memo to be reproduced by the user in the voice memo display area 1502 is determined whether or not the thumbnail is selected.
- FIG. 14 is a diagram showing a screen during voice memo playback.
- a display 1601 indicating that the voice memo is being played is displayed together with the still image of the video data of the main part.
- the display 1601 may be a blinking display.
- the playback of the moving image of the main video data may be started from the position associated with the voice memo to be played.
- the still image or the blue-back image of the last frame of the main part of the video data may be continuously output.
- the display automatically returns to the voice memo clip list screen.
- a predetermined button for example, a decision button 1103, a stop button (not shown)
- the voice memo playback operation may be interrupted when is pressed.
- the main play button (not shown) or the selection decision button 1103 is pressed during the voice memo reproduction, the reproduction of the voice memo is interrupted, and the main memo is reproduced from the position associated with the reproduced voice memo.
- the reproduction of the video and audio data may be started.
- the tuple frame offset and the voice memo data are related.
- a management table as shown in FIGS. 2 and 3 is used.
- information relating to voice memo association is described in an XML (extensible Marque Language: W3C recommendation) file.
- various information about clips such as information about video data and audio data that make up clips, is described in XML files.
- FIG. 15 shows an example of a directory structure of each content recorded on the recording medium 150.
- the Connts directory 800 is located on the recording medium 150 under the root directory. All the files that make up the clip are located under the Connects directory 800.
- the C1ip directory 810 is located under the Connects directory 800.
- the XML file describing the clip information is stored under the Cip directory 810.
- the Video directory 820 is located under the Content directory 800.
- the video data file is stored under the Video directory 820.
- the Au d i o directory 830 is located under the Connects directory 800.
- the audio data file is stored under the audio directory 830.
- the Voice directory 850 is located below the Connect directory 800.
- the voice memo data file is stored under the Voice directory 850.
- Each of the clip files 811 and 812 is an XML file in which all types of tap information such as voice memo additional information are described.
- One tuple file is created for one tulip.
- Video files 821 and 822 are the video data that make up the clip, respectively. It is Huai Nore.
- Each of the audio files 831-834 is an audio data file constituting a clip.
- Voice memo files 51 to 853 are voice memo data files associated with clips.
- Figure 16 shows the items that are specifically described in the XML description.
- the items described in FIG. 16 are examples enumerated for describing the present embodiment, and items not described in FIG. 16 may exist. May not be present. Each item may have its attribute.
- the C1ipContnt tag 900 has the following information about clips as elements.
- the C 1 ip N ame tag 901 has a name of a tulip as an element.
- the Duratation tag 902 has the number of frames of the tap as an element.
- the Ess enc eList tag 910 has a list of essences such as video and audio data as elements.
- the Video tag 920 has the following video data information as elements. Also, by adding, for example, an attribute (not shown), such as Va 1 id Audio F 1 ag, as an attribute of the video tag 920, it is determined whether or not the audio data is multiplexed in the video data. May be.
- an attribute such as Va 1 id Audio F 1 ag
- the Video Format tag 921 has a final format of video data as an element.
- the MXF file format is conceivable, but other formats may be used.
- 11 (1 1 (> tag 940) has the following audio data information as elements.
- the Au dioFormat tag 941 has a file format of audio data as an element.
- the MXF file format is conceivable, but other formats may be used.
- the SamlinGrate tag 942 has a sampling rate of audio data as an element. For example, 4800 OHz is conceivable, but any value can be used for the sampling rate.
- the BitsPerSample tag 943 has the bit rate of audio data as an element. For example, 16 bps, 24 bs, etc. can be considered, but the value of the bit rate is not limited.
- the C 1 i pMeta d a ta tag 960 has, as an element, information of metadata other than the material data associated with the tap, such as a voice memo.
- the MemoList tag 970 has a list of memos to be associated with the tulip as elements. If the memo does not exist, the Memo List tag 970 may be omitted.
- the Memo tag 980 has the following pieces of memo information as elements. Note that Memo ID 985 is added as an attribute to the Memo tag 980. Memo ID 985 is a 2-digit value that is independent for each tulip, and up to 100 memos can be associated with each tulip. Also, Memo ID 985 need not be a two-digit value, and the maximum number of notes associated with each clip need not be 100.
- the Offset tag 981 has, as an element, the frame offset of the clip to which the memo is associated.
- the Offset tag 981 may not be necessary if necessary. Without the Offset tag 981, the note should be associated with the entire clip.
- the Perso 11 tag 982 has, as elements, the name of the person who created the memo. For example, when recording a voice memo, the person who recorded the voice memo is described in the Person tag 982. This clarifies the person who recorded the voice memo, and if you want to know the situation when the voice memo was recorded, you can contact that person. In addition, it is not necessary to add the Person tag 982 unless necessary.
- the Vo ice tag 990 has the following information of voice memos as elements. If no voice memo is associated, there is no need for the Vo ice tag 990.
- the Voice Format tag 991 has a finale format of voice memo data as an element.
- tfWAVE file format can be considered.
- Other formats may be used.
- the VoiceSam1ingRate tag 992 has a voice memo data sampling rate as an element. For example, 12000 Hz is conceivable, but any value can be used for the sampling rate.
- the Voice BizPerSamle tag 993 has a bit rate of voice memo data as an element. For example, 16 bp s is conceivable, but any value can be used for the bit rate.
- the RecCont i d i on tag 994 has a state at the time of recording the voice memo as an element.
- a PLAY state, a STILL state, and the like are conceivable, but there may be other states, and those states may be subdivided. It should be noted that the RecCondition tag 994 may not be provided if it is not necessary.
- the Rec Condition i on tag 994 is set to "P LAY".
- Rec Condition tag 994 When playing a voice memo, refer to Rec Condition tag 994. If it is "PLAY”, the video data of the main part is started from the position associated with the voice memo. Play them simultaneously. On the other hand, if the Rec Condition tag is 994 power S "ST ILL", the voice memo is reproduced while the still image of the main video data at the position associated with the voice memo is continuously output. In this case, “PLAY” and “ST I LL” are set for the RecCondition tag 994. However, other states may be defined separately. For example, if a voice memo is recorded during search playback, The value indicating the search reproduction speed at that time may be described in the Rec Condition tag 994. In this case, it is also possible to acquire the search playback speed by referring to the Rec Conditio 11 tag 994, and to play back the voice memo while performing search playback of the main data at the acquired speed.
- the voice memo may be played back while the still image data of the main part is continuously output.
- the user may freely determine the relationship between the RecCondition tag 994 and the voice memo reproduction method.
- the Rec Cond i it on tag 994 need not be recorded, and in that case, the method of reproducing the voice memo may be unified.
- the note memo may be reproduced by the reproduction method set by the user.
- the number of voice memo reproducing means can be increased.
- the video and audio when playing back video and audio data of a material, the video and audio must be synchronized. At this time, it is general that the difference between the video and the audio by one frame is not allowed. If the video data and audio data are not multiplexed and the video data file and the audio data file (including multiple channels) are separate files, control to play all these files in synchronization Is complicated. Also, when playing two or more audio data files with different sampling rates, the control of playing them while synchronizing them becomes more complicated. On the other hand, if it is not necessary to play back video and audio while synchronizing them, that is, if a shift of several frames is permissible, even if the video data file and audio data file are separate files, they are independent. Playback is simpler and the control is simpler.
- voice memo data is memo data indicating what the material data is, it is not considered necessary to reproduce it in strict synchronization with the material data. Therefore, if the voice memo is played back without synchronizing with the main data, the control becomes easier.
- the voice memo is associated with a specific point on the time axis of the main data, and by recording the voice memo without synchronizing with the main data, the voice memo can be recorded for a longer time than the duration of the associated clip. I can do it. For example, for a clip of several seconds, a voice memo of several tens of seconds can be recorded.
- voice memos can be recorded in various states, such as during stoppage, during playback, during special playback, etc. (multiple search playback, reverse playback, etc.).
- the value of the Rec Condition tag 994 may be set to "STI LL" as the voice memo management information. The method will be described later.)
- a voice memo can be recorded in advance before recording the material. For example, for a scene to be shot from now, a description of the scene is recorded in advance as a voice memo, then the target scene is shot, and the previously recorded voice memo is associated with the clip. You can also. Also, when recording a voice memo during recording and playback of the main part data, the value of the Rec Condition tag 994 may be set to “P LAY”.
- the voice memo playback method at this time will be described later.
- the voice memo to be recorded does not necessarily have to be synchronized with the main data. Therefore, as shown in the second embodiment, even when the main data is recorded over a plurality of recording media, or even when the main data is recorded, the voice memo can be recorded without straddling the plurality of recording media. Also, especially when recording a voice memo during playback of the main data, it is possible to continue recording the voice memo during the voice memo recording even after the end of the main data has passed.
- the value of the Rec Condition tag 994 should be set to a value that indicates the status of each. Les ,.
- the playback method may be selected by referring to the RecCondition tag 994 added at the time of recording the voice memo.
- the value of the Rec Condition tag 994 is “PL AY”, that is, when the state of the main part at the time of recording the voice memo is recording or reproducing, the position of the voice memo associated with the voice memo is set.
- the main video data is reproduced at the same time.
- the main data and the voice memo do not always need to be synchronized, they can be reproduced with simpler control. If you want to listen to the voice memo quickly with a long-time voice memo, you can play only the voice memo at 1.5 or 2x speed while playing the main part at the same magnification. Conversely, if the voice message left in the voice memo is too fast to be heard, etc., only the voice memo may be played back at 0.5x speed while the main part is being played at 1: 1.
- the Rec Condition tag 994 has a value indicating search playback, etc., for example, a value indicating search playback at 4x speed
- the main data is searched and played back at 4x speed from the position where the voice memo is associated. You can play voice memos.
- the RecCondition tag 994 indicates a reverse playback, the voice memo can be played back while the main data is played back from the position where the voice memo is associated.
- FIG. 17 shows an example of an XML description for part of the directory structure in Figure 15.
- FIG. 15 includes a clip file # 1 (81 1) whose clip name is "0001 AB" and a clip file # 2 (812) whose clip name is S "0001 CD”.
- FIG. 17 shows an XML description of the clip file # 1 (811).
- FIG. 17 shows a part of the contents described in the clip file # 1 (81 1), and the described items are only those necessary for describing the present embodiment. Items not described in FIG. 17 may exist, and some of the items illustrated in FIG. 17 may not exist. Each item may have its attribute.
- the clip name (ClipNarae) of clip file # 1 (811) is 0001 AB ".
- the length (Duration) of clip file # 1 (811) is 1000 frames.
- File format of voice memo using MX F as file format (Video Format, AudioFormat) of main video data and main audio data
- the bit rates (BitsPerSample, VoiceBitsPerSample) of voice data and voice memo are both 16 bps.
- Clip file # 1 (811) is video file # 1 (821), audio file # 1 (831), audio file # 2 (832), voice memo file # 1 (851), and voice memo file # 2 ( 852).
- the voice memo file # 1 (851) has been created by a user having "User Name 1" as the user name (Person).
- the voice memo is in one of the following states: material recording is paused, recording is stopped, playback is paused, or playback is stopped.
- Voice memo file # 2 uses "U ser Name 2".
- Each data file is named as follows.
- the file name of clip file # 1 (811) is clip name "0001 A
- the file name of the video file # 1 (821) is obtained by adding the extension ".mxf ,,” to the clip name "0001 AB” and becomes "0001 AB.mxf".
- the file names of audio file # 1 (831) and audio file # 2 (832) are further extended to clip name "0001 AB” with 2-digit channel numbers "00" and "01” It is obtained by adding the child “.mxf", which becomes “OOO IABOO.mxf” and "OOO IABO I.mxf", respectively.
- the channel numbers of the audio data are assigned as follows: 0 channel, 1 channel, 2 channels,... ′ In the order of the list of elements of the Audio tag 940 registered in the EssecceList tag 910.
- the channel number of the audio data may be determined by adding a channel number as an attribute of the audio tag 940 shown in FIG. 9 and determined from the value. Alternatively, the channel information may be obtained from another tag.
- the means does not matter.
- voice memo file # 1 (851) and voice memo file # 2 (852) is the clip name "0001 AB" with the two-digit value "00" and "01" of each Memo ID 985 added.
- the extension ". Wa V” is added, and they become “OOO IABO O. wa v” and “OOO IAB O I. wa v”, respectively.
- the above-mentioned files are stored based on the directory structure shown in FIG. With this configuration, related information such as material data and voice memo data constituting the clip can be known only by referring to the clip file # 1 (811).
- the file name of each file may be determined by a method other than the above example. In this embodiment, only the method for associating a clip with video, audio data, and voice memo data and items necessary for explaining the effects thereof have been described. However, detailed information of each material data and a representative image of the clip are used. Various information such as information on a certain thumbnail file, information on a shooting location, user information of a photographer, and information on a shooting device may be described in a clip file. By doing so, you can get all the information of a clip just by referring to the clip file.
- XML is used as the description language of the clip file.
- XML is a language that is standardized (recommended) by the World Wide Web Consortium (W3C).
- W3C World Wide Web Consortium
- management information can be transferred to another database. Operations such as migration can also be performed, which is expected to increase versatility.
- management information can be easily added by defining a new tag, so it is considered to be highly extensible.
- XML is a text file, a user can directly and easily refer to a clip file using a general-purpose information device, and can roughly grasp clip information.
- the user can directly edit the clip file using a general-purpose information device, and can perform simple editing. Industrial applicability
- the present invention is useful for a video / audio recording / reproducing apparatus for efficiently performing an editing operation such as a non-linear editing based on a collected media captured and recorded by a memory recording camera recorder or the like.
- an editing operation such as a non-linear editing based on a collected media captured and recorded by a memory recording camera recorder or the like.
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Television Signal Processing For Recording (AREA)
- Management Or Editing Of Information On Record Carriers (AREA)
- Signal Processing For Digital Recording And Reproducing (AREA)
- Indexing, Searching, Synchronizing, And The Amount Of Synchronization Travel Of Record Carriers (AREA)
Abstract
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP04746958A EP1677531A4 (en) | 2003-10-16 | 2004-06-29 | VIDEO / AUDIO RECORDING / PLAYBACK, VIDEO / AUDIO RECORDING AND PLAYBACK METHOD |
JP2005514705A JP4271195B2 (ja) | 2003-10-16 | 2004-06-29 | 映像音声記録再生装置、映像音声記録方法及び映像音声再生方法 |
CA002542390A CA2542390A1 (en) | 2003-10-16 | 2004-06-29 | Audio and video recording and reproducing apparatus, audio and video recording method, and audio and video reproducing method |
US10/575,715 US20070127888A1 (en) | 2003-10-16 | 2004-06-29 | Audio and video recording and reproducing apparatus, audio and video recording method, and audio and video reproducing method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2003356079 | 2003-10-16 | ||
JP2003-356079 | 2003-10-16 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2005039175A1 true WO2005039175A1 (ja) | 2005-04-28 |
Family
ID=34463188
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2004/009489 WO2005039175A1 (ja) | 2003-10-16 | 2004-06-29 | 映像音声記録再生装置、映像音声記録方法及び映像音声再生方法 |
Country Status (7)
Country | Link |
---|---|
US (1) | US20070127888A1 (ja) |
EP (1) | EP1677531A4 (ja) |
JP (1) | JP4271195B2 (ja) |
KR (1) | KR100801396B1 (ja) |
CN (1) | CN100484222C (ja) |
CA (1) | CA2542390A1 (ja) |
WO (1) | WO2005039175A1 (ja) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN100409681C (zh) * | 2005-08-19 | 2008-08-06 | 上海晨兴电子科技有限公司 | 影音同步录制及播放方法 |
JP2008219345A (ja) * | 2007-03-02 | 2008-09-18 | Xing Inc | 合成装置、合成システム、合成方法及びプログラム |
JP2011090751A (ja) * | 2009-10-24 | 2011-05-06 | Korg Inc | 多重録音装置 |
WO2016151994A1 (ja) * | 2015-03-23 | 2016-09-29 | パナソニックIpマネジメント株式会社 | ウェアラブルカメラ及びウェアラブルカメラシステム |
Families Citing this family (134)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8645137B2 (en) | 2000-03-16 | 2014-02-04 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US8677377B2 (en) | 2005-09-08 | 2014-03-18 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US8301995B2 (en) * | 2006-06-22 | 2012-10-30 | Csr Technology Inc. | Labeling and sorting items of digital data by use of attached annotations |
KR100856407B1 (ko) | 2006-07-06 | 2008-09-04 | 삼성전자주식회사 | 메타 데이터를 생성하는 데이터 기록 및 재생 장치 및 방법 |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
KR101309284B1 (ko) | 2006-12-05 | 2013-09-16 | 삼성전자주식회사 | 오디오 사용자 인터페이스 처리 방법 및 장치 및 그를적용한 오디오 기기 |
CN101202873B (zh) * | 2006-12-13 | 2012-07-25 | 株式会社日立制作所 | 信息记录再现装置和信息记录再现方法 |
US8553856B2 (en) * | 2007-01-07 | 2013-10-08 | Apple Inc. | Voicemail systems and methods |
US20080167009A1 (en) * | 2007-01-07 | 2008-07-10 | Gregory Novick | Voicemail Systems and Methods |
KR101125286B1 (ko) * | 2007-03-27 | 2012-03-21 | 삼성전자주식회사 | 부가 데이터 업데이트 방법 및 재생 장치 |
US8977255B2 (en) | 2007-04-03 | 2015-03-10 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US20090062944A1 (en) * | 2007-09-04 | 2009-03-05 | Apple Inc. | Modifying media files |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US8996376B2 (en) | 2008-04-05 | 2015-03-31 | Apple Inc. | Intelligent text-to-speech conversion |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US8682848B2 (en) * | 2008-07-08 | 2014-03-25 | Broadcom Corporation | Method and system for automatic detection of multimedia settings |
US20100030549A1 (en) | 2008-07-31 | 2010-02-04 | Lee Michael M | Mobile device having human language translation capability with positional feedback |
WO2010067118A1 (en) | 2008-12-11 | 2010-06-17 | Novauris Technologies Limited | Speech recognition involving a mobile device |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US10255566B2 (en) | 2011-06-03 | 2019-04-09 | Apple Inc. | Generating and processing task items that represent tasks to perform |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US9431006B2 (en) | 2009-07-02 | 2016-08-30 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
KR101641240B1 (ko) * | 2009-12-17 | 2016-07-21 | 엘지전자 주식회사 | 영상 표시 방법 및 그 장치 |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US8682667B2 (en) | 2010-02-25 | 2014-03-25 | Apple Inc. | User profiling for selecting user specific voice input processing information |
CN102013179A (zh) * | 2010-12-02 | 2011-04-13 | 无敌科技(西安)有限公司 | 同时同步录播方式的电子学习系统及方法 |
US10762293B2 (en) | 2010-12-22 | 2020-09-01 | Apple Inc. | Using parts-of-speech tagging and named entity recognition for spelling correction |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US8994660B2 (en) | 2011-08-29 | 2015-03-31 | Apple Inc. | Text correction processing |
JP2013055569A (ja) * | 2011-09-06 | 2013-03-21 | Sony Corp | 撮像装置、情報処理装置、それらの制御方法、および、プログラム |
KR101356006B1 (ko) * | 2012-02-06 | 2014-02-12 | 한국과학기술원 | 구간설정이 가능한 음성기반 멀티미디어 컨텐츠 태깅 방법 및 장치 |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
KR102042265B1 (ko) * | 2012-03-30 | 2019-11-08 | 엘지전자 주식회사 | 이동 단말기 |
CN102682820B (zh) * | 2012-05-03 | 2015-01-28 | 福建星网视易信息系统有限公司 | 嵌入式linux平台媒体播放方法和系统 |
US9280610B2 (en) | 2012-05-14 | 2016-03-08 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US9721563B2 (en) | 2012-06-08 | 2017-08-01 | Apple Inc. | Name recognition system |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9576574B2 (en) | 2012-09-10 | 2017-02-21 | Apple Inc. | Context-sensitive handling of interruptions by intelligent digital assistant |
US9547647B2 (en) | 2012-09-19 | 2017-01-17 | Apple Inc. | Voice-based media searching |
KR101943320B1 (ko) * | 2012-09-21 | 2019-04-17 | 엘지전자 주식회사 | 이동단말기 및 그 제어방법 |
KR20150104615A (ko) | 2013-02-07 | 2015-09-15 | 애플 인크. | 디지털 어시스턴트를 위한 음성 트리거 |
US9368114B2 (en) | 2013-03-14 | 2016-06-14 | Apple Inc. | Context-sensitive handling of interruptions |
WO2014144579A1 (en) | 2013-03-15 | 2014-09-18 | Apple Inc. | System and method for updating an adaptive speech recognition model |
KR101759009B1 (ko) | 2013-03-15 | 2017-07-17 | 애플 인크. | 적어도 부분적인 보이스 커맨드 시스템을 트레이닝시키는 것 |
WO2014197334A2 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
WO2014197336A1 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
WO2014197335A1 (en) | 2013-06-08 | 2014-12-11 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
CN105264524B (zh) | 2013-06-09 | 2019-08-02 | 苹果公司 | 用于实现跨数字助理的两个或更多个实例的会话持续性的设备、方法、和图形用户界面 |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
CN105265005B (zh) | 2013-06-13 | 2019-09-17 | 苹果公司 | 用于由语音命令发起的紧急呼叫的系统和方法 |
CN105453026A (zh) | 2013-08-06 | 2016-03-30 | 苹果公司 | 基于来自远程设备的活动自动激活智能响应 |
US9620105B2 (en) | 2014-05-15 | 2017-04-11 | Apple Inc. | Analyzing audio input for efficient speech and music recognition |
US10592095B2 (en) | 2014-05-23 | 2020-03-17 | Apple Inc. | Instantaneous speaking of content on touch devices |
US9502031B2 (en) | 2014-05-27 | 2016-11-22 | Apple Inc. | Method for supporting dynamic grammars in WFST-based ASR |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
EP3149728B1 (en) | 2014-05-30 | 2019-01-16 | Apple Inc. | Multi-command single utterance input method |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US10289433B2 (en) | 2014-05-30 | 2019-05-14 | Apple Inc. | Domain specific language for encoding assistant dialog |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US9734193B2 (en) | 2014-05-30 | 2017-08-15 | Apple Inc. | Determining domain salience ranking from ambiguous words in natural speech |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US9711141B2 (en) | 2014-12-09 | 2017-07-18 | Apple Inc. | Disambiguating heteronyms in speech synthesis |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US9578173B2 (en) | 2015-06-05 | 2017-02-21 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
DK179309B1 (en) | 2016-06-09 | 2018-04-23 | Apple Inc | Intelligent automated assistant in a home environment |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10586535B2 (en) | 2016-06-10 | 2020-03-10 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
DK179415B1 (en) | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
DK179343B1 (en) | 2016-06-11 | 2018-05-14 | Apple Inc | Intelligent task discovery |
DK179049B1 (en) | 2016-06-11 | 2017-09-18 | Apple Inc | Data driven natural language event detection and classification |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US11281993B2 (en) | 2016-12-05 | 2022-03-22 | Apple Inc. | Model and ensemble compression for metric learning |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
DK201770383A1 (en) | 2017-05-09 | 2018-12-14 | Apple Inc. | USER INTERFACE FOR CORRECTING RECOGNITION ERRORS |
DK201770439A1 (en) | 2017-05-11 | 2018-12-13 | Apple Inc. | Offline personal assistant |
DK201770429A1 (en) | 2017-05-12 | 2018-12-14 | Apple Inc. | LOW-LATENCY INTELLIGENT AUTOMATED ASSISTANT |
DK179745B1 (en) | 2017-05-12 | 2019-05-01 | Apple Inc. | SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT |
DK179496B1 (en) | 2017-05-12 | 2019-01-15 | Apple Inc. | USER-SPECIFIC Acoustic Models |
DK201770431A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
DK201770432A1 (en) | 2017-05-15 | 2018-12-21 | Apple Inc. | Hierarchical belief states for digital assistants |
DK179549B1 (en) | 2017-05-16 | 2019-02-12 | Apple Inc. | FAR-FIELD EXTENSION FOR DIGITAL ASSISTANT SERVICES |
CN110740275B (zh) * | 2019-10-30 | 2022-07-19 | 中央电视台 | 一种非线性编辑系统 |
IL299335A (en) * | 2020-06-30 | 2023-02-01 | Seff Tech Corporation | A system and method for managing digital information |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001136482A (ja) * | 1999-11-09 | 2001-05-18 | Matsushita Electric Ind Co Ltd | 映像音声記録再生装置 |
JP2001285780A (ja) * | 2000-03-28 | 2001-10-12 | Pioneer Electronic Corp | 映像編集方法及び装置、並びにそのための記憶媒体 |
JP2003158697A (ja) * | 2001-11-21 | 2003-05-30 | Fuji Photo Film Co Ltd | 撮像装置及びプログラム |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3528214B2 (ja) * | 1993-10-21 | 2004-05-17 | 株式会社日立製作所 | 画像表示方法及び装置 |
US5600775A (en) * | 1994-08-26 | 1997-02-04 | Emotion, Inc. | Method and apparatus for annotating full motion video and other indexed data structures |
US6006241A (en) * | 1997-03-14 | 1999-12-21 | Microsoft Corporation | Production of a video stream with synchronized annotations over a computer network |
US6529920B1 (en) * | 1999-03-05 | 2003-03-04 | Audiovelocity, Inc. | Multimedia linking device and method |
US6378132B1 (en) * | 1999-05-20 | 2002-04-23 | Avid Sports, Llc | Signal capture and distribution system |
GB2366926A (en) * | 2000-09-06 | 2002-03-20 | Sony Uk Ltd | Combining material and data |
US20020089519A1 (en) * | 2001-01-05 | 2002-07-11 | Vm Labs, Inc. | Systems and methods for creating an annotated media presentation |
US20040216173A1 (en) * | 2003-04-11 | 2004-10-28 | Peter Horoszowski | Video archiving and processing method and apparatus |
-
2004
- 2004-06-29 US US10/575,715 patent/US20070127888A1/en not_active Abandoned
- 2004-06-29 WO PCT/JP2004/009489 patent/WO2005039175A1/ja active Application Filing
- 2004-06-29 CN CNB2004800304181A patent/CN100484222C/zh not_active Expired - Fee Related
- 2004-06-29 CA CA002542390A patent/CA2542390A1/en not_active Withdrawn
- 2004-06-29 JP JP2005514705A patent/JP4271195B2/ja not_active Expired - Fee Related
- 2004-06-29 KR KR1020067007101A patent/KR100801396B1/ko not_active IP Right Cessation
- 2004-06-29 EP EP04746958A patent/EP1677531A4/en not_active Withdrawn
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001136482A (ja) * | 1999-11-09 | 2001-05-18 | Matsushita Electric Ind Co Ltd | 映像音声記録再生装置 |
JP2001285780A (ja) * | 2000-03-28 | 2001-10-12 | Pioneer Electronic Corp | 映像編集方法及び装置、並びにそのための記憶媒体 |
JP2003158697A (ja) * | 2001-11-21 | 2003-05-30 | Fuji Photo Film Co Ltd | 撮像装置及びプログラム |
Non-Patent Citations (1)
Title |
---|
See also references of EP1677531A4 * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN100409681C (zh) * | 2005-08-19 | 2008-08-06 | 上海晨兴电子科技有限公司 | 影音同步录制及播放方法 |
JP2008219345A (ja) * | 2007-03-02 | 2008-09-18 | Xing Inc | 合成装置、合成システム、合成方法及びプログラム |
JP2011090751A (ja) * | 2009-10-24 | 2011-05-06 | Korg Inc | 多重録音装置 |
WO2016151994A1 (ja) * | 2015-03-23 | 2016-09-29 | パナソニックIpマネジメント株式会社 | ウェアラブルカメラ及びウェアラブルカメラシステム |
Also Published As
Publication number | Publication date |
---|---|
JP4271195B2 (ja) | 2009-06-03 |
JPWO2005039175A1 (ja) | 2007-02-08 |
US20070127888A1 (en) | 2007-06-07 |
CN1868209A (zh) | 2006-11-22 |
KR100801396B1 (ko) | 2008-02-05 |
CA2542390A1 (en) | 2005-04-28 |
EP1677531A4 (en) | 2009-03-04 |
CN100484222C (zh) | 2009-04-29 |
KR20060096026A (ko) | 2006-09-05 |
EP1677531A1 (en) | 2006-07-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2005039175A1 (ja) | 映像音声記録再生装置、映像音声記録方法及び映像音声再生方法 | |
KR100387387B1 (ko) | 콘텐츠를 액세싱하기 위한 방법 및 장치 | |
JPH0944394A (ja) | マルチメディアデータファイリングシステム | |
WO2004021701A1 (ja) | データ処理装置および方法、並びにプログラム | |
WO2004051999A1 (ja) | ビデオ信号処理装置、ビデオ信号記録装置、ビデオ信号再生装置、ビデオ信号処理装置の処理方法、ビデオ信号記録装置の処理方法、ビデオ信号再生装置の処理方法、記録媒体 | |
JPH11205718A (ja) | 情報再生装置及び情報記録再生装置 | |
US20150312638A1 (en) | Method and system for displaying a menu which has an icon and additional information corresponding to stored image data, wherein the icon can display the image data with the additional information | |
EP1059640A2 (en) | Information editing apparatus and information reproducing apparatus | |
JP4218029B2 (ja) | 映像記録装置 | |
WO2004095451A1 (ja) | 入力データ記録装置及び入力データ記録方法 | |
JP4807250B2 (ja) | 記録方法 | |
JP2003244637A (ja) | 再記録可能記録媒体への静止画像記録方法 | |
CN100562938C (zh) | 信息处理设备和方法 | |
JPH0965271A (ja) | 映像信号記録装置、編集方法及びそのシステム | |
JPWO2005081522A1 (ja) | データ処理装置およびデータ処理方法 | |
JP4375052B2 (ja) | 映像編集支援方法 | |
JPH11273227A (ja) | 続き再生時のダイジェスト再生機能付きdvdビデオ再生システム | |
JP4118073B2 (ja) | 画像記録装置 | |
JP2008147727A (ja) | 記録方法 | |
JP3815398B2 (ja) | 編集方法及び編集装置 | |
JP4434633B2 (ja) | 情報処理装置および方法、プログラム記録媒体、並びにプログラム | |
JP2002008353A (ja) | 再生装置、記録再生装置及び記録装置 | |
KR20010038639A (ko) | 개인의 동영상 제작 및 운용시스템 | |
JP2000217055A (ja) | 画像処理装置 | |
JP4784548B2 (ja) | 記録方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200480030418.1 Country of ref document: CN |
|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2005514705 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2542390 Country of ref document: CA |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2007127888 Country of ref document: US Ref document number: 10575715 Country of ref document: US Ref document number: 1020067007101 Country of ref document: KR |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2004746958 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 2004746958 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 10575715 Country of ref document: US |