WO2018078768A1 - コンテンツ再生プログラム及びコンテンツ再生装置 - Google Patents
コンテンツ再生プログラム及びコンテンツ再生装置 Download PDFInfo
- Publication number
- WO2018078768A1 WO2018078768A1 PCT/JP2016/081900 JP2016081900W WO2018078768A1 WO 2018078768 A1 WO2018078768 A1 WO 2018078768A1 JP 2016081900 W JP2016081900 W JP 2016081900W WO 2018078768 A1 WO2018078768 A1 WO 2018078768A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- content
- time axis
- unit
- audio
- data
- Prior art date
Links
- 230000029777 axis specification Effects 0.000 claims abstract description 36
- 238000001514 detection method Methods 0.000 claims abstract description 19
- 238000003860 storage Methods 0.000 description 15
- 238000010586 diagram Methods 0.000 description 7
- 238000013500 data storage Methods 0.000 description 6
- 238000000034 method Methods 0.000 description 6
- 230000006870 function Effects 0.000 description 4
- 238000004519 manufacturing process Methods 0.000 description 2
- 238000005401 electroluminescence Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000002360 preparation method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/4302—Content synchronisation processes, e.g. decoder synchronisation
- H04N21/4307—Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
- H04N21/43079—Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen of additional data with content streams on multiple devices
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
- G11B27/30—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording
- G11B27/3027—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording used signal is digitally coded
- G11B27/3036—Time code signal
- G11B27/3045—Time code signal superimposed on the recorded main signal, e.g. burn-in-time code
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/4302—Content synchronisation processes, e.g. decoder synchronisation
- H04N21/4307—Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/4302—Content synchronisation processes, e.g. decoder synchronisation
- H04N21/4307—Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
- H04N21/43072—Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen of multiple content streams on the same device
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/439—Processing of audio elementary streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/439—Processing of audio elementary streams
- H04N21/4394—Processing of audio elementary streams involving operations for analysing the audio stream, e.g. detecting features or characteristics in audio streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/44008—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/835—Generation of protective data, e.g. certificates
- H04N21/8358—Generation of protective data, e.g. certificates involving watermark
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/845—Structuring of content, e.g. decomposing content into time segments
- H04N21/8456—Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/11—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information not detectable on the record carrier
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/8126—Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts
- H04N21/8133—Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts specifically related to the content, e.g. biography of the actors in a movie, detailed information about an article seen in a video program
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
- H04N21/854—Content authoring
- H04N21/8547—Content authoring involving timestamps for synchronizing content
Definitions
- the present invention relates to a content reproduction program and a content reproduction apparatus.
- Patent Document 1 A method of outputting second screen information for a video with audio to a portable device capable of audio input / output and image display is disclosed (Patent Document 1).
- a time code associated with each video scene of video is embedded as watermark data in the audio.
- the time code is associated with data corresponding to the second screen information and stored in advance in the mobile device, audio is input to the mobile device in the vicinity of the video, and the second screen information is output in synchronization with the video scene. Is called.
- Patent Document 1 it is necessary to embed watermark data in audio in association with each video scene. For this reason, the burden of content production work such as a movie is high, and there are cases where watermark data cannot be embedded for convenience of the production side.
- an object of the present invention is to provide a content reproduction program and a content reproduction apparatus that easily and reliably determine a time axis of a certain content and reproduce another content in conjunction with the reproduction of the content.
- a content reproduction program includes a computer that detects a sound of a first content reproduced outside the computer, and a first content outside the computer based on the detected sound. And a time axis determination unit that determines a time axis of the second content to be reproduced, and a third content corresponding to the second content is reproduced in accordance with the reproduction of the second content based on the determined time axis. And a content playback unit to be realized.
- a content playback device plays back externally after the first content based on the detected audio and a sound detection unit that detects the sound of the first content played back externally.
- a time axis determination unit that determines a time axis of the second content, and a content reproduction unit that reproduces the third content according to the second content in accordance with the reproduction of the second content based on the determined time axis .
- the “unit” does not simply mean a physical means, but includes a case where the function of the “unit” is realized by software. Also, even if the functions of one “unit” or device are realized by two or more physical means or devices, the functions of two or more “units” or devices are realized by one physical means or device. May be.
- the present invention it is possible to provide a content reproduction program and a content reproduction apparatus that easily and reliably determine a time axis of a certain content and reproduce another content in conjunction with the reproduction of the content.
- FIG. 2 is a diagram illustrating an example of a functional configuration of a user terminal 100.
- FIG. It is a figure which shows an example of the user interface for acquiring a content. It is a figure which shows an example of the screen on which the determination state of the time axis was displayed.
- 4 is a flowchart illustrating an example of content reproduction processing in a user terminal 100.
- FIG. 1 is a diagram showing a configuration of a content reproduction system according to an embodiment of the present invention.
- the content reproduction system includes a user terminal 100 (content reproduction device), a screen 110, and a speaker 120.
- the screen 110 and the speaker 120 are installed in a facility such as a movie theater or a stadium, and the user terminal 100 is used by a user of the facility.
- a video of content such as a movie is displayed on the screen 110, and audio of the content is output from the speaker 120.
- another content such as subtitles and audio guides is played back in conjunction with the playback of the content.
- FIG. 2 is a diagram showing an outline of content reproduction in the content reproduction system.
- Content reproduced at a facility such as a movie theater includes first content 200 and second content 210 reproduced following the first content 200.
- the second content 210 is, for example, a main part of a movie.
- the first content 200 is, for example, a trailer, an advertisement, or the like (priority) that is played before the main part of the movie.
- watermark data 220 is embedded in the sound.
- the watermark data 220 is embedded, for example, for each frame of the first content 200, and includes data (first data) such as a time code for determining the start timing (time axis) of the second content 210.
- the watermark data 220 includes, for example, data (second data) indicating a facility, a region, and the like.
- the third content 230 reproduced on the user terminal 100 is, for example, subtitles of the movie that are reproduced in conjunction with the main part of the movie.
- the third content 230 is not limited to subtitles as long as it is played back in conjunction with the second content 210, and may be any content such as a voice guide, a sign language image, or an explanation image.
- the start timing (time axis) of the second content 210 is determined based on the watermark data 220, and the third content 230 is played back in conjunction with the second content 210.
- FIG. 3 is a diagram illustrating an example of a hardware configuration of the user terminal 100.
- the user terminal 100 is a computer such as a smartphone or a tablet terminal.
- the user terminal 100 includes a processor 300, a storage device 310, a communication interface (I / F) 320, a microphone 330, an audio output device 340, a display device 350, and an input device 360.
- I / F communication interface
- the processor 300 can implement various functions in the user terminal 100 by executing a program stored in the storage device 310.
- the storage device 310 is a storage area for storing programs and data.
- the storage device 310 can include a volatile memory that is a temporary storage area and a nonvolatile memory that is a permanent storage area.
- the storage device 310 may be provided outside the user terminal 100.
- the communication interface 320 is an interface device for transmitting and receiving programs and data to and from an external device.
- the communication interface 320 includes, for example, an interface for a mobile phone communication network and an interface for a wireless LAN (Local Area Network).
- the communication interface 320 may include a wired interface.
- the microphone 330 detects surrounding sounds.
- the microphone 330 can detect not only audible sound but also non-audible sound.
- the microphone 330 may be provided outside the user terminal 100.
- the audio output device 340 is an output device for outputting audio.
- the audio output device 340 is, for example, a speaker. Audio output device 340 may be a device for outputting audio to the earphone.
- the third content 230 includes sound (for example, sound guide or voice-over sound)
- the sound is output from the sound output device 340.
- the display device 350 is an output device for displaying an image (still image or moving image).
- the display device 350 is, for example, a liquid crystal display or an organic EL (Electroluminescence) display.
- the display device 350 may be provided outside the user terminal 100.
- the third content 230 includes an image (for example, caption)
- the image is displayed on the display device 350.
- the input device 360 is a device for accepting user input.
- the input device 360 may be a touch panel formed integrally with the display device 350, for example.
- the input device 360 may be provided outside the user terminal 100.
- FIG. 4 is a diagram illustrating an example of a functional configuration of the user terminal 100.
- the user terminal 100 includes a content acquisition unit 400, a content storage unit 410, an audio detection unit 420, a time axis determination unit 430, a content reproduction unit 440, a reference data acquisition unit 450, a reference data storage unit 460, a setting unit 470, and a setting data storage. Part 480.
- Each unit illustrated in FIG. 4 can be realized, for example, by using a storage area of the storage device 310 or by the processor 300 executing a program (content reproduction program) stored in the storage device 310.
- the content acquisition unit 400 acquires the third content 230 reproduced on the user terminal 100 via the Internet, for example, and stores it in the content storage unit 410.
- FIG. 5 is a diagram illustrating an example of a user interface for acquiring content.
- the screen 500 is displayed when, for example, an application for playing back the third content 230 is activated.
- Screen 500 includes areas 510 and 520 for selecting the type of content.
- subtitles, sign language, or voice guide can be selected.
- a movie theater / theatre, a theme park, a museum / art museum, a DVD / BD, or the like can be selected.
- a list of contents corresponding to the selection contents in the areas 510 and 520 is displayed in the area 530.
- the title of the second content 210 (for example, the main part) is displayed.
- the content acquisition unit 400 acquires the third content 230 corresponding to the second content 210 selected in the area 530. Note that the content acquisition unit 400 may acquire the third content 230 including content that is not a reproduction target.
- the content acquisition unit 400 may acquire the third content 230 specified by the sound (watermark data or fingerprint) detected in the first content 200. For example, when the watermark data embedded in the first content 200 includes data (second data) for specifying the third content 230, the content acquisition unit 400 determines the third content 230 based on the data. May be obtained.
- the sound detection unit 420 detects the sound of the content played outside the user terminal 100. Specifically, the sound detection unit 420 can detect the sound of the first content 200 (priority). Further, the sound detection unit 420 can also detect the sound of the second content 210 (main part). The sound detected by the sound detection unit 420 is, for example, watermark data or a fingerprint. The voice detection unit 420 can stop the operation when the time axis determination unit 430 determines the time axis of the second content 210.
- the time axis determination unit 430 determines the time axis of the second content 210 based on the sound detected by the sound detection unit 420. Specifically, for example, based on the data (first data) for determining the time axis of the second content 210 included in the watermark data embedded in the sound of the first content 200, the second content 210 The time axis can be determined. Further, the time axis determination unit 430 can determine the time axis of the second content 210 during reproduction of the second content 210 based on the audio fingerprint of the second content 210.
- determination of the time axis of the second content 210 refers to associating the reproduction timing of the second content 210 with time.
- the time axis determination unit 430 can determine a time (for example, an actual time or a relative time based on a certain time point) at which the reproduction of the second content 210 is started. For example, the time axis determination unit 430 can determine which part of the second content 210 is currently being played. The time axis determination unit 430 can stop the operation when the time axis of the second content 210 is determined.
- the content reproduction unit 440 reproduces the third content 230 in accordance with (synchronizes with) the reproduction of the second content 210 based on the time axis determined by the time axis determination unit 430. Further, the content reproduction unit 440 can output information indicating the time axis determination state by the time axis determination unit 430.
- FIG. 6 is a diagram illustrating an example of a screen on which a time axis determination state is displayed. As shown in screen 600 in FIG. 6, content reproduction unit 440 displays a message indicating that the determination of the time axis has been completed (for example, “Preparation for displaying subtitles has been completed. Please wait for a while until the main part starts.”). It can be displayed on the device 350.
- the reference data acquisition unit 450 acquires reference data for the time axis determination unit 430 to determine the time axis based on the fingerprint, for example, via the Internet, and stores the reference data in the reference data storage unit 460.
- the reference data is data in which the audio feature amount of the content is associated with the reproduction time of the content.
- the time axis determination unit 430 can determine the time axis of the second content 210 by comparing the audio feature amount of the first content 200 or the second content 210 with the reference data.
- the setting unit 470 receives a setting for playing the third content 230 from the user, and stores setting data indicating the setting in the setting data storage unit 480.
- the content reproduction unit 440 can reproduce the third content 230 based on the setting data.
- FIG. 7 is a flowchart showing an example of content reproduction processing in the user terminal 100. The process illustrated in FIG. 7 is started, for example, when the user starts an application.
- the audio detection unit 420 starts detecting watermark data embedded in the audio of the first content 200 (S701). If the watermark data is detected, the time axis determination unit 430 determines the time axis of the second content 210 based on the watermark data (S702). Note that the time axis determination unit 430 may determine the time axis of the second content 210 based on the audio fingerprint of the first content 200.
- the sound detection unit 420 executes time axis determination (S703 to S704) based on the sound of the second content 210 in parallel with determination of the time axis based on the sound of the first content 200 (S701 to S702). .
- the audio detection unit 420 starts detecting the audio fingerprint (feature amount) of the second content 210 (S703).
- the time axis determination unit 430 determines the time axis of the second content 210 based on the fingerprint (S704).
- the determination of the time axis based on the sound of the second content 210 (S703 to S704) is performed, for example, when the user starts viewing the second content 210 after the reproduction of the second content 210 is started, This is effective when the time axis determination (S701 to S702) based on the audio of the content 200 does not operate normally.
- the time axis determination process is repeatedly executed.
- the determination of the time axis is completed (S705: Y)
- the operation of the voice detection unit 420 and the time axis determination unit 430 is stopped, and the content reproduction unit 440 performs the second content 210 based on the determined time axis.
- the third content 230 is reproduced in accordance with the reproduction of (S706).
- the embodiment of the present invention has been described above. According to the present embodiment, it is possible to determine the time axis of the second content 210 (for example, the main part) to be reproduced following the first content 200 based on the sound of the first content 200 (for example, prior). Then, based on the determined time axis, the third content 230 (for example, subtitles or audio guide) corresponding to the second content 210 can be reproduced in accordance with the reproduction of the second content 210. Accordingly, even when the watermark data is not embedded in the audio of the second content 210 or when it is difficult to determine the start timing of the second content 210 with the audio fingerprint of the second content 210, the second content 210 The time axis can be easily and reliably determined.
- the third content 230 for example, subtitles or audio guide
- the time axis of the second content 210 can be determined based on the watermark data embedded in the sound of the first content 200.
- the first content 200 is often easier to embed watermark data than the second content 210. Therefore, by embedding watermark data in the sound of the first content 200, the time axis of the second content 210 can be easily and reliably determined.
- the watermark data embedded in the sound of the first content 200 can be sound data in a non-audible area.
- the time axis of the second content 210 can be easily and reliably determined without affecting the user viewing the first content 200.
- the fingerprint (feature amount) of the sound of the second content 210 is also included. Based on this, the time axis of the second content 210 may be determined. Thereby, for example, when the user starts viewing the second content 210 after starting the reproduction of the second content 210, the time axis of the second content 210 can be easily and reliably determined.
- the watermark data 220 embedded in the sound of the first content 200 is used to specify the third content 230 in addition to the data for determining the time axis (for example, the time code).
- Data for example, data indicating a facility or a region
- the third content 230 can be selected by the watermark data 220 embedded in the sound of the first content 200.
- the time axis of the second content 210 may be determined based on the audio feature amount of the first content 200. Thereby, for example, even when the watermark data 220 is not embedded in the first content 200, the time axis of the second content 210 can be easily and reliably determined.
- the voice detection unit 420 and the time axis determination unit 430 can stop the operation when the time axis of the second content 210 is determined. Thereby, the power consumption in the user terminal 100 can be suppressed.
- the time axis of the second content 210 held by the user terminal 100 and the actual time of the second content 210 after the determination of the time axis of the second content 210 is performed due to the performance of the user terminal 100 and the like. There may be a deviation from the axis. Therefore, the time axis determination unit 430, based on the fingerprint (feature amount) of the second content 210, at an appropriate timing after the determination of the time axis of the second content 210 is completed. The time axis may be corrected. For example, the time axis determination unit 430, after the determination of the time axis of the second content 210 is completed, at a predetermined time interval, based on the audio fingerprint (feature amount) of the second content 210, the second content 210.
- the time axis may be corrected. Further, for example, after the determination of the time axis of the second content 210 is completed, the time axis determination unit 430 performs the first determination based on the voice fingerprint (feature amount) of the second content 210 in accordance with an instruction from the user. The time axis of the two contents 210 may be corrected.
- information indicating the determination state of the time axis of the second content 210 can be output to the display device 350.
- the output destination of the information is not limited to the display device 350 but may be the audio output device 340. Thereby, it is possible to notify the user of the determination state of the time axis of the second content 210.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computer Security & Cryptography (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
Claims (10)
- コンピュータに、
前記コンピュータの外部で再生される第1コンテンツの音声を検出する音声検出部と、
前記検出された音声に基づいて、前記コンピュータの外部で前記第1コンテンツに続いて再生される第2コンテンツの時間軸を判定する時間軸判定部と、
前記判定された時間軸に基づいて、前記第2コンテンツの再生に合わせて、前記第2コンテンツに応じた第3コンテンツを再生するコンテンツ再生部と、
を実現させるためのコンテンツ再生プログラム。 - 前記音声は、前記第1コンテンツに埋め込まれた透かしデータである、
請求項1に記載のコンテンツ再生プログラム。 - 前記透かしデータは、非可聴領域の音声データである、
請求項2に記載のコンテンツ再生プログラム。 - 前記音声検出部は、さらに、前記コンピュータの外部で再生される前記第2コンテンツの音声の特徴量を検出し、
前記時間軸判定部は、前記第1コンテンツに埋め込まれた前記透かしデータ又は前記第2コンテンツの前記特徴量に基づいて前記時間軸を判定する、
請求項2又は3に記載のコンテンツ再生プログラム。 - 前記透かしデータは、前記時間軸の判定のための第1データと、前記第3コンテンツを特定するための第2データとを含み、
前記時間軸判定部は、前記第1データに基づいて前記時間軸を判定し、
前記コンテンツ再生部は、前記第2データに基づいて前記第3コンテンツを再生する、
請求項2から4のいずれか一項に記載のコンテンツ再生プログラム。 - 前記音声検出部は、前記音声の特徴量を検出し、
前記時間軸判定部は、前記検出された特徴量に基づいて前記時間軸を判定する、
請求項1に記載のコンテンツ再生プログラム。 - 前記音声検出部及び前記時間軸判定部は、前記時間軸が判定されると動作を停止する、請求項1から6のいずれか一項に記載のコンテンツ再生プログラム。
- 前記時間軸判定部は、前記時間軸が判定された後に、前記第2コンテンツの前記特徴量に基づいて前記時間軸を補正する、
請求項4に記載のコンテンツ再生プログラム。 - 前記コンテンツ再生部は、前記時間軸の判定状態を示す情報を出力する、
請求項1から8のいずれか一項に記載のコンテンツ再生プログラム。 - 外部で再生される第1コンテンツの音声を検出する音声検出部と、
前記検出された音声に基づいて、前記第1コンテンツに続いて外部で再生される第2コンテンツの時間軸を判定する時間軸判定部と、
前記判定された時間軸に基づいて、前記第2コンテンツの再生に合わせて、前記第2コンテンツに応じた第3コンテンツを再生するコンテンツ再生部と、
を備えるコンテンツ再生装置。
Priority Applications (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2017508106A JP6163680B1 (ja) | 2016-10-27 | 2016-10-27 | コンテンツ再生プログラム及びコンテンツ再生装置 |
CN201680090523.7A CN109983781B (zh) | 2016-10-27 | 2016-10-27 | 内容播放装置以及计算机可读存储介质 |
PCT/JP2016/081900 WO2018078768A1 (ja) | 2016-10-27 | 2016-10-27 | コンテンツ再生プログラム及びコンテンツ再生装置 |
US16/345,568 US11303951B2 (en) | 2016-10-27 | 2016-10-27 | Content reproduction program and content reproduction device |
KR1020197014929A KR102607703B1 (ko) | 2016-10-27 | 2016-10-27 | 콘텐츠 재생 프로그램 및 콘텐츠 재생 장치 |
EP16920393.2A EP3534618A4 (en) | 2016-10-27 | 2016-10-27 | CONTENT REPRODUCING PROGRAM AND CONTENT REPRODUCING APPARATUS |
TW106137069A TWI760383B (zh) | 2016-10-27 | 2017-10-27 | 內容再生程式及內容再生裝置 |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2016/081900 WO2018078768A1 (ja) | 2016-10-27 | 2016-10-27 | コンテンツ再生プログラム及びコンテンツ再生装置 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2018078768A1 true WO2018078768A1 (ja) | 2018-05-03 |
Family
ID=59351268
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2016/081900 WO2018078768A1 (ja) | 2016-10-27 | 2016-10-27 | コンテンツ再生プログラム及びコンテンツ再生装置 |
Country Status (7)
Country | Link |
---|---|
US (1) | US11303951B2 (ja) |
EP (1) | EP3534618A4 (ja) |
JP (1) | JP6163680B1 (ja) |
KR (1) | KR102607703B1 (ja) |
CN (1) | CN109983781B (ja) |
TW (1) | TWI760383B (ja) |
WO (1) | WO2018078768A1 (ja) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2015061112A (ja) | 2013-09-17 | 2015-03-30 | Npo法人メディア・アクセス・サポートセンター | 携帯デバイスへのセカンドスクリーン情報の提供方法 |
JP2015510305A (ja) * | 2012-01-06 | 2015-04-02 | トムソン ライセンシングThomson Licensing | セカンドスクリーン上のコンテンツを同期させるための方法およびシステム |
JP2016111492A (ja) * | 2014-12-05 | 2016-06-20 | 株式会社テレビ朝日 | 端末装置、サーバ装置、及びプログラム |
Family Cites Families (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102004021404B4 (de) * | 2004-04-30 | 2007-05-10 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Wasserzeicheneinbettung |
JP2007207328A (ja) * | 2006-01-31 | 2007-08-16 | Toshiba Corp | 情報記憶媒体、プログラム、情報再生方法、情報再生装置、データ転送方法、及びデータ処理方法 |
US8763029B1 (en) * | 2006-11-13 | 2014-06-24 | Visible World Inc. | Systems and methods for client-based addressable advertising |
CN101370261B (zh) * | 2007-08-15 | 2011-12-07 | 华为技术有限公司 | 从cs域向ps域切换的资源准备方法及通信设备 |
JP2010170604A (ja) * | 2009-01-21 | 2010-08-05 | Sony Corp | 光ディスク装置、光ピックアップ、プリフォーマット信号生成方法及びプログラム |
US8925024B2 (en) * | 2009-12-31 | 2014-12-30 | The Nielsen Company (Us), Llc | Methods and apparatus to detect commercial advertisements associated with media presentations |
US9026102B2 (en) * | 2010-03-16 | 2015-05-05 | Bby Solutions, Inc. | Movie mode and content awarding system and method |
CN101877625A (zh) * | 2010-06-29 | 2010-11-03 | 中兴通讯股份有限公司 | 下行语音帧的接收处理方法及装置、基带 |
JP5724283B2 (ja) * | 2010-10-15 | 2015-05-27 | ソニー株式会社 | 情報処理装置、同期方法およびプログラム |
US20130083036A1 (en) * | 2011-08-19 | 2013-04-04 | Hall of Hands Limited | Method of rendering a set of correlated events and computerized system thereof |
CN102664007B (zh) * | 2012-03-27 | 2016-08-31 | 上海量明科技发展有限公司 | 用于生成字符标识内容的方法、客户端及系统 |
CN102625164A (zh) * | 2012-04-06 | 2012-08-01 | 上海车音网络科技有限公司 | 多媒体数据处理平台及多媒体读物、系统和方法 |
US9305559B2 (en) * | 2012-10-15 | 2016-04-05 | Digimarc Corporation | Audio watermark encoding with reversing polarity and pairwise embedding |
FR3006525B1 (fr) * | 2013-06-04 | 2016-10-14 | Visiware | Synchronisation de contenus multimedia sur deuxieme ecran |
EP3651469A1 (en) * | 2013-09-20 | 2020-05-13 | Koninklijke KPN N.V. | Correlating timeline information between media streams |
JP6349977B2 (ja) * | 2013-10-21 | 2018-07-04 | ソニー株式会社 | 情報処理装置および方法、並びにプログラム |
CN110413185A (zh) * | 2013-12-09 | 2019-11-05 | 知识产权之桥一号有限责任公司 | 用于指定链接目的地和用于观看者的界面装置和记录介质 |
CN103702145A (zh) | 2013-12-26 | 2014-04-02 | 京东方科技集团股份有限公司 | 节目播放提醒方法及其系统和显示装置 |
CN105940679B (zh) * | 2014-01-31 | 2019-08-06 | 交互数字Ce专利控股公司 | 用于同步两个电子设备处的回放的方法和装置 |
CN104023251B (zh) * | 2014-06-13 | 2015-08-19 | 腾讯科技(深圳)有限公司 | 基于视频的互动方法和系统 |
US9665341B2 (en) | 2015-02-09 | 2017-05-30 | Sonos, Inc. | Synchronized audio mixing |
US20160309214A1 (en) * | 2015-04-15 | 2016-10-20 | Rovi Guides, Inc. | Method of synchronizing alternate audio content with video content |
EP3226570A1 (en) * | 2016-03-31 | 2017-10-04 | Thomson Licensing | Synchronizing audio and video signals rendered on different devices |
JP6157679B2 (ja) | 2016-04-12 | 2017-07-05 | 日本放送協会 | 信号処理装置及びプログラム |
US11432031B2 (en) * | 2018-02-23 | 2022-08-30 | Evixar Inc. | Content playback program, content playback method, and content playback system |
-
2016
- 2016-10-27 KR KR1020197014929A patent/KR102607703B1/ko active IP Right Grant
- 2016-10-27 EP EP16920393.2A patent/EP3534618A4/en active Pending
- 2016-10-27 CN CN201680090523.7A patent/CN109983781B/zh active Active
- 2016-10-27 JP JP2017508106A patent/JP6163680B1/ja active Active
- 2016-10-27 WO PCT/JP2016/081900 patent/WO2018078768A1/ja active Application Filing
- 2016-10-27 US US16/345,568 patent/US11303951B2/en active Active
-
2017
- 2017-10-27 TW TW106137069A patent/TWI760383B/zh active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2015510305A (ja) * | 2012-01-06 | 2015-04-02 | トムソン ライセンシングThomson Licensing | セカンドスクリーン上のコンテンツを同期させるための方法およびシステム |
JP2015061112A (ja) | 2013-09-17 | 2015-03-30 | Npo法人メディア・アクセス・サポートセンター | 携帯デバイスへのセカンドスクリーン情報の提供方法 |
JP2016111492A (ja) * | 2014-12-05 | 2016-06-20 | 株式会社テレビ朝日 | 端末装置、サーバ装置、及びプログラム |
Non-Patent Citations (1)
Title |
---|
See also references of EP3534618A4 |
Also Published As
Publication number | Publication date |
---|---|
JPWO2018078768A1 (ja) | 2018-11-01 |
TWI760383B (zh) | 2022-04-11 |
CN109983781A (zh) | 2019-07-05 |
CN109983781B (zh) | 2022-03-22 |
EP3534618A4 (en) | 2020-06-03 |
KR20190085938A (ko) | 2019-07-19 |
US20190253749A1 (en) | 2019-08-15 |
JP6163680B1 (ja) | 2017-07-19 |
US11303951B2 (en) | 2022-04-12 |
TW201820316A (zh) | 2018-06-01 |
KR102607703B1 (ko) | 2023-11-30 |
EP3534618A1 (en) | 2019-09-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
TWI538520B (zh) | 視頻資訊添加方法及裝置、視頻資訊顯示方法及裝置 | |
WO2016095384A1 (zh) | 一种弹幕显示方法及系统 | |
JP6271194B2 (ja) | 携帯デバイスへのセカンドスクリーン情報の提供方法 | |
WO2017113721A1 (zh) | 一种预览播放方法及装置 | |
KR20160013649A (ko) | 주변 소음에 기초하여 자막을 생성하는 동영상 디스플레이 방법 및 사용자 단말 | |
KR101789221B1 (ko) | 동영상 제공 장치, 동영상 제공 방법, 및 컴퓨터 프로그램 | |
US20120287283A1 (en) | Electronic device with voice prompt function and voice prompt method | |
US20160359932A1 (en) | Display device and method of controlling the same | |
US20170249971A1 (en) | Method for generating image in which sound source is inserted and apparatus therefor | |
JP6986036B2 (ja) | コンテンツ再生プログラム、コンテンツ再生方法及びコンテンツ再生システム | |
JP6163680B1 (ja) | コンテンツ再生プログラム及びコンテンツ再生装置 | |
KR102389040B1 (ko) | 콘텐츠 재생 프로그램, 콘텐츠 재생 방법 및 콘텐츠 재생 시스템 | |
CN114501126B (zh) | 视频播放方法、系统和存储介质 | |
US20170301328A1 (en) | Acoustic system, communication device, and program | |
WO2014097656A1 (ja) | 電子機器及び表示制御方法 | |
JP2019507978A5 (ja) | ||
US11582514B2 (en) | Source apparatus and control method therefor | |
KR101832464B1 (ko) | 동영상 제공 장치, 동영상 제공 방법, 및 컴퓨터 프로그램 | |
JP2015222865A (ja) | 携帯端末、関連データ提示方法、及びプログラム | |
JP4679462B2 (ja) | 静止画像表示切替システム | |
JP2008294513A (ja) | 映像再生装置 | |
JP2010287974A (ja) | 携帯電話及びプログラム | |
JP2006352885A (ja) | デジタルマルチメディア放送を多様な方式にて再生する方法及びそれによるデジタルマルチメディア放送受信装置 | |
JP2022090874A (ja) | 映像表示装置 | |
US20170169852A1 (en) | Method for resuming playing audio-video file after muting and electronic device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 2017508106 Country of ref document: JP |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 16920393 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 20197014929 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2016920393 Country of ref document: EP Effective date: 20190527 |