WO2014155526A1 - Information processing device and information processing method - Google Patents

Information processing device and information processing method Download PDF

Info

Publication number
WO2014155526A1
WO2014155526A1 PCT/JP2013/058791 JP2013058791W WO2014155526A1 WO 2014155526 A1 WO2014155526 A1 WO 2014155526A1 JP 2013058791 W JP2013058791 W JP 2013058791W WO 2014155526 A1 WO2014155526 A1 WO 2014155526A1
Authority
WO
WIPO (PCT)
Prior art keywords
song
search
music
result
search result
Prior art date
Application number
PCT/JP2013/058791
Other languages
French (fr)
Japanese (ja)
Inventor
剛 舘野
Original Assignee
株式会社 東芝
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社 東芝 filed Critical 株式会社 東芝
Priority to PCT/JP2013/058791 priority Critical patent/WO2014155526A1/en
Publication of WO2014155526A1 publication Critical patent/WO2014155526A1/en

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0033Recording/reproducing or transmission of music for electrophonic musical instruments
    • G10H1/0041Recording/reproducing or transmission of music for electrophonic musical instruments in coded form
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/683Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/121Musical libraries, i.e. musical databases indexed by musical parameters, wavetables, indexing schemes using musical parameters, musical rule bases or knowledge bases, e.g. for automatic composing methods
    • G10H2240/151Thumbnail, i.e. retrieving, playing or managing a short and musically relevant song preview from a library, e.g. the chorus
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/54Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for retrieval

Definitions

  • Embodiments of the present invention relate to an information processing apparatus and an information processing method for analyzing a reproduction state of music information included in recorded content.
  • receiving devices that receive television broadcasts, radio broadcasts, and the like can not only receive and record broadcast content, but also by accessing a predetermined server via a network line, A function of acquiring attribute information related to various types of information included in the recorded content has been installed.
  • the server searches for the music piece based on the feature value.
  • a so-called song search function in which attribute information corresponding to a song is returned to the receiving device has been put into practical use.
  • the information processing apparatus includes search means and analysis means.
  • the search means searches for music at a predetermined time interval for the content to be analyzed.
  • the analysis unit analyzes the reproduction state of the song included in the content based on the song search result obtained at predetermined time intervals by the search unit.
  • the song search function By using the song search function, it is possible to quickly analyze the playback status of music information (songs) included in the recorded content, that is, the tunes and song sections, etc. As a result, the user can perform various kinds of songs included in the recorded content. It is possible to provide an information processing apparatus and an information processing method that can be recognized in advance and can be handled easily when viewing recorded content.
  • FIG. It is a block block diagram shown in order to demonstrate roughly an example of the content delivery system as embodiment. It is a block block diagram shown in order to demonstrate roughly an example of the signal processing system of the receiving terminal which comprises the content delivery system in the embodiment. 3 is an external view for explaining an example of a remote controller that operates the receiving terminal in the embodiment.
  • FIG. It is a block block diagram shown in order to demonstrate an example of the music search process which the music search process part of the receiving terminal in the embodiment performs. It is a flowchart shown in order to demonstrate an example of the music search process operation which the music search process part of the receiving terminal in the embodiment performs. It is a flowchart shown in order to demonstrate an example of the music analysis process operation which the music analysis process part of the receiving terminal in the embodiment performs.
  • the information processing apparatus includes a search unit and an analysis unit.
  • the search means searches for music at predetermined time intervals for the content to be analyzed.
  • the analysis unit analyzes the reproduction state of the song included in the content based on the song search result obtained at predetermined time intervals by the search unit.
  • FIG. 1 schematically shows an example of a content distribution system 11 described in this embodiment.
  • program content distributed from a broadcasting station 12 using a broadcast wave as a medium is received by a receiving terminal 13 and used for video display, audio reproduction, and the like.
  • the receiving terminal 13 also has a function of recording and reproducing the received program content.
  • program content is supplied from the broadcasting station 12 to the server 14 by wire or wireless communication and stored.
  • the receiving terminal 13 communicates with the server 14 via a LAN (local area network) router 15 capable of wired or wireless communication, for example, a network circuit network 16 such as a fixed IP (internet protocol) communication network, and a gateway 17. It is accessible.
  • LAN local area network
  • the receiving terminal 13 obtains a program content distributed from the server 14 based on a preset program distribution schedule and performs video display, audio reproduction, or the like, or a request from the server 14. Therefore, it is possible to realize a so-called VOD (video on demand) function that performs video display, audio playback, and the like based on program content acquired from the server 14.
  • VOD video on demand
  • the broadcast station 12 supplies the server 14 with attribute information related to various program contents stored in the broadcast or the server 14 and stores the attribute information. For this reason, the receiving terminal 13 can acquire attribute information for a desired program content by accessing the server 14 and make it available for viewing by the user.
  • a so-called music search function is also realized, in which a music piece 14 searches for a music piece based on the feature amount and returns attribute information corresponding to the searched music piece to the receiving terminal 13.
  • the receiving terminal 13 can analyze the reproduction state of the music information included in the program content recorded therein, that is, the song and song section, etc. it can. Thereby, the user can recognize in advance the reproduction state of the song included in the program content recorded in the receiving terminal 13, and can make the handling when viewing the recorded program content convenient.
  • FIG. 2 schematically shows an example of the signal processing system of the receiving terminal 13. That is, the broadcast signal received by the antenna 18 is supplied to the tuner unit 20 via the input terminal 19, so that the broadcast signal of a desired channel is selected.
  • the broadcast signal selected by the tuner unit 20 is supplied to the demodulation processing unit 21 to demodulate the TS (transport stream), and the TS demodulated by the demodulation processing unit 21 is supplied to the signal processing unit 22. .
  • the signal processing unit 22 separates the input TS into a video component and an audio component, performs a decoding process on each stream, restores the digital video signal and the audio signal, and then restores the restored video signal. And predetermined digital signal processing is applied to the audio signal.
  • the signal processing unit 22 outputs the restored video signal to the synthesis processing unit 23, and outputs the restored audio signal to the audio processing unit 24.
  • the composition processing unit 23 superimposes and outputs an OSD (on-screen display) signal on the video signal supplied from the signal processing unit 22.
  • the video signal output from the synthesis processing unit 23 is supplied to the video processing unit 25 and converted into a format that can be displayed on a flat-type video display unit 26 having a liquid crystal display panel or the like at a subsequent stage, The video is supplied to the video display unit 26 and used for video display.
  • the audio processing unit 24 converts the input audio signal into an audio signal in a format that can be reproduced by the speaker 27 at the subsequent stage. Then, the audio signal output from the audio processing unit 24 is supplied to the speaker 27 for audio reproduction. Note that the audio signal output from the audio processing unit 24 is not limited to the speaker 17 and can be supplied to, for example, headphones (not shown) or the like for audio reproduction.
  • control unit 28 includes a CPU (central processing unit) 28 a and is transmitted from the operation unit 29 provided in the main body of the receiving terminal 13 or transmitted from the remote controller 30 and received by the receiving unit 31. By receiving the operation information, each part is controlled so that the operation content is reflected.
  • CPU central processing unit
  • the control unit 28 controls each unit by using the memory unit 28b.
  • the memory unit 28b mainly includes a ROM (read only ⁇ ⁇ memory) storing a control program executed by the CPU 28a, a RAM (random access memory) for providing a work area to the CPU 28a, various setting information, A device including a nonvolatile memory in which control information and the like are stored is assumed.
  • an HDD (hard disk drive) 32 is connected to the control unit 28.
  • the control unit 28 controls the video signal and the audio signal obtained from the signal processing unit 22 to be supplied to the HDD 32 and recorded on the hard disk 32a based on the operation of the operation unit 29 or the remote controller 30 by the user. be able to.
  • control unit 28 reads out the video signal and the audio signal from the hard disk 32a by the HDD 32 and supplies them to the signal processing unit 22 based on the operation of the operation unit 29 or the remote controller 30 by the user. It can be controlled to be used for the above-described video display and audio reproduction.
  • a network interface 33 is connected to the control unit 28.
  • the network interface 33 is connected to the LAN router 15 so as to be able to transmit information. Therefore, the control unit 28 can access the server 14 based on the operation of the operation unit 29 and the remote controller 30 by the user, and can acquire the program content and attribute information provided there. .
  • program content and attribute information acquired from the server 14 are also used for the above-described video display by the video display unit 26 and audio reproduction by the speaker 27. Furthermore, it goes without saying that program content and attribute information acquired from the server 14 are also used for recording and reproduction on the hard disk 32a by the HDD 32.
  • control unit 28 is provided with a music search processing unit 28c.
  • the song search processing unit 28c cuts out an audio stream of a predetermined section from the audio stream constituting the program content recorded on the hard disk 32a at regular time intervals, and decodes the cut out audio stream A feature value is generated from the processed audio signal and sent to the server 14, and the server 14 obtains a result of searching for a piece of music based on the feature value.
  • control unit 28 is provided with a music analysis processing unit 28d.
  • the music analysis processing unit 28d uses the search result acquired by the music search processing unit 28c to include the audio signal stream constituting the program content recorded on the hard disk 32a. It functions to analyze the reproduction state of the music information to be played, that is, the music piece, the music section, and the like.
  • FIG. 3 shows the external appearance of the remote controller 30.
  • the remote controller 30 mainly includes a power key 30a, a numeric key 30b, a channel up / down key 30c, a volume adjustment key 30d, a cursor up key 30e, a cursor down key 30f, a cursor left key 30g, and a cursor right key.
  • 30h an enter key 30i, a menu key 30j, a return key 30k, an end key 301, four color (blue, red, green, yellow) color keys 30m are provided.
  • the remote controller 30 is provided with a reproduction stop key 30n, a reproduction / pause key 30o, a reverse skip key 30p, a forward skip key 30q, a fast reverse key 30r, a fast forward key 30s, and the like. That is, the HDD 32 can be played, stopped, and paused by operating the playback stop key 30n or the playback / pause key 30o of the remote controller 30.
  • the fast reverse key 30r, fast forward key 30s, etc. of the remote controller 30 the data such as video and audio read from the hard disk 32a by the HDD 32 is continuously reversed or forward with respect to the reproduction direction.
  • fast reverse playback and fast forward playback can be performed.
  • FIG. 4 shows an example of a music search process performed by the music search processing unit 28c using functional blocks. That is, when a song search request is supplied to the song search instruction unit 34, the song search instruction unit 34 generates a cut-out request to the voice cut-out unit 35, and the voice cut-out unit 35 sends a request to the signal processing unit 22. And request acquisition of the audio signal.
  • the audio stream acquisition unit 36 acquires an audio stream of a predetermined section from the audio stream separated from the TS, and outputs it to the audio decoding unit 37.
  • the audio decoding unit 37 performs a decoding process on the input audio stream, generates a digital audio signal converted into, for example, PCM (pulse code modulation), and outputs the digital audio signal to the audio storage unit 38.
  • PCM pulse code modulation
  • the voice accumulation unit 38 accumulates the input voice signal for a predetermined section and outputs it to the voice cutout unit 35, whereby the voice signal cut out for the predetermined section is supplied to the voice cutout unit 35. . Thereafter, the voice cutout unit 35 outputs the input voice signal to the feature value generation unit 39.
  • the feature value generation unit 39 generates a feature value necessary for performing a song search from the input audio signal, and transmits the feature value to the server 14.
  • the server 14 searches for the music corresponding to the received feature value, and returns the search result to the receiving terminal 13.
  • the search result returned from the server 14 is acquired by the search result acquisition unit 40 of the receiving terminal 13, and after it is determined whether or not the search result determination unit 41 is valid, the search result is stored in the search result storage unit 42. They are taken out as necessary and used for display or the like.
  • FIG. 5 shows a flowchart summarizing an example of the music search processing operation performed by the music search processing unit 28c. That is, when the process is started (step S5a), the song search processing unit 28c cuts out the audio signal for a predetermined section in step S5b, and performs a song search from the cut out audio signal in step S5c. Necessary feature amounts are generated, and the generated feature amounts are transmitted to the server 14 in step S5d.
  • step S5e the server 14 searches for the music piece based on the received feature amount, and transmits the search result to the receiving terminal 13.
  • the music search processing unit 28c receives and determines the search result in step S5f, accumulates the search result in step S5g, and ends the process (step S5h).
  • the song analysis processing unit 28d uses the search result of the song search processing by the song search processing unit 28c to determine the audio signal constituting the program content recorded in the hard disk 32a.
  • the reproduction state of the music information included in the music, that is, the music piece, the music section, and the like are analyzed.
  • FIG. 6 shows an example of a processing operation in which the music analysis processing unit 28d uses the search result of the music search processing by the music search processing unit 28c to analyze the music information from the audio signal constituting the program content recorded on the hard disk 32a.
  • the flowchart which summarized these is shown. That is, when the process is started (step S6a), the song analysis processing unit 28d, in step S6b, performs a certain time interval (Intvl) from the beginning to the end of the audio signal constituting the program content to be analyzed. [Sec]: For example, in 30 seconds to 1 minute), the music search processing unit 28c is caused to execute the music search processing.
  • the audio data length (music search data length) used for the music search process is AudLen [sec]
  • the minimum audio data length necessary for the music search processing unit 28c to perform the music search process is Lmin [sec. ]
  • AudLen Lmin ⁇ 2 Satisfy the relationship.
  • the song search interval Intvl [sec] and song search data length AudiLen [sec] Intvl AudLen Satisfy the relationship.
  • the song search data length AudLen [sec] is set to twice the minimum voice data length Lmin [sec] necessary for performing the song search processing, so that the song search is performed near the change of the song.
  • FIGS. 7A, 7B, and 7C the case where the music search process is performed on the audio signal in which the music B continues after the music A is taken into consideration.
  • the music search result is one candidate for the music A only.
  • the music search data length AudLen [sec] when Lmin [sec] including only the audio signal of the music A and Lmin [sec] including only the audio signal of the music B can be secured,
  • the song search results are two candidates, song A and song B.
  • Lmin [sec] including only the audio signal of the song B can be secured in the song search data length AudLen [sec]
  • the song search result is only one candidate for the song B.
  • the song search data length AudLen [sec] is shorter than twice the minimum voice data length Lmin [sec] necessary for performing the song search process, that is, AudLen ⁇ Lmin ⁇ 2 If set to, a situation occurs in which Lmin [sec] including only the audio signal of the music A or B cannot be secured in the music search data length AudLen [sec] at the turn of the music A and B. In this case, a reliable music search result cannot be obtained. In addition, the music search may fail even when the human voice is included with the music. As a countermeasure, it has been found that it is effective to lengthen the music search data length AudLen [sec]. Therefore, AudLen [sec] is set to be twice as long as Lmin [sec].
  • the song analysis processing unit 28d refers to the song search results obtained by performing the song search processing at a constant time interval Intvl [sec] one by one, and determines whether or not the analysis is to be performed.
  • a so-called song segment identification process is performed.
  • a so-called song segment analysis process is performed for estimating a song start position and end position for a song segment determined to be analyzed. Details of the song segment identification processing and analysis processing will be described later.
  • step S6d the song analysis processing unit 28d executes the zone specifying process and the analysis process for the song section that has not been specified in the process of step S6c. Details of the specifying process and the analyzing process for the unspecified section will be described later.
  • the song analysis processing unit 28d performs a filtering process on the character string that is the song search result in step S6e. That is, the song search result is returned as a character string from the server 14. In this case, even if the same song has a plurality of notations on the database, it may be returned as a different song search result. For example, there are cases where the alphabet is in uppercase and lowercase letters, or whether the font is full-width or half-width. In this embodiment, the existing filtering process is performed on the character string that is the music search result, thereby rounding off the difference in notation.
  • step S6f the music analysis processing unit 28d stores the analysis result after the filtering process on the character string in the memory unit 28b, the hard disk 32a, etc., and ends the processing (step S6g).
  • the stored analysis results can be displayed as a list on the video processing unit 26 based on, for example, the operation of the operation unit 29 or the remote controller 30 by the user.
  • the music search processing unit 28c is caused to execute the music search processing at a constant time interval Intvl [sec]. It can be changed at a predetermined time interval for each scene, such as changing to the same program content.
  • FIG. 8 shows a flow chart summarizing an example of the song segment identification processing operation and analysis processing operation performed in step S6c. That is, when the process is started (step S8a), the song analysis processing unit 28d performs a song search process at a predetermined time interval Intvl [sec] in step S8b, among a plurality of song search results obtained. Then, one music search result is input based on a predetermined order (for example, oldest order in time).
  • a predetermined order for example, oldest order in time.
  • step S8c the music analysis processing unit 28d determines whether or not the input music search result is a so-called no result state in which no specific music is shown. If it is determined that there is no result (YES), in step S8d, the song analysis processing unit 28d sends the song search result to the analysis processing in the unspecified section performed in step S6d. The process returns to S8b.
  • step S8c determines whether or not the input song search result indicates a specific song, that is, it is determined that there is no result (NO)
  • step S8e The number of songs obtained as a song search result, that is, whether or not the number of candidate songs is two or more is determined. If it is determined that there are two or more songs (YES), the song search results are displayed in step S8d. The analysis process is performed in the unspecified section performed in step S6d, and the process returns to step S8b.
  • step S8e determines whether or not the number of candidate songs is not two or more (NO)
  • the song analysis processing unit 28d obtains the input song search result before or after that in step S8f. It is determined whether or not it is the same as the music search result, and if it is determined that it is the same (YES), in step S8g, the sections in which both music search results are obtained are combined as sections where the same music exists. Then, the process returns to step S8b.
  • step S8f If it is determined in step S8f that the input song search result is not the same as the song search result obtained before or after that time (NO), the song analysis processing unit 28d, in step S8h, A so-called song section analysis process (details will be described later) for estimating the start position or end position of the song indicated by the song search result is executed.
  • step S8i the song analysis processing unit 28d determines whether or not the song segment identification processing and analysis processing have been completed for all the song search results obtained by performing the song search processing. If it is determined that there is not (NO), the process returns to step S8b, and if it is determined that the process is completed (YES), the process ends (step S8j).
  • FIGS. 9A and 9B specifically show the song segment identification processing and analysis processing described in FIG.
  • the song search result obtained at time T1 indicates the song TNs
  • the next time T2 When the music search result obtained in step S8f also indicates the same music TNs (YES in step S8f), the music analysis processing unit 28d uses the section (Intvl [ sec]) and the section (Intvl [sec]) used to obtain the song search result (TNs) at time T2 is regarded as a section in which the same song TNs exists (step S8g). To identify.
  • the analysis processing unit 28d determines that the song TNs ends between times T2 and T3, and estimates the end position of the song TNs between times T2 and T3 (step S8h). Analysis processing is performed.
  • the music search result obtained at time T1 indicates the music TN
  • the next time T2 If the song search result obtained in step S3 indicates a song TNe different from the song TN, and the song search result obtained at the next time T3 also indicates the same song TNe, the song analysis processing unit 28d
  • the song section is identified as a section where the same song TNe exists, and the song TNe is determined to start between the times T1 and T2, and the start position of the song TNe is determined between the times T1 and T2.
  • An estimation processing of the song section is performed.
  • FIG. 10 shows a flowchart summarizing the analysis processing operation for estimating the end position of the song TNs between times T2 and T3 in FIG. 9A. That is, when the process is started (step S10a), the song analysis processing unit 28d, in step S10b, the position Ns on the audio signal corresponding to the time T2 when the song TNs was obtained as the song search result, and the song search result. As a result, an intermediate position N_mid with respect to the position N on the audio signal corresponding to the time T3 when the song TN is obtained is calculated by the following equation.
  • step S10c the music analysis processing unit 28d determines whether or not the time difference between the position Ns and the intermediate position N_mid is equal to or less than a preset threshold value (for example, 1 second). If it is determined that there is (YES), in step S10d, the position Ns or the intermediate position N_mid is determined as the end position of the music TNs, and the process is ended (step S10e).
  • a preset threshold value for example, 1 second
  • step S10c determines whether the time difference between the position Ns and the intermediate position N_mid is not less than 1 second (NO)
  • the music analysis processing unit 28d performs the music search processing unit at the intermediate position N_mid in step S10f.
  • the music search process is performed in 28c, and the music TN_mid is obtained as a music search result.
  • step S10g the song analysis processing unit 28d determines whether or not the song TN_mid indicated by the song search result at the intermediate position N_mid is the same as the song TNs indicated by the song search result at the position Ns. (YES), the position Ns is updated to the intermediate position N_mid in step S10h, and the process returns to step S10b.
  • step S10g When it is determined in step S10g that the song TN_mid indicated by the song search result at the intermediate position N_mid is not the same as the song TNs indicated by the song search result at the position Ns (NO), the song analysis processing unit 28d In step S10i, the position N is updated to the intermediate position N_mid, and the process returns to step S10b.
  • a music search is performed at an intermediate position N_mid between the position Ns and the position N, and as a result, the music TNs is obtained. It is determined whether the position Ns is updated to the intermediate position N_mid and the music search at the intermediate position with the position N is performed, or the position N is updated to the intermediate position N_mid and the music search at the intermediate position with the position Ns is performed. When the difference between the position Ns and the intermediate position N_mid is less than or equal to a predetermined threshold value, the position Ns or the intermediate position N_mid is determined as the end position of the song TNs.
  • FIG. 11 shows a flowchart summarizing the analysis processing operation for estimating the start position of the song TNe between times T1 and T2 in FIG. 9B. That is, when the process is started (step S11a), the song analysis processing unit 28d, in step S11b, the position N on the audio signal corresponding to the time T1 when the song TN was obtained as the song search result, and the song search result. As an intermediate position Ne_mid with respect to the position Ne on the audio signal corresponding to the time T2 at which the song TNe is obtained as follows.
  • step S11c the music analysis processing unit 28d determines whether or not the time difference between the position Ne and the intermediate position Ne_mid is equal to or less than a preset threshold value (for example, 1 second). If it is determined that there is (YES), in step S11d, the position Ne or the intermediate position Ne_mid is determined as the start position of the music TNe, and the process ends (step S11e).
  • a preset threshold value for example, 1 second
  • step S11c If it is determined in step S11c that the time difference between the position Ne and the intermediate position Ne_mid is not less than 1 second (NO), the music analysis processing unit 28d performs the music search processing unit at the intermediate position Ne_mid in step S11f. The music search process is performed in 28c, and the music TNE_mid is obtained as a music search result.
  • step S11g the music analysis processing unit 28d determines whether or not the music TNe_mid indicated by the music search result at the intermediate position Ne_mid is the same as the music T Ne indicated by the music search result at the position Ne. If it is determined (YES), the position Ne is updated to the intermediate position Ne_mid in step S11h, and the process returns to step S11b.
  • step S11g When it is determined in step S11g that the song TNe_mid indicated by the song search result at the intermediate position Ne_mid is not the same as the song TNe indicated by the song search result at the position Ne (NO), the song analysis processing unit 28d In step S11i, the position N is updated to the intermediate position Ne_mid, and the process returns to step S11b.
  • a music search is performed at an intermediate position Ne_mid between the position N and the position Ne, and as a result, whether or not the music TNe is obtained, It is determined whether the position Ne is updated to the intermediate position Ne_mid and the music search at the intermediate position with the position N is performed, or the position N is updated to the intermediate position Ne_mid and the music search at the intermediate position with the position Ne is performed.
  • the position Ne or the intermediate position Ne_mid is determined as the start position of the song TNe.
  • FIGS. 12 to 15 are flowcharts summarizing an example of the specifying process operation and the analyzing process operation of the unspecified music section performed in step S6d. That is, when the process is started (step S12a), the song analysis processing unit 28d determines whether or not there is a song search result that was sent to the analysis process in the unspecified section in step S8d in step S12b. If it is determined that it does not exist (NO), the process ends (step S12c).
  • step S12b If it is determined in step S12b that there is a music search result sent to the analysis process in the unspecified section (YES), the music analysis processing unit 28d determines in step S12d that a predetermined order (eg, time In accordance with the oldest etc.), and in step S12e, it is determined whether or not the input song search result indicates that there is no specific song, so-called no result. Determine.
  • a predetermined order eg, time In accordance with the oldest etc.
  • the song analysis processing unit 28d determines that the song search results obtained before and after the song search result are already in the song section or in step S13a. It is determined whether or not it has been analyzed as a non-music section. If it is determined that it has been analyzed (YES), the section used to obtain the song search result is determined as a non-music section in step S13b. Then, the process returns to step S12b.
  • step S13a If it is determined in step S13a that the song search result obtained before and after the input song search result has not been analyzed (NO), the song analysis processing unit 28d performs step S13c before and after. It is determined whether or not the obtained music search result is a so-called no result state in which no specific music is indicated.
  • the song analysis processing unit 28d estimates the end position of the non-music section using the analysis processing operation described above with reference to FIG. 10 in step S13d. The process returns to step S12b.
  • step S13c If it is determined in step S13c that the song search result obtained before and after the input song search result is in a state of no result (YES), the song analysis processing unit 28d determines in step S13e.
  • the song search processing unit 28c performs song search processing at an intermediate position between the position on the audio signal from which the song search result is obtained and the position on the audio signal from which the previous or subsequent song search result is obtained. Make it.
  • step S13f the song analysis processing unit 28d determines whether or not the song search result at the intermediate position is a so-called no-result state in which no particular song is indicated, and there is no result. If it is determined that it is not (NO), in step S13g, the section used to obtain the song search result is newly added as an unspecified section, and the process returns to step S12b.
  • step S13f If it is determined in step S13f that the result of the song search at the intermediate position is in a state of no result (YES), the song analysis processing unit 28d is used to obtain the input song search result in step S13h.
  • the obtained sections and the sections used for obtaining the music search results before and after the combined sections are combined as non-music sections, and the process returns to step S12b.
  • step S12e when it is determined in step S12e that the input song search result is not in a state of no result (NO), the song analysis processing unit 28d determines the number of songs obtained as a song search result in step S12f, That is, it is determined whether or not the number of candidate songs is one, and if it is determined that there is one song (YES), in step S12g, the analysis processing operation described above with reference to FIGS. 10 and 11 is used. Thus, the end position and start position of the candidate song are estimated, and the process returns to step S12b.
  • step S12f If it is determined in step S12f that the number of candidate songs is not one (NO), the song analysis processing unit 28d determines whether or not the number of candidate songs is two in step S12h. If it is determined that there are three or more songs (NO), in step S14a, there is the same song as the song obtained from the song search results before and after the entered song search result in step S14a. It is determined whether or not to do.
  • the song analysis processing unit 28d uses the section used for obtaining the input song search result in step S14b to obtain the song search result before and after that. Are combined as the same song section as that existing in the section, and the process returns to the process of step S12b.
  • step S14a If it is determined in step S14a that the same song as the song obtained from the song search results before and after the input song search result does not exist in the candidate song (NO), the song analysis processing unit 28d In step S14c, it is determined whether or not the candidate song includes the same song as the song obtained from the song search result before or after the input song search result.
  • the song analysis processing unit 28d obtains the song search result before or after the section used to obtain the entered song search result in step S14d.
  • the music is combined as the same music section as the music existing in the used section, and the process returns to step S12b.
  • step S14c If it is determined in step S14c that the candidate song does not include the same song as the song obtained from the song search result before or after the input song search result (NO), the song analysis processing unit In step S14e, in order to obtain an input song search result for the first candidate song determined based on a preset priority in consideration of the case where there are three or more candidate songs in step S14e. It is estimated that the song exists in the used section, and the process returns to step S12b.
  • the song analysis processing unit 28d determines a song search start position on a section where two candidate songs are searched in step S15a, and resets the time counter.
  • the song search start position is the head position of the section in which two candidate songs are searched.
  • step S15b the song analysis processing unit 28d determines whether or not the newly stored song search result has changed compared to the previous song search result, and if it is determined that it has changed (YES) In step S15c, it is determined that the changed time is the boundary of the song, and the process returns to step S12b.
  • step S15b If it is determined in step S15b that the song search result has not changed (NO), the song analysis processing unit 28d sets the song search data length AudLen [sec] to Lmin [sec] at the head position in step S15d.
  • the music search processing unit 28c performs music search processing, and the music search results are stored in step S15e.
  • the music analysis processing unit 28d moves the music search position by a predetermined amount in step S15g and returns to the processing in step S15b.
  • the music search position is moved by a certain amount toward the end position of the section in which two candidate songs are searched.
  • FIG. 16 shows an example of the result of analyzing the reproduction state of the music information using the song search results as described in FIGS. 6 to 15 for the program content recorded on the hard disk 32a by the HDD 32. Yes.
  • the analysis result shown in this example is obtained by the user operating the menu key 30j of the remote controller 30 to divide into a plurality of menu screens having a hierarchical structure and requesting the display of the analysis result, thereby the video display unit 26 Can be listed.
  • the present invention is not limited to the above-described embodiment as it is, and can be embodied by variously modifying the constituent elements without departing from the scope of the invention in the implementation stage.
  • Various inventions can be formed by appropriately combining a plurality of constituent elements disclosed in the above-described embodiments. For example, some components may be deleted from all the components shown in the embodiment. Furthermore, constituent elements according to different embodiments may be appropriately combined.

Landscapes

  • Engineering & Computer Science (AREA)
  • Library & Information Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

In an embodiment, an information processing device is provided with a searching means (28c) and an analyzing means (28d). The searching means (28c) performs a song search on content to be analyzed at prescribed time intervals. The analyzing means (28d) analyzes a playback state for songs included in the content on the basis of the song search results obtained at the prescribed intervals by the search means (28c).

Description

情報処理装置及び情報処理方法Information processing apparatus and information processing method
 この発明の実施の形態は、記録コンテンツに含まれる音楽情報の再生状態を解析する情報処理装置及び情報処理方法に関する。 Embodiments of the present invention relate to an information processing apparatus and an information processing method for analyzing a reproduction state of music information included in recorded content.
 周知のように、近年、テレビジョン放送やラジオ放送等を受信する受信機器では、放送コンテンツを受信して記録することができるだけでなく、ネットワーク回線網を介して所定のサーバにアクセスすることにより、記録したコンテンツに含まれる各種の情報に関する属性情報を取得する機能が搭載されるようになってきている。 As is well known, in recent years, receiving devices that receive television broadcasts, radio broadcasts, and the like can not only receive and record broadcast content, but also by accessing a predetermined server via a network line, A function of acquiring attribute information related to various types of information included in the recorded content has been installed.
 そして、この機能を用いることにより、現在では、受信機器が、記録コンテンツに含まれる音楽情報から特徴量を生成してサーバに送ると、サーバが特徴量に基づいて曲目を検索し、その検索した曲に対応する属性情報を受信機器に返信するという、いわゆる、曲検索機能も実用化されている。 By using this function, when the receiving device generates a feature value from the music information included in the recorded content and sends it to the server, the server searches for the music piece based on the feature value. A so-called song search function in which attribute information corresponding to a song is returned to the receiving device has been put into practical use.
特開2007-219178号公報JP 2007-219178 A
 曲検索機能を利用して記録コンテンツに含まれる音楽情報(曲)の再生状態、つまり、曲目や曲区間等を迅速に解析処理することができ、ひいては、ユーザが記録コンテンツに含まれる各種の曲を予め認識して、記録コンテンツを視聴する際の取り扱いを便利にし得るようにした情報処理装置及び情報処理方法を提供することを目的とする。 By using the song search function, it is possible to quickly analyze the playback status of music information (songs) included in the recorded content, that is, the tunes and song sections, etc. As a result, the user can perform various kinds of songs included in the recorded content. It is an object of the present invention to provide an information processing apparatus and an information processing method that can recognize information in advance and make it easier to handle when viewing recorded content.
 実施の形態によれば、情報処理装置は、検索手段と解析手段とを備える。検索手段は、 解析対象となるコンテンツに対して所定の時間間隔で曲検索を行なう。解析手段は、検索手段により所定の時間間隔で得られる曲検索結果に基づいて、コンテンツに含まれる曲の再生状態を解析する。 According to the embodiment, the information processing apparatus includes search means and analysis means. The search means searches for music at a predetermined time interval for the content to be analyzed. The analysis unit analyzes the reproduction state of the song included in the content based on the song search result obtained at predetermined time intervals by the search unit.
 曲検索機能を利用して記録コンテンツに含まれる音楽情報(曲)の再生状態、つまり、曲目や曲区間等を迅速に解析処理することができ、ひいては、ユーザが記録コンテンツに含まれる各種の曲を予め認識して、記録コンテンツを視聴する際の取り扱いを便利にし得るようにした情報処理装置及び情報処理方法を提供することができる。 By using the song search function, it is possible to quickly analyze the playback status of music information (songs) included in the recorded content, that is, the tunes and song sections, etc. As a result, the user can perform various kinds of songs included in the recorded content. It is possible to provide an information processing apparatus and an information processing method that can be recognized in advance and can be handled easily when viewing recorded content.
実施の形態としてのコンテンツ配信システムの一例を概略的に説明するために示すブロック構成図である。It is a block block diagram shown in order to demonstrate roughly an example of the content delivery system as embodiment. 同実施の形態におけるコンテンツ配信システムを構成する受信端末の信号処理系の一例を概略的に説明するために示すブロック構成図である。It is a block block diagram shown in order to demonstrate roughly an example of the signal processing system of the receiving terminal which comprises the content delivery system in the embodiment. 同実施の形態における受信端末を操作するリモートコントローラの一例を説明するために示す外観図である。3 is an external view for explaining an example of a remote controller that operates the receiving terminal in the embodiment. FIG. 同実施の形態における受信端末の曲検索処理部が行なう曲検索処理の一例を機能的に説明するために示すブロック構成図である。It is a block block diagram shown in order to demonstrate an example of the music search process which the music search process part of the receiving terminal in the embodiment performs. 同実施の形態における受信端末の曲検索処理部が行なう曲検索処理動作の一例を説明するために示すフローチャートである。It is a flowchart shown in order to demonstrate an example of the music search process operation which the music search process part of the receiving terminal in the embodiment performs. 同実施の形態における受信端末の曲解析処理部が行なう曲解析処理動作の一例を説明するために示すフローチャートである。It is a flowchart shown in order to demonstrate an example of the music analysis process operation which the music analysis process part of the receiving terminal in the embodiment performs. 同実施の形態における受信端末の曲検索処理部が行なう曲検索処理に利用する曲検索データ長を説明するために示す図である。It is a figure shown in order to demonstrate the music search data length utilized for the music search process which the music search process part of the receiving terminal in the embodiment performs. 同実施の形態における受信端末の曲解析処理部が行なう曲区間の特定処理動作及び解析処理動作の一例を説明するために示すフローチャートである。It is a flowchart shown in order to demonstrate an example of the music section specific process operation | movement and analysis process operation which the music analysis process part of the receiving terminal in the embodiment performs. 同実施の形態における受信端末の曲解析処理部が行なう曲の終了位置及び開始位置の解析処理の一例を説明するために示す図である。It is a figure shown in order to demonstrate an example of the analysis process of the end position and start position of the music which the music analysis process part of the receiving terminal in the embodiment performs. 同実施の形態における受信端末の曲解析処理部が行なう曲の終了位置の解析処理動作の一例を説明するために示すフローチャートである。6 is a flowchart shown for explaining an example of a music end position analysis processing operation performed by a music analysis processing unit of the receiving terminal in the embodiment. 同実施の形態における受信端末の曲解析処理部が行なう曲の開始位置の解析処理動作の一例を説明するために示すフローチャートである。It is a flowchart shown in order to demonstrate an example of the analysis process operation | movement of the music start position which the music analysis process part of the receiving terminal in the embodiment performs. 同実施の形態における受信端末の曲解析処理部が行なう未特定曲区間の特定処理動作及び解析処理動作の一例を説明するために示すフローチャートである。It is a flowchart shown in order to demonstrate an example of the specific process operation | movement and analysis process operation | movement of an unspecified music area which the music analysis process part of the receiving terminal in the embodiment performs. 同実施の形態における受信端末の曲解析処理部が行なう未特定曲区間の特定処理動作及び解析処理動作の一例を説明するために示すフローチャートである。It is a flowchart shown in order to demonstrate an example of the specific process operation | movement and analysis process operation | movement of an unspecified music area which the music analysis process part of the receiving terminal in the embodiment performs. 同実施の形態における受信端末の曲解析処理部が行なう未特定曲区間の特定処理動作及び解析処理動作の一例を説明するために示すフローチャートである。It is a flowchart shown in order to demonstrate an example of the specific process operation | movement and analysis process operation | movement of an unspecified music area which the music analysis process part of the receiving terminal in the embodiment performs. 同実施の形態における受信端末の曲解析処理部が行なう未特定曲区間の特定処理動作及び解析処理動作の一例を説明するために示すフローチャートである。It is a flowchart shown in order to demonstrate an example of the specific process operation | movement and analysis process operation | movement of an unspecified music area which the music analysis process part of the receiving terminal in the embodiment performs. 同実施の形態における受信端末の曲解析処理部が行なった記録コンテンツに含まれる音楽情報の再生状態を解析した結果の一例を説明するために示す図である。It is a figure shown in order to demonstrate an example of the result of having analyzed the reproduction | regeneration state of the music information contained in the recording content which the music analysis process part of the receiving terminal in the same embodiment performed. 同実施の形態における受信端末の曲解析処理部が行なった記録コンテンツに含まれる音楽情報の再生状態を解析した結果の他の例を説明するために示す図である。It is a figure shown in order to demonstrate the other example of the result of having analyzed the reproduction | regeneration state of the music information contained in the recording content which the music analysis process part of the receiving terminal in the same embodiment performed.
 以下、実施の形態について図面を参照して詳細に説明する。実施の形態によれば、情報処理装置は、検索手段と解析手段とを備える。検索手段は、解析対象となるコンテンツに対して所定の時間間隔で曲検索を行なう。解析手段は、検索手段により所定の時間間隔で得られる曲検索結果に基づいて、コンテンツに含まれる曲の再生状態を解析する。 Hereinafter, embodiments will be described in detail with reference to the drawings. According to the embodiment, the information processing apparatus includes a search unit and an analysis unit. The search means searches for music at predetermined time intervals for the content to be analyzed. The analysis unit analyzes the reproduction state of the song included in the content based on the song search result obtained at predetermined time intervals by the search unit.
 図1は、この実施の形態で説明するコンテンツ配信システム11の一例を概略的に示している。まず、このコンテンツ配信システム11では、放送局12から放送波を媒体として配信される番組コンテンツが、受信端末13に受信されて映像表示や音声再生等に供される。また、この受信端末13は、受信した番組コンテンツを記録再生する機能も備えている。 FIG. 1 schematically shows an example of a content distribution system 11 described in this embodiment. First, in the content distribution system 11, program content distributed from a broadcasting station 12 using a broadcast wave as a medium is received by a receiving terminal 13 and used for video display, audio reproduction, and the like. The receiving terminal 13 also has a function of recording and reproducing the received program content.
 さらに、このコンテンツ配信システム11では、放送局12から有線または無線通信により、番組コンテンツがサーバ14に供給されて蓄積されるようになっている。そして、上記受信端末13は、有線または無線による通信が可能なLAN(local area network)ルータ15、例えば固定IP(internet protocol)通信網等でなるネットワーク回線網16、ゲートウェイ17を介してサーバ14にアクセス可能となっている。 Furthermore, in the content distribution system 11, program content is supplied from the broadcasting station 12 to the server 14 by wire or wireless communication and stored. The receiving terminal 13 communicates with the server 14 via a LAN (local area network) router 15 capable of wired or wireless communication, for example, a network circuit network 16 such as a fixed IP (internet protocol) communication network, and a gateway 17. It is accessible.
 これにより、受信端末13では、サーバ14から予め設定された番組配信スケジュールに基づいて配信される番組コンテンツを取得して映像表示や音声再生等を行なう、いわゆるIP放送受信機能や、サーバ14に要求しサーバ14から取得した番組コンテンツに基づいて映像表示や音声再生等を行なう、いわゆるVOD(video on demand)機能等を実現することが可能となる。 As a result, the receiving terminal 13 obtains a program content distributed from the server 14 based on a preset program distribution schedule and performs video display, audio reproduction, or the like, or a request from the server 14. Therefore, it is possible to realize a so-called VOD (video on demand) function that performs video display, audio playback, and the like based on program content acquired from the server 14.
 また、このコンテンツ配信システム11では、放送局12が、放送またはサーバ14に蓄積した各種の番組コンテンツに関する属性情報を、サーバ14に供給して蓄積させるようにしている。このため、受信端末13では、サーバ14にアクセスすることにより、所望の番組コンテンツに対する属性情報を取得して、ユーザへの視聴に供させることができるようになっている。 In the content distribution system 11, the broadcast station 12 supplies the server 14 with attribute information related to various program contents stored in the broadcast or the server 14 and stores the attribute information. For this reason, the receiving terminal 13 can acquire attribute information for a desired program content by accessing the server 14 and make it available for viewing by the user.
 さらに、この属性情報の取得機能を用いることにより、このコンテンツ配信システム11では、受信端末13が、自己の記録した番組コンテンツに含まれる音楽情報から特徴量を生成してサーバ14に送ると、サーバ14が特徴量に基づいて曲目を検索し、その検索した曲に対応する属性情報を受信端末13に返信する、いわゆる、曲検索機能も実現されるようになっている。 Further, by using this attribute information acquisition function, in the content distribution system 11, when the receiving terminal 13 generates a feature amount from the music information included in the program content recorded by itself and sends it to the server 14, A so-called music search function is also realized, in which a music piece 14 searches for a music piece based on the feature amount and returns attribute information corresponding to the searched music piece to the receiving terminal 13.
 そして、この曲検索機能を利用することにより、受信端末13では、自己の記録した番組コンテンツについて、そこに含まれる音楽情報の再生状態、つまり、曲目や曲区間等を解析処理しておくことができる。これにより、ユーザは、受信端末13に記録された番組コンテンツに含まれる曲の再生状態を予め認識することができ、記録された番組コンテンツを視聴する際の取り扱いを便利にすることができる。 By using this song search function, the receiving terminal 13 can analyze the reproduction state of the music information included in the program content recorded therein, that is, the song and song section, etc. it can. Thereby, the user can recognize in advance the reproduction state of the song included in the program content recorded in the receiving terminal 13, and can make the handling when viewing the recorded program content convenient.
 図2は、上記受信端末13の信号処理系の一例を概略的に示している。すなわち、アンテナ18で受信した放送信号は、入力端子19を介してチューナ部20に供給されることにより、所望のチャンネルの放送信号が選局される。このチューナ部20で選局された放送信号は、復調処理部21に供給されてTS(transport stream)が復調され、この復調処理部21で復調されたTSは、信号処理部22に供給される。 FIG. 2 schematically shows an example of the signal processing system of the receiving terminal 13. That is, the broadcast signal received by the antenna 18 is supplied to the tuner unit 20 via the input terminal 19, so that the broadcast signal of a desired channel is selected. The broadcast signal selected by the tuner unit 20 is supplied to the demodulation processing unit 21 to demodulate the TS (transport stream), and the TS demodulated by the demodulation processing unit 21 is supplied to the signal processing unit 22. .
 この信号処理部22は、入力されたTSを映像成分と音声成分とに分離し、各ストリームに対してそれぞれデコード処理を施して、デジタルの映像信号及び音声信号を復元した後、復元した映像信号及び音声信号に対してそれぞれ所定のデジタル信号処理を施している。そして、この信号処理部22は、復元した映像信号を合成処理部23に出力し、復元した音声信号を音声処理部24に出力している。 The signal processing unit 22 separates the input TS into a video component and an audio component, performs a decoding process on each stream, restores the digital video signal and the audio signal, and then restores the restored video signal. And predetermined digital signal processing is applied to the audio signal. The signal processing unit 22 outputs the restored video signal to the synthesis processing unit 23, and outputs the restored audio signal to the audio processing unit 24.
 このうち、合成処理部23は、信号処理部22から供給される映像信号に、OSD(on screen display)信号を重畳して出力している。この合成処理部23から出力された映像信号は、映像処理部25に供給されて、後段の、例えば液晶表示パネル等を有する平面型の映像表示部26で表示可能なフォーマットに変換された後、上記映像表示部26に供給されて映像表示に供される。 Among these, the composition processing unit 23 superimposes and outputs an OSD (on-screen display) signal on the video signal supplied from the signal processing unit 22. The video signal output from the synthesis processing unit 23 is supplied to the video processing unit 25 and converted into a format that can be displayed on a flat-type video display unit 26 having a liquid crystal display panel or the like at a subsequent stage, The video is supplied to the video display unit 26 and used for video display.
 また、上記音声処理部24は、入力された音声信号を、後段のスピーカ27で再生可能なフォーマットの音声信号に変換している。そして、この音声処理部24から出力された音声信号が、上記スピーカ27に供給されることにより音声再生に供される。なお、音声処理部24から出力された音声信号は、スピーカ17に限らず、例えばヘッドホン(図示せず)等に供給して音声再生に供させることもできる。 The audio processing unit 24 converts the input audio signal into an audio signal in a format that can be reproduced by the speaker 27 at the subsequent stage. Then, the audio signal output from the audio processing unit 24 is supplied to the speaker 27 for audio reproduction. Note that the audio signal output from the audio processing unit 24 is not limited to the speaker 17 and can be supplied to, for example, headphones (not shown) or the like for audio reproduction.
 ここで、この受信端末13は、上記した各種の受信動作を含む種々の動作を制御部28によって統括的に制御されている。この制御部28は、CPU(central processing unit)28aを内蔵しており、受信端末13の本体に設けられた操作部29からの操作情報、または、リモートコントローラ30から送信され受信部31で受信された操作情報を受けることにより、その操作内容が反映されるように各部をそれぞれ制御している。 Here, in the receiving terminal 13, various operations including the above-described various receiving operations are comprehensively controlled by the control unit 28. The control unit 28 includes a CPU (central processing unit) 28 a and is transmitted from the operation unit 29 provided in the main body of the receiving terminal 13 or transmitted from the remote controller 30 and received by the receiving unit 31. By receiving the operation information, each part is controlled so that the operation content is reflected.
 この場合、制御部28は、メモリ部28bを利用することにより、各部の制御を行なっている。このメモリ部28bとしては、主に、CPU28aが実行する制御プログラムを格納したROM(read only memory)と、該CPU28aに作業エリアを提供するためのRAM(random access memory)と、各種の設定情報及び制御情報等が格納される不揮発性メモリとを備えたものを想定している。 In this case, the control unit 28 controls each unit by using the memory unit 28b. The memory unit 28b mainly includes a ROM (read only し た memory) storing a control program executed by the CPU 28a, a RAM (random access memory) for providing a work area to the CPU 28a, various setting information, A device including a nonvolatile memory in which control information and the like are stored is assumed.
 また、この制御部28には、HDD(hard disk drive)32が接続されている。この制御部28は、ユーザによる操作部29やリモートコントローラ30の操作に基づいて、上記信号処理部22から得られる映像信号及び音声信号をHDD32に供給し、そのハードディスク32aに記録させるように制御することができる。 Further, an HDD (hard disk drive) 32 is connected to the control unit 28. The control unit 28 controls the video signal and the audio signal obtained from the signal processing unit 22 to be supplied to the HDD 32 and recorded on the hard disk 32a based on the operation of the operation unit 29 or the remote controller 30 by the user. be able to.
 また、この制御部28は、ユーザによる操作部29やリモートコントローラ30の操作に基づいて、HDD32によりハードディスク32aから映像信号及び音声信号を読み出させて信号処理部22に供給することによって、以後、上記した映像表示及び音声再生に供させるように制御することができる。 Further, the control unit 28 reads out the video signal and the audio signal from the hard disk 32a by the HDD 32 and supplies them to the signal processing unit 22 based on the operation of the operation unit 29 or the remote controller 30 by the user. It can be controlled to be used for the above-described video display and audio reproduction.
 さらに、この制御部28には、ネットワークインターフェース33が接続されている。このネットワークインターフェース33は、上記LANルータ15と情報伝送可能に接続されている。このため、制御部28は、ユーザによる操作部29やリモートコントローラ30の操作に基づいてサーバ14にアクセスし、そこで提供している番組コンテンツ及び属性情報等を取得することができるようになっている。 Furthermore, a network interface 33 is connected to the control unit 28. The network interface 33 is connected to the LAN router 15 so as to be able to transmit information. Therefore, the control unit 28 can access the server 14 based on the operation of the operation unit 29 and the remote controller 30 by the user, and can acquire the program content and attribute information provided there. .
 なお、サーバ14から取得した番組コンテンツ及び属性情報等についても、上記した映像表示部26による映像表示やスピーカ27による音声再生に供されることはもちろんである。さらに、サーバ14から取得した番組コンテンツ及び属性情報等についても、上記HDD32によりハードディスク32aに対する記録再生に供されることは言うまでもないことである。 Of course, the program content and attribute information acquired from the server 14 are also used for the above-described video display by the video display unit 26 and audio reproduction by the speaker 27. Furthermore, it goes without saying that program content and attribute information acquired from the server 14 are also used for recording and reproduction on the hard disk 32a by the HDD 32.
 また、上記制御部28には、曲検索処理部28cが設けられている。この曲検索処理部28cは、詳細は後述するが、ハードディスク32aに記録された番組コンテンツを構成する音声ストリームから、一定の時間間隔毎に所定区間の音声ストリームを切り出し、その切り出した音声ストリームをデコード処理した音声信号から特徴量を生成してサーバ14に送り、サーバ14が特徴量に基づいて曲目の検索を行なった結果を取得する。 In addition, the control unit 28 is provided with a music search processing unit 28c. As will be described in detail later, the song search processing unit 28c cuts out an audio stream of a predetermined section from the audio stream constituting the program content recorded on the hard disk 32a at regular time intervals, and decodes the cut out audio stream A feature value is generated from the processed audio signal and sent to the server 14, and the server 14 obtains a result of searching for a piece of music based on the feature value.
 さらに、上記制御部28には、曲解析処理部28dが設けられている。この曲解析処理部28dは、詳細は後述するが、曲検索処理部28cが取得した検索結果を利用することにより、ハードディスク32aに記録された番組コンテンツを構成する音声信号のストリームについて、そこに含まれる音楽情報の再生状態、つまり、曲目や曲区間等を解析処理するように機能する。 Further, the control unit 28 is provided with a music analysis processing unit 28d. As will be described in detail later, the music analysis processing unit 28d uses the search result acquired by the music search processing unit 28c to include the audio signal stream constituting the program content recorded on the hard disk 32a. It functions to analyze the reproduction state of the music information to be played, that is, the music piece, the music section, and the like.
 図3は、上記リモートコントローラ30の外観を示している。このリモートコントローラ30には、主として、電源キー30a、数字キー30b、チャンネルアップダウンキー30c、音量調整キー30d、カーソル上移動キー30e、カーソル下移動キー30f、カーソル左移動キー30g、カーソル右移動キー30h、決定キー30i、メニューキー30j、戻るキー30k、終了キー30l、4色(青,赤,緑,黄)のカラーキー30m等が設けられている。 FIG. 3 shows the external appearance of the remote controller 30. The remote controller 30 mainly includes a power key 30a, a numeric key 30b, a channel up / down key 30c, a volume adjustment key 30d, a cursor up key 30e, a cursor down key 30f, a cursor left key 30g, and a cursor right key. 30h, an enter key 30i, a menu key 30j, a return key 30k, an end key 301, four color (blue, red, green, yellow) color keys 30m are provided.
 また、このリモートコントローラ30には、再生停止キー30n、再生/一時停止キー30o、逆方向スキップキー30p、順方向スキップキー30q、早戻しキー30r、早送りキー30s等が設けられている。すなわち、上記HDD32に対しては、リモートコントローラ30の再生停止キー30nや再生/一時停止キー30oを操作することによって、再生、停止、一時停止を行なうことが可能となる。 Further, the remote controller 30 is provided with a reproduction stop key 30n, a reproduction / pause key 30o, a reverse skip key 30p, a forward skip key 30q, a fast reverse key 30r, a fast forward key 30s, and the like. That is, the HDD 32 can be played, stopped, and paused by operating the playback stop key 30n or the playback / pause key 30o of the remote controller 30.
 また、リモートコントローラ30の逆方向スキップキー30pや順方向スキップキー30qを操作することによって、HDD32でハードディスク32aから読み取っている映像や音声等のデータを、その再生方向に対して逆方向や順方向に一定量ずつスキップさせる、いわゆる、逆方向スキップや順方向スキップを行なうことができる。 Further, by operating the backward skip key 30p or the forward skip key 30q of the remote controller 30, data such as video and audio read from the hard disk 32a by the HDD 32 can be reversed or forward with respect to the reproduction direction. Thus, it is possible to perform a so-called reverse skip or forward skip, which causes a predetermined amount to be skipped.
 さらに、リモートコントローラ30の早戻しキー30rや早送りキー30s等を操作することにより、HDD32でハードディスク32aから読み取っている映像や音声等のデータを、その再生方向に対して逆方向や順方向に連続的に高速で再生させる、いわゆる、早戻し再生や早送り再生を行なうことができる。 Furthermore, by operating the fast reverse key 30r, fast forward key 30s, etc. of the remote controller 30, the data such as video and audio read from the hard disk 32a by the HDD 32 is continuously reversed or forward with respect to the reproduction direction. Thus, so-called fast reverse playback and fast forward playback can be performed.
 図4は、上記曲検索処理部28cが行なう曲検索処理の一例を機能ブロックで示している。すなわち、曲検索要求が曲検索指示部34に供給されると、曲検索指示部34は、音声切り出し部35に対して切り出し要求を発生し、音声切り出し部35は、上記信号処理部22に対して音声信号の取得を要求する。 FIG. 4 shows an example of a music search process performed by the music search processing unit 28c using functional blocks. That is, when a song search request is supplied to the song search instruction unit 34, the song search instruction unit 34 generates a cut-out request to the voice cut-out unit 35, and the voice cut-out unit 35 sends a request to the signal processing unit 22. And request acquisition of the audio signal.
 すると、信号処理部22では、音声ストリーム取得部36が、TSから分離した音声ストリームの中から所定区間の音声ストリームを取得し、音声デコード部37に出力する。この音声デコード部37は、入力された音声ストリームにデコード処理を施して、例えばPCM(pulse code modulation)化されたデジタルの音声信号を生成し、音声蓄積部38に出力する。 Then, in the signal processing unit 22, the audio stream acquisition unit 36 acquires an audio stream of a predetermined section from the audio stream separated from the TS, and outputs it to the audio decoding unit 37. The audio decoding unit 37 performs a decoding process on the input audio stream, generates a digital audio signal converted into, for example, PCM (pulse code modulation), and outputs the digital audio signal to the audio storage unit 38.
 そして、この音声蓄積部38が、入力された音声信号を所定区間分蓄積して、音声切り出し部35に出力することにより、音声切り出し部35に所定区間分の切り出された音声信号が供給される。その後、音声切り出し部35は、入力された音声信号を特徴量生成部39に出力する。この特徴量生成部39は、入力された音声信号から曲目検索を行なうために必要な特徴量を生成し、サーバ14に送信する。 Then, the voice accumulation unit 38 accumulates the input voice signal for a predetermined section and outputs it to the voice cutout unit 35, whereby the voice signal cut out for the predetermined section is supplied to the voice cutout unit 35. . Thereafter, the voice cutout unit 35 outputs the input voice signal to the feature value generation unit 39. The feature value generation unit 39 generates a feature value necessary for performing a song search from the input audio signal, and transmits the feature value to the server 14.
 これにより、サーバ14は、受信した特徴量に該当する曲目の検索を行ない、その検索結果を受信端末13に返信する。そして、サーバ14から返信された検索結果は、受信端末13の検索結果取得部40に取得され、検索結果判定部41で有効か否かを判断された後、検索結果蓄積部42に蓄積され、必要に応じて取り出されて表示等に供されるようになっている。 As a result, the server 14 searches for the music corresponding to the received feature value, and returns the search result to the receiving terminal 13. The search result returned from the server 14 is acquired by the search result acquisition unit 40 of the receiving terminal 13, and after it is determined whether or not the search result determination unit 41 is valid, the search result is stored in the search result storage unit 42. They are taken out as necessary and used for display or the like.
 図5は、上記した曲検索処理部28cが行なう曲検索処理動作の一例をまとめたフローチャートを示している。すなわち、処理が開始(ステップS5a)されると、曲検索処理部28cは、ステップS5bで、所定区間分の音声信号の切り出しを行ない、ステップS5cで、切り出した音声信号から曲目検索を行なうために必要な特徴量を生成し、ステップS5dで、生成した特徴量をサーバ14に送信する。 FIG. 5 shows a flowchart summarizing an example of the music search processing operation performed by the music search processing unit 28c. That is, when the process is started (step S5a), the song search processing unit 28c cuts out the audio signal for a predetermined section in step S5b, and performs a song search from the cut out audio signal in step S5c. Necessary feature amounts are generated, and the generated feature amounts are transmitted to the server 14 in step S5d.
 すると、サーバ14は、ステップS5eで、受信した特徴量に基づいて曲目の検索を行ない、その検索結果を受信端末13に送信する。これにより、曲検索処理部28cは、ステップS5fで、検索結果を受信して判定し、ステップS5gで、検索結果を蓄積して、処理を終了(ステップS5h)する。 Then, in step S5e, the server 14 searches for the music piece based on the received feature amount, and transmits the search result to the receiving terminal 13. Thereby, the music search processing unit 28c receives and determines the search result in step S5f, accumulates the search result in step S5g, and ends the process (step S5h).
 そして、この実施の形態では、上記した曲検索処理部28cによる曲検索処理の検索結果を利用して、曲解析処理部28dが、ハードディスク32aに記録された番組コンテンツを構成する音声信号について、そこに含まれる音楽情報の再生状態、つまり、曲目や曲区間等を解析処理するようにしている。 In this embodiment, the song analysis processing unit 28d uses the search result of the song search processing by the song search processing unit 28c to determine the audio signal constituting the program content recorded in the hard disk 32a. The reproduction state of the music information included in the music, that is, the music piece, the music section, and the like are analyzed.
 図6は、曲検索処理部28cによる曲検索処理の検索結果を利用して、曲解析処理部28dがハードディスク32aに記録された番組コンテンツを構成する音声信号から音楽情報を解析する処理動作の一例をまとめたフローチャートを示している。すなわち、処理が開始(ステップS6a)されると、曲解析処理部28dは、ステップS6bで、解析対象となる番組コンテンツを構成する音声信号に対し、その先頭から終端までについて一定の時間間隔(Intvl[sec]:例えば30秒~1分等)で、曲検索処理部28cに曲検索処理を実行させる。 FIG. 6 shows an example of a processing operation in which the music analysis processing unit 28d uses the search result of the music search processing by the music search processing unit 28c to analyze the music information from the audio signal constituting the program content recorded on the hard disk 32a. The flowchart which summarized these is shown. That is, when the process is started (step S6a), the song analysis processing unit 28d, in step S6b, performs a certain time interval (Intvl) from the beginning to the end of the audio signal constituting the program content to be analyzed. [Sec]: For example, in 30 seconds to 1 minute), the music search processing unit 28c is caused to execute the music search processing.
 この場合、曲検索処理に利用する音声データ長(曲検索データ長)をAudLen[sec]とし、曲検索処理部28cが曲検索処理を行なうために必要な最低限の音声データ長をLmin[sec]とすると、
   AudLen=Lmin×2
なる関係を満たすものとする。また、曲検索の実施間隔Intvl[sec]と曲検索データ長AudLen[sec]とは、
   Intvl=AudLen
なる関係を満たすものとする。
In this case, the audio data length (music search data length) used for the music search process is AudLen [sec], and the minimum audio data length necessary for the music search processing unit 28c to perform the music search process is Lmin [sec. ]
AudLen = Lmin × 2
Satisfy the relationship. The song search interval Intvl [sec] and song search data length AudiLen [sec]
Intvl = AudLen
Satisfy the relationship.
 ここで、曲検索データ長AudLen[sec]を、曲検索処理を行なうために最低限必要な音声データ長Lmin[sec]の2倍に設定したのは、曲の変わり目付近で曲検索が行なわれる場合、つまり、図7(a),(b),(c)に示すように、曲Aの後に曲Bが連続する音声信号について曲検索処理が行なわれる場合を考慮してのことである。 Here, the song search data length AudLen [sec] is set to twice the minimum voice data length Lmin [sec] necessary for performing the song search processing, so that the song search is performed near the change of the song. In other words, as shown in FIGS. 7A, 7B, and 7C, the case where the music search process is performed on the audio signal in which the music B continues after the music A is taken into consideration.
 すなわち、AudLen[sec]をLmin[sec]の2倍に設定すれば、曲A,Bの変わり目において、曲検索データ長AudLen[sec]の中に、図7(a)に示すように曲Aの音声信号のみを含むLmin[sec]を確保することができる場合と、図7(b)に示すように曲Aの音声信号のみを含むLmin[sec]と曲Bの音声信号のみを含むLmin[sec]とを確保することができる場合と、図7(c)に示すように曲Bの音声信号のみを含むLmin[sec]を確保することができる場合とが生じることになる。 That is, if AudLen [sec] is set to be twice as long as Lmin [sec], the song A and B at the change of the song search data length AudLen [sec] as shown in FIG. Lmin [sec] including only the audio signal of the song A, and Lmin [sec] including only the audio signal of the song A and Lmin including only the audio signal of the song B as shown in FIG. There are cases where [sec] can be secured and cases where Lmin [sec] including only the audio signal of the music piece B can be secured as shown in FIG.
 そして、曲検索データ長AudLen[sec]の中に、曲Aの音声信号のみを含むLmin[sec]を確保することができた場合、その曲検索結果は曲Aのみの1候補となる。また、曲検索データ長AudLen[sec]の中に、曲Aの音声信号のみを含むLmin[sec]と曲Bの音声信号のみを含むLmin[sec]とを確保することができた場合、その曲検索結果は曲Aと曲Bとの2候補となる。さらに、曲検索データ長AudLen[sec]の中に、曲Bの音声信号のみを含むLmin[sec]を確保することができた場合、その曲検索結果は曲Bのみの1候補となる。 If Lmin [sec] including only the audio signal of the music A can be secured in the music search data length AudLen [sec], the music search result is one candidate for the music A only. In addition, in the music search data length AudLen [sec], when Lmin [sec] including only the audio signal of the music A and Lmin [sec] including only the audio signal of the music B can be secured, The song search results are two candidates, song A and song B. Furthermore, when Lmin [sec] including only the audio signal of the song B can be secured in the song search data length AudLen [sec], the song search result is only one candidate for the song B.
 なお、曲検索データ長AudLen[sec]を、曲検索処理を行なうために最低限必要な音声データ長Lmin[sec]の2倍よりも短く、つまり、
   AudLen<Lmin×2
に設定すると、曲A,Bの変わり目において、曲検索データ長AudLen[sec]の中に、曲Aまたは曲Bの音声信号のみを含むLmin[sec]を確保することができない事態が生じ、この場合、信頼性の高い曲検索結果が得られなくなる。また、音楽とともに人の喋り声が含まれているように場合にも、曲検索に失敗することがある。この対策として、曲検索データ長AudLen[sec]を長くすることが有効であることが分かっており、このことから、AudLen[sec]をLmin[sec]の2倍に設定している。
Note that the song search data length AudLen [sec] is shorter than twice the minimum voice data length Lmin [sec] necessary for performing the song search process, that is,
AudLen <Lmin × 2
If set to, a situation occurs in which Lmin [sec] including only the audio signal of the music A or B cannot be secured in the music search data length AudLen [sec] at the turn of the music A and B. In this case, a reliable music search result cannot be obtained. In addition, the music search may fail even when the human voice is included with the music. As a countermeasure, it has been found that it is effective to lengthen the music search data length AudLen [sec]. Therefore, AudLen [sec] is set to be twice as long as Lmin [sec].
 次に、曲解析処理部28dは、ステップS6cで、一定の時間間隔Intvl[sec]で曲検索処理を行なって得られた曲検索結果を1つずつ参照し、解析対象とすべきか否かを判定する、いわゆる、曲区間の特定処理を実行する。また、解析対象と判断された曲区間について、曲の開始位置と終了位置とを推定する、いわゆる、曲区間の解析処理を実行する。なお、曲区間の特定処理と解析処理とについては、詳細を後述する。 Next, in step S6c, the song analysis processing unit 28d refers to the song search results obtained by performing the song search processing at a constant time interval Intvl [sec] one by one, and determines whether or not the analysis is to be performed. A so-called song segment identification process is performed. In addition, a so-called song segment analysis process is performed for estimating a song start position and end position for a song segment determined to be analyzed. Details of the song segment identification processing and analysis processing will be described later.
 その後、曲解析処理部28dは、ステップS6dで、上記ステップS6cの処理で未特定となっている曲区間について、区間の特定処理と解析処理とを実行する。なお、この未特定区間に対する特定処理と解析処理とについても、詳細を後述する。 Thereafter, in step S6d, the song analysis processing unit 28d executes the zone specifying process and the analysis process for the song section that has not been specified in the process of step S6c. Details of the specifying process and the analyzing process for the unspecified section will be described later.
 そして、曲解析処理部28dは、ステップS6eで、曲検索結果である文字列に対してフィルタリング処理を実行する。すなわち、曲検索結果は、サーバ14から文字列で返信される。この場合、同じ曲でもデータベース上の表記が複数存在すると、異なる曲検索結果として返信される場合がある。例えば、アルファベット表記が大文字の場合と小文字の場合や、フォントが全角か半角かの違い等である。この実施の形態では、曲検索結果である文字列に対して既存のフィルタリング処理を行なうことにより、表記上の違いを丸め込むようにしている。 Then, the song analysis processing unit 28d performs a filtering process on the character string that is the song search result in step S6e. That is, the song search result is returned as a character string from the server 14. In this case, even if the same song has a plurality of notations on the database, it may be returned as a different song search result. For example, there are cases where the alphabet is in uppercase and lowercase letters, or whether the font is full-width or half-width. In this embodiment, the existing filtering process is performed on the character string that is the music search result, thereby rounding off the difference in notation.
 その後、曲解析処理部28dは、ステップS6fで、文字列に対するフィルタリング処理後の解析結果を、メモリ部28bやハードディスク32a等に保存して、処理を終了(ステップS6g)する。そして、保存された解析結果は、例えば、ユーザによる操作部29やリモートコントローラ30の操作に基づいて、上記映像処理部26に一覧表示することができる。 Thereafter, in step S6f, the music analysis processing unit 28d stores the analysis result after the filtering process on the character string in the memory unit 28b, the hard disk 32a, etc., and ends the processing (step S6g). The stored analysis results can be displayed as a list on the video processing unit 26 based on, for example, the operation of the operation unit 29 or the remote controller 30 by the user.
 これにより、上記ハードディスク32aに記録された番組コンテンツを構成する音声信号について、そこに含まれる音楽情報の再生状態、つまり、曲目や曲区間等を迅速に解析処理することができ、ひいては、ユーザが記録コンテンツに含まれる各種の曲を予め認識して、記録コンテンツを視聴する際の取り扱いを便利にすることができる。 As a result, it is possible to quickly analyze the reproduction state of the music information included in the audio signal constituting the program content recorded on the hard disk 32a, that is, the music piece, the music section, and the like. Various kinds of music included in the recorded content can be recognized in advance, and handling when viewing the recorded content can be made convenient.
 なお、図6で説明した処理動作では、一定の時間間隔Intvl[sec]で曲検索処理部28cに曲検索処理を実行させるようにしたが、曲検索処理を行なう時間間隔としては、番組コンテンツ毎に変えたり、同じ番組コンテンツ内で変えたりする等、場面毎に所定の時間間隔に変化させることができる。 In the processing operation described with reference to FIG. 6, the music search processing unit 28c is caused to execute the music search processing at a constant time interval Intvl [sec]. It can be changed at a predetermined time interval for each scene, such as changing to the same program content.
 図8は、上記ステップS6cで行なわれる曲区間の特定処理動作及び解析処理動作の一例をまとめたフローチャートを示している。すなわち、処理が開始(ステップS8a)されると、曲解析処理部28dは、ステップS8bで、一定の時間間隔Intvl[sec]毎に曲検索処理を行なって得られた複数の曲検索結果のうち、予め定められた順序(例えば時間的に古い順等)に基づいて1つの曲検索結果を入力する。 FIG. 8 shows a flow chart summarizing an example of the song segment identification processing operation and analysis processing operation performed in step S6c. That is, when the process is started (step S8a), the song analysis processing unit 28d performs a song search process at a predetermined time interval Intvl [sec] in step S8b, among a plurality of song search results obtained. Then, one music search result is input based on a predetermined order (for example, oldest order in time).
 その後、曲解析処理部28dは、ステップS8cで、入力した曲検索結果が何も特定の曲目を示していない、いわゆる、結果なしの状態であるか否かを判別する。そして、結果なしの状態であると判断された場合(YES)、曲解析処理部28dは、ステップS8dで、その曲検索結果を上記ステップS6dで行なわれる未特定区間での解析処理に回し、ステップS8bの処理に戻される。 Thereafter, in step S8c, the music analysis processing unit 28d determines whether or not the input music search result is a so-called no result state in which no specific music is shown. If it is determined that there is no result (YES), in step S8d, the song analysis processing unit 28d sends the song search result to the analysis processing in the unspecified section performed in step S6d. The process returns to S8b.
 これに対し、上記ステップS8cで、入力した曲検索結果が特定の曲目を示している、つまり、結果なしの状態でないと判断された場合(NO)、曲解析処理部28dは、ステップS8eで、曲検索結果として得られた曲の数、つまり、候補曲数が2曲以上か否かを判別し、2曲以上であると判断された場合(YES)、ステップS8dで、その曲検索結果を上記ステップS6dで行なわれる未特定区間での解析処理に回し、ステップS8bの処理に戻される。 On the other hand, if it is determined in step S8c that the input song search result indicates a specific song, that is, it is determined that there is no result (NO), the song analysis processing unit 28d, in step S8e, The number of songs obtained as a song search result, that is, whether or not the number of candidate songs is two or more is determined. If it is determined that there are two or more songs (YES), the song search results are displayed in step S8d. The analysis process is performed in the unspecified section performed in step S6d, and the process returns to step S8b.
 一方、上記ステップS8eで、候補曲数が2曲以上でないと判断された場合(NO)、曲解析処理部28dは、ステップS8fで、入力した曲検索結果がその時間的に前または後に得られた曲検索結果と同じであるか否かを判別し、同じであると判断された場合(YES)、ステップS8gで、両方の曲検索結果が得られた区間を同じ曲の存在する区間として結合し、ステップS8bの処理に戻される。 On the other hand, if it is determined in step S8e that the number of candidate songs is not two or more (NO), the song analysis processing unit 28d obtains the input song search result before or after that in step S8f. It is determined whether or not it is the same as the music search result, and if it is determined that it is the same (YES), in step S8g, the sections in which both music search results are obtained are combined as sections where the same music exists. Then, the process returns to step S8b.
 また、上記ステップS8fで、入力した曲検索結果がその時間的に前または後に得られた曲検索結果と同じでないと判断された場合(NO)、曲解析処理部28dは、ステップS8hで、その曲検索結果が示す曲の開始位置または終了位置を推定する、いわゆる、曲区間の解析処理(詳細は後述)を実行する。 If it is determined in step S8f that the input song search result is not the same as the song search result obtained before or after that time (NO), the song analysis processing unit 28d, in step S8h, A so-called song section analysis process (details will be described later) for estimating the start position or end position of the song indicated by the song search result is executed.
 その後、曲解析処理部28dは、ステップS8iで、曲検索処理を行なって得られた全ての曲検索結果について、曲区間の特定処理や解析処理が完了したか否かを判別し、完了していないと判断された場合(NO)、ステップS8bの処理に戻され、完了したと判断された場合(YES)、処理を終了(ステップS8j)する。 Thereafter, in step S8i, the song analysis processing unit 28d determines whether or not the song segment identification processing and analysis processing have been completed for all the song search results obtained by performing the song search processing. If it is determined that there is not (NO), the process returns to step S8b, and if it is determined that the process is completed (YES), the process ends (step S8j).
 図9(a),(b)は、図8で説明した曲区間の特定処理及び解析処理を具体的に示している。例えば、図9(a)に示すように、一定の時間間隔Intvl[sec]毎に得られる曲検索結果のうち、時刻T1で得られた曲検索結果が曲TNsを示し、その次の時刻T2で得られた曲検索結果も同じ曲TNsを示している場合(ステップS8fでYES)、曲解析処理部28dは、時刻T1で曲検索結果(TNs)を得るために使用された区間(Intvl[sec])と、時刻T2で曲検索結果(TNs)を得るために使用された区間(Intvl[sec])とを、同じ曲TNsの存在する区間とみなす(ステップS8g)ようにして、曲区間を特定するようにしている。 FIGS. 9A and 9B specifically show the song segment identification processing and analysis processing described in FIG. For example, as shown in FIG. 9A, among the song search results obtained at regular time intervals Intvl [sec], the song search result obtained at time T1 indicates the song TNs, and the next time T2 When the music search result obtained in step S8f also indicates the same music TNs (YES in step S8f), the music analysis processing unit 28d uses the section (Intvl [ sec]) and the section (Intvl [sec]) used to obtain the song search result (TNs) at time T2 is regarded as a section in which the same song TNs exists (step S8g). To identify.
 また、時刻T2で得られた曲検索結果が曲TNsを示し、その次の時刻T3で得られた曲検索結果が曲TNsとは異なる曲TNを示している場合(ステップS8fでNO)、曲解析処理部28dは、時刻T2とT3との間で曲TNsが終了していると判断し、時刻T2とT3との間において、曲TNsの終了位置を推定する(ステップS8h)、曲区間の解析処理を行なうようにしている。 If the song search result obtained at time T2 indicates the song TNs and the song search result obtained at the next time T3 indicates a song TN different from the song TNs (NO in step S8f), The analysis processing unit 28d determines that the song TNs ends between times T2 and T3, and estimates the end position of the song TNs between times T2 and T3 (step S8h). Analysis processing is performed.
 一方、図9(b)に示すように、一定の時間間隔Intvl[sec]毎に得られる曲検索結果のうち、時刻T1で得られた曲検索結果が曲TNを示し、その次の時刻T2で得られた曲検索結果が曲TNとは異なる曲TNeを示し、その次の時刻T3で得られた曲検索結果も同じ曲TNeを示している場合には、曲解析処理部28dは、時刻T2で曲検索結果(TNe)を得るために使用された区間(Intvl[sec])と、時刻T3で曲検索結果(TNe)を得るために使用された区間(Intvl[sec])とを、同じ曲TNeの存在する区間とみなして曲区間を特定するとともに、時刻T1とT2との間で曲TNeが開始していると判断し、時刻T1とT2との間で曲TNeの開始位置を推定する、曲区間の解析処理を行なうようにしている。 On the other hand, as shown in FIG. 9B, among the music search results obtained at regular time intervals Intvl [sec], the music search result obtained at time T1 indicates the music TN, and the next time T2 If the song search result obtained in step S3 indicates a song TNe different from the song TN, and the song search result obtained at the next time T3 also indicates the same song TNe, the song analysis processing unit 28d The section (Intvl [sec]) used to obtain the song search result (TNe) at T2 and the section (Intvl [sec]) used to obtain the song search result (TNe) at time T3, The song section is identified as a section where the same song TNe exists, and the song TNe is determined to start between the times T1 and T2, and the start position of the song TNe is determined between the times T1 and T2. An estimation processing of the song section is performed.
 図10は、図9(a)において、時刻T2とT3との間で曲TNsの終了位置を推定する解析処理動作をまとめたフローチャートを示している。すなわち、処理が開始(ステップS10a)されると、曲解析処理部28dは、ステップS10bで、曲検索結果として曲TNsが得られた時刻T2に対応する音声信号上の位置Nsと、曲検索結果として曲TNが得られた時刻T3に対応する音声信号上の位置Nとの中間位置N_midを、下式により算出する。 FIG. 10 shows a flowchart summarizing the analysis processing operation for estimating the end position of the song TNs between times T2 and T3 in FIG. 9A. That is, when the process is started (step S10a), the song analysis processing unit 28d, in step S10b, the position Ns on the audio signal corresponding to the time T2 when the song TNs was obtained as the song search result, and the song search result. As a result, an intermediate position N_mid with respect to the position N on the audio signal corresponding to the time T3 when the song TN is obtained is calculated by the following equation.
   N_mid=(Ns+N)/2
 そして、曲解析処理部28dは、ステップS10cで、位置Nsと中間位置N_midとの間の時間差が、予め設定された閾値(例えば1秒)以下であるか否かを判別し、1秒以下であると判断された場合(YES)、ステップS10dで、位置Nsまたは中間位置N_midを曲TNsの終了位置に決定して、処理を終了(ステップS10e)する。
N_mid = (Ns + N) / 2
Then, in step S10c, the music analysis processing unit 28d determines whether or not the time difference between the position Ns and the intermediate position N_mid is equal to or less than a preset threshold value (for example, 1 second). If it is determined that there is (YES), in step S10d, the position Ns or the intermediate position N_mid is determined as the end position of the music TNs, and the process is ended (step S10e).
 また、上記ステップS10cで位置Nsと中間位置N_midとの間の時間差が1秒以下でないと判断された場合(NO)、曲解析処理部28dは、ステップS10fで、中間位置N_midにおいて曲検索処理部28cに曲検索処理を行なわせ、曲検索結果として曲TN_midを得る。 On the other hand, if it is determined in step S10c that the time difference between the position Ns and the intermediate position N_mid is not less than 1 second (NO), the music analysis processing unit 28d performs the music search processing unit at the intermediate position N_mid in step S10f. The music search process is performed in 28c, and the music TN_mid is obtained as a music search result.
 その後、曲解析処理部28dは、ステップS10gで、中間位置N_midでの曲検索結果の示す曲TN_midが、位置Nsでの曲検索結果の示す曲TNsと同じか否かを判別し、同じであると判断された場合(YES)、ステップS10hで、位置Nsを中間位置N_midに更新して、ステップS10bの処理に戻される。 Thereafter, in step S10g, the song analysis processing unit 28d determines whether or not the song TN_mid indicated by the song search result at the intermediate position N_mid is the same as the song TNs indicated by the song search result at the position Ns. (YES), the position Ns is updated to the intermediate position N_mid in step S10h, and the process returns to step S10b.
 また、上記ステップS10gで、中間位置N_midでの曲検索結果の示す曲TN_midが、位置Nsでの曲検索結果の示す曲TNsと同じでないと判断された場合(NO)、曲解析処理部28dは、ステップS10iで、位置Nを中間位置N_midに更新して、ステップS10bの処理に戻される。 When it is determined in step S10g that the song TN_mid indicated by the song search result at the intermediate position N_mid is not the same as the song TNs indicated by the song search result at the position Ns (NO), the song analysis processing unit 28d In step S10i, the position N is updated to the intermediate position N_mid, and the process returns to step S10b.
 すなわち、位置Nsと位置Nとの間に曲TNsの終了位置が存在する場合、位置Nsと位置Nとの中間位置N_midで曲検索を行ない、その結果として曲TNsが得られたか否かにより、位置Nsを中間位置N_midに更新して位置Nとの中間位置での曲検索を行なうか、位置Nを中間位置N_midに更新して位置Nsとの中間位置での曲検索を行なうかを決定するという動作を繰り返し、位置Nsと中間位置N_midとの差が所定の閾値以下になったとき、位置Nsまたは中間位置N_midを曲TNsの終了位置として決定している。 That is, when the end position of the music TNs exists between the position Ns and the position N, a music search is performed at an intermediate position N_mid between the position Ns and the position N, and as a result, the music TNs is obtained. It is determined whether the position Ns is updated to the intermediate position N_mid and the music search at the intermediate position with the position N is performed, or the position N is updated to the intermediate position N_mid and the music search at the intermediate position with the position Ns is performed. When the difference between the position Ns and the intermediate position N_mid is less than or equal to a predetermined threshold value, the position Ns or the intermediate position N_mid is determined as the end position of the song TNs.
 図11は、図9(b)において、時刻T1とT2との間で曲TNeの開始位置を推定する解析処理動作をまとめたフローチャートを示している。すなわち、処理が開始(ステップS11a)されると、曲解析処理部28dは、ステップS11bで、曲検索結果として曲TNが得られた時刻T1に対応する音声信号上の位置Nと、曲検索結果として曲TNeが得られた時刻T2に対応する音声信号上の位置Neとの中間位置Ne_midを、下式により算出する。 FIG. 11 shows a flowchart summarizing the analysis processing operation for estimating the start position of the song TNe between times T1 and T2 in FIG. 9B. That is, when the process is started (step S11a), the song analysis processing unit 28d, in step S11b, the position N on the audio signal corresponding to the time T1 when the song TN was obtained as the song search result, and the song search result. As an intermediate position Ne_mid with respect to the position Ne on the audio signal corresponding to the time T2 at which the song TNe is obtained as follows.
   Ne_mid=(Ne+N)/2
 そして、曲解析処理部28dは、ステップS11cで、位置Neと中間位置Ne_midとの間の時間差が、予め設定された閾値(例えば1秒)以下であるか否かを判別し、1秒以下であると判断された場合(YES)、ステップS11dで、位置Neまたは中間位置Ne_midを曲TNeの開始位置に決定して、処理を終了(ステップS11e)する。
Ne_mid = (Ne + N) / 2
Then, in step S11c, the music analysis processing unit 28d determines whether or not the time difference between the position Ne and the intermediate position Ne_mid is equal to or less than a preset threshold value (for example, 1 second). If it is determined that there is (YES), in step S11d, the position Ne or the intermediate position Ne_mid is determined as the start position of the music TNe, and the process ends (step S11e).
 また、上記ステップS11cで位置Neと中間位置Ne_midとの間の時間差が1秒以下でないと判断された場合(NO)、曲解析処理部28dは、ステップS11fで、中間位置Ne_midにおいて曲検索処理部28cに曲検索処理を行なわせ、曲検索結果として曲TNe_midを得る。 If it is determined in step S11c that the time difference between the position Ne and the intermediate position Ne_mid is not less than 1 second (NO), the music analysis processing unit 28d performs the music search processing unit at the intermediate position Ne_mid in step S11f. The music search process is performed in 28c, and the music TNE_mid is obtained as a music search result.
 その後、曲解析処理部28dは、ステップS11gで、中間位置Ne_midでの曲検索結果の示す曲TNe_midが、位置Neでの曲検索結果の示す曲TNeと同じか否かを判別し、同じであると判断された場合(YES)、ステップS11hで、位置Neを中間位置Ne_midに更新して、ステップS11bの処理に戻される。 Thereafter, in step S11g, the music analysis processing unit 28d determines whether or not the music TNe_mid indicated by the music search result at the intermediate position Ne_mid is the same as the music T Ne indicated by the music search result at the position Ne. If it is determined (YES), the position Ne is updated to the intermediate position Ne_mid in step S11h, and the process returns to step S11b.
 また、上記ステップS11gで、中間位置Ne_midでの曲検索結果の示す曲TNe_midが、位置Neでの曲検索結果の示す曲TNeと同じでないと判断された場合(NO)、曲解析処理部28dは、ステップS11iで、位置Nを中間位置Ne_midに更新して、ステップS11bの処理に戻される。 When it is determined in step S11g that the song TNe_mid indicated by the song search result at the intermediate position Ne_mid is not the same as the song TNe indicated by the song search result at the position Ne (NO), the song analysis processing unit 28d In step S11i, the position N is updated to the intermediate position Ne_mid, and the process returns to step S11b.
 すなわち、位置Nと位置Neとの間に曲TNeの開始位置が存在する場合、位置Nと位置Neとの中間位置Ne_midで曲検索を行ない、その結果として曲TNeが得られたか否かにより、位置Neを中間位置Ne_midに更新して位置Nとの中間位置での曲検索を行なうか、位置Nを中間位置Ne_midに更新して位置Neとの中間位置での曲検索を行なうかを決定するという動作を繰り返し、位置Neと中間位置Ne_midとの差が所定の閾値以下になったとき、位置Neまたは中間位置Ne_midを曲TNeの開始位置として決定している。 That is, when the start position of the music TNe exists between the position N and the position Ne, a music search is performed at an intermediate position Ne_mid between the position N and the position Ne, and as a result, whether or not the music TNe is obtained, It is determined whether the position Ne is updated to the intermediate position Ne_mid and the music search at the intermediate position with the position N is performed, or the position N is updated to the intermediate position Ne_mid and the music search at the intermediate position with the position Ne is performed. When the difference between the position Ne and the intermediate position Ne_mid is equal to or less than a predetermined threshold value, the position Ne or the intermediate position Ne_mid is determined as the start position of the song TNe.
 図12乃至図15は、上記ステップS6dで行なわれる未特定曲区間の特定処理動作及び解析処理動作の一例をまとめたフローチャートを示している。すなわち、処理が開始(ステップS12a)されると、曲解析処理部28dは、ステップS12bで、上記ステップS8dで未特定区間での解析処理に回された曲検索結果が存在するか否かを判別し、存在しないと判断された場合(NO)、処理を終了(ステップS12c)する。 FIGS. 12 to 15 are flowcharts summarizing an example of the specifying process operation and the analyzing process operation of the unspecified music section performed in step S6d. That is, when the process is started (step S12a), the song analysis processing unit 28d determines whether or not there is a song search result that was sent to the analysis process in the unspecified section in step S8d in step S12b. If it is determined that it does not exist (NO), the process ends (step S12c).
 また、上記ステップS12bで未特定区間での解析処理に回された曲検索結果が存在すると判断された場合(YES)、曲解析処理部28dは、ステップS12dで、予め定められた順序(例えば時間的に古い順等)に基づいて1つの曲検索結果を入力し、ステップS12eで、入力した曲検索結果が何も特定の曲目を示していない、いわゆる、結果なしの状態であるか否かを判別する。 If it is determined in step S12b that there is a music search result sent to the analysis process in the unspecified section (YES), the music analysis processing unit 28d determines in step S12d that a predetermined order (eg, time In accordance with the oldest etc.), and in step S12e, it is determined whether or not the input song search result indicates that there is no specific song, so-called no result. Determine.
 そして、結果なしの状態であると判断された場合(YES)、曲解析処理部28dは、ステップS13aで、その曲検索結果の時間的に前後に得られた曲検索結果が、既に曲区間または非音楽区間として解析済みであるか否かを判別し、解析済みであると判断された場合(YES)、ステップS13bで、当該曲検索結果を得るために使用された区間を非音楽区間に確定して、ステップS12bの処理に戻される。 If it is determined that there is no result (YES), the song analysis processing unit 28d determines that the song search results obtained before and after the song search result are already in the song section or in step S13a. It is determined whether or not it has been analyzed as a non-music section. If it is determined that it has been analyzed (YES), the section used to obtain the song search result is determined as a non-music section in step S13b. Then, the process returns to step S12b.
 また、上記ステップS13aで、入力した曲検索結果の時間的に前後に得られた曲検索結果が解析済みでないと判断された場合(NO)、曲解析処理部28dは、ステップS13cで、前後に得られた曲検索結果が何も特定の曲目を示していない、いわゆる、結果なしの状態であるか否かを判別する。 If it is determined in step S13a that the song search result obtained before and after the input song search result has not been analyzed (NO), the song analysis processing unit 28d performs step S13c before and after. It is determined whether or not the obtained music search result is a so-called no result state in which no specific music is indicated.
 そして、結果なしの状態でないと判断された場合(NO)、曲解析処理部28dは、ステップS13dで、先に図10で説明した解析処理動作を利用して非音楽区間の終了位置を推定し、ステップS12bの処理に戻される。 If it is determined that there is no result (NO), the song analysis processing unit 28d estimates the end position of the non-music section using the analysis processing operation described above with reference to FIG. 10 in step S13d. The process returns to step S12b.
 また、上記ステップS13cで、入力した曲検索結果の時間的に前後に得られた曲検索結果が結果なしの状態であると判断された場合(YES)、曲解析処理部28dは、ステップS13eで、その曲検索結果が得られた音声信号上の位置と、その前または後の曲検索結果が得られた音声信号上の位置との中間位置において、曲検索処理部28cに曲検索処理を行なわせる。 If it is determined in step S13c that the song search result obtained before and after the input song search result is in a state of no result (YES), the song analysis processing unit 28d determines in step S13e. The song search processing unit 28c performs song search processing at an intermediate position between the position on the audio signal from which the song search result is obtained and the position on the audio signal from which the previous or subsequent song search result is obtained. Make it.
 そして、曲解析処理部28dは、ステップS13fで、中間位置での曲検索結果が何も特定の曲目を示していない、いわゆる、結果なしの状態であるか否かを判別し、結果なしの状態でないと判断された場合(NO)、ステップS13gで、その曲検索結果を得るために使用された区間を未特定の区間として新規に追加し、ステップS12bの処理に戻される。 Then, in step S13f, the song analysis processing unit 28d determines whether or not the song search result at the intermediate position is a so-called no-result state in which no particular song is indicated, and there is no result. If it is determined that it is not (NO), in step S13g, the section used to obtain the song search result is newly added as an unspecified section, and the process returns to step S12b.
 また、上記ステップS13fで中間位置での曲検索結果が結果なしの状態であると判断された場合(YES)、曲解析処理部28dは、ステップS13hで、入力した曲検索結果を得るために使用された区間と、その前後の曲検索結果を得るために使用された区間とを非音楽区間として結合し、ステップS12bの処理に戻される。 If it is determined in step S13f that the result of the song search at the intermediate position is in a state of no result (YES), the song analysis processing unit 28d is used to obtain the input song search result in step S13h. The obtained sections and the sections used for obtaining the music search results before and after the combined sections are combined as non-music sections, and the process returns to step S12b.
 次に、上記ステップS12eで、入力した曲検索結果が結果なしの状態でないと判断された場合(NO)、曲解析処理部28dは、ステップS12fで、曲検索結果として得られた曲の数、つまり、候補曲数が1曲であるか否かを判別し、1曲であると判断された場合(YES)、ステップS12gで、先に図10及び図11で説明した解析処理動作を利用して、候補曲の終了位置及び開始位置を推定し、ステップS12bの処理に戻される。 Next, when it is determined in step S12e that the input song search result is not in a state of no result (NO), the song analysis processing unit 28d determines the number of songs obtained as a song search result in step S12f, That is, it is determined whether or not the number of candidate songs is one, and if it is determined that there is one song (YES), in step S12g, the analysis processing operation described above with reference to FIGS. 10 and 11 is used. Thus, the end position and start position of the candidate song are estimated, and the process returns to step S12b.
 また、上記ステップS12fで候補曲数が1曲でないと判断された場合(NO)、曲解析処理部28dは、ステップS12hで、候補曲数が2曲であるか否かを判別し、2曲でない、つまり、3曲以上であると判断された場合(NO)、ステップS14aで、その候補曲の中に、入力した曲検索結果の前後の曲検索結果で得られた曲と同じ曲が存在するか否かを判別する。 If it is determined in step S12f that the number of candidate songs is not one (NO), the song analysis processing unit 28d determines whether or not the number of candidate songs is two in step S12h. If it is determined that there are three or more songs (NO), in step S14a, there is the same song as the song obtained from the song search results before and after the entered song search result in step S14a. It is determined whether or not to do.
 そして、存在すると判断された場合(YES)、曲解析処理部28dは、ステップS14bで、入力した曲検索結果を得るために使用された区間を、その前後の曲検索結果を得るために使用された区間に存在する曲と同じ曲区間として結合し、ステップS12bの処理に戻される。 If it is determined that it exists (YES), the song analysis processing unit 28d uses the section used for obtaining the input song search result in step S14b to obtain the song search result before and after that. Are combined as the same song section as that existing in the section, and the process returns to the process of step S12b.
 また、上記ステップS14aで、候補曲の中に、入力した曲検索結果の前後の曲検索結果で得られた曲と同じ曲が存在しないと判断された場合(NO)、曲解析処理部28dは、ステップS14cで、候補曲の中に、入力した曲検索結果の前後いずれかの曲検索結果で得られた曲と同じ曲が存在するか否かを判別する。 If it is determined in step S14a that the same song as the song obtained from the song search results before and after the input song search result does not exist in the candidate song (NO), the song analysis processing unit 28d In step S14c, it is determined whether or not the candidate song includes the same song as the song obtained from the song search result before or after the input song search result.
 そして、存在すると判断された場合(YES)、曲解析処理部28dは、ステップS14dで、入力した曲検索結果を得るために使用された区間を、その前後いずれかの曲検索結果を得るために使用された区間に存在する曲と同じ曲区間として結合し、ステップS12bの処理に戻される。 If it is determined that it exists (YES), the song analysis processing unit 28d obtains the song search result before or after the section used to obtain the entered song search result in step S14d. The music is combined as the same music section as the music existing in the used section, and the process returns to step S12b.
 また、上記ステップS14cで、候補曲の中に、入力した曲検索結果の前後いずれかの曲検索結果で得られた曲と同じ曲が存在しないと判断された場合(NO)、曲解析処理部28dは、ステップS14eで、候補曲数が3曲以上あった場合を考慮して、予め設定された優先度に基づいて決定される第1候補の曲を、入力した曲検索結果を得るために使用された区間に存在する曲であると推定し、ステップS12bの処理に戻される。 If it is determined in step S14c that the candidate song does not include the same song as the song obtained from the song search result before or after the input song search result (NO), the song analysis processing unit In step S14e, in order to obtain an input song search result for the first candidate song determined based on a preset priority in consideration of the case where there are three or more candidate songs in step S14e. It is estimated that the song exists in the used section, and the process returns to step S12b.
 一方、上記ステップS12hで、候補曲数が2曲であると判断された場合(YES)、曲解析処理部28dは、2曲のうちのいずれかを選択するのではなく、2つの曲が連続して存在すると判断し、2つの曲の境目を検索するように動作する。すなわち、曲解析処理部28dは、ステップS15aで、2つの候補曲が検索される区間上で曲検索開始位置を決定し、時刻カウンタをリセットする。この場合、曲検索開始位置は、2つの候補曲が検索される区間の先頭位置となる。 On the other hand, when it is determined in step S12h that the number of candidate songs is two (YES), the song analysis processing unit 28d does not select one of the two songs, but two songs are continuous. It operates to search for the boundary between two songs. That is, the song analysis processing unit 28d determines a song search start position on a section where two candidate songs are searched in step S15a, and resets the time counter. In this case, the song search start position is the head position of the section in which two candidate songs are searched.
 そして、曲解析処理部28dは、ステップS15bで、新たに蓄積された曲検索結果が、その前の曲検索結果に比べて変化したか否かを判別し、変化したと判断された場合(YES)、ステップS15cで、その変化した時点が曲の境目であると判別して、ステップS12bの処理に戻される。 Then, in step S15b, the song analysis processing unit 28d determines whether or not the newly stored song search result has changed compared to the previous song search result, and if it is determined that it has changed (YES) In step S15c, it is determined that the changed time is the boundary of the song, and the process returns to step S12b.
 また、上記ステップS15bで曲検索結果が変化していないと判断された場合(NO)、曲解析処理部28dは、ステップS15dで、上記先頭位置において曲検索データ長AudLen[sec]をLmin[sec]にして曲検索処理部28cに曲検索処理を行なわせ、ステップS15eで、その曲検索結果を蓄積する。 If it is determined in step S15b that the song search result has not changed (NO), the song analysis processing unit 28d sets the song search data length AudLen [sec] to Lmin [sec] at the head position in step S15d. The music search processing unit 28c performs music search processing, and the music search results are stored in step S15e.
 その後、曲解析処理部28dは、ステップS15fで、時刻カウンタにより1秒が計測されると、ステップS15gで、曲検索位置を一定量移動させて、ステップS15bの処理に戻される。この場合、曲検索位置は、2つの候補曲が検索される区間の終了位置に向けて一定量移動される。 Thereafter, when one second is measured by the time counter in step S15f, the music analysis processing unit 28d moves the music search position by a predetermined amount in step S15g and returns to the processing in step S15b. In this case, the music search position is moved by a certain amount toward the end position of the section in which two candidate songs are searched.
 図16は、上記HDD32によりハードディスク32aに記録された番組コンテンツについて、図6乃至図15で説明したような、曲検索結果を利用して音楽情報の再生状態を解析処理した結果の一例を示している。この一例に示す解析結果は、ユーザがリモートコントローラ30のメニューキー30jを操作して、階層構造となされた複数のメニュー画面に分け入り、解析結果の表示を要求することによって、上記映像表示部26に一覧表示することができる。 FIG. 16 shows an example of the result of analyzing the reproduction state of the music information using the song search results as described in FIGS. 6 to 15 for the program content recorded on the hard disk 32a by the HDD 32. Yes. The analysis result shown in this example is obtained by the user operating the menu key 30j of the remote controller 30 to divide into a plurality of menu screens having a hierarchical structure and requesting the display of the analysis result, thereby the video display unit 26 Can be listed.
 この例によれば、番組コンテンツの再生開始から、10分後、2時間55分後、3時間8分後、3時間29分後、4時間27分後、4時間44分後に、それぞれ、どのトラック名の曲が開始されるかが分かる。また、番組コンテンツの再生開始から、0分後、1時間48分後、3時間43分後、4時間58分後に、それぞれ、非音楽区間となることが分かる。これにより、ユーザが記録コンテンツに含まれる各種の曲を予め認識することができるので、記録コンテンツを視聴する際の取り扱いを便利にすることができる。 According to this example, 10 minutes, 2 hours and 55 minutes, 3 hours and 8 minutes, 3 hours and 29 minutes, 4 hours and 27 minutes, and 4 hours and 44 minutes after the start of program content playback, You can see if the song with the track name starts. In addition, it can be seen that after the start of program content playback, non-music intervals are reached after 0 minutes, 1 hour 48 minutes, 3 hours 43 minutes, and 4 hours 58 minutes. Thereby, since the user can recognize in advance various kinds of music included in the recorded content, it is possible to make handling convenient when viewing the recorded content.
 ところで、ハードディスク32aに記録された番組コンテンツについて、曲検索結果を利用して音楽情報の再生状態を解析処理した場合、先に述べたように、同じ時間に候補曲数が2曲ある場合がある。例えば、図17に示す解析結果のように、番組コンテンツの再生開始から1時間15分後に、2つの曲が候補として検索されることがある。このような場合に、図15で説明した処理を用いて2つの曲の境目を検索する。 By the way, for the program content recorded on the hard disk 32a, when the music information playback state is analyzed using the song search result, there may be two candidate songs at the same time as described above. . For example, as shown in the analysis result shown in FIG. 17, two songs may be searched as candidates one hour and fifteen minutes after the start of program content playback. In such a case, the boundary between two songs is searched using the processing described in FIG.
 上記した実施の形態において、検索された曲について、その開始位置と終了位置とを推定する際、全ての曲について開始位置のみの推定を行ない、その後、全ての曲について終了位置のみの推定を行なうようにすれば、終了位置の推定を行なう際に、開始位置の推定結果が利用できるので処理を容易にすることができる。また、その逆に、全ての曲について終了位置のみの推定を行ない、その後、全ての曲について開始位置のみの推定を行なうことも考えられる。 In the above-described embodiment, when estimating the start position and end position of a searched song, only the start position is estimated for all songs, and then only the end position is estimated for all songs. In this way, when the end position is estimated, the estimation result of the start position can be used, so that the processing can be facilitated. Conversely, it is also conceivable that only the end position is estimated for all songs, and then only the start position is estimated for all songs.
 なお、この発明は上記した実施の形態そのままに限定されるものではなく、実施段階ではその要旨を逸脱しない範囲で構成要素を種々変形して具体化することができる。また、上記した実施の形態に開示されている複数の構成要素を適宜に組み合わせることにより、種々の発明を形成することができる。例えば、実施の形態に示される全構成要素から幾つかの構成要素を削除しても良いものである。さらに、異なる実施の形態に係る構成要素を適宜組み合わせても良いものである。 Note that the present invention is not limited to the above-described embodiment as it is, and can be embodied by variously modifying the constituent elements without departing from the scope of the invention in the implementation stage. Various inventions can be formed by appropriately combining a plurality of constituent elements disclosed in the above-described embodiments. For example, some components may be deleted from all the components shown in the embodiment. Furthermore, constituent elements according to different embodiments may be appropriately combined.

Claims (9)

  1.  解析対象となるコンテンツに対して所定の時間間隔で曲検索を行なう検索手段と、
     前記検索手段により所定の時間間隔で得られる曲検索結果に基づいて、前記コンテンツに含まれる曲の再生状態を解析する解析手段とを具備する情報処理装置。
    Search means for searching for music at predetermined time intervals for the content to be analyzed;
    An information processing apparatus comprising: an analysis unit that analyzes a reproduction state of a song included in the content based on a song search result obtained at a predetermined time interval by the search unit.
  2.  前記検索手段は、前記コンテンツから所定長の音声信号を切り出し、その切り出した音声信号から特徴量を生成し、その特徴量に基づいて曲検索を行なった結果を取得する請求項1記載の情報処理装置。 2. The information processing according to claim 1, wherein the search unit cuts out an audio signal having a predetermined length from the content, generates a feature amount from the cut-out audio signal, and acquires a result of music search based on the feature amount. apparatus.
  3.  前記検索手段は、
     曲検索要求を受けて前記コンテンツから所定長の音声信号を切り出す切り出し手段と、
     前記切り出し手段で切り出された音声信号から特徴量を生成し、その生成した特徴量をネットワーク経由でサーバに送信する生成手段と、
     前記生成手段で生成された特徴量に基づいて前記サーバが曲検索を行なった結果を、前記ネットワーク経由で取得する取得手段とを備える請求項1記載の情報処理装置。
    The search means includes
    Cut-out means for cutting out a predetermined-length audio signal from the content in response to a song search request;
    Generating means for generating a feature quantity from the audio signal cut out by the cut-out means, and transmitting the generated feature quantity to a server via a network;
    The information processing apparatus according to claim 1, further comprising: an acquisition unit configured to acquire a result of the music search performed by the server based on the feature amount generated by the generation unit via the network.
  4.  前記解析手段は、前記検索手段により所定の時間間隔で得られる曲検索結果のうち、時間的に連続する複数の曲検索結果が同じ曲を示している場合、それらの曲検索結果が得られた前記音声信号上の各位置を同じ曲が存在する区間として結合する請求項1記載の情報処理装置。 When the plurality of song search results that are temporally continuous among the song search results obtained at predetermined time intervals by the search means indicate the same song, the analysis means has obtained those song search results. The information processing apparatus according to claim 1, wherein each position on the audio signal is combined as a section where the same music exists.
  5.  前記解析手段は、前記検索手段により所定の時間間隔で得られる曲検索結果のうち、時間的に連続して得られた2つの曲検索結果が異なる曲を示している場合、その2つの曲検索結果を用いて時間的に前の曲検索結果が示す曲の終了位置を推定する請求項1記載の情報処理装置。 When the two song search results obtained continuously in time indicate different songs among the song search results obtained by the search means at predetermined time intervals, the analysis means searches for the two songs. The information processing apparatus according to claim 1, wherein the end position of the music indicated by the previous music search result is estimated using the result.
  6.  前記解析手段は、
     時間的に連続する2つの曲検索結果が得られた各位置の中間位置で曲検索を行ない、その結果として時間的に前の曲検索結果が示す曲が得られた場合、当該中間位置と時間的に後の曲検索結果が得られた位置との中間位置で曲検索を行ない、
     時間的に連続する2つの曲検索結果が得られた各位置の中間位置で曲検索を行ない、その結果として時間的に後の曲検索結果が示す曲が得られた場合、当該中間位置と時間的に前の曲検索結果が得られた位置との中間位置で曲検索を行なう請求項5記載の情報処理装置。
    The analysis means includes
    When a song search is performed at an intermediate position between two positions where two time search results are obtained, and as a result, a song indicated by the previous song search result is obtained, the intermediate position and time The song search is performed at a position intermediate to the position where the later song search result is obtained,
    When a song search is performed at an intermediate position between two positions where two temporal search results are obtained, and as a result, a song indicated by a later song search result is obtained, the intermediate position and time 6. The information processing apparatus according to claim 5, wherein the music search is performed at an intermediate position from a position where the previous music search result is obtained.
  7.  前記解析手段は、前記検索手段により所定の時間間隔で得られる曲検索結果のうち、時間的に連続して得られた2つの曲検索結果が異なる曲を示している場合、その2つの曲検索結果を用いて時間的に後の曲検索結果が示す曲の開始位置を推定する請求項1記載の情報処理装置。 When the two song search results obtained continuously in time indicate different songs among the song search results obtained by the search means at predetermined time intervals, the analysis means searches for the two songs. The information processing apparatus according to claim 1, wherein the start position of a song indicated by a later song search result is estimated using the result.
  8.  前記解析手段は、
     時間的に連続する2つの曲検索結果が得られた各位置の中間位置で曲検索を行ない、その結果として時間的に後の曲検索結果が示す曲が得られた場合、当該中間位置と時間的に前の曲検索結果が得られた位置との中間位置で曲検索を行ない、
     時間的に連続する2つの曲検索結果が得られた各位置の中間位置で曲検索を行ない、その結果として時間的に前の曲検索結果が示す曲が得られた場合、当該中間位置と時間的に後の曲検索結果が得られた位置との中間位置で曲検索を行なう請求項7記載の情報処理装置。
    The analysis means includes
    When a song search is performed at an intermediate position between two positions where two temporal search results are obtained, and as a result, a song indicated by a later song search result is obtained, the intermediate position and time The song search is performed at an intermediate position from the position where the previous song search result was obtained,
    When a song search is performed at an intermediate position between two positions where two time search results are obtained, and as a result, a song indicated by the previous song search result is obtained, the intermediate position and time The information processing apparatus according to claim 7, wherein the music search is performed at an intermediate position from a position where a subsequent music search result is obtained.
  9.  検索手段を用いて、解析対象となるコンテンツに対して所定の時間間隔で曲検索を行ない、
     解析手段を用いて、前記検索手段により所定の時間間隔で得られる曲検索結果に基づいて、前記コンテンツに含まれる曲の再生状態を解析する情報処理方法。
    Using search means, perform a song search at a predetermined time interval for the content to be analyzed,
    An information processing method for analyzing a reproduction state of a song included in the content based on a song search result obtained at a predetermined time interval by the search unit using an analysis unit.
PCT/JP2013/058791 2013-03-26 2013-03-26 Information processing device and information processing method WO2014155526A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/JP2013/058791 WO2014155526A1 (en) 2013-03-26 2013-03-26 Information processing device and information processing method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2013/058791 WO2014155526A1 (en) 2013-03-26 2013-03-26 Information processing device and information processing method

Publications (1)

Publication Number Publication Date
WO2014155526A1 true WO2014155526A1 (en) 2014-10-02

Family

ID=51622612

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2013/058791 WO2014155526A1 (en) 2013-03-26 2013-03-26 Information processing device and information processing method

Country Status (1)

Country Link
WO (1) WO2014155526A1 (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2007171289A (en) * 2005-12-19 2007-07-05 Mitsubishi Electric Corp Music detecting apparatus
JP2007219178A (en) * 2006-02-16 2007-08-30 Sony Corp Musical piece extraction program, musical piece extraction device, and musical piece extraction method

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2007171289A (en) * 2005-12-19 2007-07-05 Mitsubishi Electric Corp Music detecting apparatus
JP2007219178A (en) * 2006-02-16 2007-08-30 Sony Corp Musical piece extraction program, musical piece extraction device, and musical piece extraction method

Similar Documents

Publication Publication Date Title
KR102091075B1 (en) Reception device, reception method, transmission device, transmission method, and program
KR102099357B1 (en) Device, system and method for providing screen shot
US9277267B2 (en) Content output system, information display apparatus, content output apparatus, and content information display method
JP4445555B2 (en) Content editing device
JP2004194294A (en) Receiving apparatus and receiving method
JP5076892B2 (en) Same scene detection device and storage medium storing program
AU2018260403A1 (en) Methods and systems for effective scrub bar navigation
KR20150068828A (en) Image Display Apparatus and Driving Method Thereof, Method for Displaying Image and Computer Readable Recording Medium
KR20150056394A (en) Picture display device and operating method thereof
WO2014155526A1 (en) Information processing device and information processing method
JP2008085934A (en) Remote reproduction system for video and method of resume reproduction
JP4507194B2 (en) Program recording / reproducing apparatus and program recording / reproducing system
JP2006019888A (en) Recorder and recording control method
JP2008147802A (en) Recording and reproducing device, and display control method
JP2009253311A (en) Receiving apparatus and control method thereof
JP2007158441A (en) Program guide generating apparatus and program guide generating method
JP2008182539A (en) Broadcast receiving, recording and reproducing apparatus, and broadcast receiving, recording and reproducing method
JP2009055518A (en) Streaming server, and streaming system
JP2019102909A (en) Electronic apparatus and video reproduction method
JP2006270793A (en) Digest video recording system
KR101270516B1 (en) Method and apparatus for obtaining/providing a related video content section during playing a live audio stream
JP6192547B2 (en) Video recording / reproducing apparatus and video recording / reproducing method
JP2011044834A (en) Content accumulation device, content accumulation method, content accumulation program and content accumulation system
KR100703338B1 (en) Method For Setting Of the Instant Recording Ending-Time in PVR
JP2009212814A (en) Program selection device, program selection program, and program selection method

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13880314

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 13880314

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP