US20080236368A1 - Recording or playback apparatus and musical piece detecting apparatus - Google Patents

Recording or playback apparatus and musical piece detecting apparatus Download PDF

Info

Publication number
US20080236368A1
US20080236368A1 US12/053,647 US5364708A US2008236368A1 US 20080236368 A1 US20080236368 A1 US 20080236368A1 US 5364708 A US5364708 A US 5364708A US 2008236368 A1 US2008236368 A1 US 2008236368A1
Authority
US
United States
Prior art keywords
cut point
audio signal
characteristic amount
music
judging unit
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US12/053,647
Other versions
US7745714B2 (en
Inventor
Satoru Matsumoto
Yuji Yamamoto
Tatsuo Koga
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sanyo Electric Co Ltd
Original Assignee
Sanyo Electric Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sanyo Electric Co Ltd filed Critical Sanyo Electric Co Ltd
Assigned to SANYO ELECTRIC CO., LTD. reassignment SANYO ELECTRIC CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KOGA, TATSUO, MATSUMOTO, SATORU, YAMAMOTO, YUJI
Publication of US20080236368A1 publication Critical patent/US20080236368A1/en
Application granted granted Critical
Publication of US7745714B2 publication Critical patent/US7745714B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0008Associated control or indicating means
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/046Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for differentiation between music and non-music signals, based on the identification of musical parameters, e.g. based on tempo detection
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/066Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for pitch analysis as part of wider processing for musical purposes, e.g. transcription, musical performance evaluation; Pitch recognition, e.g. in polyphonic sounds; Estimation or use of missing fundamental
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/011Files or data streams containing coded musical information, e.g. for transmission
    • G10H2240/046File format, i.e. specific or non-standard musical file format used in or adapted for electrophonic musical instruments, e.g. in wavetables
    • G10H2240/061MP3, i.e. MPEG-1 or MPEG-2 Audio Layer III, lossy audio compression

Definitions

  • the present invention relates to an apparatus which detects music (musical piece) sections from an audio including speech sections and music sections in a mixed manner.
  • an aired audio often includes sections carrying speeches of an announcer and music sections in a mixed manner.
  • the listener has to manually start recording the musical piece at a timing when the musical piece begins, and to manually stop recording the musical piece at a timing when the musical piece ends.
  • These manual operations are troublesome for the listener.
  • a listener suddenly decides to record a favorite musical piece which is aired, it is usually impossible to thoroughly record the musical piece from its beginning without missing any part. In such case, it is effective to record an entire aired program first, and then extract the favorite musical piece from the recorded program by editing. This editing becomes easier by separating music sections from the aired program beforehand and by playing back only the separated music sections.
  • a technology for automatically separating music sections and speech sections from each other by analyzing characteristics of each of the sections is for separating a musical piece and a speech from each other by using characteristic amounts in terms of frequencies such as mel-frequency cepstral coefficients (MFCCs).
  • MFCCs mel-frequency cepstral coefficients
  • the technology disclosed by the Publication No. 2004-258659 has a problem that a process for calculating the characteristic amount in a frequency area of an audio signal becomes vast because the process is so complicated that the workload for the process becomes large.
  • An aspect of the invention provides an apparatus implementing at least recording or playback that detects a music section from an audio signal.
  • the apparatus comprises: a cut point detector configured to detect a time point as a cut point where a level of an audio signal or an amount of change in the audio signal level is equal to or more than a predetermined value; a frequency characteristic amount calculator configured to calculate a characteristic amount in a frequency area of the audio signal; a cut point judging unit configured to judge an attribute of the cut point on a basis of the calculated characteristic amount in a frequency; and a music section detector configured to detect a start point and an end point of a music section on a basis of the attribute and an interval between sampling points.
  • the apparatus comprises: a cut point detector configured to detect a time point as a cut point where a level of an audio signal level or an amount of change in the audio signal level is equal to or more than a predetermined value; a frequency characteristic amount calculator configured to calculate a characteristic amount in a frequency area of the audio signal; and a music section detector configured to detect a start point and an end point of each music section on a basis of the calculated characteristic amount of the frequency and information on the detected cut point.
  • Still another aspect of the invention provides a musical piece detecting apparatus that detects a musical piece from an inputted audio.
  • the apparatus comprises: an audio power calculator configured to calculate an audio power from an inputted audio signal; a cut point detector configured to detect a time point as a cut point where a level of an audio signal level or an amount of change in the audio signal level is equal to or more than a predetermined value on a basis of the audio power, the cut point detector configured to output time information on the cut point; a frequency characteristic amount calculator configured to calculate a characteristic amount in a frequency area at the detected cut point of the inputted audio signal; a likelihood calculator configured to calculate a likelihood between the characteristic amount and reference data on the musical piece; a cut point judging unit configured to judge, on a basis of the likelihood, whether or not the audio signal at the cut point is the musical piece; a time length judging unit configured to judge, on a basis of the time information on the cut point, a result of the judgment made by the cut point judging unit, the time length judging unit judging,
  • the recording or playback apparatus is capable of separating the musical piece from the audio consisting of the musical piece and the speech though a simple arithmetic process.
  • FIG. 1 is a configuration diagram illustrating a musical piece detecting function in a recording or playback apparatus according to an embodiment of the present invention.
  • FIG. 2 is a functional block diagram illustrating a part of the recording or playback apparatus according to the embodiment.
  • FIGS. 3A and 3B are waveform diagrams each illustrating how a cut point detector operates.
  • FIG. 4 shows a table stored in a temporary storage memory.
  • FIG. 5 shows a final table rewritten in the temporary storage memory.
  • MPEG audio layer-3 (MP3) codec 3 includes an encoder function and a decoder function.
  • the encoder function encodes the digital audio data, and thus generates compressed coded data, as well as subsequently outputs the compressed coded data along with time information.
  • the decoder function decodes the coded data.
  • D/A (digital-to-analog) converter 4 converts the digital audio data, which is decoded by MP3 codec 3 , to analog signal data. Subsequently, this analog signal data is inputted into speaker 5 via an amplifier, whose illustration is omitted from FIG. 1 .
  • CPU 8 controls the overall operation of the recording or playback apparatus according to the present embodiment.
  • CPU 8 performs things such as a process for assuming whether the cut point belongs to the start point or the end point of the musical piece.
  • HDD (hard disc drive) 10 is a large-capacity storage in which the coded data and the time information is stored via HDD interface 9 of an ATA (advanced technology attachment) interface.
  • Memory 11 has a function of storing the execution program, and of having data generated through the arithmetic process stored temporarily, as well as of delaying the audio data for a predetermined time length right after the audio data is converted from analog to digital. It should be noted that various pieces of data are transmitted to, and received from, MP3 codec 3 , DSP 7 , CPU 8 , HDD interface 9 and memory 11 via bus 6 .
  • FIG. 2 is a functional block diagram showing a part of the recording or playback apparatus according to the present embodiment.
  • the recording or playback apparatus inputs the audio signal tuned in to by tuner 1 to A/D converter 2 , and thus converts the audio signal from analog to digital.
  • the recording or playback apparatus inputs the digital-converted audio signal along with the time information to the MP3 codec 3 , and thus compresses and encodes the digital-converted audio signal into MP3 data, as well as continuously records the MP3 data along with the time information in HDD 10 via HDD interface 9 while the musical piece is being recorded.
  • the digital audio data from A/D converter 2 is stored in delay memory 11 a for delaying the digital audio data by a time length equivalent to a time needed for DSP 7 to perform its process.
  • audio power calculator 71 in DSP 7 calculates the audio power equivalent to the audio signal level, or a value by raising the value representing the amplitude of the audio signal to the second power.
  • Cut point detector 72 in DSP 7 detects, as a cut point, a timing at which the amount of change in the audio signal level is large, or a timing at which the amount of change in the audio signal level is not smaller than the predetermined value. Thus, an output from the detection is outputted. Concurrently, the time information and the amount of change at the cut point are stored in temporary storage memory 11 c.
  • FIGS. 3A and 3B are waveform diagrams each illustrating how cut point detector 72 operates.
  • FIG. 3A shows how the audio power changes
  • FIG. 3B shows how the amount of change (differential value) changes.
  • cut point detector 72 detects, as cut points, times Tm and Tm+1 at which the differential value becomes a local maximum point exceeding a predetermined threshold value. Thereafter, a result of the detection is inputted to frequency characteristic amount calculator 73 .
  • Frequency characteristic amount calculator 73 synchronizes the audio data, which is outputted from delay memory 11 a with delay by the predetermined time, with the output from cut point detector 72 . Then, in a very short period of time between a timing slightly preceding a cut point and a timing slightly delayed from the cut point, the calculator 73 temporarily calculates the characteristic amount of the frequency, such as the MFCC. Then, the result is inputted to likelihood calculator 74 .
  • likelihood detector 74 in the DSP calculates the likelihood between the reference data and the output representing the result of the calculation of the characteristic amount at each cut point and in its proximity, which output is received from frequency characteristic amount calculator 73 . Thereafter, likelihood detector 74 inputs an output representing the calculated likelihood to cut point judging unit 81 in CPU 8 .
  • the calculated characteristic amount of the frequency does not have to be compared with the reference data.
  • another applicable method calculates the likelihood of the musical piece through assigning the characteristic amount of the frequency to an evaluation function set up beforehand.
  • cut point judging unit 81 judges whether the audio signal at the cut point belongs to the music or the speech on the basis of the output of the calculated likelihood.
  • a result of the judgment is additionally stored in temporary storage memory 11 c, in which the time information and the amount of change at the cut point which are received from the cut point detector 72 are already stored, with the result of the judgment associated with the time information and the amount of change at the cut point.
  • FIG. 4 shows a table of temporary storage memory 11 c which stores the result of the judgment in association with the time information and the amount of change at the cut point.
  • Time length judging unit 83 judges whether the audio judged, by cut point judging unit 81 , as belonging to the music section lasts for a predetermined time length or longer. Time length judging unit 83 judges that the section is not a musical piece when the music section lasts shorter than the predetermined time length. In the case shown in FIG. 4 , for instance, sections judged as the musical pieces by cut point judging unit 81 are those corresponding to times T 2 , T 3 , T 4 , T 6 , T 8 and T 9 .
  • time length judging unit 83 judges whether each of these three sections lasts for the predetermined length time or longer. In this example, if the time T 6 is shorter than the predetermined time length, time length judging unit 83 judges that the section corresponding to time T 6 is not a musical piece.
  • time length judging unit 83 judges whether or not the total time length of the one or more sections interposed in between is not shorter than the predetermined time length. If the total time length is shorter than the predetermined time length, time length judging unit 83 judges that the one or more sections interposed in between are not musical pieces.
  • the predetermined time length may be set at 100 seconds in order for time length judging unit 83 to make the judgment on the music section. However, the predetermined time length is not necessarily limited to 100 seconds.
  • time length judging unit 83 is designed not to judge the section between the two neighboring sampling points as a musical piece. The time interval between two neighboring sampling points judged as a speech or anything but a musical piece is measured, and a corresponding section which is not shorter than 100 seconds is judged as a musical piece.
  • time length judging unit 83 is designed to judge the corresponding section as no musical piece.
  • Time length judging unit 83 is designed to measure the time interval between two neighboring sampling points judged as a speech or anything but a musical piece, and to judge a corresponding section which is more than 100 seconds as a musical piece.
  • Music section detector 82 receives an output of the judgment which is obtained from time length judging unit 83 , and thus rewrites the table in temporary storage memory 11 c, accordingly changing an existing table to a table (final table) for each musical piece.
  • FIG. 5 is a diagram showing a final table obtained by rewriting an existing table in temporary storage memory 11 c.
  • the final table shows that time T 6 is removed from the table, even though time T 6 is once judged as a musical piece. This is because time T 6 is regarded as no musical piece on the basis that the time length between its preceding time T 5 and its subsequent time T 7 both judged as a speech is shorter than the predetermined time length.
  • this final table is supplied to HDD interface unit 9 via music section detector 82 , and is subsequently stored in HDD 10 .
  • each final table is stored in HDD 10 with a start point, an end point, cut points, and amounts of change left for a corresponding musical piece. These are all used to play back the chorus of the musical piece when the musical piece is going to be played back.
  • MP3 codec 3 decodes the corresponding parts in the encoded data. Subsequently, the decoded parts are converted to the audio signal by D/A converter 4 , and are thus outputted from speaker 5 . This makes it possible to detect only the musical piece from the audio signal including speech sections and the like, as well as accordingly to extract and play back the musical piece.
  • the present embodiment makes it possible to precisely detect the musical piece, because the music sections are detected by use of both information on the cut points and information on the amounts of characteristic of the respective frequencies.
  • the present embodiment also makes it possible to detect the music sections though the arithmetic process entailing only a light workload, because the music sections are detected by calculating the characteristic amount in the frequency area of the audio signal only at each cut point and in its proximity.
  • DSP 7 is designed to implement its own function whereas CPU 8 is designed to implement its own function.
  • the present embodiment is not necessarily limited to the function division therebetween. The two functions may be implemented by CPU 8 only. Otherwise, the present embodiment may have a configuration in which, through software process, CPU 8 implements the functions respectively of A/D converter 2 , MP3 codec 3 and D/A converter 4 in addition to the function of DSP 7 .
  • delay memory 11 a, external memory 11 b and temporary storage memory 11 c have been discretely shown in the foregoing example, the memories are formed in memory 11 shown in FIG. 1 .
  • the apparatus detects the music sections while recording the musical piece, so that the apparatus creates and records the final table.
  • a configuration may be adopted, which causes the apparatus to detect the music sections while sequentially playing back the recorded digital audio data from HDD 10 during an idle time after the apparatus completes recording the musical piece, so that the apparatus creates the final table.
  • a circuit configuration may be adopted, which causes the apparatus to carry out all of the operations according to the foregoing example in linkage with the playback operation. It goes without saying that these configurations are included in the present invention.
  • the audio signal level is detected as the value obtained by raising a value representing the amplitude of the audio signal to the second power.
  • the audio signal level can be similarly detected as the absolute value of the amplitude, instead.
  • the cut point is defined as a timing at which the audio signal level changes to the large extent.
  • the cut point corresponds to neither the start point nor the end point of the musical piece precisely.
  • the cut point can be sufficiently used as the playback start point or the playback end point of the musical piece.
  • the foregoing example has a configuration effective for a method with which, while editing after recording musical pieces, the operator determines whether or not each of the recorded musical pieces is what the operator wished to have by playing back a part of every recorded musical piece, and leaves only musical pieces which the operator wishes to have as a library afterward.
  • the foregoing example aims at being used regardless of whether or not the editing is carried out precisely.
  • the music sections may be detected in accordance with the following procedure.
  • the detection according to the modification makes it possible to increase the precision with which the music section is detected in comparison with the technology, disclosed in Japanese Patent Application Laid-Open Publication No. 2004-258659, for detecting a music section by use of a characteristic amount of the frequency only.

Abstract

Provided is a recording or playback apparatus capable of separating a musical piece from an audio including the musical piece and a speech through a simple arithmetic process. A cut point detector detects, as a cut point, a time point at which an audio signal level or an amount of change in the audio signal level is not lower than a predetermined value. A frequency characteristic amount calculator calculates a characteristic amount in a frequency area of the audio signal only at each cut point and in its proximity. A cut point judging unit judges an attribute of the cut point on a basis of the calculated characteristic amount of the frequency. A music section detector detects a start and end points of each music section on a basis of the attribute and an interval between sampling points.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application claims priority based on 35 USC 119 from prior Japanese Patent Application No. P2007-078956 filed on Mar. 26, 2007, the entire contents of which are incorporated herein by reference.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to an apparatus which detects music (musical piece) sections from an audio including speech sections and music sections in a mixed manner.
  • 2. Description of Related Art
  • In general, an aired audio often includes sections carrying speeches of an announcer and music sections in a mixed manner. When a listener wishes to record his/her favorite musical piece while listening to the audio, the listener has to manually start recording the musical piece at a timing when the musical piece begins, and to manually stop recording the musical piece at a timing when the musical piece ends. These manual operations are troublesome for the listener. Moreover, if a listener suddenly decides to record a favorite musical piece which is aired, it is usually impossible to thoroughly record the musical piece from its beginning without missing any part. In such case, it is effective to record an entire aired program first, and then extract the favorite musical piece from the recorded program by editing. This editing becomes easier by separating music sections from the aired program beforehand and by playing back only the separated music sections.
  • To this end, a technology for automatically separating music sections and speech sections from each other by analyzing characteristics of each of the sections. A technology disclosed by Japanese Patent Application Laid-Open Publication No. 2004-258659 is for separating a musical piece and a speech from each other by using characteristic amounts in terms of frequencies such as mel-frequency cepstral coefficients (MFCCs). However, the technology disclosed by the Publication No. 2004-258659 has a problem that a process for calculating the characteristic amount in a frequency area of an audio signal becomes vast because the process is so complicated that the workload for the process becomes large.
  • SUMMARY OF THE INVENTION
  • An aspect of the invention provides an apparatus implementing at least recording or playback that detects a music section from an audio signal. The apparatus comprises: a cut point detector configured to detect a time point as a cut point where a level of an audio signal or an amount of change in the audio signal level is equal to or more than a predetermined value; a frequency characteristic amount calculator configured to calculate a characteristic amount in a frequency area of the audio signal; a cut point judging unit configured to judge an attribute of the cut point on a basis of the calculated characteristic amount in a frequency; and a music section detector configured to detect a start point and an end point of a music section on a basis of the attribute and an interval between sampling points.
  • Another aspect of the invention provides an apparatus implementing at least recording or playback that detects a music section from an audio signal. The apparatus comprises: a cut point detector configured to detect a time point as a cut point where a level of an audio signal level or an amount of change in the audio signal level is equal to or more than a predetermined value; a frequency characteristic amount calculator configured to calculate a characteristic amount in a frequency area of the audio signal; and a music section detector configured to detect a start point and an end point of each music section on a basis of the calculated characteristic amount of the frequency and information on the detected cut point.
  • Still another aspect of the invention provides a musical piece detecting apparatus that detects a musical piece from an inputted audio. The apparatus comprises: an audio power calculator configured to calculate an audio power from an inputted audio signal; a cut point detector configured to detect a time point as a cut point where a level of an audio signal level or an amount of change in the audio signal level is equal to or more than a predetermined value on a basis of the audio power, the cut point detector configured to output time information on the cut point; a frequency characteristic amount calculator configured to calculate a characteristic amount in a frequency area at the detected cut point of the inputted audio signal; a likelihood calculator configured to calculate a likelihood between the characteristic amount and reference data on the musical piece; a cut point judging unit configured to judge, on a basis of the likelihood, whether or not the audio signal at the cut point is the musical piece; a time length judging unit configured to judge, on a basis of the time information on the cut point, a result of the judgment made by the cut point judging unit, the time length judging unit judging, on the basis of the time information on the cut point, whether or not a section between sections not judged as musical pieces lasts for a predetermined time length or longer; and a music section detector configured to detect a music section on a basis of a result of the judgment made by the time length judging unit.
  • The recording or playback apparatus is capable of separating the musical piece from the audio consisting of the musical piece and the speech though a simple arithmetic process.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a configuration diagram illustrating a musical piece detecting function in a recording or playback apparatus according to an embodiment of the present invention.
  • FIG. 2 is a functional block diagram illustrating a part of the recording or playback apparatus according to the embodiment.
  • FIGS. 3A and 3B are waveform diagrams each illustrating how a cut point detector operates.
  • FIG. 4 shows a table stored in a temporary storage memory.
  • FIG. 5 shows a final table rewritten in the temporary storage memory.
  • DETAILED DESCRIPTION OF EMBODIMENT
  • Descriptions will be provided hereinbelow for an embodiment with reference to the drawings. FIG. 1 is a configuration diagram illustrating a musical piece detecting function in a recording or playback apparatus according to the embodiment. As shown in FIG. 1, the recording or playback apparatus according to the present embodiment selects, and receives, a broadcast signal for a television program, a radio program or the like, as well as thus demodulates the broadcast signal to an audio signal. A/D (analog-to-digital) converter 2 converts an analog audio signal selected by tuner 1 to a digital signal.
  • MPEG audio layer-3 (MP3) codec 3 includes an encoder function and a decoder function. The encoder function encodes the digital audio data, and thus generates compressed coded data, as well as subsequently outputs the compressed coded data along with time information. The decoder function decodes the coded data. D/A (digital-to-analog) converter 4 converts the digital audio data, which is decoded by MP3 codec 3, to analog signal data. Subsequently, this analog signal data is inputted into speaker 5 via an amplifier, whose illustration is omitted from FIG. 1.
  • On a basis of the audio signal, DSP (digital signal processor) 7 calculates an audio power obtained by raising a value representing the amplitude of the audio signal to the second power for the purpose of detecting an audio signal level. In addition, DSP 7 calculates an amount of change in the audio power in order to detect an amount of change in the audio signal level. Furthermore, DSP 7 defines, as a cut point, a timing at which the amount of change in the audio power is not smaller than a predetermined value, and thus detects the cut point. Moreover, DSP 7 calculates a characteristic amount in a frequency area, an MFCC, for example, only at each cut point and in its proximity. Then, DSP 7 calculates a likelihood between the characteristic amount and an MFCC calculated on a basis of a sample audio signal.
  • Through bus 6, CPU (central processing unit) 8 controls the overall operation of the recording or playback apparatus according to the present embodiment. In addition, CPU 8 performs things such as a process for assuming whether the cut point belongs to the start point or the end point of the musical piece. HDD (hard disc drive) 10 is a large-capacity storage in which the coded data and the time information is stored via HDD interface 9 of an ATA (advanced technology attachment) interface. Memory 11 has a function of storing the execution program, and of having data generated through the arithmetic process stored temporarily, as well as of delaying the audio data for a predetermined time length right after the audio data is converted from analog to digital. It should be noted that various pieces of data are transmitted to, and received from, MP3 codec 3, DSP 7, CPU 8, HDD interface 9 and memory 11 via bus 6.
  • FIG. 2 is a functional block diagram showing a part of the recording or playback apparatus according to the present embodiment. As shown in FIG. 1, the recording or playback apparatus according to the present embodiment inputs the audio signal tuned in to by tuner 1 to A/D converter 2, and thus converts the audio signal from analog to digital. Subsequently, the recording or playback apparatus inputs the digital-converted audio signal along with the time information to the MP3 codec 3, and thus compresses and encodes the digital-converted audio signal into MP3 data, as well as continuously records the MP3 data along with the time information in HDD 10 via HDD interface 9 while the musical piece is being recorded.
  • The digital audio data from A/D converter 2 is stored in delay memory 11 a for delaying the digital audio data by a time length equivalent to a time needed for DSP 7 to perform its process. Concurrently, audio power calculator 71 in DSP 7 calculates the audio power equivalent to the audio signal level, or a value by raising the value representing the amplitude of the audio signal to the second power.
  • Cut point detector 72 in DSP 7 detects, as a cut point, a timing at which the amount of change in the audio signal level is large, or a timing at which the amount of change in the audio signal level is not smaller than the predetermined value. Thus, an output from the detection is outputted. Concurrently, the time information and the amount of change at the cut point are stored in temporary storage memory 11 c.
  • FIGS. 3A and 3B are waveform diagrams each illustrating how cut point detector 72 operates. FIG. 3A shows how the audio power changes, and FIG. 3B shows how the amount of change (differential value) changes. As shown in FIGS. 3A and 3B, on the basis of the value representing the audio power calculated by audio power calculator 71, cut point detector 72 detects, as cut points, times Tm and Tm+1 at which the differential value becomes a local maximum point exceeding a predetermined threshold value. Thereafter, a result of the detection is inputted to frequency characteristic amount calculator 73.
  • Frequency characteristic amount calculator 73 synchronizes the audio data, which is outputted from delay memory 11 a with delay by the predetermined time, with the output from cut point detector 72. Then, in a very short period of time between a timing slightly preceding a cut point and a timing slightly delayed from the cut point, the calculator 73 temporarily calculates the characteristic amount of the frequency, such as the MFCC. Then, the result is inputted to likelihood calculator 74.
  • In the present embodiment, it is taken into consideration that the characteristic amount of the frequency of the musical piece is different from that of the speech. For this reason, a characteristic amount of the frequency typical of the musical piece and that of the speech are both stored in external memory 11 b beforehand as reference data used for comparison between the characteristic amounts of the frequencies. As a result, likelihood detector 74 in the DSP calculates the likelihood between the reference data and the output representing the result of the calculation of the characteristic amount at each cut point and in its proximity, which output is received from frequency characteristic amount calculator 73. Thereafter, likelihood detector 74 inputs an output representing the calculated likelihood to cut point judging unit 81 in CPU 8.
  • It should be noted that the calculated characteristic amount of the frequency does not have to be compared with the reference data. Specifically, in addition to the foregoing method of calculating the likelihood of the musical piece through comparing the calculated characteristic amount of the frequency with the reference data, another applicable method calculates the likelihood of the musical piece through assigning the characteristic amount of the frequency to an evaluation function set up beforehand.
  • Subsequently, cut point judging unit 81 judges whether the audio signal at the cut point belongs to the music or the speech on the basis of the output of the calculated likelihood. A result of the judgment is additionally stored in temporary storage memory 11 c, in which the time information and the amount of change at the cut point which are received from the cut point detector 72 are already stored, with the result of the judgment associated with the time information and the amount of change at the cut point.
  • FIG. 4 shows a table of temporary storage memory 11 c which stores the result of the judgment in association with the time information and the amount of change at the cut point.
  • Time length judging unit 83 judges whether the audio judged, by cut point judging unit 81, as belonging to the music section lasts for a predetermined time length or longer. Time length judging unit 83 judges that the section is not a musical piece when the music section lasts shorter than the predetermined time length. In the case shown in FIG. 4, for instance, sections judged as the musical pieces by cut point judging unit 81 are those corresponding to times T2, T3, T4, T6, T8 and T9. In this respect, consecutive sections corresponding to times T2, T3, T4 which are judged as the musical pieces are regarded as a single musical piece; an isolated section corresponding to time T6 is regarded as another musical piece; and consecutive sections corresponding to times T8 and T9 which are judged as the musical pieces are regarded as yet another musical piece. Then, time length judging unit 83 judges whether each of these three sections lasts for the predetermined length time or longer. In this example, if the time T6 is shorter than the predetermined time length, time length judging unit 83 judges that the section corresponding to time T6 is not a musical piece. In other words, when one or more sections are judged as musical pieces with the sections between sections judged as no musical pieces, time length judging unit 83 judges whether or not the total time length of the one or more sections interposed in between is not shorter than the predetermined time length. If the total time length is shorter than the predetermined time length, time length judging unit 83 judges that the one or more sections interposed in between are not musical pieces. In this respect, the predetermined time length may be set at 100 seconds in order for time length judging unit 83 to make the judgment on the music section. However, the predetermined time length is not necessarily limited to 100 seconds.
  • As a result, in the case where the time interval between two neighboring sampling points in the speech is shorter than 100 seconds, even if a sampling point between the two sampling point is judged as a musical piece, time length judging unit 83 is designed not to judge the section between the two neighboring sampling points as a musical piece. The time interval between two neighboring sampling points judged as a speech or anything but a musical piece is measured, and a corresponding section which is not shorter than 100 seconds is judged as a musical piece.
  • It is empirically learned that a musical piece lasts more than 100 seconds. Accordingly, in the case where the time interval between two neighboring sampling points in a speech is shorter than 100 seconds, even if a sampling point between the two neighboring points may be judged as a musical piece, time length judging unit 83 is designed to judge the corresponding section as no musical piece. Time length judging unit 83 is designed to measure the time interval between two neighboring sampling points judged as a speech or anything but a musical piece, and to judge a corresponding section which is more than 100 seconds as a musical piece.
  • Music section detector 82 receives an output of the judgment which is obtained from time length judging unit 83, and thus rewrites the table in temporary storage memory 11 c, accordingly changing an existing table to a table (final table) for each musical piece.
  • FIG. 5 is a diagram showing a final table obtained by rewriting an existing table in temporary storage memory 11 c. The final table shows that time T6 is removed from the table, even though time T6 is once judged as a musical piece. This is because time T6 is regarded as no musical piece on the basis that the time length between its preceding time T5 and its subsequent time T7 both judged as a speech is shorter than the predetermined time length.
  • When the recording operation is completed, this final table is supplied to HDD interface unit 9 via music section detector 82, and is subsequently stored in HDD 10.
  • It should be noted that each final table is stored in HDD 10 with a start point, an end point, cut points, and amounts of change left for a corresponding musical piece. These are all used to play back the chorus of the musical piece when the musical piece is going to be played back.
  • Out of encoded data stored in HDD 10, only parts corresponding to music sections specified in the final table are sequentially read out in accordance with editing and playback operations, and are thus inputted into MP3 codec 3. MP3 codec 3 decodes the corresponding parts in the encoded data. Subsequently, the decoded parts are converted to the audio signal by D/A converter 4, and are thus outputted from speaker 5. This makes it possible to detect only the musical piece from the audio signal including speech sections and the like, as well as accordingly to extract and play back the musical piece.
  • The present embodiment makes it possible to precisely detect the musical piece, because the music sections are detected by use of both information on the cut points and information on the amounts of characteristic of the respective frequencies.
  • Furthermore, the present embodiment also makes it possible to detect the music sections though the arithmetic process entailing only a light workload, because the music sections are detected by calculating the characteristic amount in the frequency area of the audio signal only at each cut point and in its proximity.
  • In the present embodiment, DSP 7 is designed to implement its own function whereas CPU 8 is designed to implement its own function. However, the present embodiment is not necessarily limited to the function division therebetween. The two functions may be implemented by CPU 8 only. Otherwise, the present embodiment may have a configuration in which, through software process, CPU 8 implements the functions respectively of A/D converter 2, MP3 codec 3 and D/A converter 4 in addition to the function of DSP 7. Although delay memory 11 a, external memory 11 b and temporary storage memory 11 c have been discretely shown in the foregoing example, the memories are formed in memory 11 shown in FIG. 1.
  • In the case of the foregoing example, the apparatus detects the music sections while recording the musical piece, so that the apparatus creates and records the final table. Instead, a configuration may be adopted, which causes the apparatus to detect the music sections while sequentially playing back the recorded digital audio data from HDD 10 during an idle time after the apparatus completes recording the musical piece, so that the apparatus creates the final table. Otherwise, a circuit configuration may be adopted, which causes the apparatus to carry out all of the operations according to the foregoing example in linkage with the playback operation. It goes without saying that these configurations are included in the present invention.
  • In addition, in the foregoing example, the audio signal level is detected as the value obtained by raising a value representing the amplitude of the audio signal to the second power. The audio signal level can be similarly detected as the absolute value of the amplitude, instead.
  • Moreover, in the foregoing example, the cut point is defined as a timing at which the audio signal level changes to the large extent. As a result, the cut point corresponds to neither the start point nor the end point of the musical piece precisely. However, the cut point can be sufficiently used as the playback start point or the playback end point of the musical piece.
  • The foregoing example has a configuration effective for a method with which, while editing after recording musical pieces, the operator determines whether or not each of the recorded musical pieces is what the operator wished to have by playing back a part of every recorded musical piece, and leaves only musical pieces which the operator wishes to have as a library afterward. The foregoing example aims at being used regardless of whether or not the editing is carried out precisely.
  • (Modification)
  • The music sections may be detected in accordance with the following procedure.
    • (1) First of all, a characteristic amount of the frequency of an audio signal is calculated. Then, the likelihood between a musical piece and the calculated characteristic amount of the frequency is calculated.
    • (2) Subsequently, a time point at which a value representing the likelihood exceeds a predetermined value is judged as being a provisional start point of a music section, whereas a time point at which the value representing the likelihood is lower than the predetermined value is judged as being a provisional end point.
    • (3) Thereafter, a cut point is judged as being a true start point of the music section in a case where the cut point is equal to or close to the provisional start point, whereas a cut point is judged as being a true end point of the music section in a case where the cut point is equal to or close to the provisional end point.
    • (4) After that, it is assumed that the section from the true start point through the true end point is the music section.
  • The detection according to the modification makes it possible to increase the precision with which the music section is detected in comparison with the technology, disclosed in Japanese Patent Application Laid-Open Publication No. 2004-258659, for detecting a music section by use of a characteristic amount of the frequency only.
  • The invention includes other embodiments in addition to the above-described embodiments without departing from the spirit of the invention. The embodiments are to be considered in all respects as illustrative, and not restrictive. The scope of the invention is indicated by the appended claims rather than by the foregoing description. Hence, all configurations including the meaning and range within equivalent arrangements of the claims are intended to be embraced in the invention.

Claims (9)

1. An apparatus implementing at least recording or playback that detects a music section from an audio signal, comprising:
a cut point detector configured to detect a time point as a cut point where a level of an audio signal or an amount of change in the audio signal level is equal to or more than a predetermined value;
a frequency characteristic amount calculator configured to calculate a characteristic amount in a frequency area of the audio signal;
a cut point judging unit configured to judge an attribute of the cut point on a basis of the calculated characteristic amount in a frequency; and
a music section detector configured to detect a start point and an end point of a music section on a basis of the attribute and an interval between sampling points.
2. The apparatus of claim 1, wherein
the frequency characteristic amount calculator calculates a characteristic amount in the frequency area of the audio signal only at each cut point and in its proximity.
3. The apparatus of claim 1, wherein
on a basis of the calculated characteristic amount of the frequency, the cut point judging unit judges whether the audio signal at each cut point and in its proximity belongs to a music section or to a non-music section, and
when a time interval between two neighboring non-music sections is not shorter than a predetermined length of time, the cut point judging unit presumes that the audio signal between these non-music sections is a music section.
4. The apparatus of claim 1, wherein
on a basis of the calculated characteristic amount of the frequency, the cut point judging unit judges whether the audio signal at each cut point and in its proximity belongs to a music section or to a non-music section, and
when a time interval between two cut points respectively belonging to two neighboring non-music sections is not shorter than a predetermined length of time, the cut point judging unit assumes that the audio signal between the cut points respectively belonging to these non-music sections is a music section.
5. An apparatus implementing at least recording or playback that detects a music section from an audio signal, comprising:
a cut point detector configured to detect a time point as a cut point where a level of an audio signal level or an amount of change in the audio signal level is equal to or more than a predetermined value;
a frequency characteristic amount calculator configured to calculate a characteristic amount in a frequency area of the audio signal; and
a music section detector configured to detect a start point and an end point of each music section on a basis of the calculated characteristic amount of the frequency and information on the detected cut point.
6. A musical piece detecting apparatus that detects a musical piece from an inputted audio, comprising:
an audio power calculator configured to calculate an audio power from an inputted audio signal;
a cut point detector configured to detect a time point as a cut point where a level of an audio signal level or an amount of change in the audio signal level is equal to or more than a predetermined value on a basis of the audio power, the cut point detector configured to output time information on the cut point;
a frequency characteristic amount calculator configured to calculate a characteristic amount in a frequency area at the detected cut point of the inputted audio signal;
a likelihood calculator configured to calculate a likelihood between the characteristic amount and reference data on the musical piece;
a cut point judging unit configured to judge, on a basis of the likelihood, whether or not the audio signal at the cut point is the musical piece;
a time length judging unit configured to judge, on a basis of the time information on the cut point, a result of the judgment made by the cut point judging unit, the time length judging unit judging, on the basis of the time information on the cut point, whether or not a section between sections not judged as musical pieces lasts for a predetermined time length or longer; and
a music section detector configured to detect a music section on a basis of a result of the judgment made by the time length judging unit.
7. The apparatus of claim 6, wherein
the frequency characteristic amount calculator calculates a characteristic amount in the frequency area of the audio signal only at each cut point and in its proximity.
8. The apparatus of claim 6, wherein
on a basis of the calculated characteristic amount of the frequency, the cut point judging unit judges whether the audio signal at each cut point and in its proximity belongs to a music section or to a non-music section, and
when a time interval between two neighboring non-music sections is not shorter than a predetermined length of time, the cut point judging unit presumes that the audio signal between these non-music sections is a music section.
9. The apparatus of claim 6, wherein
on a basis of the calculated characteristic amount of the frequency, the cut point judging unit judges whether the audio signal at each cut point and in its proximity belongs to a music section or to a non-music section, and
when a time interval between two cut points respectively belonging to two neighboring non-music sections is not shorter than a predetermined length of time, the cut point judging unit assumes that the audio signal between the cut points respectively belonging to these non-music sections is a music section.
US12/053,647 2007-03-26 2008-03-24 Recording or playback apparatus and musical piece detecting apparatus Expired - Fee Related US7745714B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
JPJP2007-078956 2007-03-26
JP2007078956A JP2008241850A (en) 2007-03-26 2007-03-26 Recording or reproducing device
JP2007-078956 2007-03-26

Publications (2)

Publication Number Publication Date
US20080236368A1 true US20080236368A1 (en) 2008-10-02
US7745714B2 US7745714B2 (en) 2010-06-29

Family

ID=39792055

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/053,647 Expired - Fee Related US7745714B2 (en) 2007-03-26 2008-03-24 Recording or playback apparatus and musical piece detecting apparatus

Country Status (2)

Country Link
US (1) US7745714B2 (en)
JP (1) JP2008241850A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7745714B2 (en) * 2007-03-26 2010-06-29 Sanyo Electric Co., Ltd. Recording or playback apparatus and musical piece detecting apparatus
CN107481739A (en) * 2017-08-16 2017-12-15 成都品果科技有限公司 Audio cutting method and device
US20200160821A1 (en) * 2017-07-25 2020-05-21 Yamaha Corporation Information processing method

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8712771B2 (en) * 2009-07-02 2014-04-29 Alon Konchitsky Automated difference recognition between speaking sounds and music
JP2011090290A (en) * 2009-09-28 2011-05-06 Sanyo Electric Co Ltd Music extraction device and music recording apparatus

Citations (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5233484A (en) * 1989-08-04 1993-08-03 Canon Kabushiki Kaisha Audio signal reproducing apparatus
US5712953A (en) * 1995-06-28 1998-01-27 Electronic Data Systems Corporation System and method for classification of audio or audio/video signals based on musical content
US6169241B1 (en) * 1997-03-03 2001-01-02 Yamaha Corporation Sound source with free compression and expansion of voice independently of pitch
US6242681B1 (en) * 1998-11-25 2001-06-05 Yamaha Corporation Waveform reproduction device and method for performing pitch shift reproduction, loop reproduction and long-stream reproduction using compressed waveform samples
US20020120456A1 (en) * 2001-02-23 2002-08-29 Jakob Berg Method and arrangement for search and recording of media signals
US6570991B1 (en) * 1996-12-18 2003-05-27 Interval Research Corporation Multi-feature speech/music discrimination system
US20030101050A1 (en) * 2001-11-29 2003-05-29 Microsoft Corporation Real-time speech and music classifier
US20030171936A1 (en) * 2002-02-21 2003-09-11 Sall Mikhael A. Method of segmenting an audio stream
US20030229537A1 (en) * 2000-05-03 2003-12-11 Dunning Ted E. Relationship discovery engine
US20040069118A1 (en) * 2002-10-01 2004-04-15 Yamaha Corporation Compressed data structure and apparatus and method related thereto
US20040165730A1 (en) * 2001-04-13 2004-08-26 Crockett Brett G Segmenting audio signals into auditory events
US20050016360A1 (en) * 2003-07-24 2005-01-27 Tong Zhang System and method for automatic classification of music
US20050169114A1 (en) * 2002-02-20 2005-08-04 Hosung Ahn Digital recorder for selectively storing only a music section out of radio broadcasting contents and method thereof
US6998527B2 (en) * 2002-06-20 2006-02-14 Koninklijke Philips Electronics N.V. System and method for indexing and summarizing music videos
US20060074667A1 (en) * 2002-11-22 2006-04-06 Koninklijke Philips Electronics N.V. Speech recognition device and method
US20060085188A1 (en) * 2004-10-18 2006-04-20 Creative Technology Ltd. Method for Segmenting Audio Signals
US7120576B2 (en) * 2004-07-16 2006-10-10 Mindspeed Technologies, Inc. Low-complexity music detection algorithm and system
US7179980B2 (en) * 2003-12-12 2007-02-20 Nokia Corporation Automatic extraction of musical portions of an audio stream
US20070051230A1 (en) * 2005-09-06 2007-03-08 Takashi Hasegawa Information processing system and information processing method
US20070106406A1 (en) * 2005-10-28 2007-05-10 Victor Company Of Japan, Ltd. Music-piece classifying apparatus and method, and related computer program
US7277852B2 (en) * 2000-10-23 2007-10-02 Ntt Communications Corporation Method, system and storage medium for commercial and musical composition recognition and storage
US7315899B2 (en) * 2000-05-03 2008-01-01 Yahoo! Inc. System for controlling and enforcing playback restrictions for a media file by splitting the media file into usable and unusable portions for playback
US7336890B2 (en) * 2003-02-19 2008-02-26 Microsoft Corporation Automatic detection and segmentation of music videos in an audio/video stream
US20080097756A1 (en) * 2004-11-08 2008-04-24 Koninklijke Philips Electronics, N.V. Method of and Apparatus for Analyzing Audio Content and Reproducing Only the Desired Audio Data
US20090088878A1 (en) * 2005-12-27 2009-04-02 Isao Otsuka Method and Device for Detecting Music Segment, and Method and Device for Recording Data
US7558729B1 (en) * 2004-07-16 2009-07-07 Mindspeed Technologies, Inc. Music detection for enhancing echo cancellation and speech coding

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3388481B2 (en) * 1995-09-25 2003-03-24 日本電信電話株式会社 Automatic composition extraction of music information
US20040167767A1 (en) 2003-02-25 2004-08-26 Ziyou Xiong Method and system for extracting sports highlights from audio signals
US7446252B2 (en) * 2004-06-30 2008-11-04 Matsushita Electric Industrial Co., Ltd. Music information calculation apparatus and music reproduction apparatus
WO2006095847A1 (en) * 2005-03-11 2006-09-14 Pioneer Corporation Contents acquiring device, method used in such contents acquiring device, program used in such contents acquiring device, and recording medium with such program recorded therein
JP2006301134A (en) * 2005-04-19 2006-11-02 Hitachi Ltd Device and method for music detection, and sound recording and reproducing device
JP4201204B2 (en) * 2005-05-26 2008-12-24 Kddi株式会社 Audio information classification device
JP2008241850A (en) * 2007-03-26 2008-10-09 Sanyo Electric Co Ltd Recording or reproducing device

Patent Citations (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5402277A (en) * 1989-08-04 1995-03-28 Canon Kabushiki Kaisha Audio signal reproducing apparatus
US5233484A (en) * 1989-08-04 1993-08-03 Canon Kabushiki Kaisha Audio signal reproducing apparatus
US5712953A (en) * 1995-06-28 1998-01-27 Electronic Data Systems Corporation System and method for classification of audio or audio/video signals based on musical content
US6570991B1 (en) * 1996-12-18 2003-05-27 Interval Research Corporation Multi-feature speech/music discrimination system
US6169241B1 (en) * 1997-03-03 2001-01-02 Yamaha Corporation Sound source with free compression and expansion of voice independently of pitch
US6242681B1 (en) * 1998-11-25 2001-06-05 Yamaha Corporation Waveform reproduction device and method for performing pitch shift reproduction, loop reproduction and long-stream reproduction using compressed waveform samples
US20030229537A1 (en) * 2000-05-03 2003-12-11 Dunning Ted E. Relationship discovery engine
US7315899B2 (en) * 2000-05-03 2008-01-01 Yahoo! Inc. System for controlling and enforcing playback restrictions for a media file by splitting the media file into usable and unusable portions for playback
US7277852B2 (en) * 2000-10-23 2007-10-02 Ntt Communications Corporation Method, system and storage medium for commercial and musical composition recognition and storage
US20020120456A1 (en) * 2001-02-23 2002-08-29 Jakob Berg Method and arrangement for search and recording of media signals
US20040165730A1 (en) * 2001-04-13 2004-08-26 Crockett Brett G Segmenting audio signals into auditory events
US20030101050A1 (en) * 2001-11-29 2003-05-29 Microsoft Corporation Real-time speech and music classifier
US20050169114A1 (en) * 2002-02-20 2005-08-04 Hosung Ahn Digital recorder for selectively storing only a music section out of radio broadcasting contents and method thereof
US20030171936A1 (en) * 2002-02-21 2003-09-11 Sall Mikhael A. Method of segmenting an audio stream
US7346516B2 (en) * 2002-02-21 2008-03-18 Lg Electronics Inc. Method of segmenting an audio stream
US6998527B2 (en) * 2002-06-20 2006-02-14 Koninklijke Philips Electronics N.V. System and method for indexing and summarizing music videos
US7256340B2 (en) * 2002-10-01 2007-08-14 Yamaha Corporation Compressed data structure and apparatus and method related thereto
US20060081118A1 (en) * 2002-10-01 2006-04-20 Yamaha Corporation Compressed data structure and apparatus and method related thereto
US20040069118A1 (en) * 2002-10-01 2004-04-15 Yamaha Corporation Compressed data structure and apparatus and method related thereto
US20060074667A1 (en) * 2002-11-22 2006-04-06 Koninklijke Philips Electronics N.V. Speech recognition device and method
US7336890B2 (en) * 2003-02-19 2008-02-26 Microsoft Corporation Automatic detection and segmentation of music videos in an audio/video stream
US20050016360A1 (en) * 2003-07-24 2005-01-27 Tong Zhang System and method for automatic classification of music
US7179980B2 (en) * 2003-12-12 2007-02-20 Nokia Corporation Automatic extraction of musical portions of an audio stream
US7120576B2 (en) * 2004-07-16 2006-10-10 Mindspeed Technologies, Inc. Low-complexity music detection algorithm and system
US7558729B1 (en) * 2004-07-16 2009-07-07 Mindspeed Technologies, Inc. Music detection for enhancing echo cancellation and speech coding
US20060085188A1 (en) * 2004-10-18 2006-04-20 Creative Technology Ltd. Method for Segmenting Audio Signals
US20080097756A1 (en) * 2004-11-08 2008-04-24 Koninklijke Philips Electronics, N.V. Method of and Apparatus for Analyzing Audio Content and Reproducing Only the Desired Audio Data
US20070051230A1 (en) * 2005-09-06 2007-03-08 Takashi Hasegawa Information processing system and information processing method
US20070106406A1 (en) * 2005-10-28 2007-05-10 Victor Company Of Japan, Ltd. Music-piece classifying apparatus and method, and related computer program
US7544881B2 (en) * 2005-10-28 2009-06-09 Victor Company Of Japan, Ltd. Music-piece classifying apparatus and method, and related computer program
US20090088878A1 (en) * 2005-12-27 2009-04-02 Isao Otsuka Method and Device for Detecting Music Segment, and Method and Device for Recording Data

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7745714B2 (en) * 2007-03-26 2010-06-29 Sanyo Electric Co., Ltd. Recording or playback apparatus and musical piece detecting apparatus
US20200160821A1 (en) * 2017-07-25 2020-05-21 Yamaha Corporation Information processing method
US11568244B2 (en) * 2017-07-25 2023-01-31 Yamaha Corporation Information processing method and apparatus
CN107481739A (en) * 2017-08-16 2017-12-15 成都品果科技有限公司 Audio cutting method and device

Also Published As

Publication number Publication date
US7745714B2 (en) 2010-06-29
JP2008241850A (en) 2008-10-09

Similar Documents

Publication Publication Date Title
US7062442B2 (en) Method and arrangement for search and recording of media signals
US8190441B2 (en) Playback of compressed media files without quantization gaps
US7745714B2 (en) Recording or playback apparatus and musical piece detecting apparatus
US20090012637A1 (en) Chorus position detection device
US8682132B2 (en) Method and device for detecting music segment, and method and device for recording data
US20040059570A1 (en) Feature quantity extracting apparatus
JP3886372B2 (en) Acoustic inflection point extraction apparatus and method, acoustic reproduction apparatus and method, acoustic signal editing apparatus, acoustic inflection point extraction method program recording medium, acoustic reproduction method program recording medium, acoustic signal editing method program recording medium, acoustic inflection point extraction method Program, sound reproduction method program, sound signal editing method program
JP2006202127A (en) Recommended information presentation device and recommended information presentation method or the like
JP4877811B2 (en) Specific section extraction device, music recording / playback device, music distribution system
US20110235811A1 (en) Music track extraction device and music track recording device
US8069177B2 (en) Information selecting method, information selecting device and so on
KR20050003457A (en) Signal processing method and arrangement
US20040073422A1 (en) Apparatus and methods for surreptitiously recording and analyzing audio for later auditioning and application
JP4278667B2 (en) Music composition apparatus, music composition method, and music composition program
JP2010078984A (en) Musical piece extraction device and musical piece recording device
JP2004334160A (en) Characteristic amount extraction device
JP4633022B2 (en) Music editing device and music editing program.
JPWO2005093750A1 (en) Digital dubbing device
WO2009101808A1 (en) Music recorder
KR102431737B1 (en) Method of searching highlight in multimedia data and apparatus therof
KR20110064901A (en) Audio apparatus for vehicle having capability of automatically setting equalization effect and method for setting equalization effect
JP4961300B2 (en) Music match determination device, music recording device, music match determination method, music recording method, music match determination program, and music recording program
JP2010027115A (en) Music recording and reproducing device
EP1417583B1 (en) Method for receiving a media signal
JP2001085962A (en) Equalizer setting device in audio equipment

Legal Events

Date Code Title Description
AS Assignment

Owner name: SANYO ELECTRIC CO., LTD., JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MATSUMOTO, SATORU;YAMAMOTO, YUJI;KOGA, TATSUO;REEL/FRAME:020691/0187

Effective date: 20080313

Owner name: SANYO ELECTRIC CO., LTD.,JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MATSUMOTO, SATORU;YAMAMOTO, YUJI;KOGA, TATSUO;REEL/FRAME:020691/0187

Effective date: 20080313

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552)

Year of fee payment: 8

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20220629