US20150220629A1 - Sound Melody as Web Search Query - Google Patents

Sound Melody as Web Search Query Download PDF

Info

Publication number
US20150220629A1
US20150220629A1 US14/611,609 US201514611609A US2015220629A1 US 20150220629 A1 US20150220629 A1 US 20150220629A1 US 201514611609 A US201514611609 A US 201514611609A US 2015220629 A1 US2015220629 A1 US 2015220629A1
Authority
US
United States
Prior art keywords
sound
melody
internet
sound clip
search
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/611,609
Inventor
Darren Nolf
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US14/611,609 priority Critical patent/US20150220629A1/en
Publication of US20150220629A1 publication Critical patent/US20150220629A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • G06F17/30705
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/54Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for retrieval
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/683Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/951Indexing; Web crawling techniques
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0008Associated control or indicating means
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/056Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for extraction or identification of individual instrumental parts, e.g. melody, chords, bass; Identification or separation of instrumental parts by their characteristic voices or timbres
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/121Musical libraries, i.e. musical databases indexed by musical parameters, wavetables, indexing schemes using musical parameters, musical rule bases or knowledge bases, e.g. for automatic composing methods
    • G10H2240/131Library retrieval, i.e. searching a database or selecting a specific musical piece, segment, pattern, rule or parameter set
    • G10H2240/141Library retrieval matching, i.e. any of the steps of matching an inputted segment or phrase with musical database contents, e.g. query by humming, singing or playing; the steps may include, e.g. musical analysis of the input, musical feature extraction, query formulation, or details of the retrieval process

Definitions

  • the invention relates generally to Internet search infrastructure and in particular to an Internet search from a sound melody.
  • MP3 MPEG-1
  • the proliferation of audio data files in the MPEG-1 (MP3) format has opened a new world of data analysis and filing, based on the song titles, album titles, performers, melodies, and lyrics of the world of music.
  • Data collections such as WikipediaTM and YouTubeTM provide an enormous wealth of searchable data.
  • Acquiring the title of a song, or the name of performer or band, based on nothing more than a few bars of a melody presents a substantial challenge. Fortunately, software is now available which will recognize such melodies.
  • An Internet search infrastructure in the form of a mobile application, which provides melody recognition and voice-to-text software to analyze captured melodies from a smartphone or tablet, turns the data into a textual search and then searches the Internet for similar lyrics and melodies, would resolve this problem.
  • the invention is directed to Internet search infrastructure which searches from a sound melody.
  • the user may capture a melody or sing a known portion of the melody, and record it with the mobile application.
  • the infrastructure employs melody recognition and voice-to-text software to analyze and recognize a melody which has been captured from a smartphone or tablet, then searches the Internet for text or data containing the portion of the lyrics or melody which was captured.
  • the search infrastructure is controlled from a mobile application installed in a tablet or smartphone.
  • the FIGURE is a system diagram displaying the electronic system components, features, communications, and their interconnections, and other relevant processes and events during the course of using the invention.
  • the invention is directed to Internet search infrastructure which searches from a sound melody.
  • the user may capture a melody or sing a known portion of the melody, and record it with the mobile application.
  • the infrastructure employs melody recognition and voice-to-text software to analyze and recognize a melody which has been captured from a smartphone or tablet, then searches the Internet for text or data containing the portion of the lyrics or melody which was captured.
  • the search infrastructure is controlled from a mobile application installed in a tablet or smartphone (generically, a mobile computing device).
  • Components of the first exemplary embodiment include the user's tablet or smartphone; a mobile application; and an Internet website.
  • the mobile application may be used to select and capture a portion of a melody, whether it is sung by the user or previously recorded.
  • the search function then identifies salient characteristics of the captured melody, using melody recognition software and voice-to-text software, and correlates the captured melody with other melodies on the Internet sharing the same salient characteristics, and with images and text which share elements of the captured melody; the salient characteristics may, but need not be programmatically defined, and may be generated or learned through applicable well-known recognition algorithms.
  • the search function may rely upon voice-to-text analysis exclusively, or upon melody characteristics exclusively, or any combination of these or other characteristics.
  • the search results may include the title of the song, the titles of any albums containing the song, the performers and bands (generically, artists) who performed, recorded, or wrote (generically, produced) the song, the complete lyrics of the song, and any background information which may be available.
  • the search results may also include links to digital copies of the song which may be obtained and played, such as a YouTube music video or an iTunes® selection.
  • the user is preferably able to record up to about fifteen seconds of the song portion or pause streaming music, such as a satellite radio broadcast or podcast, and record a portion of the music being played to enable the most accurate result.
  • a sound clip is generated either from a sound source, typically either the device microphone or from any external source, including local audio files as well as broadcasts via satellite, Internet, or terrestrial radio.
  • the system analyzes the recorded portion and either presents the salient characteristics to a search engine as a search query, or consults a melody recognition database provided by the website, thereby converting the image into a textual or data search query.
  • the search includes data services such as WikipediaTM, and music video hosting services such as YouTubeTM.
  • the search function then delivers search result pages to the user via the visual display of a user's electronic device, such as a tablet or smartphone.
  • a user's electronic device such as a tablet or smartphone.
  • Foreign language selections are also preferably searchable.
  • the system also preferably features the ability to narrow the search by adjourning the partial lyric simultaneously in the mix of the rhythm provided to enable more accurate results.
  • the user downloads and installs the related mobile application on the user's mobile electronic device, such as a tablet or smartphone.
  • the user then activates the application, and may select and capture a portion of a melody being played, or sing and record a known portion of the melody.
  • the user may use the application to pause streaming audio to conduct the search.
  • the application analyzes the captured melody for salient characteristics and conducts an Internet search.
  • the search result pages may be displayed as a pop-up dialog box on the visual display of the user's electronic device.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Multimedia (AREA)
  • General Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Library & Information Science (AREA)
  • Signal Processing (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

An Internet search infrastructure features searches from a sound melody. The user may capture a melody or sing a known portion of the melody, and record it with the mobile application. The infrastructure employs melody recognition and voice-to-text software to analyze and recognize a melody which has been captured from a smartphone or tablet, then searches the Internet for text or data containing the portion of the lyrics or melody which was captured. The search infrastructure is controlled from a mobile application installed in a tablet or smartphone.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application claims the benefit of U.S. Provisional Application No. 61/933,903, filed Jan. 31, 2014, which is hereby incorporated by reference.
  • STATEMENT REGARDING FEDERALLY SPONSORED RESEARCH OR DEVELOPMENT
  • Not Applicable
  • PARTIES TO A JOINT RESEARCH AGREEMENT
  • Not Applicable
  • REFERENCE TO SEQUENCE LISTING, A TABLE, OR A COMPUTER PROGRAM LISTING COMPACT DISK APPENDIX
  • Not Applicable
  • BACKGROUND OF THE INVENTION
  • The invention relates generally to Internet search infrastructure and in particular to an Internet search from a sound melody. The proliferation of audio data files in the MPEG-1 (MP3) format has opened a new world of data analysis and filing, based on the song titles, album titles, performers, melodies, and lyrics of the world of music. But the problem of identifying a song based on a few words of the lyrics, or a few bars of the melody, is still a challenge. Data collections such as Wikipedia™ and YouTube™ provide an enormous wealth of searchable data. Acquiring the title of a song, or the name of performer or band, based on nothing more than a few bars of a melody presents a substantial challenge. Fortunately, software is now available which will recognize such melodies. An Internet search infrastructure in the form of a mobile application, which provides melody recognition and voice-to-text software to analyze captured melodies from a smartphone or tablet, turns the data into a textual search and then searches the Internet for similar lyrics and melodies, would resolve this problem.
  • SUMMARY OF THE INVENTION
  • Accordingly, the invention is directed to Internet search infrastructure which searches from a sound melody. The user may capture a melody or sing a known portion of the melody, and record it with the mobile application. The infrastructure employs melody recognition and voice-to-text software to analyze and recognize a melody which has been captured from a smartphone or tablet, then searches the Internet for text or data containing the portion of the lyrics or melody which was captured. The search infrastructure is controlled from a mobile application installed in a tablet or smartphone.
  • Additional features and advantages of the invention will be set forth in the description which follows, and will be apparent from the description, or may be learned by practice of the invention. The foregoing general description and the following detailed description are exemplary and explanatory and are intended to provide further explanation of the invention.
  • BRIEF DESCRIPTION OF THE DRAWING
  • The accompanying drawing is included to provide a further understanding of the invention and is incorporated into and constitutes a part of the specification. It illustrate one embodiment of the invention and, together with the description, serves to explain the principles of the invention.
  • The FIGURE is a system diagram displaying the electronic system components, features, communications, and their interconnections, and other relevant processes and events during the course of using the invention.
  • DETAILED DESCRIPTION OF THE INVENTION
  • Referring now to the invention in more detail, the invention is directed to Internet search infrastructure which searches from a sound melody. The user may capture a melody or sing a known portion of the melody, and record it with the mobile application. The infrastructure employs melody recognition and voice-to-text software to analyze and recognize a melody which has been captured from a smartphone or tablet, then searches the Internet for text or data containing the portion of the lyrics or melody which was captured. The search infrastructure is controlled from a mobile application installed in a tablet or smartphone (generically, a mobile computing device).
  • Components of the first exemplary embodiment include the user's tablet or smartphone; a mobile application; and an Internet website. The mobile application may be used to select and capture a portion of a melody, whether it is sung by the user or previously recorded. The search function then identifies salient characteristics of the captured melody, using melody recognition software and voice-to-text software, and correlates the captured melody with other melodies on the Internet sharing the same salient characteristics, and with images and text which share elements of the captured melody; the salient characteristics may, but need not be programmatically defined, and may be generated or learned through applicable well-known recognition algorithms. Optionally, the search function may rely upon voice-to-text analysis exclusively, or upon melody characteristics exclusively, or any combination of these or other characteristics. The search results may include the title of the song, the titles of any albums containing the song, the performers and bands (generically, artists) who performed, recorded, or wrote (generically, produced) the song, the complete lyrics of the song, and any background information which may be available. The search results may also include links to digital copies of the song which may be obtained and played, such as a YouTube music video or an iTunes® selection.
  • The user is preferably able to record up to about fifteen seconds of the song portion or pause streaming music, such as a satellite radio broadcast or podcast, and record a portion of the music being played to enable the most accurate result. Generically, a sound clip is generated either from a sound source, typically either the device microphone or from any external source, including local audio files as well as broadcasts via satellite, Internet, or terrestrial radio. The system then analyzes the recorded portion and either presents the salient characteristics to a search engine as a search query, or consults a melody recognition database provided by the website, thereby converting the image into a textual or data search query. The search includes data services such as Wikipedia™, and music video hosting services such as YouTube™. The search function then delivers search result pages to the user via the visual display of a user's electronic device, such as a tablet or smartphone. Foreign language selections are also preferably searchable. The system also preferably features the ability to narrow the search by adjourning the partial lyric simultaneously in the mix of the rhythm provided to enable more accurate results.
  • To use the first exemplary embodiment, the user downloads and installs the related mobile application on the user's mobile electronic device, such as a tablet or smartphone. The user then activates the application, and may select and capture a portion of a melody being played, or sing and record a known portion of the melody. The user may use the application to pause streaming audio to conduct the search. The application analyzes the captured melody for salient characteristics and conducts an Internet search. The search result pages may be displayed as a pop-up dialog box on the visual display of the user's electronic device.
  • Components, component sizes, and materials listed above are preferable, but artisans will recognize that alternate components and materials could be selected without altering the scope of the invention.
  • While the foregoing written description of the invention enables one of ordinary skill to make and use what is presently considered to be the best mode thereof, those of ordinary skill in the art will understand and appreciate the existence of variations, combinations, and equivalents of the specific embodiment, method, and examples herein. The invention should, therefore, not be limited by the above described embodiment, method, and examples, but by all embodiments and methods within the scope and spirit of the invention.

Claims (20)

I claim:
1. A method for performing an internet search from a sound melody comprising a method of operating a mobile computing device:
(a) obtaining a sound clip from a sound source;
(b) generating a search query based on at least one of the group of performing voice-to-text analysis on said sound clip or identifying the salient characteristics of a melody contained within said sound clip;
(c) searching the internet using said search query; and
(d) returning, as search results including information relating to any of the group of the title of a song contained in said sound clip, the titles of albums containing the song contained in said sound clip, the artists who produced the song contained in said sound clip, the lyrics of the song contained in said sound clip, background information relating to the song contained in said sound clip, recordings of the song contained in said sound clip, or media relating to the song contained in said sound clip.
2. The method for performing an internet search from a sound melody of claim 1 wherein said search query is generated exclusively by performing voice-to-text analysis.
3. The method for per performing an internet search from a sound melody of claim 1 wherein said search query is generated by performing voice-to-text analysis in combination with identifying the salient characteristics of said melody.
4. The method for per performing an internet search from a sound melody of claim 1 wherein performing voice-to-text analysis is enabled for one or more foreign languages.
5. The method for per performing an internet search from a sound melody of claim 2 wherein performing voice-to-text analysis is enabled for one or more foreign languages.
6. The method for per performing an internet search from a sound melody of claim 3 wherein performing voice-to-text analysis is enabled for one or more foreign languages.
7. The method for per performing an internet search from a sound melody of claim 1 wherein said sound source is selected from the group of a microphone, a local audio file, a satellite broadcast, an internet broadcast, or a terrestrial radio broadcast.
8. The method for per performing an internet search from a sound melody of claim 2 wherein said sound source is selected from the group of a microphone, a local audio file, a satellite broadcast, an internet broadcast, or a terrestrial radio broadcast.
9. The method for per performing an internet search from a sound melody of claim 3 wherein said sound source is selected from the group of a microphone, a local audio file, a satellite broadcast, an internet broadcast, or a terrestrial radio broadcast.
10. The method for per performing an internet search from a sound melody of claim 4 wherein said sound source is selected from the group of a microphone, a local audio file, a satellite broadcast, an internet broadcast, or a terrestrial radio broadcast.
11. The method for per performing an internet search from a sound melody of claim 5 wherein said sound source is selected from the group of a microphone, a local audio file, a satellite broadcast, an internet broadcast, or a terrestrial radio broadcast.
12. The method for per performing an internet search from a sound melody of claim 6 wherein said sound source is selected from the group of a microphone, a local audio file, a satellite broadcast, an internet broadcast, or a terrestrial radio broadcast.
13. A computer system, the system comprising a mobile computing device, said mobile computing device being in electronic communication with at least one of the group of a microphone or an external audio source, and said mobile computing device being in electronic communication with the internet; said mobile computing device being configured for:
(a) obtaining a sound clip from a sound source;
(b) generating a search query based on at least one of the group of performing voice-to-text analysis on said sound clip or identifying the salient characteristics of a melody contained within said sound clip;
(c) searching the internet using said search query; and
(d) returning, as search results including information relating to any of the group of the title of a song contained in said sound clip, the titles of albums containing the song contained in said sound clip, the artists who produced the song contained in said sound clip, the lyrics of the song contained in said sound clip, background information relating to the song contained in said sound clip, recordings of the song contained in said sound clip, or media relating to the song contained in said sound clip.
14. The computer system of claim 13 wherein said search query is generated exclusively by performing voice-to-text analysis.
15. The computer system of claim 13 wherein said search query is generated by performing voice-to-text analysis in combination with identifying the salient characteristics of said melody.
16. The computer system of claim 13 wherein said search query is generated by performing voice-to-text analysis in combination with identifying the salient characteristics of said melody.
17. A computer program product, said computer program product being embodied in a tangible non-transitory computer readable storage medium and comprising computer instructions for:
(a) obtaining a sound clip from a sound source;
(b) generating a search query based on at least one of the group of performing voice-to-text analysis on said sound clip or identifying the salient characteristics of a melody contained within said sound clip;
(c) searching the internet using said search query; and
(d) returning, as search results including information relating to any of the group of the title of a song contained in said sound clip, the titles of albums containing the song contained in said sound clip, the artists who produced the song contained in said sound clip, the lyrics of the song contained in said sound clip, background information relating to the song contained in said sound clip, recordings of the song contained in said sound clip, or media relating to the song contained in said sound clip.
18. The computer program product of claim 17 wherein said search query is generated exclusively by performing voice-to-text analysis.
19. The computer program product of claim 17 wherein said search query is generated by performing voice-to-text analysis in combination with identifying the salient characteristics of said melody.
20. The computer system of claim 17 wherein said search query is generated by performing voice-to-text analysis in combination with identifying the salient characteristics of said melody.
US14/611,609 2014-01-31 2015-02-02 Sound Melody as Web Search Query Abandoned US20150220629A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/611,609 US20150220629A1 (en) 2014-01-31 2015-02-02 Sound Melody as Web Search Query

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201461933903P 2014-01-31 2014-01-31
US14/611,609 US20150220629A1 (en) 2014-01-31 2015-02-02 Sound Melody as Web Search Query

Publications (1)

Publication Number Publication Date
US20150220629A1 true US20150220629A1 (en) 2015-08-06

Family

ID=53755029

Family Applications (1)

Application Number Title Priority Date Filing Date
US14/611,609 Abandoned US20150220629A1 (en) 2014-01-31 2015-02-02 Sound Melody as Web Search Query

Country Status (1)

Country Link
US (1) US20150220629A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017028116A1 (en) * 2015-08-16 2017-02-23 胡丹丽 Intelligent desktop speaker and method for controlling intelligent desktop speaker
CN107341158A (en) * 2016-05-03 2017-11-10 斑马网络技术有限公司 Information acquisition method, device and the equipment of multimedia file

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020038597A1 (en) * 2000-09-29 2002-04-04 Jyri Huopaniemi Method and a system for recognizing a melody
US20030040904A1 (en) * 2001-08-27 2003-02-27 Nec Research Institute, Inc. Extracting classifying data in music from an audio bitstream
US8140329B2 (en) * 2003-07-28 2012-03-20 Sony Corporation Method and apparatus for automatically recognizing audio data
US20130238336A1 (en) * 2012-03-08 2013-09-12 Google Inc. Recognizing speech in multiple languages

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020038597A1 (en) * 2000-09-29 2002-04-04 Jyri Huopaniemi Method and a system for recognizing a melody
US20030040904A1 (en) * 2001-08-27 2003-02-27 Nec Research Institute, Inc. Extracting classifying data in music from an audio bitstream
US8140329B2 (en) * 2003-07-28 2012-03-20 Sony Corporation Method and apparatus for automatically recognizing audio data
US20130238336A1 (en) * 2012-03-08 2013-09-12 Google Inc. Recognizing speech in multiple languages

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017028116A1 (en) * 2015-08-16 2017-02-23 胡丹丽 Intelligent desktop speaker and method for controlling intelligent desktop speaker
CN107341158A (en) * 2016-05-03 2017-11-10 斑马网络技术有限公司 Information acquisition method, device and the equipment of multimedia file

Similar Documents

Publication Publication Date Title
CN106960051B (en) Audio playing method and device based on electronic book and terminal equipment
CA2899657C (en) Method and device for audio recognition
US20130007043A1 (en) Voice description of time-based media for indexing and searching
US20150169747A1 (en) Systems and methods for automatically suggesting media accompaniments based on identified media content
CN107591149A (en) Audio synthetic method, device and storage medium
WO2013003854A2 (en) Method and system for communicating between a sender and a recipient via a personalized message including an audio clip extracted from a pre-existing recording
US9558272B2 (en) Method of and a system for matching audio tracks using chromaprints with a fast candidate selection routine
US20160336026A1 (en) Deep tagging background noises
US9224385B1 (en) Unified recognition of speech and music
US20140114656A1 (en) Electronic device capable of generating tag file for media file based on speaker recognition
WO2016112841A1 (en) Information processing method and client, and computer storage medium
JP2011048729A (en) Apparatus, method and program for query extraction
EP3839952A1 (en) Masking systems and methods
US11785276B2 (en) Event source content and remote content synchronization
US20150220629A1 (en) Sound Melody as Web Search Query
WO2016024171A1 (en) Method of and a system for indexing audio tracks using chromaprints
Kurth et al. Syncplayer-An Advanced System for Multimodal Music Access.
KR20080107143A (en) System and method for recommendation of music and moving video based on audio signal processing
Raimond et al. Automated semantic tagging of speech audio
JP5474591B2 (en) Image selection apparatus, image selection method, and image selection program
US11023520B1 (en) Background audio identification for query disambiguation
Simou et al. Towards blind quality assessment of concert audio recordings using deep neural networks
Phang et al. Tools and technologies for enhancing access to audiovisual archives: The Singapore journey
Brinkman et al. Online music recognition: the Echoprint system
JP5713775B2 (en) Music search device

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION