US20150220629A1 - Sound Melody as Web Search Query - Google Patents
Sound Melody as Web Search Query Download PDFInfo
- Publication number
- US20150220629A1 US20150220629A1 US14/611,609 US201514611609A US2015220629A1 US 20150220629 A1 US20150220629 A1 US 20150220629A1 US 201514611609 A US201514611609 A US 201514611609A US 2015220629 A1 US2015220629 A1 US 2015220629A1
- Authority
- US
- United States
- Prior art keywords
- sound
- melody
- internet
- sound clip
- search
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 claims description 17
- 238000004458 analytical method Methods 0.000 claims description 15
- 238000004590 computer program Methods 0.000 claims description 5
- 230000006870 function Effects 0.000 description 3
- 230000008901 benefit Effects 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 238000007405 data analysis Methods 0.000 description 1
- 238000013480 data collection Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000035755 proliferation Effects 0.000 description 1
- 230000033764 rhythmic process Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
-
- G06F17/30705—
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/54—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for retrieval
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/60—Information retrieval; Database structures therefor; File system structures therefor of audio data
- G06F16/68—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/683—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/951—Indexing; Web crawling techniques
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0008—Associated control or indicating means
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2210/00—Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
- G10H2210/031—Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
- G10H2210/056—Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for extraction or identification of individual instrumental parts, e.g. melody, chords, bass; Identification or separation of instrumental parts by their characteristic voices or timbres
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2240/00—Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
- G10H2240/121—Musical libraries, i.e. musical databases indexed by musical parameters, wavetables, indexing schemes using musical parameters, musical rule bases or knowledge bases, e.g. for automatic composing methods
- G10H2240/131—Library retrieval, i.e. searching a database or selecting a specific musical piece, segment, pattern, rule or parameter set
- G10H2240/141—Library retrieval matching, i.e. any of the steps of matching an inputted segment or phrase with musical database contents, e.g. query by humming, singing or playing; the steps may include, e.g. musical analysis of the input, musical feature extraction, query formulation, or details of the retrieval process
Definitions
- the invention relates generally to Internet search infrastructure and in particular to an Internet search from a sound melody.
- MP3 MPEG-1
- the proliferation of audio data files in the MPEG-1 (MP3) format has opened a new world of data analysis and filing, based on the song titles, album titles, performers, melodies, and lyrics of the world of music.
- Data collections such as WikipediaTM and YouTubeTM provide an enormous wealth of searchable data.
- Acquiring the title of a song, or the name of performer or band, based on nothing more than a few bars of a melody presents a substantial challenge. Fortunately, software is now available which will recognize such melodies.
- An Internet search infrastructure in the form of a mobile application, which provides melody recognition and voice-to-text software to analyze captured melodies from a smartphone or tablet, turns the data into a textual search and then searches the Internet for similar lyrics and melodies, would resolve this problem.
- the invention is directed to Internet search infrastructure which searches from a sound melody.
- the user may capture a melody or sing a known portion of the melody, and record it with the mobile application.
- the infrastructure employs melody recognition and voice-to-text software to analyze and recognize a melody which has been captured from a smartphone or tablet, then searches the Internet for text or data containing the portion of the lyrics or melody which was captured.
- the search infrastructure is controlled from a mobile application installed in a tablet or smartphone.
- the FIGURE is a system diagram displaying the electronic system components, features, communications, and their interconnections, and other relevant processes and events during the course of using the invention.
- the invention is directed to Internet search infrastructure which searches from a sound melody.
- the user may capture a melody or sing a known portion of the melody, and record it with the mobile application.
- the infrastructure employs melody recognition and voice-to-text software to analyze and recognize a melody which has been captured from a smartphone or tablet, then searches the Internet for text or data containing the portion of the lyrics or melody which was captured.
- the search infrastructure is controlled from a mobile application installed in a tablet or smartphone (generically, a mobile computing device).
- Components of the first exemplary embodiment include the user's tablet or smartphone; a mobile application; and an Internet website.
- the mobile application may be used to select and capture a portion of a melody, whether it is sung by the user or previously recorded.
- the search function then identifies salient characteristics of the captured melody, using melody recognition software and voice-to-text software, and correlates the captured melody with other melodies on the Internet sharing the same salient characteristics, and with images and text which share elements of the captured melody; the salient characteristics may, but need not be programmatically defined, and may be generated or learned through applicable well-known recognition algorithms.
- the search function may rely upon voice-to-text analysis exclusively, or upon melody characteristics exclusively, or any combination of these or other characteristics.
- the search results may include the title of the song, the titles of any albums containing the song, the performers and bands (generically, artists) who performed, recorded, or wrote (generically, produced) the song, the complete lyrics of the song, and any background information which may be available.
- the search results may also include links to digital copies of the song which may be obtained and played, such as a YouTube music video or an iTunes® selection.
- the user is preferably able to record up to about fifteen seconds of the song portion or pause streaming music, such as a satellite radio broadcast or podcast, and record a portion of the music being played to enable the most accurate result.
- a sound clip is generated either from a sound source, typically either the device microphone or from any external source, including local audio files as well as broadcasts via satellite, Internet, or terrestrial radio.
- the system analyzes the recorded portion and either presents the salient characteristics to a search engine as a search query, or consults a melody recognition database provided by the website, thereby converting the image into a textual or data search query.
- the search includes data services such as WikipediaTM, and music video hosting services such as YouTubeTM.
- the search function then delivers search result pages to the user via the visual display of a user's electronic device, such as a tablet or smartphone.
- a user's electronic device such as a tablet or smartphone.
- Foreign language selections are also preferably searchable.
- the system also preferably features the ability to narrow the search by adjourning the partial lyric simultaneously in the mix of the rhythm provided to enable more accurate results.
- the user downloads and installs the related mobile application on the user's mobile electronic device, such as a tablet or smartphone.
- the user then activates the application, and may select and capture a portion of a melody being played, or sing and record a known portion of the melody.
- the user may use the application to pause streaming audio to conduct the search.
- the application analyzes the captured melody for salient characteristics and conducts an Internet search.
- the search result pages may be displayed as a pop-up dialog box on the visual display of the user's electronic device.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Databases & Information Systems (AREA)
- Multimedia (AREA)
- General Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Library & Information Science (AREA)
- Signal Processing (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
An Internet search infrastructure features searches from a sound melody. The user may capture a melody or sing a known portion of the melody, and record it with the mobile application. The infrastructure employs melody recognition and voice-to-text software to analyze and recognize a melody which has been captured from a smartphone or tablet, then searches the Internet for text or data containing the portion of the lyrics or melody which was captured. The search infrastructure is controlled from a mobile application installed in a tablet or smartphone.
Description
- This application claims the benefit of U.S. Provisional Application No. 61/933,903, filed Jan. 31, 2014, which is hereby incorporated by reference.
- Not Applicable
- Not Applicable
- Not Applicable
- The invention relates generally to Internet search infrastructure and in particular to an Internet search from a sound melody. The proliferation of audio data files in the MPEG-1 (MP3) format has opened a new world of data analysis and filing, based on the song titles, album titles, performers, melodies, and lyrics of the world of music. But the problem of identifying a song based on a few words of the lyrics, or a few bars of the melody, is still a challenge. Data collections such as Wikipedia™ and YouTube™ provide an enormous wealth of searchable data. Acquiring the title of a song, or the name of performer or band, based on nothing more than a few bars of a melody presents a substantial challenge. Fortunately, software is now available which will recognize such melodies. An Internet search infrastructure in the form of a mobile application, which provides melody recognition and voice-to-text software to analyze captured melodies from a smartphone or tablet, turns the data into a textual search and then searches the Internet for similar lyrics and melodies, would resolve this problem.
- Accordingly, the invention is directed to Internet search infrastructure which searches from a sound melody. The user may capture a melody or sing a known portion of the melody, and record it with the mobile application. The infrastructure employs melody recognition and voice-to-text software to analyze and recognize a melody which has been captured from a smartphone or tablet, then searches the Internet for text or data containing the portion of the lyrics or melody which was captured. The search infrastructure is controlled from a mobile application installed in a tablet or smartphone.
- Additional features and advantages of the invention will be set forth in the description which follows, and will be apparent from the description, or may be learned by practice of the invention. The foregoing general description and the following detailed description are exemplary and explanatory and are intended to provide further explanation of the invention.
- The accompanying drawing is included to provide a further understanding of the invention and is incorporated into and constitutes a part of the specification. It illustrate one embodiment of the invention and, together with the description, serves to explain the principles of the invention.
- The FIGURE is a system diagram displaying the electronic system components, features, communications, and their interconnections, and other relevant processes and events during the course of using the invention.
- Referring now to the invention in more detail, the invention is directed to Internet search infrastructure which searches from a sound melody. The user may capture a melody or sing a known portion of the melody, and record it with the mobile application. The infrastructure employs melody recognition and voice-to-text software to analyze and recognize a melody which has been captured from a smartphone or tablet, then searches the Internet for text or data containing the portion of the lyrics or melody which was captured. The search infrastructure is controlled from a mobile application installed in a tablet or smartphone (generically, a mobile computing device).
- Components of the first exemplary embodiment include the user's tablet or smartphone; a mobile application; and an Internet website. The mobile application may be used to select and capture a portion of a melody, whether it is sung by the user or previously recorded. The search function then identifies salient characteristics of the captured melody, using melody recognition software and voice-to-text software, and correlates the captured melody with other melodies on the Internet sharing the same salient characteristics, and with images and text which share elements of the captured melody; the salient characteristics may, but need not be programmatically defined, and may be generated or learned through applicable well-known recognition algorithms. Optionally, the search function may rely upon voice-to-text analysis exclusively, or upon melody characteristics exclusively, or any combination of these or other characteristics. The search results may include the title of the song, the titles of any albums containing the song, the performers and bands (generically, artists) who performed, recorded, or wrote (generically, produced) the song, the complete lyrics of the song, and any background information which may be available. The search results may also include links to digital copies of the song which may be obtained and played, such as a YouTube music video or an iTunes® selection.
- The user is preferably able to record up to about fifteen seconds of the song portion or pause streaming music, such as a satellite radio broadcast or podcast, and record a portion of the music being played to enable the most accurate result. Generically, a sound clip is generated either from a sound source, typically either the device microphone or from any external source, including local audio files as well as broadcasts via satellite, Internet, or terrestrial radio. The system then analyzes the recorded portion and either presents the salient characteristics to a search engine as a search query, or consults a melody recognition database provided by the website, thereby converting the image into a textual or data search query. The search includes data services such as Wikipedia™, and music video hosting services such as YouTube™. The search function then delivers search result pages to the user via the visual display of a user's electronic device, such as a tablet or smartphone. Foreign language selections are also preferably searchable. The system also preferably features the ability to narrow the search by adjourning the partial lyric simultaneously in the mix of the rhythm provided to enable more accurate results.
- To use the first exemplary embodiment, the user downloads and installs the related mobile application on the user's mobile electronic device, such as a tablet or smartphone. The user then activates the application, and may select and capture a portion of a melody being played, or sing and record a known portion of the melody. The user may use the application to pause streaming audio to conduct the search. The application analyzes the captured melody for salient characteristics and conducts an Internet search. The search result pages may be displayed as a pop-up dialog box on the visual display of the user's electronic device.
- Components, component sizes, and materials listed above are preferable, but artisans will recognize that alternate components and materials could be selected without altering the scope of the invention.
- While the foregoing written description of the invention enables one of ordinary skill to make and use what is presently considered to be the best mode thereof, those of ordinary skill in the art will understand and appreciate the existence of variations, combinations, and equivalents of the specific embodiment, method, and examples herein. The invention should, therefore, not be limited by the above described embodiment, method, and examples, but by all embodiments and methods within the scope and spirit of the invention.
Claims (20)
1. A method for performing an internet search from a sound melody comprising a method of operating a mobile computing device:
(a) obtaining a sound clip from a sound source;
(b) generating a search query based on at least one of the group of performing voice-to-text analysis on said sound clip or identifying the salient characteristics of a melody contained within said sound clip;
(c) searching the internet using said search query; and
(d) returning, as search results including information relating to any of the group of the title of a song contained in said sound clip, the titles of albums containing the song contained in said sound clip, the artists who produced the song contained in said sound clip, the lyrics of the song contained in said sound clip, background information relating to the song contained in said sound clip, recordings of the song contained in said sound clip, or media relating to the song contained in said sound clip.
2. The method for performing an internet search from a sound melody of claim 1 wherein said search query is generated exclusively by performing voice-to-text analysis.
3. The method for per performing an internet search from a sound melody of claim 1 wherein said search query is generated by performing voice-to-text analysis in combination with identifying the salient characteristics of said melody.
4. The method for per performing an internet search from a sound melody of claim 1 wherein performing voice-to-text analysis is enabled for one or more foreign languages.
5. The method for per performing an internet search from a sound melody of claim 2 wherein performing voice-to-text analysis is enabled for one or more foreign languages.
6. The method for per performing an internet search from a sound melody of claim 3 wherein performing voice-to-text analysis is enabled for one or more foreign languages.
7. The method for per performing an internet search from a sound melody of claim 1 wherein said sound source is selected from the group of a microphone, a local audio file, a satellite broadcast, an internet broadcast, or a terrestrial radio broadcast.
8. The method for per performing an internet search from a sound melody of claim 2 wherein said sound source is selected from the group of a microphone, a local audio file, a satellite broadcast, an internet broadcast, or a terrestrial radio broadcast.
9. The method for per performing an internet search from a sound melody of claim 3 wherein said sound source is selected from the group of a microphone, a local audio file, a satellite broadcast, an internet broadcast, or a terrestrial radio broadcast.
10. The method for per performing an internet search from a sound melody of claim 4 wherein said sound source is selected from the group of a microphone, a local audio file, a satellite broadcast, an internet broadcast, or a terrestrial radio broadcast.
11. The method for per performing an internet search from a sound melody of claim 5 wherein said sound source is selected from the group of a microphone, a local audio file, a satellite broadcast, an internet broadcast, or a terrestrial radio broadcast.
12. The method for per performing an internet search from a sound melody of claim 6 wherein said sound source is selected from the group of a microphone, a local audio file, a satellite broadcast, an internet broadcast, or a terrestrial radio broadcast.
13. A computer system, the system comprising a mobile computing device, said mobile computing device being in electronic communication with at least one of the group of a microphone or an external audio source, and said mobile computing device being in electronic communication with the internet; said mobile computing device being configured for:
(a) obtaining a sound clip from a sound source;
(b) generating a search query based on at least one of the group of performing voice-to-text analysis on said sound clip or identifying the salient characteristics of a melody contained within said sound clip;
(c) searching the internet using said search query; and
(d) returning, as search results including information relating to any of the group of the title of a song contained in said sound clip, the titles of albums containing the song contained in said sound clip, the artists who produced the song contained in said sound clip, the lyrics of the song contained in said sound clip, background information relating to the song contained in said sound clip, recordings of the song contained in said sound clip, or media relating to the song contained in said sound clip.
14. The computer system of claim 13 wherein said search query is generated exclusively by performing voice-to-text analysis.
15. The computer system of claim 13 wherein said search query is generated by performing voice-to-text analysis in combination with identifying the salient characteristics of said melody.
16. The computer system of claim 13 wherein said search query is generated by performing voice-to-text analysis in combination with identifying the salient characteristics of said melody.
17. A computer program product, said computer program product being embodied in a tangible non-transitory computer readable storage medium and comprising computer instructions for:
(a) obtaining a sound clip from a sound source;
(b) generating a search query based on at least one of the group of performing voice-to-text analysis on said sound clip or identifying the salient characteristics of a melody contained within said sound clip;
(c) searching the internet using said search query; and
(d) returning, as search results including information relating to any of the group of the title of a song contained in said sound clip, the titles of albums containing the song contained in said sound clip, the artists who produced the song contained in said sound clip, the lyrics of the song contained in said sound clip, background information relating to the song contained in said sound clip, recordings of the song contained in said sound clip, or media relating to the song contained in said sound clip.
18. The computer program product of claim 17 wherein said search query is generated exclusively by performing voice-to-text analysis.
19. The computer program product of claim 17 wherein said search query is generated by performing voice-to-text analysis in combination with identifying the salient characteristics of said melody.
20. The computer system of claim 17 wherein said search query is generated by performing voice-to-text analysis in combination with identifying the salient characteristics of said melody.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/611,609 US20150220629A1 (en) | 2014-01-31 | 2015-02-02 | Sound Melody as Web Search Query |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201461933903P | 2014-01-31 | 2014-01-31 | |
US14/611,609 US20150220629A1 (en) | 2014-01-31 | 2015-02-02 | Sound Melody as Web Search Query |
Publications (1)
Publication Number | Publication Date |
---|---|
US20150220629A1 true US20150220629A1 (en) | 2015-08-06 |
Family
ID=53755029
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/611,609 Abandoned US20150220629A1 (en) | 2014-01-31 | 2015-02-02 | Sound Melody as Web Search Query |
Country Status (1)
Country | Link |
---|---|
US (1) | US20150220629A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017028116A1 (en) * | 2015-08-16 | 2017-02-23 | 胡丹丽 | Intelligent desktop speaker and method for controlling intelligent desktop speaker |
CN107341158A (en) * | 2016-05-03 | 2017-11-10 | 斑马网络技术有限公司 | Information acquisition method, device and the equipment of multimedia file |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020038597A1 (en) * | 2000-09-29 | 2002-04-04 | Jyri Huopaniemi | Method and a system for recognizing a melody |
US20030040904A1 (en) * | 2001-08-27 | 2003-02-27 | Nec Research Institute, Inc. | Extracting classifying data in music from an audio bitstream |
US8140329B2 (en) * | 2003-07-28 | 2012-03-20 | Sony Corporation | Method and apparatus for automatically recognizing audio data |
US20130238336A1 (en) * | 2012-03-08 | 2013-09-12 | Google Inc. | Recognizing speech in multiple languages |
-
2015
- 2015-02-02 US US14/611,609 patent/US20150220629A1/en not_active Abandoned
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020038597A1 (en) * | 2000-09-29 | 2002-04-04 | Jyri Huopaniemi | Method and a system for recognizing a melody |
US20030040904A1 (en) * | 2001-08-27 | 2003-02-27 | Nec Research Institute, Inc. | Extracting classifying data in music from an audio bitstream |
US8140329B2 (en) * | 2003-07-28 | 2012-03-20 | Sony Corporation | Method and apparatus for automatically recognizing audio data |
US20130238336A1 (en) * | 2012-03-08 | 2013-09-12 | Google Inc. | Recognizing speech in multiple languages |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017028116A1 (en) * | 2015-08-16 | 2017-02-23 | 胡丹丽 | Intelligent desktop speaker and method for controlling intelligent desktop speaker |
CN107341158A (en) * | 2016-05-03 | 2017-11-10 | 斑马网络技术有限公司 | Information acquisition method, device and the equipment of multimedia file |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106960051B (en) | Audio playing method and device based on electronic book and terminal equipment | |
CA2899657C (en) | Method and device for audio recognition | |
US20130007043A1 (en) | Voice description of time-based media for indexing and searching | |
US20150169747A1 (en) | Systems and methods for automatically suggesting media accompaniments based on identified media content | |
CN107591149A (en) | Audio synthetic method, device and storage medium | |
WO2013003854A2 (en) | Method and system for communicating between a sender and a recipient via a personalized message including an audio clip extracted from a pre-existing recording | |
US9558272B2 (en) | Method of and a system for matching audio tracks using chromaprints with a fast candidate selection routine | |
US20160336026A1 (en) | Deep tagging background noises | |
US9224385B1 (en) | Unified recognition of speech and music | |
US20140114656A1 (en) | Electronic device capable of generating tag file for media file based on speaker recognition | |
WO2016112841A1 (en) | Information processing method and client, and computer storage medium | |
JP2011048729A (en) | Apparatus, method and program for query extraction | |
EP3839952A1 (en) | Masking systems and methods | |
US11785276B2 (en) | Event source content and remote content synchronization | |
US20150220629A1 (en) | Sound Melody as Web Search Query | |
WO2016024171A1 (en) | Method of and a system for indexing audio tracks using chromaprints | |
Kurth et al. | Syncplayer-An Advanced System for Multimodal Music Access. | |
KR20080107143A (en) | System and method for recommendation of music and moving video based on audio signal processing | |
Raimond et al. | Automated semantic tagging of speech audio | |
JP5474591B2 (en) | Image selection apparatus, image selection method, and image selection program | |
US11023520B1 (en) | Background audio identification for query disambiguation | |
Simou et al. | Towards blind quality assessment of concert audio recordings using deep neural networks | |
Phang et al. | Tools and technologies for enhancing access to audiovisual archives: The Singapore journey | |
Brinkman et al. | Online music recognition: the Echoprint system | |
JP5713775B2 (en) | Music search device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |