US20120213383A1 - Automatic resource retrieval and use - Google Patents

Automatic resource retrieval and use Download PDF

Info

Publication number
US20120213383A1
US20120213383A1 US13/459,497 US201213459497A US2012213383A1 US 20120213383 A1 US20120213383 A1 US 20120213383A1 US 201213459497 A US201213459497 A US 201213459497A US 2012213383 A1 US2012213383 A1 US 2012213383A1
Authority
US
United States
Prior art keywords
content
commercial location
resource
method
audio signal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/459,497
Inventor
Hans M. Glitsch
Ajay Shah
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
IVIU TECHNOLOGIES LLC
Original Assignee
IVIU TECHNOLOGIES LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to US18147209P priority Critical
Priority to US12/789,419 priority patent/US8805723B2/en
Application filed by IVIU TECHNOLOGIES LLC filed Critical IVIU TECHNOLOGIES LLC
Priority to US13/459,497 priority patent/US20120213383A1/en
Assigned to ARSH TECHNOLOGIES, LLC reassignment ARSH TECHNOLOGIES, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GLITSCH, HANS M., SHAH, AJAY
Assigned to INTERACTION CAPITAL GROUP, LLC, SHAH, AJAY, SHIOZAKI, RANDOLPH reassignment INTERACTION CAPITAL GROUP, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ARSH TECHNOLOGIES, LLC
Assigned to IVIU TECHNOLOGIES, LLC reassignment IVIU TECHNOLOGIES, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: INTERACTION CAPITAL GROUP, LLC, SHAH, AJAY, SHIOZAKI, RANDOLPH
Publication of US20120213383A1 publication Critical patent/US20120213383A1/en
Application status is Abandoned legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/24Monitoring of processes or resources, e.g. monitoring of server load, available bandwidth, upstream requests
    • H04N21/2407Monitoring of transmitted content, e.g. distribution time, number of downloads
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying
    • G06F16/432Query formulation
    • G06F16/433Query formulation using audio data
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying
    • G06F16/438Presentation of query results
    • G06F16/4387Presentation of query results by the use of playlists
    • G06F16/4393Multimedia presentations, e.g. slide shows, multimedia albums
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network, synchronizing decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • H04N21/4394Processing of audio elementary streams involving operations for analysing the audio stream, e.g. detecting features or characteristics in audio streams

Abstract

At a commercial location, broadcasting a content acquisition cue via an audio signal. The audio signal has a frequency that renders the signal inaudible to a human; corresponds to the commercial location; and bears information that causes the device to obtain content relevant to the commercial location.

Description

    RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 12/789,419, filed on May 27, 2010, which claims the benefit of U.S. Provisional Pat. App. 61/181,472 filed May 27, 2009, each of which the entire content is hereby incorporated by reference.
  • TECHNICAL FIELD
  • This disclosure relates automatic resource retrieval and use. More specifically, this disclosure relates to generating, providing, and processing cues for resource retrieval and use.
  • BACKGROUND
  • Computers and other computing devices are useful for retrieving and displaying resources such as multimedia content, interactive content, executable programs, and the like. Such resources are often useful in providing information to a user of the computer. Often, that content can augment or enhance other information, performances, or experiences that the user is presented with from a source other than his computer.
  • SUMMARY
  • In general, in one aspect: At a commercial location, broadcasting a content acquisition cue via an audio signal. The audio signal has a frequency that renders the signal inaudible to a human; corresponds to the commercial location; and bears information that causes the device to obtain content relevant to the commercial location.
  • In general, in another aspect, at a commercial location, receiving a content acquisition cue via an audio signal. The audio signal has a frequency that renders the signal inaudible to a human and corresponds to the commercial location. In response to receiving the content acquisition cue, obtaining content relevant to the commercial location.
  • Implementations may have one or more of the following features. Content is also provided to the device. The content includes a coupon for an item sold at the commercial location. The content includes an advertisement for an item sold at the commercial location. The content acquisition cue is broadcast over a plurality of substantially separate audio channels. The content acquisition cue is broadcast over three separate audio channels. The commercial location is at an entrance of a store. The audio signal further includes a content use cue that causes the content to be used on the device.
  • Other aspects include other combinations of the features recited above and other features, expressed as methods, apparatus, systems, program products, and in other ways. Other features and advantages will be apparent from the description and from the claims.
  • DESCRIPTION OF DRAWINGS
  • Embodiments of the invention described herein may be understood by reference to the following figures, which are provided by way of example and not of limitation:
  • FIG. 1 is schematic depiction of a resource retrieval system.
  • FIGS. 2-3 are flowcharts for a resource retrieval method.
  • FIG. 4 is a schematic depiction of an encoding process.
  • FIG. 5 is a flowchart for encoding a resource identifier
  • FIGS. 6A and 6B are schematic depictions of broadcasting an audio signal.
  • FIG. 7 is a flowchart for decoding a resource identifier.
  • FIG. 8 is a schematic depiction of a venue with several content use sites.
  • DETAILED DESCRIPTION
  • Computing devices (including personal computers, laptops, netbooks, smart phones, special-purpose computing hardware, and other portable or non-portable equipment, collectively referred to in this document as “computers”) are useful for retrieving information. For example, computers often can electronically communicate (e.g., via the Internet) with a vast number of information sources, and can therefore retrieve a vast amount of information.
  • Moreover, computers are capable of displaying a vast amount of information in a variety of forms. Even computers with relatively modest capabilities can routinely display a variety of video, audio, document, and other data formats, and can similarly execute a wide variety of computer programs.
  • It is often useful or entertaining to employ these capabilities of a computer to enhance events that are occurring in the computer user's environment. For example, if the computer user is attending a presentation on a topic, it may enhance the user's experience of the presentation if the user was simultaneously viewing pertinent content, e.g., on his computer.
  • To some degree, this may be accomplished by traditional means. For example, the speaker at a presentation can display a visual aid to the audience. However, this does not allow audience members to individually interact with the visual aid. Alternatively, the speaker may disseminate an electronic copy of the visual aid to conference attendants in advance of the presentation. Although this allows audience members to individually interact with the visual aid, it is relatively difficult to focus the audience's collective attention on a desired feature of the visual aid. For example, if the speaker desired audience members to activate a function on a program by clicking on one of many available buttons, the speaker's options would be relatively inefficient. For example, one option is for the speaker to attempt to verbally describe which of the several buttons to click.
  • Among other things, the techniques described in this document allow for the automatic retrieval and/or use of content or other resources. These techniques may advantageously be employed, e.g., by a speaker at a conference to cause the participants' computers to automatically retrieve, load, or use content or other resources. Other exemplary uses are discusses below.
  • FIG. 1 is a schematic depiction of a resource retrieval system. The resource retrieval system 10 may include one or more client computers 12 and a signaling computer 14. The signaling computer 14 may include a loudspeaker 16, and each of the client computers 12 may include a microphone 20. Each client computer 12 may also be in data communication via a computer network 22 to one or more resource identification servers 24 and resource providers 28. The computer network 22 can include any known networking type or structure, including local area networks, ad-hoc networks, mesh networks, wide-area networks such as the Internet, telephone networks, satellite networks, etc. The data communication may be implemented in any known fashion, including hypertext transfer protocol (HTTP), file transfer protocol (FTP), BitTorrent protocol, e-mail, short message service (SMS), or any other standard or proprietary fashion.
  • The client computer 12 may more generally include any computing device that includes a microphone, networking communications capability, and processing capability adequate for the signal processing described herein. By way of example and not limitation, a client computer 12 as described herein may include a laptop computer, a notebook computer, a netbook computer, a cellular phone (including smart phones such as an iPhone or a Blackberry), a personal digital device, a media player (e.g., an iPod), an electronic messaging device, and so forth.
  • As described more thoroughly below, the presenter may emit suitable audio signals 18 using the loudspeaker 16, which can subsequently be detected by the microphone 20 of a client computer 12. The audio signal 18 can be decoded to reveal a resource identifier. As described more fully below, the resource identifier can consist of a relatively small amount of data; e.g., a 30 bit integer.
  • The resource identifier can be used to determine a corresponding resource location. For example, this can be accomplished by submitting the resource identifier over the computer network 22 to a resource identification server 24. A server 24 can include a data store 26 that lists a resource location corresponding to each resource identifier. The client computer 12 can then obtain the resource from a resource provider 28 at the specified resource location.
  • As used in this document, the word “resource” can refer to any computer-stored information or instructions that can be transferred, described, or implemented over a computer network. In particular, a “resource” can include audio, video, images, documents, markup language documents, hypertext links, scripts, macros, source code, executable code, or any combinations of the foregoing. The words “content” and “resource” are synonymous in this document.
  • FIG. 2 is a method of associating an audio signal with a resource. The method 30 begins by selecting a resource (step 32) with which to associate an audio signal 18. For example, this resource can include a visual aid for use by a speaker during a presentation. Many more examples are given below.
  • In step 34, a resource identifier can be associated with the resource. In some implementations, the resource identifier can consist solely of a number; for example, a 30 (or fewer) bit number. Additionally, a resource identifier can also include other information, such as information that describes a context or a geographic location in which the resource will be used. Further considerations pertinent to selecting a resource identifier are discussed in connection with FIGS. 6A and 6B.
  • Step 34 can be carried out at any location, on any computer. For example, step 34 can be carried out at the signaling computer 14, a resource identification server 24, a resource provider 28, or at another location outside the resource retrieval system 10.
  • In step 36, the resource is sent to known location at a resource provider 28. In some implementations, this can be done via any electronic communication, such as file transfer protocol, secure copying, via e-mail attachment, etc. In some implementations, the performer of the method 30 is also the resource provider 28, in which case the resource need not be sent to another location.
  • In step 37, the resource identifier and the location of the corresponding resource are stored in a data store 26 within a server 24. This information can be stored in any manner. For example, the resource identifier and resource location can be stored in corresponding fields in a record of a database, as corresponding entries in an index or table, as concurrent entries in a list, etc.
  • In step 38, the resource identifier can be encoded as an audio signal 18. Further details about possible encoding techniques are described below in connection with FIGS. 4 and 5. The encoding can take place at any location on any computer. For example, step 38 can be carried out at the signaling computer 14, a server 24, a resource provider 28, or at another location outside the resource retrieval system 10.
  • In step 40, the audio signal 18 can be broadcasted. In some implementations, the audio signal 18 is broadcasted by the signaling computer 14 using the loudspeaker 16, or other hardware capable of broadcasting an audio signal. In some implementations, a computer different from the signaling computer 14 broadcasts the audio signal.
  • FIG. 3 is a method for retrieving a resource. The method 42 of FIG. 3 can begin with receiving an audio signal (step 44). For example, an audio signal 18 that had previously been broadcasted using the method 30 or another method can be received.
  • In step 46, the audio signal can be decoded. Further details about decoding techniques are described below, in connection with FIG. 7. As a result of the decoding in step 46, a resource identifier can be obtained (step 48).
  • In step 49, a resource location can be ascertained based on the resource identifier. For example, the resource identifier can be submitted to a server 24, in response to which the server 24 provides the corresponding resource location. In some implementations, the resource identifier can be used as a query for a database, as an argument for a lookup routine, etc.
  • In step 50, the resource can be obtained from the resource location. In some implementations, the resource can be obtained directly by the performer of the retrieval method 42, such as a client computer 12. For example, the client computer 12 can initiate a data connection to the resource location and obtain the resource. In some implementations, the performer of the retrieval method 42 can indirectly obtain the resource. For example, the resource provider can send the resource as an attachment in an e-mail to an e-mail server, from which the performer of the method 42 subsequently obtains the resource.
  • After obtaining the resource, the client computer 12 may take a specified action with respect to the resource (e.g., display or execute the resource), or may simply store the resource for later use. Although the resource provider 28 is illustrated separately from the client computer 12 in FIG. 1, in some implementations the resource provider 28 can be the same as the client computer 12. In this case, the resource is “local” to the client computer 12.
  • Referring to FIGS. 4 and 5, a schematic illustration of encoding a resource identifier 52 is shown (FIG. 4), together with a method 58 of encoding a resource identifier (FIG. 5). In the schematic illustration, a resource identifier 52 is illustrated as a string of bits. Although only nine bits are illustrated in FIG. 4, in principal any number of bits can be used, subject to limitations discussed below. In step 60, the resource identifier 52 can be obtained. For example, this can be the result of step 34 of process 30 (FIG. 2).
  • In step 62, a number of audio channels 54 are identified. Although three audio channels 54 a, 54 b, and 54 c are illustrated in FIG. 4, in principal any number of audio channels can be used. An audio channel is characterized by a central frequency 55 and a channel width (or “band width”) that describes the range of frequencies in the channel. The channel width is illustrated in FIG. 4 by the geometric width of the channels 54 a, 54 b, 54 c.
  • If the audio channels 54 a, 54 b, 54 c do not overlap (i.e., are separate), then there is no interference between the audio signals sent over the audio channels. However, if the audio channels 54 a, 54 b, 54 c do overlap, then the audio signals may, but need not, interfere. In some implementations, the audio channels 54 a, 54 b, and 54 c are separate. In some implementations, one or more pairs of audio channels 54 a, 54 b, and 54 c are only substantially separate, with a small amount of overlap permissible based on an acceptable level of interference-based errors. The rate at which these errors occurs depends on many factors, including the size of the overlap between channels, the type of encoding used, the data being sent during a particular transmission, and other factors.
  • As discussed below, some encoding techniques require a minimum bandwidth for a communication channel 54. In some implementations, the number of audio channels 54 can be chosen by identifying a total frequency range available for transmission, and partitioning the available frequency range into separate (or substantially separate) channels of the minimum-sized bandwidth necessary for the selected encoding technique. This approach maximizes the number of audio channels 54 available for transmission.
  • In principle, there is no upper limit for the total frequency range available for transmission. However, the sampling rate of the microphone 20 can determine the maximum possible frequency it can receive. In particular, the maximum possible identifiable frequency is equal to half the sampling rate. In some implementations, commonly available microphones have a sampling rate of 44.1 kilo-samples per second. Consequently, the maximum available frequency such a microphone can detect is 22.05 kHz.
  • In principle, there is no lower limit for the total frequency range available for transmission. However, in some implementations, it is desirable for the audio signals 18 to be in inaudible to humans or other animals. This has the advantage that people would not be distracted by audio signals 18 that are broadcast in their vicinity. According to one estimate, the audible range for most humans has an upper limit of 15 kHz. This, in turn, can be used as a lower limit of the total available frequency range.
  • Once the number of channels 54 is identified in step 62, the resource identifier 52 can be segmented into as many parts as there are channels (step 64). For example, in FIG. 4, the nine-bit resource identifier 52 is segmented into three parts, one for each of the communication channels 54 a, 54 b, 54 c. If the number of bits in the resource identifier 52 is not evenly divisible by the number of communication channels 54, then the remainder bits can be associated with the higher-frequency channels. This can be advantageous for reasons of echo prevention, described more fully below (see FIGS. 6A and 6B).
  • After segmenting the resource identifier 52, each segment can be processed as follows. A first segment is identified (step 64). That segment is encoded according to a chosen algorithm (step 66). There are a number of modulation algorithms that can be employed, including analog modulation, digital modulation, spread spectrum modulation, and others. Each algorithm has potential strengths and weaknesses, depending on the particular constraints that are important in the deployment of a particular resource retrieval system 10.
  • Among the possible encoding techniques is the class of continuous spectrum, frequency shift keying algorithms. One such algorithm, known as Gaussian minimum shift keying (“GMSK”), has the advantageous property of requiring a relatively narrow bandwidth, e.g., approximately 1.5 kHz) for each of three audio channels to transmit about ten bits (or thirty bits in aggregate) within two to five milliseconds (without any cyclic redundancy check). Consequently, using GMSK for encoding accommodates a relatively high number of audio channels 54 that can be used for transmission. It will be understood that other modulation techniques may be suitably adapted to encode data for high-frequency audio transmission. Thus, while GMSK has been usefully demonstrated to provide adequate data rates and robustness in a variety of physical contexts, this disclosure is not limited to GMSK coding unless explicitly stated to the contrary. For example, other shift-keyed encoding techniques include minimum shift keying, frequency shift keying, phase shift keying, quadrature phase shift keying, and so forth. Still more generally, other modulation techniques based upon frequency modulation, amplitude modulation, phase modulation, and various combinations of the foregoing, are generally known in the art and all such techniques that may be usefully adapted to reliably encode data onto a high-frequency audio carrier are intended to fall within the scope of this disclosure.
  • Referring to FIGS. 6A and 6B, concurrently broadcasting several audio signals 18 a, 18 b, 18 c helps reduce the possibility of errors related to echoes as compared to broadcasting a single audio signal 18. In FIG. 6A, a single audio signal 18 is broadcast from a loudspeaker 16; In FIG. 6B, three audio signals 18 a, 18 b, and 18 c are broadcast from a loudspeaker 16. If the signal 18 carries the same information content as the signals 18 a, 18 b, and 18 c, then the signal 18 is three times as long as any individual signal 18 a, 18 b, or 18 c.
  • An echo-related error may occur when a signal interferes with itself where a microphone 20 is located. In FIG. 6A, for example, the signal 18 reflects off a wall 67, so that both the last portion and the (reflected) initial portion of the signal arrive at the microphone. By contrast, in FIG. 6B when the loudspeaker 16 and microphone 20 are in the same relative positions, there are no echo-related errors, because the signal lengths are sufficiently short.
  • The higher the frequency of an audio signal, the shorter its duration needs to be in order to send a certain fixed amount of data. Thus, allocating any extra bits of the resource identifier to higher-frequency channels helps reduce echo-related errors by keeping signals relatively short.
  • Using a larger resource identifier has the advantage of being able to accommodate more distinct resources, but has the disadvantage of requiring longer audio signals 18, thereby increasing the risk of echo-related error, as described above. In some implementations, the resource identifier 52 consists of an integer with a length between twenty four and thirty bits, which are segmented for transmission across three audio channels and encoded using GMSK modulation. In some implementations, the resource identifier 52 can also include other data, e.g., data describing a geographic or other context to which the resource relates.
  • Referring back to FIG. 5, if there are any unencoded segments remaining (decision 68), then the encoding process 58 returns back to step 64, where the next unencoded segment is selected. Once all segments have been encoded, process 58 concludes, having produced a number of audio signals that collectively encode the resource identifier.
  • Although the use of several separate (or substantially separate) audio channels to broadcast several audio signals is advantageous for reasons described herein, these techniques can still be successful with a single audio signal broadcast on a single audio channel.
  • FIG. 7 is a flowchart for a decoding process. Initially, an audio signal can be received (step 72). For example, the audio signal can be the result of broadcasting step 40 of method 30. The received audio signal can then be decoded (step 74). The decoding process utilizes the compatible techniques of those used to encode the audio signal. For example, if GMSK techniques were used to encode the audio signal, those techniques are used to decode the signal.
  • Upon decoding the signal, the step 74 produces one or more resource identifier segments. If more than one segment is produced, those segments are subsequently reassembled (step 76), thereby producing a resource identification 52.
  • FIG. 8 is a schematic depiction of a venue with several content display points. The venue 80 includes several content display points 82, each with a corresponding effective display area 84. For example, there may be a signaling computer 14 at each display point 82, and each effective display area 84 corresponds to the range of the signaling computer's loudspeaker 16. Other examples are possible.
  • As a person navigates the venue 80, they encounter the various content display points 82. Each content display point 82 provides a content use cue to cause a user's computer to display content relevant to that particular content display point. For example, the content use cue can include an audio signal 18 as described above. More generally, the content use cue can include any transmission that causes the user's computer to display specific content. For example, the content use cue can include a visual signal, a conventional data transmission over a network connection, etc.
  • In the foregoing scenario, it is sometimes convenient for the user to have obtained the pertinent content for the various content display points 82 in advance of encountering those display points. To facilitate this, one or more content acquisition sites 86 can be established in the venue 80. For example, a content acquisition site can be established near an entrance 88 of the venue 80.
  • A content acquisition site 86 can provide a content acquisition cue to a user. For example, the content acquisition cue can include an audio signal 18 as describe above. More generally, the content acquisition cue can include any transmission that causes the user's computer to acquire pertinent content. For example, the content acquisition cue can include a visual signal, a conventional data transmission over a network connection, etc. A content acquisition cue can simply include a verbal instruction to load a particular piece of physical media, such as a DVD, that bears the content pertinent to the content display points 82.
  • In response to receiving the content acquisition cue, the user's computer acquires the content pertinent to the various content display sites. For example, the user's computer can perform the method 42 of FIG. 3. The user's computer can obtain the content in other ways, such as acquiring the content via a conventional network connection to a local or remote computer, or by loading files from a local storage device.
  • Acquiring content prior to use can have advantages. For example, having content locally available when receiving a content use cue tends to allow the user's computer to use the content relatively quickly. Moreover, obtaining all the content for the various content use points 82 in the venue 80 in one network transmission can provide power or bandwidth savings over obtaining content from the different content use points 82 on an as-needed basis.
  • In addition to the example discussed above involving a presenter and audience members at a conference utilizing a resource retrieval system 10 to retrieve visual aides, interactive surveys, and the like, numerous other examples exist.
  • In one example, the resource retrieval system 10 can be deployed at a retail venue. At various locations near specified products, an audio signal source can be deployed to cause computers that detect the audio signal to retrieve content pertinent to one or more products in the vicinity. Such content can include, e.g., advertisement materials, promotions, coupons, product descriptions, product reviews, links to product-related web sites, etc. Additionally, a store-wide audio signal can be broadcast periodically alerting shoppers to promotions or specials occurring in specified departments or locations in the retail venue.
  • In another example, the resource retrieval system 10 can be deployed in a museum or other venue amenable to self-guided tours. Near points of interest (e.g., exhibits, historical sites, etc.), the resource retrieval system 10 can be used to cause users' computers to automatically retrieve information about the point of interest. Moreover, different information can be displayed to different users. For example, an “advanced” presentation or a “beginner” presentation can be displayed, depending on the user's preferences.
  • In another example, the resource retrieval system 10 can be deployed in a sporting event. At various times, the resource retrieval system 10 can be used to cause users' computers to automatically retrieve information pertinent to a particular competitor, a team, a play, etc.
  • In another example, the resource retrieval system 10 can be deployed in a classroom setting. The teacher can broadcast audio signals to cause students' computers to display specific resources pertinent to a lecture. Similarly, the teacher can broadcast audio signals to cause students' computers to begin or end an examination.
  • In another example, the resource retrieval system 10 can be deployed in any venue. At various locations, the resource retrieval system 10 can be used to cause users' computers to automatically retrieve information about the venue such as a map of the venue or surrounding areas. For example a loudspeaker can be deployed at information kiosks, orientation booths, entrances to the venue, etc.
  • In another example, at any venue, the resource retrieval system can be used to cause users' computers to automatically retrieve information about emergency procedures. For example, in the event of an emergency, a venue-wide broadcast can be sounded over the venue's alarm system to alert occupants about the emergency, where the venue-wide broadcast includes an audio signal according to the techniques described above. Emergency procedures can include evacuation routes, first aid information, safety information, etc.
  • Many of the above systems, devices, methods, processes, and the like may be realized in hardware, software, or any combination of these suitable for the control, data acquisition, and data processing described herein. This includes realization in one or more microprocessors, microcontrollers, embedded microcontrollers, programmable digital signal processors or other programmable devices or processing circuitry, along with internal and/or external memory. This may also, or instead, include one or more application specific integrated circuits, programmable gate arrays, programmable array logic components, or any other device or devices that may be configured to process electronic signals. It will further be appreciated that a realization of the processes or devices described above may include computer-executable code created using a structured programming language such as C, an object oriented programming language such as C++, or any other high-level or low-level programming language (including assembly languages, hardware description languages, and database programming languages and technologies) that may be stored, compiled or interpreted to run on one of the above devices, as well as heterogeneous combinations of processors, processor architectures, or combinations of different hardware and software. At the same time, processing may be distributed across devices such as the various systems described above, or all of the functionality may be integrated into a dedicated, standalone device. All such permutations and combinations are intended to fall within the scope of the present disclosure.
  • In embodiments, disclosed herein are computer program products comprising computer-executable code or computer-usable code that, when executing on one or more computing devices (such as the devices/systems described above), performs any and/or all of the steps described above. The code may be stored in a non-transitory fashion in a computer memory, which may be a memory from which the program executes (such as random access memory associated with a processor), or a storage device such as a disk drive, flash memory or any other optical, electromagnetic, magnetic, infrared or other device or combination of devices. In another aspect, any of the processes described above may be embodied in any suitable transmission or propagation medium carrying the computer-executable code described above and/or any inputs or outputs from same.
  • It will be appreciated that the methods and systems described above are set forth by way of example and not of limitation. Numerous variations, additions, omissions, and other modifications will be apparent to one of ordinary skill in the art. In addition, the order or presentation of method steps in the description and drawings above is not intended to require this order of performing the recited steps unless a particular order is expressly required or otherwise clear from the context.
  • While particular embodiments of the present invention have been shown and described, it will be apparent to those skilled in the art that various changes and modifications in form and details may be made therein without departing from the spirit and scope of the invention as defined by the following claims. The claims that follow are intended to include all such variations and modifications that might fall within their scope, and should be interpreted in the broadest sense allowable by law.

Claims (19)

1. A method comprising:
At a commercial location, broadcasting a content acquisition cue via an audio signal, in which the audio signal:
has a frequency that renders the signal inaudible to a human;
corresponds to the commercial location; and
bears information that causes the device to obtain content relevant to the commercial location.
2. The method of claim 1, further comprising providing the content to the device.
3. The method of claim 1, in which the content includes a coupon for an item sold at the commercial location.
4. The method of claim 1, in which the content includes an advertisement for an item sold at the commercial location.
5. The method of claim 1, in which the content acquisition cue is broadcast over a plurality of substantially separate audio channels.
6. The method of claim 5, in which the content acquisition cue is broadcast over three separate audio channels.
7. The method of claim 1, in which the commercial location is at an entrance of a store.
8. The method of claim 1, in which the audio signal further includes a content use cue that causes the content to be used on the device.
9. A method comprising:
at a commercial location, receiving a content acquisition cue via an audio signal, in which the audio signal:
has a frequency that renders the signal inaudible to a human; and
corresponds to the commercial location; and
in response to receiving the content acquisition cue, obtaining content relevant to the commercial location.
10. The method of claim 7, wherein the content includes a coupon for an item sold at the commercial location.
11. The method of claim 7, wherein the content includes an advertisement for an item sold at the commercial location.
12. The method of claim 7, wherein receiving the audio signal includes receiving audio signals simultaneously broadcast over three substantially separate audio channels.
13. A system comprising:
a controller, and
a loudspeaker in data communication with a controller,
wherein the controller is configured to cause the loudspeaker to broadcast a content acquisition cue that:
has a frequency that renders the signal inaudible to a human;
corresponds to the commercial location; and
bears information that causes a device to automatically obtain content relevant to the commercial location.
14. The system of claim 13 wherein the content includes a coupon for an item sold at the commercial location.
15. The system of claim 13 wherein the content includes an advertisement for an item sold at the commercial location.
16. The system of claim 13, in which the content acquisition cue is broadcast over a plurality of substantially separate audio channels.
17. The system of claim 16, in which the content acquisition cue is broadcast over three substantially separate audio channels.
18. The system of claim 13, in which the commercial location is at an entrance of a store.
19. The system of claim 13, in which the controller is further configured to cause the loudspeaker to broadcast a content use cue that causes the content to be used on the device.
US13/459,497 2009-05-27 2012-04-30 Automatic resource retrieval and use Abandoned US20120213383A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US18147209P true 2009-05-27 2009-05-27
US12/789,419 US8805723B2 (en) 2009-05-27 2010-05-27 Acoustically transmitting a resource identifier in multiple concurrent segments
US13/459,497 US20120213383A1 (en) 2009-05-27 2012-04-30 Automatic resource retrieval and use

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13/459,497 US20120213383A1 (en) 2009-05-27 2012-04-30 Automatic resource retrieval and use

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US12/789,419 Continuation US8805723B2 (en) 2009-05-27 2010-05-27 Acoustically transmitting a resource identifier in multiple concurrent segments

Publications (1)

Publication Number Publication Date
US20120213383A1 true US20120213383A1 (en) 2012-08-23

Family

ID=43221121

Family Applications (8)

Application Number Title Priority Date Filing Date
US12/789,377 Expired - Fee Related US8718805B2 (en) 2009-05-27 2010-05-27 Audio-based synchronization to media
US12/789,419 Expired - Fee Related US8805723B2 (en) 2009-05-27 2010-05-27 Acoustically transmitting a resource identifier in multiple concurrent segments
US13/096,920 Abandoned US20110202156A1 (en) 2009-05-27 2011-04-28 Device with audio-based media synchronization
US13/096,943 Abandoned US20110208333A1 (en) 2009-05-27 2011-04-28 Pre-processing media for audio-based synchronization
US13/096,965 Abandoned US20110202687A1 (en) 2009-05-27 2011-04-28 Synchronizing audience feedback from live and time-shifted broadcast views
US13/096,977 Expired - Fee Related US8789084B2 (en) 2009-05-27 2011-04-28 Identifying commercial breaks in broadcast media
US13/096,950 Abandoned US20110208334A1 (en) 2009-05-27 2011-04-28 Audio-based synchronization server
US13/459,497 Abandoned US20120213383A1 (en) 2009-05-27 2012-04-30 Automatic resource retrieval and use

Family Applications Before (7)

Application Number Title Priority Date Filing Date
US12/789,377 Expired - Fee Related US8718805B2 (en) 2009-05-27 2010-05-27 Audio-based synchronization to media
US12/789,419 Expired - Fee Related US8805723B2 (en) 2009-05-27 2010-05-27 Acoustically transmitting a resource identifier in multiple concurrent segments
US13/096,920 Abandoned US20110202156A1 (en) 2009-05-27 2011-04-28 Device with audio-based media synchronization
US13/096,943 Abandoned US20110208333A1 (en) 2009-05-27 2011-04-28 Pre-processing media for audio-based synchronization
US13/096,965 Abandoned US20110202687A1 (en) 2009-05-27 2011-04-28 Synchronizing audience feedback from live and time-shifted broadcast views
US13/096,977 Expired - Fee Related US8789084B2 (en) 2009-05-27 2011-04-28 Identifying commercial breaks in broadcast media
US13/096,950 Abandoned US20110208334A1 (en) 2009-05-27 2011-04-28 Audio-based synchronization server

Country Status (2)

Country Link
US (8) US8718805B2 (en)
WO (2) WO2010138777A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100305730A1 (en) * 2009-05-27 2010-12-02 Glitsch Hans M Automatic resource retrieval and use

Families Citing this family (76)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7644282B2 (en) 1998-05-28 2010-01-05 Verance Corporation Pre-processed information embedding system
US6737957B1 (en) 2000-02-16 2004-05-18 Verance Corporation Remote control signaling using audio watermarks
EP2782337A3 (en) 2002-10-15 2014-11-26 Verance Corporation Media monitoring, management and information system
US20060239501A1 (en) 2005-04-26 2006-10-26 Verance Corporation Security enhancements of digital watermarks for multi-media content
US8020004B2 (en) 2005-07-01 2011-09-13 Verance Corporation Forensic marking using a common customization function
US8781967B2 (en) 2005-07-07 2014-07-15 Verance Corporation Watermarking in an encrypted domain
US8316308B2 (en) 2007-06-08 2012-11-20 Google Inc. Adaptive user interface for multi-source systems
US8321261B2 (en) 2007-12-14 2012-11-27 John Nicholas and Kristin Gross Integrated gourmet item data collection, recommender and vending system and method
US8205148B1 (en) 2008-01-11 2012-06-19 Bruce Sharpe Methods and apparatus for temporal alignment of media
US8259938B2 (en) 2008-06-24 2012-09-04 Verance Corporation Efficient and secure forensic marking in compressed
US20100205628A1 (en) 2009-02-12 2010-08-12 Davis Bruce L Media processing methods and arrangements
US8489774B2 (en) 2009-05-27 2013-07-16 Spot411 Technologies, Inc. Synchronized delivery of interactive content
US20110063503A1 (en) * 2009-07-06 2011-03-17 Brand Steven M Synchronizing secondary content to a multimedia presentation
US8463875B2 (en) * 2009-08-20 2013-06-11 Google Inc. Synchronized playback of media players
US8457118B2 (en) 2010-05-17 2013-06-04 Google Inc. Decentralized system and method for voice and video sessions
US9607131B2 (en) 2010-09-16 2017-03-28 Verance Corporation Secure and efficient content screening in a networked environment
US8392498B2 (en) * 2010-10-08 2013-03-05 Lumi Technologies Limited Distribution of content and behavior to disparate platforms
US20130331972A1 (en) * 2010-11-15 2013-12-12 Egonocast System and method for synchronizing an application with audio content
US9277292B2 (en) * 2010-12-09 2016-03-01 Cox Communications, Inc. System, method and device for providing a mobile application across smartphone platforms to enable consumer connectivity and control of media
US20120198319A1 (en) 2011-01-28 2012-08-02 Giovanni Agnoli Media-Editing Application with Video Segmentation and Caching Capabilities
US8886015B2 (en) 2011-01-28 2014-11-11 Apple Inc. Efficient media import
US8621355B2 (en) 2011-02-02 2013-12-31 Apple Inc. Automatic synchronization of media clips
US10324605B2 (en) 2011-02-16 2019-06-18 Apple Inc. Media-editing application with novel editing tools
US9997196B2 (en) 2011-02-16 2018-06-12 Apple Inc. Retiming media presentations
US8798777B2 (en) * 2011-03-08 2014-08-05 Packetvideo Corporation System and method for using a list of audio media to create a list of audiovisual media
US8965026B2 (en) * 2011-06-10 2015-02-24 Canopy Co. Method and apparatus for remote capture of audio in a handheld device
US9160837B2 (en) 2011-06-29 2015-10-13 Gracenote, Inc. Interactive streaming content apparatus, systems and methods
US8615104B2 (en) 2011-11-03 2013-12-24 Verance Corporation Watermark extraction based on tentative watermarks
US8923548B2 (en) 2011-11-03 2014-12-30 Verance Corporation Extraction of embedded watermarks from a host content using a plurality of tentative watermarks
US8682026B2 (en) 2011-11-03 2014-03-25 Verance Corporation Efficient extraction of embedded watermarks in the presence of host content distortions
US8533481B2 (en) 2011-11-03 2013-09-10 Verance Corporation Extraction of embedded watermarks from a host content based on extrapolation techniques
US20130124999A1 (en) 2011-11-14 2013-05-16 Giovanni Agnoli Reference clips in a media-editing application
US8745403B2 (en) 2011-11-23 2014-06-03 Verance Corporation Enhanced content management based on watermark extraction records
US9547753B2 (en) 2011-12-13 2017-01-17 Verance Corporation Coordinated watermarking
US9323902B2 (en) 2011-12-13 2016-04-26 Verance Corporation Conditional access using embedded watermarks
US9418669B2 (en) * 2012-05-13 2016-08-16 Harry E. Emerson, III Discovery of music artist and title for syndicated content played by radio stations
US20130325601A1 (en) * 2012-06-05 2013-12-05 Yahoo! Inc. System for providing content
US8930457B2 (en) * 2012-06-19 2015-01-06 International Business Machines Corporation Proximity initiated co-browsing sessions
US8806518B2 (en) * 2012-06-27 2014-08-12 Intel Corporation Performance analysis for combining remote audience responses
CN104584571B (en) * 2012-06-28 2018-01-09 谷歌公司 Audio-frequency fingerprint sequence is produced at set top box
US9113203B2 (en) 2012-06-28 2015-08-18 Google Inc. Generating a sequence of audio fingerprints at a set top box
US8843952B2 (en) 2012-06-28 2014-09-23 Google Inc. Determining TV program information based on analysis of audio fingerprints
US9088406B2 (en) * 2012-07-29 2015-07-21 Qualcomm Incorporated Frame sync across multiple channels
US9571606B2 (en) 2012-08-31 2017-02-14 Verance Corporation Social media viewing system
US8869222B2 (en) 2012-09-13 2014-10-21 Verance Corporation Second screen content
US20140075469A1 (en) 2012-09-13 2014-03-13 Verance Corporation Content distribution including advertisements
US8726304B2 (en) 2012-09-13 2014-05-13 Verance Corporation Time varying evaluation of multimedia content
US9661361B2 (en) 2012-09-19 2017-05-23 Google Inc. Systems and methods for live media content matching
US20140122588A1 (en) * 2012-10-31 2014-05-01 Alain Nimri Automatic Notification of Audience Boredom during Meetings and Conferences
US9318116B2 (en) 2012-12-14 2016-04-19 Disney Enterprises, Inc. Acoustic data transmission based on groups of audio receivers
US9262793B2 (en) 2013-03-14 2016-02-16 Verance Corporation Transactional video marking system
US9499128B2 (en) 2013-03-14 2016-11-22 The Crawford Group, Inc. Mobile device-enhanced user selection of specific rental vehicles for a rental vehicle reservation
US10142700B2 (en) 2013-03-15 2018-11-27 DISH Technologies L.L.C. Television service provided social networking service
US20140280752A1 (en) * 2013-03-15 2014-09-18 Time Warner Cable Enterprises Llc System and method for seamless switching between data streams
US20140336797A1 (en) * 2013-05-12 2014-11-13 Harry E. Emerson, III Audio content monitoring and identification of broadcast radio stations
US20140336799A1 (en) * 2013-05-13 2014-11-13 Harry E. Emerson, III Discovery of music artist and title via companionship between a cellular phone and a broadcast radio receiver
US9930404B2 (en) 2013-06-17 2018-03-27 Echostar Technologies L.L.C. Event-based media playback
US9251549B2 (en) 2013-07-23 2016-02-02 Verance Corporation Watermark extractor enhancements based on payload ranking
US9380383B2 (en) * 2013-09-06 2016-06-28 Gracenote, Inc. Modifying playback of content using pre-processed profile information
US9208334B2 (en) 2013-10-25 2015-12-08 Verance Corporation Content management using multiple abstraction layers
WO2015138798A1 (en) 2014-03-13 2015-09-17 Verance Corporation Interactive content acquisition using embedded codes
GB2541784B (en) 2014-05-08 2018-03-28 Mewt Ltd Synchronisation of audio and video playback
GB2525913B (en) 2014-05-08 2018-03-28 Mewt Ltd Synchronisation of audio and video playback
GB2525912B (en) 2014-05-08 2018-02-28 Mewt Ltd Synchronisation of audio and video playback
US9786291B2 (en) 2014-06-18 2017-10-10 Google Technology Holdings LLC Communicating information between devices using ultra high frequency audio
US9912973B2 (en) 2014-08-07 2018-03-06 Echostar Technologies L.L.C. Systems and methods for facilitating content discovery based on viewer ratings
US10306319B2 (en) 2014-09-09 2019-05-28 Google Llc Collaboration between a broadcaster and an audience for a broadcast
US9792424B1 (en) * 2014-09-18 2017-10-17 Amazon Technologies, Inc. Chained authentication using musical transforms
US9565474B2 (en) 2014-09-23 2017-02-07 Echostar Technologies L.L.C. Media content crowdsource
US10506295B2 (en) 2014-10-09 2019-12-10 Disney Enterprises, Inc. Systems and methods for delivering secondary content to viewers
US10432296B2 (en) 2014-12-31 2019-10-01 DISH Technologies L.L.C. Inter-residence computing resource sharing
US10218684B2 (en) * 2015-01-30 2019-02-26 Ncr Corporation Secure data transmission
US10306332B2 (en) * 2016-06-14 2019-05-28 DISH Technologies L.L.C. Use of audio signals to provide interactive content to end users via smart devices
US10129586B2 (en) * 2016-12-19 2018-11-13 Google Llc Detecting and isolating television program content from recordings of television airings
US10397287B2 (en) 2017-03-01 2019-08-27 Microsoft Technology Licensing, Llc Audio data transmission using frequency hopping
WO2019195931A1 (en) * 2018-04-09 2019-10-17 Tunevu Inc. System and method for management and delivery of secondary syndicated companion conent of discovered primary digital media presentations

Family Cites Families (107)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US734672A (en) * 1900-10-06 1903-07-28 Jacob C Counts Grain-cradle.
USRE37131E1 (en) * 1991-02-19 2001-04-10 Roy J. Mankovitz Apparatus and methods for music and lyrics broadcasting
US6505160B1 (en) * 1995-07-27 2003-01-07 Digimarc Corporation Connected audio and other media objects
US6209028B1 (en) * 1997-03-21 2001-03-27 Walker Digital, Llc System and method for supplying supplemental audio information for broadcast television programs
IL132060D0 (en) * 1997-03-31 2001-03-19 Broadband Associates Method and system for providing a presentation on a network
US7490169B1 (en) * 1997-03-31 2009-02-10 West Corporation Providing a presentation on a network having a plurality of synchronized media types
EP2267720A3 (en) * 1998-02-23 2011-03-09 Kabushiki Kaisha Toshiba Information storage medium, information playback method and apparatus and information recording method
US20060041639A1 (en) * 1999-04-21 2006-02-23 Interactual Technologies, Inc. Platform detection
US6674993B1 (en) * 1999-04-30 2004-01-06 Microvision, Inc. Method and system for identifying data locations associated with real world observations
US7562392B1 (en) * 1999-05-19 2009-07-14 Digimarc Corporation Methods of interacting with audio and ambient music
GB9917985D0 (en) * 1999-07-30 1999-09-29 Scient Generics Ltd Acoustic communication system
US6738803B1 (en) * 1999-09-03 2004-05-18 Cisco Technology, Inc. Proxy browser providing voice enabled web application audio control for telephony devices
US7423983B1 (en) * 1999-09-20 2008-09-09 Broadcom Corporation Voice and data exchange over a packet based network
US7194752B1 (en) * 1999-10-19 2007-03-20 Iceberg Industries, Llc Method and apparatus for automatically recognizing input audio and/or video streams
US6834308B1 (en) 2000-02-17 2004-12-21 Audible Magic Corporation Method and apparatus for identifying media content presented on a media playing device
US6453252B1 (en) * 2000-05-15 2002-09-17 Creative Technology Ltd. Process for identifying audio content
WO2002008948A2 (en) 2000-07-24 2002-01-31 Vivcom, Inc. System and method for indexing, searching, identifying, and editing portions of electronic multimedia files
US7853664B1 (en) 2000-07-31 2010-12-14 Landmark Digital Services Llc Method and system for purchasing pre-recorded music
US6990453B2 (en) * 2000-07-31 2006-01-24 Landmark Digital Services Llc System and methods for recognizing sound and music signals in high noise and distortion
US7346472B1 (en) * 2000-09-07 2008-03-18 Blue Spike, Inc. Method and device for monitoring and analyzing signals
US7103906B1 (en) 2000-09-29 2006-09-05 International Business Machines Corporation User controlled multi-device media-on-demand system
US6947966B1 (en) * 2000-10-13 2005-09-20 Road Runner Holdco Llc System and method for influencing dynamic community shared elements of audio, video, and text programming via a polling system
US7277766B1 (en) * 2000-10-24 2007-10-02 Moodlogic, Inc. Method and system for analyzing digital audio files
AU2211102A (en) * 2000-11-30 2002-06-11 Scient Generics Ltd Acoustic communication system
US20020072982A1 (en) * 2000-12-12 2002-06-13 Shazam Entertainment Ltd. Method and system for interacting with a user in an experiential environment
KR100893671B1 (en) * 2001-02-12 2009-04-20 그레이스노트, 인크. Generating and matching hashes of multimedia content
US7359889B2 (en) * 2001-03-02 2008-04-15 Landmark Digital Services Llc Method and apparatus for automatically creating database for use in automated media recognition system
US6907570B2 (en) * 2001-03-29 2005-06-14 International Business Machines Corporation Video and multimedia browsing while switching between views
US6973574B2 (en) * 2001-04-24 2005-12-06 Microsoft Corp. Recognizer of audio-content in digital signals
US7421376B1 (en) * 2001-04-24 2008-09-02 Auditude, Inc. Comparison of data signals using characteristic electronic thumbprints
DE60236161D1 (en) * 2001-07-20 2010-06-10 Gracenote Inc Automatic identification of sound records
US20030036408A1 (en) * 2001-08-17 2003-02-20 Johansson Lars Olof High-density radio access system
US7171367B2 (en) * 2001-12-05 2007-01-30 Ssi Corporation Digital audio with parameters for real-time time scaling
US20030120749A1 (en) * 2001-12-21 2003-06-26 Kotzin Michael D. Method and apparatus for obtaining internet content for a wireless device
US7142565B2 (en) * 2002-02-21 2006-11-28 Qualcomm, Incorporated Method and apparatus for radio link control of signaling messages and short messages data services in a communication system
US20030182663A1 (en) * 2002-03-25 2003-09-25 Sony Corporation System and method for sharing user comments on TV screens
AT405924T (en) 2002-04-25 2008-09-15 Landmark Digital Services Llc Robust and invariant audio computer comparison
US7231351B1 (en) * 2002-05-10 2007-06-12 Nexidia, Inc. Transcript alignment
US7899915B2 (en) 2002-05-10 2011-03-01 Richard Reisman Method and apparatus for browsing using multiple coordinated device sets
KR101018320B1 (en) * 2003-02-11 2011-03-04 엔디에스 리미티드 Apparatus and methods for handling interactive applications in broadcast networks
US7359006B1 (en) * 2003-05-20 2008-04-15 Micronas Usa, Inc. Audio module supporting audio signature
US7941554B2 (en) * 2003-08-01 2011-05-10 Microsoft Corporation Sparse caching for streaming media
US7013301B2 (en) * 2003-09-23 2006-03-14 Predixis Corporation Audio fingerprinting system and method
US20050147256A1 (en) * 2003-12-30 2005-07-07 Peters Geoffrey W. Automated presentation of entertainment content in response to received ambient audio
US8010986B2 (en) * 2004-02-04 2011-08-30 Ericsson Television, Inc. Synchronization and automation in an ITV environment
US7768234B2 (en) * 2004-02-28 2010-08-03 Janik Craig M System and method for automatically synchronizing and acquiring content for battery powered devices
US7325013B2 (en) * 2004-04-15 2008-01-29 Id3Man, Inc. Database with efficient fuzzy matching
US8688248B2 (en) * 2004-04-19 2014-04-01 Shazam Investments Limited Method and system for content sampling and identification
US20070067297A1 (en) * 2004-04-30 2007-03-22 Kublickis Peter J System and methods for a micropayment-enabled marketplace with permission-based, self-service, precision-targeted delivery of advertising, entertainment and informational content and relationship marketing to anonymous internet users
EP1770543A3 (en) * 2004-05-03 2007-05-09 LG Electronics Inc. Method and apparatus for managing bookmark information for content stored in a network media server
US20050289264A1 (en) * 2004-06-08 2005-12-29 Daniel Illowsky Device and method for interoperability instruction set
US20070276733A1 (en) * 2004-06-23 2007-11-29 Frank Geshwind Method and system for music information retrieval
US8190680B2 (en) 2004-07-01 2012-05-29 Netgear, Inc. Method and system for synchronization of digital media playback
US20090138108A1 (en) * 2004-07-06 2009-05-28 Kok Keong Teo Method and System for Identification of Audio Input
US7725318B2 (en) * 2004-07-30 2010-05-25 Nice Systems Inc. System and method for improving the accuracy of audio searching
US20070016918A1 (en) * 2005-05-20 2007-01-18 Alcorn Allan E Detecting and tracking advertisements
US7399006B2 (en) * 2004-09-22 2008-07-15 Jungers Jon W Method and apparatus for assembling HVAC ductwork components
US20060068818A1 (en) 2004-09-28 2006-03-30 Amir Leitersdorf Audience participation method and apparatus
US20060068816A1 (en) * 2004-09-30 2006-03-30 Lucent Technologies Inc. Network based contacts with optional DTMF exchange of the same
US7567899B2 (en) * 2004-12-30 2009-07-28 All Media Guide, Llc Methods and apparatus for audio recognition
WO2006091636A2 (en) * 2005-02-23 2006-08-31 Digital Intelligence, L.L.C. Signal decomposition and reconstruction
US20060200037A1 (en) * 2005-03-02 2006-09-07 Falasco Marianne R System and method for selectively collecting exhaled air
US8356021B2 (en) * 2005-03-11 2013-01-15 Ross Neil Williams Method and apparatus for indexing in a reduced-redundancy storage system
US8140954B2 (en) * 2005-06-05 2012-03-20 International Business Machines Corporation System and method for providing on demand captioning of broadcast programs
US7599840B2 (en) * 2005-07-15 2009-10-06 Microsoft Corporation Selectively using multiple entropy models in adaptive coding and decoding
US7516074B2 (en) * 2005-09-01 2009-04-07 Auditude, Inc. Extraction and matching of characteristic fingerprints from audio signals
US8290810B2 (en) * 2005-09-14 2012-10-16 Jumptap, Inc. Realtime surveying within mobile sponsored content
US7873259B2 (en) * 2005-09-15 2011-01-18 International Business Machines Corporation Synching a recording time of a program to the actual program broadcast time for the program
EP1955139A4 (en) * 2005-11-10 2009-11-11 Melodis Corp System and method for storing and retrieving non-text-based information
KR100803206B1 (en) * 2005-11-11 2008-02-14 삼성전자주식회사 Apparatus and method for generating audio fingerprint and searching audio data
US9319720B2 (en) 2005-12-13 2016-04-19 Audio Pod Inc. System and method for rendering digital content using time offsets
WO2007068119A1 (en) 2005-12-13 2007-06-21 Audio Pod Inc. Segmentation and transmission of audio streams
US8214516B2 (en) * 2006-01-06 2012-07-03 Google Inc. Dynamic media serving infrastructure
US20100138561A1 (en) * 2006-01-27 2010-06-03 Michael Wayne Church Simulcast internet media distribution system and method
US7627890B2 (en) * 2006-02-21 2009-12-01 At&T Intellectual Property, I,L.P. Methods, systems, and computer program products for providing content synchronization or control among one or more devices
US7735101B2 (en) * 2006-03-28 2010-06-08 Cisco Technology, Inc. System allowing users to embed comments at specific points in time into media presentation
US20070260634A1 (en) * 2006-05-04 2007-11-08 Nokia Corporation Apparatus, system, method, and computer program product for synchronizing the presentation of media content
WO2008006100A2 (en) * 2006-07-07 2008-01-10 Redlasso Corporation Search engine for audio data
WO2008011422A2 (en) * 2006-07-21 2008-01-24 Melodis Corporation System and method for producing scored search results on a database using approximate search queries
ES2433966T3 (en) * 2006-10-03 2013-12-13 Shazam Entertainment, Ltd. Method for high flow rate of distributed broadcast content identification
EP1933482A1 (en) * 2006-12-13 2008-06-18 Taylor Nelson Sofres Plc audience measurement system, fixed and portable monitoring devices
US9253222B2 (en) * 2007-02-22 2016-02-02 Match.Com. L.L.C. Synchronous delivery of media content in a collaborative environment
US7667126B2 (en) * 2007-03-12 2010-02-23 The Tc Group A/S Method of establishing a harmony control signal controlled in real-time by a guitar input signal
CN101682461A (en) * 2007-03-22 2010-03-24 尼尔森(美国)有限公司 Digital rights management and audience measurement systems and methods
JP4935468B2 (en) * 2007-04-03 2012-05-23 ソニー株式会社 Information processing apparatus, information processing method, and computer program
WO2008129482A2 (en) * 2007-04-18 2008-10-30 Koninklijke Philips Electronics, N.V. Apparatus and methods for rendering personal stories to medical patients
US8054382B2 (en) * 2007-05-21 2011-11-08 International Business Machines Corporation Apparatus, method and system for synchronizing a common broadcast signal among multiple television units
GB0710211D0 (en) 2007-05-29 2007-07-11 Intrasonics Ltd AMR Spectrography
US20080306824A1 (en) * 2007-06-08 2008-12-11 Parkinson David C Empty Space Advertising Engine
US8027365B2 (en) * 2007-07-27 2011-09-27 Realnetworks, Inc. System and method for combining media data
US8314688B2 (en) * 2007-08-22 2012-11-20 Tagarray, Inc. Method and apparatus for low power modulation and massive medium access control
US8311828B2 (en) * 2007-08-28 2012-11-13 Nexidia Inc. Keyword spotting using a phoneme-sequence index
US9662578B2 (en) * 2007-09-21 2017-05-30 Sony Interactive Entertainment America Llc Methods and apparatuses for recording and utilizing an audio signature within an application
US8595342B2 (en) * 2007-10-17 2013-11-26 Reazer Investments L.L.C. Synchronized media playback using autonomous clients over standard Internet protocols
US7899807B2 (en) * 2007-12-20 2011-03-01 Yahoo! Inc. System and method for crawl ordering by search impact
US8229160B2 (en) * 2008-01-03 2012-07-24 Apple Inc. Systems and methods for identifying objects and providing information related to identified objects
US8161160B2 (en) * 2008-02-28 2012-04-17 Microsoft Corporation XML-based web feed for web access of remote resources
US8019882B2 (en) * 2008-06-27 2011-09-13 Microsoft Corporation Content identification for peer-to-peer content retrieval
KR101008853B1 (en) 2008-07-28 2011-01-19 주식회사 삼천리기계 A Hydraulic Chuck
US8239359B2 (en) * 2008-09-23 2012-08-07 Disney Enterprises, Inc. System and method for visual search in a video media player
US8635645B2 (en) * 2008-09-30 2014-01-21 Qualcomm Incorporated Apparatus and methods of providing and receiving venue level transmissions and services
US20100088726A1 (en) * 2008-10-08 2010-04-08 Concert Technology Corporation Automatic one-click bookmarks and bookmark headings for user-generated videos
US20100299131A1 (en) 2009-05-21 2010-11-25 Nexidia Inc. Transcript alignment
US8489774B2 (en) * 2009-05-27 2013-07-16 Spot411 Technologies, Inc. Synchronized delivery of interactive content
US20120102233A1 (en) * 2009-05-27 2012-04-26 Ajay Shah Synchronized delivery of interactive content using standardized vectors
US20130097632A1 (en) * 2009-05-27 2013-04-18 Ajay Shah Synchronization to broadcast media
US8718805B2 (en) * 2009-05-27 2014-05-06 Spot411 Technologies, Inc. Audio-based synchronization to media

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Gerasimov V., and Bender W., "Things that talk: using sound for device-to-device and device-to-human communication," IBM Systems Journal, Vol. 39, Nos. 3 & 4, pp. 530-546, December 2000 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100305730A1 (en) * 2009-05-27 2010-12-02 Glitsch Hans M Automatic resource retrieval and use
US8805723B2 (en) 2009-05-27 2014-08-12 Iviu Technologies, Llc Acoustically transmitting a resource identifier in multiple concurrent segments

Also Published As

Publication number Publication date
US20100305729A1 (en) 2010-12-02
US20110202687A1 (en) 2011-08-18
US20110202949A1 (en) 2011-08-18
US20110208334A1 (en) 2011-08-25
WO2010138777A1 (en) 2010-12-02
US20110202156A1 (en) 2011-08-18
US8718805B2 (en) 2014-05-06
US8805723B2 (en) 2014-08-12
US8789084B2 (en) 2014-07-22
US20100305730A1 (en) 2010-12-02
WO2010138776A2 (en) 2010-12-02
US20110208333A1 (en) 2011-08-25
WO2010138776A3 (en) 2011-02-03

Similar Documents

Publication Publication Date Title
JP5843207B2 (en) Intuitive computing method and system
US9996628B2 (en) Providing audio-activated resource access for user devices based on speaker voiceprint
US8060407B1 (en) Method for providing personalized, targeted advertisements during playback of media
US8718805B2 (en) Audio-based synchronization to media
US10104495B2 (en) Location-based ticket books
US9288254B2 (en) Dynamic playlist for mobile computing device
EP2679042B1 (en) Audio localization using audio signal encoding and recognition
US10027772B2 (en) Location-based content aggregation and distribution
US8346867B2 (en) Dynamic playlist for mobile computing device
US8588824B2 (en) Transferring media context information based on proximity to a mobile device
CN102714780B (en) Tagging of multimedia content with geographical location by the user of a wireless communications device
US8521316B2 (en) Coordinated group musical experience
CA2701717C (en) Gathering research data
US8972496B2 (en) Content sharing
US9979691B2 (en) Watermarking and signal recognition for managing and sharing captured content, metadata discovery and related arrangements
US8948895B2 (en) System and method for engaging a person in the presence of ambient audio
KR20160108348A (en) Digital personal assistant interaction with impersonations and rich multimedia in responses
US9202513B2 (en) Systems and methods for providing access to resources through enhanced signals
KR20140024434A (en) Methods and systems for performing comparisons of received data and providing a follow-on service based on the comparisons
CA2823602A1 (en) Matching techniques for cross-platform monitoring and information
US8026805B1 (en) Media tracker
EP2193487A1 (en) Method and system of providing event content sharing by mobile communication devices
US20100280965A1 (en) Method and apparatus for intuitive management of privacy settings
US8909256B2 (en) Method and apparatus for mediating among a plurality of profiles associated with users positioned in a shared location
US20050273330A1 (en) Anti-terrorism communications systems and devices

Legal Events

Date Code Title Description
AS Assignment

Owner name: SHAH, AJAY, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ARSH TECHNOLOGIES, LLC;REEL/FRAME:028309/0973

Effective date: 20100923

Owner name: INTERACTION CAPITAL GROUP, LLC, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ARSH TECHNOLOGIES, LLC;REEL/FRAME:028309/0973

Effective date: 20100923

Owner name: SHIOZAKI, RANDOLPH, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ARSH TECHNOLOGIES, LLC;REEL/FRAME:028309/0973

Effective date: 20100923

Owner name: ARSH TECHNOLOGIES, LLC, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GLITSCH, HANS M.;SHAH, AJAY;REEL/FRAME:028309/0877

Effective date: 20100922

Owner name: IVIU TECHNOLOGIES, LLC, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SHAH, AJAY;SHIOZAKI, RANDOLPH;INTERACTION CAPITAL GROUP, LLC;REEL/FRAME:028310/0018

Effective date: 20100923

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION