US20190019521A1 - System and method for utilizing audio encoding for measuring media exposure with environmental masking - Google Patents

System and method for utilizing audio encoding for measuring media exposure with environmental masking Download PDF

Info

Publication number
US20190019521A1
US20190019521A1 US16/002,775 US201816002775A US2019019521A1 US 20190019521 A1 US20190019521 A1 US 20190019521A1 US 201816002775 A US201816002775 A US 201816002775A US 2019019521 A1 US2019019521 A1 US 2019019521A1
Authority
US
United States
Prior art keywords
computing device
sound signals
encoded
processor
audio
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/002,775
Inventor
Wendell Lynch
Michael Tenbrock
Alan Neuhauser
Anand Jain
William K. Krug
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Citibank NA
Original Assignee
Nielsen Co US LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US12/425,556 external-priority patent/US20100268573A1/en
Priority claimed from US12/425,464 external-priority patent/US20100268540A1/en
Application filed by Nielsen Co US LLC filed Critical Nielsen Co US LLC
Priority to US16/002,775 priority Critical patent/US20190019521A1/en
Publication of US20190019521A1 publication Critical patent/US20190019521A1/en
Assigned to NIELSEN AUDIO, INC. reassignment NIELSEN AUDIO, INC. CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: ARBITRON, INC
Assigned to THE NIELSEN COMPANY (US), LLC. reassignment THE NIELSEN COMPANY (US), LLC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: NIELSEN AUDIO, INC.
Assigned to ARBITRON, INC reassignment ARBITRON, INC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: JAIN, ANAND, LYNCH, WENDELL, TENBROCK, MICHAEL, NEUHAUSER, ALAN R., KRUG, WILLIAM K.
Assigned to CITIBANK, N.A. reassignment CITIBANK, N.A. SUPPLEMENTAL SECURITY AGREEMENT Assignors: A. C. NIELSEN COMPANY, LLC, ACN HOLDINGS INC., ACNIELSEN CORPORATION, ACNIELSEN ERATINGS.COM, AFFINNOVA, INC., ART HOLDING, L.L.C., ATHENIAN LEASING CORPORATION, CZT/ACN TRADEMARKS, L.L.C., Exelate, Inc., GRACENOTE DIGITAL VENTURES, LLC, GRACENOTE MEDIA SERVICES, LLC, GRACENOTE, INC., NETRATINGS, LLC, NIELSEN AUDIO, INC., NIELSEN CONSUMER INSIGHTS, INC., NIELSEN CONSUMER NEUROSCIENCE, INC., NIELSEN FINANCE CO., NIELSEN FINANCE LLC, NIELSEN HOLDING AND FINANCE B.V., NIELSEN INTERNATIONAL HOLDINGS, INC., NIELSEN MOBILE, LLC, NIELSEN UK FINANCE I, LLC, NMR INVESTING I, INC., NMR LICENSING ASSOCIATES, L.P., TCG DIVESTITURE INC., THE NIELSEN COMPANY (US), LLC, THE NIELSEN COMPANY B.V., TNC (US) HOLDINGS, INC., VIZU CORPORATION, VNU INTERNATIONAL B.V., VNU MARKETING INFORMATION, INC.
Assigned to CITIBANK, N.A reassignment CITIBANK, N.A CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001. ASSIGNOR(S) HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT. Assignors: A.C. NIELSEN (ARGENTINA) S.A., A.C. NIELSEN COMPANY, LLC, ACN HOLDINGS INC., ACNIELSEN CORPORATION, ACNIELSEN ERATINGS.COM, AFFINNOVA, INC., ART HOLDING, L.L.C., ATHENIAN LEASING CORPORATION, CZT/ACN TRADEMARKS, L.L.C., Exelate, Inc., GRACENOTE DIGITAL VENTURES, LLC, GRACENOTE MEDIA SERVICES, LLC, GRACENOTE, INC., NETRATINGS, LLC, NIELSEN AUDIO, INC., NIELSEN CONSUMER INSIGHTS, INC., NIELSEN CONSUMER NEUROSCIENCE, INC., NIELSEN FINANCE CO., NIELSEN FINANCE LLC, NIELSEN HOLDING AND FINANCE B.V., NIELSEN INTERNATIONAL HOLDINGS, INC., NIELSEN MOBILE, LLC, NMR INVESTING I, INC., NMR LICENSING ASSOCIATES, L.P., TCG DIVESTITURE INC., THE NIELSEN COMPANY (US), LLC, THE NIELSEN COMPANY B.V., TNC (US) HOLDINGS, INC., VIZU CORPORATION, VNU INTERNATIONAL B.V., VNU MARKETING INFORMATION, INC.
Assigned to BANK OF AMERICA, N.A. reassignment BANK OF AMERICA, N.A. SECURITY AGREEMENT Assignors: GRACENOTE DIGITAL VENTURES, LLC, GRACENOTE MEDIA SERVICES, LLC, GRACENOTE, INC., THE NIELSEN COMPANY (US), LLC, TNC (US) HOLDINGS, INC.
Assigned to CITIBANK, N.A. reassignment CITIBANK, N.A. SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GRACENOTE DIGITAL VENTURES, LLC, GRACENOTE MEDIA SERVICES, LLC, GRACENOTE, INC., THE NIELSEN COMPANY (US), LLC, TNC (US) HOLDINGS, INC.
Assigned to ARES CAPITAL CORPORATION reassignment ARES CAPITAL CORPORATION SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GRACENOTE DIGITAL VENTURES, LLC, GRACENOTE MEDIA SERVICES, LLC, GRACENOTE, INC., THE NIELSEN COMPANY (US), LLC, TNC (US) HOLDINGS, INC.
Assigned to GRACENOTE MEDIA SERVICES, LLC, Exelate, Inc., NETRATINGS, LLC, THE NIELSEN COMPANY (US), LLC, GRACENOTE, INC., A. C. NIELSEN COMPANY, LLC reassignment GRACENOTE MEDIA SERVICES, LLC RELEASE (REEL 053473 / FRAME 0001) Assignors: CITIBANK, N.A.
Assigned to GRACENOTE, INC., NETRATINGS, LLC, Exelate, Inc., THE NIELSEN COMPANY (US), LLC, A. C. NIELSEN COMPANY, LLC, GRACENOTE MEDIA SERVICES, LLC reassignment GRACENOTE, INC. RELEASE (REEL 054066 / FRAME 0064) Assignors: CITIBANK, N.A.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/018Audio watermarking, i.e. embedding inaudible data in the audio signal
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/683Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F17/30743
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0201Market modelling; Market analysis; Collecting market data

Definitions

  • the present disclosure relates to systems and processes for communicating and processing data, and, more specifically, to communicate media data exposure that may include coding that provides media and/or market research.
  • watermarking has been known in the art for incorporating information signals into media signals or executable code.
  • Typical watermarks may include encoded indications of authorship, content, lineage, existence of copyright, or the like.
  • other information may be incorporated into audio signals, either concerning the signal itself, or unrelated to it.
  • the information may be incorporated in an audio signal for various purposes, such as identification or as an address or command, whether or not related to the signal itself.
  • CBET Critical Band Encoding Technology
  • This embedded audio signal or ID code is then picked up by small (pager-size) specially designed receiving stations called Portable People Meters (PPM), which capture the encoded identifying signal, and store the information along with a time stamp in memory for retrieval at a later time.
  • PPM Portable People Meters
  • a microphone contained within the PPM receives the audio signal, which contains within it the ID code.
  • the encoded audio signal described above is suitable for broadcast transmission and reception and may be adapted for Internet transmission, reception, recording and reproduction.
  • the audio signal is processed to detect the presence of the multiple-frequency code signal.
  • a portion of the multiple-frequency code signal e.g., a number of single frequency code components, inserted into the original audio signal, is detected in the received audio signal.
  • the information signal itself may be recovered.
  • an audio beacon system, apparatus and method for collecting information on a panelist's exposure to media.
  • the audio beacon is configured as on-device encoding technology that is operative in a panelist's processing device (e.g., cell phone, PDA, PC) to enable the device to acoustically transmit user/panelist data for a predetermined period of time.
  • a panelist's processing device e.g., cell phone, PDA, PC
  • PPM Personal People Meter
  • the acoustic transmissions are configured to utilize environmental sounds that are advantageous in being less obtrusive to users.
  • FIG. 1A is a block diagram illustrating a portion of an audio beaconing system under one exemplary embodiment
  • FIG. 1B is a block diagram illustrating another portion of an audio beaconing system under the embodiment illustrated in FIG. 1A ;
  • FIG. 2 is a tabular illustration of an audio beaconing and audio matching process under another exemplary embodiment
  • FIG. 3 illustrates a block diagram of a server-side encoding process under yet another exemplary embodiment
  • FIG. 4 illustrates an exemplary watermarking process for a digital media file suitable for use in the embodiment of FIGS. 1A-B ;
  • FIG. 5 illustrates a block diagram of a client-side encoding process under yet another exemplary embodiment
  • FIG. 6 illustrates an exemplary audio waveform having encoded data therein
  • FIG. 7 illustrates an exemplary device configured to select environmental sounds for audio transmission.
  • FIG. 1A is an exemplary block diagram illustrating a portion of an audio beaconing system 150 under one embodiment, where a web page 110 is provided by a page developer and published on content server 100 .
  • the web page preferably contains an embedded video player 111 and audio player 112 (that is preferably not visible), together with an application programming interface (API) 113 .
  • Other content 114 e.g., HTML, text, etc.
  • API 113 is preferably embodied as a set of routines, data structures, object classes and/or protocols provided by libraries and/or operating system services in order to support the video player 111 and audio player 112 . Additionally, the API 113 may be language-dependent (i.e.
  • suitable API's include Windows API, Java Platform API, OpenGL, DirectX, Simple DirectMedia Layer (SDL), YouTube API, Facebook API and iPhone API, among others.
  • API 113 is configured as a beaconing API object.
  • the API object may reside on an Audience Measurement (AM) server 120 , so that the object may be remotely initialized, thus minimizing the objects software's exposure to possible tampering and to maintain security.
  • the API object can reside on the content server 100 , where the API object may be initialized under increased performance conditions.
  • API 113 can communicate the following properties: (1) the URL of the page playing the media, (2) URL of the media being served on the page, (3) any statically available media metadata, and (3) a timestamp. It is understood that additional properties may be communicated in API 113 as well.
  • an initialization request is received by API 113 , to create a code tone that is preferably unique for each website and encode it on a small inaudible audio stream.
  • the AM server 120 could generate a pre-encoded audio clip 101 , with a code tone, for each site and forward it on the content server 100 in advance.
  • the encoded audio stream would then travel from content server 100 to the web page 110 holding audio player 110 .
  • audio player 110 may be set by the page developer as an object instance, where the visible property of player 110 is oriented as “false” or set to a one-by-one dimension in order to minimize the visual interference of the audio player with the web page.
  • the encoded audio stream may then be played out in parallel with the media content being received from the web page 110 .
  • the encoded audio stream would preferably repeat at predetermined time periods through an on-device beacon 131 resident on a user device 130 as long as the user is on the same website.
  • the beacon 131 would enable device 130 to acoustically transmit the encoded audio stream so that a suitably configured portable device 140 (e.g., PPM) can receive and process the encoded information.
  • a suitably configured portable device 140 e.g., PPM
  • Beacon 131 could be embedded into an audio player resident on a web page being viewed inside the browser on user device 130 , or may be a stand-alone application on user device 130 .
  • FIGS. 1A-B A simplified example further illustrates the operation of the system 150 of FIGS. 1A-B under an alternate embodiment.
  • User device 130 requests content (e.g., http://www.hulu.com/) from server 100 .
  • content e.g., http://www.hulu.com/
  • PC meter software 132 collects and transmits web measurement data to Internet measurement database 141 .
  • One example of a PC meter is comScore's Media MetrixTM software; further exemplary processes of web metering may be found in U.S. Pat. No. 7,493,655, titled “Systems for and methods of placing user identification in the header of data packets usable in user demographic reporting and collecting usage data” and U.S. Pat. No. 7,260,837, titled “Systems and methods for user identification, user demographic reporting and collecting usage data usage biometrics”, both of which are incorporated by reference in their entirety herein.
  • beacon 131 acoustically transmits encoded audio, which is received by portable device 140 .
  • the encoding for the beacon transmission may include data such as a timestamp, portable device ID, user device ID, household ID, or any similar information.
  • portable device 140 additionally receives multimedia data such as television and radio transmissions 142 , which may or may not be encoded, at different times. If encoded (e.g., CBET encoding), portable device can forward transmissions 142 to audio matching server 160 ( FIG. 1B ) for decoding and matching with audio matching database 161 . If transmissions 142 are not encoded, portable device 140 may employ sampling techniques for creating audio patterns or signatures, which may also be transmitted to audio matching server 160 for pattern matching using techniques known in the art.
  • Audio beacon server 150 receives and processes/decodes beacon data from portable device 140 . Under an alternate embodiment, it is possible to combine audio matching server 160 and audio beacon server 150 to collectively process both types of data. Data from Audio beacon server 150 and audio matching server 160 is transmitted to Internet measurement database 141 , where the web measurement data could be combined with audio beacon data and data from the audio matching server to provide a comprehensive collection of panelist media exposure data.
  • the video and audio players of webpage 110 are configured to operate as Flash Video, which is a file format used to deliver video over the Internet using AdobeTM Flash Player.
  • the Flash Player typically executes Shockwave Flash “SWF” files and has support for a scripting language called ActionScript, which can be used to display Flash Video from an SWF file.
  • ActionScript a scripting language
  • Flash Video files contain video bit streams which are a variant of the H.263 video standard, and include support for H.264 video standard (i.e., “MPEG-4 part 10”, or “AVC”).
  • Audio in Flash Video files (“FLV”) is usually encoded as MP3, but can also accommodate uncompressed audio or ADPCM format audio.
  • video beacons can be embedded within an action script that will be running within the video Flash Player's run time environment on web page 110 .
  • an action script associated with web page 110 gets loaded as a result of the access to the page, the script gets activated and triggers a “video beacon”, which extracts and store URL information on a server (e.g., content server 100 ), and launches the video Flash Player.
  • a server e.g., content server 100
  • the audio beacon will be triggered by the video player.
  • the audio beacon may access AM server 120 to load a pre-recorded audio file containing a special embedded compatible code (e.g., CBET). This pre-recoiled audio file would be utilized for beacon 131 to transmit for a given period of time (e.g., every x seconds).
  • a special embedded compatible code e.g., CBET
  • the beacon 131 audio player runs as a “shadow player” in parallel to the video Flash Player. If a portable device 140 is in proximity to user device 130 , portable device 140 will detect the code and reports it to audio beacon server 150 . Depending on the level of cooperation between the audio and video beacon, the URL information can also be deposited onto beacon server 150 along with codes that would allow an audience measurement entity to correlate and/or calibrate various measurements with demographic data.
  • each user device 130 may install on-device measurement software (PC meter 132 ) which includes one or more web activity monitoring applications, as well as beacon software 131 . It is understood that the web activity monitoring application and the beacon software may be individual applications, or may be merged into a single application.
  • PC meter 132 on-device measurement software
  • the web activity monitoring application collects web activities data from the user device 130 (e.g., site ID, video page URL, video file URL, start and end timestamp and any additional metadata about videosite information, URL information, time, etc.) and additionally assigns a unique ID, such as a globally unique identifier or “GUID”, to each device.
  • a unique composite ID may be assigned including a household ID (“HHID”) and a unique user device ID for each device in the household up to 10 devices for a family), as well as a portable device ID (PPMID).
  • Panelist demographic data may be included for each web activity on the device.
  • beacon 131 emits an audio beacon code (ABC) for device in the household by encoding an assigned device ID number and acoustically sending it to portable device 140 to identify the device. Further details on the encoding is provided below.
  • Portable device 140 collects the device ID and sends it to a database along with HHID and/or PPM ID and the timestamp.
  • a PPMID is always mapped to a in the backend; alternately an HHID can be set within each PPMID.
  • the web activity monitoring and beacon applications may pass information to each other as needed. Both can upload information to a designated server for additional processing.
  • a directory of panelists' devices is built to contain the GUID, HHID, and device ID for panel, and the directory could be used to correlate panelist demographic data and web measurement data.
  • FIG. 2 a tabular illustration of an audio beaconing and audio matching process under another an exemplary embodiment is provided.
  • the table illustrates a combination of audio beaconing and audio matching and its application to track a video on a content site, such as Hulu.com.
  • FIG. 2 Timeline 200 shows in sections a scenario where a user/panelist plays a ten minute video on Hulu.com.
  • Activities 201 shows actions taken in user system 150 where a video is loaded in the user device 130 , and played.
  • At the 5 minute mark (301 sec.) a 15 second advertisement is served.
  • the video continues to play until its conclusion (600 sec.).
  • beacon activities 202 are illustrated, where, under one embodiment, on-device beacon 131 transmits continuous audio representing the website (Hulu.com).
  • beacon also transmits a timestamp, portable device ID, user device ID, household ID and/or any other data in accordance with the techniques described above.
  • additional data may be transmitted in the beacon to include URLs and video ID's when a video is loaded and played.
  • an event beacon which may include advertisement URL data, is transmitted.
  • a video end beacon is transmitted to indicate the user/panelist is no longer viewing specific media.
  • portable device data 205 and end-user experience 206 of FIG. 2 portable device data (e.g., demographic ID data) is overlayed along with site information (URL, video ID, etc.) when a video is loaded.
  • site information URL, video ID, etc.
  • audio signatures may be sampled periodically by portable device 140 , until a content match is achieved. The audio signatures may be obtained through encoding, pattern matching, or any other suitable technique.
  • portable device data is overlayed to indicate that a content match exists.
  • signature sampling/audio matching allows the system 150 to identify and incorporate additional data on the users/panelists and the content being viewed.
  • the content provider media e.g., Hulu, Facebook, etc.
  • a matching database e.g., audio matching server 160 .
  • the portable device 140 would be equipped with audio matching software, so that, when a panelist is in the vicinity of user device 130 , audio matching techniques are used to collect the signature, or “audio fingerprint” for the incoming stream. The signatures would then be matched against the signatures in the matching database to identify the content.
  • encoding techniques may also be employed to identify content data.
  • content is encoded prior to transmission to include data relating to the content itself and the originating content site.
  • data relating to possible referral sites e.g., Facebook, MySpace, etc.
  • a content management system may be arranged for content distributors to choose specific files for a corresponding referral site.
  • Still other suitable encoding techniques are the subject of PCI Publication WO 00/04662 to Srinivasan, U.S. Pat. No. 5,319,735 to Preuss, et al., U.S. Pat. No. 6,175,627 to Petrovich, et al., U.S. Pat. No. 5,828,325 to Wolosewicz, et al., U.S. Pat. No. 6,154,484 to Lee, et al., U.S. Pat. No.
  • the encoder may be based on a Streaming Audio Encoding System (SAES) that operates under a set of sample rates and is integrated with media transcoding automation technology, such as Telestream's FlipFactoryTM software.
  • SAES Streaming Audio Encoding System
  • the encoder may be embodied as a console mode application, written in a general-purpose computer programming language such as “C”.
  • the encoder may be implemented as a Java Native Interface (JNI) to allow code running in a virtual machine to call and be called by native applications, where the JNI would include a JNI shared library for control using Java classes.
  • the encoder payloads would be configured using specially written Java classes.
  • the encoder would use the information hiding abstractions of an encoder payload which defines a single message.
  • the JNI encoder would operate using a 44.1 kHz sample rate.
  • One exemplary symbol configuration uses four data symbols and one end symbol defined for a total of five symbols. Each symbol may comprise five tones, with one tone coming from each of five standard Barks.
  • Bark scale edges in Hertz
  • the bins are preferably spaced on a 4 ⁇ 3.90625 grid in order to provide lighter processing demands, particularly in cases using decoders based on 512 point fast Fourier transform (FFT).
  • FFT 512 point fast Fourier transform
  • an exemplary message would comprise 20 symbols, each being 400 milliseconds in duration, for a total duration of 8 seconds.
  • the first 3 symbols could be designated as match/check criteria symbols, which could be the simple sum of the data symbols or could be derived from an error correction or cyclical redundancy check algorithm.
  • the following 16 symbols would then he designated as data symbols, leaving the last symbol as an end symbol used for a marker.
  • the total number of possible symbols would be 4 16 or 4,294,967,296 symbols.
  • a core sampling rate of 5.5125 kHz may be used instead of 8 kHz to allow down-sampling from 44.1 kHz to be efficiently performed without pre-filter (to eliminate aliasing components) followed by conversion filter to 48 kHz.
  • Such a configuration should have no effect on code tone grid spacing since the output frequency generation is independent of the core sampling rate. Additionally, this configuration would limit the top end of the usable frequency span to about 2 kHz (as opposed to 3 kHz under conventional techniques) since frequency space should be left for filters with practical numbers of taps.
  • a 16 point overlap of a 256 point large FFT is used, resulting in amplitude updates every 2.9 milliseconds for encoding instead of every 2 milliseconds for standard CBET techniques. Accordingly, fewer large FFTs are calculated under a tighter bin resolution of 21.5 Hz instead of 31.25 Hz.
  • bin spans of the clumps may be set by Bark boundaries instead of being wholly based on Critical Bandwidth criteria. By using Bark boundaries, a specific bin will not contribute to the encoding power level of multiple clumps, which provides less coupling between code amplitudes of adjacent clumps.
  • a comparison may he made of the most recent 16 point Small FFT results to a history of squared sums to simplify calculations.
  • the encoding algorithm under the present disclosure would preferably use 3 bin values over a clump: the minimum bin power (MIN), the maximum bin power (MAX), and the average bin power (AVG). Under this arrangement, the bin values could be modeled as follows:
  • PWR MIN ELSE
  • AVG PWR may be scaled by a predetermined factor to produce masking energy.
  • a similar algorithm could also be used to create a 48 kHz native encoder using a core sample rate of 6 kHz and a large FFT bin resolution of 23.4375 Hz calculated every 2.67 milliseconds.
  • Such a configuration would differ slightly in detection efficiency and inaudibility from the embodiments described above, but it is anticipated that the differences would be slight.
  • an exemplary configuration would include a software decoder based on a JNI shared library, which performs calculations up through the bin signal-to-noise ratios. Such a configuration would allow an external application to define the symbols and perform pattern matching. Such steps would be handled in a Java environment using an information hiding extraction of a decoder payload, where decoder payloads are created using specially written Java classes.
  • content server 100 has content 320 , which includes a media file 302 configured to be requested and played on media player 301 residing on user device 130 .
  • media file 302 is initialized, audio is extracted from the media file and, if the audio is encoded (e.g., MP3 audio), subjected to audio decoding in 304 to produce raw audio 305 .
  • audio is encoded (e.g., MP3 audio), subjected to audio decoding in 304 to produce raw audio 305 .
  • device ID, HHID anchor PPMID data is provided for first encoding 306 the data into the raw audio 305 , using any suitable technique described above.
  • the audio data is then subjected to a second encoding to transform the audio into a suitable format (e.g., MP3) to produce fully encoded audio 308 , which is subsequently transmitted to media player 301 and beaconed to portable device 140 .
  • a suitable format e.g., MP3
  • encoded audio 308 may be produced in advance and stored as part of media file 302 .
  • care must be taken to account for processing delays to ensure that the encoded audio is properly synchronized with any video content in media file 302 .
  • the server-side encoding may be implemented under a number of different options.
  • a first option would be to implement a pre-encoded beacon, where the encoder ( 306 ) would be configured to perform real-time encoding of the audio beacon based on the content being served to the users/panelists.
  • the user device would be equipped with a software decoder as described above which is invoked when media is played.
  • the pre-encoded beacon would establish a message link which could be used, along with an identifier from the capturing portable device 140 , in order to assign credit.
  • the encoding shared library would preferably be resident at the content site ( 100 ) as part of the encoding engine. Such a configuration would allow the transcoding and encoding to be fit into the content site workflow.
  • server-side encoding could include a pre-encoded data load, where the audio is encoded with a message that is based on the metadata or the assigned URL. This establishes a message link which can be used, along with an identifier from the capturing portable device 140 , in order to assign credit.
  • the encoding shared library is preferably resident at the content site ( 100 ), as part of the encoding engine. Again, this configuration would allow the transcoding and encoding to be fit into the content site workflow.
  • server-side encoding could include “on-the-fly” encoding. If a video is being streamed to a panelist, encoding may be inserted in the stream along with a transcoding object. The encoding may be used to encode the audio with a simple one of N beacon, and the panelist user device 130 would contain software decoding which is invoked when the video is played. This also establishes a message link which can be used, along with an identifier from the capturing portable device 140 , in order to assign credit.
  • the encoding shared library is preferably resident at the content site ( 100 ), as part of the encoding engine. Under a preferred embodiment, an ActionScript would invoke the decoding along with a suitable transcoding object.
  • FIG. 4 illustrates one embodiment for encoding media under a Flash Video platform 410 , where the content is preferably encoded in advance.
  • the audio is subjected to water mark encoding 401 , which may include techniques described above for the encoding.
  • water mark encoding 401 may include techniques described above for the encoding.
  • the audio is formatted as a Flash file using Adobe Tools 402 such as FLV Creator and SWF Compiler.
  • the file is further formatted using Flash-supported codecs (e.g., H.264, VP6, MPEG-4 ASP, Sorenson H.263) and compression 403 to produce a watermarked A/V stream or file 404 .
  • Flash-supported codecs e.g., H.264, VP6, MPEG-4 ASP, Sorenson H.263
  • FIG. 5 provides another alternate embodiment that illustrates client-side encoding and processing.
  • user device 130 requests media data.
  • a media file 531 residing on content server 100 is subsequently streamed to the device's browser 520 arranged on user's workspace 510 .
  • Media player 521 plays the streamed content and produces raw audio 511 .
  • a client-side ActionScript notifies browser 522 and encoder 522 to capture the raw audio on the device's sound mixer, or microphone (not shown), and to encode data using a suitable encoding technique described above.
  • the encoding constructs the data for an independent audio beacon using the captured audio and other data (e.g., device ID, HHID, etc.) where portable device 140 picks up the beacon and forwards the data to an appropriate server for further processing and panel data evaluation.
  • Actionscript interface APIs such as “ExternalInterface”, which is an application programming interface that enables straightforward communication between ActionScript and a Flash Player container; for example, an HTML page with JavaScript, or a desktop application with Flash Player embedded, along with encoder application 522 .
  • an ActionScript interface could be used to call code in the container application, including a web page or desktop application.
  • ActionScript code could be called from code in the container application.
  • a proxy could be created to simplify calling ActionScript code from the container application.
  • a beacon embodiment may be enabled by having an encoding message being one from a relatively small set (e.g., 1 of 12), and where each user device 130 is assigned a different message.
  • the encoding message may be a hash of the site and/or URL information gleaned from the metadata.
  • a reverse hash can be used to identify the site, where the hash could be resolved on one or more remote server (e.g., sever 160 ).
  • a simplified beaconing configuration may be arranged where the beacon operates as a complement to media data, independent of the media data, or providing a beacon where no specially encoded data exists.
  • the simplified beaconing comprises a constant amplitude acoustic signal or tone that is generated on user device 130 .
  • This acoustic tone is then automatically encoded, preferably with identification data (e.g., device ID, HHID and/or PPM ID) and a timestamp.
  • the encoded acoustic tone would then be forwarded to portable device 140 for processing and identification.
  • the acoustic tone used for the twinkle is preferably embodied as a pre-recorded constant amplitude tone that is transmitted at predetermined times.
  • the encoding is preferably performed using any of the techniques described above.
  • the simplified beaconing process would only forward the encoded, pre-recorded tone, independently of any audio data being received.
  • user device 130 receives only other content 114 from content server 110 in the form of text-based HTML.
  • the encoded tone is transmitted to portable device 140 , where after further processing (see FIG. 1B ), the user identification data is merged into internet measurement database 141 .
  • user device 140 may also receive audio data (encoded or unencoded) separately and in addition to other content. While the techniques described above would encode and forward audio data received, the simplified beacon (“twinkle”) would also transmit ID information to portable device 140 , which, in conjunction with PC meter 132 , would subsequently merge panelist data into a common database.
  • audio data encoded or unencoded
  • FIG. 6 illustrates audio signal 600 represented as a spectrum of audio 610 over a period of time (e.g., 0.25 seconds), where the energy intervals vary with frequency between 1200 and 2200 Hz.
  • Overlaid in black are discreet, narrowband code tones 602 (e.g., CBET) opportunistically inserted into the audio using the principles of psychoacoustic masking.
  • narrowband code tones 602 e.g., CBET
  • the energy of the inserted code tone varies with the level of audio, so more quiet portions of the frequency spectrum (e.g., 604 ) receive little encoding energy and compared to louder portions 605 ), which get proportionally more.
  • the simplified encoding (“twinkle”) 603 is encoded and inserted at constant levels across the frequency spectrum, where the levels are independent of the audio levels. This allows the simplified encoding to be pre-recorded, easily generated and capable of being reused accross various and/or different content.
  • the simplified encoding could have the same message structure as the CBET encoding described above, utilizing a 10-tone symbol set. Alternately, other message structures are possible as well.
  • the twinkle may be transmitted automatically at regular intervals. Alternately the twinkle may be invoked by an ActionScript.
  • the ActionScript could relay a beacon for the media from user device 130 to portable device 140 , while simultaneously requesting a second (preferably invisible) Flash Player in the user device 130 to transmit the twinkle to portable device 140 .
  • the ActionScript should invoke both players at a common volume setting.
  • the beacon may be arranged to have audio characteristics that make it easier and more robust to encode. For example, simulated environmental sounds, such as power supply fans, air vent exhaust, crowd/audience noise, ocean waves and such may be used as the audio beacon sound in which encoded messages may be inserted.
  • many computing devices such as personal computers, phones, tablets and laptops, contain a pre-stored library of audio that is used to alert or notify users. By creating and storing sounds advantageously suited to encode data in the on-device sound library, a more robust beacon may be utilized.
  • the encoding on the device may be simplified, since the time, frequency, masking and other encoding processes will be known.
  • FIG. 7 an exemplary embodiment is illustrated where user device 130 is configured with a control panel 700 that allows the device to control various aspects, including sound 701 that includes a resident sound library 720 .
  • a script operating alone or as part of meter software (see FIG. 1A , ref. 132 ), may control aspects of sound library 720 , such that specific sounds are used for audio beacon encoding.
  • Sound library 720 comprises audio sounds ( 704 - 707 ) that are associated with one or more software applications 700 - 701 and/or events 702 - 703 .
  • “Events,” for the purposes of FIG. 7 include notifications (e.g., receipt of email, social networking software status update, etc.) and alerts (e.g., start-up, shut-down, application error, etc.) that are made pursuant to the device's operating system and associated software running on the device.
  • a first application (APP 1 ) 700 and a second application (APP 2 ) 701 are associated with respective sounds 704 - 705 that may be triggered when the application is activated, or at predetermined times determined by the specific application.
  • Applications 700 - 701 may include metering software 132 discussed above in connection with FIG. 1A .
  • each sound 704 , 705 is associated with a respective application 700 , 701 .
  • the sounds ( 704 - 705 ) are used for encoding messages to form respective beacons ( 708 - 709 ).
  • the sound e.g., 704
  • the sound is copied, data is encoded into it and the encoded copy is stored in a buffer or other suitable memory.
  • another copy of the original sound ( 704 ) is made, and the further data is encoded and stored into the buffer/memory. This process repeats for as long as necessary to form a string of encoded sounds.
  • the sounds may be arranged sequentially or in other suitable formats.
  • a first beacon 708 is audibly transmitted.
  • the next beacon is audibly transmitted, and so on, until the buffer/memory is empty, or a predetermined amount of time has expired.
  • one application can control a plurality of sounds ( 704 - 705 ) and produce a plurality of encoded beacons ( 708 - 709 ).
  • events 702 - 703 may be associated with respective sounds 706 - 707 , similar to applications 700 - 701 discussed above, to produce beacons 710 - 711 .
  • This configuration may be particularly advantageous to beacon information when a user performs an act on the device, such as open/close a browser window, open/close a tab on a browser, open/close an application, etc.
  • data for the audio beacon is being collected (e.g., user device ID, web hash, etc.)
  • the sound e.g., 706
  • data is encoded into it and the encoded copy is stored in a buffer or other suitable memory.
  • the sounds 704 - 707 are preferably predetermined and may simulate an environmental sound so as not to be intrusive or distracting to the user. Additionally, the sound may be selected to contain audio characteristics (e.g., having high masking levels in critical frequency bands) that makes it conducive to robust audio encoding. By using a predetermined sound for the audio beacon encoding, designers can have more flexibility in audibly beaconing data. In addition to audio characteristics, the predetermined sounds may have different lengths as well. In an example where an application ( 700 ) controls multiple sounds, the sounds may be the same instance of one sound, but having different lengths (e.g., 5 sec., 10 sec., etc.).
  • the application may default to a longer sound to increase the probability of the beacon code being detected. If the volume increases, the device 130 can switch to a shorter sound. This configuration has the added benefit of ensuring that users maintain a sufficient volume on their device to avoid longer (and possibly more intrusive) beacons.
  • Various embodiments disclosed herein provide devices, systems and methods for performing various functions using an audience measurement system that includes audio beaconing. Although specific embodiments are described herein, those skilled in the art recognize that other embodiments may be substituted for the specific embodiments shown to achieve the same purpose. As an example, although terms like “portable” are used to describe different components, it is understood that other, fixed, devices may perform the same or equivalent functions. Also, while specific communication protocols are mentioned in this document, one skilled in the art would appreciate that other protocols may be used or substituted. This application covers any adaptations or variations of the present invention. Therefore, the present invention is limited only by the claims and all available equivalents.

Abstract

An audio beacon system, apparatus and method for collecting information on a panelist's exposure to media. An audio beacon is configured as on-device encoding technology that is operative in a processing device (e.g., cell phone, PDA, PC) to enable the device to encode an environmental sound and transmit it for a predetermined period of time. The acoustically transmitted data is received and processed by a portable audience measurement device, such as Arbitron's Personal People Meter™ (“PPM”), or other specially equipped portable device to enable audience measurement systems to achieve higher levels of detail on panel member activity and greater association of measurement devices to their respective users.

Description

    RELATED APPLICATIONS
  • The present application is a continuation-in-part of U.S. patent application Ser. No. 12/425,464, titled “System and Method for Utilizing Audio Beaconing in Audience Measurement” filed Apr. 17, 2009, and U.S. patent application Ser. No. 12/425,556, titled “System and Method for Utilizing Supplemental Audio Beaconing in Audience Measurement,” also filed on Apr. 17, 2009. Both applications are assigned to the assignee of the present application and are incorporated by reference in their entireties herein
  • TECHNICAL FIELD
  • The present disclosure relates to systems and processes for communicating and processing data, and, more specifically, to communicate media data exposure that may include coding that provides media and/or market research.
  • BACKGROUND INFORMATION
  • The use of global distribution systems such as the Internet for distribution of digital assets such as music, film, computer programs, pictures, games and other content continues to grow. In many instances, media offered via traditional broadcast mediums is supplemented through similar media offerings through computer networks and the Internet. It is estimated that Internet-related media offerings will rival and even surpass traditional broadcast offerings in the coming years.
  • Techniques such as “watermarking” have been known in the art for incorporating information signals into media signals or executable code. Typical watermarks may include encoded indications of authorship, content, lineage, existence of copyright, or the like. Alternatively, other information may be incorporated into audio signals, either concerning the signal itself, or unrelated to it. The information may be incorporated in an audio signal for various purposes, such as identification or as an address or command, whether or not related to the signal itself.
  • There is considerable interest in encoding audio signals with information to produce encoded audio signals having substantially the same perceptible characteristics as the original unencoded audio signals. Recent successful techniques exploit the psychoacoustic masking effect of the human auditory system whereby certain sounds are humanly imperceptible when received along with other sounds.
  • Arbitron has developed a new and innovative technology called Critical Band Encoding Technology (CBET) that encompasses all forms of audio and video broadcasts in the measurement of audience participation. This technology dramatically increases the both the accuracy of the measurement and the quantity of useable and effective data across all types of signal broadcasts. CBET is an encoding technique that Arbitron developed and that embeds identifying information (ID code) or other information within the audio portion of a broadcast. An audio signal is broadcast within the actual audio signal of the program, in a manner that makes the ID code inaudible, to all locations the program is broadcast, for example, a car radio, home stereo, computer network, television. etc. This embedded audio signal or ID code is then picked up by small (pager-size) specially designed receiving stations called Portable People Meters (PPM), which capture the encoded identifying signal, and store the information along with a time stamp in memory for retrieval at a later time. A microphone contained within the PPM receives the audio signal, which contains within it the ID code.
  • Further disclosures related to CBET encoding may be found in U.S. Pat. No. 5,450,490 and U.S. Pat. No. 5,764,763 (Jensen et at.) in which information is represented by a multiple-frequency code signal which is incorporated into an audio signal based upon the masking ability of the audio signal. Additional examples include U.S. Pat. No. 6,871,180 (Neuhauser et al.) and U.S. Pat. No. 6,845,360 (Jensen et al.), where numerous messages represented by multiple frequency code signals are incorporated to produce and encoded audio signal. Other examples include U.S. Pat. No. 7,239,981 (Kolessar et al.). Each of the above-mentioned patents is incorporated by reference in its entirety herein.
  • The encoded audio signal described above is suitable for broadcast transmission and reception and may be adapted for Internet transmission, reception, recording and reproduction. When received, the audio signal is processed to detect the presence of the multiple-frequency code signal. Sometimes, only a portion of the multiple-frequency code signal, e.g., a number of single frequency code components, inserted into the original audio signal, is detected in the received audio signal. However, if a sufficient quantity of code components is detected, the information signal itself may be recovered.
  • Other means of watermarking have been used in various forms to track multimedia over computer networks and to detect if a user is authorized to access and play the multimedia. For certain digital media, metadata is transmitted along with media signals. This metadata can be used to catty one or more identifiers that are mapped to metadata or actions. The metadata can be encoded at the time of broadcast or prior to broadcasting. Decoding of the identifier may be performed at a digital receiver. Other means of watermarking include the combination of digital watermarking with various encryption techniques known in the art.
  • While various encoding and watermarking techniques have been used to track and protect digital data, there have been insufficient advances in the fields of cross-platform digital media monitoring. Specifically, in cases where a person's exposure to Internet digital media is monitored in addition to exposure to other forms of digital media (e.g., radio, television, etc.), conventional watermarking systems have shown themselves unable to effectively monitor and track media exposure. Furthermore, there is a need to integrate exposure to digital media across platforms where the digital media includes formats that are not traditionally subject to audio encoding. Moreover, there is a need in the art to properly “mask” such signals using environmental sounds and/or sounds native to a device that is conducting beaconing processes.
  • SUMMARY
  • Accordingly, an audio beacon system, apparatus and method is disclosed for collecting information on a panelist's exposure to media. Under a preferred embodiment, the audio beacon is configured as on-device encoding technology that is operative in a panelist's processing device (e.g., cell phone, PDA, PC) to enable the device to acoustically transmit user/panelist data for a predetermined period of time. The acoustically transmitted data is received and processed by a portable audience measurement device, such as Arbitron's Personal People Meter™ (“PPM”) or specially equipped cell phone, laptop etc., to enable audience measurement systems to achieve higher levels of detail on panel member activity and greater association of measurement devices to their respective panelists. Additionally, the acoustic transmissions are configured to utilize environmental sounds that are advantageous in being less obtrusive to users.
  • Additional features and advantages of the various aspects of the present disclosure will become apparent from the following description of the preferred embodiments, which description should be taken in conjunction with the accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1A is a block diagram illustrating a portion of an audio beaconing system under one exemplary embodiment;
  • FIG. 1B is a block diagram illustrating another portion of an audio beaconing system under the embodiment illustrated in FIG. 1A;
  • FIG. 2 is a tabular illustration of an audio beaconing and audio matching process under another exemplary embodiment;
  • FIG. 3 illustrates a block diagram of a server-side encoding process under yet another exemplary embodiment;
  • FIG. 4 illustrates an exemplary watermarking process for a digital media file suitable for use in the embodiment of FIGS. 1A-B;
  • FIG. 5 illustrates a block diagram of a client-side encoding process under yet another exemplary embodiment;
  • FIG. 6 illustrates an exemplary audio waveform having encoded data therein; and
  • FIG. 7 illustrates an exemplary device configured to select environmental sounds for audio transmission.
  • DETAILED DESCRIPTION
  • FIG. 1A is an exemplary block diagram illustrating a portion of an audio beaconing system 150 under one embodiment, where a web page 110 is provided by a page developer and published on content server 100. The web page preferably contains an embedded video player 111 and audio player 112 (that is preferably not visible), together with an application programming interface (API) 113. Other content 114 (e.g., HTML, text, etc.) is also provided on web page 110, which may or may not be coupled through API 113. API 113 is preferably embodied as a set of routines, data structures, object classes and/or protocols provided by libraries and/or operating system services in order to support the video player 111 and audio player 112. Additionally, the API 113 may be language-dependent (i.e. available only in a particular programming language) or language-independent (i.e., can be called from several programming languages, preferably an assembly/C-level interface). Examples of suitable API's include Windows API, Java Platform API, OpenGL, DirectX, Simple DirectMedia Layer (SDL), YouTube API, Facebook API and iPhone API, among others.
  • In one preferred embodiment, API 113 is configured as a beaconing API object. Depending on the features desired, the API object may reside on an Audience Measurement (AM) server 120, so that the object may be remotely initialized, thus minimizing the objects software's exposure to possible tampering and to maintain security. Alternately, the API object can reside on the content server 100, where the API object may be initialized under increased performance conditions.
  • When initialized, API 113 can communicate the following properties: (1) the URL of the page playing the media, (2) URL of the media being served on the page, (3) any statically available media metadata, and (3) a timestamp. It is understood that additional properties may be communicated in API 113 as well. In one configuration of FIG. 1 A an initialization request is received by API 113, to create a code tone that is preferably unique for each website and encode it on a small inaudible audio stream. Alternatively, the AM server 120 could generate a pre-encoded audio clip 101, with a code tone, for each site and forward it on the content server 100 in advance.
  • The encoded audio stream would then travel from content server 100 to the web page 110 holding audio player 110. In a preferred embodiment, audio player 110 may be set by the page developer as an object instance, where the visible property of player 110 is oriented as “false” or set to a one-by-one dimension in order to minimize the visual interference of the audio player with the web page. The encoded audio stream may then be played out in parallel with the media content being received from the web page 110. The encoded audio stream would preferably repeat at predetermined time periods through an on-device beacon 131 resident on a user device 130 as long as the user is on the same website. The beacon 131, would enable device 130 to acoustically transmit the encoded audio stream so that a suitably configured portable device 140 (e.g., PPM) can receive and process the encoded information. Beacon 131 could be embedded into an audio player resident on a web page being viewed inside the browser on user device 130, or may be a stand-alone application on user device 130.
  • A simplified example further illustrates the operation of the system 150 of FIGS. 1A-B under an alternate embodiment. User device 130 requests content (e.g., http://www.hulu.com/) from server 100. When the content is received in user device 130, PC meter software 132 collects and transmits web measurement data to Internet measurement database 141. One example of a PC meter is comScore's Media Metrix™ software; further exemplary processes of web metering may be found in U.S. Pat. No. 7,493,655, titled “Systems for and methods of placing user identification in the header of data packets usable in user demographic reporting and collecting usage data” and U.S. Pat. No. 7,260,837, titled “Systems and methods for user identification, user demographic reporting and collecting usage data usage biometrics”, both of which are incorporated by reference in their entirety herein.
  • As web measurement data is collected by PC meter 132, beacon 131 acoustically transmits encoded audio, which is received by portable device 140. In the exemplary embodiment, the encoding for the beacon transmission may include data such as a timestamp, portable device ID, user device ID, household ID, or any similar information. In addition to the beacon data, portable device 140 additionally receives multimedia data such as television and radio transmissions 142, which may or may not be encoded, at different times. If encoded (e.g., CBET encoding), portable device can forward transmissions 142 to audio matching server 160 (FIG. 1B) for decoding and matching with audio matching database 161. If transmissions 142 are not encoded, portable device 140 may employ sampling techniques for creating audio patterns or signatures, which may also be transmitted to audio matching server 160 for pattern matching using techniques known in the art.
  • Audio beacon server 150, shown in FIG. 1B, receives and processes/decodes beacon data from portable device 140. Under an alternate embodiment, it is possible to combine audio matching server 160 and audio beacon server 150 to collectively process both types of data. Data from Audio beacon server 150 and audio matching server 160 is transmitted to Internet measurement database 141, where the web measurement data could be combined with audio beacon data and data from the audio matching server to provide a comprehensive collection of panelist media exposure data.
  • Under another exemplary embodiment, the video and audio players of webpage 110 are configured to operate as Flash Video, which is a file format used to deliver video over the Internet using Adobe™ Flash Player. The Flash Player typically executes Shockwave Flash “SWF” files and has support for a scripting language called ActionScript, which can be used to display Flash Video from an SWF file. Because the Flash Player runs as a browser plug-in, it is possible to embed Flash Video in web pages and view the video within a web browser. Commonly, Flash Video files contain video bit streams which are a variant of the H.263 video standard, and include support for H.264 video standard (i.e., “MPEG-4 part 10”, or “AVC”). Audio in Flash Video files (“FLV”) is usually encoded as MP3, but can also accommodate uncompressed audio or ADPCM format audio.
  • Continuing with the embodiment, video beacons can be embedded within an action script that will be running within the video Flash Player's run time environment on web page 110. When an action script associated with web page 110 gets loaded as a result of the access to the page, the script gets activated and triggers a “video beacon”, which extracts and store URL information on a server (e.g., content server 100), and launches the video Flash Player. By inserting an audio beacon in the same action script, the audio beacon will be triggered by the video player. Once triggered, the audio beacon may access AM server 120 to load a pre-recorded audio file containing a special embedded compatible code (e.g., CBET). This pre-recoiled audio file would be utilized for beacon 131 to transmit for a given period of time (e.g., every x seconds).
  • As a result, the beacon 131 audio player runs as a “shadow player” in parallel to the video Flash Player. If a portable device 140 is in proximity to user device 130, portable device 140 will detect the code and reports it to audio beacon server 150. Depending on the level of cooperation between the audio and video beacon, the URL information can also be deposited onto beacon server 150 along with codes that would allow an audience measurement entity to correlate and/or calibrate various measurements with demographic data.
  • Under the present disclosure, media data may be processed in a myriad of ways for conducting customized panel research. As an example, each user device 130 may install on-device measurement software (PC meter 132) which includes one or more web activity monitoring applications, as well as beacon software 131. It is understood that the web activity monitoring application and the beacon software may be individual applications, or may be merged into a single application.
  • The web activity monitoring application collects web activities data from the user device 130 (e.g., site ID, video page URL, video file URL, start and end timestamp and any additional metadata about videosite information, URL information, time, etc.) and additionally assigns a unique ID, such as a globally unique identifier or “GUID”, to each device. For the beacon 131, a unique composite ID may be assigned including a household ID (“HHID”) and a unique user device ID for each device in the household up to 10 devices for a family), as well as a portable device ID (PPMID). Panelist demographic data may be included for each web activity on the device.
  • Continuing with the example, beacon 131 emits an audio beacon code (ABC) for device in the household by encoding an assigned device ID number and acoustically sending it to portable device 140 to identify the device. Further details on the encoding is provided below. Portable device 140 collects the device ID and sends it to a database along with HHID and/or PPM ID and the timestamp. Preferably, a PPMID is always mapped to a in the backend; alternately an HHID can be set within each PPMID.
  • The web activity monitoring and beacon applications may pass information to each other as needed. Both can upload information to a designated server for additional processing. A directory of panelists' devices is built to contain the GUID, HHID, and device ID for panel, and the directory could be used to correlate panelist demographic data and web measurement data.
  • Turning to FIG. 2, a tabular illustration of an audio beaconing and audio matching process under another an exemplary embodiment is provided. Specifically, the table illustrates a combination of audio beaconing and audio matching and its application to track a video on a content site, such as Hulu.com. FIG. 2. Timeline 200 shows in sections a scenario where a user/panelist plays a ten minute video on Hulu.com. Activities 201 shows actions taken in user system 150 where a video is loaded in the user device 130, and played. At the 5 minute mark (301 sec.) a 15 second advertisement is served. At the conclusion of the advertisement (316 sec.), the video continues to play until its conclusion (600 sec.).
  • During this time, audio beacon activities 202 are illustrated, where, under one embodiment, on-device beacon 131 transmits continuous audio representing the website (Hulu.com). In addition, beacon also transmits a timestamp, portable device ID, user device ID, household ID and/or any other data in accordance with the techniques described above. Under an alternate embodiment shown in 203, additional data may be transmitted in the beacon to include URLs and video ID's when a video is loaded and played. As the advertisement is served, an event beacon, which may include advertisement URL data, is transmitted. At the conclusion of the video, a video end beacon is transmitted to indicate the user/panelist is no longer viewing specific media.
  • When the video and advertisement is loaded and played, additional audio matching may occur in the portable device 140, in addition with audio matching processes explained above in relation to FIGS. 1A-B. Referring to audio matching events 204, portable device data 205 and end-user experience 206 of FIG. 2, portable device data (e.g., demographic ID data) is overlayed along with site information (URL, video ID, etc.) when a video is loaded. When the video is played, audio signatures may be sampled periodically by portable device 140, until a content match is achieved. The audio signatures may be obtained through encoding, pattern matching, or any other suitable technique. When a match is found, portable device data is overlayed to indicate that a content match exists. Further signature samples are taken to ensure that the same content is being viewed. When an advertisement is served, the sampled signature will indicate that different content is being viewed, at which point the portable device data is overlayed in the system. When the video resumes, the audio signature indicates the same video is played, and portable device data is overlayed through the end of the video as shown in FIG. 2.
  • As explained above, signature sampling/audio matching allows the system 150 to identify and incorporate additional data on the users/panelists and the content being viewed. Under a typical configuration, the content provider media (e.g., Hulu, Facebook, etc.) may be sampled in advance to establish respective signatures for content and stored in a matching database (e.g., audio matching server 160). The portable device 140 would be equipped with audio matching software, so that, when a panelist is in the vicinity of user device 130, audio matching techniques are used to collect the signature, or “audio fingerprint” for the incoming stream. The signatures would then be matched against the signatures in the matching database to identify the content.
  • It is understood by those skilled in the art however, that encoding techniques may also be employed to identify content data. Under such a configuration, content is encoded prior to transmission to include data relating to the content itself and the originating content site. Additionally, data relating to possible referral sites (e.g., Facebook, MySpace, etc.) may be included. Under one embodiment, a content management system may be arranged for content distributors to choose specific files for a corresponding referral site.
  • For the media data encoding, several advantageous and suitable techniques for encoding audience measurement data in audio data are disclosed in U.S. Pat. No. 5,764,763 to James M. Jensen, et al., which is assigned to the assignee of the present application, and which is incorporated by reference herein. Other appropriate encoding techniques are disclosed in U.S. Pat. No. 5,579,124 to Aijala, et al., U.S. Pat. Nos. 5,574,962, 5,581,800 and 5,787,334 to Fardeau, et al., U.S. Pat. No. 5,450,490 to Jensen, et al., and U.S. patent application Ser. No. 09/318,045, in the names of Neuhauser, et al., each of which is assigned to the assignee of the present application and all of which are incorporated by reference in their entirety herein.
  • Still other suitable encoding techniques are the subject of PCI Publication WO 00/04662 to Srinivasan, U.S. Pat. No. 5,319,735 to Preuss, et al., U.S. Pat. No. 6,175,627 to Petrovich, et al., U.S. Pat. No. 5,828,325 to Wolosewicz, et al., U.S. Pat. No. 6,154,484 to Lee, et al., U.S. Pat. No. 5,945,932 to Smith, et al., PCI Publication WO 99/59275 to Lu, et al., PCT Publication WO 98/26529 to Lu, et al., and PCT Publication WO 96/27264 to Lu, et al, all of which are incorporated by reference in their entirety herein.
  • Variations on the encoding techniques described above are also possible. Under one embodiment, the encoder may be based on a Streaming Audio Encoding System (SAES) that operates under a set of sample rates and is integrated with media transcoding automation technology, such as Telestream's FlipFactory™ software. Also, the encoder may be embodied as a console mode application, written in a general-purpose computer programming language such as “C”. Alternately, the encoder may be implemented as a Java Native Interface (JNI) to allow code running in a virtual machine to call and be called by native applications, where the JNI would include a JNI shared library for control using Java classes. The encoder payloads would be configured using specially written Java classes. Under this embodiment, the encoder would use the information hiding abstractions of an encoder payload which defines a single message. Under a preferred embodiment, the JNI encoder would operate using a 44.1 kHz sample rate.
  • Examples of symbol configurations and message structures are provided below. One exemplary symbol configuration uses four data symbols and one end symbol defined for a total of five symbols. Each symbol may comprise five tones, with one tone coming from each of five standard Barks. One exemplary illustration of Bark scale edges (in Hertz), would be {920, 1080, 1270, 1480, 1720, 2000}. The bins are preferably spaced on a 4×3.90625 grid in order to provide lighter processing demands, particularly in cases using decoders based on 512 point fast Fourier transform (FFT). an exemplary bin structure is provided below:
      • Symbol 0: {248, 292, 344, 400, 468}
      • Symbol 1: {252, 296, 348, 404, 472}
      • Symbol 2: {256, 300, 352, 408, 476}
      • Symbol 3: {260, 304, 356, 412, 480}
      • End Marker Symbol: {264, 308, 360, 416, 484}
  • Regarding message structure, an exemplary message would comprise 20 symbols, each being 400 milliseconds in duration, for a total duration of 8 seconds. Under this embodiment, the first 3 symbols could be designated as match/check criteria symbols, which could be the simple sum of the data symbols or could be derived from an error correction or cyclical redundancy check algorithm. The following 16 symbols would then he designated as data symbols, leaving the last symbol as an end symbol used for a marker. Under this configuration, the total number of possible symbols would be 416 or 4,294,967,296 symbols.
  • Variations in the algorithmic process for encoding are possible as well under the present disclosure. For example, a core sampling rate of 5.5125 kHz may be used instead of 8 kHz to allow down-sampling from 44.1 kHz to be efficiently performed without pre-filter (to eliminate aliasing components) followed by conversion filter to 48 kHz. Such a configuration should have no effect on code tone grid spacing since the output frequency generation is independent of the core sampling rate. Additionally, this configuration would limit the top end of the usable frequency span to about 2 kHz (as opposed to 3 kHz under conventional techniques) since frequency space should be left for filters with practical numbers of taps.
  • Under one embodiment, a 16 point overlap of a 256 point large FFT is used, resulting in amplitude updates every 2.9 milliseconds for encoding instead of every 2 milliseconds for standard CBET techniques. Accordingly, fewer large FFTs are calculated under a tighter bin resolution of 21.5 Hz instead of 31.25 Hz.
  • The psychoacoustic model calculations used for the encoding algorithm under the present disclosure may vary from traditional techniques as well. In one embodiment, bin spans of the clumps may be set by Bark boundaries instead of being wholly based on Critical Bandwidth criteria. By using Bark boundaries, a specific bin will not contribute to the encoding power level of multiple clumps, which provides less coupling between code amplitudes of adjacent clumps. When producing Equivalent Large FFTs, a comparison may he made of the most recent 16 point Small FFT results to a history of squared sums to simplify calculations.
  • For noise power computation, the encoding algorithm under the present disclosure would preferably use 3 bin values over a clump: the minimum bin power (MIN), the maximum bin power (MAX), and the average bin power (AVG). Under this arrangement, the bin values could be modeled as follows:
  • IF (MAX > (2 * MIN))
     PWR = MIN
    ELSE
     PWR = AVG

    Here, PWR may be scaled by a predetermined factor to produce masking energy.
  • A similar algorithm could also be used to create a 48 kHz native encoder using a core sample rate of 6 kHz and a large FFT bin resolution of 23.4375 Hz calculated every 2.67 milliseconds. Such a configuration would differ slightly in detection efficiency and inaudibility from the embodiments described above, but it is anticipated that the differences would be slight.
  • With regards to decoding, an exemplary configuration would include a software decoder based on a JNI shared library, which performs calculations up through the bin signal-to-noise ratios. Such a configuration would allow an external application to define the symbols and perform pattern matching. Such steps would be handled in a Java environment using an information hiding extraction of a decoder payload, where decoder payloads are created using specially written Java classes.
  • Turning to FIG. 3, an exemplary server-side encoding embodiment is illustrated. In this example, content server 100 has content 320, which includes a media file 302 configured to be requested and played on media player 301 residing on user device 130. When media file 302 is initialized, audio is extracted from the media file and, if the audio is encoded (e.g., MP3 audio), subjected to audio decoding in 304 to produce raw audio 305. To encode the audio for beaconing, device ID, HHID anchor PPMID data is provided for first encoding 306 the data into the raw audio 305, using any suitable technique described above.
  • After the first encoding, the audio data is then subjected to a second encoding to transform the audio into a suitable format (e.g., MP3) to produce fully encoded audio 308, which is subsequently transmitted to media player 301 and beaconed to portable device 140. Alternately, encoded audio 308 may be produced in advance and stored as part of media file 302. During the encoding process illustrated in FIG. 3, care must be taken to account for processing delays to ensure that the encoded audio is properly synchronized with any video content in media file 302.
  • The server-side encoding may be implemented under a number of different options. A first option would be to implement a pre-encoded beacon, where the encoder (306) would be configured to perform real-time encoding of the audio beacon based on the content being served to the users/panelists. The user device would be equipped with a software decoder as described above which is invoked when media is played. The pre-encoded beacon would establish a message link which could be used, along with an identifier from the capturing portable device 140, in order to assign credit. The encoding shared library would preferably be resident at the content site (100) as part of the encoding engine. Such a configuration would allow the transcoding and encoding to be fit into the content site workflow.
  • Another option for server-side encoding could include a pre-encoded data load, where the audio is encoded with a message that is based on the metadata or the assigned URL. This establishes a message link which can be used, along with an identifier from the capturing portable device 140, in order to assign credit. The encoding shared library is preferably resident at the content site (100), as part of the encoding engine. Again, this configuration would allow the transcoding and encoding to be fit into the content site workflow.
  • Yet another option for server-side encoding could include “on-the-fly” encoding. If a video is being streamed to a panelist, encoding may be inserted in the stream along with a transcoding object. The encoding may be used to encode the audio with a simple one of N beacon, and the panelist user device 130 would contain software decoding which is invoked when the video is played. This also establishes a message link which can be used, along with an identifier from the capturing portable device 140, in order to assign credit. The encoding shared library is preferably resident at the content site (100), as part of the encoding engine. Under a preferred embodiment, an ActionScript would invoke the decoding along with a suitable transcoding object.
  • FIG. 4 illustrates one embodiment for encoding media under a Flash Video platform 410, where the content is preferably encoded in advance. As raw audio from a video file or other source 400 is received, the audio is subjected to water mark encoding 401, which may include techniques described above for the encoding. Once encoded, the audio is formatted as a Flash file using Adobe Tools 402 such as FLV Creator and SWF Compiler. Once compiled, the file is further formatted using Flash-supported codecs (e.g., H.264, VP6, MPEG-4 ASP, Sorenson H.263) and compression 403 to produce a watermarked A/V stream or file 404.
  • FIG. 5 provides another alternate embodiment that illustrates client-side encoding and processing. In this example, user device 130 requests media data. In response to the request, a media file 531 residing on content server 100 is subsequently streamed to the device's browser 520 arranged on user's workspace 510. Media player 521 plays the streamed content and produces raw audio 511. A client-side ActionScript notifies browser 522 and encoder 522 to capture the raw audio on the device's sound mixer, or microphone (not shown), and to encode data using a suitable encoding technique described above. The encoding constructs the data for an independent audio beacon using the captured audio and other data (e.g., device ID, HHID, etc.) where portable device 140 picks up the beacon and forwards the data to an appropriate server for further processing and panel data evaluation.
  • Similar to the server-side embodiment disclosed in FIG. 3, care must be taken in the software to account for processing delays in audio pickup and (CBET) encoding of the audio beacon. Preferably, synchronization between audio beacon playback and audio playback (specifically FLV playback) should be accounted for. In alternate embodiments, communication between media player 521 and encoder 522 could be through Actionscript interface APIs, such as “ExternalInterface”, which is an application programming interface that enables straightforward communication between ActionScript and a Flash Player container; for example, an HTML page with JavaScript, or a desktop application with Flash Player embedded, along with encoder application 522. To get information on the container application, an ActionScript interface could be used to call code in the container application, including a web page or desktop application. Additionally, ActionScript code could be called from code in the container application. Also, a proxy could be created to simplify calling ActionScript code from the container application.
  • For the panel-side encoding, a beacon embodiment may be enabled by having an encoding message being one from a relatively small set (e.g., 1 of 12), and where each user device 130 is assigned a different message. When portable device 140 detects the encoded message, it identifies the user device 130. Alternately, the encoding message may be a hash of the site and/or URL information gleaned from the metadata. When a panelist portable device 140 detects and reports the encoded message, a reverse hash can be used to identify the site, where the hash could be resolved on one or more remote server (e.g., sever 160).
  • In addition to the encoding techniques described above in connection with media content, a simplified beaconing configuration may be arranged where the beacon operates as a complement to media data, independent of the media data, or providing a beacon where no specially encoded data exists. Referred to herein as a “twinkle,” the simplified beaconing comprises a constant amplitude acoustic signal or tone that is generated on user device 130. This acoustic tone is then automatically encoded, preferably with identification data (e.g., device ID, HHID and/or PPM ID) and a timestamp. The encoded acoustic tone would then be forwarded to portable device 140 for processing and identification.
  • The acoustic tone used for the twinkle is preferably embodied as a pre-recorded constant amplitude tone that is transmitted at predetermined times. The encoding is preferably performed using any of the techniques described above. Under one embodiment, the simplified beaconing process would only forward the encoded, pre-recorded tone, independently of any audio data being received. Thus, referring back to FIG. 1, it is possible that user device 130 receives only other content 114 from content server 110 in the form of text-based HTML. As PC meter 132 records browsing information, the encoded tone is transmitted to portable device 140, where after further processing (see FIG. 1B), the user identification data is merged into internet measurement database 141. It is understood, that user device 140 may also receive audio data (encoded or unencoded) separately and in addition to other content. While the techniques described above would encode and forward audio data received, the simplified beacon (“twinkle”) would also transmit ID information to portable device 140, which, in conjunction with PC meter 132, would subsequently merge panelist data into a common database.
  • In another exemplary embodiment, FIG. 6 illustrates audio signal 600 represented as a spectrum of audio 610 over a period of time (e.g., 0.25 seconds), where the energy intervals vary with frequency between 1200 and 2200 Hz. Overlaid in black are discreet, narrowband code tones 602 (e.g., CBET) opportunistically inserted into the audio using the principles of psychoacoustic masking. For encoded tones, one of which is illustrated as 602 in FIG. 6, the energy of the inserted code tone varies with the level of audio, so more quiet portions of the frequency spectrum (e.g., 604) receive little encoding energy and compared to louder portions 605), which get proportionally more.
  • In contrast, the simplified encoding (“twinkle”) 603 is encoded and inserted at constant levels across the frequency spectrum, where the levels are independent of the audio levels. This allows the simplified encoding to be pre-recorded, easily generated and capable of being reused accross various and/or different content. The simplified encoding could have the same message structure as the CBET encoding described above, utilizing a 10-tone symbol set. Alternately, other message structures are possible as well. As mentioned above, the twinkle may be transmitted automatically at regular intervals. Alternately the twinkle may be invoked by an ActionScript. If two players are utilized (i.e., one for the media, and one for the twinkle), the ActionScript could relay a beacon for the media from user device 130 to portable device 140, while simultaneously requesting a second (preferably invisible) Flash Player in the user device 130 to transmit the twinkle to portable device 140. Under a preferred embodiment, the ActionScript should invoke both players at a common volume setting.
  • In certain embodiments, it is advantageous to configure the audio beacon or “twinkle” on a device so that it is not intrusive and/or distracting to the user. Additionally, the beacon may be arranged to have audio characteristics that make it easier and more robust to encode. For example, simulated environmental sounds, such as power supply fans, air vent exhaust, crowd/audience noise, ocean waves and such may be used as the audio beacon sound in which encoded messages may be inserted. In another example, it is known that many computing devices, such as personal computers, phones, tablets and laptops, contain a pre-stored library of audio that is used to alert or notify users. By creating and storing sounds advantageously suited to encode data in the on-device sound library, a more robust beacon may be utilized. Furthermore, since the sound used for the beacon is known a priori, the encoding on the device may be simplified, since the time, frequency, masking and other encoding processes will be known.
  • Turning to FIG. 7, an exemplary embodiment is illustrated where user device 130 is configured with a control panel 700 that allows the device to control various aspects, including sound 701 that includes a resident sound library 720. In a preferred embodiment, a script, operating alone or as part of meter software (see FIG. 1A, ref. 132), may control aspects of sound library 720, such that specific sounds are used for audio beacon encoding. Of course, it is possible for a user to manually change these features either directly through user device 130, or through a remote connection.
  • Sound library 720 comprises audio sounds (704-707) that are associated with one or more software applications 700-701 and/or events 702-703. “Events,” for the purposes of FIG. 7, include notifications (e.g., receipt of email, social networking software status update, etc.) and alerts (e.g., start-up, shut-down, application error, etc.) that are made pursuant to the device's operating system and associated software running on the device. In this example, a first application (APP1) 700 and a second application (APP2) 701 are associated with respective sounds 704-705 that may be triggered when the application is activated, or at predetermined times determined by the specific application. Applications 700-701 may include metering software 132 discussed above in connection with FIG. 1A.
  • In one embodiment, each sound 704, 705 is associated with a respective application 700, 701. The sounds (704-705) are used for encoding messages to form respective beacons (708-709). As data for the audio beacon is being collected on user device 130 (e.g., user device ID, web hash, etc.), the sound (e.g., 704) is copied, data is encoded into it and the encoded copy is stored in a buffer or other suitable memory. As further data for an audio beacon is collected, another copy of the original sound (704) is made, and the further data is encoded and stored into the buffer/memory. This process repeats for as long as necessary to form a string of encoded sounds. The sounds may be arranged sequentially or in other suitable formats. When an application 700 triggers a sound, a first beacon 708 is audibly transmitted. When the next trigger occurs, the next beacon is audibly transmitted, and so on, until the buffer/memory is empty, or a predetermined amount of time has expired. In another embodiment, one application can control a plurality of sounds (704-705) and produce a plurality of encoded beacons (708-709).
  • Continuing with FIG. 7, events 702-703 may be associated with respective sounds 706-707, similar to applications 700-701 discussed above, to produce beacons 710-711. This configuration may be particularly advantageous to beacon information when a user performs an act on the device, such as open/close a browser window, open/close a tab on a browser, open/close an application, etc. As data for the audio beacon is being collected (e.g., user device ID, web hash, etc.), the sound (e.g., 706) is copied, data is encoded into it and the encoded copy is stored in a buffer or other suitable memory. As further data for an audio beacon is collected, another copy of the original sound (706) is made, and the further data is encoded and stored into the buffer/memory. This process repeats for as long as necessary to form a string of encoded sounds. Again, the sounds may be arranged sequentially or in other suitable formats. When an event (702) is detected, a first beacon 710 is audibly transmitted. When the next event occurs, the next beacon is audibly transmitted, and so on, until the buffer/memory is empty, or a predetermined amount of time has expired.
  • As mentioned previously, the sounds 704-707 are preferably predetermined and may simulate an environmental sound so as not to be intrusive or distracting to the user. Additionally, the sound may be selected to contain audio characteristics (e.g., having high masking levels in critical frequency bands) that makes it conducive to robust audio encoding. By using a predetermined sound for the audio beacon encoding, designers can have more flexibility in audibly beaconing data. In addition to audio characteristics, the predetermined sounds may have different lengths as well. In an example where an application (700) controls multiple sounds, the sounds may be the same instance of one sound, but having different lengths (e.g., 5 sec., 10 sec., etc.). In cases where a device's (130) volume is lower, the application may default to a longer sound to increase the probability of the beacon code being detected. If the volume increases, the device 130 can switch to a shorter sound. This configuration has the added benefit of ensuring that users maintain a sufficient volume on their device to avoid longer (and possibly more intrusive) beacons.
  • Various embodiments disclosed herein provide devices, systems and methods for performing various functions using an audience measurement system that includes audio beaconing. Although specific embodiments are described herein, those skilled in the art recognize that other embodiments may be substituted for the specific embodiments shown to achieve the same purpose. As an example, although terms like “portable” are used to describe different components, it is understood that other, fixed, devices may perform the same or equivalent functions. Also, while specific communication protocols are mentioned in this document, one skilled in the art would appreciate that other protocols may be used or substituted. This application covers any adaptations or variations of the present invention. Therefore, the present invention is limited only by the claims and all available equivalents.

Claims (21)

1-18. (Canceled)
19. A computing device comprising:
memory including computer readable instructions; and
a processor to execute the instructions to at least:
select, based on a volume setting of the computing device, one of a plurality of sound signals stored at the computing device, the plurality of sound signals having different lengths;
encode a digital message in the selected one of the plurality of sound signals to form an encoded sound signal; and
cause the encoded sound signal to be audibly output from the computing device in response to a trigger.
20. The computing device of claim 19, wherein the processor is further to collect data to include in the digital message, the data to include at least one of an identifier associated the computing device or hash of a data associated with a web site accessed by the computing device.
21. The computing device of claim 19, wherein to select the one of the plurality of sound signals, the processor is to:
select a first one of the plurality of sound signals having a first length when the volume setting of the computing device is a first volume setting; and
select a second one of the plurality of sound signals having a second length that is longer than the first length when the volume setting of the computing device is a second volume setting that is lower than the first volume setting.
22. The computing device of claim 19, wherein the processor is further to:
store the encoded sound signal in a buffer including a plurality of encoded sound signals encoded with respective digital messages; and
cause respective ones of the plurality of encoded sound signals included in the buffer to be audibly output sequentially from the computing device in response to a sequence of triggers.
23. The computing device of claim 22, wherein the processor is to continue to cause respective ones of the plurality of encoded sound signals included in the buffer to be audibly output sequentially from the computing device in response to the sequence of triggers until at least one of the buffer is empty or an amount of time has expired.
24. The computing device of claim 19, wherein the plurality of sound signals is associated with an application to execute on the computing device, and the application is to trigger the processor to cause the encoded sound signal to be audibly output from the computing device.
25. The computing device of claim 19, wherein the plurality of sound signals is associated with an operating system event of the computing device, and occurrence of the operating system event is to trigger the processor to cause the encoded sound signal to be audibly output from the computing device.
26. A computer readable storage device comprising computer readable instructions that, when executed by a processor of a computing device, cause the processor to at least:
select, based on a volume setting of the computing device, one of a plurality of sound signals stored at the computing device, the plurality of sound signals having different lengths;
encode a digital message in the selected one of the plurality of sound signals to form an encoded sound signal; and
cause the encoded sound signal to be audibly output from the computing device in response to a trigger.
27. The storage device of claim 26, wherein the instructions further cause the processor to collect data to include in the digital message, the data to include at least one of an identifier associated the computing device or hash of a data associated with a web site accessed by the computing device.
28. The storage device of claim 26, wherein to select the one of the plurality of sound signals, the instructions cause the processor to:
select a first one of the plurality of sound signals having a first length when the volume setting of the computing device is a first volume setting; and
select a second one of the plurality of sound signals having a second length that is longer than the first length when the volume setting of the computing device is a second volume setting that is lower than the first volume setting.
29. The storage device of claim 26, wherein the instructions further cause the processor to:
store the encoded sound signal in a buffer including a plurality of encoded sound signals encoded with respective digital messages; and
cause respective ones of the plurality of encoded sound signals included in the buffer to be audibly output sequentially from the computing device in response to a sequence of triggers.
30. The storage device of claim 29, wherein the instructions cause the processor to continue to cause respective ones of the plurality of encoded sound signals included in the buffer to be audibly output sequentially from the computing device in response to the sequence of triggers until at least one of the buffer is empty or an amount of time has expired.
31. The storage device of claim 26, wherein the plurality of sound signals is associated with an application to execute on the computing device, and the application is to trigger the processor to cause the encoded sound signal to be audibly output from the computing device.
32. The storage device of claim 26, wherein the plurality of sound signals is associated with an operating system event of the computing device, and occurrence of the operating system event is to trigger the processor to cause the encoded sound signal to be audibly output from the computing device.
33. A method for a computing device, the method comprising:
selecting, by executing an instruction with a processor of the computing device and based on a volume setting of the computing device, one of a plurality of sound signals stored at the computing device, the plurality of sound signals having different lengths;
encoding, by executing an instruction with a processor, a digital message in the selected one of the plurality of sound signals to form an encoded sound signal; and
audibly outputting the encoded sound signal from the computing device in response to a trigger.
34. The method of claim 33, further including collecting data to include in the digital message, the data including at least one of an identifier associated the computing device or hash of a data associated with a web site accessed by the computing device.
35. The method of claim 33, wherein the selecting of the one of the plurality of sound signals includes:
selecting a first one of the plurality of sound signals having a first length when the volume setting of the computing device is a first volume setting; and
selecting a second one of the plurality of sound signals having a second length that is longer than the first length when the volume setting of the computing device is a second volume setting that is lower than the first volume setting.
36. The method of claim 33, further including:
storing the encoded sound signal in a buffer including a plurality of encoded sound signals encoded with respective digital messages; and
audibly outputting respective ones of the plurality of encoded sound signals from the buffer sequentially in response to a sequence of triggers until at least one of the buffer is empty or an amount of time has expired.
37. The method of claim 33, wherein the plurality of sound signals is associated with an application to execute on the computing device, and the application is to trigger the encoded sound signal to be audibly output from the computing device.
38. The method of claim 33, wherein the plurality of sound signals is associated with an operating system event of the computing device, and occurrence of the operating system event is to trigger the processor to cause the encoded sound signal to be audibly output from the computing device.
US16/002,775 2009-04-17 2018-06-07 System and method for utilizing audio encoding for measuring media exposure with environmental masking Abandoned US20190019521A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/002,775 US20190019521A1 (en) 2009-04-17 2018-06-07 System and method for utilizing audio encoding for measuring media exposure with environmental masking

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US12/425,556 US20100268573A1 (en) 2009-04-17 2009-04-17 System and method for utilizing supplemental audio beaconing in audience measurement
US12/425,464 US20100268540A1 (en) 2009-04-17 2009-04-17 System and method for utilizing audio beaconing in audience measurement
US13/338,588 US10008212B2 (en) 2009-04-17 2011-12-28 System and method for utilizing audio encoding for measuring media exposure with environmental masking
US16/002,775 US20190019521A1 (en) 2009-04-17 2018-06-07 System and method for utilizing audio encoding for measuring media exposure with environmental masking

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US13/338,588 Continuation US10008212B2 (en) 2009-04-17 2011-12-28 System and method for utilizing audio encoding for measuring media exposure with environmental masking

Publications (1)

Publication Number Publication Date
US20190019521A1 true US20190019521A1 (en) 2019-01-17

Family

ID=48698633

Family Applications (2)

Application Number Title Priority Date Filing Date
US13/338,588 Active 2032-04-17 US10008212B2 (en) 2009-04-17 2011-12-28 System and method for utilizing audio encoding for measuring media exposure with environmental masking
US16/002,775 Abandoned US20190019521A1 (en) 2009-04-17 2018-06-07 System and method for utilizing audio encoding for measuring media exposure with environmental masking

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US13/338,588 Active 2032-04-17 US10008212B2 (en) 2009-04-17 2011-12-28 System and method for utilizing audio encoding for measuring media exposure with environmental masking

Country Status (2)

Country Link
US (2) US10008212B2 (en)
WO (1) WO2013102016A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022150612A1 (en) * 2021-01-08 2022-07-14 The Nielsen Company (Us), Llc Engagement measurement of media consumers based on the acoustic environment

Families Citing this family (64)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101711388B (en) 2007-03-29 2016-04-27 神经焦点公司 The effect analysis of marketing and amusement
US8392253B2 (en) 2007-05-16 2013-03-05 The Nielsen Company (Us), Llc Neuro-physiology and neuro-behavioral based stimulus targeting system
US8533042B2 (en) 2007-07-30 2013-09-10 The Nielsen Company (Us), Llc Neuro-response stimulus and stimulus attribute resonance estimator
US8386313B2 (en) 2007-08-28 2013-02-26 The Nielsen Company (Us), Llc Stimulus placement system using subject neuro-response measurements
US8392255B2 (en) 2007-08-29 2013-03-05 The Nielsen Company (Us), Llc Content based selection and meta tagging of advertisement breaks
US9667365B2 (en) 2008-10-24 2017-05-30 The Nielsen Company (Us), Llc Methods and apparatus to perform audio watermarking and watermark detection and extraction
US8359205B2 (en) 2008-10-24 2013-01-22 The Nielsen Company (Us), Llc Methods and apparatus to perform audio watermarking and watermark detection and extraction
US20100250325A1 (en) 2009-03-24 2010-09-30 Neurofocus, Inc. Neurological profiles for market matching and stimulus presentation
US8655437B2 (en) 2009-08-21 2014-02-18 The Nielsen Company (Us), Llc Analysis of the mirror neuron system for evaluation of stimulus
US10987015B2 (en) 2009-08-24 2021-04-27 Nielsen Consumer Llc Dry electrodes for electroencephalography
US20110106750A1 (en) 2009-10-29 2011-05-05 Neurofocus, Inc. Generating ratings predictions using neuro-response data
US8209224B2 (en) 2009-10-29 2012-06-26 The Nielsen Company (Us), Llc Intracluster content management using neuro-response priming data
US9560984B2 (en) 2009-10-29 2017-02-07 The Nielsen Company (Us), Llc Analysis of controlled and automatic attention for introduction of stimulus material
US8769614B1 (en) * 2009-12-29 2014-07-01 Akamai Technologies, Inc. Security framework for HTTP streaming architecture
US8684742B2 (en) 2010-04-19 2014-04-01 Innerscope Research, Inc. Short imagery task (SIT) research method
US8655428B2 (en) 2010-05-12 2014-02-18 The Nielsen Company (Us), Llc Neuro-response data synchronization
US8307006B2 (en) 2010-06-30 2012-11-06 The Nielsen Company (Us), Llc Methods and apparatus to obtain anonymous audience measurement data from network server data for particular demographic and usage profiles
US8392250B2 (en) 2010-08-09 2013-03-05 The Nielsen Company (Us), Llc Neuro-response evaluated stimulus in virtual reality environments
US8392251B2 (en) 2010-08-09 2013-03-05 The Nielsen Company (Us), Llc Location aware presentation of stimulus material
US8396744B2 (en) 2010-08-25 2013-03-12 The Nielsen Company (Us), Llc Effective virtual reality environments for presentation of marketing materials
CN103119565B (en) 2010-09-22 2016-05-11 尼尔森(美国)有限公司 Utilize distributed demographics information to determine the method and apparatus of impression
AU2011349435B2 (en) 2010-12-20 2016-08-18 The Nielsen Company (Us), Llc Methods and apparatus to determine media impressions using distributed demographic information
US8538333B2 (en) 2011-12-16 2013-09-17 Arbitron Inc. Media exposure linking utilizing bluetooth signal characteristics
US9569986B2 (en) 2012-02-27 2017-02-14 The Nielsen Company (Us), Llc System and method for gathering and analyzing biometric user feedback for use in social media and advertising applications
AU2013204865B2 (en) 2012-06-11 2015-07-09 The Nielsen Company (Us), Llc Methods and apparatus to share online media impressions data
US8744995B1 (en) 2012-07-30 2014-06-03 Google Inc. Alias disambiguation
US8520807B1 (en) 2012-08-10 2013-08-27 Google Inc. Phonetically unique communication identifiers
US8583750B1 (en) 2012-08-10 2013-11-12 Google Inc. Inferring identity of intended communication recipient
US8571865B1 (en) * 2012-08-10 2013-10-29 Google Inc. Inference-aided speaker recognition
US8989835B2 (en) 2012-08-17 2015-03-24 The Nielsen Company (Us), Llc Systems and methods to gather and analyze electroencephalographic data
AU2013204953B2 (en) 2012-08-30 2016-09-08 The Nielsen Company (Us), Llc Methods and apparatus to collect distributed user information for media impressions
US9224292B2 (en) * 2012-09-21 2015-12-29 Kerry L. Davis Method for controlling a computing device over existing broadcast media acoustic channels
US9609110B2 (en) * 2013-01-22 2017-03-28 Mobius Connective Technologies, Ltd. Apparatus for bi-directional communication with medical and wellness devices
US9099080B2 (en) 2013-02-06 2015-08-04 Muzak Llc System for targeting location-based communications
US9320450B2 (en) 2013-03-14 2016-04-26 The Nielsen Company (Us), Llc Methods and apparatus to gather and analyze electroencephalographic data
TWI557727B (en) 2013-04-05 2016-11-11 杜比國際公司 An audio processing system, a multimedia processing system, a method of processing an audio bitstream and a computer program product
US9697533B2 (en) 2013-04-17 2017-07-04 The Nielsen Company (Us), Llc Methods and apparatus to monitor media presentations
US9519914B2 (en) 2013-04-30 2016-12-13 The Nielsen Company (Us), Llc Methods and apparatus to determine ratings information for online media presentations
US10068246B2 (en) 2013-07-12 2018-09-04 The Nielsen Company (Us), Llc Methods and apparatus to collect distributed user information for media impressions
US9711152B2 (en) 2013-07-31 2017-07-18 The Nielsen Company (Us), Llc Systems apparatus and methods for encoding/decoding persistent universal media codes to encoded audio
US20150039321A1 (en) * 2013-07-31 2015-02-05 Arbitron Inc. Apparatus, System and Method for Reading Codes From Digital Audio on a Processing Device
US9313294B2 (en) 2013-08-12 2016-04-12 The Nielsen Company (Us), Llc Methods and apparatus to de-duplicate impression information
US10333882B2 (en) 2013-08-28 2019-06-25 The Nielsen Company (Us), Llc Methods and apparatus to estimate demographics of users employing social media
US9332035B2 (en) 2013-10-10 2016-05-03 The Nielsen Company (Us), Llc Methods and apparatus to measure exposure to streaming media
AU2014353157B2 (en) 2013-11-19 2017-09-07 The Nielsen Company (Us), Llc Methods and apparatus to measure a cross device audience
US10956947B2 (en) 2013-12-23 2021-03-23 The Nielsen Company (Us), Llc Methods and apparatus to measure media using media object characteristics
US9852163B2 (en) 2013-12-30 2017-12-26 The Nielsen Company (Us), Llc Methods and apparatus to de-duplicate impression information
US9237138B2 (en) 2013-12-31 2016-01-12 The Nielsen Company (Us), Llc Methods and apparatus to collect distributed user information for media impressions and search terms
US10147114B2 (en) 2014-01-06 2018-12-04 The Nielsen Company (Us), Llc Methods and apparatus to correct audience measurement data
US20150193816A1 (en) 2014-01-06 2015-07-09 The Nielsen Company (Us), Llc Methods and apparatus to correct misattributions of media impressions
US9953330B2 (en) 2014-03-13 2018-04-24 The Nielsen Company (Us), Llc Methods, apparatus and computer readable media to generate electronic mobile measurement census data
EP3117385A4 (en) 2014-03-13 2017-08-02 The Nielsen Company (US), LLC Methods and apparatus to compensate impression data for misattribution and/or non-coverage by a database proprietor
US9622702B2 (en) 2014-04-03 2017-04-18 The Nielsen Company (Us), Llc Methods and apparatus to gather and analyze electroencephalographic data
US10410643B2 (en) 2014-07-15 2019-09-10 The Nielson Company (Us), Llc Audio watermarking for people monitoring
US10311464B2 (en) 2014-07-17 2019-06-04 The Nielsen Company (Us), Llc Methods and apparatus to determine impressions corresponding to market segments
US20160063539A1 (en) 2014-08-29 2016-03-03 The Nielsen Company (Us), Llc Methods and apparatus to associate transactions with media impressions
US20160189182A1 (en) 2014-12-31 2016-06-30 The Nielsen Company (Us), Llc Methods and apparatus to correct age misattribution in media impressions
US9826359B2 (en) 2015-05-01 2017-11-21 The Nielsen Company (Us), Llc Methods and apparatus to associate geographic locations with user devices
US9936250B2 (en) 2015-05-19 2018-04-03 The Nielsen Company (Us), Llc Methods and apparatus to adjust content presented to an individual
US10380633B2 (en) 2015-07-02 2019-08-13 The Nielsen Company (Us), Llc Methods and apparatus to generate corrected online audience measurement data
US10045082B2 (en) 2015-07-02 2018-08-07 The Nielsen Company (Us), Llc Methods and apparatus to correct errors in audience measurements for media accessed using over-the-top devices
US9838754B2 (en) 2015-09-01 2017-12-05 The Nielsen Company (Us), Llc On-site measurement of over the top media
US10205994B2 (en) 2015-12-17 2019-02-12 The Nielsen Company (Us), Llc Methods and apparatus to collect distributed user information for media impressions
US10785329B2 (en) * 2017-01-05 2020-09-22 The Nielsen Company (Us), Llc Methods and apparatus to facilitate meter to meter matching for media identification

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6272176B1 (en) * 1998-07-16 2001-08-07 Nielsen Media Research, Inc. Broadcast encoding system and method
WO2007126992A2 (en) * 2006-03-27 2007-11-08 Nielsen Media Research, Inc. Methods and systems to meter media content presented on a wireless communication device
US20070291848A1 (en) * 1992-11-16 2007-12-20 Aijala Victor A Method and Apparatus for Encoding/Decoding Broadcast or Recorded Segments and Monitoring Audience Exposure Thereto
US20080037658A1 (en) * 2005-03-14 2008-02-14 Lois Price Compressed domain encoding apparatus and methods for use with media signals
US20080101454A1 (en) * 2004-01-23 2008-05-01 Luff Robert A Variable encoding and detection apparatus and methods
US20090187409A1 (en) * 2006-10-10 2009-07-23 Qualcomm Incorporated Method and apparatus for encoding and decoding audio signals
US20090253457A1 (en) * 2008-04-04 2009-10-08 Apple Inc. Audio signal processing for certification enhancement in a handheld wireless communications device

Family Cites Families (84)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4930011A (en) * 1988-08-02 1990-05-29 A. C. Nielsen Company Method and apparatus for identifying individual members of a marketing and viewing audience
GB8824969D0 (en) * 1988-10-25 1988-11-30 Emi Plc Thorn Identification codes
FR2681997A1 (en) * 1991-09-30 1993-04-02 Arbitron Cy METHOD AND DEVICE FOR AUTOMATICALLY IDENTIFYING A PROGRAM COMPRISING A SOUND SIGNAL
US5319735A (en) * 1991-12-17 1994-06-07 Bolt Beranek And Newman Inc. Embedded signalling
ES2229214T3 (en) * 1992-11-16 2005-04-16 Arbitron Inc. METHOD AND APPARATUS FOR CODING / DECODING BROADCASTED OR RECORDED SEGMENTS AND TO MONITOR THE EXHIBITION OF THE HEARING TO THEM.
US5483276A (en) * 1993-08-02 1996-01-09 The Arbitron Company Compliance incentives for audience monitoring/recording devices
US5481294A (en) 1993-10-27 1996-01-02 A. C. Nielsen Company Audience measurement system utilizing ancillary codes and passive signatures
US5450490A (en) * 1994-03-31 1995-09-12 The Arbitron Company Apparatus and methods for including codes in audio signals and decoding
PL177808B1 (en) 1994-03-31 2000-01-31 Arbitron Co Apparatus for and methods of including codes into audio signals and decoding such codes
US5737025A (en) * 1995-02-28 1998-04-07 Nielsen Media Research, Inc. Co-channel transmission of program signals and ancillary signals
US6760463B2 (en) * 1995-05-08 2004-07-06 Digimarc Corporation Watermarking methods and media
US6505160B1 (en) * 1995-07-27 2003-01-07 Digimarc Corporation Connected audio and other media objects
US6650761B1 (en) * 1999-05-19 2003-11-18 Digimarc Corporation Watermarked business cards and methods
US6154484A (en) * 1995-09-06 2000-11-28 Solana Technology Development Corporation Method and apparatus for embedding auxiliary data in a primary data signal using frequency and time domain processing
US5917425A (en) * 1996-01-22 1999-06-29 Wireless Communiations Products, Llc IR/RF locator
US5828325A (en) * 1996-04-03 1998-10-27 Aris Technologies, Inc. Apparatus and method for encoding and decoding information in analog signals
US5940135A (en) * 1997-05-19 1999-08-17 Aris Technologies, Inc. Apparatus and method for encoding and decoding information in analog signals
US6353929B1 (en) * 1997-06-23 2002-03-05 One River Worldtrek, Inc. Cooperative system for measuring electronic media
US5945932A (en) * 1997-10-30 1999-08-31 Audiotrack Corporation Technique for embedding a code in an audio signal and for detecting the embedded code
US20020002039A1 (en) * 1998-06-12 2002-01-03 Safi Qureshey Network-enabled audio device
BR9810699A (en) * 1998-05-12 2000-09-05 Nielsen Media Res Inc Television audience measurement system, process and device to identify a television program selected by a viewer, and software agent stored in memory in association with digital television equipment
US6975835B1 (en) * 1998-09-08 2005-12-13 Sonigistix Corporation Method and apparatus for an interactive Web Radio system that broadcasts a digital markup language
AU1704900A (en) * 1998-10-13 2000-05-01 Radiowave.Com, Inc. System and method for determining the audience of digital radio programmes broadcast through the internet
US6160245A (en) * 1999-05-19 2000-12-12 Maytag Corporation Variable volume signaling device for an appliance
US6871180B1 (en) * 1999-05-25 2005-03-22 Arbitron Inc. Decoding of information in audio signals
US7020285B1 (en) * 1999-07-13 2006-03-28 Microsoft Corporation Stealthy audio watermarking
US6539393B1 (en) * 1999-09-30 2003-03-25 Hill-Rom Services, Inc. Portable locator system
EP2448155A3 (en) * 1999-11-10 2014-05-07 Pandora Media, Inc. Internet radio and broadcast method
US6737957B1 (en) * 2000-02-16 2004-05-18 Verance Corporation Remote control signaling using audio watermarks
US6968564B1 (en) * 2000-04-06 2005-11-22 Nielsen Media Research, Inc. Multi-band spectral audio encoding
US20010055391A1 (en) * 2000-04-27 2001-12-27 Jacobs Paul E. System and method for extracting, decoding, and utilizing hidden data embedded in audio signals
US8121843B2 (en) * 2000-05-02 2012-02-21 Digimarc Corporation Fingerprint methods and systems for media signals
WO2002015111A1 (en) * 2000-08-17 2002-02-21 Parthus Technologies Plc An integrated battery and media decoder for a portable host device, and methods of operating
AU2002220858A1 (en) * 2000-11-30 2002-06-11 Scientific Generics Limited Communication system
GB2369955B (en) * 2000-12-07 2004-01-07 Hewlett Packard Co Encoding of hyperlinks in sound signals
US8572640B2 (en) * 2001-06-29 2013-10-29 Arbitron Inc. Media data use measurement with remote decoding/pattern matching
US7877438B2 (en) * 2001-07-20 2011-01-25 Audible Magic Corporation Method and apparatus for identifying new media content
US6862355B2 (en) * 2001-09-07 2005-03-01 Arbitron Inc. Message reconstruction from partial detection
US7471987B2 (en) * 2002-03-08 2008-12-30 Arbitron, Inc. Determining location of an audience member having a portable media monitor
US7711791B2 (en) * 2002-06-26 2010-05-04 Clear Channel Management Services, Inc. Using multiple media players to insert data items into a media stream of a streaming media
US7239981B2 (en) * 2002-07-26 2007-07-03 Arbitron Inc. Systems and methods for gathering audience measurement data
US7460827B2 (en) * 2002-07-26 2008-12-02 Arbitron, Inc. Radio frequency proximity detection and identification system and method
US7627872B2 (en) * 2002-07-26 2009-12-01 Arbitron Inc. Media data usage measurement and reporting systems and methods
US7222071B2 (en) 2002-09-27 2007-05-22 Arbitron Inc. Audio data receipt/exposure measurement with code monitoring and signature extraction
US6845360B2 (en) * 2002-11-22 2005-01-18 Arbitron Inc. Encoding multiple messages in audio data and detecting same
US7483835B2 (en) * 2002-12-23 2009-01-27 Arbitron, Inc. AD detection using ID code and extracted signature
US7174151B2 (en) 2002-12-23 2007-02-06 Arbitron Inc. Ensuring EAS performance in audio signal encoding
US7012652B1 (en) * 2003-01-09 2006-03-14 Sam Stuff Property Trust Audio hush for entertainment equipment and particularly television receivers
US8225194B2 (en) * 2003-01-09 2012-07-17 Kaleidescape, Inc. Bookmarks and watchpoints for selection and presentation of media streams
WO2004073217A1 (en) * 2003-02-10 2004-08-26 Nielsen Media Research, Inc. Methods and apparatus to adaptively gather audience information data
US7420464B2 (en) * 2004-03-15 2008-09-02 Arbitron, Inc. Methods and systems for gathering market research data inside and outside commercial establishments
US20050203798A1 (en) * 2004-03-15 2005-09-15 Jensen James M. Methods and systems for gathering market research data
US7483975B2 (en) * 2004-03-26 2009-01-27 Arbitron, Inc. Systems and methods for gathering data concerning usage of media data
US8738763B2 (en) * 2004-03-26 2014-05-27 The Nielsen Company (Us), Llc Research data gathering with a portable monitor and a stationary device
US8135606B2 (en) * 2004-04-15 2012-03-13 Arbitron, Inc. Gathering data concerning publication usage and exposure to products and/or presence in commercial establishment
US7623823B2 (en) * 2004-08-31 2009-11-24 Integrated Media Measurement, Inc. Detecting and measuring exposure to media content items
CA2581982C (en) * 2004-09-27 2013-06-18 Nielsen Media Research, Inc. Methods and apparatus for using location information to manage spillover in an audience monitoring system
US7818350B2 (en) * 2005-02-28 2010-10-19 Yahoo! Inc. System and method for creating a collaborative playlist
CN101053256B (en) * 2005-08-04 2010-09-01 日本电信电话株式会社 Digital watermark padding method, digital watermark padding device
GB2431837A (en) * 2005-10-28 2007-05-02 Sony Uk Ltd Audio processing
KR101488317B1 (en) * 2005-12-20 2015-02-04 아비트론 인코포레이티드 Methods and systems for conducting research operations
US20070157228A1 (en) * 2005-12-30 2007-07-05 Jason Bayer Advertising with video ad creatives
US8224018B2 (en) * 2006-01-23 2012-07-17 Digimarc Corporation Sensing data from physical objects
EP2293222A1 (en) * 2006-01-23 2011-03-09 Digimarc Corporation Methods, systems, and subcombinations useful with physical articles
JP2009529753A (en) * 2006-03-09 2009-08-20 グレースノート インコーポレイテッド Media navigation method and system
EP2059924A4 (en) * 2006-08-28 2010-08-25 Shaul Shalev Systems and methods for audio-marking of information items for identifying and activating links to information or processes related to the marked items
US20080077469A1 (en) * 2006-09-27 2008-03-27 Philport Joseph C Method and system for determining media exposure
KR20080029446A (en) * 2006-09-29 2008-04-03 (주)이스트랩 System for providing information using audio watermarking and method thereof
US20080167968A1 (en) * 2007-01-07 2008-07-10 Eddy Cue Creating and Purchasing Ringtones
US20080243590A1 (en) * 2007-01-31 2008-10-02 Michael Rich Methods and systems for measuring exposure to media
US20080187188A1 (en) * 2007-02-07 2008-08-07 Oleg Beletski Systems, apparatuses and methods for facilitating efficient recognition of delivered content
US7888582B2 (en) * 2007-02-08 2011-02-15 Kaleidescape, Inc. Sound sequences with transitions and playlists
US8218812B2 (en) * 2007-05-28 2012-07-10 Mitsubishi Electric Corporation Digital watermark embedding device and method, and digital watermark detection device and method
US20090077579A1 (en) * 2007-09-14 2009-03-19 Att Knowledge Ventures L.P. System and method for estimating an effectivity index for targeted advertising data in a communitcation system
US8554551B2 (en) * 2008-01-28 2013-10-08 Qualcomm Incorporated Systems, methods, and apparatus for context replacement by audio level
US8180677B2 (en) * 2008-03-11 2012-05-15 At&T Intellectual Property I, Lp System and method for compensating users for advertising data in a community of end users
US8330170B2 (en) * 2008-12-05 2012-12-11 Micron Technology, Inc. Semiconductor device structures including transistors with energy barriers adjacent to transistor channels and associated methods
US8977974B2 (en) * 2008-12-08 2015-03-10 Apple Inc. Ambient noise based augmentation of media playback
US8375404B2 (en) * 2008-12-30 2013-02-12 The Nielsen Company (Us), Llc Methods and apparatus to enforce a power off state of an audience measurement device during shipping
US20100268573A1 (en) * 2009-04-17 2010-10-21 Anand Jain System and method for utilizing supplemental audio beaconing in audience measurement
US20100268540A1 (en) * 2009-04-17 2010-10-21 Taymoor Arshi System and method for utilizing audio beaconing in audience measurement
US20100315236A1 (en) * 2009-06-10 2010-12-16 Sakargayan Anupam User activated alarm for communication devices
US9094726B2 (en) * 2009-12-04 2015-07-28 At&T Intellectual Property I, Lp Apparatus and method for tagging media content and managing marketing
US20110214143A1 (en) * 2010-03-01 2011-09-01 Rits Susan K Mobile device application

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070291848A1 (en) * 1992-11-16 2007-12-20 Aijala Victor A Method and Apparatus for Encoding/Decoding Broadcast or Recorded Segments and Monitoring Audience Exposure Thereto
US6272176B1 (en) * 1998-07-16 2001-08-07 Nielsen Media Research, Inc. Broadcast encoding system and method
US20080101454A1 (en) * 2004-01-23 2008-05-01 Luff Robert A Variable encoding and detection apparatus and methods
US20080037658A1 (en) * 2005-03-14 2008-02-14 Lois Price Compressed domain encoding apparatus and methods for use with media signals
WO2007126992A2 (en) * 2006-03-27 2007-11-08 Nielsen Media Research, Inc. Methods and systems to meter media content presented on a wireless communication device
US20090187409A1 (en) * 2006-10-10 2009-07-23 Qualcomm Incorporated Method and apparatus for encoding and decoding audio signals
US20090253457A1 (en) * 2008-04-04 2009-10-08 Apple Inc. Audio signal processing for certification enhancement in a handheld wireless communications device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Arbitron, Critical Band Encoding Technology Audio Encoding System from Arbitron, Technical Overview, 1050-1054 E, February 2008 v3 http://www.davidgleason.com/Archive-Arbitron/Arbitron_PPM_Encoding_White_Paper.pdf (Year: 2008) *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022150612A1 (en) * 2021-01-08 2022-07-14 The Nielsen Company (Us), Llc Engagement measurement of media consumers based on the acoustic environment

Also Published As

Publication number Publication date
US10008212B2 (en) 2018-06-26
US20120239407A1 (en) 2012-09-20
WO2013102016A1 (en) 2013-07-04

Similar Documents

Publication Publication Date Title
US20190019521A1 (en) System and method for utilizing audio encoding for measuring media exposure with environmental masking
AU2010236208B2 (en) System and method for utilizing supplemental audio beaconing in audience measurement
US20100268540A1 (en) System and method for utilizing audio beaconing in audience measurement
US11831950B2 (en) Methods and apparatus to measure exposure to streaming media
AU2016219688B2 (en) Matching techniques for cross-platform monitoring and information
US9100132B2 (en) Systems and methods for gathering audience measurement data
US20120203363A1 (en) Apparatus, system and method for activating functions in processing devices using encoded audio and audio signatures
US11848030B2 (en) Audio encoding for functional interactivity
US20190215169A1 (en) Methods and apparatus to identify media
US11854556B2 (en) Methods and apparatus for supplementing partially readable and/or inaccurate codes in media
AU2015224478A1 (en) System and method for utilizing supplemental audio beaconing in audience measurement
US11842422B2 (en) Methods and apparatus to extend a timestamp range supported by a watermark without breaking backwards compatibility
AU2014250673B2 (en) Methods and apparatus to identify media
AU2016216648A1 (en) Methods and apparatus to identify media

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

AS Assignment

Owner name: ARBITRON, INC, MARYLAND

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LYNCH, WENDELL;TENBROCK, MICHAEL;JAIN, ANAND;AND OTHERS;SIGNING DATES FROM 20090723 TO 20110320;REEL/FRAME:052560/0238

Owner name: NIELSEN AUDIO, INC., NEW YORK

Free format text: CHANGE OF NAME;ASSIGNOR:ARBITRON, INC;REEL/FRAME:052560/0296

Effective date: 20131011

Owner name: THE NIELSEN COMPANY (US), LLC., NEW YORK

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NIELSEN AUDIO, INC.;REEL/FRAME:052560/0275

Effective date: 20140325

AS Assignment

Owner name: CITIBANK, N.A., NEW YORK

Free format text: SUPPLEMENTAL SECURITY AGREEMENT;ASSIGNORS:A. C. NIELSEN COMPANY, LLC;ACN HOLDINGS INC.;ACNIELSEN CORPORATION;AND OTHERS;REEL/FRAME:053473/0001

Effective date: 20200604

AS Assignment

Owner name: CITIBANK, N.A, NEW YORK

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001. ASSIGNOR(S) HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT;ASSIGNORS:A.C. NIELSEN (ARGENTINA) S.A.;A.C. NIELSEN COMPANY, LLC;ACN HOLDINGS INC.;AND OTHERS;REEL/FRAME:054066/0064

Effective date: 20200604

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

AS Assignment

Owner name: BANK OF AMERICA, N.A., NEW YORK

Free format text: SECURITY AGREEMENT;ASSIGNORS:GRACENOTE DIGITAL VENTURES, LLC;GRACENOTE MEDIA SERVICES, LLC;GRACENOTE, INC.;AND OTHERS;REEL/FRAME:063560/0547

Effective date: 20230123

AS Assignment

Owner name: CITIBANK, N.A., NEW YORK

Free format text: SECURITY INTEREST;ASSIGNORS:GRACENOTE DIGITAL VENTURES, LLC;GRACENOTE MEDIA SERVICES, LLC;GRACENOTE, INC.;AND OTHERS;REEL/FRAME:063561/0381

Effective date: 20230427

AS Assignment

Owner name: ARES CAPITAL CORPORATION, NEW YORK

Free format text: SECURITY INTEREST;ASSIGNORS:GRACENOTE DIGITAL VENTURES, LLC;GRACENOTE MEDIA SERVICES, LLC;GRACENOTE, INC.;AND OTHERS;REEL/FRAME:063574/0632

Effective date: 20230508

AS Assignment

Owner name: NETRATINGS, LLC, NEW YORK

Free format text: RELEASE (REEL 053473 / FRAME 0001);ASSIGNOR:CITIBANK, N.A.;REEL/FRAME:063603/0001

Effective date: 20221011

Owner name: THE NIELSEN COMPANY (US), LLC, NEW YORK

Free format text: RELEASE (REEL 053473 / FRAME 0001);ASSIGNOR:CITIBANK, N.A.;REEL/FRAME:063603/0001

Effective date: 20221011

Owner name: GRACENOTE MEDIA SERVICES, LLC, NEW YORK

Free format text: RELEASE (REEL 053473 / FRAME 0001);ASSIGNOR:CITIBANK, N.A.;REEL/FRAME:063603/0001

Effective date: 20221011

Owner name: GRACENOTE, INC., NEW YORK

Free format text: RELEASE (REEL 053473 / FRAME 0001);ASSIGNOR:CITIBANK, N.A.;REEL/FRAME:063603/0001

Effective date: 20221011

Owner name: EXELATE, INC., NEW YORK

Free format text: RELEASE (REEL 053473 / FRAME 0001);ASSIGNOR:CITIBANK, N.A.;REEL/FRAME:063603/0001

Effective date: 20221011

Owner name: A. C. NIELSEN COMPANY, LLC, NEW YORK

Free format text: RELEASE (REEL 053473 / FRAME 0001);ASSIGNOR:CITIBANK, N.A.;REEL/FRAME:063603/0001

Effective date: 20221011

Owner name: NETRATINGS, LLC, NEW YORK

Free format text: RELEASE (REEL 054066 / FRAME 0064);ASSIGNOR:CITIBANK, N.A.;REEL/FRAME:063605/0001

Effective date: 20221011

Owner name: THE NIELSEN COMPANY (US), LLC, NEW YORK

Free format text: RELEASE (REEL 054066 / FRAME 0064);ASSIGNOR:CITIBANK, N.A.;REEL/FRAME:063605/0001

Effective date: 20221011

Owner name: GRACENOTE MEDIA SERVICES, LLC, NEW YORK

Free format text: RELEASE (REEL 054066 / FRAME 0064);ASSIGNOR:CITIBANK, N.A.;REEL/FRAME:063605/0001

Effective date: 20221011

Owner name: GRACENOTE, INC., NEW YORK

Free format text: RELEASE (REEL 054066 / FRAME 0064);ASSIGNOR:CITIBANK, N.A.;REEL/FRAME:063605/0001

Effective date: 20221011

Owner name: EXELATE, INC., NEW YORK

Free format text: RELEASE (REEL 054066 / FRAME 0064);ASSIGNOR:CITIBANK, N.A.;REEL/FRAME:063605/0001

Effective date: 20221011

Owner name: A. C. NIELSEN COMPANY, LLC, NEW YORK

Free format text: RELEASE (REEL 054066 / FRAME 0064);ASSIGNOR:CITIBANK, N.A.;REEL/FRAME:063605/0001

Effective date: 20221011

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION