WO2011060106A1 - Dynamic audio playback of soundtracks for electronic visual works - Google Patents
Dynamic audio playback of soundtracks for electronic visual works Download PDFInfo
- Publication number
- WO2011060106A1 WO2011060106A1 PCT/US2010/056281 US2010056281W WO2011060106A1 WO 2011060106 A1 WO2011060106 A1 WO 2011060106A1 US 2010056281 W US2010056281 W US 2010056281W WO 2011060106 A1 WO2011060106 A1 WO 2011060106A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- audio
- cue
- electronic visual
- visual work
- electronic book
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
- G09B5/06—Electrically-operated educational appliances with both visual and audible presentation of the material to be studied
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/166—Editing, e.g. inserting or deleting
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
- G09B5/06—Electrically-operated educational appliances with both visual and audible presentation of the material to be studied
- G09B5/062—Combinations of audio and printed presentations, e.g. magnetically striped cards, talking books, magnetic tapes with printed texts thereon
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
Definitions
- Electronic books are a kind of multimedia work that is primarily comprised of text, but also may include other visual media such as graphics and images. While text in an electronic book may be accompanied by other visual media, generally an electronic book is intended to be read from start to finish, although not necessarily in one sitting.
- file formats used for electronic books including but not limited to various types of markup language document types (e.g., SGML, HTML, XML, LaTex and the like), and other data file types, such as .pdf files, plain text files, etc.
- Various file formats are used with electronic book readers, such as the KINDLE reader from Amazon.com.
- Such a book reader generally is a computer program designed to run on a platform such as a personal computer, notebook computer, laptop computer, tablet computer, mobile device or dedicated hardware system for reading electronic books (such as the KINDLE reader).
- An electronic book is provided with a soundtrack, to which a reader can listen while reading the electronic book.
- the purpose of the soundtrack is to accompany and enhance the reading experience, in which readers generally have images in their minds based on the story or other visual media that is part of the electronic book. Playback of the soundtrack is synchronized with the visual display of the electronic book.
- audio cues are associated with different points in the text and these cues are dynamically played back in synchronization with the visual display of the electronic book based on the interaction of the user with the electronic book.
- the dynamic playback involves editing and playing an audio cue so that it has a duration that is based on a prediction of the duration of the portion of the electronic book with which the cue is synchronized.
- the system starts playing an audio cue, it predicts when the next audio cue should start. The current cue is played for predicted duration and a transition to the next audio cue is initiated at an appropriate time.
- Such a soundtrack generally is not just any music or sound; some music and sound could be distracting to the reader instead of enhancing the reading experience. Instead, the soundtrack includes music and sound designed to evoke emotions in the reader similar to those emotions that would be evoked by the text.
- a soundtrack for an electronic book benefits when there are few bright transient sounds, no vocals, and a spare, somewhat hypnotic feel to the music. Genre-wise, music that is too fast or too intense can be distracting and difficult to read to.
- the invention can be embodied in a computer-implemented process, a machine (such as an electronic device, or a general purpose computer or other device that provides a platform on which computer programs can be executed), processes performed by these machines, or an article of manufacture.
- a machine such as an electronic device, or a general purpose computer or other device that provides a platform on which computer programs can be executed
- Such articles can include a computer program product or digital information product in which a computer readable storage medium containing computer program instructions or computer readable data stored thereon, and processes and machines that create and use these articles of manufacture.
- dynamic playback of audio involves receiving data about user interaction with a portion of an electronic visual work.
- a section of audio to be played back associated with the portion of the electronic visual work is dynamically adjusted in length according to the user interaction with the electronic visual work.
- the duration of the visual display of the portion of the electronic visual work is estimated according to the received data about user interaction with the portion of the electronic visual work.
- a sequence of sub-mixes of audio associated with the portion of the electronic visual work is selected so as to provide audio elements that will match the estimated duration. This estimation can be done using a history of reading speeds.
- a soundtrack is played in synchronization with display of an electronic visual work.
- the electronic visual work is received into memory.
- Information associating portions of the electronic visual work with tags also is received into memory. Portions of the electronic visual work are displayed in response to user interaction.
- Audio files with tags are accessed. Audio files to be associated with portions of the electronic visual work are selected according to the tags associated with the portions of the electronic visual work.
- Data about user interaction with the portion of an electronic visual work is received and the duration of playback of audio associated with that specific portion of the electronic visual work is dynamically adjusted according to the user interaction.
- a soundtrack for an electronic visual work is generated.
- the electronic visual work is received into memory.
- the electronic visual work is processed in the memory such that portions of the electronic visual work are marked with tags that will associate to specific portions of tagged audio files. Audio files with the appropriate tags are then accessed, and the target audio files for portions of the electronic visual work are selected and associated to create and play back the resulting soundtrack.
- the electronic visual work can includes text and the processing includes processing the text.
- the tags can include emotional descriptors.
- a cue list includes for each portion of an electronic visual work, an emotional descriptor, wherein the emotional descriptors correspond to emotional descriptors also associated with audio data.
- an audio cue includes audio data for a plurality of submixes of the musical work (called “stems”) that can be mixed to provide audio data and information indicative of how the stems can be repeated and combined to create the final result heard by the reader.
- stems submixes of the musical work
- distribution of a soundtrack and its associated electronic visual work is done in a manner that enables the electronic visual work to be viewed in the same manner as if the soundtrack was not available.
- a cue list is identified and read.
- audio data is downloaded while a first cue in the soundtrack is played in synchronization with the display of the electronic visual work.
- FIG. 1 is a dataflow diagram of an electronic book reader with a dynamic audio player.
- Fig. 2 is a dataflow diagram of more details of the dynamic audio player of Fig. 1.
- Fig. 3 is an illustration of a cue list.
- Fig. 4 is an illustration of an audio cue file.
- Fig. 5 is a flow chart of the setup process when an electronic book is opened.
- Fig. 6 is a flow chart describing how an audio cue file is used to create audio data of a desired duration.
- Fig. 7 is a flow chart describing how reading speed is calculated.
- Fig. 8 is a data flow diagram describing how a soundtrack can be
- Fig. 9 is a block diagram of a computer system.
- the embodiments may be described as a process that is depicted as a flowchart, a flow diagram, a structure diagram, or a block diagram. Although a flowchart may describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently. In addition, the order of the operations may be rearranged.
- a process is terminated when its operations are completed.
- a process may correspond to a method, a function, a procedure, a subroutine, a subprogram, etc., in a computer program. When a process corresponds to a function, its termination corresponds to a return of the function to the calling function or a main function.
- mobile device includes, but is not limited to, a wireless device, a mobile phone, a mobile communication device, a user communication device, personal digital assistant, mobile hand-held computer, a laptop computer, an electronic book reader and reading devices capable of reading electronic contents and/or other types of mobile devices typically carried by individuals and/or having some form of communication capabilities (e.g., wireless, infrared, short-range radio, etc.).
- FIG. 9 is a block diagram illustrating the internal functional architecture of a computer system 900 usable with one or more aspects of the systems and methods described in further detail below.
- the computer system 900 may include a central processing unit (CPU) 914 for executing computer-executable process steps and interfaces with a computer bus 916.
- CPU central processing unit
- FIG. 9 Also shown in FIG. 9 are a network interface 918, a display device interface 920, a keyboard or input interface 922, a pointing device interface 924, an audio interface 926, a video interface 932, and a hard disk drive 934 or other persistent storage.
- the disk 934 may store operating system program files, application program files, web browsers, and other files. Some of these files may be stored on the disk 934 using an installation program.
- the CPU 914 may execute computer-executable process steps of an installation program so that the CPU 914 can properly execute the application program.
- a random access main memory (“RAM”) 936 may also interface to the computer bus 916 to provide the CPU 914 with access to memory storage.
- the CPU 914 stores and executes the process steps out of the RAM 936.
- Data to be processed also can be read from such memory 936 or storage 934, and stored in such memory 936 or storage 934.
- Read only memory (“ROM”) 938 may be provided to store invariant instruction sequences such as startup instruction sequences or basic input/output operating system (BIOS) sequences for operation of the keyboard 922.
- BIOS basic input/output operating system
- An electronic book reader, or other application for providing visual displays of electronic books and other multimedia works can be implemented on a platform such as described in Fig. 9.
- an electronic book and an electronic book reader are used as examples of the kind of multimedia work and corresponding viewer with which playback of a soundtrack can be synchronized.
- Other kinds of multimedia works in which the duration of the visual display of a portion of the work is dependent on user interaction with the work also can use this kind of synchronization.
- the term electronic book is intended to encompass books, magazines, newsletters, newspapers, periodicals, maps, articles, and other works that are primarily text or text with accompanying graphics or other visual media.
- Soundtracks can be associated with any of a variety of electronic visual works, including electronic books. The types of music or audio that could be used also likely would depend on the type of work.
- the soundtrack will be similar in purpose to a movie soundtrack, i.e., to support the story - creating suspense, underpinning a love interest, or reaching a big climax.
- the music may be similar to that used for cartoons, possibly including more sound effects, such as for when a page is being turned.
- the soundtrack may include rhythms and tonalities known to enhance knowledge retention, such as material at about 128 or 132 beats per minute and using significant modal tonalities.
- Some books designed to support meditation could have a soundtrack with sounds of nature, ambient sparse music, instruments with soft tones, and the like. Travel books could have music and sounds that are native to the locations being described.
- an electronic book 110 is input to an electronic book reader 112, which provides a visual display of the electronic book to an end user or reader.
- the electronic book 1 10 is one or more computer data files that contain at least text and are in a file format designed to enable a computer program to read, format and display the text.
- file formats for electronic books including but not limited to various types of markup language document types (e.g., SGML, HTML, XML, LaTex and the like), and other document types, examples of which include, but are not limited to, EPUB, FictionBook, plucker, PalmDoc, zTxt, TCR, CHM, RTF, OEB, PDF, mobipocket, Calibre, Stanza, and plain-text.
- markup language document types e.g., SGML, HTML, XML, LaTex and the like
- other document types examples of which include, but are not limited to, EPUB, FictionBook, plucker, PalmDoc, zTxt, TCR, CHM, RTF, OEB, PDF, mobipocket, Calibre, Stanza, and plain-text.
- Some file formats are proprietary and are designed to be used with dedicated electronic book readers. The invention is not limited to any particular file format.
- the electronic book reader 112 can be any computer program designed to run on a computer platform, such as described above in connection with Fig. 9, examples of which include, but are not limited to, a personal computer, tablet computer, mobile device or dedicated hardware system for reading electronic books and that receives and displays the contents of the electronic book 110.
- a computer platform such as described above in connection with Fig. 9, examples of which include, but are not limited to, a personal computer, tablet computer, mobile device or dedicated hardware system for reading electronic books and that receives and displays the contents of the electronic book 110.
- the invention is not limited to any particular electronic book reader.
- the electronic book reader 112 also outputs data 114 indicative of the user interaction with the electronic book reader 112, so that such data can be used by a dynamic audio player 116.
- data 114 indicative of the user interaction with the electronic book reader 112, so that such data can be used by a dynamic audio player 116.
- Commercially or publicly available electronic book readers can be modified in accordance with the description herein to provide such outputs.
- the data about the user interaction with the text can come in a variety of forms.
- an identifier of the book being read such as an ISBN, e-ISBN number or hash code
- the current position in the text can be provided.
- the current position is tracked by the electronic book reader as the current "page" or portion of the electronic book that is being displayed.
- the electronic book reader can output this
- information when it changes includes, but is not limited to the word count for a current range of the document being display, an indication of when the user has exited the electronic book reader application, and an indication of whether the reader has paused reading or resumed reading after a pause.
- API programming interface
- the dynamic audio player can be programmed to implement this API as well.
- An example implementation of the API includes, but is not limited to, two interfaces, one for calls from the electronic book reader application, and another for calls to the electronic book reader application.
- Example calls that the electronic book reader can make to the dynamic audio player include:
- ReadingResumed This function is called when the user has resumed reading after an extended period of inactivity, which the electronic book reader detects by receiving any of a variety of inputs from the user (such as a page turn command) after reading has been determined to be "paused.”
- "fetchSoundtrack” This function is called by the electronic book reader to instruct the dynamic audio player to fetch and import the soundtrack file, or cue list, for the electronic book with a specified unique identifier (provided as a parameter of this function).
- Audio Volume This function is called by the electronic book reader to instruct the dynamic audio player to set the volume of the audio playback.
- cueListEnabled This function is called by the electronic book reader to instruct the dynamic audio player to enable or disable a particular cue list, e.g., an alternative soundtrack, sound effects, a recorded reader or text-to-speech conversion.
- audiolntensity This function is called by the electronic book reader to instruct the dynamic audio player to set the intensity of the audio playback, e.g., to make the audio composition quieter or mute a drum stem (submix).
- AudioPreloadDefault This function is called to set a default number of hours of audio to download and keep on hand generally for electronic books.
- AudioPreloadForEbook This function is called to set a number of hours of audio to download and keep for a specific ebook.
- ReadingPaused This function is called by the dynamic audio player if it has not received a "displayedPositionRangeChanged” call from the electronic book reader within an expected time. From this information, it is assumed by the dynamic audio player that the user is no longer reading. After calling this function, the electronic book reader should call the "readingResumed” function when the user starts reading again.
- wordCountForRange This function is called by the dynamic audio player to instruct the electronic book reader to provide a number of words for a specified range of the electronic book, to be used in scheduling playlists and tracking reading speed as described in more detail below.
- the electronic book 110 has an associated cue list 118, described in more detail below in connection with Fig. 3, which associates portions of the text with audio cues 120.
- an identifier used to uniquely identify the electronic book 110 is used to associate the cue list 118 to the book by either embedding the identifier in the cue list or having a form of lookup table or map that associates the identifier of the book with the cue list 118.
- An audio cue 120 is a computer data file that includes audio data.
- an audio cue 120 associated with a portion of the text by the cue list 118 is played back while the reader is reading that portion of the text.
- a portion of the text may be designated by a point in the text around which the audio cue should start playing, or a range in the text during which the audio cue should play.
- the dynamic audio player 116 determines when and how to stop playing one audio cue and start playing another.
- the dynamic audio player 116 receives data 114 about the user interaction with the electronic book reader 112, as well as cues 120 and the cue list 118. As will be described in more detail below, the dynamic audio player 116 uses the user interaction data 114 and the cue list 118 to select the audio cues 120 to be played, and when and how to play them, to provide an output audio signal 122.
- the dynamic audio player plays a current cue, associated with the portion of the text currently being read, and determines how and when to transition the next cue to be played, based on the data about the user interaction with the text.
- the dynamic audio player 200 thus uses a current cue 204 and a next cue 210 to generate audio 206.
- the cues 204 and 210 to be played are determined through a cue lookup 208, using the data 212 about the user interaction, and the cue list 202.
- the dynamic audio player While the dynamic audio player is playing the current cue 204, it monitors the incoming data 212 to determine when the next cue should be played.
- the current cue 204 may need to be played for a longer or shorter time than the cue's actual duration.
- the dynamic audio player lengthens or shortens the current cue so as to fit the amount of time the user is taking to read the associated portion of the text, and then implements a transition, such as a cross fade, at the estimated time at which the user reaches the text associated with the next cue.
- Audio cues e.g., 120 in Fig. 1 and 204, 210 in Fig. 2, are assigned to portions of the text. This assignment can be done using a meta-tag
- the association with an audio file may be direct or indirect, and may be statically or dynamically defined. For example, different portions of the text can be assigned different words or other labels indicative of emotions, moods or styles of music to be associated with those portions of the text. Audio files then can be associated with such words or labels.
- the audio files can be selected and statically associated with the text, or they can be selected dynamically at the time of playback, as described in more detail below. Alternatively, different points in the text may be associated directly with an audio file.
- the meta-tag information file is a list 300 of pairs 302 of data representing a cue. Each pair 302
- the pair 302 also includes data 306 that specifies the cue.
- This data may be a word or label, such as an emotive tag, or an indication of an audio file, such as a file name, or any other data that may be used to select an audio file. How a composer or a computer program can create such cue lists will be described in more detail below.
- the meta-tag information file can be implemented as a file that is an archive containing several metadata files. These files can be in JavaScript Object Notation (JSON) format.
- JSON JavaScript Object Notation
- the meta-tag information file can include a manifest file that contains general information about the soundtrack, such as the unique identifier of the electronic book with which it is associated, the title of the electronic book, a schema version, (for compatibility purposed, in case the format changes in the future), and a list of other files in the archive, with checksums for integrity checking.
- the meta-tag information file also includes a cuelists file which contains the list of cue list descriptors available in the soundtrack.
- Each cue list descriptor includes a display name, a unique identifier for lookup purposes and an optional group name of the cue list.
- These cue lists might have a group name of "main,” whereas with a sound effects or "read to me” cue list it would be ok to play them all at that same time, and thus would not utilize the group name.
- the meta-tag information file also includes a cues file that contains the list of cue descriptors for all of the cue lists.
- Each cue descriptor includes a descriptive name given to the cue descriptor by a producer. This descriptor could be entered using another application for this purpose, and could include information such as a cue file name that is used to look up the location of the cue file in the list of cue files, and in and out points in the electronic book.
- the meta-tag information file includes a "cuefiles" file that contains the list of cue file descriptors.
- the cuefiles file specifies the network location of the cue files.
- Each cue file descriptor includes a descriptive name given to the cuefile by a producer and used as the cue file name in the cue descriptor, a uniform resource locator (URL) for retrieving the cue file and the original file name of the cue file.
- URL uniform resource locator
- the audio cues (120 in Fig. 1) referred to in such a cue list contain audio data, which may be stored in audio file formats, such as AIFF, MP3, AAC, m4a or other file types.
- An audio cue file 400 can include multiple "stems" (submixes) 402, each of which is a separate audio file that provides one part of a multipart audio mix for the cue. The use of such stems allows the dynamic audio player to select from among the stems to repeat in order to lengthen the playback time of the cue.
- An audio cue file also can include information that is helpful to the dynamic audio player to modify the duration for which the audio cue is played, such as loop markers 404, bar locations 406 and recommended mix information 408.
- the recommended mix information includes a list of instructions for combining the audio stems, where each instruction indicates the stems and sections to be used, and any audio effects processing to be applied.
- Other information such as a word or label indicative of the emotion or mood intended to be evoked by the audio or data indicative of genre, style, instruments, emotion, atmosphere, place, era— called descriptors 410— also can be provided.
- descriptors 410 also can be provided.
- Even more additional information such as alternative keywords, cue volume, cross-fade or fade-in/out shape/intensity and recommended harmonic progression for successive cues also can be included.
- the audio cue file can be implemented as an archive containing a metadata file in JSON format and one or more audio files for stems of the cue.
- the metadata file contains a descriptor for the metadata associated with the audio files, which includes bar locations, loop markers, recommended mix information, emodes (emotional content meta-tags), audio dynamics control metadata (dynamic range compression), instruments, atmospheres and genres.
- the audio files can include data compressed audio files and high resolution original audio files for each stem. Retaining the high resolution versions of each stem supports later editing using music production tools. A copy of the audio cue files without the original audio files can be made to provide for smaller downloads to electronic book readers.
- the cue file contains the compressed audio files for the stems, which are the files used for playback in the end user applications.
- the cue files can be created using a software tool that inputs a set of standard audio stems, adds descriptor, loop point and recommended mix meta information as a separate text file, optimizes and compresses the audio for network delivery and outputs a single package file that can be uploaded to a database.
- An audio file can be analyzed using various analytic techniques to locate sections, beats, loudness information, fades, loop points and the link.
- Cues can be selected using the descriptors "genre, style, instruments, emotion, place, era” and delivered over the network as they are used by the reader.
- the cue lists and cue files can be individually encrypted and linked to a specific work for which they are the soundtrack. The same key would be used to access the work and its soundtrack. Thus files could be tied to the specific work or the specific viewing device through which the work was accessed, and can use digital rights management information associated with the work.
- the electronic book reader calls 502 the "ebookOpenedwithUniquelD" function, indicating the book's unique identifier and whether the book had been opened before.
- the dynamic audio player receives 504 the identifier of the electronic book, and downloads or reads 506 the cue list for the identified book.
- the electronic book reader prompts the dynamic audio player for information about the cue list, by calling 508 the "getCueLists" function.
- the dynamic audio player sends 510 the cue list, which the electronic book reader presents to the user to select 512 one of the soundtracks (if there is more than one soundtrack) for the book.
- Such a selection could be enhanced by using a customer feedback rating system that allows users to rate soundtracks, and these ratings could be displayed to users when a selection of a soundtrack is requested by the system.
- the "cueListEnabled” function is then called 514 to inform the dynamic audio player of the selected cue list, which the dynamic audio player receives 516 through the function call.
- the "fetchSoundtrack” function is called 518 to instruct the dynamic audio player to fetch 520 the cues for playback.
- the dynamic audio player has the starting cue and the cue list, and thus the current cue, for initiating playback. Playback can be started at about the time this portion of the electronic book is displayed by the electronic book reader.
- the dynamic player determines, based on the data about the user interaction with the book, the next cue to play, when to play the cue, and how to transition to the next cue from the current cue.
- the dynamic audio player extends or shortens the playback time of a cue's audio stem files to fit the estimated total cue duration.
- This estimated cue duration can be computed in several ways. An example implementation uses an estimate of the reading speed, the computation of which is described in more detail below.
- the current cue duration is updated in response to the data that describes the user interaction with the electronic book reader, such as provided at every page turn through the "displayedPositionRangeChanged" function call.
- the playback time of a cue's audio stem files is modified by automatically looping sections of the audio stem files, varying the individual stem mixes and dynamically adding various effects such as reverb, delays and chorus.
- the loop points and other mix automation data specific to the audio stem files are stored in the cue file's metadata.
- the sections of the audio stems can be selected so that, when looped and remixed, they provide the most effective and interesting musical end user experience. This process avoids generating music that has obvious repetitions and maximizes the musical content to deliver a musically pleasing result that can have a duration many times that of the original piece(s) of audio.
- the transition between the outgoing and the incoming audio is also managed by the same process, using the cue file metadata to define the style and placement of an appropriate cross fade to create a seamless musical transition.
- a cue file contains four audio stems (a melody track, a sustained chordal or "pad” track, a rhythmic percussive (often drums) track and a rhythmic harmonic track) that would run for 4 minutes if played in a single pass. Further assume that this recording has 3 distinct sections, A, B and C.
- the meta information in the cue file will include:
- transition style i.e., slow, medium or quick fade-in, or stop previous cue with reverb tail and start new cue from beginning of cue
- musical bar and beat markers so that the cross fade will be musically seamless
- the duration of time until the next cue is to be played is determined (600).
- An example way to compute this duration is provided in more detail below.
- the cue producer's input is processed to produce a playlist of the desired duration.
- the first instruction in the remix information is selected 602 and added to playlist. If this section of the audio stems has a duration less than the desired duration, determined at 604, then the next instruction is selected 606, and the process repeats until a playlist of the desired duration is completed 608.
- the transition information in the metadata for the next cue is used to select 610 a starting point in the current playlist to implement a cross-fade from the current cue to the next cue.
- One way to estimate the duration of a cue is to estimate the reading speed of the reader (in words per minute) and, given the number of words in the cue, determine how much time the reader is likely to take to complete reading this portion of the book. This estimate can be computed from a history of reading speed information for the reader.
- an initial reading speed of a certain number words per minute is assumed. This initial speed can be calculated from a variety of data about a user's previous reading speed history from reading previous books, which can be organized by author, by genre, by time of day, by location, and across all books. If no previous reading history is available, then an anonymous global tally of how other users have read this title can be used. If no other history is available a typical average of 400 words per minute is used.
- the reading speed for the user is tracked each time the displayed position range is changed, as indicated by the "displayedPositionRangeChanged" function call. If this function call is received (700), then several conditions are checked 702. These conditions can include, but are not limited to nor are all required: the user is actively reading, i.e., not in the reading paused state; the new displayed position range is greater than the previously displayed position range; the start of the newly displayed position range touches the end of the previously displayed position range; and the word count is above a minimum amount (currently 150 words). The time since the last change also should be within a sensible range, such as the standard deviation of the average reading speed to check the speed is within the normal expected variance.
- the current time is recorded 704.
- the time since the last change to the displayed position range is computed and stored 706, together with the word count for the previously displayed position range.
- the reading speed for this section is computed 708. From this historic data of measured reading speeds, an average reading speed can be computed and used to estimate cue durations.
- Wp is the word count for the page and Tp is the time taken to read the page, in seconds.
- the statistic used for the average reading speed is a 20 period exponential moving average (EMA), which smoothes out fluctuations in speed, while still considering recent page speeds more important.
- EMA exponential moving average
- n is the number of periods, i.e., 20.
- This reading speed information can be stored locally on the user's electronic book reader application platform. Such information for multiple users can be compiled and stored on a server in an anonymous fashion. The application could look up reading speed information statistics to determine how fast others have read a work or portions of a work.
- the data about the user interaction with the electronic book indicates that the reader has started reading from a point within the book. This happens often, as a reader generally does not read a book from start to finish in one sitting. In some cases, when a reader restarts reading at a point within the book, the audio level, or other level of "excitement,” of the audio in the soundtrack at that point might not be appropriate. That is, the audio could actually be distracting at that point.
- the dynamic audio player can use an indication that the reader has started reading from a position within the book as an
- the reader may be reading the book by skipping around from section to section.
- Other multimedia works may encourage such a manner of reading.
- the audio cue associated with a section of a work is played when display of that section is initiated.
- a brief cross-fade from the audio of the previously displayed section to the audio for the newly displayed section can be performed.
- the dynamic playback engine can simply presume that the duration is indefinite and it can continue to generate audio based on the instructions in the cue file until an instruction is received to start another audio cue.
- the audio cue files could be used to playback different sections of a cue file in response to user inputs.
- popular songs could be divided into sections.
- a user interface could be provided for controlling audio playback that would instruct the player to jump to a next section or to a specified section in response to a user input.
- Creating a soundtrack for an electronic book involves associating audio files with portions of the text of the electronic book. There are several ways in which the soundtrack can be created.
- a composer writes and records original music for each portion of the text.
- Each portion of the text can be associated with individual audio files that are so written and recorded.
- previously recorded music can be selected and associated directly with the portions of the text.
- the audio file is statically and directly assigned to portions of the text.
- audio files are indirectly assigned to portions of the text.
- Tags such as words or other labels, are associated with portions of the text. Such tags may be stored in a computer data file or database and associated with the electronic book, similar to the cue list described above.
- Corresponding tags also are associated with audio files. One or more composers write and record original music that is intended to evoke particular emotions or moods. Alternatively, previously recorded music can be selected.
- These audio files also are associated with such tags, and can be stored in a database.
- the tags associated with the portions of the text can be used to automatically select corresponding audio files with the same tags. In the event that multiple audio files are identified for a tag in the book, one of the audio files can be selected either by a computer or through human intervention.
- This implementation allows audio files to be collected in a database, and the creation of a soundtrack to be completed semi-automatically, by automating the process of selecting audio files given the tags associated with the electronic book and with audio files.
- the audio files also can be dynamically selected using the tags at a time closer to playback.
- the process of associating tags with the electronic book also can be automated.
- the text can be processed by a computer to associate emotional descriptors to portions of the text based on a semantic analysis of the words of the text.
- Example techniques for such semantic analysis include, but are not limited to, those described in "Emotions from text: machine learning for text-based emotion prediction," by Cecilia Ovesdotter Aim et al., in Proceedings of Human Language Technology Conference and Conference on Empirical Methods in Natural Language Processing (October 2005), pp. 579-586, and which is hereby incorporated by reference.
- These tags can describe the emotional feeling or other sentiment that supports the section of the work being viewed.
- these emotional feelings can include, but are not limited to, medium tension, love interest, tension, jaunty, macho, dark, brooding, ghostly, happy, sad, wistful, sexy moments, bright and sunny.
- Fig. 8 is a data flow diagram that illustrates an example of a fully automated process for creating a soundtrack for an electronic book, given audio files that have tags associated with them.
- An electronic book 800 is input to an emotional descriptor generator 802 that outputs the emotional descriptors and text ranges 804 for the book.
- the emotional descriptors are used to lookup, in an audio database 806, audio files 810 that match the emotional descriptors for each range in the book.
- the audio selector 808 allows for automated, random or semi-automated selection of an audio file for each text range to generate a cue list 812.
- a unique identifier can be generated for the electronic book and stored with the cue list 812.
- Such electronic books and their soundtracks can be distributed in any of variety of ways, including but not limited to currently used ways for commercial distribution of electronic books.
- the electronic book and the electronic book reader are distributed to end users using conventional techniques.
- the distribution of the additional soundtrack and dynamic audio player is completed separately.
- the distribution of the soundtrack is generally completed in two steps: first the cue list is downloaded, and then each audio file is downloaded.
- the audio files can be downloaded on demand.
- the dynamic audio player can include a file manager that maintains information about available cue files that may be stored on the same device on which the electronic book reader operates, or that may be stored remotely.
- the electronic book is distributed to end users along with the cue list and dynamic audio player.
- the electronic book and its associated cue list are distributed together.
- the cue list is then used to download the audio files for the soundtrack as a background task.
- the electronic book is downloaded first and the download of the cue list is initiated as a background task, and then the first audio file for the first cue is immediately downloaded.
- the electronic book reader is a device with local storage that includes local generic cues, having a variety of emotional descriptors that can be selected for a playback in accordance with the cue list. These generic cues would allow playback of audio if a remote audio file became unavailable.
- the electronic book reader application is loaded on a platform that has access to a network, such as the internet, through which it can communicate with a distributor of electronic media.
- a distributor may receive a request to purchase and/or download electronic media from users.
- the distributor may retrieve the requested work and its accompanying soundtrack information from a database.
- the retrieved electronic media can be encrypted and sent to the user of the electronic book reader application.
- the electronic media may be encrypted such that the electronic media may be played only on a single electronic book reader.
- the digital rights management information associated with the work also is applied to the soundtrack information.
- a storage medium may represent one or more devices for storing data, including read-only memory (ROM), random access memory (RAM), magnetic disk storage mediums, optical storage mediums, flash memory devices and/or other machine readable mediums for storing information.
- ROM read-only memory
- RAM random access memory
- magnetic disk storage mediums magnetic disk storage mediums
- optical storage mediums flash memory devices and/or other machine readable mediums for storing information.
- machine readable medium and “computer readable medium” include, but are not limited to portable or fixed storage devices, optical storage devices, and/or various other mediums capable of storing, containing or carrying instruction(s) and/or data.
- embodiments may be implemented by hardware, software, firmware, middleware, microcode, or any combination thereof.
- the program code or code segments to perform the necessary tasks may be stored in a machine-readable medium such as a storage medium or other storage(s).
- a processor may perform the necessary tasks.
- a code segment may represent a procedure, a function, a subprogram, a program, a routine, a subroutine, a module, a software package, a class, or any combination of instructions, data structures, or program statements.
- a code segment may be coupled to another code segment or a hardware circuit by passing and/or receiving information, data, arguments, parameters, or memory contents. Information, arguments, parameters, data, etc. may be passed, forwarded, or transmitted via any suitable means including memory sharing, message passing, token passing, network transmission, etc.
- DSP digital signal processor
- ASIC application specific integrated circuit
- FPGA field programmable gate array
- a general purpose processor may be a microprocessor, but in the alternative, the processor may be any conventional processor, controller, microcontroller, circuit, and/or state machine.
- a processor may also be implemented as a combination of computing components, e.g., a combination of a DSP and a microprocessor, a number of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.
- a software module may reside in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, hard disk, a removable disk, a CD- ROM, or any other form of storage medium known in the art.
- a storage medium may be coupled to the processor such that the processor can read information from, and write information to, the storage medium. In the alternative, the storage medium may be integral to the processor.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Business, Economics & Management (AREA)
- Educational Administration (AREA)
- Educational Technology (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- User Interface Of Digital Computer (AREA)
- Electrically Operated Instructional Devices (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Signal Processing For Digital Recording And Reproducing (AREA)
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2012538959A JP2013511214A (en) | 2009-11-10 | 2010-11-10 | Dynamic audio playback of soundtracks for electronic visual works |
EP10810833A EP2499626A1 (en) | 2009-11-10 | 2010-11-10 | Dynamic audio playback of soundtracks for electronic visual works |
CN2010800607993A CN102687182A (en) | 2009-11-10 | 2010-11-10 | Dynamic audio playback of soundtracks for electronic visual works |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US25999509P | 2009-11-10 | 2009-11-10 | |
US61/259,995 | 2009-11-10 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2011060106A1 true WO2011060106A1 (en) | 2011-05-19 |
Family
ID=43868382
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2010/056281 WO2011060106A1 (en) | 2009-11-10 | 2010-11-10 | Dynamic audio playback of soundtracks for electronic visual works |
Country Status (6)
Country | Link |
---|---|
US (2) | US8527859B2 (en) |
EP (1) | EP2499626A1 (en) |
JP (1) | JP2013511214A (en) |
KR (1) | KR20120091325A (en) |
CN (1) | CN102687182A (en) |
WO (1) | WO2011060106A1 (en) |
Cited By (174)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013015694A1 (en) * | 2011-07-26 | 2013-01-31 | Booktrack Holdings Limited | Soundtrack for electronic text |
JP2014519058A (en) * | 2011-06-03 | 2014-08-07 | アップル インコーポレイテッド | Automatic creation of mapping between text data and audio data |
EP2851901A1 (en) * | 2013-09-18 | 2015-03-25 | Booktrack Holdings Limited | Playback system for synchronised soundtracks for electronic media content |
CN104603734A (en) * | 2012-06-18 | 2015-05-06 | 奥德伯公司 | Selecting and conveying supplemental content |
EP2827333A3 (en) * | 2013-07-17 | 2015-06-03 | Booktrack Holdings Limited | Delivery of synchronised soundtracks for electronic media content |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9535906B2 (en) | 2008-07-31 | 2017-01-03 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
US9620104B2 (en) | 2013-06-07 | 2017-04-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9626955B2 (en) | 2008-04-05 | 2017-04-18 | Apple Inc. | Intelligent text-to-speech conversion |
US9633674B2 (en) | 2013-06-07 | 2017-04-25 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
US9633660B2 (en) | 2010-02-25 | 2017-04-25 | Apple Inc. | User profiling for voice input processing |
US9646614B2 (en) | 2000-03-16 | 2017-05-09 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US9798393B2 (en) | 2011-08-29 | 2017-10-24 | Apple Inc. | Text correction processing |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
EP3276623A1 (en) * | 2016-07-29 | 2018-01-31 | Booktrack Holdings Limited | System and methods for automatic-creation of soundtracks for text |
EP3276624A1 (en) * | 2016-07-29 | 2018-01-31 | Booktrack Holdings Limited | Systems and methods for automatic-creation of soundtracks for speech audio |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9953088B2 (en) | 2012-05-14 | 2018-04-24 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US9966068B2 (en) | 2013-06-08 | 2018-05-08 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US9966065B2 (en) | 2014-05-30 | 2018-05-08 | Apple Inc. | Multi-command single utterance input method |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US9971774B2 (en) | 2012-09-19 | 2018-05-15 | Apple Inc. | Voice-based media searching |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10079014B2 (en) | 2012-06-08 | 2018-09-18 | Apple Inc. | Name recognition system |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US10083690B2 (en) | 2014-05-30 | 2018-09-25 | Apple Inc. | Better resolution when referencing to concepts |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10089072B2 (en) | 2016-06-11 | 2018-10-02 | Apple Inc. | Intelligent device arbitration and control |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10169329B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Exemplar-based natural language processing |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
US10185542B2 (en) | 2013-06-09 | 2019-01-22 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
CN109543671A (en) * | 2018-09-30 | 2019-03-29 | 与德科技有限公司 | Reading method, talking pen, server data processing method and point-of-reading system |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US10269345B2 (en) | 2016-06-11 | 2019-04-23 | Apple Inc. | Intelligent task discovery |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US10283110B2 (en) | 2009-07-02 | 2019-05-07 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US10297253B2 (en) | 2016-06-11 | 2019-05-21 | Apple Inc. | Application integration with a digital assistant |
US10303715B2 (en) | 2017-05-16 | 2019-05-28 | Apple Inc. | Intelligent automated assistant for media exploration |
US10311144B2 (en) | 2017-05-16 | 2019-06-04 | Apple Inc. | Emoji word sense disambiguation |
US10318871B2 (en) | 2005-09-08 | 2019-06-11 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US10332518B2 (en) | 2017-05-09 | 2019-06-25 | Apple Inc. | User interface for correcting recognition errors |
US10356243B2 (en) | 2015-06-05 | 2019-07-16 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10354011B2 (en) | 2016-06-09 | 2019-07-16 | Apple Inc. | Intelligent automated assistant in a home environment |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US10395654B2 (en) | 2017-05-11 | 2019-08-27 | Apple Inc. | Text normalization based on a data-driven learning network |
US10403278B2 (en) | 2017-05-16 | 2019-09-03 | Apple Inc. | Methods and systems for phonetic matching in digital assistant services |
US10403283B1 (en) | 2018-06-01 | 2019-09-03 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US10410637B2 (en) | 2017-05-12 | 2019-09-10 | Apple Inc. | User-specific acoustic models |
US10417266B2 (en) | 2017-05-09 | 2019-09-17 | Apple Inc. | Context-aware ranking of intelligent response suggestions |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US10445429B2 (en) | 2017-09-21 | 2019-10-15 | Apple Inc. | Natural language understanding using vocabularies with compressed serialized tries |
US10474753B2 (en) | 2016-09-07 | 2019-11-12 | Apple Inc. | Language identification using recurrent neural networks |
US10482874B2 (en) | 2017-05-15 | 2019-11-19 | Apple Inc. | Hierarchical belief states for digital assistants |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10496705B1 (en) | 2018-06-03 | 2019-12-03 | Apple Inc. | Accelerated task performance |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US10521466B2 (en) | 2016-06-11 | 2019-12-31 | Apple Inc. | Data driven natural language event detection and classification |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US10568032B2 (en) | 2007-04-03 | 2020-02-18 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US10592604B2 (en) | 2018-03-12 | 2020-03-17 | Apple Inc. | Inverse text normalization for automatic speech recognition |
US10607140B2 (en) | 2010-01-25 | 2020-03-31 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10636424B2 (en) | 2017-11-30 | 2020-04-28 | Apple Inc. | Multi-turn canned dialog |
US10643611B2 (en) | 2008-10-02 | 2020-05-05 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US10657328B2 (en) | 2017-06-02 | 2020-05-19 | Apple Inc. | Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US10684703B2 (en) | 2018-06-01 | 2020-06-16 | Apple Inc. | Attention aware virtual assistant dismissal |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10699717B2 (en) | 2014-05-30 | 2020-06-30 | Apple Inc. | Intelligent assistant for home automation |
US10706373B2 (en) | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10714117B2 (en) | 2013-02-07 | 2020-07-14 | Apple Inc. | Voice trigger for a digital assistant |
US10726832B2 (en) | 2017-05-11 | 2020-07-28 | Apple Inc. | Maintaining privacy of personal information |
US10733375B2 (en) | 2018-01-31 | 2020-08-04 | Apple Inc. | Knowledge-based framework for improving natural language understanding |
US10733982B2 (en) | 2018-01-08 | 2020-08-04 | Apple Inc. | Multi-directional dialog |
US10733993B2 (en) | 2016-06-10 | 2020-08-04 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10748546B2 (en) | 2017-05-16 | 2020-08-18 | Apple Inc. | Digital assistant services based on device capabilities |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10755703B2 (en) | 2017-05-11 | 2020-08-25 | Apple Inc. | Offline personal assistant |
US10755051B2 (en) | 2017-09-29 | 2020-08-25 | Apple Inc. | Rule-based natural language processing |
US10789945B2 (en) | 2017-05-12 | 2020-09-29 | Apple Inc. | Low-latency intelligent automated assistant |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US10791176B2 (en) | 2017-05-12 | 2020-09-29 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US10789959B2 (en) | 2018-03-02 | 2020-09-29 | Apple Inc. | Training speaker recognition models for digital assistants |
US10810274B2 (en) | 2017-05-15 | 2020-10-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
US10818288B2 (en) | 2018-03-26 | 2020-10-27 | Apple Inc. | Natural assistant interaction |
US10839159B2 (en) | 2018-09-28 | 2020-11-17 | Apple Inc. | Named entity normalization in a spoken dialog system |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
US10909331B2 (en) | 2018-03-30 | 2021-02-02 | Apple Inc. | Implicit identification of translation payload with neural machine translation |
US10928918B2 (en) | 2018-05-07 | 2021-02-23 | Apple Inc. | Raise to speak |
US10984780B2 (en) | 2018-05-21 | 2021-04-20 | Apple Inc. | Global semantic word embeddings using bi-directional recurrent neural networks |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US11010561B2 (en) | 2018-09-27 | 2021-05-18 | Apple Inc. | Sentiment prediction from textual data |
US11010127B2 (en) | 2015-06-29 | 2021-05-18 | Apple Inc. | Virtual assistant for media playback |
US11023513B2 (en) | 2007-12-20 | 2021-06-01 | Apple Inc. | Method and apparatus for searching using an active ontology |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US11070949B2 (en) | 2015-05-27 | 2021-07-20 | Apple Inc. | Systems and methods for proactively identifying and surfacing relevant content on an electronic device with a touch-sensitive display |
US11069336B2 (en) | 2012-03-02 | 2021-07-20 | Apple Inc. | Systems and methods for name pronunciation |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
US11145294B2 (en) | 2018-05-07 | 2021-10-12 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US11170166B2 (en) | 2018-09-28 | 2021-11-09 | Apple Inc. | Neural typographical error modeling via generative adversarial networks |
US11204787B2 (en) | 2017-01-09 | 2021-12-21 | Apple Inc. | Application integration with a digital assistant |
US11217251B2 (en) | 2019-05-06 | 2022-01-04 | Apple Inc. | Spoken notifications |
US11227589B2 (en) | 2016-06-06 | 2022-01-18 | Apple Inc. | Intelligent list reading |
US11231904B2 (en) | 2015-03-06 | 2022-01-25 | Apple Inc. | Reducing response latency of intelligent automated assistants |
US11237797B2 (en) | 2019-05-31 | 2022-02-01 | Apple Inc. | User activity shortcut suggestions |
US11269678B2 (en) | 2012-05-15 | 2022-03-08 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
US11281993B2 (en) | 2016-12-05 | 2022-03-22 | Apple Inc. | Model and ensemble compression for metric learning |
US11289073B2 (en) | 2019-05-31 | 2022-03-29 | Apple Inc. | Device text to speech |
US11301477B2 (en) | 2017-05-12 | 2022-04-12 | Apple Inc. | Feedback analysis of a digital assistant |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
US11314370B2 (en) | 2013-12-06 | 2022-04-26 | Apple Inc. | Method for extracting salient dialog usage from live data |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
US11360641B2 (en) | 2019-06-01 | 2022-06-14 | Apple Inc. | Increasing the relevance of new available information |
US11386266B2 (en) | 2018-06-01 | 2022-07-12 | Apple Inc. | Text correction |
US11388291B2 (en) | 2013-03-14 | 2022-07-12 | Apple Inc. | System and method for processing voicemail |
US11423908B2 (en) | 2019-05-06 | 2022-08-23 | Apple Inc. | Interpreting spoken requests |
US11462215B2 (en) | 2018-09-28 | 2022-10-04 | Apple Inc. | Multi-modal inputs for voice commands |
US11468282B2 (en) | 2015-05-15 | 2022-10-11 | Apple Inc. | Virtual assistant in a communication session |
US11475898B2 (en) | 2018-10-26 | 2022-10-18 | Apple Inc. | Low-latency multi-speaker speech recognition |
US11475884B2 (en) | 2019-05-06 | 2022-10-18 | Apple Inc. | Reducing digital assistant latency when a language is incorrectly determined |
US11488406B2 (en) | 2019-09-25 | 2022-11-01 | Apple Inc. | Text detection using global geometry estimators |
US11496600B2 (en) | 2019-05-31 | 2022-11-08 | Apple Inc. | Remote execution of machine-learned models |
US11495218B2 (en) | 2018-06-01 | 2022-11-08 | Apple Inc. | Virtual assistant operation in multi-device environments |
US11532306B2 (en) | 2017-05-16 | 2022-12-20 | Apple Inc. | Detecting a trigger of a digital assistant |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US11638059B2 (en) | 2019-01-04 | 2023-04-25 | Apple Inc. | Content playback on multiple devices |
US11657813B2 (en) | 2019-05-31 | 2023-05-23 | Apple Inc. | Voice identification in digital assistant systems |
US11755276B2 (en) | 2020-05-12 | 2023-09-12 | Apple Inc. | Reducing description length based on confidence |
US11765209B2 (en) | 2020-05-11 | 2023-09-19 | Apple Inc. | Digital assistant hardware abstraction |
US11798547B2 (en) | 2013-03-15 | 2023-10-24 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
US11809483B2 (en) | 2015-09-08 | 2023-11-07 | Apple Inc. | Intelligent automated assistant for media search and playback |
US11853536B2 (en) | 2015-09-08 | 2023-12-26 | Apple Inc. | Intelligent automated assistant in a media environment |
US11886805B2 (en) | 2015-11-09 | 2024-01-30 | Apple Inc. | Unconventional virtual assistant interactions |
US12010262B2 (en) | 2013-08-06 | 2024-06-11 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
Families Citing this family (57)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102124523B (en) | 2008-07-04 | 2014-08-27 | 布克查克控股有限公司 | Method and system for making and playing soundtracks |
KR101702659B1 (en) * | 2009-10-30 | 2017-02-06 | 삼성전자주식회사 | Appratus and method for syncronizing moving picture contents and e-book contents and system thereof |
US8527859B2 (en) * | 2009-11-10 | 2013-09-03 | Dulcetta, Inc. | Dynamic audio playback of soundtracks for electronic visual works |
US20130297599A1 (en) * | 2009-11-10 | 2013-11-07 | Dulcetta Inc. | Music management for adaptive distraction reduction |
US20120001923A1 (en) * | 2010-07-03 | 2012-01-05 | Sara Weinzimmer | Sound-enhanced ebook with sound events triggered by reader progress |
KR101058268B1 (en) * | 2010-08-03 | 2011-08-22 | 안명환 | Mobile terminal with non-reading part |
US9535884B1 (en) | 2010-09-30 | 2017-01-03 | Amazon Technologies, Inc. | Finding an end-of-body within content |
US9645986B2 (en) | 2011-02-24 | 2017-05-09 | Google Inc. | Method, medium, and system for creating an electronic book with an umbrella policy |
US8935259B2 (en) | 2011-06-20 | 2015-01-13 | Google Inc | Text suggestions for images |
US8612584B2 (en) * | 2011-08-29 | 2013-12-17 | Google Inc. | Using eBook reading data to generate time-based information |
US9141404B2 (en) | 2011-10-24 | 2015-09-22 | Google Inc. | Extensible framework for ereader tools |
GB2510744A (en) * | 2011-11-04 | 2014-08-13 | Barnesandnoble Com Llc | System and method for creating recordings associated with electronic publication |
US20130117670A1 (en) * | 2011-11-04 | 2013-05-09 | Barnesandnoble.Com Llc | System and method for creating recordings associated with electronic publication |
US9031493B2 (en) | 2011-11-18 | 2015-05-12 | Google Inc. | Custom narration of electronic books |
US20130131849A1 (en) * | 2011-11-21 | 2013-05-23 | Shadi Mere | System for adapting music and sound to digital text, for electronic devices |
US9183807B2 (en) | 2011-12-07 | 2015-11-10 | Microsoft Technology Licensing, Llc | Displaying virtual data as printed content |
US9229231B2 (en) | 2011-12-07 | 2016-01-05 | Microsoft Technology Licensing, Llc | Updating printed content with personalized virtual data |
US9182815B2 (en) | 2011-12-07 | 2015-11-10 | Microsoft Technology Licensing, Llc | Making static printed content dynamic with virtual data |
US9213705B1 (en) | 2011-12-19 | 2015-12-15 | Audible, Inc. | Presenting content related to primary audio content |
EP2608002A1 (en) * | 2011-12-21 | 2013-06-26 | France Telecom | Method for determining a reading speed of a section of an electronic content |
US20130268826A1 (en) * | 2012-04-06 | 2013-10-10 | Google Inc. | Synchronizing progress in audio and text versions of electronic books |
US9165381B2 (en) | 2012-05-31 | 2015-10-20 | Microsoft Technology Licensing, Llc | Augmented books in a mixed reality environment |
US8933312B2 (en) * | 2012-06-01 | 2015-01-13 | Makemusic, Inc. | Distribution of audio sheet music as an electronic book |
US9535885B2 (en) | 2012-06-28 | 2017-01-03 | International Business Machines Corporation | Dynamically customizing a digital publication |
US20140040715A1 (en) * | 2012-07-25 | 2014-02-06 | Oliver S. Younge | Application for synchronizing e-books with original or custom-created scores |
CN103680562B (en) * | 2012-09-03 | 2017-03-22 | 腾讯科技(深圳)有限公司 | Point distribution method and device for audio file |
US9047356B2 (en) | 2012-09-05 | 2015-06-02 | Google Inc. | Synchronizing multiple reading positions in electronic books |
US9544204B1 (en) * | 2012-09-17 | 2017-01-10 | Amazon Technologies, Inc. | Determining the average reading speed of a user |
US9575960B1 (en) * | 2012-09-17 | 2017-02-21 | Amazon Technologies, Inc. | Auditory enhancement using word analysis |
US9123053B2 (en) | 2012-12-10 | 2015-09-01 | Google Inc. | Analyzing reading metrics to generate action information |
CN103020226A (en) * | 2012-12-12 | 2013-04-03 | 北京百度网讯科技有限公司 | Method and device for acquiring search result |
US20140173441A1 (en) * | 2012-12-19 | 2014-06-19 | Hcl Technologies Limited | Method and system for inserting immersive contents into eBook |
WO2014160717A1 (en) * | 2013-03-28 | 2014-10-02 | Dolby Laboratories Licensing Corporation | Using single bitstream to produce tailored audio device mixes |
CN104168174A (en) * | 2013-05-16 | 2014-11-26 | 阿里巴巴集团控股有限公司 | Method and apparatus for information transmission |
US9317486B1 (en) | 2013-06-07 | 2016-04-19 | Audible, Inc. | Synchronizing playback of digital content with captured physical content |
US20150269133A1 (en) * | 2014-03-19 | 2015-09-24 | International Business Machines Corporation | Electronic book reading incorporating added environmental feel factors |
US10049477B1 (en) | 2014-06-27 | 2018-08-14 | Google Llc | Computer-assisted text and visual styling for images |
CN104166689B (en) * | 2014-07-28 | 2018-04-06 | 小米科技有限责任公司 | The rendering method and device of e-book |
US20160182435A1 (en) * | 2014-12-23 | 2016-06-23 | Sony Corporation | Socially acceptable display of messaging |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
CN105261374A (en) * | 2015-09-23 | 2016-01-20 | 海信集团有限公司 | Cross-media emotion correlation method and system |
US20170116047A1 (en) * | 2015-10-25 | 2017-04-27 | Khozem Z. Dohadwala | Further applications of Reading State control - A method for repositioning reading material on electronic devices |
CN106708894B (en) * | 2015-11-17 | 2019-12-20 | 腾讯科技(深圳)有限公司 | Method and device for configuring background music for electronic book |
US9959343B2 (en) | 2016-01-04 | 2018-05-01 | Gracenote, Inc. | Generating and distributing a replacement playlist |
US10042880B1 (en) * | 2016-01-06 | 2018-08-07 | Amazon Technologies, Inc. | Automated identification of start-of-reading location for ebooks |
US20180032611A1 (en) * | 2016-07-29 | 2018-02-01 | Paul Charles Cameron | Systems and methods for automatic-generation of soundtracks for live speech audio |
CN106960051B (en) * | 2017-03-31 | 2019-12-10 | 掌阅科技股份有限公司 | Audio playing method and device based on electronic book and terminal equipment |
DK180048B1 (en) | 2017-05-11 | 2020-02-04 | Apple Inc. | MAINTAINING THE DATA PROTECTION OF PERSONAL INFORMATION |
DK201770411A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | Multi-modal interfaces |
EP3646322A1 (en) * | 2017-06-29 | 2020-05-06 | Dolby International AB | Methods, systems, devices and computer program products for adapting external content to a video stream |
CN109493839B (en) * | 2018-11-12 | 2024-01-23 | 平安科技(深圳)有限公司 | Air quality display method and device based on voice synthesis and terminal equipment |
CN109994000B (en) * | 2019-03-28 | 2021-10-19 | 掌阅科技股份有限公司 | Reading accompanying method, electronic equipment and computer storage medium |
US11227599B2 (en) | 2019-06-01 | 2022-01-18 | Apple Inc. | Methods and user interfaces for voice-based control of electronic devices |
US11061543B1 (en) | 2020-05-11 | 2021-07-13 | Apple Inc. | Providing relevant data items based on context |
US11490204B2 (en) | 2020-07-20 | 2022-11-01 | Apple Inc. | Multi-device audio adjustment coordination |
US11438683B2 (en) | 2020-07-21 | 2022-09-06 | Apple Inc. | User identification using headphones |
US12051133B2 (en) * | 2021-06-01 | 2024-07-30 | Apple Inc. | Color treatment and color normalization for digital assets |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070183753A1 (en) | 2006-01-24 | 2007-08-09 | Sharp Kabushiki Kaisha | Data outputting device, data outputting method, data outputting program, and recording medium |
WO2009024626A1 (en) | 2007-08-23 | 2009-02-26 | Tunes4Books, S.L. | Method and system for adapting the reproduction speed of a soundtrack associated with a text to the reading speed of a user |
US20090191531A1 (en) | 2007-12-21 | 2009-07-30 | Joseph Saccocci | Method and Apparatus for Integrating Audio and/or Video With a Book |
WO2010002275A2 (en) * | 2008-07-04 | 2010-01-07 | Isoundtrack Limited | Method and system for making and playing soundtracks |
Family Cites Families (29)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4305131A (en) * | 1979-02-05 | 1981-12-08 | Best Robert M | Dialog between TV movies and human viewers |
US5484292A (en) * | 1989-08-21 | 1996-01-16 | Mctaggart; Stephen I. | Apparatus for combining audio and visual indicia |
US5209665A (en) * | 1989-10-12 | 1993-05-11 | Sight & Sound Incorporated | Interactive audio visual work |
JP3120085B2 (en) * | 1991-11-21 | 2000-12-25 | 株式会社セガ | Electronic devices and information carriers |
US5437552A (en) * | 1993-08-13 | 1995-08-01 | Western Publishing Co., Inc. | Interactive audio-visual work |
US5649234A (en) * | 1994-07-07 | 1997-07-15 | Time Warner Interactive Group, Inc. | Method and apparatus for encoding graphical cues on a compact disc synchronized with the lyrics of a song to be played back |
US6199076B1 (en) * | 1996-10-02 | 2001-03-06 | James Logan | Audio program player including a dynamic program selection controller |
US5897324A (en) * | 1997-02-03 | 1999-04-27 | Atop Technologies, Inc. | Multimedia-book operable with removable data storage media implemented with universal interfacing book-adapting processor |
US6424996B1 (en) * | 1998-11-25 | 2002-07-23 | Nexsys Electronics, Inc. | Medical network system and method for transfer of information |
US6529920B1 (en) * | 1999-03-05 | 2003-03-04 | Audiovelocity, Inc. | Multimedia linking device and method |
US7149690B2 (en) * | 1999-09-09 | 2006-12-12 | Lucent Technologies Inc. | Method and apparatus for interactive language instruction |
US6598074B1 (en) * | 1999-09-23 | 2003-07-22 | Rocket Network, Inc. | System and method for enabling multimedia production collaboration over a network |
WO2001091028A1 (en) * | 2000-05-20 | 2001-11-29 | Leem Young Hie | On demand contents providing method and system |
AU2002232928A1 (en) * | 2000-11-03 | 2002-05-15 | Zoesis, Inc. | Interactive character system |
FI20002828A (en) * | 2000-12-22 | 2002-06-23 | Nokia Corp | A method and system for detecting emotional state in telephone communications |
US20060292049A1 (en) * | 2003-09-04 | 2006-12-28 | Tetsuya Nagasaka | Process for producing rutile |
US20050186548A1 (en) * | 2004-02-25 | 2005-08-25 | Barbara Tomlinson | Multimedia interactive role play system |
NO320758B1 (en) * | 2004-07-23 | 2006-01-23 | Telenor Asa | Device and method for reproducing audiovisual content |
US7735012B2 (en) * | 2004-11-04 | 2010-06-08 | Apple Inc. | Audio user interface for computing devices |
US20080120342A1 (en) * | 2005-04-07 | 2008-05-22 | Iofy Corporation | System and Method for Providing Data to be Used in a Presentation on a Device |
US20080141180A1 (en) * | 2005-04-07 | 2008-06-12 | Iofy Corporation | Apparatus and Method for Utilizing an Information Unit to Provide Navigation Features on a Device |
US7790974B2 (en) * | 2006-05-01 | 2010-09-07 | Microsoft Corporation | Metadata-based song creation and editing |
US9386139B2 (en) * | 2009-03-20 | 2016-07-05 | Nokia Technologies Oy | Method and apparatus for providing an emotion-based user interface |
US20100332225A1 (en) * | 2009-06-29 | 2010-12-30 | Nexidia Inc. | Transcript alignment |
US8527859B2 (en) * | 2009-11-10 | 2013-09-03 | Dulcetta, Inc. | Dynamic audio playback of soundtracks for electronic visual works |
US20110153330A1 (en) * | 2009-11-27 | 2011-06-23 | i-SCROLL | System and method for rendering text synchronized audio |
US8634701B2 (en) * | 2009-12-04 | 2014-01-21 | Lg Electronics Inc. | Digital data reproducing apparatus and corresponding method for reproducing content based on user characteristics |
US8564421B2 (en) * | 2010-04-30 | 2013-10-22 | Blackberry Limited | Method and apparatus for generating an audio notification file |
US8434001B2 (en) * | 2010-06-03 | 2013-04-30 | Rhonda Enterprises, Llc | Systems and methods for presenting a content summary of a media item to a user based on a position within the media item |
-
2010
- 2010-11-10 US US12/943,917 patent/US8527859B2/en not_active Expired - Fee Related
- 2010-11-10 KR KR1020127014737A patent/KR20120091325A/en not_active Application Discontinuation
- 2010-11-10 EP EP10810833A patent/EP2499626A1/en not_active Withdrawn
- 2010-11-10 JP JP2012538959A patent/JP2013511214A/en active Pending
- 2010-11-10 WO PCT/US2010/056281 patent/WO2011060106A1/en active Application Filing
- 2010-11-10 CN CN2010800607993A patent/CN102687182A/en active Pending
-
2013
- 2013-08-26 US US14/010,286 patent/US20130346838A1/en not_active Abandoned
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070183753A1 (en) | 2006-01-24 | 2007-08-09 | Sharp Kabushiki Kaisha | Data outputting device, data outputting method, data outputting program, and recording medium |
WO2009024626A1 (en) | 2007-08-23 | 2009-02-26 | Tunes4Books, S.L. | Method and system for adapting the reproduction speed of a soundtrack associated with a text to the reading speed of a user |
US20100149933A1 (en) | 2007-08-23 | 2010-06-17 | Leonard Cervera Navas | Method and system for adapting the reproduction speed of a sound track to a user's text reading speed |
US20090191531A1 (en) | 2007-12-21 | 2009-07-30 | Joseph Saccocci | Method and Apparatus for Integrating Audio and/or Video With a Book |
WO2010002275A2 (en) * | 2008-07-04 | 2010-01-07 | Isoundtrack Limited | Method and system for making and playing soundtracks |
Cited By (287)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9646614B2 (en) | 2000-03-16 | 2017-05-09 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US11928604B2 (en) | 2005-09-08 | 2024-03-12 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US10318871B2 (en) | 2005-09-08 | 2019-06-11 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US11671920B2 (en) | 2007-04-03 | 2023-06-06 | Apple Inc. | Method and system for operating a multifunction portable electronic device using voice-activation |
US10568032B2 (en) | 2007-04-03 | 2020-02-18 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US11023513B2 (en) | 2007-12-20 | 2021-06-01 | Apple Inc. | Method and apparatus for searching using an active ontology |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US10381016B2 (en) | 2008-01-03 | 2019-08-13 | Apple Inc. | Methods and apparatus for altering audio output signals |
US9865248B2 (en) | 2008-04-05 | 2018-01-09 | Apple Inc. | Intelligent text-to-speech conversion |
US9626955B2 (en) | 2008-04-05 | 2017-04-18 | Apple Inc. | Intelligent text-to-speech conversion |
US10108612B2 (en) | 2008-07-31 | 2018-10-23 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US9535906B2 (en) | 2008-07-31 | 2017-01-03 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US11348582B2 (en) | 2008-10-02 | 2022-05-31 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US10643611B2 (en) | 2008-10-02 | 2020-05-05 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US11080012B2 (en) | 2009-06-05 | 2021-08-03 | Apple Inc. | Interface for a virtual digital assistant |
US10795541B2 (en) | 2009-06-05 | 2020-10-06 | Apple Inc. | Intelligent organization of tasks items |
US10475446B2 (en) | 2009-06-05 | 2019-11-12 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US10283110B2 (en) | 2009-07-02 | 2019-05-07 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US12087308B2 (en) | 2010-01-18 | 2024-09-10 | Apple Inc. | Intelligent automated assistant |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US11423886B2 (en) | 2010-01-18 | 2022-08-23 | Apple Inc. | Task flow identification based on user intent |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10741185B2 (en) | 2010-01-18 | 2020-08-11 | Apple Inc. | Intelligent automated assistant |
US9548050B2 (en) | 2010-01-18 | 2017-01-17 | Apple Inc. | Intelligent automated assistant |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US10706841B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Task flow identification based on user intent |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US10984327B2 (en) | 2010-01-25 | 2021-04-20 | New Valuexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US11410053B2 (en) | 2010-01-25 | 2022-08-09 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10607140B2 (en) | 2010-01-25 | 2020-03-31 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10607141B2 (en) | 2010-01-25 | 2020-03-31 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10984326B2 (en) | 2010-01-25 | 2021-04-20 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10692504B2 (en) | 2010-02-25 | 2020-06-23 | Apple Inc. | User profiling for voice input processing |
US10049675B2 (en) | 2010-02-25 | 2018-08-14 | Apple Inc. | User profiling for voice input processing |
US9633660B2 (en) | 2010-02-25 | 2017-04-25 | Apple Inc. | User profiling for voice input processing |
US10417405B2 (en) | 2011-03-21 | 2019-09-17 | Apple Inc. | Device access using voice authentication |
US10102359B2 (en) | 2011-03-21 | 2018-10-16 | Apple Inc. | Device access using voice authentication |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US11120372B2 (en) | 2011-06-03 | 2021-09-14 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US10706373B2 (en) | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US10672399B2 (en) | 2011-06-03 | 2020-06-02 | Apple Inc. | Switching between text data and audio data based on a mapping |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
JP2014519058A (en) * | 2011-06-03 | 2014-08-07 | アップル インコーポレイテッド | Automatic creation of mapping between text data and audio data |
US11350253B2 (en) | 2011-06-03 | 2022-05-31 | Apple Inc. | Active transport based notifications |
CN103782342B (en) * | 2011-07-26 | 2016-08-31 | 布克查克控股有限公司 | The sound channel of e-text |
US9613654B2 (en) | 2011-07-26 | 2017-04-04 | Booktrack Holdings Limited | Soundtrack for electronic text |
CN103782342A (en) * | 2011-07-26 | 2014-05-07 | 布克查克控股有限公司 | Soundtrack for electronic text |
WO2013015694A1 (en) * | 2011-07-26 | 2013-01-31 | Booktrack Holdings Limited | Soundtrack for electronic text |
US9666227B2 (en) | 2011-07-26 | 2017-05-30 | Booktrack Holdings Limited | Soundtrack for electronic text |
US9613653B2 (en) | 2011-07-26 | 2017-04-04 | Booktrack Holdings Limited | Soundtrack for electronic text |
EP2737481A4 (en) * | 2011-07-26 | 2016-06-22 | Booktrack Holdings Ltd | Soundtrack for electronic text |
US9798393B2 (en) | 2011-08-29 | 2017-10-24 | Apple Inc. | Text correction processing |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US11069336B2 (en) | 2012-03-02 | 2021-07-20 | Apple Inc. | Systems and methods for name pronunciation |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
US9953088B2 (en) | 2012-05-14 | 2018-04-24 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US11269678B2 (en) | 2012-05-15 | 2022-03-08 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
US11321116B2 (en) | 2012-05-15 | 2022-05-03 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
US10079014B2 (en) | 2012-06-08 | 2018-09-18 | Apple Inc. | Name recognition system |
EP2862048A4 (en) * | 2012-06-18 | 2016-02-17 | Audible Inc | Selecting and conveying supplemental content |
CN104603734A (en) * | 2012-06-18 | 2015-05-06 | 奥德伯公司 | Selecting and conveying supplemental content |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9971774B2 (en) | 2012-09-19 | 2018-05-15 | Apple Inc. | Voice-based media searching |
US10978090B2 (en) | 2013-02-07 | 2021-04-13 | Apple Inc. | Voice trigger for a digital assistant |
US11636869B2 (en) | 2013-02-07 | 2023-04-25 | Apple Inc. | Voice trigger for a digital assistant |
US10714117B2 (en) | 2013-02-07 | 2020-07-14 | Apple Inc. | Voice trigger for a digital assistant |
US11388291B2 (en) | 2013-03-14 | 2022-07-12 | Apple Inc. | System and method for processing voicemail |
US11798547B2 (en) | 2013-03-15 | 2023-10-24 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
US9620104B2 (en) | 2013-06-07 | 2017-04-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9966060B2 (en) | 2013-06-07 | 2018-05-08 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9633674B2 (en) | 2013-06-07 | 2017-04-25 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
US9966068B2 (en) | 2013-06-08 | 2018-05-08 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US10657961B2 (en) | 2013-06-08 | 2020-05-19 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US11048473B2 (en) | 2013-06-09 | 2021-06-29 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10185542B2 (en) | 2013-06-09 | 2019-01-22 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
US12073147B2 (en) | 2013-06-09 | 2024-08-27 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US11727219B2 (en) | 2013-06-09 | 2023-08-15 | Apple Inc. | System and method for inferring user intent from speech inputs |
US10769385B2 (en) | 2013-06-09 | 2020-09-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
EP2827333A3 (en) * | 2013-07-17 | 2015-06-03 | Booktrack Holdings Limited | Delivery of synchronised soundtracks for electronic media content |
US9836271B2 (en) | 2013-07-17 | 2017-12-05 | Booktrack Holdings Limited | Delivery of synchronised soundtracks for electronic media content |
US12010262B2 (en) | 2013-08-06 | 2024-06-11 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
US9898077B2 (en) | 2013-09-18 | 2018-02-20 | Booktrack Holdings Limited | Playback system for synchronised soundtracks for electronic media content |
EP2851901A1 (en) * | 2013-09-18 | 2015-03-25 | Booktrack Holdings Limited | Playback system for synchronised soundtracks for electronic media content |
US11314370B2 (en) | 2013-12-06 | 2022-04-26 | Apple Inc. | Method for extracting salient dialog usage from live data |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US11699448B2 (en) | 2014-05-30 | 2023-07-11 | Apple Inc. | Intelligent assistant for home automation |
US10878809B2 (en) | 2014-05-30 | 2020-12-29 | Apple Inc. | Multi-command single utterance input method |
US10714095B2 (en) | 2014-05-30 | 2020-07-14 | Apple Inc. | Intelligent assistant for home automation |
US10497365B2 (en) | 2014-05-30 | 2019-12-03 | Apple Inc. | Multi-command single utterance input method |
US10169329B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Exemplar-based natural language processing |
US11257504B2 (en) | 2014-05-30 | 2022-02-22 | Apple Inc. | Intelligent assistant for home automation |
US10083690B2 (en) | 2014-05-30 | 2018-09-25 | Apple Inc. | Better resolution when referencing to concepts |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US10699717B2 (en) | 2014-05-30 | 2020-06-30 | Apple Inc. | Intelligent assistant for home automation |
US9966065B2 (en) | 2014-05-30 | 2018-05-08 | Apple Inc. | Multi-command single utterance input method |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US11670289B2 (en) | 2014-05-30 | 2023-06-06 | Apple Inc. | Multi-command single utterance input method |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US10417344B2 (en) | 2014-05-30 | 2019-09-17 | Apple Inc. | Exemplar-based natural language processing |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US11810562B2 (en) | 2014-05-30 | 2023-11-07 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US10657966B2 (en) | 2014-05-30 | 2020-05-19 | Apple Inc. | Better resolution when referencing to concepts |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US11516537B2 (en) | 2014-06-30 | 2022-11-29 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US9668024B2 (en) | 2014-06-30 | 2017-05-30 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10904611B2 (en) | 2014-06-30 | 2021-01-26 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US10431204B2 (en) | 2014-09-11 | 2019-10-01 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US9986419B2 (en) | 2014-09-30 | 2018-05-29 | Apple Inc. | Social reminders |
US10438595B2 (en) | 2014-09-30 | 2019-10-08 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US10453443B2 (en) | 2014-09-30 | 2019-10-22 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10390213B2 (en) | 2014-09-30 | 2019-08-20 | Apple Inc. | Social reminders |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US11556230B2 (en) | 2014-12-02 | 2023-01-17 | Apple Inc. | Data detection |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US11231904B2 (en) | 2015-03-06 | 2022-01-25 | Apple Inc. | Reducing response latency of intelligent automated assistants |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US10529332B2 (en) | 2015-03-08 | 2020-01-07 | Apple Inc. | Virtual assistant activation |
US10930282B2 (en) | 2015-03-08 | 2021-02-23 | Apple Inc. | Competing devices responding to voice triggers |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US11087759B2 (en) | 2015-03-08 | 2021-08-10 | Apple Inc. | Virtual assistant activation |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US10311871B2 (en) | 2015-03-08 | 2019-06-04 | Apple Inc. | Competing devices responding to voice triggers |
US11842734B2 (en) | 2015-03-08 | 2023-12-12 | Apple Inc. | Virtual assistant activation |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US11468282B2 (en) | 2015-05-15 | 2022-10-11 | Apple Inc. | Virtual assistant in a communication session |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US11127397B2 (en) | 2015-05-27 | 2021-09-21 | Apple Inc. | Device voice control |
US11070949B2 (en) | 2015-05-27 | 2021-07-20 | Apple Inc. | Systems and methods for proactively identifying and surfacing relevant content on an electronic device with a touch-sensitive display |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10681212B2 (en) | 2015-06-05 | 2020-06-09 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10356243B2 (en) | 2015-06-05 | 2019-07-16 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US11947873B2 (en) | 2015-06-29 | 2024-04-02 | Apple Inc. | Virtual assistant for media playback |
US11010127B2 (en) | 2015-06-29 | 2021-05-18 | Apple Inc. | Virtual assistant for media playback |
US11550542B2 (en) | 2015-09-08 | 2023-01-10 | Apple Inc. | Zero latency digital assistant |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US11126400B2 (en) | 2015-09-08 | 2021-09-21 | Apple Inc. | Zero latency digital assistant |
US11853536B2 (en) | 2015-09-08 | 2023-12-26 | Apple Inc. | Intelligent automated assistant in a media environment |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US11809483B2 (en) | 2015-09-08 | 2023-11-07 | Apple Inc. | Intelligent automated assistant for media search and playback |
US11500672B2 (en) | 2015-09-08 | 2022-11-15 | Apple Inc. | Distributed personal assistant |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US11526368B2 (en) | 2015-11-06 | 2022-12-13 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US11886805B2 (en) | 2015-11-09 | 2024-01-30 | Apple Inc. | Unconventional virtual assistant interactions |
US10354652B2 (en) | 2015-12-02 | 2019-07-16 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10942703B2 (en) | 2015-12-23 | 2021-03-09 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US11227589B2 (en) | 2016-06-06 | 2022-01-18 | Apple Inc. | Intelligent list reading |
US11069347B2 (en) | 2016-06-08 | 2021-07-20 | Apple Inc. | Intelligent automated assistant for media exploration |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
US10354011B2 (en) | 2016-06-09 | 2019-07-16 | Apple Inc. | Intelligent automated assistant in a home environment |
US10733993B2 (en) | 2016-06-10 | 2020-08-04 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US11037565B2 (en) | 2016-06-10 | 2021-06-15 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US11657820B2 (en) | 2016-06-10 | 2023-05-23 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10942702B2 (en) | 2016-06-11 | 2021-03-09 | Apple Inc. | Intelligent device arbitration and control |
US10297253B2 (en) | 2016-06-11 | 2019-05-21 | Apple Inc. | Application integration with a digital assistant |
US10521466B2 (en) | 2016-06-11 | 2019-12-31 | Apple Inc. | Data driven natural language event detection and classification |
US10580409B2 (en) | 2016-06-11 | 2020-03-03 | Apple Inc. | Application integration with a digital assistant |
US11152002B2 (en) | 2016-06-11 | 2021-10-19 | Apple Inc. | Application integration with a digital assistant |
US11809783B2 (en) | 2016-06-11 | 2023-11-07 | Apple Inc. | Intelligent device arbitration and control |
US11749275B2 (en) | 2016-06-11 | 2023-09-05 | Apple Inc. | Application integration with a digital assistant |
US10089072B2 (en) | 2016-06-11 | 2018-10-02 | Apple Inc. | Intelligent device arbitration and control |
US10269345B2 (en) | 2016-06-11 | 2019-04-23 | Apple Inc. | Intelligent task discovery |
US10698951B2 (en) | 2016-07-29 | 2020-06-30 | Booktrack Holdings Limited | Systems and methods for automatic-creation of soundtracks for speech audio |
EP3276623A1 (en) * | 2016-07-29 | 2018-01-31 | Booktrack Holdings Limited | System and methods for automatic-creation of soundtracks for text |
EP3276624A1 (en) * | 2016-07-29 | 2018-01-31 | Booktrack Holdings Limited | Systems and methods for automatic-creation of soundtracks for speech audio |
US10474753B2 (en) | 2016-09-07 | 2019-11-12 | Apple Inc. | Language identification using recurrent neural networks |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US10553215B2 (en) | 2016-09-23 | 2020-02-04 | Apple Inc. | Intelligent automated assistant |
US11281993B2 (en) | 2016-12-05 | 2022-03-22 | Apple Inc. | Model and ensemble compression for metric learning |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US11204787B2 (en) | 2017-01-09 | 2021-12-21 | Apple Inc. | Application integration with a digital assistant |
US11656884B2 (en) | 2017-01-09 | 2023-05-23 | Apple Inc. | Application integration with a digital assistant |
US10417266B2 (en) | 2017-05-09 | 2019-09-17 | Apple Inc. | Context-aware ranking of intelligent response suggestions |
US10741181B2 (en) | 2017-05-09 | 2020-08-11 | Apple Inc. | User interface for correcting recognition errors |
US10332518B2 (en) | 2017-05-09 | 2019-06-25 | Apple Inc. | User interface for correcting recognition errors |
US10755703B2 (en) | 2017-05-11 | 2020-08-25 | Apple Inc. | Offline personal assistant |
US10726832B2 (en) | 2017-05-11 | 2020-07-28 | Apple Inc. | Maintaining privacy of personal information |
US10395654B2 (en) | 2017-05-11 | 2019-08-27 | Apple Inc. | Text normalization based on a data-driven learning network |
US10847142B2 (en) | 2017-05-11 | 2020-11-24 | Apple Inc. | Maintaining privacy of personal information |
US11599331B2 (en) | 2017-05-11 | 2023-03-07 | Apple Inc. | Maintaining privacy of personal information |
US11301477B2 (en) | 2017-05-12 | 2022-04-12 | Apple Inc. | Feedback analysis of a digital assistant |
US10789945B2 (en) | 2017-05-12 | 2020-09-29 | Apple Inc. | Low-latency intelligent automated assistant |
US11580990B2 (en) | 2017-05-12 | 2023-02-14 | Apple Inc. | User-specific acoustic models |
US11380310B2 (en) | 2017-05-12 | 2022-07-05 | Apple Inc. | Low-latency intelligent automated assistant |
US10410637B2 (en) | 2017-05-12 | 2019-09-10 | Apple Inc. | User-specific acoustic models |
US11405466B2 (en) | 2017-05-12 | 2022-08-02 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US10791176B2 (en) | 2017-05-12 | 2020-09-29 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US10810274B2 (en) | 2017-05-15 | 2020-10-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
US10482874B2 (en) | 2017-05-15 | 2019-11-19 | Apple Inc. | Hierarchical belief states for digital assistants |
US10303715B2 (en) | 2017-05-16 | 2019-05-28 | Apple Inc. | Intelligent automated assistant for media exploration |
US10403278B2 (en) | 2017-05-16 | 2019-09-03 | Apple Inc. | Methods and systems for phonetic matching in digital assistant services |
US11675829B2 (en) | 2017-05-16 | 2023-06-13 | Apple Inc. | Intelligent automated assistant for media exploration |
US10909171B2 (en) | 2017-05-16 | 2021-02-02 | Apple Inc. | Intelligent automated assistant for media exploration |
US10748546B2 (en) | 2017-05-16 | 2020-08-18 | Apple Inc. | Digital assistant services based on device capabilities |
US10311144B2 (en) | 2017-05-16 | 2019-06-04 | Apple Inc. | Emoji word sense disambiguation |
US11532306B2 (en) | 2017-05-16 | 2022-12-20 | Apple Inc. | Detecting a trigger of a digital assistant |
US11217255B2 (en) | 2017-05-16 | 2022-01-04 | Apple Inc. | Far-field extension for digital assistant services |
US10657328B2 (en) | 2017-06-02 | 2020-05-19 | Apple Inc. | Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling |
US10445429B2 (en) | 2017-09-21 | 2019-10-15 | Apple Inc. | Natural language understanding using vocabularies with compressed serialized tries |
US10755051B2 (en) | 2017-09-29 | 2020-08-25 | Apple Inc. | Rule-based natural language processing |
US10636424B2 (en) | 2017-11-30 | 2020-04-28 | Apple Inc. | Multi-turn canned dialog |
US10733982B2 (en) | 2018-01-08 | 2020-08-04 | Apple Inc. | Multi-directional dialog |
US10733375B2 (en) | 2018-01-31 | 2020-08-04 | Apple Inc. | Knowledge-based framework for improving natural language understanding |
US10789959B2 (en) | 2018-03-02 | 2020-09-29 | Apple Inc. | Training speaker recognition models for digital assistants |
US10592604B2 (en) | 2018-03-12 | 2020-03-17 | Apple Inc. | Inverse text normalization for automatic speech recognition |
US11710482B2 (en) | 2018-03-26 | 2023-07-25 | Apple Inc. | Natural assistant interaction |
US10818288B2 (en) | 2018-03-26 | 2020-10-27 | Apple Inc. | Natural assistant interaction |
US10909331B2 (en) | 2018-03-30 | 2021-02-02 | Apple Inc. | Implicit identification of translation payload with neural machine translation |
US11854539B2 (en) | 2018-05-07 | 2023-12-26 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US11169616B2 (en) | 2018-05-07 | 2021-11-09 | Apple Inc. | Raise to speak |
US10928918B2 (en) | 2018-05-07 | 2021-02-23 | Apple Inc. | Raise to speak |
US11487364B2 (en) | 2018-05-07 | 2022-11-01 | Apple Inc. | Raise to speak |
US11900923B2 (en) | 2018-05-07 | 2024-02-13 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US11145294B2 (en) | 2018-05-07 | 2021-10-12 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US10984780B2 (en) | 2018-05-21 | 2021-04-20 | Apple Inc. | Global semantic word embeddings using bi-directional recurrent neural networks |
US11360577B2 (en) | 2018-06-01 | 2022-06-14 | Apple Inc. | Attention aware virtual assistant dismissal |
US10720160B2 (en) | 2018-06-01 | 2020-07-21 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US11495218B2 (en) | 2018-06-01 | 2022-11-08 | Apple Inc. | Virtual assistant operation in multi-device environments |
US10684703B2 (en) | 2018-06-01 | 2020-06-16 | Apple Inc. | Attention aware virtual assistant dismissal |
US12080287B2 (en) | 2018-06-01 | 2024-09-03 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US11009970B2 (en) | 2018-06-01 | 2021-05-18 | Apple Inc. | Attention aware virtual assistant dismissal |
US10984798B2 (en) | 2018-06-01 | 2021-04-20 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US11386266B2 (en) | 2018-06-01 | 2022-07-12 | Apple Inc. | Text correction |
US10403283B1 (en) | 2018-06-01 | 2019-09-03 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US11431642B2 (en) | 2018-06-01 | 2022-08-30 | Apple Inc. | Variable latency device coordination |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
US10504518B1 (en) | 2018-06-03 | 2019-12-10 | Apple Inc. | Accelerated task performance |
US10944859B2 (en) | 2018-06-03 | 2021-03-09 | Apple Inc. | Accelerated task performance |
US10496705B1 (en) | 2018-06-03 | 2019-12-03 | Apple Inc. | Accelerated task performance |
US11010561B2 (en) | 2018-09-27 | 2021-05-18 | Apple Inc. | Sentiment prediction from textual data |
US10839159B2 (en) | 2018-09-28 | 2020-11-17 | Apple Inc. | Named entity normalization in a spoken dialog system |
US11462215B2 (en) | 2018-09-28 | 2022-10-04 | Apple Inc. | Multi-modal inputs for voice commands |
US11170166B2 (en) | 2018-09-28 | 2021-11-09 | Apple Inc. | Neural typographical error modeling via generative adversarial networks |
CN109543671A (en) * | 2018-09-30 | 2019-03-29 | 与德科技有限公司 | Reading method, talking pen, server data processing method and point-of-reading system |
CN109543671B (en) * | 2018-09-30 | 2021-03-02 | 阿里巴巴(中国)有限公司 | Touch reading method, touch reading pen, server data processing method and touch reading system |
US11475898B2 (en) | 2018-10-26 | 2022-10-18 | Apple Inc. | Low-latency multi-speaker speech recognition |
US11638059B2 (en) | 2019-01-04 | 2023-04-25 | Apple Inc. | Content playback on multiple devices |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
US11423908B2 (en) | 2019-05-06 | 2022-08-23 | Apple Inc. | Interpreting spoken requests |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
US11475884B2 (en) | 2019-05-06 | 2022-10-18 | Apple Inc. | Reducing digital assistant latency when a language is incorrectly determined |
US11705130B2 (en) | 2019-05-06 | 2023-07-18 | Apple Inc. | Spoken notifications |
US11217251B2 (en) | 2019-05-06 | 2022-01-04 | Apple Inc. | Spoken notifications |
US11888791B2 (en) | 2019-05-21 | 2024-01-30 | Apple Inc. | Providing message response suggestions |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
US11289073B2 (en) | 2019-05-31 | 2022-03-29 | Apple Inc. | Device text to speech |
US11360739B2 (en) | 2019-05-31 | 2022-06-14 | Apple Inc. | User activity shortcut suggestions |
US11237797B2 (en) | 2019-05-31 | 2022-02-01 | Apple Inc. | User activity shortcut suggestions |
US11657813B2 (en) | 2019-05-31 | 2023-05-23 | Apple Inc. | Voice identification in digital assistant systems |
US11496600B2 (en) | 2019-05-31 | 2022-11-08 | Apple Inc. | Remote execution of machine-learned models |
US11360641B2 (en) | 2019-06-01 | 2022-06-14 | Apple Inc. | Increasing the relevance of new available information |
US11488406B2 (en) | 2019-09-25 | 2022-11-01 | Apple Inc. | Text detection using global geometry estimators |
US11924254B2 (en) | 2020-05-11 | 2024-03-05 | Apple Inc. | Digital assistant hardware abstraction |
US11765209B2 (en) | 2020-05-11 | 2023-09-19 | Apple Inc. | Digital assistant hardware abstraction |
US11755276B2 (en) | 2020-05-12 | 2023-09-12 | Apple Inc. | Reducing description length based on confidence |
Also Published As
Publication number | Publication date |
---|---|
US8527859B2 (en) | 2013-09-03 |
US20130346838A1 (en) | 2013-12-26 |
KR20120091325A (en) | 2012-08-17 |
EP2499626A1 (en) | 2012-09-19 |
CN102687182A (en) | 2012-09-19 |
US20110195388A1 (en) | 2011-08-11 |
JP2013511214A (en) | 2013-03-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8527859B2 (en) | Dynamic audio playback of soundtracks for electronic visual works | |
US20130297599A1 (en) | Music management for adaptive distraction reduction | |
US11635936B2 (en) | Audio techniques for music content generation | |
EP2126726B1 (en) | Browser interpretable document for controlling a plurality of media players and systems and methods related thereto | |
US9355627B2 (en) | System and method for combining a song and non-song musical content | |
US8644971B2 (en) | System and method for providing music based on a mood | |
US9607655B2 (en) | System and method for seamless multimedia assembly | |
US20080235588A1 (en) | Media player playlist creation and editing within a browser interpretable document | |
US20080235142A1 (en) | System and methods for obtaining rights in playlist entries | |
US10529312B1 (en) | System and method for delivering dynamic user-controlled musical accompaniments | |
US10694222B2 (en) | Generating video content items using object assets | |
JP2019091014A (en) | Method and apparatus for reproducing multimedia | |
CN102163220B (en) | Song transition metadata | |
EP4134947A1 (en) | Music customization user interface | |
US20110314370A1 (en) | Tiered pageview generation for computing devices | |
US20140244576A1 (en) | System capable of providing multiple digital contents and method using the system | |
Goto | OngaCREST Project: building a similarity-aware information environment for a content-symbiotic society | |
US9471205B1 (en) | Computer-implemented method for providing a media accompaniment for segmented activities | |
Meng | MashupMuse: A Web Application for Easier Music Mashup Creation | |
JP2015038760A (en) | System capable of providing a plurality of digital contents and method using the same | |
CN116434729A (en) | Audio content distribution method, device, equipment, storage medium and program product | |
KR101295197B1 (en) | Player for e-book, editing and playing method thereof | |
Hoffman et al. | Multimedia Recipes | |
Breen | Make mine MIDI | |
Takahashi et al. | bog: instrumental aliens |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201080060799.3 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10810833 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012538959 Country of ref document: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010810833 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 20127014737 Country of ref document: KR Kind code of ref document: A |