US20050187773A1 - Voice synthesis system - Google Patents

Voice synthesis system Download PDF

Info

Publication number
US20050187773A1
US20050187773A1 US11047556 US4755605A US2005187773A1 US 20050187773 A1 US20050187773 A1 US 20050187773A1 US 11047556 US11047556 US 11047556 US 4755605 A US4755605 A US 4755605A US 2005187773 A1 US2005187773 A1 US 2005187773A1
Authority
US
Grant status
Application
Patent type
Prior art keywords
text
synthesized
voice
server
voice synthesis
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11047556
Inventor
Pascal Filoche
Paul Miquel
Edouard Hinard
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Orange SA
Original Assignee
Orange SA
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/487Arrangements for providing information services, e.g. recorded voice services, time announcements
    • H04M3/493Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals
    • H04M3/4938Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals comprising a voice browser which renders and interprets, e.g. VoiceXML
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2201/00Electronic components, circuits, software, systems or apparatus used in telephone systems
    • H04M2201/40Electronic components, circuits, software, systems or apparatus used in telephone systems using speech recognition
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2201/00Electronic components, circuits, software, systems or apparatus used in telephone systems
    • H04M2201/60Medium conversion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2203/00Aspects of automatic or semi-automatic exchanges
    • H04M2203/20Aspects of automatic or semi-automatic exchanges related to features of supplementary services
    • H04M2203/2061Language aspects

Abstract

A voice synthesis system for interactive voice services comprises a voice server connected to a packet network dispensing a voice service to a user terminal by executing a service file associated with the voice service. An HTTP client in the voice server transmits a request containing a text to be synthesized during execution of the service file. The service file includes an address designating a resource in a voice synthesis server connected to the packet network and a command responsive to the audio format for commanding the transmitting of the request to the voice synthesis server. An HTTP server in the voice synthesis server transmits to the voice server an audio response including the text that has been synthesized by the voice synthesis server independently of the voice server.

Description

    CROSS-REFERENCE TO RELATED APPLICATION
  • This application claims priority under 35 U.S.C. §119 based on French Application No. 0400958, filed Feb. 2, 2004, the disclosure of which is incorporated by reference herein in its entirety.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to a system and a method of voice synthesis. The invention relates more particularly to a system and a method of voice synthesis for interactive voice services conceived in a voice services management server and dispensed to a user terminal by an interactive voice server.
  • 2. Description of the Prior Art
  • Interactive voice servers known in the art directly integrate voice synthesizers that synthesize text conventionally included in VXML (Voice extensible Markup Language) files. Specific VXML flags indicate text portions to be synthesized to the interactive voice server.
  • At present, although emergent languages such as SSML (Speech Synthesis Markup Language) control certain characteristics at the voice synthesis level and at the voice recognition level, no voice synthesis system has completely dispensed with synthesizers in interactive voice servers. Consequently, voice service providers must conform to the characteristics of existing voice server synthesizers, which considerably limits the field of application of voice synthesis. For example, a text formatted specifically for a particular use, such as RFC822 electronic mail (e-mail), cannot be synthesized directly by an interactive voice server without modifying the voice server itself, which obliges service providers to be dependent on voice service providers.
  • OBJECT OF THE INVENTION
  • An object of the present invention is to render voice synthesis independent of an interactive voice server in order to be able to carry out voice synthesis specific to a text to be synthesized without calling on a voice server.
  • SUMMARY OF THE INVENTION
  • Accordingly, a voice synthesis system for interactive voice services comprises an interactive voice server connected to a packet network dispensing a voice service to a user terminal by executing a service file associated with said voice service, and a voice synthesis server connected to the packet network and including voice synthesis means. The voice synthesis system is characterized in that it comprises:
      • means in the interactive voice server for transmitting a request containing a text to be synthesized during the execution of the service file, the service file including an address designating a resource in the voice synthesis server and a command responsive to the audio format for commanding transmitting of the request to the voice synthesis server,
      • means in the voice synthesis server for transforming the text to be synthesized into a transformed text as a function of a formatting file that is a parameter of the address of the resource in order for the voice synthesis means to synthesize the transformed text into synthesized text, and
      • means in the voice synthesis server for transmitting to the interactive voice server an audio response to said request including the synthesized text.
  • The service file includes the address designating a resource in the voice synthesis server and the command responsive to the audio format for commanding transmitting of the request in order for the interactive voice server to accept only one audio response to said request. Because the text to be synthesized is a parameter of the address of the resource, voice synthesis in accordance with the invention is easier and faster.
  • The text to be synthesized may also be located by another resource address that is a parameter of the resource address.
  • Before the voice synthesis means synthesizes the text to be synthesized, the transforming means transforms the text to be synthesized as a function of characteristics of the text to be synthesized. The characteristics of the text to be synthesized may be a type, a format and a language of the text. The type of the text to be synthesized may indicate an electronic mail, a short message or a multimedia message.
  • The transformation means can also transform the text to be synthesized as a function of characteristics of the voice synthesis means before the voice synthesis means synthesizes the text to be synthesized.
  • According to one advantageous aspect of the invention, the voice synthesis server may also comprise means for determining the language of the text to be synthesized and means for translating the text to be synthesized into a translation language different from the language of the text to be synthesized that has been determined. The voice synthesis means then synthesizes the translated text into a synthesized text in the translation language.
  • Preprocessing of the text such as transforming and translating it are advantageously effected just before voice synthesis of the text in order to prepare the text to be synthesized for specific voice synthesis, for example.
  • The voice synthesis system may comprise plural voice synthesis means, one of which may be included in the voice synthesis server, and which are divided between voice synthesis servers connected via the packet network. The voice synthesis server then selects one of the voice synthesizing means to synthesize the text to be synthesized as a function of characteristics of the text to be synthesized.
  • The invention also relates to a voice synthesis method for interactive voice services comprising execution of a service file in an interactive voice server connected to a packet network in order to dispense to a user terminal a voice service associated with said service file. The method of the invention is characterized in that it comprises the following steps:
      • transmitting a request containing a text to be synthesized to a voice synthesis server connected to the packet network during the execution of the service file, the service file including an address designating a resource in the voice synthesis server and a command responsive to an audio format to command transmitting of the request,
      • transforming the text to be synthesized into a transformed text as a function of a formatting file that is a parameter of the address of the resource in order for voice synthesis means in the voice synthesis server to synthesize the transformed text into a synthesized text, and
      • transmitting an audio response to said request including the synthesized text to the interactive voice server.
  • The invention also relates to a voice synthesis server for interactive voice services connected via a packet network to an interactive voice server dispensing a voice service to a user terminal by executing a service file associated with said voice service and including voice synthesis means. The voice synthesis server is characterized in that it comprises:
      • means for transforming a text to be synthesized, transmitted by the interactive voice server during the execution of the service file in a request, the service file also containing an address designating a resource in the voice synthesis server and a command responsive to the audio format for commanding transmitting of the request, into a transformed text as a function of a formatting file that is a parameter of the address of the resource in order for the voice synthesis means to synthesize the transformed text into a synthesized text, and
      • means for transmitting to the interactive voice server an audio response to said request including the synthesized text.
    BRIEF DESCRIPTION OF THE DRAWINGS
  • The foregoing and other features and advantages of the present invention will be apparent from the following detailed description of several embodiments of the invention with reference to the corresponding accompanying drawings, in which:
  • FIG. 1 is a block schematic of a voice synthesis system for interactive voice services provided by a voice services management server and dispensed by an interactive voice server of the invention;
  • FIG. 2 is an algorithm of consultation of a voice service from a user terminal in accordance with the invention; and
  • FIG. 3 is an algorithm of the method of the invention of voice synthesis of a text.
  • DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • Referring to FIG. 1, the voice synthesis system of the invention comprises mainly an interactive voice server SVI, a voice services management server SGS coupled to an administrator terminal TA, at least one voice synthesis service SSV, and at least one user terminal T. FIG. 1 shows three voice synthesis servers SSV1, SSV2 and SSV3 and two user terminals T1 and T2 respectively and interchangeably designated SSV and T in the remainder of the description.
  • The interactive voice server SVI communicates with the voice services management server SGS and the voice synthesis server SSV via a high bit rate packet network RP of the Internet type and with user terminals T connected via an access network RA.
  • In the embodiment shown in FIG. 1, the terminal T is connected to the access network RA by a connection LT.
  • For example, the terminal T is a cellular mobile radio communication terminal T1, the connection LT is a radio communication channel and the access network RA comprises the fixed network of a radio communication network, for example of the GSM (Global System for Mobile communications) type with a GPRS (General Packet Radio Service) facility, or of the UMTS (Universal Mobile Telecommunications System) type.
  • In another embodiment, the terminal T is a fixed telecommunication terminal T2, the connection LT is a telephone line and the access network RA is the switched telephone network.
  • In other embodiments, the user terminal T comprises an electronic telecommunication device or object personal to the user, for example a communicating personal digital assistant PDA. The terminal T may be any other portable or non-portable domestic terminal such as a personal computer having a loudspeaker and connected directly by modem to the connection LT, a video games console or an intelligent television receiver cooperating via an infrared link with a remote controller comprising a display or an alphanumeric keyboard and serving also as a mouse.
  • In other variants, the connection LT is an xDSL (Digital Subscriber Line) or ISDN (Integrated Services Digital Network) line connected to the corresponding access network RA.
  • The user terminals T and the access network RA are not limited to the above examples and may consist of other terminals and access networks known in the art.
  • The administrator terminal TA is typically a personal computer connected to the packet network RP through which it communicates with the voice services management server SGS. The administrator terminal TA makes a software interface available to a user with administrator status after connection of the terminal TA to the voice services management server SGS for the latter to edit the voice service that the administrator user wishes to enable. The voice services management server SGS then generates a service file FS containing the description of a voice service SV, generally in VXML (Voice extensible Markup Language), and stores the service file FS in order to make it available to the interactive voice server SVI.
  • The services management server SGS comprises mainly an HTTP server, a database and software modules.
  • The interactive voice server SVI comprises mainly and conventionally a VXML interpreter IVX, a voice recognition module MRV, a DTMF (Dual Tone MultiFrequency) interpreter DT, an audio module MA, a voice synthesizer SYV and an HTTP (HyperText Transfer Protocol) client CH.
  • The voice synthesizer SYV is not used in the present invention and is shown in FIG. 1 to illustrate the known context of the invention. Consequently, the voice synthesizer SYV could be dispensed with.
  • The interactive voice server SVI also comprises at least one call processing unit for managing voice service calls from the user terminals T. For example, a user terminal T selects a voice service SV of the interactive voice server SVI that executes the VXML service file FS associated with the selected voice service SV and transmitted by the voice services management server SGS at the request of the interactive voice server SVI, as explained in the description of the algorithm for consulting the voice service SV.
  • According to the invention, the voice synthesis server SSV comprises mainly a transformation unit UTR, a language determination module MDL, at least one translator TR, at least one synthesizer SY, an audio processing unit UTA and an HTTP server SH.
  • Following reception of a voice service file by the HTTP client CH of the interactive voice service SVI, the HTTP client CH transmits a request REQ containing at least one text to be synthesized TX to the HTTP server SH. The synthesizer SY synthesizes the text TX into a synthesized text TXS which the HTTP server transmits to the interactive voice server SVI in an audio response REPA.
  • As shown in FIG. 2, the consultation of a voice server SV from a user terminal T essentially comprises steps E1 to E8.
  • In the step E1, the user terminal T conventionally calls the interactive voice server SVI via the access network RA, for example via the switched telephone network, after the user has entered on the keypad of the terminal T a service telephone number NSV to call directly the voice service SV of his choice in the server SVI. Thus the telephone number NSV is transmitted to the server SVI. The server SVI matches the service number NSV to an identifier IDSV of the voice service SV in the step E2.
  • The server SVI stores the identifier IDSV of the voice service SV in association with the telephone number NTU of the user terminal T in the step E3 and transmits them in an IP (Internet Protocol) call packet to the services management server SGS via the packet network RP in the step E4.
  • In the step E5, the services management server SGS stores the pair IDSV-NTU in a table TB1 of the database of the management server SGS and then verifies if the user designated by the number NTU is authorized to consult the voice service SV designated by the identifier IDSV in a table TB2 of the database in the step E6, data relating to a profile of the user is stored beforehand in the table TB2. If the number NTU is not found to match the identifier IDSV in the table TB2, the user is not authorized to consult the selected service and the management server SGS breaks off the call with the voice server SVI which breaks off the call with the user terminal T in the step E7. In the contrary situation, where applicable, the user is invited to enter a confidential access code that the management server SGS receives via the voice server SVI in order to compare it to the one stored in the table TB2 in corresponding relationship to the identifier IDSV. The call is broken off if the code entered is incorrect.
  • Otherwise, if the user is authorized to consult the voice service SV designated by the identifier IDSV, and where applicable has entered the confidential code correctly the voice services management server SGS transmits, by means of IP packets, the VXML service file FS in corresponding relationship to the voice service SV to the voice server SVI in the step E8, in order for a dialog to be instigated between the terminal T and the voice server SVI for the purpose of browsing the voice service SV.
  • During execution of the VXML voice service SV in the voice server SVI, and thus during browsing of the voice service SV by the user, the voice server SVI may be invoked conventionally to call a prerecorded sound file designated by a URL (Uniform Resource Locator) address. The URL address refers to a resource situated in the management server SGS or in any server connected to the packet network RP.
  • In the prior art, the voice server SVI was invoked to synthesize a text or a text file in the voice synthesizer SYV.
  • In the present invention, the voice server SVI is invoked to transmit a text to be synthesized to the voice synthesis server SSV different from the voice server SVI and connected to the packet network RP.
  • Referring to FIG. 3, the voice synthesis method of the invention comprises mainly steps S1 to S8.
  • When editing the voice service SV beforehand, the administrator at the administrator terminal TA references the text TX to be synthesized in the synthesis server SSV by introducing a resource address and a command into the service file FS generated by the management server SGS. The address designates a resource in the voice synthesis server SSV. The command is responsive to the audio format and commands transmitting of the request REQ from the voice server SVI in order for the voice server SVI to accept only one audio response REPA to the request REQ.
  • Appendix 1 shows one example of the VXML command code included in the service file FS, which invokes the VXML “<audio>” flag. The text TX to be synthesized is then a parameter “text” of the resource address.
  • Alternatively, the text TX to be synthesized is located by a parameter “text” of the resource address comprising a resource address of the text to be synthesized. The voice synthesis server then consults this resource address of the text to be synthesized in order to recover the text TX to be synthesized. The resource address of the text TX to be synthesized points to any server connected to the packet network RP. In this variant, the text TX to be synthesized may be generated dynamically.
  • Characteristics of the text may constitute additional parameters of the address, such as the type of text to be synthesized (“type”), the translation language (“ltraduc”), the audio format (“format”), the formatting file (“fmf”), etc. The text type defines the text TX to be synthesized, for example a basic text, an electronic mail (e-mail), an SMS (Short Message Service) short message, an MMS (Multimedia Messaging Service) multimedia message, a postal address, etc. The parameter “fmf” defines, in the same way as the parameter “text”, either the content of the formatting file directly or a formatting file resource address enabling the voice synthesis server SSV subsequently to recover the content of the formatting file. The additional parameters are specified by the administrator at the terminal TA when editing the voice service SV. The parameters are automatically coded by the management server SGS for transmitting over the packet network RP in accordance with the HTTP protocol.
  • During execution of the service file FS, the VXML interpreter IVX in the server SVI comes across the command. At this time, the HTTP client CH transmits the request REQ containing the text TX to be synthesized to the voice synthesis server SSV in the step S1.
  • The HTTP server SH receives the request REQ and the transformation unit UTR transforms the text TX to be synthesized into a transformed text TXT in the step S2. This transformation consists in modifying the text to be synthesized as a function of characteristics of the text TX to be synthesized and/or characteristics of the synthesizer or synthesizers SY.
  • If the text TX to be synthesized is an e-mail, it comprises an e-mail that conforms to the RFC822 standard, i.e. the text TX to be synthesized specifies fields such as the sender, the receiver, the subject and the body. The transformation unit UTR then extracts these different fields in order to eliminate the names of the fields explicitly designated in the text TX to be synthesized and reformulates all of the fields into a transformed text TXT that is coherent for voice presentation of the e-mail. Appendix 2 gives one example of this transformation of an e-mail type text TX to be synthesized.
  • If the text TX to be synthesized is an SMS short message, it is often written using abbreviations, like a telegram. The transformation unit UTR corrects the text TX to be synthesized in order to recompose the text TX to be synthesized into a corrected text TXT including terms in the language of the text to be synthesized known to the synthesizer SY of the synthesis server SSV. Appendix 3 gives an example of the transformation of a short message (SMS) text TX to be synthesized.
  • Another example of a type of text to be synthesized is a mailing address, for example “13 av. Champs Elysées”. This is transformed by the transformation unit UTR into “thirteen avenue Champs Elysées”.
  • In a variant, the text TX to be synthesized is either presented directly in an XML (extensible Markup Language) format document or transformed by the transformation unit UTR into an XML format document.
  • In another variant, the type of the text TX to be synthesized is not transmitted as a parameter but is instead determined automatically by the transformation unit UTR carrying out a textual analysis of the text TX to be synthesized.
  • In another variant, the transformation does not depend on characteristics of the text TX to be synthesized, but on characteristics of the synthesizer or synthesizers SY, such as SSML (Speech Synthesis Markup Language) flags added to the text TX to be synthesized with a view to preparing the text TX for a synthesizer SY that can interpret SSML.
  • In another variant, the transformation unit UTR transforms the text TX to be synthesized (or the associated file containing the text to be synthesized) as a function of the formatting file that is a parameter of the resource address. This file is generally an XSLT (extensible Stylesheet Language Transformations) file if the text TX to be synthesized is an XML document. If the text TX to be synthesized is not an XML document, but has an implicit tree structure, the formatting file is based on that structure.
  • For example, in the case of a “database entry” text TX to be synthesized in an XML document, the XSLT formatting file specifies elements of the XML format document to be synthesized, the order of those elements and parameters of the voice synthesizer that in particular define a particular voice synthesis voice.
  • In another example, the text TX to be synthesized is an e-mail. An e-mail does not conform to the XML format but has an implicit tree structure comprising a header composed of fields such as the receiver, the sender, the subject, the body. The body may be composed of a plurality of elements such as paragraphs, a signature, another e-mail, etc. The formatting file specifies at the transformation level (for example in a manner specific to the type concerned) the order and/or the presence of the fields and/or the elements, as well as adding time delays and/or sound elements.
  • The text TX to be synthesized may be subjected to a plurality of transformations.
  • In the step S3, the language determination module MDL of the voice synthesis server SSV determines the language of the transformed text TXT to be synthesized in order for the translator TR, in the step S4, to translate the text TXT into a to-be-synthesized transformed text translated in the language that is a parameter of the resource address included in the service file FS.
  • Alternatively, the text TX or TXT to be synthesized, where applicable after it is transformed in the unit UTR, is again translated into a predetermined unique language if the language of the text TXT to be synthesized is different from the unique language. In this latter variant, it is not necessary to transmit the translation language as a parameter.
  • In another variant, the text TXT to be synthesized is not translated.
  • After the translation step S4, in the step S5 the voice synthesis server SSV selects the synthesizer SY most appropriate for voice synthesis of the text TX, TXT to be synthesized in order for the predetermined characteristics of the selected synthesizer SY to correspond to the characteristics of the text to be synthesized. These characteristics may be lumped with certain parameters in the service file FS, such as the translation language, or determined by analyzing the text TX, TXT to be synthesized, for example the number of characters, the context, etc.
  • In a variant, the synthesizers SY are distributed between the voice synthesis servers SSV1 to SSV3 represented in FIG. 1 and connected via the packet network RP. The location address of the voice synthesis server SSV1 to SSV3 that includes the most appropriate synthesizer SY is a characteristic of the synthesizer SY.
  • In a variant, the transformed text TXT to be synthesized is composed of terms in more than one language. The language determination module MDL recognizes the languages in the text TX, TXT to be synthesized and segments the latter into respective consecutive segments progressively as a function of the languages that have been recognized. The voice synthesis server SSV selects for each segment one of a plurality of synthesizers SY in the voice synthesis server SSV or distributed between the voice synthesis servers SSV1 to SSV3, as a function of the language of the segment, in order for the segment to be synthesized in the language of the segment.
  • The text TX to be synthesized or the transformed text TXT to be synthesized is transmitted to the selected synthesizer SY in order for the text TX, TXT to be synthesized, whether it has been translated or not, to be synthesized as a synthesized text TXS in the step S6.
  • In the step S7, the audio processing unit UTA processes the synthesized text TXS as a conventional sound file in order to modify the format of the sound file according to the format specified in the corresponding parameter in the service file FS, such as “MP3”, “WMA” or “WAV”, for example. In a variant, the format is not specified as a parameter of the resource address in the service file FS and the audio processing unit UTA always modifies the sound file associated with the synthesized text TXS according to a unique format.
  • In the step S8, the HTTP server SH transmits the voice server SVI the synthesized text TXS in the audio response REPA to the request REQ. The VXML interpreter IVX therefore has access to the sound file associated with the voice synthesis of the text TXT to be synthesized.
  • In a variant, the characteristics of the text TX, TXT to be synthesized, such as the type or the audio format, do not constitute additional parameters of the address but are determined automatically by the voice synthesis server SSV analyzing the text to be synthesized.
  • In another variant, certain parameters, such as the type or the audio format, are stored in a database of the voice synthesis server SSV in corresponding relationship to a client identifier and in this case the only parameter transmitted in the resource address is the client identifier, from which the parameters previously stored can be deduced.
  • In another variant the management server SGS and the synthesis server SSV are implemented in a unique server.
  • Appendix 1
  • Syntax of the VXML command
    <form>
    <block>
    <prompt>
    <audio
    src=“http://@IP_TTS/webCVOX.cgi?text=
    ‘Hello Word’&
    type=‘e-mail’&
    ltraduc=‘English’&
    format=‘ ’”>
    </audio>
    </prompt>
    </block>
    </form>
  • Appendix 2 Transformation of an e-mail Text to be Synthesized
  • Source Text to be Synthesized:
      • From: “Dupont Henri” <henri_dupont@wanadoo.fr>
      • To: paul_lanou@wanadoo.fr
      • Subject: holiday
      • Date: Wed, 7 Jan. 2004 17:07:15+0100
      • MIME-Version: 1.0
      • Content-Type: multipart/alternative
      • X-Priority: 3
      • Content: Hi Paul, I hope you are well. I am writing about our planned winter holiday in February . . . .
  • Transformed Text:
      • You received an e-mail from Henri Dupont on 7 Jan. 2004 at 17:07.
      • The subject of this e-mail is “holiday”.
      • Here is the content of the e-mail: “Hi Paul, I hope you are well. I am writing about our planned winter holiday in February . . . ”
    Appendix 3 Transformation of a Short Message Text to be Synthesized
  • Source Text TX to be Synthesized:
      • 1) Ive bought sme cofy
      • 2) sry bout dis arvo
      • 3) film lol
      • 4) Y? avent U cllD
      • 5) hi Julien dis S Elodie I got my mob dis arvo Iz goin awy 2moz
      • 6) w@ cnI do 4u 2 4give me
      • 7) sry but I cnot cum dis evng HAGN :) fran
      • 8) I cnot cll U, we'll do w@ we Z: 3h20 pm undR r trE n D prk! QSL or rng 1s f ur OK X lee.
  • Corresponding Transformed Text TXT:
      • 1) I have bought some coffee
      • 2) sorry about this afternoon
      • 3) film very funny
      • 4) why haven't you called
      • 5) hi Julien this is Elodie I got my mobile this afternoon I am going away tomorrow
      • 6) what can I do for you to forgive me
      • 7) sorry but I cannot come this evening have a good night <audio src=“audio/up.wav”/>francs In this short message the “smiley” “:)” is replaced by the sound of laughter.
      • 8) I cannot call you, we will do what we said: 15h20 under our tree in the park! reply or ring once if you're OK kiss lee.

Claims (15)

  1. 1. A voice synthesis system for interactive voice services comprising an interactive voice server connected to a packet network dispensing a voice service to a user terminal by executing a service file associated with said voice service, and a voice synthesis server connected to the packet network and including voice synthesis means,
    said interactive voice server comprising means for transmitting a request containing a text to be synthesized during the execution of said service file, said service file including an address designating a resource in said voice synthesis server and a command responsive to an audio format for commanding transmitting of said request to said voice synthesis server, and
    said voice synthesis server comprising means for transforming said text to be synthesized into a transformed text as a function of a formatting file that is a parameter of the resource address in order for said voice synthesis means to synthesize said transformed text into a synthesized text, and means for transmitting an audio response including said synthesized text to said interactive voice server.
  2. 2. A system according to the claim 1, wherein said text to be synthesized is located by another resource address that is a parameter of said resource address.
  3. 3. A system according to claim 1, wherein the transforming means transforms said text to be synthesized as a function of characteristics of said text to be synthesized before said voice synthesis means synthesizes said text to be synthesized.
  4. 4. A system according to claim 3, wherein said characteristics of said text to be synthesized are a type, a format and a language of said text to be synthesized.
  5. 5. A system according to claim 4, wherein said type of said text to be synthesized may indicates one of an electronic mail, a short message and a multimedia message.
  6. 6. A system according to claim 1, wherein said transforming means transforms said text to be synthesized as a function of characteristics of said voice synthesis means before the voice synthesis means synthesizes said text to be synthesized.
  7. 7. A system according to claim 1, wherein said voice synthesis server comprises means for determining the language of said text to be synthesized and means for translating said text to be synthesized into a translated text in a translation language different from said language of said text to be synthesized that has been determined, said voice synthesis means synthesizing said translated text into a synthesized text in said translation language.
  8. 8. A system according to claim 1, comprising plural voice synthesis means in order for said voice synthesizer server to select one of said plural voice synthesis means to synthesize said text to be synthesized as a function of characteristics of said text to be synthesized.
  9. 9. A system according to claim 1, comprising a plural voice synthesis means, and wherein said voice synthesis server comprises means for segmenting said text to be synthesized into respective consecutive segments progressively as a function of recognized languages and selects one of said plural voice synthesis means for each segment as a function of the language of said segment in order for said segment to be synthesized in the language of said segment.
  10. 10. A system according to claim 8, wherein said plural voice synthesis means are divided between voice synthesis servers connected via said packet network.
  11. 11. A voice synthesis method for interactive voice services comprising execution of a service file in an interactive voice server connected to a packet network in order to dispense to a user terminal a voice service associated with said service file, said method comprising the following steps:
    transmitting a request containing a text to be synthesized to a voice synthesis server connected to said packet network during the execution of said service file, said service file including an address designating a resource in said voice synthesis server and a command responsive to an audio format to command transmitting of said request,
    transforming said text to be synthesized into a transformed text as a function of a formatting file that is a parameter of the resource address in order for voice synthesis means in said voice synthesis server to synthesize said transformed text into a synthesized text, and
    transmitting an audio response including said synthesized text to the interactive voice server.
  12. 12. A method according to claim 11, wherein said transformation of said text to be synthesized into said transformed text is effected as a function of characteristics of said text to be synthesized before said voice synthesis server synthesizes said text to be synthesized.
  13. 13. A method according to claim 11, wherein said transformation of said text to be synthesized into said transformed text is effected as a function of characteristics of said voice synthesis means before said voice synthesis server synthesizes said text to be synthesized.
  14. 14. A method according to claim 12, wherein said transformation of said text to be synthesized into said transformed text is effected as a function of characteristics of said voice synthesis means before said synthesis server synthesizes said text to be synthesized.
  15. 15. A voice synthesis server for interactive voice services connected via a packet network to an interactive voice server dispensing a voice service to a user terminal by executing a service file associated with said voice service,
    said voice synthesis server including:
    voice synthesis means,
    means for transforming a text to be synthesized, transmitted by said interactive voice server during execution of said service file in a request, said service file containing an address designating a resource in said voice synthesis server and a command responsive to an audio format for commanding transmitting of the request, into a transformed text as a function of a formatting file that is a parameter of the address of the resource in order for said voice synthesis means to synthesize said transformed text into a synthesized text, and
    means for transmitting an audio response including said synthesized text to said interactive voice server.
US11047556 2004-02-02 2005-02-02 Voice synthesis system Abandoned US20050187773A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
FR0400958A FR2865846A1 (en) 2004-02-02 2004-02-02 Voice synthesis system
FR0400958 2004-02-02

Publications (1)

Publication Number Publication Date
US20050187773A1 true true US20050187773A1 (en) 2005-08-25

Family

ID=34639826

Family Applications (1)

Application Number Title Priority Date Filing Date
US11047556 Abandoned US20050187773A1 (en) 2004-02-02 2005-02-02 Voice synthesis system

Country Status (3)

Country Link
US (1) US20050187773A1 (en)
EP (1) EP1560198A1 (en)
FR (1) FR2865846A1 (en)

Cited By (62)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050136955A1 (en) * 2003-12-23 2005-06-23 Mumick Inderpal S. Techniques for combining voice with wireless text short message services
US20070078656A1 (en) * 2005-10-03 2007-04-05 Niemeyer Terry W Server-provided user's voice for instant messaging clients
US20080004046A1 (en) * 2004-08-14 2008-01-03 Mumick Inderpal S Methods for Identifying Messages and Communicating with Users of a Multimodal Message Service
US20100169096A1 (en) * 2008-12-31 2010-07-01 Alibaba Group Holding Limited Instant communication with instant text data and voice data
US20100228549A1 (en) * 2009-03-09 2010-09-09 Apple Inc Systems and methods for determining the language to use for speech generated by a text to speech engine
CN102169689A (en) * 2011-03-25 2011-08-31 深圳Tcl新技术有限公司 Realization method of speech synthesis plug-in
US8892446B2 (en) 2010-01-18 2014-11-18 Apple Inc. Service orchestration for intelligent automated assistant
US9195656B2 (en) 2013-12-30 2015-11-24 Google Inc. Multilingual prosody generation
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US9300784B2 (en) 2013-06-13 2016-03-29 Apple Inc. System and method for emergency calls initiated by voice command
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US9535906B2 (en) 2008-07-31 2017-01-03 Apple Inc. Mobile device having human language translation capability with positional feedback
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
US9606986B2 (en) 2014-09-29 2017-03-28 Apple Inc. Integrated word N-gram and class M-gram language models
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US9620104B2 (en) 2013-06-07 2017-04-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9626955B2 (en) 2008-04-05 2017-04-18 Apple Inc. Intelligent text-to-speech conversion
US9633660B2 (en) 2010-02-25 2017-04-25 Apple Inc. User profiling for voice input processing
US9633674B2 (en) 2013-06-07 2017-04-25 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9646614B2 (en) 2000-03-16 2017-05-09 Apple Inc. Fast, language-independent method for user authentication by voice
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US9697822B1 (en) 2013-03-15 2017-07-04 Apple Inc. System and method for updating an adaptive speech recognition model
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9798393B2 (en) 2011-08-29 2017-10-24 Apple Inc. Text correction processing
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9922641B1 (en) * 2012-10-01 2018-03-20 Google Llc Cross-lingual speaker adaptation for multi-lingual speech synthesis
US9922642B2 (en) 2013-03-15 2018-03-20 Apple Inc. Training an at least partial voice command system
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9953088B2 (en) 2012-05-14 2018-04-24 Apple Inc. Crowd sourcing information to fulfill user requests
US9959870B2 (en) 2008-12-11 2018-05-01 Apple Inc. Speech recognition involving a mobile device
US9966068B2 (en) 2013-06-08 2018-05-08 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US9971774B2 (en) 2012-09-19 2018-05-15 Apple Inc. Voice-based media searching
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10074360B2 (en) 2015-08-24 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5832433A (en) * 1996-06-24 1998-11-03 Nynex Science And Technology, Inc. Speech synthesis method for operator assistance telecommunications calls comprising a plurality of text-to-speech (TTS) devices
US6243681B1 (en) * 1999-04-19 2001-06-05 Oki Electric Industry Co., Ltd. Multiple language speech synthesizer
US20020091528A1 (en) * 1997-04-14 2002-07-11 Daragosh Pamela Leigh System and method for providing remote automatic speech recognition and text to speech services via a packet network
US6574598B1 (en) * 1998-01-19 2003-06-03 Sony Corporation Transmitter and receiver, apparatus and method, all for delivery of information
US20030187658A1 (en) * 2002-03-29 2003-10-02 Jari Selin Method for text-to-speech service utilizing a uniform resource identifier
US20050091058A1 (en) * 2002-02-13 2005-04-28 France Telecom Interactive telephone voice services
US7003463B1 (en) * 1998-10-02 2006-02-21 International Business Machines Corporation System and method for providing network coordinated conversational services

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6587822B2 (en) * 1998-10-06 2003-07-01 Lucent Technologies Inc. Web-based platform for interactive voice response (IVR)
EP1241600A1 (en) * 2001-03-13 2002-09-18 Siemens Schweiz AG Method and communication system for the generation of responses to questions

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5832433A (en) * 1996-06-24 1998-11-03 Nynex Science And Technology, Inc. Speech synthesis method for operator assistance telecommunications calls comprising a plurality of text-to-speech (TTS) devices
US20020091528A1 (en) * 1997-04-14 2002-07-11 Daragosh Pamela Leigh System and method for providing remote automatic speech recognition and text to speech services via a packet network
US6574598B1 (en) * 1998-01-19 2003-06-03 Sony Corporation Transmitter and receiver, apparatus and method, all for delivery of information
US7003463B1 (en) * 1998-10-02 2006-02-21 International Business Machines Corporation System and method for providing network coordinated conversational services
US6243681B1 (en) * 1999-04-19 2001-06-05 Oki Electric Industry Co., Ltd. Multiple language speech synthesizer
US20050091058A1 (en) * 2002-02-13 2005-04-28 France Telecom Interactive telephone voice services
US20030187658A1 (en) * 2002-03-29 2003-10-02 Jari Selin Method for text-to-speech service utilizing a uniform resource identifier

Cited By (81)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9646614B2 (en) 2000-03-16 2017-05-09 Apple Inc. Fast, language-independent method for user authentication by voice
US7184786B2 (en) * 2003-12-23 2007-02-27 Kirusa, Inc. Techniques for combining voice with wireless text short message services
US20050136955A1 (en) * 2003-12-23 2005-06-23 Mumick Inderpal S. Techniques for combining voice with wireless text short message services
US8688150B2 (en) * 2004-08-14 2014-04-01 Kirusa Inc. Methods for identifying messages and communicating with users of a multimodal message service
US20080004046A1 (en) * 2004-08-14 2008-01-03 Mumick Inderpal S Methods for Identifying Messages and Communicating with Users of a Multimodal Message Service
US20070078656A1 (en) * 2005-10-03 2007-04-05 Niemeyer Terry W Server-provided user's voice for instant messaging clients
US8428952B2 (en) 2005-10-03 2013-04-23 Nuance Communications, Inc. Text-to-speech user's voice cooperative server for instant messaging clients
US8224647B2 (en) * 2005-10-03 2012-07-17 Nuance Communications, Inc. Text-to-speech user's voice cooperative server for instant messaging clients
US9026445B2 (en) 2005-10-03 2015-05-05 Nuance Communications, Inc. Text-to-speech user's voice cooperative server for instant messaging clients
US8930191B2 (en) 2006-09-08 2015-01-06 Apple Inc. Paraphrasing of user requests and results by automated digital assistant
US9117447B2 (en) 2006-09-08 2015-08-25 Apple Inc. Using event alert text as input to an automated assistant
US8942986B2 (en) 2006-09-08 2015-01-27 Apple Inc. Determining user intent based on ontologies of domains
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US9865248B2 (en) 2008-04-05 2018-01-09 Apple Inc. Intelligent text-to-speech conversion
US9626955B2 (en) 2008-04-05 2017-04-18 Apple Inc. Intelligent text-to-speech conversion
US9535906B2 (en) 2008-07-31 2017-01-03 Apple Inc. Mobile device having human language translation capability with positional feedback
US9959870B2 (en) 2008-12-11 2018-05-01 Apple Inc. Speech recognition involving a mobile device
US20100169096A1 (en) * 2008-12-31 2010-07-01 Alibaba Group Holding Limited Instant communication with instant text data and voice data
US8751238B2 (en) 2009-03-09 2014-06-10 Apple Inc. Systems and methods for determining the language to use for speech generated by a text to speech engine
US8380507B2 (en) * 2009-03-09 2013-02-19 Apple Inc. Systems and methods for determining the language to use for speech generated by a text to speech engine
US20100228549A1 (en) * 2009-03-09 2010-09-09 Apple Inc Systems and methods for determining the language to use for speech generated by a text to speech engine
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US8892446B2 (en) 2010-01-18 2014-11-18 Apple Inc. Service orchestration for intelligent automated assistant
US9548050B2 (en) 2010-01-18 2017-01-17 Apple Inc. Intelligent automated assistant
US8903716B2 (en) 2010-01-18 2014-12-02 Apple Inc. Personalized vocabulary for digital assistant
US9633660B2 (en) 2010-02-25 2017-04-25 Apple Inc. User profiling for voice input processing
US10049675B2 (en) 2010-02-25 2018-08-14 Apple Inc. User profiling for voice input processing
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
CN102169689A (en) * 2011-03-25 2011-08-31 深圳Tcl新技术有限公司 Realization method of speech synthesis plug-in
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US9798393B2 (en) 2011-08-29 2017-10-24 Apple Inc. Text correction processing
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US9953088B2 (en) 2012-05-14 2018-04-24 Apple Inc. Crowd sourcing information to fulfill user requests
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9971774B2 (en) 2012-09-19 2018-05-15 Apple Inc. Voice-based media searching
US9922641B1 (en) * 2012-10-01 2018-03-20 Google Llc Cross-lingual speaker adaptation for multi-lingual speech synthesis
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
US9922642B2 (en) 2013-03-15 2018-03-20 Apple Inc. Training an at least partial voice command system
US9697822B1 (en) 2013-03-15 2017-07-04 Apple Inc. System and method for updating an adaptive speech recognition model
US9633674B2 (en) 2013-06-07 2017-04-25 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9966060B2 (en) 2013-06-07 2018-05-08 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
US9620104B2 (en) 2013-06-07 2017-04-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9966068B2 (en) 2013-06-08 2018-05-08 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US9300784B2 (en) 2013-06-13 2016-03-29 Apple Inc. System and method for emergency calls initiated by voice command
US9195656B2 (en) 2013-12-30 2015-11-24 Google Inc. Multilingual prosody generation
US9905220B2 (en) 2013-12-30 2018-02-27 Google Llc Multilingual prosody generation
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9668024B2 (en) 2014-06-30 2017-05-30 Apple Inc. Intelligent automated assistant for TV user interactions
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US9606986B2 (en) 2014-09-29 2017-03-28 Apple Inc. Integrated word N-gram and class M-gram language models
US9986419B2 (en) 2014-09-30 2018-05-29 Apple Inc. Social reminders
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US10074360B2 (en) 2015-08-24 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant

Also Published As

Publication number Publication date Type
FR2865846A1 (en) 2005-08-05 application
EP1560198A1 (en) 2005-08-03 application

Similar Documents

Publication Publication Date Title
US6738803B1 (en) Proxy browser providing voice enabled web application audio control for telephony devices
US6990514B1 (en) Unified messaging system using web based application server for management of messages using standardized servers
US6574599B1 (en) Voice-recognition-based methods for establishing outbound communication through a unified messaging system including intelligent calendar interface
US7519359B2 (en) Voice tagging of automated menu location
US7103548B2 (en) Audio-form presentation of text messages
US20040176123A1 (en) Communicating with public safety answering points via character based messages
US20020055350A1 (en) Apparatus and method of toggling between text messages and voice messages with a wireless communication device
US20050021826A1 (en) Gateway controller for a multimodal system that provides inter-communication among different data and voice servers through various mobile devices, and interface for that controller
US7443964B2 (en) Caller ID messaging
US8428228B1 (en) Unified communication system
US20040208302A1 (en) Caller ID messaging device
US6125284A (en) Communication system with handset for distributed processing
US6181781B1 (en) Voice mail system that downloads an applet for managing voice mail messages
US6535586B1 (en) System for the remote notification and retrieval of electronically stored messages
US20010048676A1 (en) Methods and apparatus for executing an audio attachment using an audio web retrieval telephone system
US6687341B1 (en) Network and method for the specification and delivery of customized information content via a telephone interface
US20030186722A1 (en) Method and device for real time GSM user device profile interrogation and registration
US20040252679A1 (en) Stored voice message control extensions
US20030009337A1 (en) Enhanced media gateway control protocol
US20100150321A1 (en) System and method for voice activated dialing from a home phone
US6006189A (en) Method and apparatus for storing and forwarding voice signals
US6629130B2 (en) Method and apparatus for processing electronic mail
US20040209604A1 (en) Caller ID messaging telecommunications services
US7092370B2 (en) Method and system for wireless voice channel/data channel integration
US20030006912A1 (en) Location and event triggered notification services

Legal Events

Date Code Title Description
AS Assignment

Owner name: FRANCE TELECOM, FRANCE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:FILOCHE, PASCAL;MIQUEL, PAUL;HINARD, EDOUARD;REEL/FRAME:016201/0686

Effective date: 20050207

AS Assignment

Owner name: FRANCE TELECOM, FRANCE

Free format text: CORRECTIVE ASSIGNMENT ON REEL 016201/FRAME 0686;ASSIGNORS:FILOCHE, PASCAL;MIQUEL, PAUL;HINARD, EDOUARD;REEL/FRAME:016918/0483

Effective date: 20050207