EP3809671A1 - Procédé et terminal de lecture de message - Google Patents

Procédé et terminal de lecture de message Download PDF

Info

Publication number
EP3809671A1
EP3809671A1 EP18925355.2A EP18925355A EP3809671A1 EP 3809671 A1 EP3809671 A1 EP 3809671A1 EP 18925355 A EP18925355 A EP 18925355A EP 3809671 A1 EP3809671 A1 EP 3809671A1
Authority
EP
European Patent Office
Prior art keywords
terminal
message
voice
user
keyword
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
EP18925355.2A
Other languages
German (de)
English (en)
Other versions
EP3809671A4 (fr
Inventor
Yue Zhang
Qiang TAO
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Publication of EP3809671A1 publication Critical patent/EP3809671A1/fr
Publication of EP3809671A4 publication Critical patent/EP3809671A4/fr
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • G10L13/04Details of speech synthesis systems, e.g. synthesiser structure or memory management
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/06Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
    • G10L15/065Adaptation
    • G10L15/07Adaptation to the speaker
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/04Real-time or near real-time messaging, e.g. instant messaging [IM]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/42Mailbox-related aspects, e.g. synchronisation of mailboxes
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L2015/088Word spotting
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/21Monitoring or handling of messages
    • H04L51/226Delivery according to priorities

Definitions

  • This application relates to the field of communications technologies, and in particular, to a message playing method and a terminal.
  • an instant communication application for example, Messages, WeChat, or QQ
  • a terminal receives a message from the instant communication application, the user needs to manually operate the mobile phone to view and process the message.
  • the terminal may interact with the user in a voice manner, to determine whether to play the message in the voice manner.
  • the terminal in a process in which the terminal interacts with the user in the voice manner, the terminal usually cannot recognize a real intention of the user in the voice, and consequently the terminal cannot perform processing, for example, message playing, according to the user intention. This affects user experience.
  • This application provides a message playing method and a terminal, to learn a voice command of a user, so as to recognize an intention of the user and perform a corresponding operation. This helps improve efficiency of interaction between the user and the terminal, and improve user experience.
  • a method provided in an embodiment of this application is applicable to a terminal, and the method includes: receiving, by the terminal, a first message, where the first message is text information; playing, by the terminal, a first voice in response to receiving the first message, where the first voice is used to ask a user whether to play the first message in a voice manner; detecting, by the terminal, a second voice of the user; converting, by the terminal, the second voice into a first text; if the first text does not match a first keyword, continuing, by the terminal, to detect a voice of the user, where the first keyword is a positive keyword; when the terminal detects a third voice of the user, converting, by the terminal, the third voice into a second text; if the second text matches the first keyword, playing, by the terminal, the first message in the voice manner, and recording, by the terminal, a quantity of times of using the first text; and if the quantity of times of using the first text is greater than a first threshold, adding, by the terminal,
  • a non-preset reply of a user can be learned, to determine whether the reply is a positive reply, to be specific, whether the user expects to play a message. This increases accuracy of executing a command by the terminal, and a success rate of playing a message in the voice manner by the terminal, so that the terminal is more intelligent, and user experience of using the terminal is improved.
  • the method further includes: converting, by the terminal, the first message into a fourth voice; and the broadcasting, by the terminal, the first message in the voice manner is specifically: playing, by the terminal, the fourth voice.
  • the terminal may convert the text information of the first message into a voice message (namely, the fourth voice), and then play the voice message.
  • the terminal may convert the text information of the first message into a voice message (namely, the fourth voice).
  • the terminal may directly play the voice message. This reduces a time period in which the user waits for the terminal to play the first message in the voice manner, and improves user experience.
  • the terminal may convert the text information of the first message into the fourth voice.
  • a time at which the terminal converts the text information of the first message into the voice message is not limited in this embodiment of this application.
  • the method further includes: receiving, by the terminal, a second message, where the second message is text information; playing, by the terminal, a fifth voice in response to receiving the second message, where the fifth voice is used to ask the user whether to play the second message in the voice manner; detecting, by the terminal, a sixth voice of the user; converting, by the terminal, the sixth voice into the third text; and if the third text matches an added first keyword, broadcasting, by the terminal, the second message in the voice manner.
  • the terminal can quickly recognize the intention of the user, and play the second message in the voice manner. This improves efficiency of interaction between the user and the terminal and improves user experience.
  • the method before the playing, by the terminal, a first voice, the method further includes: if the terminal determines that the first message belongs to a preset application, and/or a sender of the first message belongs to a preset contact group, and/or the first message includes a second keyword, determining, by the terminal, to play the first voice.
  • the terminal may further filter the message.
  • the user selects, based on a requirement, a specific message to play the message in the voice manner, to prevent excessive messages from playing in the voice manner and disturbing the user. This improves user experience.
  • the method before the playing, by the terminal, a first voice, the method further includes: receiving, by the terminal, a third message when receiving the first message; and determining, by the terminal based on a preset priority sequence, that a priority of the first message is higher than a priority of the third message.
  • the terminal may determine a message playing sequence based on a preset priority sequence, to meet diversified requirements of the user and improve user experience.
  • the method further includes: displaying, by the terminal, prompt information, where the prompt information is used to notify the terminal that the first keyword has been updated.
  • the method further includes: if the terminal has not detected the voice of the user within a preset time period, or if the terminal has not detected the voice that is of the user and that matches the first keyword within a preset time period, determining, by the terminal, not to play the first message in the voice manner.
  • the method further includes: if the terminal detects, within a preset time period, that a quantity of times of using the voice that is of the user and that does not match the first keyword is greater than a second threshold, determining, by the terminal, not to play the first message in the voice manner.
  • the first message is a message of an instant communication application.
  • a message playing method is applicable to a terminal, and the method includes: receiving, by the terminal, a first message, where the first message is text information; playing, by the terminal, a first voice in response to receiving the first message, where the first voice is used to ask a user whether to play the first message in a voice manner; detecting, by the terminal, a second voice of the user; converting, by the terminal, the second voice into a first text; if the first text does not match a first keyword, continuing, by the terminal, to detect a voice of the user, where the first keyword includes a positive keyword and a negative keyword; when the terminal detects a third voice of the user, converting, by the terminal, the third voice into a second text; and if the second text matches the positive keyword, playing, by the terminal, the first message in the voice manner, and recording, by the terminal, a quantity of times of using the first text; and if the quantity of times of using the first text is greater than a first
  • a non-preset reply of a user can be learned, to determine whether the reply is a positive reply or a negative reply, to be specific, whether the user expects to play a message. This increases accuracy of executing a command by the terminal, and a success rate of playing a message in the voice manner by the terminal, so that the terminal is more intelligent, and user experience of using the terminal is improved.
  • the method further includes: receiving, by the terminal, a second message, where the second message is text information; playing, by the terminal, a fourth voice in response to receiving the second message, where the fourth voice is used to ask the user whether to play the second message in the voice manner; detecting, by the terminal, a fifth voice of the user; converting, by the terminal, the fifth voice into the third text; and if the third text matches an added positive keyword, playing, by the terminal, the second message in the voice manner; or if the third text matches an added negative keyword, determining, by the terminal, not to broadcast the second message in the voice manner.
  • the method before the playing, by the terminal, a first voice, the method further includes: if the terminal determines that the first message belongs to a preset application, and/or a sender of the first message belongs to a preset contact group, and/or the first message includes a second keyword, determining, by the terminal, to play the first voice.
  • the method before the playing, by the terminal, a first voice, the method further includes: receiving, by the terminal, a third message when receiving the first message; and determining, by the terminal based on a preset priority sequence, that a priority of the first message is higher than a priority of the third message.
  • the method further includes: displaying, by the terminal, prompt information, where the prompt information is used to notify the user that the first keyword has been updated.
  • the method further includes: if the terminal has not detected the voice of the user within a preset time period, or if the terminal has not detected the voice that is of the user and that matches the first keyword within a preset time period, determining, by the terminal, not to play the first message in the voice manner.
  • the method further includes: if the terminal detects, within a preset time period, that a quantity of times of using the voice that is of the user and that does not match the first keyword is greater than a second threshold, determining, by the terminal, not to play the first message in the voice manner.
  • the first message is a message of an instant communication application.
  • a terminal includes a processor, a memory, and a touchscreen.
  • the memory and the touchscreen are coupled to the processor, the memory is configured to store computer program code, and the computer program code includes a computer instruction.
  • the processor reads the computer instruction in the memory to perform the method in any one of the first aspect and the possible implementations of the first aspect.
  • a terminal includes a processor, a memory, and a touchscreen.
  • the memory and the touchscreen are coupled to the processor, the memory is configured to store computer program code, and the computer program code includes a computer instruction.
  • the processor reads the computer instruction in the memory to perform the method in any one of the second aspect and the possible implementations of the second aspect.
  • a computer storage medium includes a computer instruction.
  • the terminal is enabled to perform the method in any one of the first aspect and the possible implementations of the first aspect.
  • a computer storage medium includes a computer instruction.
  • the terminal is enabled to perform the method in any one of the second aspect and the possible implementations of the second aspect.
  • a computer program product is provided.
  • the computer program product runs on a computer, the computer is enabled to perform the method in any one of the first aspect and the possible implementations of the first aspect.
  • a computer program product is provided.
  • the computer program product runs on a computer, the computer is enabled to perform the method in any one of the second aspect and the possible implementations of the second aspect.
  • first and second are merely intended for a purpose of description, and shall not be understood as an indication or implication of relative importance or implicit indication of a quantity of indicated technical features. Therefore, a feature limited by “first” or “second” may explicitly or implicitly include one or more features. In descriptions of embodiments of this application, unless otherwise stated, "a plurality of' means two or more than two.
  • a terminal in this application may be a mobile phone, a tablet computer, a personal computer (Personal Computer, PC), a personal digital assistant (personal digital assistant, PDA), a smartwatch, a netbook, a wearable electronic device, an augmented reality (Augmented Reality, AR) technology device, a virtual reality (Virtual Reality, VR) device, or the like.
  • a specific form of the terminal is not specially limited in this application.
  • FIG. 1 is an example of a structural block diagram of a terminal 100 according to an embodiment of the present invention.
  • the terminal 100 may include a processor 110, an external memory interface 120, an internal memory 121, a USB interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2, a radio frequency module 150, a communications module 160, an audio module 170, a speaker 170A, a telephone receiver 170B, a microphone 170C, a headset interface 170D, a sensor module 180, a key 190, a motor 191, an indicator 192, a camera 193, a display screen 194, a SIM card interface 195, and the like.
  • a processor 110 an external memory interface 120, an internal memory 121, a USB interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2, a radio frequency module 150, a communications module 160, an audio module 170, a speaker 170A, a telephone receiver 170B, a microphone 170C, a headset interface 170D, a sensor module 180
  • the sensor module may include a pressure sensor 180A, a gyro sensor 180B, a barometric pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a range sensor 180F, an optical proximity sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, an ambient light sensor 180L, a bone conduction sensor 180M, and the like.
  • a schematic structure in this embodiment of the present invention constitutes no limitation on the terminal 100.
  • the terminal 100 may include more or fewer components than those shown in the figure, or combine some components, or split some components, or have different component arrangements.
  • the components in the figure may be implemented by using hardware, software, or a combination of software and hardware.
  • the processor 110 may include one or more processing units.
  • the processor 110 may include an application processor (application processor, AP), a modem processor, a graphics processing unit (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), a controller, a memory, a video codec, a digital signal processor (digital signal processor, DSP), a baseband processor, and/or a neural processing unit (Neural-network Processing Unit, NPU).
  • application processor application processor, AP
  • modem processor graphics processing unit
  • ISP image signal processor
  • controller a memory
  • video codec digital signal processor
  • DSP digital signal processor
  • baseband processor baseband processor
  • NPU neural processing unit
  • Different processing units may be independent devices, or may be integrated into a same processor.
  • the controller may be a decision maker that directs each component of the terminal 100 to carry out coordinate work according to an instruction.
  • the controller is a nerve center and a command center of the terminal 100.
  • the controller generates an operation control signal based on an instruction operation code and a time sequence signal, to complete control of instruction fetching and instruction execution.
  • the application processor is configured to: obtain a voice of a user, convert the obtained voice of the user into a text, and further match the converted text with a pre-stored keyword and record a quantity of times of using the text, and when the quantity of times of using the text reaches a preset quantity of times, add the text to the corresponding keyword.
  • the application processor may be further configured to: obtain, by using the radio frequency module, the communications module, or the like, a text message sent by another terminal or server to the terminal, and convert the received text message into a voice or the like.
  • the memory may be further disposed in the processor 110 to store an instruction and data.
  • the memory in the processor is a cache.
  • the memory may store an instruction or data that is just used or cyclically used by the processor. If the processor needs to use the instruction or the data again, the processor may directly invoke the instruction or the data from the memory, to avoid repeated access and reduce a waiting time of the processor. Therefore, system efficiency is improved.
  • the terminal may store, in the memory in the processor 110, a keyword preset by the user, for example, a keyword of a positive reply and/or a keyword of a negative reply.
  • the terminal may further store, in the memory, recorded content of a voice command, a quantity of times of using the voice command, and the like.
  • the terminal may alternatively store data in the internal memory 121 or an external memory. This is not specifically limited in this embodiment of this application.
  • the processor 110 may include an interface.
  • the interface may include an integrated circuit (inter-integrated circuit, I2C) interface, an inter-integrated circuit sound (inter-integrated circuit sound, I2S) interface, a pulse-code modulation (pulse code modulation, PCM) interface, a universal asynchronous receiver/transmitter (universal asynchronous receiver/transmitter, UART) interface, a mobile industry processor interface (mobile industry processor interface, MIPI), a general-purpose input/output (general-purpose input/output, GPIO) interface, a subscriber identity module (subscriber identity module, SIM) interface, a universal serial bus (universal serial bus, USB) interface, and/or the like.
  • I2C integrated circuit
  • I2S inter-integrated circuit sound
  • PCM pulse-code modulation
  • PCM pulse-code modulation
  • UART universal asynchronous receiver/transmitter
  • MIPI mobile industry processor interface
  • GPIO general-purpose input/output
  • SIM subscriber identity module
  • the I2C interface is a bidirectional synchronous serial bus, and includes a serial data line (serial data line, SDA) and a serial clock line (derail clock line, SCL).
  • the processor may include a plurality of groups of I2C buses.
  • the processor may be coupled to the touch sensor, a charger, a flash light, the camera, and the like through different I2C bus interfaces.
  • the processor may be coupled to the touch sensor through the I2C interface, so that the processor communicates with the touch sensor through the I2C bus interface, to implement a touch function of the terminal 100.
  • the I2S interface may be configured for audio communication.
  • the processor may include a plurality of groups of I2S buses.
  • the processor 110 may be coupled to the audio module through the I2S bus, to implement communication between the processor and the audio module.
  • the audio module may transfer an audio signal to the communications module through the I2S interface, to implement a function of answering a call by using a Bluetooth headset.
  • the PCM interface may also be configured for audio communication, and sample, quantize, and code an analog signal.
  • the audio module may be coupled to the communications module through the PCM bus interface.
  • the audio module may alternatively transfer an audio signal to the communications module through the PCM interface, to implement the function of answering a call by using the Bluetooth headset.
  • Both the I2S interface and the PCM interface may be configured for audio communication, and sampling rates of the two interfaces are different.
  • the UART interface is a universal serial data bus, and is configured for asynchronous communication.
  • the bus is a bidirectional communications bus.
  • the bus converts to-be-transmitted data between serial communication and parallel communication.
  • the UART interface is usually configured to connect the processor to the communications module 160.
  • the processor communicates with a Bluetooth module through the UART interface, to implement a Bluetooth function.
  • the audio module may transfer an audio signal to the communications module through the UART interface, to implement a function of playing music by using the Bluetooth headset.
  • the terminal can implement voice playing of a message through any one or more of the I2S interface, the PCM interface, and the UART interface, transfer a recorded user voice to the processor, and the like.
  • the MIPI interface may be configured to connect the processor to a peripheral device, for example, the display screen or the camera.
  • the MIPI interface includes a camera serial interface (camera serial interface, CSI), a display serial interface (display serial interface, DSI), and the like.
  • the processor communicates with the camera through the CSI interface, to implement a photographing function of the terminal 100.
  • the processor communicates with the display screen through the DSI interface, to implement a display function of the terminal 100.
  • the terminal may display, through the MIPI interface, an interface diagram, for example, a setting interface of the user, mentioned in a process of performing voice playing by the terminal.
  • the GPIO interface may be configured by using software.
  • the GPIO interface may be configured as a control signal, or may be configured as a data signal.
  • the GPIO interface may be configured to connect the processor to the camera, the display screen, the communications module, the audio module, a sensor, and the like.
  • the GPIO interface may alternatively be configured as the I2C interface, the I2S interface, the UART interface, the MIPI interface, or the like.
  • the USB interface 130 may be a mini USB interface, a micro USB interface, a USB Type C interface, or the like.
  • the USB interface may be configured to connect to the charger to charge the terminal 100, or may be configured to transmit data between the terminal 100 and a peripheral device.
  • the USB interface may alternatively be connected to a headset, to play audio by using the headset.
  • the USB interface may alternatively be connected to another electronic device, for example, an AR device.
  • An interface connection relationship between the modules shown in this embodiment of the present invention is merely an example for description, and does not constitute a limitation on the structure of the terminal 100.
  • the terminal 100 may use different interface connection manners or a combination of a plurality of interface connection manners in this embodiment of the present invention.
  • the charging management module 140 is configured to receive charging input from the charger.
  • the charger may be a wireless charger or a wired charger.
  • the power management module 141 is connected to the battery 142, the charging management module 140, and the processor 110.
  • the power management module receives input of the battery and/or the charging management module, and supplies power to the processor, the internal memory, the external memory, the display screen, the camera, the communications module, and the like.
  • a wireless communication function of the terminal 100 may be implemented by using the antenna 1, the antenna 2, the radio frequency module 150, the communications module 160, the modem processor, the baseband processor, and the like.
  • the antenna 1 and the antenna 2 are configured to transmit and receive an electromagnetic wave signal.
  • Each antenna in the terminal 100 may be configured to cover one or more communication frequency bands.
  • the radio frequency module 150 may provide a communications processor module that is applied to the terminal 100 and that includes a wireless communication solution, for example, 2G/3G/4G/5G.
  • the radio frequency module 150 may include at least one filter, a switch, a power amplifier, a low noise amplifier (Low Noise Amplifier, LNA), and the like.
  • the modem processor may include a modulator and a demodulator.
  • the modulator is configured to modulate a to-be-sent low-frequency baseband signal into a medium-high frequency signal.
  • the demodulator is configured to demodulate a received electromagnetic wave signal into a low-frequency baseband signal.
  • the demodulator transfers the low-frequency baseband signal obtained through demodulation to the baseband processor for processing.
  • the low-frequency baseband signal is processed by the baseband processor, and is then transferred to the application processor.
  • the application processor outputs a sound signal by using an audio device (not limited to the speaker, the telephone receiver, or the like), or displays an image or a video by using the display screen.
  • the communications module 160 may provide a communications processor module of wireless communication solutions, including a wireless local area network (wireless local area networks, WLAN), Bluetooth (bluetooth, BT), a global navigation satellite system (global navigation satellite system, GNSS), frequency modulation (frequency modulation, FM), a near field communication (near field communication, NFC) technology, an infrared (infrared, IR) technology, and the like that are applied to the terminal 100.
  • the communications module 160 may be one or more devices integrating at least one communications processor module.
  • the communications module receives an electromagnetic wave by using the antenna 2, performs frequency modulation and filtering processing on an electromagnetic wave signal, and sends a processed signal to the processor.
  • the communications module 160 may alternatively receive a to-be-sent signal from the processor, perform frequency modulation and amplification on the signal, and convert the signal into an electromagnetic wave for radiation by using the antenna 2.
  • the voice of the user may be recorded by using the microphone in the Bluetooth headset (or a Bluetooth sound box, or the like), and the recorded voice is transferred to the processor 110 by using a Bluetooth communications processor module and the audio module 170.
  • the terminal may further play the voice by using the Bluetooth headset (or the Bluetooth sound box, or the like) through the audio module 170 and the Bluetooth communication processor module.
  • the antenna 1 of the terminal 100 is coupled to the radio frequency module, and the antenna 2 is coupled to the communications module.
  • the wireless communications technology may include a global system for mobile communications (global system for mobile communications, GSM), a general packet radio service (general packet radio service, GPRS), code division multiple access (code division multiple access, CDMA), wideband code division multiple access (wideband code division multiple access, WCDMA), time-division code division multiple access (time-division code division multiple access, TD-SCDMA), long term evolution (long term evolution, LTE), the BT, the GNSS, the WLAN, the NFC, the FM, the IR technology, and/or the like.
  • GSM global system for mobile communications
  • GPRS general packet radio service
  • code division multiple access code division multiple access
  • CDMA wideband code division multiple access
  • WCDMA wideband code division multiple access
  • time-division code division multiple access time-division code division multiple access
  • TD-SCDMA time-division code division multiple access
  • the GNSS may include a global positioning system (global positioning system, GPS), the global navigation satellite system (global navigation satellite system, GLONASS), a Beidou navigation satellite system (beidou navigation satellite system, BDS), a quasi-zenith satellite system (quasi-zenith satellite system, QZSS), and/or a satellite-based augmentation system (satellite based augmentation systems, SBAS).
  • GPS global positioning system
  • GLONASS global navigation satellite system
  • Beidou navigation satellite system beidou navigation satellite system
  • BDS Beidou navigation satellite system
  • QZSS quasi-zenith satellite system
  • SBAS satellite-based augmentation system
  • the terminal may receive, by using the antenna 1 and the radio frequency module, a message sent by another terminal, for example, a Messages message.
  • the terminal may alternatively receive, by using the antenna 2 and the communications module, a message sent by another terminal, for example, a WeChat message or a QQ message.
  • the message is not specifically limited in this embodiment of this application.
  • the terminal 100 implements the display function by using the GPU, the display screen 194, the application processor, and the like.
  • the GPU is a microprocessor for image processing, and is connected to the display screen and the application processor.
  • the GPU is configured to perform mathematical and geometric calculation, and is configured to render an image.
  • the processor 110 may include one or more GPUs that execute a program instruction to generate or change display information.
  • the display screen 194 is configured to display an image, a video, and the like.
  • the display screen includes a display panel.
  • the display panel may use an LCD (liquid crystal display, liquid crystal display screen), an OLED (organic light-emitting diode, organic light-emitting diode), an active-matrix organic light-emitting diode or an active-matrix organic light-emitting diode (active-matrix organic light emitting diode, AMOLED), a flexible light-emitting diode (flex light-emitting diode, FLED), a miniled, a microLed, a micro-oLed, a quantum dot light-emitting diode (quantum dot light emitting diodes, QLED), or the like.
  • the terminal 100 may include one display or N display screens, where N is a positive integer greater than 1.
  • the mobile terminal 100 can implement the photographing function by using the ISP, the camera 193, the video codec, the GPU, the display screen, the application processor, and the like.
  • the external memory interface 120 may be connected to an external storage card, for example, a micro SD card, to extend a storage capability of the terminal 100.
  • the external storage card communicates with the processor through the external memory interface, to implement a data storage function. For example, a file, for example, music or a video, is stored into the external storage card.
  • the internal memory 121 may be configured to store computer-executable program code, and the executable program code includes an instruction.
  • the processor 110 runs the instruction stored in the internal memory 121, to implement various function applications and data processing of the terminal 100.
  • the memory 121 may include a program storage area and a data storage area.
  • the program storage area may store an operating system, an application program required by at least one function (for example, a sound playing function or an image playing function), and the like.
  • the data storage area may store data (for example, audio data or a phone book) created during use of the terminal 100, and the like.
  • the memory 121 may include a high-speed random access memory, and may further include a nonvolatile memory, for example, at least one magnetic disk storage device, a flash storage device, another volatile solid-state storage device, a universal flash storage (universal flash storage, UFS), or the like.
  • a nonvolatile memory for example, at least one magnetic disk storage device, a flash storage device, another volatile solid-state storage device, a universal flash storage (universal flash storage, UFS), or the like.
  • the terminal 100 can implement an audio function by using the audio module 170, the speaker 170A, the telephone receiver 170B, the microphone 170C, the headset interface 170D, the application processor, and the like, for example, music playing or recording.
  • the audio module 170 is configured to convert digital audio information into an analog audio signal for output, or is configured to convert analog audio input into a digital audio signal.
  • the audio module may be further configured to encode and decode an audio signal.
  • the audio module may be disposed in the processor 110, or some functional modules in the audio module are disposed in the processor 110.
  • the speaker 170A also referred to as a "loudspeaker" is configured to convert an audio electrical signal into a sound signal.
  • the terminal 100 may play music by using the speaker, or receive a hands-free call.
  • the telephone receiver 170B also referred to as an "earpiece" is configured to convert an audio electrical signal into a sound signal.
  • the telephone receiver may be put close to a human ear to receive a voice.
  • the microphone 170C also referred to as a "mic” or a " sound conducting device" is configured to convert a sound signal into an electrical signal.
  • a user may make a sound by moving a mouth of the user near the microphone to input a sound signal to the microphone.
  • At least one microphone may be disposed in the terminal 100.
  • two microphones may be disposed in the terminal 100, to collect a sound signal and further implement a noise reduction function.
  • three, four, or more microphones may alternatively be disposed in the terminal 100, to collect a sound signal, reduce noise, and further recognize a sound source, implement a directional recording function, and the like.
  • the headset interface 170D is connected to a wired headset.
  • the headset interface may be a USB interface, or may be a 3.5 mm open mobile terminal platform (open mobile terminal platform, OMTP) standard interface or a cellular telecommunications industry association of the USA (cellular telecommunications industry association of the USA, CTIA) standard interface.
  • OMTP open mobile terminal platform
  • CTIA cellular telecommunications industry association of the USA
  • the terminal may record the voice of the user by using the microphone 170C, and transfer the recorded voice to the processor 110 by using the audio module 170. After the processor 110 converts the received text message into a voice, the terminal may further play the voice by using the speaker through the audio module 170. In some other embodiments of this application, the terminal may record the voice of the user by using a microphone in the wired headset, and transfer the recorded voice to the processor 110 by using the headset interface 170D and the audio module 170. After the processor 110 converts the received text message into a voice, the terminal may further play the voice by using the wired headset through the audio module 170 and the headset interface 170D.
  • the touch sensor 180K is also referred to as a "touch panel”, may be set on the display screen, and is configured to detect a touch operation performed on or near the touch sensor 180K.
  • the touch sensor 180K may transfer the detected touch operation to the application processor, to determine a touch event type, and provide corresponding visual output by using the display screen.
  • the key 190 includes a power key, a volume key, and the like.
  • the key may be a mechanical key, or may be a touch key.
  • the terminal 100 receives key input, and generates key signal input related to a user setting and function control of the terminal 100.
  • the motor 191 may generate a vibration prompt.
  • the motor may be used for an incoming call vibration prompt, or may be used for a touch vibration feedback.
  • touch operations performed on different applications may correspond to different vibration feedback effects.
  • Different touch operations performed in different areas of the display screen may also correspond to different vibration feedback effects.
  • Different application scenarios for example, a time reminder, information receiving, an alarm clock, a game, and the like
  • Customization of a touch vibration feedback effect may further be supported.
  • the indicator 192 may be an indicator light that may be configured to indicate a charging status and a battery power change, or may be configured to indicate a message, a missed call, a notification, and the like.
  • the SIM card interface 195 is connected to a subscriber identity module (subscriber identity module, SIM).
  • SIM subscriber identity module
  • a SIM card may be inserted into the SIM card interface or removed from the SIM card interface, to implement contact with or separation from the terminal 100.
  • the terminal 100 may support one SIM card interface or N SIM card interfaces, where N is a positive integer greater than 1.
  • the SIM card interface may support a nano SIM card, a micro SIM card, a SIM card, and the like.
  • a plurality of cards may be inserted into a same SIM card interface simultaneously.
  • the plurality of cards may be of a same type or different types.
  • the SIM card interface may also be compatible with different types of SIM cards, and the SIM card interface may alternatively be compatible with the external storage card.
  • the terminal 100 interacts with a network by using the SIM card, to implement functions such as calling and data communication.
  • the terminal 100 uses an eSIM, namely, an embedded SIM card.
  • the eSIM card may be embedded in the terminal 100, and cannot be separated from the terminal 100.
  • a software system of the terminal 100 may use a layered architecture, an event-driven architecture, a microkernel architecture, a microservice architecture, or a cloud architecture.
  • an Android system with the layered architecture is used as an example to describe a software structure of the terminal 100.
  • the layered architecture software is divided into several layers, and each layer has a clear role and task.
  • the layers communicate with each other through an interface.
  • the Android system is divided into four layers, namely, an application program layer, an application program framework layer, Android runtime (Android runtime) and a system library, and a kernel layer from top to bottom.
  • the application program layer may include a series of application program packages.
  • an application program package may include application programs such as camera, gallery, calendar, phone, map, navigation, WLAN, Bluetooth, music, video, Messages, WeChat, QQ, and setting.
  • the application program package mainly includes instant communication applications, including but not limited to the application programs such as Messages, WeChat, and QQ.
  • a setting application is further involved, to provide an interface for setting a message played in a voice manner for a user.
  • Content of setting includes but is not limited to a preset application, a preset contact, a preset contact group, a preset second keyword, a playing priority, and the like.
  • the application program framework layer provides an application programming interface (application programming interface, API) and a programming framework for an application program at the application program layer.
  • the application program framework layer includes some predefined functions.
  • the application program framework layer may include a window manager, a content provider, a view system, a phone manager, a resource manager, a notification manager, and the like.
  • the window manager is configured to manage a window program.
  • the window manager may obtain a size of a display screen, determine whether there is a status bar, lock a screen, take a screenshot, and the like.
  • the content provider is configured to: store and obtain data, and make the data accessible to an application program.
  • the data may include a video, an image, audio, calls that are made and answered, a browsing history and a bookmark, a phone book, and the like.
  • the view system includes visual controls such as a control for displaying a character and a control for display a picture, and the view system may be configured to construct an application program.
  • a display interface may include one or more views.
  • a display interface including a Messages notification icon may include a view for displaying a character and a view for displaying a picture.
  • the phone manager is configured to provide a communication function of the terminal 100, for example, management of call statuses (including a connected state, a disconnected state, and the like).
  • the resource manager provides various resources such as a localized character string, an icon, a picture, a layout file, and a video file for an application program.
  • the notification manager enables an application program to display notification information in a status bar, and may be configured to convey a notification-type message.
  • the notification-type message may automatically disappear after the message is displayed for a short period of time without user interaction.
  • the notification manager is configured to provide a notification of download completion, a message reminder, and the like.
  • the notification manager may alternatively be a notification that appears on the top of a status bar of a system in a text form of a graph or a scroll bar, for example, a notification of an application program running in the background or a notification that appears on a screen in a form of a dialog window. For example, text information is prompted in the status bar, a prompt tone is produced, the terminal vibrates, or the indicator light blinks.
  • the application program framework layer may further include a voice playing system, and the voice playing system provides a voice playing service for an instant message.
  • the voice playing system may be an independent module at the application framework layer, or the voice playing system may invoke another module at the application program framework layer to jointly complete a voice playing function of the instant message. This is not specifically limited in this embodiment of this application.
  • the Android runtime includes a kernel library and a virtual machine, and the Android runtime is responsible for scheduling and management of the Android system.
  • the kernel library includes two parts: a function that needs to be invoked by a Java language and a kernel library of Android.
  • the application program layer and the application program framework layer run on the virtual machine.
  • the virtual machine executes a Java file at the application program layer and the application program framework layer as a binary file.
  • the virtual machine is configured to perform functions such as object lifecycle management, stack management, thread management, security and exception management, and garbage collection.
  • the system library may include a plurality of functional modules, for example, a surface manager (surface manager), a media library (Media Libraries), a three-dimensional graphics processing library (OpenGL ES), and a 2D graphics engine (SGL).
  • a surface manager surface manager
  • Media Libraries media libraries
  • OpenGL ES three-dimensional graphics processing library
  • SGL 2D graphics engine
  • the surface manager is configured to: manage a display subsystem, and provide fusion of 2D and 3D layers for a plurality of application programs.
  • the media library supports playback and recording of a plurality of commonly used audio and video formats, static image files, and the like.
  • the media library may support a plurality of audio and video coding formats, for example, MPEG4, H.264, MP3, AAC, AMR, JPG, and PNG.
  • OpenGL ES is configured to implement three-dimensional graphics drawing, image rendering, composition, layer processing, and the like.
  • SGL is a 2D drawing engine.
  • the kernel layer is a layer between hardware and software.
  • the kernel layer includes at least a display driver, a camera driver, an audio driver, and a sensor driver.
  • the following uses an example in which a message of a Messages application is processed to describe a working process of software and hardware of the terminal 100.
  • the Messages application at the application program layer may invoke the display driver at the kernel layer, to display prompt information of the message on a touchscreen at a hardware layer, so as to prompt a user to view the message.
  • the touchscreen may be triggered to report, to the kernel layer by using a corresponding driver, a touch event (for example, a parameter including a touch point location, a time, or the like) generated by a touch action of the user.
  • a touch event for example, a parameter including a touch point location, a time, or the like
  • the kernel layer invokes a corresponding API to distribute the touch event to the Messages application.
  • the terminal starts the Messages application, and displays an interface for viewing the message. In this way, the user can view content of the message.
  • the Messages application at the application program layer may invoke the voice playing system at the framework layer.
  • the voice playing system may play, by invoking the audio driver at the kernel layer and by using an audio output device (for example, a Bluetooth headset or a speaker), a voice for asking the user whether to play the message.
  • an audio input device for example, a Bluetooth headset or a microphone
  • the kernel layer invokes a corresponding API to distribute the event to the voice playing system of the framework layer.
  • the voice playing system determines, based on the event, whether to play the message in a voice manner. Specifically, the voice playing system may convert the reported voice of the user into a text, and match the converted text with a pre-stored keyword (a keyword of a positive reply and/or a keyword of a negative reply). If the converted text matches the keyword of the positive reply, it is determined that the message is played in the voice manner. In this case, the voice playing system converts the message into a voice message, invokes the audio driver at the kernel layer, and plays the voice message by using the audio output device. If the converted text matches the keyword of the negative reply, it is determined that the message is not played in the voice manner. In this way, when the user is inconvenient to manually operate the terminal, the user can also process the message.
  • a pre-stored keyword a keyword of a positive reply and/or a keyword of a negative reply
  • the voice playing system may further record a quantity of times of using the converted text.
  • the quantity of times of using the converted text reaches a predetermined quantity of times, the text may be further added to the keyword, to learn the voice of the user.
  • the voice playing system may further invoke the display driver at the kernel layer, to display, by using the touchscreen, an interface in embodiments of this application, for example, interface diagrams shown in FIG. 7(1) to FIG. 7(4) and FIG. 8(1) to FIG. 8(6) .
  • an embodiment of this application provides a method for playing an instant message in a voice manner. Further, in consideration of factors such as a factor that the user may forget a preset voice command or a language habit of the user, the voice command of the user is not a voice command preset by the terminal. Therefore, the terminal cannot recognize an intention of the user, and cannot perform an operation expected by the user. Therefore, in the technical solutions provided in the embodiments of this application, the voice command of the user can be learned, to automatically recognize a real meaning of the voice command of the user. This improves use efficiency of the terminal and improves user experience.
  • a terminal asks a user whether to play a newly received message in a voice manner.
  • FIG. 3A and FIG. 3B are a schematic flowchart of a method for playing a message in a voice manner according to an embodiment of this application. The method specifically includes the following steps.
  • S301 A terminal receives a first message.
  • the terminal receives the first message sent by another terminal or a server.
  • the first message may be a message of an instant communication application, for example, a message of a Messages application, a message of a WeChat application, or a message of a QQ application.
  • S302 The terminal asks, in a voice manner, whether to play the first message.
  • the terminal After newly receiving the first message, the terminal displays prompt information of the first message in a terminal interface.
  • the terminal may ask, in the voice manner, the user whether to play the first message before, when, or after displaying the prompt information of the first message.
  • the terminal may perform voice playing by using an audio device, for example, a speaker, a wired headset, a wireless headset, a Bluetooth sound box, or a Bluetooth vehicle-mounted device. This is not specifically limited in this embodiment of this application.
  • FIG. 8 (1) shows an interface 801 displayed by the terminal.
  • the interface 801 may display a status bar 802, a message prompt box 803, a pattern 804, a time widget, and the like.
  • the status bar 802 may include a name of an operator (for example, China Mobile), a time, a WiFi icon, signal strength, current remaining power, and the like.
  • the interface 801 is an interface diagram in which the terminal asks, in the voice manner, the user whether to play the first message.
  • the terminal may dynamically display the pattern 804, or change a color, a gray scale, or the like of the pattern 804, to notify the user that the terminal is playing a voice.
  • the terminal may display text information, to notify that the terminal is asking the user whether to play the first message.
  • a prompt manner of the terminal is not limited in this embodiment of this application.
  • S303 The terminal detects a first voice of the user.
  • the terminal records the first voice of the user by using an audio input device, and sends the recorded first voice of the user to an application processor of the terminal for processing.
  • FIG. 8 (2) shows an interface 805 displayed by the terminal.
  • the interface 805 may display a status bar 802, a message prompt box 803, a pattern 806, a time widget, and the like.
  • the interface 805 is an interface in which the terminal detects a voice of the user.
  • the terminal may dynamically display the pattern 806, or change a color, a gray scale, or the like of the pattern 806, to notify the user that the terminal detects the voice of the user or that the terminal is processing the detected voice of the user.
  • the terminal may display text information, to notify that the terminal detects the voice of the user and is processing the detected voice of the user.
  • a prompt manner in which the terminal detects the voice of the user is not limited in this embodiment of this application.
  • S304 The terminal converts the first voice into text information, and records the text information as a first command.
  • S305 The terminal matches the first command with a first keyword pre-stored in the terminal.
  • the first keyword may include a command preset by the terminal, for example, a command of a positive reply or a command of a negative reply.
  • the first keyword may be a default keyword of the terminal, or may be set by the user.
  • the first keyword may be learned by the terminal. For a specific learning method, refer to the following description.
  • the preset first keyword may be a keyword of a positive reply.
  • the first command matches the first keyword, it can be determined that the first command is a command that the user expects to play the first message in the voice manner. If the first command does not match the first keyword, the first command needs to be learned by using the method (as shown in FIG. 3A and FIG. 3B ) in this embodiment of this application, to determine whether the first command is a positive reply.
  • the preset first keyword may alternatively be a keyword of a negative reply.
  • the preset first keyword may alternatively include both a keyword of a positive reply and a keyword of a negative reply. In this case, the terminal needs to separately perform processing based on a type of first keyword that matches the first command. This case is explained below. This is not limited in this embodiment of this application.
  • step S306 to S313 an example in which the first keyword is the positive reply is used to describe a process in which the terminal learns the first command. If the first command matches the first keyword, it is determined that the first command is a positive reply, and the user expects to play the first message in the voice manner. Therefore, the terminal plays the first message in the voice manner, that is, performs step S306. If the first command does not match the first keyword, step S307 is performed.
  • S306 The terminal plays the first message in the voice manner.
  • the step of playing the first message in the voice manner specifically includes: playing content of the first message in the voice manner, or playing a name of an application to which the first message belongs, a name of a sender of the first message, or the like.
  • the terminal may convert the text information of the first message into a voice message, and then play the voice message.
  • the terminal may convert the text information of the first message into a voice message.
  • the terminal may directly play the voice message. This reduces a time period in which the user waits for the terminal to play the first message in the voice manner, and improves user experience. For example, after receiving the first message, or after receiving the first voice of the user, or after converting the first voice of the user into the first command, or after the user matches the first command with the preset first keyword, the terminal may convert the text information of the first message into a voice message.
  • a time at which the terminal converts the text information of the first message into the voice message is not limited in this embodiment of this application.
  • FIG. 8 (3) shows an interface 807 displayed by the terminal.
  • the interface 807 may display a status bar 802, a message prompt box 803, a pattern 804, a time widget, and the like.
  • the interface 807 is an interface in which the terminal is playing the first message.
  • the terminal may dynamically display the pattern 804, or change a color, a gray scale, or the like of the pattern 804, to notify the user that the terminal is playing the first message.
  • the terminal may display prompt information, to notify the user that the terminal is playing the first message. It should be noted that a prompt manner in which the terminal is playing the first message is not limited in this embodiment of this application.
  • S307 The terminal does not play the first message temporarily, and continues to listen to a voice of the user.
  • FIG. 8 (5) shows an interface 809 displayed by the terminal.
  • the interface 809 may display a status bar 802, a message prompt box 803, a pattern 806, a time widget, and the like.
  • the interface 809 is an interface in which the terminal continues to listen to a voice of the user because the terminal fails to recognize the voice command of the user.
  • the terminal may dynamically display the pattern 806, or change a color, a gray scale, or the like of the pattern 806, to notify the user that the terminal fails to recognize the voice command of the user.
  • the terminal may display text information, to notify the user that the terminal continues to listen to the voice of the user because the terminal fails to recognize the voice command of the user.
  • the terminal may notify, in the voice manner, the user that the terminal fails to recognize an indication of the user, and then continues to listen to the voice of the user. It should be noted that a specific prompt form of the terminal is not limited in this embodiment of this application.
  • S308 The terminal detects a second voice of the user.
  • S309 The terminal converts the detected second voice into text information, and records the text information as a second command.
  • step S310 The terminal matches the second command with the first keyword preset by the terminal. If the second command does not match the preset first keyword, step S311 is performed. If the second command matches the preset first keyword, step S312 is performed.
  • S311 The terminal does not play the first message temporarily, and continues to listen to a voice of the user.
  • the terminal may end this procedure. In other words, the terminal considers by default that the user does not expect to play the first message in the voice manner.
  • the terminal may end this procedure. In other words, the terminal considers by default that the user does not expect to play the first message in the voice manner.
  • FIG. 8 (4) shows an interface 808 displayed by the terminal.
  • the interface 808 may display a status bar 802, a message prompt box 803, a pattern 804, a time widget, and the like.
  • the interface 808 is an interface in which the terminal determines not to play the first message.
  • the terminal may change a color, a gray scale, or the like of the pattern 804, to notify the user that the terminal does not play the first message.
  • the terminal may display text information, to notify the user that the terminal does not play the first message. It should be noted that a manner in which the terminal notifies the user that the terminal does not play the first message is not limited in this embodiment of this application.
  • S312 The terminal plays the first message in the voice manner.
  • the terminal records content of the first command and a quantity of times of using the first command.
  • the first command converted from the first voice is different from the preset positive reply.
  • the first voice of the user is not a reply made for inquiry of the terminal.
  • the first voice may be a dialogue between the user and another person. After receiving the first voice, the terminal mistakenly considers that the first voice is the first command of the user.
  • the terminal needs to record the content of the first command and the quantity of times that the user uses the first command.
  • the terminal may consider that the first command of the user is the positive reply and that the user expects the terminal to play the message in the voice manner. Therefore, the terminal learns that the first command of the user is the positive reply.
  • FIG. 8 (6) shows an interface 810 displayed by the terminal.
  • the interface 810 may display a status bar 802, a message prompt box 803, a pattern 806, a time widget, and the like.
  • the interface 810 is an interface in which the terminal has successfully learned the first command.
  • the terminal may change a color, a gray scale, or the like of the pattern 806, to notify the user that the terminal has learned the first command.
  • the terminal may display text information, to notify the user that the terminal has successfully learned the first command, or to add the first command to the keyword of the positive reply.
  • a prompt manner of the terminal is not limited in this embodiment of this application.
  • the terminal may display prompt information indicating that the first command has successfully learned after playing the message, or may display prompt information indicating that the first command has successfully learned before playing the message, or may not display prompt information indicating that the first command has successfully learned. This is not limited in this embodiment of this application.
  • the terminal asks, in the voice manner, whether to play the message.
  • the terminal detects a voice of the user, and the voice of the user is converted into the first command (to be specific, in this case, content of the voice of the user is the same as content of the first voice)
  • the terminal matches the first command with the set first keyword.
  • the first keyword set by the terminal includes the first command (a learning result)
  • matching succeeds.
  • the terminal determines that the user expects to play the message in the voice manner, and therefore the terminal plays the third message in the voice manner.
  • the user sets the positive reply to "yes" and "play".
  • the terminal matches "please say” with the set positive reply. If it is determined that "please say” is not the set positive reply, the terminal does not play the message in the voice manner temporarily, and continues to listen to a reply of the user. If the user replies "play” for the second time, the terminal matches "play” with the preset positive reply. If it is determined that "play” is the set positive reply, the terminal plays the message in the voice manner. In addition, the terminal records "please say” once. Then, after the terminal asks the user whether the message needs to be played in the voice manner, if the user still replies "please say”, and then replies the positive reply set by the terminal.
  • the terminal records "please say” twice. After a quantity of times of answering "please say” recorded by the terminal reaches a preset quantity of times, the terminal learns that "please say” is the positive reply. Then, the terminal may set “please say” to the positive reply. Afterwards, when the terminal receives "please say” again used by the user, and the terminal matches "please say” with the set positive reply, it can be determined that "please say” is the positive reply, and the terminal plays the message in the voice manner.
  • a non-preset reply of a user can be learned, to determine a real intention of the user and whether the user expects to play a message. This increases accuracy of executing a command by the terminal, and a success rate of playing a message in the voice manner by the terminal, so that the terminal is more intelligent, and user experience of using the terminal is improved.
  • FIG. 4A and FIG. 4B are a schematic flowchart of another method for playing a message in a voice manner according to an embodiment of this application.
  • the method specifically includes S101 to S113.
  • S101 to S113 are as follows:
  • time is used for timekeeping.
  • the terminal ends this procedure, and considers by default that the user does not expect to play the message in the voice manner.
  • the terminal may alternatively initialize time in step S101, and start timekeeping from step S101. This is not limited in this embodiment of this application.
  • m is used to mark a recorded quantity of times of learning a text of the voice in a process in which the terminal learns the voice command of the user.
  • S103 The terminal detects a voice of the user, and records a quantity n of times that the voice of the user is detected.
  • an initial value of n in this embodiment is 0.
  • S104 The terminal converts the voice detected for the n th time into a text.
  • step S105 The terminal matches the converted text with a set first keyword. If the converted text does not match the set first keyword, step S106 is performed. If the converted text matches the set first keyword, step S108 is performed.
  • the first keyword set herein is a positive reply. To be specific, if the converted text matches a first keyword of a positive reply, it is considered that the user expects to play the message in the voice manner. If the converted text does not match the first keyword of the positive reply, the voice of the user needs to be further learned to determine whether the user expects to play the message in the voice manner.
  • S106 The terminal does not play the message in the voice manner temporarily.
  • step S107 The terminal determines whether the time (time) reaches a preset time. If the time reaches the preset time, the procedure ends. If the time does not reach the preset time, step S103 continues to be performed.
  • S108 The terminal plays the message in the voice manner.
  • step S109 The terminal determines whether n is greater than or equal to 2. If n is greater than or equal to 2, step S110 is performed. If n is not greater than or equal to 2, this procedure ends.
  • n is not greater than or equal to 2, it indicates that after receiving the message, when sending the voice command for the first time, the user makes a reply by using the preset first keyword of the positive reply.
  • the terminal recognizes the voice command of the user, and plays the message in the voice manner. Therefore, there is no need to learn a last voice command of the user, to be specific, the following learning process does not need to be performed, and the procedure ends.
  • n is greater than or equal to 2, it indicates that after receiving the message, when sending the voice command for the second time or later time, the user makes a reply by using the preset first keyword of the positive reply.
  • the terminal needs to learn the voice command sent for the at least one time, to determine whether the voice command sent for the at least one time is a positive reply, that is, S110 and subsequent steps are performed.
  • step S110 The terminal performs semantic analysis on a text of a voice detected last time (namely, the (n-1) th time), to determine whether the text of the voice detected last time is a positive reply. If the text of the voice detected last time is the positive reply, step S111 is performed. If the text of the voice detected last time is not the positive reply, this procedure ends.
  • whether the text of the voice detected last time is the positive reply, a negative reply, or neither may be determined through the semantic analysis. This can provide more basis for the terminal to determine whether the text of the voice detected last time is the positive reply.
  • the terminal may perform semantic analysis on texts of voices detected before this time (from the first time to the (n-1) th time). This is because the voices detected from the first time to the (n-1) th time may be different expressions that the user expects to play the message in the voice manner. In this way, a learning capability and learning efficiency of the terminal are improved.
  • the terminal may directly perform Sill without performing step S110. This is not specifically limited in this embodiment of this application.
  • the terminal may separately record the texts of all the voices detected before this time (from the first time to the (n-1) th time).
  • a text of a voice detected each time corresponds to one m value, and m corresponding to the text of the voice detected each time is increased by 1.
  • step S112 The terminal determines whether m is a predetermined quantity of times. If m is the predetermined quantity of times, step S113 is performed. If m is not the predetermined quantity of times, this procedure ends.
  • the terminal records the texts of the voices detected from the first time to the (n-1) th time.
  • the terminal may determine, for an m value corresponding to a text of a voice detected each time, whether a quantity of times of using the m value corresponding to the text of the voice detected each time reaches the predetermined quantity of times.
  • S113 The terminal adds, to the first keyword, the text of the voice detected last time (namely, the (n-1) th time).
  • the terminal records the texts of the voices detected from the first time to the (n-1) th time.
  • the terminal may set, as the first keyword, a text of a voice detected for one or more times when a corresponding value of m meets the predetermined quantity of times.
  • FIG. 5A to FIG. 5C are a schematic flowchart of another method for playing a message in a voice manner according to an embodiment of this application.
  • the procedure specifically includes steps S101 to S104 and steps S201 to S215.
  • Steps S201 to S215 are specifically as follows:
  • step S201 A terminal determines whether a text of a voice detected for the n th time matches a set key. If it is determined that the text of the voice detected for the n th time does not match a first keyword of a positive reply or a first keyword of a negative reply, step S202 is performed. If it is determined that the text of the voice detected for the n th time matches the first keyword of the positive reply, step S204 is performed. If it is determined that the text of the voice detected for the n th time matches the first keyword of the negative reply, step S210 is performed.
  • S202 The terminal does not play a message in a voice manner temporarily.
  • step S203 The terminal determines whether a time (time) reaches a preset time. If the time reaches the preset time, the procedure ends. If the time does not reach the preset time, step S103 continues to be performed.
  • S204 The terminal plays the message in the voice manner.
  • step S205 The terminal determines whether n is greater than or equal to 2. If n is greater than or equal to 2, step S206 is performed. If n is not greater than or equal to 2, this procedure ends.
  • step S206 The terminal performs semantic analysis on a text of a voice detected for the (n-1) th time, to determine whether the text of the voice detected for the (n-1) th time is a positive reply. If the text of the voice detected for the (n-1) th time is the positive reply, step S207 is performed. If the text of the voice detected for the (n-1) th time is not the positive reply, this procedure ends.
  • step S208 The terminal determines whether m is a predetermined quantity of times. If m is the predetermined quantity of times, step S209 is performed. If m is not the predetermined quantity of times, this procedure ends.
  • S209 The terminal adds the text of the voice detected for the (n-1) th time to the first keyword of the positive reply.
  • steps S204 to S209 refer to steps S108 to S113. Details are not described.
  • S210 The terminal does not play the message in the voice manner.
  • step S211 The terminal determines whether n is greater than or equal to 2. If n is greater than or equal to 2, step S212 is performed. If n is not greater than or equal to 2, this procedure ends.
  • step S212 The terminal performs semantic analysis on a text of a voice detected for the (n-1) th time, to determine whether the text of the voice detected for the (n-1) th time is a positive reply. If the text of the voice detected for the (n-1) th time is the positive reply, step S213 is performed. If the text of the voice detected for the (n-1) th time is not the positive reply, this procedure ends.
  • whether the text of the voice detected for the (n-1) th time is the positive reply, a negative reply, or neither may be determined through the voice analysis. This can provide more basis for the terminal to determine whether the text of the voice detected for the (n-1) th time is the negative reply.
  • the terminal may perform semantic analysis on texts of voices detected before this time (from the first time to the (n-1) th time). This is because the voices detected from the first time to the (n-1) th time may be different expressions that a user expects not to play the message in the voice manner. In this way, a learning capability and learning efficiency of the terminal are improved.
  • the terminal may directly perform S213 without performing this step. This is not specifically limited in this embodiment of this application.
  • the terminal may separately record texts of voices detected before this time (from the first time to the (n-1) th time).
  • a text of a voice detected each time corresponds to one m value, and m corresponding to the text of the voice detected each time is increased by 1.
  • step S214 The terminal determines whether m is the predetermined quantity of times. If m is the predetermined quantity of times, step S215 is performed. If m is not the predetermined quantity of times, this procedure ends.
  • the terminal separately records texts of voices detected from the first time to the (n-1) th time.
  • the terminal may determine, for an m value corresponding to a text of a voice detected each time, whether a quantity of times of using the m value corresponding to the text of the voice detected each time reaches the predetermined quantity of times.
  • S215 The terminal adds the text of the voice detected for the (n-1) th time to the first keyword of the negative reply.
  • the terminal separately records texts of voices detected from the first time to the (n-1) th time.
  • the terminal may set, as the first keyword, a text of a voice detected for one or more times when a corresponding value of m meets the predetermined quantity of times. This procedure ends.
  • the terminal may learn the voice command of the user, to recognize an intention of the user and perform a corresponding operation. In this way, interaction between the user and the terminal is more personalized and intelligent. This helps improve use efficiency of the terminal and improve user experience.
  • a terminal automatically plays a newly received message in a voice manner.
  • FIG. 6 is a schematic flowchart of a method for playing a message in a voice manner according to an embodiment of this application.
  • the method includes steps S501a to S505, and steps S501a to S505 are specifically as follows:
  • the user may set the message automatically played by the terminal.
  • the terminal may set an application (namely, a preset application) for automatically playing the message.
  • the terminal may automatically play the message.
  • the terminal may set a contact (namely, a preset contact) corresponding to the automatically played message, or a group (namely, a preset contact group) of a contact corresponding to the automatically played message.
  • the terminal may automatically play the message in the voice manner.
  • the terminal may set a second keyword (a preset second keyword) included in content of the automatically played message.
  • the terminal when the message received by the terminal includes the preset second keyword, the terminal automatically plays the message.
  • the terminal may alternatively set a type (for example, a chat message, a Moments message, and a system message in WeChat) of the automatically played message, a time period of the automatically played message, a location of the automatically played message, or the like. Details are not described in this embodiment of this application.
  • the terminal may alternatively set a playing priority of the message.
  • the user may determine the playing priority of the message based on a usage frequency of each application, importance of each contact or contact group, or specific setting content of the second keyword. For example, if the user uses WeChat more frequently, a priority of WeChat may be set to be higher than a priority of Messages. For another example, a priority of a starred contact in WeChat is set to be higher than a priority of a common contact in WeChat. For another example, if the second keyword is set to "urgent", a message priority including the second keyword may be set to the highest. This is not specifically limited in this embodiment of this application.
  • FIG. 7(1) to FIG. 7(4) are a schematic diagram of some interfaces of a terminal.
  • An interface 701 shown in FIG. 7(1) may be used to set an application that can automatically play a message.
  • the interface 701 may include a status bar 712, a plurality of controls 702, a plurality of controls 703, and a control 704.
  • the status bar 712 may include a name of an operator (for example, China Mobile), a time, a WiFi icon, signal strength, current remaining power, and the like.
  • the control 702 may be used to delete the application for automatically playing a message
  • the control 703 may be used to add the application for automatically playing a message
  • the control 704 may be used to further set a selected application.
  • the terminal when the terminal receives an operation performed by the user on tapping the control 704 corresponding to "WeChat", for example, a tapping operation, the terminal displays an interface 705 shown in FIG. 7 (2).
  • the interface 705 may be used to set a contact in "WeChat".
  • Contact settings include a group setting and a specific contact setting.
  • the group setting is used as an example for description.
  • a control 706 may be used to enable a function of automatically playing a message in a group in contacts.
  • the terminal may set the group to automatically play a message.
  • a control 707 is used to further set of the selected group.
  • the terminal when the terminal receives an operation performed by the user on the control 707 corresponding to the group, for example, tapping, the terminal displays an interface 708 shown in FIG. 7 (3).
  • the user may select the group to automatically play a message, and specifically set the selected group.
  • the terminal receives an operation performed by the user on a control 709 corresponding to "family", for example, tapping, the terminal displays an interface 710 shown in FIG. 7 (4).
  • a function of whether to enable a keyword may be enabled. After the function of enabling a keyword is enabled, if content of a message includes the keyword, it may be considered that the message needs to be automatically played.
  • a control 711 may be used to receive a keyword input by the user.
  • the function of enabling a keyword may not be associated with an application or a contact.
  • the terminal may alternatively be set to automatically play content of a message if the content of the message includes some keywords. This is irrelevant to an application to which the message belongs and whether the message is sent by a contact. This is not specifically limited in this embodiment of this application.
  • S502 The terminal receives a fourth message.
  • step S503 The terminal determines whether the fourth message belongs to the preset application. If the fourth message belongs to the preset application, step S505 is performed. If the fourth message does not belong to the preset application, step S504 is performed.
  • S504 The terminal plays a prompt tone, to notify the user that the fourth message is received.
  • step S505 The terminal determines whether the fourth message is sent by the preset contact. If the fourth message is sent by the preset contact, step S506 is performed. If the fourth message is not sent by the preset contact, step S504 is performed.
  • step S506 The terminal determines whether content of the fourth message includes the second keyword. If the content of the fourth message includes the second keyword, step S507 is performed. If the content of the fourth message does not include the second keyword, step S504 is performed.
  • steps S503 and S504 and specific determining content in the steps are not limited in this embodiment of this application.
  • a specific determining process and specific determining content of the terminal are related to a specific setting of the user in step S501. This is not specifically limited in this embodiment of this application.
  • S507 The terminal plays the fourth message in the voice manner.
  • the terminal may alternatively ask, in the voice manner, the user whether to play the fourth message.
  • the terminal may alternatively ask, in the voice manner, the user whether to play the fourth message.
  • For an inquiry process refer to related content in the application scenario 1. Details are not described herein.
  • the terminal may further receive a setting of the user about whether to enable a voice playing function. If the user enables the voice playing function, the terminal has permission to play a message in the voice manner, and may play the message in the voice manner when a condition is met. If the user does not enable the voice playing function, the terminal does not have the permission to play a message in the voice manner, and cannot play the message in the voice manner.
  • the terminal may further receive a voice command of the user, and answer the message.
  • the voice command of the user may use a third keyword as a prefix of a voice, to recognize that the voice of the user is an answer of the user to the message.
  • the third keyword may be "please answer", "please reply”, or the like.
  • the terminal After playing the first message or the fourth message, the terminal receives a fourth voice of the user.
  • the terminal converts the fourth voice into text information.
  • the terminal answers, to a contact that sends the message, voice information following the third keyword in the fourth voice.
  • the terminal may alternatively convert the voice information following the third keyword in the fourth voice into text information, and answer the text information to the contact that sends the message. This is not specifically limited in this embodiment of this application.
  • the terminal may convert the fourth voice into the text information.
  • the terminal may convert a part of the received fourth voice into text information.
  • the terminal may learn the third keyword.
  • a learning method is similar to a method for learning the first keyword by the terminal in the application scenario 1. Details are not described herein.
  • the terminal may answer to a sender with some pictures (for example, a smiling face picture or an angry face picture in an input method) based on a voice of the user.
  • the user may preset a correspondence between a fourth keyword and a picture in the input method. For example, “smile” corresponds to the smiling face picture, and "anger” corresponds to the angry face picture.
  • the terminal when the terminal has determined that the fourth voice is an answer of the user to the first message or the fourth message, and detects that the fourth voice of the user includes the fourth keyword, the terminal answers a corresponding picture to the sender according to the correspondence between the fourth keyword and the picture. In this way, diversity of answered messages of the user is enriched, and user experience is improved.
  • the terminal when receiving the message, may alternatively perform natural semantic analysis on content of the message.
  • the terminal may invoke a related application module or functional module based on a result of the natural semantic analysis, and provide more related information for the user in a voice playing manner.
  • the content of the message is "I am going to your place on business tomorrow. How is the weather over there?"
  • the terminal performs semantic analysis on inquiry about the weather, and may invoke a weather-related application in the terminal, to obtain weather information from the weather-related application, and play the weather information to the user.
  • the content of the message is "Where to eat today?"
  • the terminal performs natural semantic analysis on inquiry about a restaurant.
  • the terminal may invoke a map to ask a restaurant near the terminal, or the terminal may invoke, for example, a Dazhong Dianping application, to ask information, for example, a restaurant that the user frequently goes to, and play the information to the user.
  • a Dazhong Dianping application to ask information, for example, a restaurant that the user frequently goes to, and play the information to the user.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Artificial Intelligence (AREA)
  • Telephone Function (AREA)
EP18925355.2A 2018-07-04 2018-07-04 Procédé et terminal de lecture de message Pending EP3809671A4 (fr)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2018/094517 WO2020006711A1 (fr) 2018-07-04 2018-07-04 Procédé et terminal de lecture de message

Publications (2)

Publication Number Publication Date
EP3809671A1 true EP3809671A1 (fr) 2021-04-21
EP3809671A4 EP3809671A4 (fr) 2021-06-02

Family

ID=69060593

Family Applications (1)

Application Number Title Priority Date Filing Date
EP18925355.2A Pending EP3809671A4 (fr) 2018-07-04 2018-07-04 Procédé et terminal de lecture de message

Country Status (4)

Country Link
US (1) US11837217B2 (fr)
EP (1) EP3809671A4 (fr)
CN (1) CN112154640B (fr)
WO (1) WO2020006711A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115204127A (zh) * 2022-09-19 2022-10-18 深圳市北科瑞声科技股份有限公司 基于远程流调的表单填写方法、装置、设备及介质

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113206779A (zh) * 2021-03-31 2021-08-03 广州朗国电子科技有限公司 消息播放的优化方法、终端以及存储装置
CN113364669B (zh) * 2021-06-02 2023-04-18 中国工商银行股份有限公司 消息处理方法、装置、电子设备及介质
CN115695636A (zh) * 2021-07-27 2023-02-03 华为技术有限公司 一种智能语音交互的方法及电子设备
CN114124860A (zh) * 2021-11-26 2022-03-01 中国联合网络通信集团有限公司 会话管理方法、装置、设备及存储介质
CN114822506A (zh) * 2022-04-15 2022-07-29 广州易而达科技股份有限公司 一种消息播报方法、装置、移动终端及存储介质
US11777882B1 (en) * 2022-07-06 2023-10-03 ph7, Ltd. Notification sound processing for push notification
CN115499397B (zh) * 2022-09-08 2023-11-17 亿咖通(湖北)技术有限公司 一种信息回复方法、装置、设备及存储介质

Family Cites Families (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101378530A (zh) 2007-08-30 2009-03-04 乐金电子(中国)研究开发中心有限公司 一种短信收发方法、系统及短信服务器
CN101800800A (zh) 2009-02-06 2010-08-11 沈阳晨讯希姆通科技有限公司 实现语音化短信接收的移动终端及其工作方法
CN101778154A (zh) 2009-12-28 2010-07-14 中兴通讯股份有限公司 一种短信语音播报屏蔽的方法和装置
US8798995B1 (en) * 2011-09-23 2014-08-05 Amazon Technologies, Inc. Key word determinations from voice data
US9026176B2 (en) * 2013-05-12 2015-05-05 Shyh-Jye Wang Message-triggered voice command interface in portable electronic devices
US9218804B2 (en) * 2013-09-12 2015-12-22 At&T Intellectual Property I, L.P. System and method for distributed voice models across cloud and device for embedded text-to-speech
US10296160B2 (en) * 2013-12-06 2019-05-21 Apple Inc. Method for extracting salient dialog usage from live data
CN103929537B (zh) 2014-04-03 2017-02-15 北京深思数盾科技股份有限公司 基于不同级别信息的实时提醒方法
CN104159206A (zh) 2014-08-19 2014-11-19 广州市久邦数码科技有限公司 一种可穿戴式设备的短信系统及其短信处理方法
CN106156022B (zh) * 2015-03-23 2020-02-21 联想(北京)有限公司 一种信息处理方法及电子设备
CN104991894A (zh) 2015-05-14 2015-10-21 深圳市万普拉斯科技有限公司 即时聊天信息浏览方法和系统
CN105245729B (zh) * 2015-11-02 2019-02-26 北京奇虎科技有限公司 移动终端消息阅读方法和装置
CN106656732A (zh) 2015-11-04 2017-05-10 陈包容 一种基于场景信息获取聊天回复内容的方法及装置
CN106899946A (zh) * 2015-12-17 2017-06-27 北京奇虎科技有限公司 消息的语音播放处理方法、装置及系统
CN105825856B (zh) * 2016-05-16 2019-11-08 四川长虹电器股份有限公司 车载语音识别模块的自主学习方法
CN107644640A (zh) * 2016-07-22 2018-01-30 佛山市顺德区美的电热电器制造有限公司 一种信息处理方法及家电设备
EP3506256A4 (fr) * 2016-08-26 2019-08-21 Sony Corporation Dispositif de traitement d'informations et procédé de traitement d'informations
CN106412282B (zh) 2016-09-26 2019-08-20 维沃移动通信有限公司 一种实时消息语音提示方法及移动终端
CN106506804B (zh) * 2016-09-29 2020-02-21 维沃移动通信有限公司 一种通知消息的提醒方法及移动终端
CN107220292A (zh) * 2017-04-25 2017-09-29 上海庆科信息技术有限公司 智能对话装置、反馈式智能语音控制系统及方法
CN107360320A (zh) * 2017-06-30 2017-11-17 维沃移动通信有限公司 一种移动终端控制方法及移动终端
CN107452373A (zh) * 2017-07-26 2017-12-08 上海与德通讯技术有限公司 机器人交互方法及系统
CN107612814A (zh) 2017-09-08 2018-01-19 北京百度网讯科技有限公司 用于生成候选回复信息的方法和装置

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115204127A (zh) * 2022-09-19 2022-10-18 深圳市北科瑞声科技股份有限公司 基于远程流调的表单填写方法、装置、设备及介质

Also Published As

Publication number Publication date
CN112154640A (zh) 2020-12-29
US20210210068A1 (en) 2021-07-08
EP3809671A4 (fr) 2021-06-02
WO2020006711A1 (fr) 2020-01-09
CN112154640B (zh) 2024-04-30
US11837217B2 (en) 2023-12-05

Similar Documents

Publication Publication Date Title
US11837217B2 (en) Message playing method and terminal
EP3952263A1 (fr) Procédé de prévisualisation des messages de notification et dispositif électronique
CN113542485B (zh) 一种通知处理方法、电子设备及计算机可读存储介质
WO2020207326A1 (fr) Procédé d'envoi de message de dialogue et dispositif électronique
CN110825469A (zh) 语音助手显示方法及装置
EP3876506B1 (fr) Procédé de présentation d'une vidéo sur un dispositif électronique lors de l'arrivée d'un appel entrant et dispositif électronique
CN110910872A (zh) 语音交互方法及装置
EP3923617A1 (fr) Procédé de réduction de la consommation d'énergie d'un terminal mobile et terminal mobile
CN111602108B (zh) 一种应用图标的显示方法及终端
EP4187891A1 (fr) Procédé et système de réponse à un message, et dispositif électronique
CN113170279B (zh) 基于低功耗蓝牙的通信方法及相关装置
EP4213489A1 (fr) Procédé de recommandation de dispositif et dispositif électronique
CN113170019A (zh) 一种基于应用的来电显示方法和终端设备
CN114116085A (zh) 一种通知消息管理方法及电子设备
WO2021042881A1 (fr) Procédé de notification par message et dispositif électronique
CN114971107A (zh) 一种隐私风险反馈方法、装置及第一终端设备
CN114650330A (zh) 一种添加操作序列的方法、电子设备和系统
EP4152736A1 (fr) Procédé de commutation de canal de sortie audio, appareil et dispositif électronique
CN113467821A (zh) 应用程序的修复方法、装置、设备及可读存储介质
CN114664306A (zh) 一种编辑文本的方法、电子设备和系统
CN114449103B (zh) 提醒方法、图形用户界面及终端
CN114079642B (zh) 一种邮件处理方法及电子设备
CN117118970A (zh) 文件的下载方法及装置
CN117041465A (zh) 一种视频通话的优化方法、电子设备及存储介质
CN114900583A (zh) 控制视频彩振播放的方法和装置

Legal Events

Date Code Title Description
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE

PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

17P Request for examination filed

Effective date: 20210112

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

A4 Supplementary search report drawn up and despatched

Effective date: 20210504

RIC1 Information provided on ipc code assigned before grant

Ipc: H04L 29/08 20060101AFI20210428BHEP

Ipc: G10L 15/06 20130101ALI20210428BHEP

Ipc: G10L 15/22 20060101ALI20210428BHEP

Ipc: G10L 15/07 20130101ALI20210428BHEP

Ipc: G10L 15/08 20060101ALN20210428BHEP

DAV Request for validation of the european patent (deleted)
DAX Request for extension of the european patent (deleted)
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

17Q First examination report despatched

Effective date: 20230502