WO2024114233A1 - Procédé d'appel et dispositif électronique - Google Patents
Procédé d'appel et dispositif électronique Download PDFInfo
- Publication number
- WO2024114233A1 WO2024114233A1 PCT/CN2023/127971 CN2023127971W WO2024114233A1 WO 2024114233 A1 WO2024114233 A1 WO 2024114233A1 CN 2023127971 W CN2023127971 W CN 2023127971W WO 2024114233 A1 WO2024114233 A1 WO 2024114233A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- user
- electronic device
- mobile phone
- voice
- call
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 103
- 230000015654 memory Effects 0.000 claims description 47
- 230000004044 response Effects 0.000 claims description 47
- 230000006870 function Effects 0.000 claims description 28
- 230000000694 effects Effects 0.000 claims description 14
- 238000004590 computer program Methods 0.000 claims description 11
- 238000013473 artificial intelligence Methods 0.000 description 120
- 238000004891 communication Methods 0.000 description 41
- 238000010586 diagram Methods 0.000 description 31
- 238000010295 mobile communication Methods 0.000 description 17
- 238000005516 engineering process Methods 0.000 description 11
- 230000008569 process Effects 0.000 description 11
- 230000005236 sound signal Effects 0.000 description 10
- 238000007726 management method Methods 0.000 description 8
- 230000001360 synchronised effect Effects 0.000 description 5
- 241000394635 Acetomicrobium mobile Species 0.000 description 4
- 101100233916 Saccharomyces cerevisiae (strain ATCC 204508 / S288c) KAR5 gene Proteins 0.000 description 4
- 230000007774 longterm Effects 0.000 description 4
- 229920001621 AMOLED Polymers 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000013500 data storage Methods 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 230000005855 radiation Effects 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 101001121408 Homo sapiens L-amino-acid oxidase Proteins 0.000 description 1
- 102100026388 L-amino-acid oxidase Human genes 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000003416 augmentation Effects 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000010267 cellular communication Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000037361 pathway Effects 0.000 description 1
- 239000002096 quantum dot Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72484—User interfaces specially adapted for cordless or mobile telephones wherein functions are triggered by incoming communication events
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
- G10L13/08—Text analysis or generation of parameters for speech synthesis out of text, e.g. grapheme to phoneme translation, prosody generation or stress or intonation determination
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/64—Automatic arrangements for answering calls; Automatic arrangements for recording messages for absent subscribers; Arrangements for recording conversations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72403—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
- H04M1/7243—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
- H04M1/72433—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for voice messaging, e.g. dictaphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72403—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
- H04M1/7243—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
- H04M1/72436—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for text messaging, e.g. short messaging services [SMS] or e-mails
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72469—User interfaces specially adapted for cordless or mobile telephones for operating the device by selecting functions from two or more displayed items, e.g. menus or icons
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/725—Cordless telephones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/487—Arrangements for providing information services, e.g. recorded voice services or time announcements
- H04M3/493—Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals
Definitions
- the present application relates to the field of terminal technology, and in particular to a communication method and electronic equipment.
- Audio calls have become a common way of making calls in people's daily life and work.
- users are unable or inconvenient to answer the phone (for example, when they are in a meeting, working, or driving), but do not want to miss the call, they can use artificial intelligence (AI) technology to make smart calls.
- AI artificial intelligence
- the current AI smart call technology does not provide a good user experience.
- the present application provides a calling method and an electronic device, which can improve the user's calling experience and call efficiency.
- the present application provides a calling method, which is applied to a first electronic device, and includes: displaying a calling interface, the calling interface includes an identification of a second user, the second user is a user using a second electronic device, and the calling is a calling between the second electronic device and the first electronic device; sending a first voice message of a first timbre to the second electronic device, the first voice message being determined by the first electronic device; receiving first information input by the first user through the calling interface; and sending a second voice message of a second timbre to the second electronic device based on the first information; the first timbre is different from the second timbre.
- the first electronic device can use voices with different tones to play the voice corresponding to the automatic reply content and the voice corresponding to the manual reply content to the other party of the call, so that the other party of the call can distinguish the manual reply information from the automatic reply information, thereby improving the user's call experience and call efficiency.
- the identifier of the second user in the call interface can be contact A in the call interface 701 shown in FIG7A.
- the call is between the second electronic device (mobile phone 2) and the first electronic device (mobile phone 1); mobile phone 1 sends a first voice message of a first timbre (timbre 1) to mobile phone 2 for automatic reply by intelligent AI.
- a first voice message of a first timbre (timbre 1)
- a second voice message of a second timbre (timbre 2) is sent to the second electronic device; the first timbre is different from the second timbre.
- the method before sending the first voice message to the second electronic device, the method further includes: displaying first prompt information, where the first prompt information is used to prompt that the first electronic device is in an automatic reply mode.
- the words “auto answer mode” are displayed in the interface 701 to prompt the user that the mobile phone 1 is in the auto reply mode.
- the user can intuitively understand that the call mode of the mobile phone 1 at this time is the auto reply mode, thereby improving the user's call experience.
- the method before sending the second voice message to the second electronic device, the method further includes: displaying second prompt information, where the second prompt information is used to prompt that the first electronic device is in a manual reply mode.
- the words “manual answering mode” are displayed in the interface 701 to prompt the user that the mobile phone 1 is in the manual reply mode.
- displaying the second prompt information specifically includes: in response to the first operation, displaying the second prompt information, the second prompt information is used to prompt that the first electronic device is in a manual reply mode; the first operation includes an input operation of the first user acting on the call interface.
- the first operation includes an operation of the first user acting on the text input box of the call interface.
- the first operation may be an operation of the mobile phone cursor hitting the text input box 707, the user clicking the text input box 707, calling out the mobile phone keyboard, the user typing the first word, and other manual reply operations of the user, which are not limited in the embodiments of the present application.
- the intelligent AI automatic reply mode and the user manual reply mode in the first electronic device can be switched adaptively without the user having to click on the control to switch the mode, thereby improving call efficiency.
- the method after receiving the first information input by the first user through the call interface, the method further includes: Displaying first information on the call interface with a first user interface UI effect; after sending a first voice message with a first tone to a second electronic device, the method also includes: displaying second information on the call interface with a second user interface UI effect, the second information being text information corresponding to the first voice message; wherein the first user interface UI effect is different from the second user interface UI effect.
- mobile phone 1 can also display the text content of automatic replies and the text content of text replies manually entered by the user through different interfaces (user interface, UI), and more clearly distinguish the text content of automatic replies and the text content of text replies manually entered by the user, which is convenient for users to view and understand the call content, improves the user's call experience, and improves call efficiency.
- UI user interface
- the content of the automatic reply by the intelligent AI of mobile phone 1 is displayed in a text box 7061 with a black background in Song font format
- the content of the manual reply by the user is displayed in a text box 7064 with a black background in Lishu font format.
- sending a second voice message of a second timbre to a second electronic device based on the first information includes: if it is detected that the first user operates a send control of the call interface within a first time period, sending a second voice message of a second timbre to the second electronic device based on the first information; and the first time period is less than a first threshold.
- the user completes the input within the first time period and sends it to the mobile phone 2.
- the interface display module 312 displays the text information input by the user in the text box 7064 in the mobile phone interface 701, and the mobile phone 1 converts the text information into voice information of the tone 2 through the personalized text-to-speech module 313 and sends it to the mobile phone 2.
- the method further includes: if the first user does not detect an operation of the send control on the call interface within the first time period, sending a first voice prompt message to the second electronic device, the first voice prompt message being used to prompt the first user that the first user is inputting the first message.
- the first electronic device can remind the other party of the call that the user is manually inputting the reply message, thereby preventing the other party from hanging up the call due to not receiving a reply for a long time.
- the mobile phone 1 sends a first voice prompt message.
- the first voice prompt message can be displayed in the interface shown in FIG8.
- the interface display module 312 can display the first voice prompt message in the text box 7063 in FIG8 (a).
- the first voice prompt message is converted into a voice message of tone 1 by the universal text-to-speech module 314 and sent to the mobile phone 2.
- the method further includes: if the first user does not detect an operation of the send control on the call interface within the second duration, a fourth voice message of the first tone is sent to the second electronic device, the second duration is greater than the first duration; the fourth voice message is determined by the first electronic device.
- the manual reply mode can be automatically switched to the automatic reply mode without the user clicking the control to switch the mode, thereby improving the call efficiency.
- the user does not click the send control on the call interface of mobile phone 1, and mobile phone 1 switches from manual reply mode to automatic reply mode, and sends a fourth voice message in the first tone to mobile phone 2.
- the content of the fourth voice message may be as shown in text box 7066 of Figure 9 (b).
- the method further includes at least one of the following: displaying text information corresponding to the second voice information on the call interface; displaying text information corresponding to the first voice prompt information on the call interface; and displaying text information corresponding to the fourth voice information on the call interface.
- the call records can be displayed in the form of text on the mobile phone call interface, making it convenient for users to view the call records.
- the text information in the text box 7061 is the text information corresponding to the first voice prompt information.
- the text information in the text box 7064 is the text information corresponding to the second voice information.
- the text information in the text box 7066 is the text information corresponding to the fourth voice information.
- the method further includes: receiving a third voice message from a second electronic device; the call interface further includes a first control associated with the third voice message; and detecting an operation of the first user on the first control, enabling a voice playback function.
- the user can voice play the message of the other party, thereby improving the user's call experience.
- the mobile phone 1 can play the voice message 3 of the contact A according to the playing requirement of the user of the mobile phone 1.
- the user can click the text box 7062 of the mobile phone 2 reply of the contact A in the dialog box 706.
- the mobile phone 1 can play the voice message 3 replied by the mobile phone 2 corresponding to the content in the text box 7062.
- the method when the first information is voice information, the method further includes: in response to the first user operating a second control associated with the voice information in the call interface, converting the voice information into corresponding text information; Text information is displayed on the interface.
- the mobile phone 1 can input the first information by voice.
- the user can long press the voice control 707’ to perform voice input.
- the user can long press the voice control 707’ shown in (b) of FIG11 , and the voice control 707’ can be converted into the form shown in (c) of FIG11 , indicating that the user is inputting voice.
- the user voice inputs “Why do you think so, I’m in a meeting, do you have anything to do?”.
- the content of the user’s voice input can be displayed in the form of a voice box 7067 in the dialog box 706, and the duration of the voice content in the voice box is 5s.
- the mobile phone 1 in response to the user clicking on the voice box 7067, the mobile phone 1 can play the voice information.
- the user long presses the voice box 7067 to convert the voice information into text information and display it in the text box 7068.
- the method further includes: displaying a first interface, the first interface including a timbre setting option; and setting the first timbre and/or the second timbre in response to the first user's operation on the timbre setting option.
- the mobile phone 1 can pre-set the first tone to "intellectual female voice" and the second tone to "my voice".
- the first tone is different from the second tone, which makes it easier for the other party to distinguish between the content of the automatic reply and the content of the manual reply.
- the first tone and the second tone can also be the same, and this embodiment of the application does not impose specific restrictions on this.
- the present application provides a call method, which is applied to a second electronic device, including: displaying a call interface, the call interface including an identifier of a first user, the first user is a user using the first electronic device, and the call is a call between the second electronic device and the first electronic device; receiving a first voice message of a first timbre from the first electronic device; receiving a second voice message of a second timbre from the first electronic device; wherein the first timbre is different from the second timbre.
- the second electronic device receives voice messages of different timbres, so that the user of the second electronic device can distinguish between manually replied messages and automatically replied messages, thereby improving the user's call experience and call efficiency.
- the method further includes: receiving a first voice prompt message sent from the first electronic device, the first voice prompt message being used to prompt the first user that the first message is being input, the first message being a text message or a voice message corresponding to the second voice message.
- the user of the second electronic device can know that the user of the first electronic device is manually inputting a reply message, thereby preventing the user of the second electronic device from hanging up the phone due to not receiving a reply for a long time.
- the method further includes: sending third voice information to the first electronic device, where the third voice information is voice information input by a second user, and the second user is a user using the second electronic device.
- the present application provides an electronic device comprising: a processor and a memory, the memory being coupled to the processor, the memory being used to store computer program code, the computer program code comprising computer instructions, and when the processor reads the computer instructions from the memory, the electronic device executes a method as described in any one of the first aspect and a method as described in any one of the second aspect.
- the present application provides a computer storage medium, comprising computer instructions, which, when executed on a computer, enable the computer to execute the method described in any one of the first aspect and the method described in any one of the second aspect.
- the present application provides a chip system comprising at least one processor and at least one interface circuit, wherein the at least one interface circuit is used to perform transceiver functions and send instructions to at least one processor, and when the at least one processor executes the instructions, the at least one processor executes the method described in any one of the first aspect and the method described in any one of the second aspect.
- the present application also provides a computer program product, comprising instructions, which, when executed on a computer, enable the computer to execute the method described in any one of the first aspect and the method described in any one of the second aspect.
- an embodiment of the present application provides a circuit system, the circuit system includes a processing circuit, and the processing circuit is configured to execute the method of the first aspect or any one of the implementations of the first aspect and any one of the methods of the second aspect.
- FIG1 is a schematic diagram of a call interface provided in an embodiment of the present application.
- FIG2A is a schematic diagram of the structure of an electronic device provided in an embodiment of the present application.
- FIG2B is a schematic diagram of the structure of another electronic device provided in an embodiment of the present application.
- FIG2C is a schematic diagram of the structure of another electronic device provided in an embodiment of the present application.
- FIG3 is a schematic diagram of a module of an electronic device provided in an embodiment of the present application.
- FIG4A is a schematic diagram of an electronic device interface provided by an embodiment of the present application.
- FIG4B is a schematic diagram of another electronic device interface provided by an embodiment of the present application.
- FIG5 is a schematic diagram of another electronic device interface provided by an embodiment of the present application.
- FIG6 is a schematic diagram of another electronic device interface provided in an embodiment of the present application.
- FIG7A is a schematic diagram of another call interface provided in an embodiment of the present application.
- FIG7B is a schematic diagram of another call interface provided in an embodiment of the present application.
- FIG8 is a schematic diagram of another call interface provided in an embodiment of the present application.
- FIG9 is a schematic diagram of another call interface provided in an embodiment of the present application.
- FIG10 is a schematic diagram of another call interface provided in an embodiment of the present application.
- FIG11 is a schematic diagram of another call interface provided in an embodiment of the present application.
- FIG12A is another schematic diagram of an interface provided in an embodiment of the present application.
- FIG12B is a schematic diagram of another interface provided in an embodiment of the present application.
- FIG13 is another schematic diagram of an interface provided in an embodiment of the present application.
- FIG14 is another schematic diagram of an interface provided in an embodiment of the present application.
- FIG15 is a schematic diagram of another call interface provided in an embodiment of the present application.
- FIG16 is another schematic diagram of an interface provided in an embodiment of the present application.
- FIG17 is a flow chart of a calling method provided in an embodiment of the present application.
- FIG18 is a schematic diagram of another call interface provided in an embodiment of the present application.
- FIG19 is a schematic diagram of another interface call provided in an embodiment of the present application.
- FIG20 is another schematic diagram of a module provided in an embodiment of the present application.
- FIG21 is a schematic diagram of the structure of an electronic device provided in an embodiment of the present application.
- FIG. 22 is a schematic diagram of the structure of a chip system provided in an embodiment of the present application.
- a and/or B can mean: A exists alone, A and B exist at the same time, and B exists alone.
- AI intelligent call technology can realize functions such as recording and screen recording of the call process, automatically generating transcribed text from the other party's voice and displaying it on the electronic device screen, automatically replying to the electronic device and displaying the reply text on the screen, manually inputting text replies and displaying the reply text on the screen, and playing the content of automatic replies and manually input text replies to the other party of the call. And the automatic reply of the electronic device and the manual text reply can be switched at any time.
- a mobile phone intelligent AI call interface 301 which includes an automatic answer button 302, a phone subtitle button, a manual answer button 303, and a dialog box 304.
- the mobile phone call mode is the automatic answer mode, and the words "automatic answer mode" are displayed in the dialog box 304.
- Intelligent AI can automatically reply to the other party. If the user wants to switch to the manual answer mode, as shown in (b) of Figure 1, the user needs to click the manual answer button 303 in the mobile phone interface with his finger. In response to the user clicking the manual answer button 303, the mobile phone call mode changes from the automatic answer mode to the manual answer mode.
- the mode is switched to manual answering mode, and the words "manual answering mode" are displayed in the dialog box 304.
- a text input box 305 and a text sending button 306 are displayed below the dialog box.
- the user enters the text "What's the matter?” in the text input box 305 and clicks the send button 306.
- the mobile phone sends the text "What's the matter?”
- the content of the manual reply and the content of the smart AI automatic reply are displayed in the dialog box with the same UI display effect. The other party cannot find out that the call mode has been switched from automatic answering to manual answering based on the sound of the answering.
- the embodiments of the present application propose a calling method and an electronic device.
- the electronic device uses intelligent AI to make a call, it can use voices with different tones to play the voices corresponding to the automatic reply content and the voices corresponding to the manual reply content to the other party of the call.
- the electronic device can also display the text content of the automatic reply and the text content of the manual reply input by the user through different interfaces (user interface, UI).
- the intelligent AI automatic reply mode and the user manual reply mode in the electronic device can be switched adaptively to improve the user's call experience and call efficiency.
- the embodiment of the present application can be applied to a system composed of multiple electronic devices, and the system can be shown in FIG2A.
- the system includes a first electronic device 100 and a second electronic device 200.
- the first electronic device 100 and the second electronic device 200 can implement the above-mentioned call method.
- the first electronic device 100 and the second electronic device 200 can be a personal computer (PC), a mobile phone, a tablet computer (Pad), a laptop computer, a desktop computer, a laptop computer, a computer with a transceiver function, a virtual reality (VR) terminal device, an augmented reality (AR) terminal device, a wireless terminal in industrial control (industrial control), a wireless terminal in self-driving, a wireless terminal in remote medical, a wireless terminal in smart grid (smart grid), a wireless terminal in transportation safety (transportation safety), a wireless terminal in smart city (smart city), a wireless terminal in smart home (smart home), a wearable device, a vehicle-mounted device and other terminal devices.
- PC personal computer
- a mobile phone a tablet computer (Pad)
- a laptop computer a desktop computer
- a laptop computer a computer with a transceiver function
- VR virtual reality
- AR augmented reality
- wireless terminal in industrial control industrial control
- a wireless terminal in self-driving
- the first electronic device 100 can interact with the second electronic device 200 through an operator network, such as a 4th generation (4G) mobile communication system, such as a long term evolution (LTE) system, a 5th generation (5G) mobile communication system, such as a new radio (NR) system, and future communication systems, such as a 6th generation (6G) mobile communication system, to achieve a call.
- 4G 4th generation
- LTE long term evolution
- 5G 5th generation
- NR new radio
- 6G 6th generation
- the first electronic device 100 can interact with the second electronic device 200 through a non-operator network to implement a call.
- the non-operator network may include but is not limited to a wireless fidelity (Wi-Fi) network.
- Wi-Fi wireless fidelity
- the second electronic device 200 may initiate a call with the first electronic device 100 , and the first electronic device 100 may also initiate a call with the second electronic device 200 .
- FIG2B is a schematic diagram of the hardware structure of an electronic device provided in an embodiment of the present application.
- the electronic device may be the first electronic device and/or the second electronic device described above.
- the electronic device includes at least one processor 201, a communication line 202, a memory 203, and at least one communication interface 204.
- the memory 203 may also be included in the processor 201.
- the processor 201 may be a central processing unit (CPU), or other general-purpose processors, digital signal processors (DSP), application specific integrated circuits (ASIC), field programmable gate arrays (FPGA), or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
- a general-purpose processor may be a microprocessor or any conventional processor, etc.
- the communication link 202 may include a pathway to transmit information between the above-mentioned components.
- the communication interface can be a module, a circuit, a bus, an interface, a transceiver or other device capable of realizing a communication function, used to communicate with other devices.
- the transceiver can be an independently arranged transmitter, which can be used to send information to other devices, and the transceiver can also be an independently arranged receiver, which is used to receive information from other devices.
- the transceiver can also be a component that integrates the functions of sending and receiving information, and the embodiment of the present application does not limit the specific implementation of the transceiver.
- the memory 203 may be a volatile memory or a nonvolatile memory, or may include both volatile and nonvolatile memories.
- the nonvolatile memory may be a read-only memory (ROM), a programmable ROM (PROM), an erasable programmable ROM (EPROM), an electrically erasable programmable ROM (EEPROM), or a flash memory.
- the volatile memory may be a random access memory.
- Memory random access memory, RAM, which is used as an external high-speed cache.
- RAM random access memory
- static RAM static random access memory
- dynamic RAM dynamic random access memory
- DRAM dynamic random access memory
- SDRAM synchronous dynamic random access memory
- double data rate synchronous dynamic random access memory double data rate SDRAM, DDR SDRAM
- enhanced synchronous dynamic random access memory enhanced SDRAM, ESDRAM
- serial link DRAM SLDRAM
- direct memory bus random access memory direct rambus RAM, DR RAM
- the memory can be independent and connected to the processor 201 through the communication line 202.
- the memory 203 can also be integrated with the processor 201.
- the memory 203 is used to store computer-executable instructions for implementing the solution of the present application, and the execution is controlled by the processor 201.
- the processor 201 is used to execute the computer-executable instructions stored in the memory 203, thereby implementing the carrier transmission method provided in the following embodiments of the present application.
- the computer-executable instructions in the embodiments of the present application may also be referred to as application code, instructions, computer program or other names, which are not specifically limited in the embodiments of the present application.
- the processor 201 may include one or more CPUs, such as CPU0 and CPU1 in FIG. 2B .
- the electronic device may include multiple processors, such as processor 201 and processor 205 in FIG. 2B .
- processors may be a single-core (single-CPU) processor or a multi-core (multi-CPU) processor.
- the processor here may refer to one or more devices, circuits, and/or processing cores for processing data (e.g., computer program instructions).
- the above-mentioned electronic device may be a general device or a special device, and the embodiment of the present application does not limit the type of the electronic device.
- the structure illustrated in the embodiment of the present application does not constitute a specific limitation on the first electronic device.
- the first electronic device may include more or fewer components than shown in the figure, or combine certain components, or split certain components, or arrange the components differently.
- the components shown in the figure may be implemented in hardware, software, or a combination of software and hardware.
- the first electronic device and the second electronic device are described as mobile phones, but this does not constitute a limitation on the form, function, etc. of the first electronic device.
- FIG. 2C a schematic diagram of the structure of a mobile phone provided in the embodiments of the present application is shown. The methods in the following embodiments can be implemented in a mobile phone having the above hardware structure.
- the mobile phone may include a processor 110, an external memory interface 120, an internal memory 121, a universal serial bus (USB) interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2, a wireless communication module 160, an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, a camera 193, and a display screen 194.
- the mobile phone may also include a mobile communication module 150, etc.
- the structure illustrated in this embodiment does not constitute a specific limitation on the mobile phone.
- the mobile phone may include more or fewer components than shown in the figure, or combine some components, or split some components, or arrange the components differently.
- the components shown in the figure may be implemented in hardware, software, or a combination of software and hardware.
- the processor 110 may include one or more processing units, for example, the processor 110 may include an application processor (AP), a modem processor, a graphics processor (GPU), an image signal processor (ISP), a controller, a memory, a video codec, a digital signal processor (DSP), a baseband processor, and/or a neural-network processing unit (NPU), etc.
- AP application processor
- GPU graphics processor
- ISP image signal processor
- controller a memory
- video codec a digital signal processor
- DSP digital signal processor
- NPU neural-network processing unit
- Different processing units may be independent devices or integrated in one or more processors.
- the controller can be the nerve center and command center of the mobile phone.
- the controller can generate operation control signals according to the instruction operation code and timing signal to complete the control of fetching and executing instructions.
- the processor 110 may also be provided with a memory for storing instructions and data.
- the memory in the processor 110 is a cache memory.
- the memory may store instructions or data that the processor 110 has just used or cyclically used. If the processor 110 needs to use the instruction or data again, it may be directly called from the memory. This avoids repeated access, reduces the waiting time of the processor 110, and thus improves the efficiency of the system.
- processor 110 may include one or more interfaces.
- the interface may include an integrated circuit (inter-integrated circuit, I2C) interface, integrated circuit built-in audio (inter-integrated circuit sound, I2S) interface, pulse code modulation (pulse code modulation, PCM) interface, universal asynchronous receiver/transmitter (universal asynchronous receiver/transmitter, UART) interface, mobile industry processor interface (mobile industry processor interface, MIPI), general-purpose input/output (general-purpose input/output, GPIO) interface, USB interface, etc.
- I2C integrated circuit
- I2S integrated circuit built-in audio
- PCM pulse code modulation
- PCM pulse code modulation
- UART universal asynchronous receiver/transmitter
- MIPI mobile industry processor interface
- GPIO general-purpose input/output
- USB interface etc.
- the charging management module 140 is used to receive charging input from the charger. While the charging management module 140 is charging the battery 142, it can also power the mobile phone through the power management module 141.
- the power management module 141 is used to connect the battery 142, the charging management module 140 and the processor 110.
- the power management module 141 can also receive input from the battery 142 to power the mobile phone.
- the wireless communication function of the mobile phone can be realized through antenna 1, antenna 2, mobile communication module 150, wireless communication module 160, modem processor and baseband processor.
- Antenna 1 and antenna 2 are used to transmit and receive electromagnetic wave signals.
- Each antenna in the mobile phone can be used to cover a single or multiple communication frequency bands. Different antennas can also be reused to improve the utilization of the antennas.
- antenna 1 can be reused as a diversity antenna for a wireless local area network.
- the antenna can be used in combination with a tuning switch.
- the mobile communication module 150 can provide solutions for wireless communications including 2G/3G/4G/5G applied on the mobile phone.
- the mobile communication module 150 may include at least one filter, a switch, a power amplifier, a low noise amplifier (LNA), etc.
- the mobile communication module 150 may receive electromagnetic waves from the antenna 1, filter, amplify, and process the received electromagnetic waves, and transmit them to the modulation and demodulation processor for demodulation.
- the mobile communication module 150 may also amplify the signal modulated by the modulation and demodulation processor, and convert it into electromagnetic waves for radiation through the antenna 1.
- at least some of the functional modules of the mobile communication module 150 may be arranged in the processor 110.
- At least some of the functional modules of the mobile communication module 150 may be arranged in the same device as at least some of the modules of the processor 110.
- the mobile phone 1 and the mobile phone 2 may communicate through the mobile communication module 150.
- the modem processor may include a modulator and a demodulator.
- the modulator is used to modulate the low-frequency baseband signal to be sent into a medium-high frequency signal.
- the demodulator is used to demodulate the received electromagnetic wave signal into a low-frequency baseband signal.
- the demodulator then transmits the demodulated low-frequency baseband signal to the baseband processor for processing.
- the application processor outputs a sound signal through an audio device (not limited to a speaker 170A, a receiver 170B, etc.), or displays an image or video through a display screen 194.
- the modem processor may be an independent device.
- the modem processor may be independent of the processor 110 and be set in the same device as the mobile communication module 150 or other functional modules.
- the wireless communication module 160 can provide wireless communication solutions including wireless local area networks (WLAN) (such as Wi-Fi network), Bluetooth (BT), global navigation satellite system (GNSS), frequency modulation (FM), NFC, infrared technology (IR), etc. applied on the mobile phone.
- WLAN wireless local area networks
- BT Bluetooth
- GNSS global navigation satellite system
- FM frequency modulation
- NFC infrared technology
- the wireless communication module 160 can be one or more devices integrating at least one communication processing module.
- the wireless communication module 160 receives electromagnetic waves via the antenna 2, modulates the frequency of the electromagnetic wave signal and performs filtering, and sends the processed signal to the processor 110.
- the wireless communication module 160 can also receive the signal to be sent from the processor 110, modulate the frequency of the signal, amplify it, and convert it into electromagnetic waves for radiation through the antenna 2.
- the first electronic device and the second electronic device can communicate through the wireless communication module 160.
- the antenna 1 of the mobile phone is coupled to the mobile communication module 150, and the antenna 2 is coupled to the wireless communication module 160, so that the mobile phone can communicate with the network and other devices through wireless communication technology.
- the wireless communication technology may include global system for mobile communications (GSM), general packet radio service (GPRS), code division multiple access (CDMA), wideband code division multiple access (WCDMA), time-division code division multiple access (TD-SCDMA), long term evolution (LTE), BT, GNSS, WLAN, NFC, FM, and/or IR technology.
- the GNSS may include a global positioning system (GPS), a global navigation satellite system (GLONASS), a Beidou navigation satellite system (BDS), a quasi-zenith satellite system (QZSS) and/or a satellite based augmentation system (SBAS).
- GPS global positioning system
- GLONASS global navigation satellite system
- BDS Beidou navigation satellite system
- QZSS quasi-zenith satellite system
- SBAS satellite based augmentation system
- the mobile phone realizes the display function through the GPU, the display screen 194, and the application processor.
- the GPU is a microprocessor for image processing, which connects the display screen 194 and the application processor.
- the processor 110 may include one or more GPUs, which execute program instructions to generate or modify Change the displayed information.
- the display screen 194 is used to display images, videos, etc.
- the display screen 194 includes a display panel.
- the display panel can be a liquid crystal display (LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode or an active-matrix organic light-emitting diode (AMOLED), a flexible light-emitting diode (FLED), Miniled, MicroLed, Micro-oLed, quantum dot light-emitting diodes (QLED), etc.
- the mobile phone may include 1 or N display screens 194, where N is a positive integer greater than 1.
- the display screen 194 displays a call interface as shown in FIG. 7A , FIG. 7B , etc.
- the mobile phone can realize the shooting function through the ISP, the camera 193, the video codec, the GPU, the display screen 194 and the application processor, etc.
- the mobile phone can include 1 or N cameras 193, where N is a positive integer greater than 1.
- the external memory interface 120 can be used to connect an external memory card to expand the storage capacity of the mobile phone.
- the external memory card communicates with the processor 110 through the external memory interface 120 to implement a data storage function. For example, files such as music and videos can be stored in the external memory card.
- the internal memory 121 can be used to store computer executable program codes, which include instructions.
- the processor 110 executes various functional applications and data processing of the mobile phone by running the instructions stored in the internal memory 121.
- the internal memory 121 may include a program storage area and a data storage area.
- the program storage area can store an operating system, an application required for at least one function (such as a sound playback function, an image playback function, etc.), etc.
- the data storage area can store data created during the use of the mobile phone (such as audio data, a phone book, etc.), etc.
- the internal memory 121 may include a high-speed random access memory, and may also include a non-volatile memory, such as at least one disk storage device, a flash memory device, a universal flash storage (UFS), etc.
- UFS universal flash storage
- the mobile phone can realize audio functions such as music playing and recording through the audio module 170, the speaker 170A, the receiver 170B, the microphone 170C, the earphone interface 170D, and the application processor.
- the audio module 170 is used to convert digital audio information into analog audio signal output, and is also used to convert analog audio input into digital audio signals.
- the audio module 170 can also be used to encode and decode audio signals.
- the audio module 170 can be arranged in the processor 110, or some functional modules of the audio module 170 can be arranged in the processor 110.
- the audio module 170 of the first electronic device can convert the sound collected by the microphone 170C into corresponding voice information. Voice information from the second electronic device can also be converted into audio information.
- the speaker 170A also called a "speaker" is used to convert an audio electrical signal into a sound signal.
- the terminal 100 can listen to music or listen to a hands-free call through the speaker 170A.
- the voice message can be played through the speaker 170A or the receiver.
- the receiver 170B also called a "handset" is used to convert audio electrical signals into sound signals.
- the voice can be received by placing the receiver 170B close to the ear.
- Microphone 170C also known as “microphone” or “microphone” is used to convert sound signals into electrical signals. When making a call or sending a voice message, the user can make a sound by approaching the microphone 170C with his mouth to input the sound signal into the microphone 170C.
- the terminal 100 can be provided with at least one microphone 170C. In other embodiments, the terminal 100 can be provided with two microphones 170C, which can not only collect sound signals but also realize noise reduction function. In other embodiments, the terminal 100 can also be provided with three, four or more microphones 170C to realize the collection of sound signals, noise reduction, identification of sound sources, and realization of directional recording functions, etc.
- the first electronic device can collect the user's voice through the microphone 170C, and the sound is processed by the audio module 170, the processor, etc. to obtain the corresponding voice information.
- the electronic device can send the voice information to the opposite device.
- FIG3 is another schematic diagram of the structure of an electronic device provided in an embodiment of the present application.
- the electronic device may be the first electronic device and/or the second electronic device described above.
- the electronic device includes a call module 310, a semantic understanding module 311, an interface display module 312, a personalized text-to-speech module 313, a general text-to-speech module 314, a speech-to-text module 315, a text input module 316, a personalized timbre selection module 317, a personalized timbre generation module 318 and other modules.
- the call module 310 can be used to monitor the call status such as an incoming call and communicate with the second electronic device.
- the semantic understanding module 311 is used to understand the call information of the second electronic device from the call module 310, and generate reply information according to a preset reply template and/or understood semantics.
- the interface display module 312 is used to display the interface information of the first electronic device.
- the call interface of the first electronic device the call mode information of the first electronic device, the information automatically replied by the first electronic device according to the understanding of the semantic understanding module 311, the information manually replied by the user received by the first electronic device, the reply information of the second electronic device, etc.
- the call mode includes an automatic answering mode or a manual answering mode.
- the personalized text-to-speech module 313 is used to convert the text information received by the first electronic device into a voice information with a personalized tone, for example, converting the text information manually replied by the user on the interface display module into a voice information with a personalized tone.
- the universal text-to-speech module 314 is used to convert the text information received by the first electronic device into a voice information with a universal tone. For example, the text information of the intelligent AI automatic reply displayed on the interface display module is converted into a voice information with a universal AI tone.
- the voice-to-text module 315 is used to convert the voice information from the second electronic device into text information, and display it on the interface display module of the first electronic device.
- the text input module 316 the first electronic device receives the text information input by the user through the text input module 316, and displays it in the interface display module of the first electronic device.
- the text input module 316 is also used to trigger the first electronic device to enter the user automatic reply mode. For example, when the first electronic device detects that the user clicks the text input module 316, it triggers the first electronic device to enter the user automatic reply mode.
- the text input module 316 is also used to detect whether the user has completed the text input.
- the personalized timbre selection module 317 is used for the first electronic device to select a personalized timbre.
- the personalized timbre generating module 318 is used for the first electronic device to generate a personalized timbre for selection by the personalized timbre selecting module 317 .
- FIG3 is only an exemplary illustration of the functional modules in the electronic device and the connection relationship between the modules. It should be understood that the electronic device may include more or fewer modules than shown in the figure, or combine some modules, or split some modules, or have different module arrangements and connection relationships.
- the modules shown in the figure may be implemented in hardware, software, or a combination of software and hardware.
- a call scenario between the user's mobile phone 1 and the contact A's mobile phone 2 is taken as an example for introduction.
- the mobile phone 1 may display a main interface 401.
- the main interface 401 may be an interface displayed after the mobile phone 1 is turned on, but it is not limited thereto and may also be other interfaces, such as a background interface, such as a negative one screen.
- the main interface 401 may include various applications (applications, APPs) installed on the mobile phone 1.
- APPs applications
- it may include icons of APPs that come with the mobile phone 1, such as a clock, calendar, gallery, memo, application store, settings, music player, calculator, sports health, weather, camera, phone, information, address book 402, etc.
- it may also include third-party APPs, such as WeChat, Alipay, online games, etc.
- mobile phone 1 can initiate a call. If the user wants to initiate a call with other users, the user can click on the address book 402 in the main interface 401 of mobile phone 1. As shown in Figure 4B, in response to this operation, mobile phone 1 can jump from the main interface 401 to the contact selection interface 403.
- the contact selection interface 403 may include icons of various contacts, such as "Contact A” 404, "Contact B" 405, "Contact C", etc., so that the user can select the contact to be called in the contact selection interface 403.
- the user can select the contact in the contact selection interface 403 and click on initiate call 406.
- mobile phone 1 can initiate a call to mobile phone 2 of "Contact A" 404.
- mobile phone 1 can receive calls. If the user receives a call from another user (such as contact A), as shown in Figure 5, the interface of mobile phone 1 can display interface 501.
- interface 501 may include the name of the caller contact A, answer button 502, hang up button 503, reply button 504, and smart AI answer button 505.
- the mobile phone 1 can display a call interface 601, and the user can talk to the contact A.
- the call interface 601 may include an icon of the selected contact A, and may also include icons related to the call, such as: call duration, recording, waiting, video call, mute, contact, speaker, hang up 602, more buttons 603, etc.
- the user wants to end the call he can click the hang up button 602.
- the mobile phone 1 can end the call with the mobile phone 2 of the contact A.
- the user can click the smart AI answer button 505 in the mobile phone interface 501 as shown in FIG5 .
- the mobile phone 1 can perform a smart AI call.
- the mobile phone 1 can automatically talk to the mobile phone 2 through the smart AI technology, or the mobile phone 1 can receive the reply information input by the user in the mobile phone 1, and convert the information into voice and send it to the mobile phone 2.
- FIG. 7A shows an example of mobile phone 1 executing intelligent AI call.
- mobile phone 1 answers the call from mobile phone 2.
- the interface 701 includes an icon of contact A, and can also include icons related to the call, such as: call duration, hang up button 702, microphone button 703, speaker button 704, return call button 705, dialog box 706, text input box 707, text send button 708, voice input button 709, card 710, etc.
- the icon of contact A, call duration, hang up button 702, microphone button 703, speaker button 704, and return call button 705 are located in card 710, and card 710 can be expanded and suspended above interface 701, or can be shrunk and suspended above interface 701.
- card 710 is expanded and suspended above interface 701, and the user can click on a blank space of the expanded card 710.
- card 710 can be shrunk and suspended above interface 701, as shown in (b) of FIG. 7A .
- the user can click on a blank space of the shrunk card 710, and in response to the user clicking on a blank space of the shrunk card 710, card 710 can be expanded and suspended above interface 701.
- the text automatically replied by the intelligent AI of mobile phone 1, the text manually replied by the user, and the text converted from the voice of the user (contact A) of mobile phone 2 can all be displayed in the text box of dialog box 706.
- the user can manually enter the text to be replied to contact A in the text input box 706.
- the words "auto answer mode” can be displayed in the dialog box 706 of the interface 701, indicating that the mobile phone 1 is in the intelligent AI automatic reply mode at this time.
- the mobile phone 1 can determine the text that needs to be replied to the contact A by itself, and display the text of the automatic reply in the dialog box 706.
- the mobile phone 1 can display the content of the automatic reply in the text box 7061 in the dialog box 706, "Hello, I am the owner's intelligent AI assistant, he is not convenient to answer the phone now, you can tell me what you want".
- the mobile phone 1 converts the automatic reply text into a voice message of the voice 1, and sends the voice message of the voice 1 to the mobile phone 2 of the contact A.
- the mobile phone 2 receives the voice message and plays the voice message using the voice 1.
- the contact A can hear the voice message of the voice 1, "Hello, I am the owner's intelligent AI assistant, he is not convenient to answer the phone now, you can tell me what you want".
- the content of the automatic reply can be a reply template pre-set by the user, or a reply content automatically generated by the intelligent AI according to the semantics of the reply message of the other party on the call.
- the user can set the intelligent AI to first reply to the reply template pre-set by contact A after answering the call.
- the template is "Hello, I am the owner's intelligent AI assistant. He is not convenient to answer the phone now. You can tell me if you have anything.”
- the reply content is automatically generated according to the semantics of the reply message of contact A. For example, if contact A replies "What time is it now?", the intelligent AI can reply to contact A according to the time in mobile phone 1, "It is 8 o'clock Beijing time now.”
- mobile phone 1 can recognize the call voice of contact A and convert the voice content into text to be displayed in dialog box 706 of interface 701. For example, mobile phone 1 recognizes that the call voice content of contact A is "Hello, why don't you answer the phone?", and mobile phone 1 can convert the voice content into text "Hello, why don't you answer the phone?". As shown in (a) of Figure 7B, mobile phone 1 can display the converted text information in text box 7062 of dialog box 706. In this way, the user of mobile phone 1 can view the call content of contact A through dialog box 706.
- text box 7061 and text box 7062 are displayed in dialog box 706 with different UIs to distinguish the content of the reply by the user of mobile phone 1 from the content of the reply by contact A of mobile phone 2.
- the content of the reply by the user of mobile phone 1 is displayed in Songti font format in text box 7061 with a black background
- the content of the reply by contact A of mobile phone 2 is displayed in Songti font format in text box 7062 with a white background. This is convenient for the user to check and understand the content of the call later.
- the user of mobile phone 1 can directly manually enter text in the text input box of the call interface to reply.
- the user of mobile phone 1 after viewing the call content of contact A "Hello, why don't you answer the phone", the user of mobile phone 1 hopes to manually enter information to reply to contact A, and can enter the information that he wants to reply to contact A in the input box 805 (such as "Why do you think so, I'm in a meeting, what do you want?").
- mobile phone 1 responds to the user's manual reply operation (an example of the first operation), triggering mobile phone 1 to switch from automatic reply mode to manual reply mode.
- the mobile phone 1 can automatically switch to manual answering mode, and can display the words "manual answering mode" in the dialog box 706, indicating that the mobile phone 1 is in the user manual reply mode at this time.
- the above-mentioned first operation can also be an operation in which the cursor of the mobile phone hits the text input box 707, the user types the first word, etc., and the embodiment of the present application does not limit this.
- the duration manually input by the user is less than the first threshold (e.g., 10 seconds), and after the user enters the reply text, the reply text can be sent (an example of the second operation).
- the send button 708 To trigger mobile phone 1 to send the input information to dialog box 706.
- mobile phone 1 can display the content of the manual reply in the text box 7064 in the dialog box 706 "Why do you think so, I'm in a meeting, what do you want?".
- Mobile phone 1 converts the manually replied text into a voice message of tone 2, and sends the voice message of tone 2 to mobile phone 2 of contact A.
- Mobile phone 2 receives the voice message and plays the voice message using tone 2, and contact A can hear the voice message of tone 2 "Why do you think so, I'm in a meeting, what do you want?".
- the duration of manual input by the user can be the time difference between the user clicking the text input box 707 to call out the mobile phone keyboard and the user clicking the send button 708 after completing the information input.
- text box 7064 and text box 7061 are displayed in dialog box 706 with different UI effects (such as different fonts, different text box shapes, colors) to distinguish the content of the automatic reply of the intelligent AI of mobile phone 1 from the content of the manual reply of the user.
- the fonts of the text box of the automatic reply of the intelligent AI and the text box of the manual reply of the user are both in Kaiti
- the background color of the text box of the automatic reply of the intelligent AI is white
- the background color of the text box of the manual reply of the user is green.
- the text box of the automatic reply of the intelligent AI is a square box
- the text box of the manual reply of the user is in the shape of a bubble.
- the content of the automatic reply of the intelligent AI of mobile phone 1 is displayed in the text box 7061 with a black background in the font format of Songti
- the content of the manual reply of the user is displayed in the text box 7064 with a black background in the font format of Lishu.
- UI effects listed in the above-mentioned intelligent AI automatic reply text box and the user manual reply text box are merely examples, and the intelligent AI automatic reply text box and the user manual reply text box may also include other UI effects, which are not limited in this application.
- mobile phone 1 may automatically reply with a prompt message: "The owner is manually inputting content, the content is long, please wait”.
- the prompt message is displayed in text box 7063 in dialog box 706.
- the prompt message is displayed in text box 7063 with a black background in Songti font format.
- Mobile phone 1 converts the prompt message into a voice message of timbre 1, and sends the voice message of timbre 1 to mobile phone 2 of contact A.
- Mobile phone 2 receives the voice message and plays the voice message using timbre 1.
- Contact A can hear the voice prompt message of timbre 1 "The owner is manually inputting content, the content is long, please wait”. In this way, contact A can be reminded that the user is manually inputting a reply message, preventing contact A from hanging up the phone due to not receiving a reply for a long time.
- mobile phone 1 can automatically reply to mobile phone 2 with a prompt message "The owner is manually inputting content, the content is long, please wait.”
- a second threshold e.g. 20 seconds
- the mobile phone 1 may display the content of the manual reply "Why do you think so, I'm in a meeting, what do you want?" in the text box 7064 in the dialog box 706.
- the mobile phone 1 converts the manually replied text into a voice message of the voice 2, and sends the voice message of the voice 2 to the mobile phone 2 of the contact A.
- the mobile phone 2 receives the voice message and plays the voice message using the voice 2.
- the contact A can hear the voice message of the voice 2 "Why do you think so, I'm in a meeting, what do you want?".
- mobile phone 2 can use different tones to play the information automatically replied by the intelligent AI of mobile phone 1 and the information manually replied by the user of mobile phone 1, so that the user of mobile phone 2 (such as contact A) can distinguish whether the content replied by mobile phone 1 is automatically replied by the intelligent AI or manually replied by the user.
- contact A judges by the tone that the content replied by mobile phone 1 is the content manually replied by the user, it can be known that the current call is with the user, not with the intelligent AI, which improves the call efficiency and the call experience of both parties.
- mobile phone 1 continues to recognize the conversation voice of contact A of mobile phone 2, and converts the voice content into text and displays it in dialog box 706 of interface 701.
- mobile phone 1 recognizes that the conversation voice content of contact A is "Oh, I want to ask you out to play", and mobile phone 1 can convert the voice content into text "Oh, I want to ask you out to play”.
- mobile phone 1 can display the converted text information in text box 7065 of dialog box 706. In this way, the user of mobile phone 1 can continue to view the conversation content of contact A through dialog box 706.
- the user clicks the text input box 707 within the first threshold (e.g., 10 seconds) to call out the mobile phone keyboard.
- the text input is still not completed after a long time, for example, the manual input content is not completed within the third threshold (e.g., 60 seconds), and the send button 708 is clicked to send the manual reply information to the dialog box 706, or the user does not click the text input box 707 within the first threshold time to call out the mobile phone keyboard.
- the mobile phone 1 can automatically switch from the manual reply mode to the intelligent AI automatic reply mode.
- the third threshold is greater than the first threshold.
- mobile phone 1 can automatically switch to the automatic reply mode to quickly respond to mobile phone 2.
- mobile phone 1 can display the words “automatic answer mode” again in the dialog box 706 of interface 701, indicating that mobile phone 1 is in the intelligent AI automatic reply mode at this time, and mobile phone 1 can determine the text to be replied to contact A according to the content of the reply of mobile phone 2, and display the text of the automatic reply in the dialog box 706.
- mobile phone 1 can display the content of the automatic reply “Feedback the time and place, I will record it for you” in the text box 7066 in the dialog box 706 according to the content of the text box 7065.
- Mobile phone 1 converts the text of the automatic reply into the voice information of timbre 1, and sends the voice information of timbre 1 to mobile phone 2 of contact A.
- Mobile phone 2 receives the voice information and plays the voice information using timbre 1.
- Contact A can hear the voice information of timbre 1 “Feedback the time and place, I will record it for you”.
- the intelligent AI assistant of mobile phone 1 can continue to communicate with contact A until the call is ended. Alternatively, the user can intervene in the call again and enter the manual reply mode.
- the mobile phone 1 can play the voice information of contact A alone.
- the mobile phone 1 can start the function of playing the other party's voice.
- FIG10 (b) after the function of playing the other party's voice is turned on, the information of contact A received by the mobile phone 1 can be played in real time.
- the user can click on the text box 7062 of the contact A reply in the dialog box 706.
- the mobile phone 1 can play the voice message of the contact A corresponding to the content in the text box 7062.
- the voice message of contact A can be played in other ways, for example, the user can long press the text box 7062, and in response to the user's long press of the text box 7062, the mobile phone 1 can play the voice message of contact A corresponding to the content in the text box 7062.
- the embodiment of the present application is not limited to this.
- the user can input the content to be replied by voice.
- the user can turn on the microphone of the mobile phone 1 to input voice.
- the position of the text input box 707 presents a voice control 707', and the user can input voice through the voice control 707'.
- the user can long press the voice control 707' to input voice.
- the voice control 707' can be converted into the form shown in (c) of FIG. 11, indicating that the user is inputting voice.
- the user inputs "Why do you think so? I'm in a meeting. What do you want?" by voice.
- the content of the user's voice input can be displayed in the form of a voice box 7067 in the dialog box 706, and the duration of the voice content in the voice box is 5s.
- the mobile phone 1 can play the voice message.
- the user can long press the voice box to convert the voice into text and display it in the text box.
- the user long presses the voice box 7067 to convert the voice message into text
- the dialog box 706 displays the text box 7068 corresponding to the voice box 7067
- the text box 7068 displays the text "Why do you think so, I'm in a meeting, what do you want?".
- mobile phone 1 can send the voice message input by the user's voice to mobile phone 2 of contact A, and mobile phone 2 receives the voice message and plays the voice message using tone 2, so that contact A can hear the voice message of tone 2 "Why do you think so, I'm in a meeting, what do you want?".
- the mobile phone 1 can also play all the information during the call between the contact A of the mobile phone 2 and the user of the mobile phone 1.
- the embodiment of the present application is not limited to this.
- the user can click the hang up button 702 in the mobile phone 1 interface 701.
- the mobile phone 1 can end the call with the mobile phone 2.
- the user can click the return call button 705 in the mobile phone 1 interface 701.
- the mobile phone 1 can have a normal voice call with the mobile phone 2.
- the mobile phone 1 can display a call interface 601 such as shown in FIG. 6, and the interface 601 may not include a dialog box 706 such as shown in FIG. 7B.
- mobile phone 1 uses different tones to send voice messages automatically replied by the intelligent AI and voice messages manually replied by the user.
- mobile phone 1 can convert the text message automatically replied or the text message manually replied by the user into a voice message with the corresponding tone and send it to mobile phone 2.
- Mobile phone 2 receives the voice message and plays it to the user of mobile phone 2 (such as contact A). For example, mobile phone 1 will automatically reply to the text message "Hello, I am the owner's intelligent AI assistant. He is not available to answer the phone now. Please Is there anything you can tell me?" is converted into a voice message of the corresponding tone 1 and sent to mobile phone 2.
- Mobile phone 2 receives the voice message of tone 1 and plays the voice message of tone 1.
- mobile phone 1 uses tone 1 to send the voice message of the smart AI automatic reply, and uses tone 2 to send the voice message of the user's manual reply, so that contact A can judge whether the received voice message is the smart AI automatic reply message or the user's manual reply message through voice messages of different tones, which is convenient for improving call efficiency.
- the user can customize the timbre 1 corresponding to the voice message automatically replied by the intelligent AI and the timbre 2 corresponding to the voice message manually replied by the user in the mobile phone 1.
- the timbre 1 and the timbre 2 can be generated by the personalized text to speech (TTS) technology of the mobile phone 1.
- TTS personalized text to speech
- the mobile phone 1 can also collect the user's timbre as the timbre 2 corresponding to the voice message manually replied by the user.
- the user can click the icon of the setting application in the mobile phone 1 to trigger the mobile phone 1 to jump to the interface 801 shown in (b) of FIG12A .
- the interface 801 displays WLAN controls, Bluetooth controls, mobile network controls, smart AI controls 802, and the like.
- the user can click on the smart AI control 802 of the interface 801.
- the mobile phone 1 can jump to the interface 803 shown in (c) of FIG. 12A.
- the interface 803 displays the smart AI call tone selection control 804 and multiple controls related to smart AI, such as the smart AI scene control 805, the smart AI search control, and the smart AI suggestion control.
- the user can click the smart AI call tone selection button 804 in the interface 803.
- the mobile phone 1 can jump to the interface 806.
- the interface 806 is a smart AI call tone selection interface, and the interface 806 may include a smart AI tone selection button 807 and a user tone selection button 808.
- the user can click the smart AI tone selection button 807 of the interface 803.
- the mobile phone 1 can jump to the tone interface 809 shown in (e) of Figure 12A.
- the tone interface 809 may include an add sound button 812 and official sound controls and custom sound controls.
- the official sound control is one or more.
- One official sound control corresponds to one sound type.
- the official sound control may include one or more of an intellectual female voice control, an innocent child voice control, and a clear male voice control.
- the custom sound control may include my sound control 811 and/or other user's sound control as shown in (e) of FIG. 12A above.
- My sound control 811 may be a sound pre-recorded by a user.
- Other users may include friends, relatives, and the like.
- the custom sound control includes the sound control of the other user, and the mobile phone 1 has recorded the voice of friend Xiao Ming.
- the custom sound control may include the custom sound control of Xiao Ming's voice as shown in (e) of FIG. 12A.
- the timbre of the intelligent AI automatic reply can be set.
- the user can select the intellectual female voice control 810 in the official sound control.
- the mobile phone 1 can set the intellectual female voice as the timbre 1 of the intelligent AI automatic reply.
- the intelligent AI timbre selection control 807 in the interface 806 can display that the selected voice is an intellectual female voice. In this way, the user can intuitively see what the timbre of the intelligent AI is, improving the user experience.
- the user's timbre can be set.
- the user can click the user timbre selection button 808 of the interface 806.
- the mobile phone 1 can jump to the timbre interface 809 shown in (b) of Figure 12B.
- the user can select the my sound control 811 in the custom sound control, and in response to the user clicking the my sound control 811, the mobile phone 1 can set the my sound control 811 to the user's timbre 2.
- the user timbre selection control 808 in the interface 806 can display that the selected sound is my sound. In this way, the user can intuitively see what kind of timbre the user's timbre is, thereby improving the user experience.
- the mobile phone 1 can record a custom sound.
- the user can click Add Sound Button 812.
- the mobile phone 1 jumps to the interface 813 shown in (b) of Figure 13.
- Interface 813 includes a self-recording control 814 and an invite others to record control 815.
- the user can click the self-recording control 814, and in response to the user clicking the self-recording control 814, the mobile phone 1 can jump to the self-recording interface 816 shown in (c) of Figure 13.
- Interface 813 includes a "self-recording" prompt and a recording completion control 817. The user can record his own voice by talking, singing, etc.
- the user can click the recording completion control 817.
- the mobile phone 1 can jump to the interface 809 shown in (d) of Figure 13, and the "my voice" button 811 is displayed in the interface 809.
- the mobile phone can play the sound recorded by the user.
- the user can invite Xiao Ming to record Xiao Ming's timbre. The embodiments of this application are not repeated.
- the above-mentioned mobile phone 1 can be used to set a user call scenario, so that when the user is busy (such as the user is in a meeting), when the mobile phone 1 receives an incoming call, the intelligent AI can be used directly to answer the call, avoiding missed calls and improving the calling user's call experience.
- the setting process of the above user scene may include: first, the user may click the smart AI scene button 803 in the interface 801 shown in (a) of FIG. 14.
- the mobile phone 1 may jump to the smart AI scene interface 811 shown in (b) of FIG. 14.
- the smart AI scene interface 811 includes a scene button 812 in a meeting, a scene button in a call, etc.
- the user can click the in-meeting scene button 812 in the intelligent AI scene interface 811.
- the mobile phone 1 can be set to the user's in-meeting scene.
- the call can be answered directly based on the intelligent AI. In this way, even if the call cannot be answered in time because the user is in a meeting, there will be no missed calls, which ensures the timeliness of answering the call, and also enables the caller to know the user's current situation, thereby improving the call experience of both parties.
- the intelligent AI automatic reply mode, the user manual reply mode and the voice call mode can be switched freely.
- the user can click the smart AI answer button 505 of the interface 501.
- the mobile phone 1 can pop up a call mode option box 506 on the interface 501.
- the call mode option box 506 includes a smart AI reply option 507 and a manual reply option 508.
- the mobile phone 1 enters the smart AI automatic reply mode, and displays the interface 701 shown in (b) of FIG.
- the dialog box 706 in the interface 701 displays the words "automatic answer mode”.
- the mobile phone 1 enters the user manual reply mode, and displays the interface 701 shown in (c) of FIG. 15 , and the dialog box 706 in the interface 701 displays the words "manual answer mode”.
- the user can click the more button 603 in the voice call interface 601, and in response to the user clicking the more button 603, the mobile phone 1 can pop up a call mode option box 604 on the interface 601.
- the call mode option box 604 includes an intelligent AI reply option 605 and a manual reply option 606.
- the mobile phone 1 enters the intelligent AI automatic reply mode, and displays the interface 701 shown in (b) of FIG15, and the dialog box 706 in the interface 701 displays the words "automatic answer mode".
- the mobile phone 1 enters the user manual reply mode, and can display the interface 701 shown in (c) of FIG15.
- the words "manual answer mode" are displayed in the dialog box 706 in the interface 701.
- the mobile phone 1 can switch freely between the intelligent AI automatic reply mode, the user manual reply mode and the voice call mode.
- the normal progress of the call can be guaranteed.
- users can intervene in the call in time to ensure the efficiency of the call and meet the user's personalized call needs.
- mobile phone 1 receives a call from mobile phone 2 .
- a user receives a call from another user (such as contact A).
- Mobile phone 1 is the user's mobile phone
- mobile phone 2 is contact A's mobile phone.
- Mobile phone 1 displays an incoming call interface.
- an interface 501 as shown in FIG. 5 may be displayed.
- the user can directly click the answer button 502 to answer the call.
- the mobile phone 1 can display the call interface 601 shown in FIG. 6, and the user can make a call with the contact A.
- the user can click the smart AI answer button 505 in the mobile phone interface 501 as shown in FIG. 5 .
- the user can click on the smart AI reply option 507, and the mobile phone 1 enters the smart AI automatic reply mode, displaying the interface 701 shown in (b) of Figure 15, and the words "automatic answer mode" are displayed in the dialog box 706 in the interface 701.
- the user may click on the manual reply option 508, and the mobile phone 1 enters the user manual reply mode, displaying the interface 701 shown in FIG. 15(c), in which the words "manual answer mode" are displayed in the dialog box 706.
- Mobile phone 1 displays an AI call interface according to the user's answering instruction.
- the user can click the smart AI answer button 505 in the mobile phone interface 501 as shown in FIG5 .
- the mobile phone 1 can answer the call from the mobile phone 2 through the smart AI call, and can display an AI call interface such as shown in FIG7A (a) or FIG7A (b).
- Mobile phone 1 receives the reply information input by the user on the AI call interface.
- the mobile phone 1 can receive a reply message manually replied by the user.
- the user of the mobile phone 1 can enter the reply message he wants to reply to the contact A in the input box 805. For example, "Why do you think so? I'm in a meeting. What can I do for you?".
- the user can input the reply information of the automatic reply by voice, and the reply information of the automatic reply by voice input is displayed in the form of a voice box 7067 in the dialog box 706 .
- the intelligent AI of mobile phone 1 can automatically reply to the message.
- the content of the automatic reply of the intelligent AI of mobile phone 1 can be “Hello, I am the intelligent AI assistant of the owner. He is not available to answer the phone now. You can talk to me if you have anything to say.”
- Mobile phone 1 can display the content in the text box 7061 in the dialog box 706.
- the user clicks on the text input box 707 to call out the mobile phone keyboard and prepare to enter text, which means that the user wants to manually enter a reply message, and the mobile phone can display the words "manual answering mode" in the dialog box 706, indicating that the mobile phone 1 is in the user manual reply mode.
- the manual answering mode the user can input voice or text into the mobile phone 1, and the voice or text can be used as a reply message.
- the mobile phone 1 can convert the manual reply content "Why do you think so, I'm in a meeting, what do you want?" displayed in the text box 7064 in the dialog box 706 into a voice message 2 of timbre 2, and send the voice message 2 of timbre 2 to the mobile phone 2 of contact A.
- the content of the manual reply by the user is displayed in the text box 7064 with a black background in the font format of Lishu. In this way, when the user checks the content of the call again, it is easy to understand, which improves the user's experience.
- the mobile phone 2 receives the voice message 2 and plays the voice message 2 of timbre 2, and the contact A can listen to the voice message 2 of timbre 2 "Why do you think so, I'm in a meeting, what do you want?".
- the reply message input by the user is a voice message.
- mobile phone 1 can perform voice input.
- a voice control 707’ can be presented at the position of the text input box 707 of mobile phone 1, and the user can perform voice input through the voice control 707’.
- the user can perform voice input by long pressing the voice control 707’.
- the voice control 707’ can be converted into a form as shown in (c) of FIG. 11 , indicating that the user is inputting voice.
- Mobile phone 1 can use the user’s voice message as voice message 2 of timbre 2. Alternatively, mobile phone 1 can convert the user’s voice message into voice message 2 of timbre 2. Mobile phone 1 sends voice message 2 of timbre 2 to mobile phone 2 of contact A. Afterwards, mobile phone 2 receives voice message 2 and plays voice message 2 in tone 2, and contact A can hear voice message 2 in tone 2 "Why do you think so? I'm in a meeting. What can I do for you?".
- the voice of tone 2 is quite different from the voice of tone 1 during automatic reply, so that contact A of mobile phone 2 can more clearly distinguish the content of manual ⁇ automatic reply.
- mobile phone 1 detects the time length of the user manually inputting information in text box 706. If the user completes the input within a period of time, it is deemed that the manual reply condition is met, and mobile phone 1 sends a reply message to mobile phone 2 according to the reply message manually input by the user. Voice message 2 of voice 2. If the user fails to complete the input within a period of time, it is deemed that the manual reply condition is not met, and mobile phone 1 may first send a prompt message to mobile phone 2, "The owner is manually inputting content, the content is long, please wait", to prevent contact A of mobile phone 2 from hanging up due to not receiving a reply from mobile phone 1 for a long time. Mobile phone 1 can continue to receive the reply information manually input by the user in text box 706.
- mobile phone 1 may send a prompt message to mobile phone 2 again until it is detected that the user has completed the input.
- Mobile phone 1 can send voice message 2 of voice 2 to mobile phone 2 according to the reply information input by the user, "Why do you think so? I'm in a meeting. What can I do for you?"
- mobile phone 1 receives voice message 3 from mobile phone 2 .
- Mobile phone 1 presents voice message 3 .
- the mobile phone 1 can convert the voice message 3 into corresponding text and present the text corresponding to the voice message 3. For example, as shown in (a) of FIG. 7B , the mobile phone 1 recognizes the voice message 3 of the contact A as “Hello, why don’t you answer the phone?” and converts the voice message 3 into text “Hello, why don’t you answer the phone?” and displays it in the text box 7062 of the dialog box 706. In this way, the user of the mobile phone 1 can view the voice message 3 replied by the contact A through the dialog box 706.
- the reply voice message 3 of the mobile phone 2 of the contact A is displayed in a text box 7062 with a white background in Songti font format, so that the user can check and understand the content of the call later.
- mobile phone 1 after receiving the voice message 3 of contact A from mobile phone 2, mobile phone 1 can play the voice message 3 of contact A according to the playing requirements of the user of mobile phone 1. For example, as shown in FIG10(b), after the function of playing the other party's voice is turned on, the user can click on the text box 7062 of contact A's mobile phone 2 reply in the dialog box 706. In response to the user's operation of clicking on the text box 7062, mobile phone 1 can play the voice message 3 of mobile phone 2 reply corresponding to the content in the text box 7062.
- mobile phone 1 may first enter the automatic reply mode, and then, according to the user's manual reply intention, mobile phone 1 may enter the manual reply mode.
- mobile phone 1 After answering the call from mobile phone 2, mobile phone 1 first enters the manual reply mode according to the user's manual reply intention, and then, when the automatic reply condition is met, mobile phone 1 enters the automatic reply mode.
- mobile phone 1 can adaptively switch between the automatic reply mode and the manual reply mode.
- mobile phone 1 after mobile phone 1 answers a call from mobile phone 2, mobile phone 1 first enters the automatic reply mode, as shown in (a) of FIG7B , in response to the user clicking the smart AI answer button 502 of the interface 501 shown in FIG5 , mobile phone 1 answers the call from mobile phone 2 through smart AI, and can display the smart AI call interface 701.
- mobile phone 1 can first automatically enter the smart AI automatic reply mode, and can display the words "automatic answer mode" in the dialog box 706 of the interface 701, indicating that mobile phone 1 is in the smart AI automatic reply mode at this time. Afterwards, according to the user's manual reply intention, mobile phone 1 can switch to the manual reply mode.
- mobile phone 1 after mobile phone 1 answers the call of mobile phone 2, mobile phone 1 first enters the automatic reply mode.
- the reply message of the automatic reply of the intelligent AI of mobile phone 1 may be "Hello, I am the intelligent AI assistant of the owner. He is not convenient to answer the phone now. You can talk to me if you have anything to say”.
- Mobile phone 1 can display the content in the text box 7061 in the dialog box 706, and convert the reply information into a voice message 1 of tone 1.
- the content of the automatic reply of the intelligent AI of mobile phone 1 is displayed in the text box 7061 with a black background in Songti font format.
- mobile phone 1 first enters manual reply mode according to the user's manual reply intention, the user clicks on the text input box 707 within the first threshold time to call out the mobile phone keyboard.
- the manual input content is not completed within the third threshold time (for example, 60s), and the send button 708 is clicked to send the manual reply information to the dialog box 706.
- mobile phone 1 can switch from manual reply mode to intelligent AI automatic reply mode.
- mobile phone 1 can display the words "automatic answer mode" again in the dialog box 706 of interface 701, indicating that mobile phone 1 is in intelligent AI automatic reply mode at this time.
- mobile phone 1 After mobile phone 1 receives the message “Oh, I want to invite you out to play” from mobile phone 2, the user of mobile phone 1 does not manually reply to the message for a long time. In order to avoid contact A of mobile phone 2 waiting for a long time, mobile phone 1 can automatically switch to automatic reply mode to quickly respond to mobile phone 2. In some examples, mobile phone 1 can display the content of the automatic reply “Feedback the time and place, I will record it for you” in the text box 7066 in the dialog box 706 according to the content of the text box 7065, and convert the text of the automatic reply into a voice message of tone 1.
- mobile phone 1 sends voice message 1 of tone 1 to mobile phone 2 .
- mobile phone 1 sends a voice message 1 with tone 1, "Hello, I am the owner's smart AI assistant. He is not available to answer the phone right now. You can talk to me if you have anything to say" to mobile phone 2 of contact A.
- mobile phone 2 receives the voice message and plays it using tone 1. In this way, contact A can hear the voice message of tone 1 "Hello, I am the owner's smart AI assistant. He is not available to answer the phone now. You can tell me what you want to say.”
- mobile phone 1 sends voice message 1 of tone 1, “Please tell me the time and place, I will record it for you”, to mobile phone 2 of contact A.
- mobile phone 2 receives the voice message and plays the voice message using tone 1, and contact A can hear the voice message of tone 1 "Feed back the time and place, I will record it for you".
- mobile phone 1 can obtain the authorized user portrait and automatically reply to mobile phone 2 according to the user portrait. In this way, contact A of mobile phone 2 can know the current status of the user according to the reply information of mobile phone 1.
- the words "automatic answering mode” can be displayed in the dialog box 706 of the interface 701.
- Mobile phone 1 can determine that the user is in a meeting based on the user portrait, and can display the content of the automatic reply in the text box 7160 in the dialog box 706, "I am the owner's intelligent AI assistant, he is in a meeting now, it is not convenient to answer the phone, you can tell me if there is anything.”
- Mobile phone 1 can convert the automatic reply text into a voice message of timbre 1, and send the voice message of timbre 1 to mobile phone 2 of contact A.
- Mobile phone 2 receives the voice message and plays the voice message using timbre 1.
- Contact A can hear the voice message of timbre 1 "I am the owner's intelligent AI assistant, he is in a meeting now, it is not convenient to answer the phone, you can tell me if there is anything", and can know the current status of the user of mobile phone 1 according to the voice message, so that the call efficiency can be improved.
- contact A of mobile phone 2 can reply the message "In a meeting again” according to the current status of the user of mobile phone 1.
- Mobile phone 1 may display the reply in a text box 7161 in dialog box 706 .
- the user of mobile phone 1 can click text box 707 to trigger mobile phone 1 to switch to manual answering mode.
- mobile phone 1 in manual answering mode, can display the words "manual answering mode" in dialog box 706 of interface 701.
- the user of mobile phone 1 can manually input "recently there is a lot of work and regular meetings" in text box 707, and mobile phone 1 can send the text input by the user to text box 7162 of dialog box 706 for display.
- the information that mobile phone 1 automatically replies to mobile phone 2 is not limited to voice and text, and the information that the user replies through mobile phone 1 is not limited to voice and text, for example, it can also be pictures, videos, emoticons, business cards, etc.
- the interface 701 may further include a control 711 , and the user may click the control 711 and add pictures, business cards, audio, and video to be sent.
- the mobile phone 1 sends the information to be sent selected by the user to the mobile phone 2 .
- FIG3 above introduces the functions of the modules included in the electronic device, and FIG20 shows the process of multiple modules cooperating with each other to implement the calling method.
- mobile phone 1 receives a call request from mobile phone 2 through call module 310, and displays it in answering interface 501 of mobile phone 1 as shown in FIG5 through interface display module 312. If the user clicks answer button 502 of answering interface 501, mobile phone 1 enters voice call mode. If the user clicks smart AI answer button 505 of answering interface 501, mobile phone 1 enters automatic reply mode by default.
- the mobile phone 1 receives a first operation from the user and can switch from the automatic reply mode to the manual reply mode. After the mobile phone 1 enters the manual reply mode, if the first operation from the user is not detected within a first time period (e.g., 10 seconds) or the second operation from the user is not detected within a second time period (e.g., 30 seconds), the mobile phone 1 can switch from the manual reply mode to the automatic reply mode.
- the second time period is greater than the first time period.
- the interface display module 312 displays the words "automatic answer mode" as shown in (a) of Figure 7B, prompting the user that mobile phone 1 is currently in automatic reply mode, and mobile phone 1 can automatically talk to the user of mobile phone 2 (contact A).
- mobile phone 1 can convert the voice information from contact A into text information through the voice-to-text module 315, and display the text information in the interface 701 of mobile phone 1 as shown in (a) of Figure 7B through the interface display module 312.
- the text information is displayed in the text box 7062.
- the intelligent AI of mobile phone 1 understands the information from contact A through the semantic understanding module 311 and generates an automatic reply text.
- Mobile phone 1 converts the automatic reply text into a voice message of tone 1 through the general text-to-speech module 314 and sends it to mobile phone 2; the automatic reply text can be displayed in the interface shown in (b) of Figure 9. For example, the automatic reply text is displayed in the text box 7066 in (b) of FIG. 9 .
- mobile phone 1 receives the first operation from the user, mobile phone 1 is in automatic reply mode, and the interface display module 312 displays the words "manual answer mode" as shown in (a) of Figure 7B, prompting the user that mobile phone 1 is currently in manual reply mode, and the user can manually talk to the user (contact A) of mobile phone 2 through mobile phone 1.
- mobile phone 1 can convert the voice information from contact A into text information through voice-to-text module 315, and display the text information in the interface of mobile phone 1 as shown in (a) of Figure 9 through interface display module 312.
- the text information is displayed in text box 7065 in (a) of Figure 9.
- the user enters the reply text information in mobile phone 1 through text input module 316, and detects whether the second operation is received within the second time length.
- the user completes the input within the first time period and sends it to mobile phone 2.
- the interface display module 312 displays the text information input by the user in the text box 7064 in the mobile phone interface 701, and the mobile phone 1 converts the text information into voice information of tone 2 through the personalized text-to-speech module 313 and sends it to the mobile phone 2.
- the mobile phone 1 sends a first voice prompt message.
- the first voice prompt message can be displayed in the interface shown in FIG8.
- the interface display module 312 can display the first voice prompt message in the text box 7063 in (a) of FIG8.
- the first voice prompt message is converted into a voice message of tone 1 by the universal text-to-speech module 314 and sent to the mobile phone 2.
- the mobile phone 1 switches from the manual reply mode to the automatic reply mode, wherein the second time period is greater than the first time period.
- the above one or more interfaces are exemplary. In other embodiments, there may be other interface design methods.
- steps in the method embodiment may be equivalently replaced by other possible steps.
- some steps in the method embodiment may be optional and may be deleted in certain usage scenarios.
- other possible steps may be added to the method embodiment.
- Some other embodiments of the present application provide a device, which may be the above-mentioned second electronic device or the first electronic device or a component in the first electronic device or a component in the second electronic device (such as a chip system).
- the device may include: a display screen, a memory, and one or more processors.
- the display screen, the memory, and the processor are coupled.
- the memory is used to store computer program code, and the computer program code includes computer instructions.
- the processor executes the computer instructions, the electronic device may execute the various functions or steps executed by the mobile phone in the above method embodiment.
- the structure of the electronic device may refer to the structure of the electronic device shown in FIG. 2C.
- the core structure of the electronic device can be represented as the structure shown in Figure 21.
- the electronic device includes: a processing module 151, an input module 152, a storage module 153, a display module 154 and a communication module 155.
- the processing module 151 may include at least one of a central processing unit (CPU), an application processor (AP) or a communication processor (CP).
- the processing module 151 may perform operations or data processing related to the control and/or communication of at least one of other elements of the user electronic device.
- the processing module 151 is used to support the first electronic device 100 to execute S101-S109 in Figure 17.
- the input module 152 is used to obtain the instructions or data input by the user and transmit the obtained instructions or data to other modules of the electronic device.
- the input method of the input module 152 may include touch, gesture, approaching the screen, etc., or voice input.
- the input module can be the screen of the electronic device, obtain the user's input operation and generate an input signal according to the obtained input operation, and transmit the input signal to the processing module 151.
- the input module 152 is used to obtain text information or voice information input by the user, and reference can be made to the input interface schematic diagrams shown in Figures 7B, 8 and 11.
- the storage module 153 may include a volatile memory and/or a non-volatile memory.
- the storage module is used to store the user terminal device
- the storage module 153 is used for the first electronic device 100 to store preset template reply information, personalized tone information, and call recording information.
- the display module 154 may include, for example, a liquid crystal display (LCD), a light emitting diode (LED) display, an organic light emitting diode (OLED) display, a micro-electromechanical system (MEMS) display, or an electronic paper display. It is used to display content (e.g., text, images, videos, icons, symbols, etc.) that can be viewed by a user. Optionally, the display module 154 is used to display the content shown in FIG. 7A on the first electronic device 100.
- LCD liquid crystal display
- LED light emitting diode
- OLED organic light emitting diode
- MEMS micro-electromechanical system
- the communication module 155 is used to support the personal terminal to communicate with other personal terminals (through the communication network).
- the communication module can be connected to the network via wireless communication or wired communication to communicate with other personal terminals or network servers.
- the wireless communication can adopt at least one of the cellular communication protocols, such as Long Term Evolution (LTE), Advanced Long Term Evolution (LTE-A), Code Division Multiple Access (CDMA), Wideband Code Division Multiple Access (WCDMA), Universal Mobile Telecommunications System (UMTS), Wireless Broadband (WiBro) or Global System for Mobile Communications (GSM).
- Wireless communication may include, for example, short-range communication.
- Short-range communication may include at least one of wireless fidelity (Wi-Fi), Bluetooth, near field communication (NFC), magnetic stripe transmission (MST) or GNSS.
- the communication module 155 is used to support the first electronic device to communicate with the second electronic device.
- the communication module 155 is used to support the first electronic device to communicate with the second electronic device.
- the communication module 155 is used to support
- the device shown in Figure 21 may also include more or fewer components, or split some components, or have other ways of laying out components, which is not limited in this embodiment of the present application.
- the embodiment of the present application also provides a chip system, as shown in Figure 22, the chip system includes at least one processor 161 and at least one interface circuit 162.
- the processor 161 and the interface circuit 162 can be interconnected through lines.
- the interface circuit 162 can be used to receive signals from other devices (such as the memory of the electronic device).
- the interface circuit 162 can be used to send signals to other devices (such as the processor 161).
- the interface circuit 162 can read the instructions stored in the memory and send the instructions to the processor 161.
- the electronic device can perform the various steps in the above embodiments.
- the chip system can also include other discrete devices, which are not specifically limited in the embodiment of the present application.
- An embodiment of the present application also provides a computer storage medium, which includes computer instructions.
- the computer instructions When the computer instructions are executed on the above-mentioned electronic device, the electronic device executes each function or step executed by the mobile phone in the above-mentioned method embodiment.
- the embodiment of the present application also provides a computer program product.
- the computer program product When the computer program product is run on a computer, the computer is enabled to execute each function or step executed by the mobile phone in the above method embodiment.
- the disclosed devices and methods can be implemented in other ways.
- the device embodiments described above are only schematic.
- the division of the modules or units is only a logical function division. There may be other division methods in actual implementation, such as multiple units or components can be combined or integrated into another device, or some features can be ignored or not executed.
- Another point is that the mutual coupling or direct coupling or communication connection shown or discussed can be through some interfaces, indirect coupling or communication connection of devices or units, which can be electrical, mechanical or other forms.
- the units described as separate components may or may not be physically separated, and the components shown as units may be one physical unit or multiple physical units, that is, they may be located in one place or distributed in multiple different places. Some or all of the units may be selected according to actual needs to achieve the purpose of the present embodiment.
- each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.
- the above-mentioned integrated unit may be implemented in the form of hardware or in the form of software functional units.
- the integrated unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a readable storage medium.
- the technical solution of the embodiment of the present application is essentially or the part that contributes to the prior art or all or part of the technical solution can be embodied in the form of a software product, which is stored in a storage medium and includes a number of instructions for making a device (which can be a single-chip microcomputer, chip, etc.) or a processor (processor) Execute all or part of the steps of the method described in each embodiment of the present application.
- the aforementioned storage medium includes: a USB flash drive, a mobile hard disk, a read only memory (ROM), a random access memory (RAM), a magnetic disk or an optical disk, and other media that can store program codes.
Landscapes
- Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- Business, Economics & Management (AREA)
- General Business, Economics & Management (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Multimedia (AREA)
- Acoustics & Sound (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Telephone Function (AREA)
Abstract
La présente demande concerne un procédé d'appel et un dispositif électronique, qui sont appliqués au domaine technique des terminaux, et peuvent améliorer l'expérience d'appel d'utilisateur et l'efficacité d'appel. Le procédé consiste à : afficher une interface d'appel, l'interface d'appel comprenant l'identifiant d'un second utilisateur, le second utilisateur étant un utilisateur utilisant un second dispositif électronique, et un appel étant un appel entre le second dispositif électronique et un premier dispositif électronique ; envoyer des premières informations vocales d'une première tonalité au second dispositif électronique, les premières informations vocales étant déterminées par le premier dispositif électronique ; recevoir des premières informations entrées par un premier utilisateur au moyen de l'interface d'appel ; et envoyer des secondes informations vocales d'une seconde tonalité au second dispositif électronique selon les premières informations, la première tonalité étant différente de la seconde tonalité.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211521752.3A CN118118593A (zh) | 2022-11-30 | 2022-11-30 | 通话方法及电子设备 |
CN202211521752.3 | 2022-11-30 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2024114233A1 true WO2024114233A1 (fr) | 2024-06-06 |
Family
ID=91207536
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2023/127971 WO2024114233A1 (fr) | 2022-11-30 | 2023-10-30 | Procédé d'appel et dispositif électronique |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN118118593A (fr) |
WO (1) | WO2024114233A1 (fr) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110401777A (zh) * | 2019-08-02 | 2019-11-01 | 上海尊源通讯技术有限公司 | 一种基于移动通讯终端的ai电话秘书系统 |
CN111131592A (zh) * | 2018-10-31 | 2020-05-08 | 奇酷互联网络科技(深圳)有限公司 | 一种自动回复方法、通信设备及具有存储功能的装置 |
CN111683175A (zh) * | 2020-04-22 | 2020-09-18 | 北京捷通华声科技股份有限公司 | 一种自动应答来电的方法、装置、设备及存储介质 |
CN113726956A (zh) * | 2021-08-04 | 2021-11-30 | 北京小米移动软件有限公司 | 一种来电接听控制方法、装置、终端设备及存储介质 |
-
2022
- 2022-11-30 CN CN202211521752.3A patent/CN118118593A/zh active Pending
-
2023
- 2023-10-30 WO PCT/CN2023/127971 patent/WO2024114233A1/fr unknown
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111131592A (zh) * | 2018-10-31 | 2020-05-08 | 奇酷互联网络科技(深圳)有限公司 | 一种自动回复方法、通信设备及具有存储功能的装置 |
CN110401777A (zh) * | 2019-08-02 | 2019-11-01 | 上海尊源通讯技术有限公司 | 一种基于移动通讯终端的ai电话秘书系统 |
CN111683175A (zh) * | 2020-04-22 | 2020-09-18 | 北京捷通华声科技股份有限公司 | 一种自动应答来电的方法、装置、设备及存储介质 |
CN113726956A (zh) * | 2021-08-04 | 2021-11-30 | 北京小米移动软件有限公司 | 一种来电接听控制方法、装置、终端设备及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
CN118118593A (zh) | 2024-05-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11722449B2 (en) | Notification message preview method and electronic device | |
CN112154640B (zh) | 一种消息的播放方法及终端 | |
CN112422874B (zh) | 一种摄像头的控制方法及电子设备 | |
WO2020078337A1 (fr) | Procédé de traduction et dispositif électronique | |
CN109286725B (zh) | 翻译方法及终端 | |
US20230299806A1 (en) | Bluetooth Communication Method, Wearable Device, and System | |
JP2022518656A (ja) | 着信があるときに電子デバイス上に映像を提示するための方法、および電子デバイス | |
WO2023024852A1 (fr) | Procédé de notification de message court et dispositif terminal électronique | |
JP2022537012A (ja) | マルチ端末マルチメディアデータ通信方法及びシステム | |
CN114115770A (zh) | 显示控制的方法及相关装置 | |
WO2023088209A1 (fr) | Procédé de transmission de données audio inter-dispositifs et dispositifs électroniques | |
CN114449110A (zh) | 一种电子设备的控制方法和装置 | |
CN113918110A (zh) | 投屏交互方法、设备、系统、存储介质和产品 | |
US20240272865A1 (en) | Audio playing method, electronic device, and system | |
WO2022127670A1 (fr) | Procédé et système d'appel, et dispositif associé | |
CN113301544B (zh) | 一种音频设备间语音互通的方法及设备 | |
CN115408492A (zh) | 资源显示的方法、终端及服务器 | |
WO2021042881A1 (fr) | Procédé de notification par message et dispositif électronique | |
WO2024114233A1 (fr) | Procédé d'appel et dispositif électronique | |
CN113805825B (zh) | 设备之间的数据通信方法、设备及可读存储介质 | |
CN116055633A (zh) | 来电处理方法、系统、电子设备及存储介质 | |
CN115242994A (zh) | 视频通话系统、方法和装置 | |
CN113656099A (zh) | 一种应用快捷启动方法、装置和终端设备 | |
CN113672404A (zh) | 一种显示方法和电子终端设备 | |
WO2022183941A1 (fr) | Procédé et dispositif de réponse de message |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 23896387 Country of ref document: EP Kind code of ref document: A1 |