CN111369995A - Call method, device, electronic equipment and medium - Google Patents

Call method, device, electronic equipment and medium Download PDF

Info

Publication number
CN111369995A
CN111369995A CN202010110891.1A CN202010110891A CN111369995A CN 111369995 A CN111369995 A CN 111369995A CN 202010110891 A CN202010110891 A CN 202010110891A CN 111369995 A CN111369995 A CN 111369995A
Authority
CN
China
Prior art keywords
language
data
text data
voice data
mobile terminal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010110891.1A
Other languages
Chinese (zh)
Inventor
梁永毅
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Yulong Computer Telecommunication Scientific Shenzhen Co Ltd
Original Assignee
Yulong Computer Telecommunication Scientific Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Yulong Computer Telecommunication Scientific Shenzhen Co Ltd filed Critical Yulong Computer Telecommunication Scientific Shenzhen Co Ltd
Priority to CN202010110891.1A priority Critical patent/CN111369995A/en
Publication of CN111369995A publication Critical patent/CN111369995A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • G10L13/04Details of speech synthesis systems, e.g. synthesiser structure or memory management
    • G10L13/047Architecture of speech synthesisers
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/005Language recognition
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72433User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for voice messaging, e.g. dictaphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72436User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for text messaging, e.g. short messaging services [SMS] or e-mails
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72448User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72484User interfaces specially adapted for cordless or mobile telephones wherein functions are triggered by incoming communication events
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/74Details of telephonic subscriber devices with voice recognition means

Landscapes

  • Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Telephone Function (AREA)

Abstract

The application discloses a conversation method, a conversation device, electronic equipment and a conversation medium. In the method, when a call event aiming at a preset number is detected, first voice data corresponding to a first language of the call event can be acquired, the first voice data is converted into first character data corresponding to a second language, and then the first character data is displayed. By applying the technical scheme of the application, the voice data during the call can be converted into the text data of the specific language, and the text data is displayed during the call. This also avoids the problem that the corresponding information cannot be effectively acquired because the foreign user cannot understand the voice operation prompt of the local language in the related art.

Description

Call method, device, electronic equipment and medium
Technical Field
The present application relates to data processing technologies, and in particular, to a method, an apparatus, an electronic device, and a medium for communication.
Background
Due to the rise of the communications era and society, mobile terminals have been continuously developed with the use of more and more users.
With the rapid development of the communication era, it has become a normal state that people use mobile terminals to make calls to obtain information required by themselves. Further, taking the example of calling a customer service telephone, the speech language currently served by each customer service telephone is generated by default for the native language or the common language (such as english, etc.). However, this approach also has the problem of not being able to serve all users, which in turn degrades the user experience.
Disclosure of Invention
The embodiment of the application provides a conversation method, a conversation device, electronic equipment and a conversation medium; the method and the device can solve the problem that the user cannot effectively acquire the corresponding information because the user cannot understand the voice operation prompt of the specific language in the related technology.
According to an aspect of an embodiment of the present application, a method for calling is provided, including:
when a call event aiming at a preset number is detected, acquiring first voice data corresponding to the call event, wherein the first voice data corresponds to a first language;
converting the first voice data into first character data, wherein the first character data corresponds to a second language;
and displaying the first text data.
Optionally, in another embodiment based on the foregoing method of the present application, after the displaying the first text data, the method further includes:
when a first operation instruction is received, converting the first text data into the second text data, wherein the first operation instruction is used for converting the first text data into the second text data, the second text data corresponds to a third language, and the third language and the second language are different languages;
and displaying the second text data.
Optionally, in another embodiment based on the foregoing method of the present application, after the displaying the first text data, the method further includes:
when a second operation instruction is received, converting the first character data into second voice data, wherein the second voice data corresponds to a fourth language;
and when a playing instruction is received, playing the second voice data.
Optionally, in another embodiment based on the foregoing method of the present application, after the displaying the first text data, the method further includes:
and when a third operation instruction is received, storing the first character data into a target storage area.
Optionally, in another embodiment based on the foregoing method of the present application, after the displaying the first text data, the method further includes:
and when a fourth operation instruction is received, clearing the first text data, wherein the fourth operation instruction is an operation performed on the first voice data.
Optionally, in another embodiment based on the foregoing method of the present application, before the converting the first voice data into the first text data, the method further includes:
and determining the second language based on the language set by an operating system in the mobile terminal or the language corresponding to the attribution of the mobile terminal, wherein the mobile terminal is a terminal corresponding to the call event.
Optionally, in another embodiment based on the foregoing method of the present application, before the acquiring the first voice data corresponding to the call event, the method further includes:
determining a preset area based on the equipment information of the mobile terminal; wherein the equipment information comprises at least one of operator information, registered network information and IMEI information;
and determining that the mobile terminal is positioned outside the preset area according to the position of the mobile terminal.
According to another aspect of the embodiments of the present application, there is provided an apparatus for calling, including:
the device comprises an acquisition module, a processing module and a processing module, wherein the acquisition module is used for acquiring first voice data corresponding to a call event when the call event aiming at a preset number is detected, and the first voice data corresponds to a first language;
a conversion module configured to convert the first voice data into first text data, the first text data corresponding to a second language;
a display module configured to display the first text data.
According to another aspect of the embodiments of the present application, there is provided an electronic device including:
a memory for storing executable instructions; and
a display for displaying with the memory to execute the executable instructions to accomplish the operations of any of the above-mentioned methods of calling.
According to yet another aspect of the embodiments of the present application, a computer-readable storage medium is provided for storing computer-readable instructions, which when executed perform the operations of any of the above-mentioned methods for calling.
In the method and the device, when a call event aiming at a preset number is detected, first voice data corresponding to a first language of the call event can be acquired, the first voice data is converted into first character data corresponding to a second language, and then the first character data is displayed. By applying the technical scheme of the application, the voice data during the call can be converted into the text data of the specific language, and the text data is displayed during the call. This also avoids the problem that the corresponding information cannot be effectively acquired because the foreign user cannot understand the voice operation prompt of the local language in the related art.
The technical solution of the present application is further described in detail by the accompanying drawings and examples.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments of the application and together with the description, serve to explain the principles of the application.
The present application may be more clearly understood from the following detailed description with reference to the accompanying drawings, in which:
FIG. 1 is a schematic diagram of a video call system architecture according to the present application;
fig. 2 is a schematic diagram of a method for calling according to the present application;
fig. 3 is a schematic diagram of a call method proposed in the present application;
fig. 4 a-4 c are schematic diagrams of a mobile phone display illustrating a method for calling according to the present application;
FIG. 5 is a schematic diagram of the apparatus for conversation according to the present application;
fig. 6 is a schematic view of an electronic device according to the present application.
Detailed Description
Various exemplary embodiments of the present application will now be described in detail with reference to the accompanying drawings. It should be noted that: the relative arrangement of the components and steps, the numerical expressions, and numerical values set forth in these embodiments do not limit the scope of the present application unless specifically stated otherwise.
Meanwhile, it should be understood that the sizes of the respective portions shown in the drawings are not drawn in an actual proportional relationship for the convenience of description.
The following description of at least one exemplary embodiment is merely illustrative in nature and is in no way intended to limit the application, its application, or uses.
Techniques, methods, and apparatus known to those of ordinary skill in the relevant art may not be discussed in detail but are intended to be part of the specification where appropriate.
It should be noted that: like reference numbers and letters refer to like items in the following figures, and thus, once an item is defined in one figure, further discussion thereof is not required in subsequent figures.
In addition, technical solutions between the various embodiments of the present application may be combined with each other, but it must be based on the realization of the technical solutions by a person skilled in the art, and when the technical solutions are contradictory or cannot be realized, such a combination of technical solutions should be considered to be absent and not within the protection scope of the present application.
It should be noted that all the directional indicators (such as upper, lower, left, right, front and rear … …) in the embodiment of the present application are only used to explain the relative position relationship between the components, the motion situation, etc. in a specific posture (as shown in the drawings), and if the specific posture is changed, the directional indicator is changed accordingly.
A method for conducting a call according to an exemplary embodiment of the present application is described below in conjunction with fig. 1-3. It should be noted that the following application scenarios are merely illustrated for the convenience of understanding the spirit and principles of the present application, and the embodiments of the present application are not limited in this respect. Rather, embodiments of the present application may be applied to any scenario where applicable.
The embodiments of the present application can be applied to a wireless communication system, and it should be noted that the wireless communication system mentioned in the embodiments of the present application includes, but is not limited to, a 5G Mobile communication system and an Enhanced Mobile broadband (eMBB) of a next generation Mobile communication system, a URLLC, and a mass machine-Type Communications (mtc).
In the embodiment of the present application, a Terminal (Terminal device) includes, but is not limited to, a Mobile Station (MS), a Mobile Terminal (Mobile Terminal), a Mobile phone (Mobile phone), a handset (handset), a portable device (portable equipment), and the like, and the Terminal may communicate with one or more core networks through a Radio Access Network (RAN), for example, the Terminal may be a Mobile phone (or referred to as a "cellular" phone), a computer with a wireless communication function, and the Terminal may also be a portable, pocket, hand-held, computer-embedded, or vehicle-mounted Mobile device or apparatus.
Fig. 1 is a schematic diagram of a communication system architecture provided in the present application.
Referring to fig. 1, a communication system 01 includes a network device 101 and a terminal 102; wherein the network device 101 is deployed using NSA mode. When the communication system 01 includes a core network, the network device 101 may also be connected to the core network. The network device 101 may also communicate with an Internet Protocol (IP) network 200, such as the Internet (Internet), a private IP network, or other data network. The network device provides services for terminals within the coverage area. For example, referring to fig. 1, network device 101 provides wireless access to one or more terminals within the coverage area of network device 101. In addition, the network devices can also communicate with each other.
The network device 101 may be a device for communicating with a terminal. The network device may be a relay station, an access point, a vehicle-mounted device, etc. In an end-to-end (D2D) communication system, the network Device may also be a terminal that functions as a base station. A terminal may include various handheld devices, vehicle-mounted devices, wearable devices, computing devices, or other processing devices connected to a wireless modem with wireless communication capabilities, as well as various forms of User Equipment (UE), Mobile Station (MS), and the like.
In the related art, for a foreign user, the service phone cannot be considered and effectively covers the language corresponding to the foreign user, so that when the user needs to query the service phones of local operators, car renters, hotels and the like abroad, the voice operation prompt of the local language cannot be understood, so that the corresponding information cannot be effectively acquired, and further the user experience is reduced.
The application also provides a conversation method, a conversation device, a target terminal and a medium.
Fig. 2 schematically shows a flow chart of a method for calling according to an embodiment of the present application. As shown in fig. 2, the method includes:
s101, when a call event aiming at a preset number is detected, first voice data corresponding to the call event is obtained, and the first voice data corresponds to a first language.
It should be noted that, in the present application, the device for detecting a call event is not specifically limited, and may be, for example, an intelligent device or a server. The smart device may be a PC (Personal Computer), a smart phone, a tablet PC, an e-book reader, an MP3(Moving Picture Experts Group audio Layer iii, motion Picture Experts compression standard audio Layer 3) communicator, an MP4(Moving Picture Experts Group audio Layer IV, motion Picture Experts compression standard audio Layer 4) communicator, a portable terminal device with a display function such as a portable Computer, and the like.
The preset number is not specifically limited in the present application, and may be, for example, a number that enables a user to query related information. Such as a local rental car company number, a local rental hotel number, a local carrier number, etc. Or, the preset number may also be a number marked in advance by the user. Or, the preset number may also be a number belonging to a specific area or a specific operator, etc.
In addition, the present application does not specifically limit the timing of detecting the call event, and for example, each call event of the mobile terminal may be monitored in real time. The call event monitoring of the mobile terminal may be started after the mobile terminal is detected to be located in a specific area, or whether the call event is generated for a preset number may be monitored when the call event occurs in the mobile terminal within a specific time period.
Further, the first voice data in the present application may be voice data occurring in the call event, for example, dialing a customer service call, and the first voice data may be non-artificial voice data occurring in the process of dialing a customer service call. The user can inquire the information needed by the user according to the prompt of the non-artificial voice data. Or, the first voice data can also be artificial voice data which appears in the process of calling the customer service telephone. It can be understood that when the artificial voice data is non-native voice for the user who makes a call, the problem that the user cannot inquire the required information due to the fact that the user cannot understand the meaning of the artificial voice data also occurs.
It should be noted that the present application does not specifically limit the first language corresponding to the first voice data. For example, chinese, english, japanese, etc.
S102, the first voice data is converted into first character data, and the first character data corresponds to the second language.
According to the method and the device, after the first voice data are obtained, the voice data can be converted into the corresponding first character data by using a voice character conversion technology. It should be noted that the first text data is text data corresponding to the second language. The second language may be the same as or different from the first language. In addition, the second language is not specifically limited in this application, and may be, for example, chinese, english, japanese, etc.
Further, the method for converting the first voice data into the first text data is not specifically limited in the present application, and for example, the voice data may be converted into the text data by using a voice recognition algorithm. A speech recognition algorithm is an algorithm that converts the vocabulary content in human speech into a computer-readable binary code or character sequence. The algorithm can adopt the existing dynamic time reduction algorithm to convert the voice content into the text content, thereby carrying out subsequent information processing.
And S103, displaying the first character data.
Furthermore, after the first voice data are converted into the first text data, the first text data can be displayed in a preset area. For example, may be displayed on the current display screen of the mobile terminal.
In the method and the device, when a call event aiming at a preset number is detected, first voice data corresponding to a first language of the call event can be acquired, the first voice data is converted into first character data corresponding to a second language, and then the first character data is displayed. By applying the technical scheme of the application, the voice data during the call can be converted into the text data of the specific language, and the text data is displayed during the call. This also avoids the problem that the corresponding information cannot be effectively acquired because the foreign user cannot understand the voice operation prompt of the local language in the related art.
In another possible embodiment of the present application, after S103 (displaying the first text data), the following various manners may be implemented:
the first mode is as follows:
when a first operation instruction is received, converting the first character data into second character data, wherein the first operation instruction is used for converting the first character data into the second character data, the second character data corresponds to a third language, and the third language is different from the second language;
and displaying the second text data.
Further, after the first text data is displayed, whether a first operation instruction for converting the first text data into the second text data is received or not can be monitored. And after the first operation instruction is determined to be received, converting the first character data into second character data corresponding to a third language.
It should be noted that the third language in the present application should be a different language from the second language. For example, when the second language is english, the third language may correspond to other languages such as chinese or japanese. And when the second language is chinese, the third language may correspond to other languages such as english or korean.
It can be understood that, after the first voice data is converted into the first text data corresponding to the second language, if the user is not familiar with the second language, and thus cannot understand the first text data, the user may further generate the first operation instruction. For converting the first text data of the second language into second text data of a third language. The third language may be a language specified by a user, or the language may be a language determined according to a preset rule.
It should be noted that, in the process of converting the first text data into the second text data, the first text data may be uniformly converted into the text data of the specific language, and then the text data of the specific language is converted into the second text data.
The second mode is as follows:
when a second operation instruction is received, converting the first character data into second voice data, wherein the second voice data corresponds to a fourth language;
and when a playing instruction is received, playing the second voice data.
Furthermore, after the first text data are displayed, whether a second operation instruction for converting the first text data into the second voice data is received or not can be monitored. And after the second operation instruction is determined to be received, converting the first character data into second voice data corresponding to a fourth language.
It should be noted that the fourth language in the present application may be a different language from the second language and the third language. For example, when the second language is English, the fourth language may correspond to other languages such as Chinese or Japanese. And when the third language is chinese, the fourth language may correspond to other languages such as english or korean. Alternatively, the fourth language in the present application may be the same language as the second language or the third language. For example, when the second language is english, the fourth language may also correspond to english. And when the third language is chinese, the fourth language may correspond to chinese.
It can be understood that, after the first voice data is converted into the first text data corresponding to the second language, if the user is inconvenient to view the screen at this time, the user may further generate the second operation instruction. For converting the first text data into second voice data. The fourth language may be a language specified by a user, or the language may be a language determined according to a preset rule.
It should be noted that, in the present application, in the process of converting the first text data into the second voice data, the second voice data is not specifically limited, and may be, for example, voice data generated according to any user, or the like.
The third mode is as follows:
and when a third operation instruction is received, storing the first character data into the target storage area.
Further, after the first text data is displayed, whether a third operation instruction for storing the first text data in the target storage area is received or not can be monitored. And after determining that the third operation instruction is received, storing the first character data into a target storage area.
The target storage area is not specifically limited in the present application, and may be any storage area of the mobile terminal, or any storage area in the server, for example. Any storage area of other intelligent devices and the like.
In addition, the storage form of the first character data is not particularly limited, and for example, the first character data may be stored in a target storage area in a character format. Or converting the first text data into voice data and storing the voice data into the target storage area.
The fourth mode is that:
and when a fourth operation instruction is received, clearing the first character data, wherein the fourth operation instruction is an operation performed on the first voice data.
Further, after the first text data is displayed, whether a fourth operation instruction for clearing the first text data is received or not can be monitored. And after the fourth operation instruction is determined to be received, the first character data is cleared.
It should be noted that, the fourth operation instruction is not specifically limited in this application, and for example, the fourth operation instruction may be an operation instruction generated by a user, or may also be an operation instruction automatically generated when it is detected that a preset condition is met.
It can be understood that, after the first text data is displayed, since the space of the display area is limited, when the user views all the text data for convenience, the user may generate a fourth operation instruction for clearing the displayed first text data. Or, when the mobile terminal detects that the remaining display space of the current display area is small, a fourth operation instruction for clearing the displayed first character data can be automatically generated.
It is noted that, the present application may first backup the first text data before the first text data is cleared. And after the backup, the first character data is cleared in the display area.
Further optionally, in an embodiment of the present application, before S102 (converting the first voice data into the first text data), a specific embodiment is further included, as shown in fig. 3, including:
s201, determining a second language based on the language set by an operating system in the mobile terminal or the language corresponding to the attribution of the mobile terminal, wherein the mobile terminal is a terminal corresponding to the call event.
Furthermore, the problem that corresponding information cannot be effectively acquired due to the fact that voice operation prompts used for understanding the local language cannot be generated is avoided. Before performing text conversion on voice data, the second language displayed for a user needs to be determined first. So that the voice data can be directly viewed by the user when being converted into the character data corresponding to the second language subsequently.
Therefore, in the process of determining the second language, the second language can be determined by the display text data of the mobile terminal. It should be noted that the mobile terminal may be a mobile terminal that dials the call.
It can be understood that, when the language information corresponding to the displayed text data of the mobile terminal is the chinese language information, the application may regard the user generating the call event as the user familiar with the chinese language. Therefore, the present application can use the chinese language as the second language. Similarly, when it is detected that the language information corresponding to the displayed text data of the mobile terminal is english information, the application may consider that the user generating the call event is a user familiar with english language, and therefore the application may use english as the second language.
S202, when a call event aiming at a preset number is detected, first voice data corresponding to the call event is acquired.
Further, before detecting whether a call event for a preset number exists, the following steps may be first implemented:
determining a preset area based on the equipment information of the mobile terminal; the equipment information comprises at least one of operator information, registered network information and IMEI information;
and determining that the mobile terminal is positioned outside the preset area according to the position of the mobile terminal.
Further, in order to avoid the problem of system resource consumption caused by always monitoring whether the mobile terminal has a call event for a preset number, the method and the device for detecting the call event can start to detect whether the call event exists currently or not when detecting that the mobile terminal is not located in a preset area (for example, abroad).
Further, when determining the preset area for the mobile terminal, the determination may be made according to at least one of operator information, registered network information, and IMEI information of the mobile terminal. For example, according to the present application, when it is detected that the mobile terminal is china unicom according to the operator information of the mobile terminal, it may be determined that the preset area may be an area in china when the mobile terminal is a mobile terminal used in china. Or, when it is detected that the mobile terminal is AT & T (american telephone and telecommunications company) according to the operator information of the mobile terminal, it may be determined that the mobile terminal is a mobile terminal used in the united states, and the preset area may be an area within the united states.
Furthermore, according to the IMEI information of the mobile terminal, when the mobile terminal is determined to be a mobile terminal used in the united states, the preset area may be an area within the united states.
The IMEI information may be an International Mobile Equipment Identity (IMEI), which is a serial number of the mobile terminal, and may be used to identify each independent mobile communication device such as a mobile phone in the mobile phone network, which is equivalent to an Identity card of the mobile terminal.
Further, the area of the mobile terminal can be determined according to the monitored GPS information of the mobile terminal. For example, when it is determined that the mobile terminal is always in china in the past preset time period according to the GPS information of the mobile terminal, it may be determined that the mobile terminal is a mobile terminal used in china, and the preset area may be an area in the united states.
S204, the first voice data is converted into first character data, and the first character data corresponds to the second language.
And S205, displaying the first character data.
Furthermore, after the first voice data are converted into the first text data, the first text data can be displayed in a preset area. For example, may be displayed on the current display screen of the mobile terminal.
As shown in fig. 4a, taking the mobile terminal as a mobile phone for illustration, when a call event of the mobile phone for a preset number is detected, first voice data of a first language corresponding to the call event is acquired, the first voice data is converted into first text data corresponding to a chinese language (a second language), and the first text data of the chinese language is displayed through the display unit. So that the subsequent user can implement corresponding operations according to the first character data 1011, 1012, 1013. Further, as shown in fig. 4b, when the mobile phone detects that the call event is ended, the display text for storing the first text data in the target storage area may be displayed (as 1014). So that the subsequent user can implement the corresponding operation (e.g. 1015, 1016) according to the displayed text. Further, when the user clicks 1015 (stores the first text data in the target storage area), the application may generate corresponding text data (as shown in fig. 4 c) and store the text data in a specific area.
In the method and the device, when a call event aiming at a preset number is detected, first voice data corresponding to a first language of the call event can be acquired, the first voice data is converted into first character data corresponding to a second language, and then the first character data is displayed. By applying the technical scheme of the application, the voice data during the call can be converted into the text data of the specific language, and the text data is displayed during the call. This also avoids the problem that the corresponding information cannot be effectively acquired because the foreign user cannot understand the voice operation prompt of the local language in the related art.
In another embodiment of the present application, as shown in fig. 5, the present application further provides a device for conversation. The device comprises an acquisition module 301, a conversion module 302 and a display module 303, wherein:
the acquiring module 301 is configured to acquire first voice data corresponding to a call event when the call event for a preset number is detected, where the first voice data corresponds to a first language;
a conversion module 302 configured to convert the first voice data into first text data, the first text data corresponding to a second language;
a display module 303 configured to display the first text data.
In the method and the device, when a call event aiming at a preset number is detected, first voice data corresponding to a first language of the call event can be acquired, the first voice data is converted into first character data corresponding to a second language, and then the first character data is displayed. By applying the technical scheme of the application, the voice data during the call can be converted into the text data of the specific language, and the text data is displayed during the call. This also avoids the problem that the corresponding information cannot be effectively acquired because the foreign user cannot understand the voice operation prompt of the local language in the related art.
In another embodiment of the present application, the display module 303 further includes:
a display module 303, configured to convert the first text data into the second text data when receiving a first operation instruction, where the first operation instruction is used to convert the first text data into second text data, the second text data corresponds to a third language, and the third language is a different language from the second language;
a display module 303 configured to display the second text data.
In another embodiment of the present application, a display module 303, wherein:
the display module 303 is configured to, when a second operation instruction is received, convert the first text data into second voice data, where the second voice data corresponds to a fourth language;
a display module 303 configured to play the second voice data when receiving a play instruction.
In another embodiment of the present application, a display module 303, wherein:
and the display module 303 is configured to store the first text data into the target storage area when receiving a third operation instruction.
In another embodiment of the present application, the display module 303 further includes:
the display module 303 is configured to clear the first text data when a fourth operation instruction is received, where the fourth operation instruction is an operation performed on the first voice data.
In another embodiment of the present application, the method further includes determining module 304, wherein:
a determining module 304, configured to obtain display text data of a mobile terminal, where the mobile terminal is a terminal corresponding to the call event;
a determining module 304 configured to determine language information corresponding to the display text data;
a determining module 304, configured to use language information corresponding to the display text data as the second language.
In another embodiment of the present application, the determining module 304 further includes:
a determining module 304 configured to determine a preset region based on device information of the mobile terminal, the device information including at least one of operator information, registered network information, and IMEI information;
a determining module 304 configured to detect whether the call event exists when it is detected that the mobile terminal is not located in the preset area.
FIG. 6 is a block diagram illustrating a logical structure of an electronic device in accordance with an exemplary embodiment. For example, the electronic device 400 may be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a game console, a tablet device, a medical device, an exercise device, a personal digital assistant, and the like.
Referring to fig. 6, electronic device 400 may include one or more of the following components: a processor 401 and a memory 402.
Processor 401 may include one or more processing cores, such as a 4-core processor, an 8-core processor, or the like. The processor 401 may be implemented in at least one hardware form of a DSP (Digital Signal Processing), an FPGA (Field-Programmable Gate Array), and a PLA (Programmable Logic Array). The processor 401 may also include a main processor and a coprocessor, where the main processor is a processor for processing data in an awake state, and is also called a Central Processing Unit (CPU); a coprocessor is a low power processor for processing data in a standby state. In some embodiments, the processor 401 may be integrated with a GPU (Graphics Processing Unit), which is responsible for rendering and drawing the content required to be displayed by the display screen. In some embodiments, the processor 401 may further include an AI (artificial intelligence) processor for processing computing operations related to machine learning.
Memory 402 may include one or more computer-readable storage media, which may be non-transitory. Memory 402 may also include high speed random access memory, as well as non-volatile memory, such as one or more magnetic disk storage devices, flash memory storage devices. In some embodiments, a non-transitory computer readable storage medium in the memory 402 is configured to store at least one instruction for execution by the processor 401 to implement the interactive special effect calibration method provided by the method embodiments of the present application.
In some embodiments, the electronic device 400 may further optionally include: a peripheral interface 403 and at least one peripheral. The processor 401, memory 402 and peripheral interface 403 may be connected by bus or signal lines. Each peripheral may be connected to the peripheral interface 403 via a bus, signal line, or circuit board. Specifically, the peripheral device includes: at least one of radio frequency circuitry 404, touch screen display 405, camera 406, audio circuitry 407, positioning components 408, and power supply 409.
The peripheral interface 403 may be used to connect at least one peripheral related to I/O (Input/Output) to the processor 401 and the memory 402. In some embodiments, processor 401, memory 402, and peripheral interface 403 are integrated on the same chip or circuit board; in some other embodiments, any one or two of the processor 401, the memory 402 and the peripheral interface 403 may be implemented on a separate chip or circuit board, which is not limited by this embodiment.
The Radio Frequency circuit 404 is used for receiving and transmitting RF (Radio Frequency) signals, also called electromagnetic signals. The radio frequency circuitry 404 communicates with communication networks and other communication devices via electromagnetic signals. The rf circuit 404 converts an electrical signal into an electromagnetic signal to transmit, or converts a received electromagnetic signal into an electrical signal. Optionally, the radio frequency circuit 404 includes: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a subscriber identity module card, and so forth. The radio frequency circuitry 404 may communicate with other terminals via at least one wireless communication protocol. The wireless communication protocols include, but are not limited to: metropolitan area networks, various generation mobile communication networks (2G, 3G, 4G, and 5G), Wireless local area networks, and/or WiFi (Wireless Fidelity) networks. In some embodiments, the rf circuit 404 may further include NFC (Near Field Communication) related circuits, which are not limited in this application.
The display screen 405 is used to display a UI (User Interface). The UI may include graphics, text, icons, video, and any combination thereof. When the display screen 405 is a touch display screen, the display screen 405 also has the ability to capture touch signals on or over the surface of the display screen 405. The touch signal may be input to the processor 401 as a control signal for processing. At this point, the display screen 405 may also be used to provide virtual buttons and/or a virtual keyboard, also referred to as soft buttons and/or a soft keyboard. In some embodiments, the display screen 405 may be one, providing the front panel of the electronic device 400; in other embodiments, the display screen 405 may be at least two, respectively disposed on different surfaces of the electronic device 400 or in a folded design; in still other embodiments, the display screen 405 may be a flexible display screen disposed on a curved surface or a folded surface of the electronic device 400. Even further, the display screen 405 may be arranged in a non-rectangular irregular pattern, i.e. a shaped screen. The Display screen 405 may be made of LCD (Liquid Crystal Display), OLED (Organic Light-Emitting Diode), and other materials.
The camera assembly 406 is used to capture images or video. Optionally, camera assembly 406 includes a front camera and a rear camera. Generally, a front camera is disposed at a front panel of the terminal, and a rear camera is disposed at a rear surface of the terminal. In some embodiments, the number of the rear cameras is at least two, and each rear camera is any one of a main camera, a depth-of-field camera, a wide-angle camera and a telephoto camera, so that the main camera and the depth-of-field camera are fused to realize a background blurring function, and the main camera and the wide-angle camera are fused to realize panoramic shooting and VR (Virtual Reality) shooting functions or other fusion shooting functions. In some embodiments, camera assembly 406 may also include a flash. The flash lamp can be a monochrome temperature flash lamp or a bicolor temperature flash lamp. The double-color-temperature flash lamp is a combination of a warm-light flash lamp and a cold-light flash lamp, and can be used for light compensation at different color temperatures.
The audio circuit 407 may include a microphone and a speaker. The microphone is used for collecting sound waves of a user and the environment, converting the sound waves into electric signals, and inputting the electric signals to the processor 401 for processing, or inputting the electric signals to the radio frequency circuit 404 for realizing voice communication. For stereo capture or noise reduction purposes, the microphones may be multiple and disposed at different locations of the electronic device 400. The microphone may also be an array microphone or an omni-directional pick-up microphone. The speaker is used to convert electrical signals from the processor 401 or the radio frequency circuit 404 into sound waves. The loudspeaker can be a traditional film loudspeaker or a piezoelectric ceramic loudspeaker. When the speaker is a piezoelectric ceramic speaker, the speaker can be used for purposes such as converting an electric signal into a sound wave audible to a human being, or converting an electric signal into a sound wave inaudible to a human being to measure a distance. In some embodiments, audio circuitry 407 may also include a headphone jack.
The positioning component 408 is used to locate a current geographic location of the electronic device 400 to implement navigation or LBS (location based Service). The positioning component 408 may be a positioning component based on the GPS (global positioning System) of the united states, the beidou System of china, the graves System of russia, or the galileo System of the european union.
The power supply 409 is used to supply power to the various components in the electronic device 400. The power source 409 may be alternating current, direct current, disposable or rechargeable. When power source 409 comprises a rechargeable battery, the rechargeable battery may support wired or wireless charging. The rechargeable battery may also be used to support fast charge technology.
In some embodiments, the electronic device 400 also includes one or more sensors 410. The one or more sensors 410 include, but are not limited to: acceleration sensor 411, gyro sensor 412, pressure sensor 413, fingerprint sensor 414, optical sensor 415, and proximity sensor 416.
The acceleration sensor 411 may detect the magnitude of acceleration in three coordinate axes of a coordinate system established with the electronic apparatus 400. For example, the acceleration sensor 411 may be used to detect components of the gravitational acceleration in three coordinate axes. The processor 401 may control the touch display screen 405 to display the user interface in a landscape view or a portrait view according to the gravitational acceleration signal collected by the acceleration sensor 411. The acceleration sensor 411 may also be used for acquisition of motion data of a game or a user.
The gyro sensor 412 may detect a body direction and a rotation angle of the electronic device 400, and the gyro sensor 412 may cooperate with the acceleration sensor 411 to acquire a 3D motion of the user on the electronic device 400. From the data collected by the gyro sensor 412, the processor 401 may implement the following functions: motion sensing (such as changing the UI according to a user's tilting operation), image stabilization at the time of photographing, game control, and inertial navigation.
The pressure sensors 413 may be disposed on a side bezel of the electronic device 400 and/or on a lower layer of the touch display screen 405. When the pressure sensor 413 is arranged on the side frame of the electronic device 400, a holding signal of the user to the electronic device 400 can be detected, and the processor 401 performs left-right hand identification or shortcut operation according to the holding signal collected by the pressure sensor 413. When the pressure sensor 413 is disposed at the lower layer of the touch display screen 405, the processor 401 controls the operability control on the UI interface according to the pressure operation of the user on the touch display screen 405. The operability control comprises at least one of a button control, a scroll bar control, an icon control and a menu control.
The fingerprint sensor 414 is used for collecting a fingerprint of the user, and the processor 401 identifies the identity of the user according to the fingerprint collected by the fingerprint sensor 414, or the fingerprint sensor 414 identifies the identity of the user according to the collected fingerprint. Upon recognizing that the user's identity is a trusted identity, processor 401 authorizes the user to perform relevant sensitive operations including unlocking the screen, viewing encrypted information, downloading software, paying, and changing settings, etc. The fingerprint sensor 414 may be disposed on the front, back, or side of the electronic device 400. When a physical button or vendor Logo is provided on the electronic device 400, the fingerprint sensor 414 may be integrated with the physical button or vendor Logo.
The optical sensor 415 is used to collect the ambient light intensity. In one embodiment, the processor 401 may control the display brightness of the touch display screen 405 based on the ambient light intensity collected by the optical sensor 415. Specifically, when the ambient light intensity is high, the display brightness of the touch display screen 405 is increased; when the ambient light intensity is low, the display brightness of the touch display screen 405 is turned down. In another embodiment, the processor 401 may also dynamically adjust the shooting parameters of the camera assembly 406 according to the ambient light intensity collected by the optical sensor 415.
Proximity sensor 416, also known as a distance sensor, is typically disposed on the front panel of electronic device 400. The proximity sensor 416 is used to capture the distance between the user and the front of the electronic device 400. In one embodiment, the processor 401 controls the touch display screen 405 to switch from the bright screen state to the dark screen state when the proximity sensor 416 detects that the distance between the user and the front surface of the electronic device 400 gradually decreases; when the proximity sensor 416 detects that the distance between the user and the front of the electronic device 400 is gradually increased, the processor 401 controls the touch display screen 405 to switch from the breath screen state to the bright screen state.
Those skilled in the art will appreciate that the configuration shown in fig. 6 does not constitute a limitation of the electronic device 400, and may include more or fewer components than those shown, or combine certain components, or employ a different arrangement of components.
In an exemplary embodiment, there is also provided a non-transitory computer-readable storage medium, such as the memory 404, including instructions executable by the processor 420 of the electronic device 400 to perform a method of the above-described call, the method including: when a call event aiming at a preset number is detected, acquiring first voice data corresponding to the call event, wherein the first voice data corresponds to a first language; converting the first voice data into first character data, wherein the first character data corresponds to a second language; and displaying the first text data. Optionally, the instructions may also be executable by the processor 420 of the electronic device 400 to perform other steps involved in the exemplary embodiments described above. Optionally, the instructions may also be executable by the processor 420 of the electronic device 400 to perform other steps involved in the exemplary embodiments described above. For example, the non-transitory computer readable storage medium may be a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
In an exemplary embodiment, there is also provided an application/computer program product comprising one or more instructions executable by the processor 420 of the electronic device 400 to perform the method of telephony described above, the method comprising: when a call event aiming at a preset number is detected, acquiring first voice data corresponding to the call event, wherein the first voice data corresponds to a first language; converting the first voice data into first character data, wherein the first character data corresponds to a second language; and displaying the first text data. Optionally, the instructions may also be executable by the processor 420 of the electronic device 400 to perform other steps involved in the exemplary embodiments described above. Other embodiments of the present application will be apparent to those skilled in the art from consideration of the specification and practice of the invention disclosed herein. This application is intended to cover any variations, uses, or adaptations of the invention following, in general, the principles of the application and including such departures from the present disclosure as come within known or customary practice within the art to which the invention pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the application being indicated by the following claims.
It will be understood that the present application is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the application is limited only by the appended claims.

Claims (10)

1. A method for calling, comprising:
when a call event aiming at a preset number is detected, acquiring first voice data corresponding to the call event, wherein the first voice data corresponds to a first language;
converting the first voice data into first character data, wherein the first character data corresponds to a second language;
and displaying the first character data through a display unit.
2. The method of claim 1, wherein after said displaying said first textual data, further comprising:
when a first operation instruction is received, converting the first text data into the second text data; the first operation instruction is used for converting the first text data into second text data, the second text data corresponds to a third language, and the third language is different from the second language;
and displaying the second text data through the display unit.
3. The method of claim 1 or 2, wherein after said displaying said first textual data, further comprising:
when a second operation instruction is received, converting the first text data into second voice data; wherein the second speech data corresponds to a fourth language;
and when a playing instruction is received, playing the second voice data.
4. The method of claim 3, wherein after said displaying said first textual data, further comprising:
and when a third operation instruction is received, storing the first character data into a target storage area.
5. The method of claim 4, wherein after said displaying said first textual data, further comprising:
when a fourth operation instruction is received, clearing the first character data; wherein the fourth operation instruction is an operation performed on the first voice data.
6. The method of claim 1, wherein prior to said converting said first speech data to first textual data, further comprising:
and determining the second language based on the language set by an operating system in the mobile terminal or the language corresponding to the attribution of the mobile terminal, wherein the mobile terminal is a terminal corresponding to the call event.
7. The method of claim 6, wherein before the obtaining the first voice data corresponding to the call event, further comprising:
determining a preset area based on the equipment information of the mobile terminal; wherein the equipment information comprises at least one of operator information, registered network information and IMEI information;
and determining that the mobile terminal is positioned outside the preset area according to the position of the mobile terminal.
8. An apparatus for conversation, comprising:
the device comprises an acquisition module, a processing module and a processing module, wherein the acquisition module is used for acquiring first voice data corresponding to a call event when the call event aiming at a preset number is detected, and the first voice data corresponds to a first language;
a conversion module configured to convert the first voice data into first text data, the first text data corresponding to a second language;
a display module configured to display the first text data.
9. An electronic device, comprising:
a memory for storing executable instructions; and the number of the first and second groups,
a processor for display with the memory to execute the executable instructions to perform the operations of the method of calling of any of claims 1-7.
10. A computer-readable storage medium storing computer-readable instructions that, when executed, perform the operations of the method of any of claims 1-7.
CN202010110891.1A 2020-02-24 2020-02-24 Call method, device, electronic equipment and medium Pending CN111369995A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010110891.1A CN111369995A (en) 2020-02-24 2020-02-24 Call method, device, electronic equipment and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010110891.1A CN111369995A (en) 2020-02-24 2020-02-24 Call method, device, electronic equipment and medium

Publications (1)

Publication Number Publication Date
CN111369995A true CN111369995A (en) 2020-07-03

Family

ID=71209710

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010110891.1A Pending CN111369995A (en) 2020-02-24 2020-02-24 Call method, device, electronic equipment and medium

Country Status (1)

Country Link
CN (1) CN111369995A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113873078A (en) * 2021-09-28 2021-12-31 维沃移动通信有限公司 Call control method and call control device

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1360303A (en) * 2000-12-22 2002-07-24 王南 Method for carrying out automatic interpretation in telephone communication, creating language library and writing interpretation codes
CN101867632A (en) * 2009-06-12 2010-10-20 刘越 Mobile phone speech instant translation system and method
CN104394265A (en) * 2014-10-31 2015-03-04 小米科技有限责任公司 Automatic session method and device based on mobile intelligent terminal
CN104754536A (en) * 2013-12-27 2015-07-01 中国移动通信集团公司 Method and system for realizing communication between different languages
CN106486125A (en) * 2016-09-29 2017-03-08 安徽声讯信息技术有限公司 A kind of simultaneous interpretation system based on speech recognition technology
CN106791157A (en) * 2017-01-04 2017-05-31 宇龙计算机通信科技(深圳)有限公司 A kind of phonetics transfer method and device
CN107833577A (en) * 2017-11-29 2018-03-23 宋晓成 A kind of voice real-time multilingual speech translation recording method
CN107909994A (en) * 2017-10-31 2018-04-13 珠海市魅族科技有限公司 Communication control method, device, computer installation and computer-readable recording medium
CN207720194U (en) * 2017-01-05 2018-08-10 陈伯妤 A kind of simultaneous interpretation mobile phone

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1360303A (en) * 2000-12-22 2002-07-24 王南 Method for carrying out automatic interpretation in telephone communication, creating language library and writing interpretation codes
CN101867632A (en) * 2009-06-12 2010-10-20 刘越 Mobile phone speech instant translation system and method
CN104754536A (en) * 2013-12-27 2015-07-01 中国移动通信集团公司 Method and system for realizing communication between different languages
CN104394265A (en) * 2014-10-31 2015-03-04 小米科技有限责任公司 Automatic session method and device based on mobile intelligent terminal
CN106486125A (en) * 2016-09-29 2017-03-08 安徽声讯信息技术有限公司 A kind of simultaneous interpretation system based on speech recognition technology
CN106791157A (en) * 2017-01-04 2017-05-31 宇龙计算机通信科技(深圳)有限公司 A kind of phonetics transfer method and device
CN207720194U (en) * 2017-01-05 2018-08-10 陈伯妤 A kind of simultaneous interpretation mobile phone
CN107909994A (en) * 2017-10-31 2018-04-13 珠海市魅族科技有限公司 Communication control method, device, computer installation and computer-readable recording medium
CN107833577A (en) * 2017-11-29 2018-03-23 宋晓成 A kind of voice real-time multilingual speech translation recording method

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113873078A (en) * 2021-09-28 2021-12-31 维沃移动通信有限公司 Call control method and call control device

Similar Documents

Publication Publication Date Title
CN110795236B (en) Method, device, electronic equipment and medium for adjusting capacity of server
CN110636477B (en) Device connection method, device, terminal and storage medium
CN110308956B (en) Application interface display method and device and mobile terminal
CN109151044B (en) Information pushing method and device, electronic equipment and storage medium
CN112965683A (en) Volume adjusting method and device, electronic equipment and medium
CN111031170A (en) Method, apparatus, electronic device and medium for selecting communication mode
CN110109608B (en) Text display method, text display device, text display terminal and storage medium
CN110944374A (en) Communication mode selection method and device, electronic equipment and medium
CN108401194B (en) Time stamp determination method, apparatus and computer-readable storage medium
CN110827830B (en) Voiceprint recognition method, voiceprint recognition device, terminal and storage medium based on voice data
CN110708742B (en) Method, device, electronic equipment and medium for selecting frequency points
CN112118353A (en) Information display method, device, terminal and computer readable storage medium
CN112866470A (en) Incoming call processing method and device, electronic equipment and medium
CN109107163B (en) Analog key detection method and device, computer equipment and storage medium
CN111369995A (en) Call method, device, electronic equipment and medium
CN110908802A (en) Method, device, equipment and storage medium for calling service
CN112992127A (en) Voice recognition method and device
CN111008083A (en) Page communication method and device, electronic equipment and storage medium
CN110941458B (en) Method, device, equipment and storage medium for starting application program
CN110708582B (en) Synchronous playing method, device, electronic equipment and medium
CN113099378A (en) Positioning method, device, equipment and storage medium
CN110933237A (en) Call method, device, electronic equipment and medium
CN111148181A (en) Method, apparatus, electronic device and medium for using 5G communication network
CN110798572A (en) Method, device, electronic equipment and medium for lighting screen
CN111010732A (en) Network registration method, device, electronic equipment and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20200703

RJ01 Rejection of invention patent application after publication