CN111711562A - Message processing method and device, computer storage medium and electronic equipment - Google Patents

Message processing method and device, computer storage medium and electronic equipment Download PDF

Info

Publication number
CN111711562A
CN111711562A CN202010687702.7A CN202010687702A CN111711562A CN 111711562 A CN111711562 A CN 111711562A CN 202010687702 A CN202010687702 A CN 202010687702A CN 111711562 A CN111711562 A CN 111711562A
Authority
CN
China
Prior art keywords
voice
shielding
data
message
voice data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010687702.7A
Other languages
Chinese (zh)
Inventor
罗树浩
赵宇杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Netease Hangzhou Network Co Ltd
Original Assignee
Netease Hangzhou Network Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Netease Hangzhou Network Co Ltd filed Critical Netease Hangzhou Network Co Ltd
Priority to CN202010687702.7A priority Critical patent/CN111711562A/en
Publication of CN111711562A publication Critical patent/CN111711562A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/21Monitoring or handling of messages
    • H04L51/212Monitoring or handling of messages using filtering or selective blocking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/07User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail characterised by the inclusion of specific contents
    • H04L51/10Multimedia information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/07User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail characterised by the inclusion of specific contents
    • H04L51/18Commands or executable codes

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Telephonic Communication Services (AREA)

Abstract

The present disclosure relates to the field of communications, and provides a message processing method, apparatus, computer storage medium and electronic device, which are applied to a message receiving end, where the message receiving end includes a voice processing unit, and the method includes: receiving voice data sent by a server; and the voice processing unit is used for shielding the voice data according to a shielding rule and sending the voice data after shielding to a voice playing interface corresponding to the message receiving end so that the voice playing interface plays the voice data after shielding. The voice shielding method and the voice shielding device can shield voice at the message receiving end, can shield voice differently according to different message receiving ends, and achieve individuation and differentiation of shielding effects.

Description

Message processing method and device, computer storage medium and electronic equipment
Technical Field
The present disclosure relates to the field of communications technologies, and in particular, to a message processing method, a message processing apparatus, a computer-readable storage medium, and an electronic device.
Background
With the development of communication technology, many application software provide a real-time voice system for a user, and the user can perform voice chat with other users through the real-time voice system. In a multi-user real-time voice system, a plurality of users speak and listen through a terminal, and the purpose of real-time communication is achieved. However, in the process of real-time communication, there may be abnormal situations that the speech of some users does not conform to the legal regulation or regulation, such as existence of an expletor, a dirty speech, etc.
In the prior art, the speech content of the user at the speech sending end is processed by using a shielding rule meeting the regulation, and then the processed speech data is uploaded and distributed to the user at the receiving end. However, this method of processing only at the transmitting end cannot implement personalized shielding on one hand, and on the other hand, causes a large system loss at the transmitting end.
In view of the above, there is a need in the art to develop a new message processing method and apparatus.
It is to be noted that the information disclosed in the above background section is only for enhancement of understanding of the background of the present disclosure, and thus may include information that does not constitute prior art known to those of ordinary skill in the art.
Disclosure of Invention
The present disclosure is directed to a message processing method, a message processing apparatus, a computer-readable storage medium, and an electronic device, so as to implement personalized voice screening at least to a certain extent.
Additional features and advantages of the disclosure will be set forth in the detailed description which follows, or in part will be obvious from the description, or may be learned by practice of the disclosure.
According to an aspect of the present disclosure, there is provided a message processing method applied to a message receiving end, the message receiving end including a voice processing unit, the method including: receiving voice data sent by a server; and the voice processing unit is used for shielding the voice data according to a shielding rule and sending the voice data after shielding to a voice playing interface corresponding to the message receiving end so that the voice playing interface plays the voice data after shielding.
In some exemplary embodiments of the present disclosure, the method further comprises: the voice processing unit comprises a main process and at least one sub-process, and the sub-process is started through the main process.
In some exemplary embodiments of the present disclosure, receiving voice data transmitted by a server includes: and sending a transmission instruction to the subprocess through the main process so as to enable the subprocess to receive the voice data sent by the server.
In some exemplary embodiments of the present disclosure, before the voice data is masked by the voice processing unit according to a masking rule, the method further includes: and sending the shielding rule to the subprocess through the main process so that the subprocess shields the voice data according to the shielding rule.
In some exemplary embodiments of the present disclosure, the voice data includes audio data, text data corresponding to the audio data, and a correspondence between the audio data and the text data.
In some exemplary embodiments of the present disclosure, the correspondence relationship includes a one-to-one correspondence between a text sequence number of the text data and a start-stop sequence number of the audio data, where the start-stop sequence number includes a start audio sequence number and an end audio sequence number.
In some exemplary embodiments of the present disclosure, the masking, by the voice processing unit, the voice data according to a masking rule includes: and matching the character data with the shielding rule through the subprocess, and shielding the voice data according to a matching result.
In some exemplary embodiments of the present disclosure, the masking rule includes one or more keywords; matching the character data with the shielding rule through the subprocess, and shielding the voice data according to a matching result, wherein the shielding process comprises the following steps: comparing the character data with the keywords one by one through the subprocess; and if the character data contains the keyword, determining a target character in the character data, and shielding the voice data according to the target character.
In some exemplary embodiments of the present disclosure, the masking the voice data according to the target word includes: acquiring a target character serial number corresponding to the target character, and determining a target starting audio serial number and a target ending audio serial number corresponding to the target character serial number according to the corresponding relation; and acquiring target audio data corresponding to the target starting audio serial number and the target ending audio serial number, and processing the target audio data according to a preset rule.
In some exemplary embodiments of the present disclosure, after the voice data is masked by the voice processing unit according to a masking rule, the method further includes: and sending the voice data after the shielding processing to the main process through the sub-process.
In some exemplary embodiments of the present disclosure, sending the voice data after the shielding processing to a voice playing interface corresponding to the message receiving end includes: and receiving the voice data after shielding processing sent by the plurality of sub-processes through the main process, carrying out sound mixing processing on the voice data after shielding processing, and sending the voice data after sound mixing processing to the voice playing interface.
In some exemplary embodiments of the present disclosure, the voice data is obtained by performing a voice recognition process on the received voice message by the message transmitting end.
In some exemplary embodiments of the present disclosure, the shielding rule is set by the message receiving end in a customized manner.
According to an aspect of the present disclosure, there is provided a message processing apparatus applied to a message receiving end, the message receiving end including a voice processing unit, including: the voice receiving module is used for receiving voice data sent by the server; and the shielding processing module is used for shielding the voice data according to a shielding rule through the voice processing unit and sending the voice data after shielding processing to the voice playing interface corresponding to the message receiving end so that the voice playing interface plays the voice data after shielding processing.
According to an aspect of the present disclosure, there is provided a message processing system, the system comprising: the message sending terminal is used for receiving voice messages, carrying out voice recognition processing on the voice messages to obtain audio data and character data corresponding to the voice messages, and assembling the audio data, the character data and the corresponding relation between the audio data and the character data into voice data; the message receiving end is used for receiving the voice data, shielding the voice data according to a shielding rule through a voice processing unit, and sending the voice data after shielding to a voice playing interface corresponding to the message receiving end; and the server is used for receiving the voice data uploaded by the message sending end and sending the voice data to the message receiving end.
According to an aspect of the present disclosure, there is provided a computer readable medium, on which a computer program is stored, which when executed by a processor, implements the message processing method as described in the above embodiments.
According to an aspect of the present disclosure, there is provided an electronic device including: one or more processors; a storage device for storing one or more programs which, when executed by the one or more processors, cause the one or more processors to implement the message processing method as described in the above embodiments.
As can be seen from the foregoing technical solutions, the message processing method and apparatus, the computer-readable storage medium, and the electronic device in the exemplary embodiments of the present disclosure have at least the following advantages and positive effects:
the message processing method is applied to a message receiving end which comprises a voice processing unit, the message processing method firstly receives voice data sent by a server, then carries out shielding processing on the voice data through the voice processing unit according to shielding rules, and sends the voice data after the shielding processing to a voice playing interface corresponding to the message receiving end so that the voice playing interface plays the voice data after the shielding processing. The message processing method in the disclosure can realize shielding processing on voice data at a message receiving end, realize different shielding effects aiming at different message receiving ends, and realize individuation and difference shielding processing on the voice data; on the other hand, the voice data can be shielded through the voice processing unit, so that the transmission of the voice data in the multi-user voice chat can be controlled instantly and quickly, and the voice chat environment is effectively optimized.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the disclosure.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present disclosure and together with the description, serve to explain the principles of the disclosure. It is to be understood that the drawings in the following description are merely exemplary of the disclosure, and that other drawings may be derived from those drawings by one of ordinary skill in the art without the exercise of inventive faculty.
Fig. 1 schematically shows a flow diagram of a message processing method according to an embodiment of the present disclosure;
fig. 2 schematically shows a detailed flowchart of message processing at a message sending end according to an embodiment of the present disclosure;
fig. 3 schematically shows a detailed flowchart of a message processing method according to an embodiment of the present disclosure;
FIG. 4 schematically shows a block diagram of a message processing apparatus according to an embodiment of the present disclosure;
FIG. 5 schematically shows a block diagram of a message processing system according to an embodiment of the present disclosure;
FIG. 6 schematically shows a block schematic of an electronic device according to an embodiment of the present disclosure;
fig. 7 schematically shows a program product schematic according to an embodiment of the present disclosure.
Detailed Description
Example embodiments will now be described more fully with reference to the accompanying drawings. Example embodiments may, however, be embodied in many different forms and should not be construed as limited to the examples set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the concept of example embodiments to those skilled in the art.
Furthermore, the described features, structures, or characteristics may be combined in any suitable manner in one or more embodiments. In the following description, numerous specific details are provided to give a thorough understanding of embodiments of the disclosure. One skilled in the relevant art will recognize, however, that the subject matter of the present disclosure can be practiced without one or more of the specific details, or with other methods, components, devices, steps, and so forth. In other instances, well-known methods, devices, implementations, or operations have not been shown or described in detail to avoid obscuring aspects of the disclosure.
The block diagrams shown in the figures are functional entities only and do not necessarily correspond to physically separate entities. I.e. these functional entities may be implemented in the form of software, or in one or more hardware modules or integrated circuits, or in different networks and/or processor means and/or microcontroller means.
The flow charts shown in the drawings are merely illustrative and do not necessarily include all of the contents and operations/steps, nor do they necessarily have to be performed in the order described. For example, some operations/steps may be decomposed, and some operations/steps may be combined or partially combined, so that the actual execution sequence may be changed according to the actual situation.
In the related technology in the field, a speech recognition technology is adopted for the speaking content of a user at a sending end, the speech content is converted into text content, then shielding processing is carried out according to a shielding rule, and finally the processed audio data is uploaded and distributed to the user at a receiving end. The voice processing method in the prior art can only be singly processed at a sending end and cannot be individualized according to different receiving ends. For example, an administrator role exists in a chat group, the administrator can access complete conversation content, and a common user can only hear the content after the shielding processing, so that personalized shielding level setting is realized according to different users.
Based on the problems in the related art, a message processing method is proposed in one embodiment of the present disclosure. Fig. 1 shows a flow diagram of a message processing method, which, as shown in fig. 1, at least comprises the following steps:
step S110: receiving voice data sent by a server;
step S120: and the voice processing unit is used for shielding the voice data according to the shielding rule and sending the voice data after shielding processing to a voice playing interface corresponding to the message receiving end so that the voice playing interface plays the voice data after shielding processing.
In the embodiment of the disclosure, on one hand, the message processing method can realize the shielding processing of the voice data at the message receiving end, and realize different shielding effects aiming at different message receiving ends, thereby realizing the individuation and difference shielding processing of the voice data; on the other hand, the voice data can be shielded through the voice processing unit, so that the transmission of the voice data in the multi-user voice chat can be controlled instantly and quickly, and the voice chat environment is effectively optimized.
It should be noted that the message processing method is applied to a message receiving end, and the message receiving end includes a voice processing unit.
In order to make the technical solution of the present disclosure clearer, each step of the voice processing method is explained next.
In step S110, voice data transmitted by the server is received.
In the exemplary embodiment of the present disclosure, the voice data is obtained by uploading the received voice message to the server after the message sending end performs voice recognition processing on the received voice message, and the message receiving end also performs voice recognition processing in a multi-process manner. The message format of the voice message received by the message sending end from the user may include the following format:
|CMD|LEN|DATA|
wherein CMD is a voice message type field and represents the actual meaning of the current voice message; LEN is the message body length of the voice message; DATA is the actual content of the message body of the voice message. By using the format, interactive transmission of voice messages with different lengths can be realized.
For example, in an exemplary embodiment of the present disclosure, fig. 2 shows a specific flowchart of message processing performed at a message sending end, where the message sending end includes a voice recognition unit, as shown in fig. 2, in step S210, a main process of the voice recognition unit is started in real time, and a process executed by the voice recognition unit is a main process at this time; in step S220, the uplink sub-process is started by the main process. Wherein, the uplink subprocess means that the subprocess is used for realizing the transmission of the voice message to the server; in step S230, the main process sends a network connection instruction to the sub-process, so that the sub-process establishes the voice message uplink network connection. The network connection command can be a CMD ═ START command, the network connection is established so as to send a voice message to the server through the subprocess, and when the server receives the voice message, the server feeds back the successful receiving to the sending end through the network connection; in step S240, when the voice of the user is detected, the main process collects a voice message from the microphone and transmits the voice message to the sub-process for processing. Wherein, the transmission type of the message may be a message type of CMD ═ MIC _ DATA type; in step S250, when the sub-process receives the voice message, reading the voice message, and inputting the voice message into the voice recognition module for processing, so as to obtain a recognized voice message; in step S260, after acquiring the recognized voice message, the sub-process records the recognized text data and the corresponding relationship between the text data and the audio data. Wherein, the corresponding relation comprises that the character serial number of the character data is in one-to-one correspondence with the start-stop serial number of the audio data; in step S270, the audio data is encoded, and the encoded audio data, text data, and correspondence relationship are assembled into voice data and sent to the server. The purpose of coding the audio data is that the audio data needs to be coded into proper size suitable for transmission before transmission because the microphone collects PCM audio data which has large memory and is not suitable for network transmission; in step S280, the main process sends a network disconnection instruction to the sub-process to cause the sub-process to disconnect the network and exit the process.
In an exemplary embodiment of the present disclosure, a voice message is subjected to voice recognition processing by a message sending end to form voice data, and the voice data is sent to a server, the server sends the voice data to a corresponding message receiving end, and the message receiving end receives the voice data sent by the server. The voice data comprises audio data, character data corresponding to the audio data and a corresponding relation between the audio data and the character data, the corresponding relation between the audio data and the character data comprises character serial numbers of the character data and start-stop serial numbers of the audio data in one-to-one correspondence, and the start-stop serial numbers comprise start audio serial numbers and end audio serial numbers.
For example, the voice DATA after the voice recognition process is loaded in the DATA field, and the DATA field includes audio DATA, text DATA, and correspondence. Wherein the audio data may comprise the following structure:
|C|len1|audio1|len2|audio2|…|lenC|audioC|
where C is the number of frames of audio data, leniThe length of the audio data of the ith frame is the length of the audio data of each frame, and the length of the audio data after the voice recognition processing is not fixed, so the audio data of each frame needs to be marked; audioiIs the ith frame of audio data.
The textual data may include the following structure:
|N|text1,text2,…,textN|
wherein N is the number of characters in the character data; textiIs the text content of the ith text data.
The correspondence relationship between the audio data and the text data may include the following structure:
|P_start1,P_end1|P_start2,P_end2|…|P_startN,P_endN|
wherein, P _ starti is the initial audio sequence number corresponding to the ith character data; p _ endi is the ending audio sequence number corresponding to the ith text data.
In an exemplary embodiment of the present disclosure, the speech processing unit includes a main process and at least one sub-process, each sub-process being initiated by the main process. The communication mechanism between the main process and the sub-process may be different according to different specific platforms, for example, taking a windows platform as an example, there are various ways that may be implemented, including socket communication, named pipes, shared memory, and the like. After the message receiving end receives the voice data sent by the server, the message receiving end starts a main process of the voice processing unit in real time, and starts a downlink sub-process through the main process, wherein the sub-process can comprise a plurality of sub-processes. The technology of the main process starting subprocess may be different according to different specific platforms, for example, taking a windows platform as an example, the main process can create and run a subprocess by using a system API of CreateProcess, and this disclosure does not specifically limit this.
In an exemplary embodiment of the present disclosure, a transmission instruction is sent to the sub-process by the main process, so that the sub-process receives the voice data sent by the server. After receiving the instruction, the sub-process establishes a network connection for downlink of the voice data in the sub-process, that is, the sub-process receives the voice data sent by the server. The type of the network connection may include TCP, UDP, WebSocket, and the like, which is not specifically limited in this disclosure. According to the voice processing method, the network connection is established in the subprocess, so that on one hand, the occupation of system resources such as a network, a memory and a CPU of the main process can be effectively reduced; and on the other hand, processing voice data in a sub-process is facilitated.
In step S120, the voice processing unit performs a masking process on the voice data according to a masking rule, and sends the voice data after the masking process to a voice playing interface corresponding to the message receiving end, so that the voice playing interface plays the voice data after the masking process.
In an exemplary embodiment of the present disclosure, the shielding rule is set by the message receiving end in a self-defined manner, specifically, the message receiving end may set the number and content of the keywords in a self-defined manner, or set the preset rule in a self-defined manner, which is not specifically limited by the present disclosure.
In an exemplary embodiment of the present disclosure, before the voice data is subjected to the masking processing according to the masking rule by the voice processing unit, the masking rule is transmitted to the subprocess by the main process, so that the subprocess performs the masking processing on the voice data according to the masking rule.
In an exemplary embodiment of the present disclosure, performing a masking process on voice data according to a masking rule by a voice processing unit includes: and matching the character data with the shielding rule through the subprocess, and shielding the voice data according to the matching result.
Wherein the masking rule may include one or more keywords. The method for shielding the voice data according to the shielding rule through the subprocess specifically comprises the following steps: comparing the character data with the keywords one by one through a subprocess; and if the character data contains the keywords, determining target characters in the character data, and shielding the voice data according to the target characters. The keywords included in the text data may be used as target text, and the target text may be one or multiple, which is not specifically limited by the present disclosure.
In an exemplary embodiment of the present disclosure, the shielding processing of the voice data according to the target word includes: acquiring a target character serial number corresponding to a target character, and determining a target starting audio serial number and a target ending audio serial number corresponding to the target character serial number according to the corresponding relation; and acquiring target audio data corresponding to the target starting audio serial number and the target ending audio serial number, and processing the target audio data according to a preset rule.
The preset rule can be set differently by different message receiving ends, and can also be set by the message receiving ends in a user-defined manner. For example, when the user corresponding to the message receiving end is a chat group administrator, the message receiving end is an administrator, the preset rule is that the target audio data is not processed at all, and the voice data is completely sent to the voice playing interface, so that the group administrator can receive all messages sent by the message sending end. For another example, when the user corresponding to the message receiving end is a normal user of a certain chat group, the preset rule may replace the target audio data with other audio data, where the other audio data may be audio data expressed as "beep" sound, or audio data expressed as "creak" sound, and the preset rule may also delete the target audio data or replace the target audio data with blank audio data, and the disclosure does not specifically limit the preset rule.
In an exemplary embodiment of the present disclosure, after the voice data is subjected to the masking process according to the masking rule by the voice processing unit, the voice data after the masking process is transmitted to the main process by the sub-process.
In the exemplary embodiment of the present disclosure, after the message receiving end receives the voice data sent by the plurality of message sending ends, the main process may issue each voice data to the plurality of sub-processes for processing, that is, one sub-process correspondingly processes the voice data sent by one message sending end. And after the plurality of sub-processes process the corresponding voice data, sending the voice data after the shielding processing to the main process. After the voice data after shielding processing sent by the plurality of sub-processes is received through the main process, the voice data after shielding processing is subjected to sound mixing processing, and the voice data after sound mixing processing is sent to the voice playing interface. The audio mixing process is performed on the plurality of masked audio data, and includes integrating the plurality of masked audio data into one audio track.
In an exemplary embodiment of the present disclosure, fig. 3 shows a specific flowchart of the message processing method of the present disclosure, as shown in fig. 3, in step S310, a voice processing main process is started, and a sub-process is started through the main process; in step S320, the main process sends a network connection instruction to the sub-process, so that the sub-process receives the voice data sent by the server; in step S330, the main process sends the masking rule to the sub-process; in step S340, the sub-process receives the voice data sent by the server, and decodes and caches the voice data; in step S350, the subprocess reads the voice data, and performs a masking process on the voice data according to a masking rule; in step S360, the main process receives the plurality of voice data after the shielding processing sent by the plurality of sub-processes, and performs audio mixing processing on the plurality of voice data after the shielding processing; in step S370, the host process sends the mixed voice data to the voice playing interface, so that the voice playing interface plays the mixed voice data; in step S380, the main process of the voice processing unit stops working, and sends a network disconnection instruction to the sub-process, so that the sub-process exits the process.
The following describes embodiments of an apparatus of the present disclosure, which may be used to perform the above-mentioned message processing method of the present disclosure. For details that are not disclosed in the embodiments of the apparatus of the present disclosure, refer to the embodiments of the message processing method described above in the present disclosure.
Fig. 4 schematically shows a block diagram of a message processing apparatus according to one embodiment of the present disclosure.
Referring to fig. 4, a message processing apparatus 400 according to an embodiment of the present disclosure, the voice processing apparatus 400 being configured at a message receiving end, the message receiving end including a voice processing unit, the message processing apparatus 400 including: the voice receiving module 401 and the shielding processing module 402 specifically include:
a voice receiving module 401, configured to receive voice data sent by a server;
the shielding processing module 402 is configured to perform shielding processing on the voice data according to a shielding rule through the voice processing unit, and send the voice data after the shielding processing to a voice playing interface corresponding to the message receiving end, so that the voice playing interface plays the voice data after the shielding processing.
In an exemplary embodiment of the present disclosure, the voice receiving module 401 may further be configured to send a transmission instruction to the sub-process through the main process, so that the sub-process receives the voice data sent by the server, where the voice processing unit includes the main process and at least one sub-process, and each sub-process is started through the main process.
In an exemplary embodiment of the present disclosure, the shielding processing module 402 may be further configured to compare the text data with the keywords one by one through a sub-process; and if the character data comprise the keywords, shielding the voice data according to the character data.
In an exemplary embodiment of the present disclosure, the shielding processing module 402 may be further configured to obtain a text serial number corresponding to the text data, and determine a target starting audio serial number and a target ending audio serial number corresponding to the text serial number according to the corresponding relationship; and acquiring target audio data corresponding to the target starting audio serial number and the target ending audio serial number, and processing the target audio data according to a preset rule.
In an exemplary embodiment of the present disclosure, the message processing apparatus 400 further includes a mixing module (not shown in the figure), configured to receive, by the host process, the voice data after the shielding processing sent by the multiple sub-processes, perform mixing processing on each voice data after the shielding processing, and send the voice data after the mixing processing to the voice playing interface.
The details of each message processing apparatus are already described in detail in the corresponding message processing method, and therefore are not described herein again.
FIG. 5 schematically shows a block diagram of a message processing system according to one embodiment of the present disclosure.
Referring to fig. 5, a message processing system 500 according to one embodiment of the present disclosure, the message processing system 500 includes: the message sending terminal 501, the message receiving terminal 502, and the server terminal 503 specifically include:
a message sending end 501, configured to receive a voice message, perform voice recognition processing on the voice message to obtain audio data and text data corresponding to the voice message, and assemble the audio data, the text data, and a corresponding relationship between the audio data and the text data into voice data;
a message receiving end 502, configured to receive the voice data, perform shielding processing on the voice data according to a shielding rule through a voice processing unit, and send the voice data after the shielding processing to a voice playing interface corresponding to the message receiving end;
the server 503 is configured to receive the voice data uploaded by the message sending end, and send the voice data to the message receiving end.
The details of the above message processing systems have been described in detail in the corresponding message processing methods, and therefore are not described herein again.
It should be noted that although in the above detailed description several modules or units of the apparatus for performing are mentioned, such a division is not mandatory. Indeed, the features and functionality of two or more modules or units described above may be embodied in one module or unit, according to embodiments of the present disclosure. Conversely, the features and functions of one module or unit described above may be further divided into embodiments by a plurality of modules or units.
In an exemplary embodiment of the present disclosure, an electronic device capable of implementing the above method is also provided.
As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or program product. Thus, various aspects of the invention may be embodied in the form of: an entirely hardware embodiment, an entirely software embodiment (including firmware, microcode, etc.) or an embodiment combining hardware and software aspects that may all generally be referred to herein as a "circuit," module "or" system.
An electronic device 600 according to this embodiment of the invention is described below with reference to fig. 6. The electronic device 600 shown in fig. 6 is only an example, and should not bring any limitation to the functions and the scope of use of the embodiments of the present invention.
As shown in fig. 6, the electronic device 600 is embodied in the form of a general purpose computing device. The components of the electronic device 600 may include, but are not limited to: the at least one processing unit 610, the at least one memory unit 620, a bus 630 connecting different system components (including the memory unit 620 and the processing unit 610), and a display unit 640.
Wherein the storage unit stores program code that is executable by the processing unit 610 to cause the processing unit 610 to perform steps according to various exemplary embodiments of the present invention as described in the above section "exemplary methods" of the present specification. For example, the processing unit 610 may execute step S110 shown in fig. 1, receiving voice data transmitted by a server; and step S120, shielding the voice data according to the shielding rule through the voice processing unit, and sending the voice data after shielding to a voice playing interface corresponding to the message receiving end, so that the voice playing interface plays the voice data after shielding.
The storage unit 620 may include readable media in the form of volatile memory units, such as a random access memory unit (RAM)6201 and/or a cache memory unit 6202, and may further include a read-only memory unit (ROM) 6203.
The memory unit 620 may also include a program/utility 6204 having a set (at least one) of program modules 6205, such program modules 6205 including, but not limited to: an operating system, one or more application programs, other program modules, and program data, each of which, or some combination thereof, may comprise an implementation of a network environment.
Bus 630 may be one or more of several types of bus structures, including a memory unit bus or memory unit controller, a peripheral bus, an accelerated graphics port, a processing unit, or a local bus using any of a variety of bus architectures.
The electronic device 600 may also communicate with one or more external devices 800 (e.g., keyboard, pointing device, bluetooth device, etc.), with one or more devices that enable a viewer to interact with the electronic device 600, and/or with any devices (e.g., router, modem, etc.) that enable the electronic device 600 to communicate with one or more other computing devices. Such communication may occur via an input/output (I/O) interface 650. Also, the electronic device 600 may communicate with one or more networks (e.g., a Local Area Network (LAN), a Wide Area Network (WAN), and/or a public network such as the Internet) via the network adapter 660. As shown, the network adapter 660 communicates with the other modules of the electronic device 600 over the bus 630. It should be appreciated that although not shown in the figures, other hardware and/or software modules may be used in conjunction with the electronic device 600, including but not limited to: microcode, device drivers, redundant processing units, external disk drive arrays, RAID systems, tape drives, and data backup storage systems, among others.
Through the above description of the embodiments, those skilled in the art will readily understand that the exemplary embodiments described herein may be implemented by software, or by software in combination with necessary hardware. Therefore, the technical solution according to the embodiments of the present disclosure may be embodied in the form of a software product, which may be stored in a non-volatile storage medium (which may be a CD-ROM, a usb disk, a removable hard disk, etc.) or on a network, and includes several instructions to enable a computing device (which may be a personal computer, a server, a terminal device, or a network device, etc.) to execute the method according to the embodiments of the present disclosure.
In an exemplary embodiment of the present disclosure, there is also provided a computer-readable storage medium having stored thereon a program product capable of implementing the above-described method of the present specification. In some possible embodiments, aspects of the invention may also be implemented in the form of a program product comprising program code means for causing a terminal device to carry out the steps according to various exemplary embodiments of the invention described in the above section "exemplary methods" of the present description, when said program product is run on the terminal device.
Referring to fig. 7, a program product 700 for implementing the above method according to an embodiment of the present invention is described, which may employ a portable compact disc read only memory (CD-ROM) and include program code, and may be run on a terminal device, such as a personal computer. However, the program product of the present invention is not limited in this regard and, in the present document, a readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
The program product may employ any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. A readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples (a non-exhaustive list) of the readable storage medium include: an electrical connection having one or more wires, a portable disk, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
A computer readable signal medium may include a propagated data signal with readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A readable signal medium may also be any readable medium that is not a readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on a readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, C + + or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computing device, partly on the user's device, as a stand-alone software package, partly on the user's computing device and partly on a remote computing device, or entirely on the remote computing device or server. In the case of a remote computing device, the remote computing device may be connected to the user computing device through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computing device (e.g., through the internet using an internet service provider).
Furthermore, the above-described figures are merely schematic illustrations of processes involved in methods according to exemplary embodiments of the invention, and are not intended to be limiting. It will be readily understood that the processes shown in the above figures are not intended to indicate or limit the chronological order of the processes. In addition, it is also readily understood that these processes may be performed synchronously or asynchronously, e.g., in multiple modules.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure disclosed herein. This application is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.
It will be understood that the present disclosure is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the present disclosure is to be limited only by the terms of the appended claims.

Claims (17)

1. A message processing method applied to a message receiving end, wherein the message receiving end includes a voice processing unit, and the method includes:
receiving voice data sent by a server;
and the voice processing unit is used for shielding the voice data according to a shielding rule and sending the voice data after shielding to a voice playing interface corresponding to the message receiving end so that the voice playing interface plays the voice data after shielding.
2. The message processing method of claim 1, wherein the method further comprises:
the voice processing unit comprises a main process and at least one subprocess, and each subprocess is started through the main process.
3. The message processing method of claim 2, wherein receiving the voice data sent by the server comprises:
and sending a transmission instruction to the subprocess through the main process so as to enable the subprocess to receive the voice data sent by the server.
4. The message processing method according to claim 2, wherein before the voice data is masked by the voice processing unit according to a masking rule, the method further comprises:
and sending the shielding rule to the subprocess through the main process so that the subprocess shields the voice data according to the shielding rule.
5. The message processing method according to claim 2, wherein the voice data includes audio data, text data corresponding to the audio data, and a correspondence between the audio data and the text data.
6. The message processing method according to claim 5, wherein the correspondence relationship comprises a one-to-one correspondence between a text sequence number of the text data and a start-stop sequence number of the audio data, and the start-stop sequence number comprises a start audio sequence number and an end audio sequence number.
7. The message processing method according to claim 6, wherein the masking the voice data according to the masking rule by the voice processing unit comprises:
and matching the character data with the shielding rule through the subprocess, and shielding the voice data according to a matching result.
8. The message processing method of claim 7, wherein the masking rule comprises one or more keywords;
matching the character data with the shielding rule through the subprocess, and shielding the voice data according to a matching result, wherein the shielding process comprises the following steps:
comparing the character data with the keywords one by one through the subprocess;
and if the character data contains the keyword, determining a target character in the character data, and shielding the voice data according to the target character.
9. The message processing method of claim 8, wherein the masking the voice data according to the target text comprises:
acquiring a target character serial number corresponding to the target character, and determining a target starting audio serial number and a target ending audio serial number corresponding to the target character serial number according to the corresponding relation;
and acquiring target audio data corresponding to the target starting audio serial number and the target ending audio serial number, and processing the target audio data according to a preset rule.
10. The message processing method according to claim 2, wherein after the voice data is masked by the voice processing unit according to a masking rule, the method further comprises:
and sending the voice data after the shielding processing to the main process through the sub-process.
11. The message processing method according to claim 10, wherein sending the voice data after the masking processing to a voice playing interface corresponding to the message receiving end comprises:
and receiving the voice data after shielding processing sent by the plurality of sub-processes through the main process, carrying out sound mixing processing on the voice data after shielding processing, and sending the voice data after sound mixing processing to the voice playing interface.
12. The message processing method according to claim 1, wherein the voice data is obtained by performing voice recognition processing on the received voice message by the message sending end.
13. The message processing method according to claim 1, wherein the masking rule is set by the message receiving end in a self-defined manner.
14. A message processing apparatus, applied to a message receiving end, the message receiving end including a voice processing unit, comprising:
the voice receiving module is used for receiving voice data sent by the server;
and the shielding processing module is used for shielding the voice data according to a shielding rule through the voice processing unit and sending the voice data after shielding processing to the voice playing interface corresponding to the message receiving end so that the voice playing interface plays the voice data after shielding processing.
15. A message processing system, the system comprising:
the message sending terminal is used for receiving voice messages, carrying out voice recognition processing on the voice messages to obtain audio data and character data corresponding to the voice messages, and assembling the audio data, the character data and the corresponding relation between the audio data and the character data into voice data;
the message receiving end is used for receiving the voice data, shielding the voice data according to a shielding rule through a voice processing unit, and sending the voice data after shielding to a voice playing interface corresponding to the message receiving end;
and the server is used for receiving the voice data uploaded by the message sending end and sending the voice data to the message receiving end.
16. A computer-readable storage medium, on which a computer program is stored, which, when being executed by a processor, carries out the message processing method according to any one of claims 1 to 13.
17. An electronic device, comprising:
one or more processors;
storage means for storing one or more programs which, when executed by the one or more processors, cause the one or more processors to carry out a message processing method according to any one of claims 1 to 13.
CN202010687702.7A 2020-07-16 2020-07-16 Message processing method and device, computer storage medium and electronic equipment Pending CN111711562A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010687702.7A CN111711562A (en) 2020-07-16 2020-07-16 Message processing method and device, computer storage medium and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010687702.7A CN111711562A (en) 2020-07-16 2020-07-16 Message processing method and device, computer storage medium and electronic equipment

Publications (1)

Publication Number Publication Date
CN111711562A true CN111711562A (en) 2020-09-25

Family

ID=72546545

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010687702.7A Pending CN111711562A (en) 2020-07-16 2020-07-16 Message processing method and device, computer storage medium and electronic equipment

Country Status (1)

Country Link
CN (1) CN111711562A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112860213A (en) * 2021-03-09 2021-05-28 腾讯科技(深圳)有限公司 Audio processing method, storage medium and electronic equipment

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN201118703Y (en) * 2007-06-19 2008-09-17 华为技术有限公司 Device for filtering the information sent or received by the communication terminal
CN102148848A (en) * 2010-02-10 2011-08-10 中兴通讯股份有限公司 Data management method and system
CN103516915A (en) * 2012-06-27 2014-01-15 百度在线网络技术(北京)有限公司 Method, system and device for replacing sensitive words in call process of mobile terminal
US20150350148A1 (en) * 2014-06-02 2015-12-03 pMDsoft, Inc. Healthcare secure messaging and reminder system
CN105681212A (en) * 2016-03-21 2016-06-15 广州盈可视电子科技有限公司 Video playing state monitoring method and device
US20160379639A1 (en) * 2015-06-29 2016-12-29 Google Inc. Privacy-preserving training corpus selection
CN107911361A (en) * 2017-11-14 2018-04-13 网易(杭州)网络有限公司 Support voice management method, apparatus, terminal device and the storage medium of more sessions
CN110413386A (en) * 2019-06-27 2019-11-05 深圳市富途网络科技有限公司 Multiprocessing method, apparatus, terminal device and computer readable storage medium
CN110534113A (en) * 2019-08-26 2019-12-03 深圳追一科技有限公司 Audio data desensitization method, device, equipment and storage medium

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN201118703Y (en) * 2007-06-19 2008-09-17 华为技术有限公司 Device for filtering the information sent or received by the communication terminal
CN102148848A (en) * 2010-02-10 2011-08-10 中兴通讯股份有限公司 Data management method and system
CN103516915A (en) * 2012-06-27 2014-01-15 百度在线网络技术(北京)有限公司 Method, system and device for replacing sensitive words in call process of mobile terminal
US20150350148A1 (en) * 2014-06-02 2015-12-03 pMDsoft, Inc. Healthcare secure messaging and reminder system
US20160379639A1 (en) * 2015-06-29 2016-12-29 Google Inc. Privacy-preserving training corpus selection
CN105681212A (en) * 2016-03-21 2016-06-15 广州盈可视电子科技有限公司 Video playing state monitoring method and device
CN107911361A (en) * 2017-11-14 2018-04-13 网易(杭州)网络有限公司 Support voice management method, apparatus, terminal device and the storage medium of more sessions
CN110413386A (en) * 2019-06-27 2019-11-05 深圳市富途网络科技有限公司 Multiprocessing method, apparatus, terminal device and computer readable storage medium
CN110534113A (en) * 2019-08-26 2019-12-03 深圳追一科技有限公司 Audio data desensitization method, device, equipment and storage medium

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112860213A (en) * 2021-03-09 2021-05-28 腾讯科技(深圳)有限公司 Audio processing method, storage medium and electronic equipment
CN112860213B (en) * 2021-03-09 2023-08-25 腾讯科技(深圳)有限公司 Audio processing method and device, storage medium and electronic equipment

Similar Documents

Publication Publication Date Title
US10425365B2 (en) System and method for relaying messages
EP3084633B1 (en) Attribute-based audio channel arbitration
CN106487757A (en) Carry out method, conference client and the system of voice conferencing
CN110138654B (en) Method and apparatus for processing speech
CN110545236A (en) Instant messaging message modification method and device, storage medium and electronic equipment
CN110290344B (en) Online translation method, system, equipment and storage medium based on teleconference
JP2022137187A (en) Use of structured audio output to detect playback and/or adapt to inconsistent playback in wireless speaker
US20170286755A1 (en) Facebot
CN110223694B (en) Voice processing method, system and device
CN111462726B (en) Method, device, equipment and medium for answering out call
CN113488048B (en) Information interaction method and device
CN112364144A (en) Interaction method, device, equipment and computer readable medium
CN111711562A (en) Message processing method and device, computer storage medium and electronic equipment
CN110740212B (en) Call answering method and device based on intelligent voice technology and electronic equipment
US11184477B2 (en) Gapless audio communication via discourse gap recovery model
CN107608718B (en) Information processing method and device
US20220206884A1 (en) Systems and methods for conducting an automated dialogue
CN113852835A (en) Live broadcast audio processing method and device, electronic equipment and storage medium
CN112965593A (en) AI algorithm-based method and device for realizing multi-mode control digital human interaction
KR102349620B1 (en) Method and apparatus for providing interpretation and translation services using user terminal
KR20190098398A (en) Method for providing image communicaiton service
KR20180131098A (en) Method and apparatus for outputting a message
CN114582339A (en) Voice interaction method and device, electronic equipment and medium
US20230005202A1 (en) Speech image providing method and computing device for performing the same
CN108874354B (en) Game platform and audio processing method thereof

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200925