CN112735375A - Voice broadcasting method, device, equipment and storage medium - Google Patents
Voice broadcasting method, device, equipment and storage medium Download PDFInfo
- Publication number
- CN112735375A CN112735375A CN202011563561.4A CN202011563561A CN112735375A CN 112735375 A CN112735375 A CN 112735375A CN 202011563561 A CN202011563561 A CN 202011563561A CN 112735375 A CN112735375 A CN 112735375A
- Authority
- CN
- China
- Prior art keywords
- user
- comment information
- information
- sound parameters
- comment
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 62
- 230000004044 response Effects 0.000 claims description 13
- 238000004590 computer program Methods 0.000 claims description 11
- 238000004458 analytical method Methods 0.000 claims description 2
- 238000013473 artificial intelligence Methods 0.000 abstract description 4
- 230000015572 biosynthetic process Effects 0.000 abstract description 2
- 238000003786 synthesis reaction Methods 0.000 abstract description 2
- 238000004891 communication Methods 0.000 description 8
- 238000010586 diagram Methods 0.000 description 8
- 238000012545 processing Methods 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 4
- 230000006870 function Effects 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 230000003993 interaction Effects 0.000 description 2
- 238000003491 array Methods 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000001953 sensory effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
- G10L13/02—Methods for producing synthetic speech; Speech synthesisers
- G10L13/04—Details of speech synthesis systems, e.g. synthesiser structure or memory management
- G10L13/047—Architecture of speech synthesisers
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
The application discloses a voice broadcasting method, a voice broadcasting device, voice broadcasting equipment and a storage medium, and relates to the technical field of artificial intelligence, in particular to the technical field of voice synthesis. One embodiment of the method comprises: obtaining comment information of a current browsing page of a user; configuring corresponding sound parameters for the comment information based on the user information corresponding to the comment information; and converting the comment information into voice, and broadcasting the voice by using the sound parameters corresponding to the comment information. The embodiment provides a voice broadcasting method, which presents the comment content in a dialogue mode, so that the comment content is easier to understand for the user, and the user experience when the user browses the comment content is enriched.
Description
Technical Field
The embodiment of the application relates to the field of computers, in particular to the field of artificial intelligence such as voice synthesis, and particularly relates to a voice broadcasting method, device, equipment and storage medium.
Background
In recent years, with the progress and application of technologies such as big data and artificial intelligence, a text-to-speech (TTS) technology also tends to be mature, and a basic text-to-speech technology also extends to different applications and can be personalized and optimized, such as application scenes in audio reading, news broadcasting, order broadcasting, webpage broadcasting and some intelligent hardware; meanwhile, with the technical development of text-to-speech, technical parameters such as speech rate, volume, pitch and the like can be set. It can be seen that the application scenarios of text and speech based technologies are welcomed better.
Disclosure of Invention
The embodiment of the application provides a voice broadcasting method, a voice broadcasting device, voice broadcasting equipment and a storage medium.
In a first aspect, an embodiment of the present application provides a voice broadcast method, including: obtaining comment information of a current browsing page of a user; configuring corresponding sound parameters for the comment information based on the user information corresponding to the comment information; and converting the comment information into voice, and broadcasting the voice by using the sound parameters corresponding to the comment information.
In a second aspect, an embodiment of the present application provides a voice broadcast device, including: the acquisition module is configured to acquire comment information of a current browsing page of a user; the configuration module is configured to configure corresponding sound parameters for the comment information based on the user information corresponding to the comment information; the broadcasting module is configured to convert the comment information into voice, and the voice is broadcasted according to the sound parameters corresponding to the comment information.
In a third aspect, an embodiment of the present application provides an electronic device, including: at least one processor; and a memory communicatively coupled to the at least one processor; wherein the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method as described in any one of the implementations of the first aspect.
In a fourth aspect, embodiments of the present application propose a non-transitory computer-readable storage medium storing computer instructions for causing a computer to perform the method as described in any one of the implementations of the first aspect.
In a fifth aspect, the present application provides a computer program product, which includes a computer program that, when executed by a processor, implements the method as described in any implementation manner of the first aspect.
According to the voice broadcasting method, the voice broadcasting device, the voice broadcasting equipment and the voice broadcasting storage medium, firstly, comment information of a current browsing page of a user is obtained; then configuring corresponding sound parameters for the comment information based on the user information corresponding to the comment information; and finally, converting the comment information into voice, and broadcasting the voice by using the sound parameters corresponding to the comment information. The application provides a voice broadcasting method, which presents the comment content in a dialogue mode, so that a user can understand the comment content more easily, and the user experience of the user in browsing the comment content is enriched.
It should be understood that the statements in this section do not necessarily identify key or critical features of the embodiments of the present disclosure, nor do they limit the scope of the present disclosure. Other features of the present disclosure will become apparent from the following description.
Drawings
Other features, objects and advantages of the present application will become more apparent upon reading of the following detailed description of non-limiting embodiments thereof, made with reference to the accompanying drawings. The drawings are included to provide a better understanding of the present solution and are not intended to limit the present application. Wherein:
fig. 1 is a flowchart of one embodiment of a voice broadcast method according to the present application;
fig. 2 is a flowchart of another embodiment of a voice broadcast method according to the present application;
fig. 3 is a flowchart of yet another embodiment of a voice announcement method according to the present application;
fig. 4 is a block diagram of an implementation of the voice broadcast method of the present application;
fig. 5 is a block diagram of another implementation of the voice broadcasting method of the present application;
fig. 6 is a schematic structural diagram of an embodiment of a voice broadcast device according to the present application;
fig. 7 is a block diagram of an electronic device for implementing a voice broadcast method according to an embodiment of the present application.
Detailed Description
The following description of the exemplary embodiments of the present application, taken in conjunction with the accompanying drawings, includes various details of the embodiments of the application for the understanding of the same, which are to be considered exemplary only. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the present application. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
It should be noted that the embodiments and features of the embodiments in the present application may be combined with each other without conflict. The present application will be described in detail below with reference to the embodiments with reference to the attached drawings.
Fig. 1 shows a flow 100 of one embodiment of a voice broadcast method according to the present application. The voice broadcasting method comprises the following steps:
In this embodiment, the execution subject of the voice broadcast method may obtain comment information of a currently browsed page of a user. Specifically, the execution subject may obtain comment information of the currently browsed page from a comment server in the cloud. And the comment server at the cloud end can store comment information corresponding to each page, and when the execution main body detects a user browsing page, the comment information corresponding to the current browsing page can be acquired from the comment server.
And 102, configuring corresponding sound parameters for the comment information based on the user information corresponding to the comment information.
In this embodiment, the execution subject may configure corresponding sound parameters for the comment information based on the user information corresponding to the comment information. Each piece of comment information has corresponding user information, the user information may include a user ID (Identity document), the user ID indicates the user Identity that issued the piece of comment information, each user corresponds to one user ID, and the user IDs of different users are different from each other. The user information may also include attribute information of the user, such as a geographical location where the user is located, an age of the user, or gender information of the user. When the user information comprises the user attribute information, the sound parameters can be adjusted and optimized according to the attribute information of the user. The execution main body may configure corresponding sound parameters for each piece of comment information based on the user ID and/or the attribute information of the user corresponding to the comment information.
In some optional implementation manners of the present embodiment, the execution subject may configure, for the comment information, a corresponding sound parameter based on the user information corresponding to the comment information and the time information of the comment information. Specifically, after obtaining the comment information of the current browsing page of the user, the execution main body analyzes the obtained browsing information to obtain time information of each piece of comment information and user information corresponding to each piece of comment information. When corresponding sound parameters are configured for the comment information, the sound parameters are configured for the comment information according to the time sequence of the comment information, when the number of users is small, the comment information corresponding to each user is configured to be the same sound parameter, and the sound parameters of the users are different; when the number of the users is too large, the multiple pieces of comment information of the users who issue the multiple pieces of comment information are preferentially configured into the same sound parameter, and then corresponding sound parameters are configured for comments of other users according to actual conditions.
And step 103, converting the comment information into voice, and broadcasting the voice by using the sound parameters corresponding to the comment information.
In this embodiment, the execution subject may convert the comment information into a voice, and broadcast the converted voice with a sound parameter corresponding to the comment information. In step 102, corresponding sound parameters are configured for each piece of comment information, so that after the comment information is converted into a voice, the voice is broadcasted for the configured sound parameters. Different comment information is broadcasted by using different sound parameters, so that a user can better understand the content of the comment information and establish a complete context, for example, when the user is a blind person, the comment information is broadcasted by using different sound parameters, the user can understand the content of the comment information more easily, but if the same sound parameter is used for broadcasting all the comment information, the user can not distinguish different publishing users of different comments, and the content of the comment information can not be understood.
The voice broadcasting method provided by the embodiment of the application comprises the steps of firstly obtaining comment information of a current browsing page of a user; then configuring corresponding sound parameters for the comment information based on the user information corresponding to the comment information; and finally, converting the comment information into voice, and broadcasting the voice by using the sound parameters corresponding to the comment information. The application provides a method for carrying out voice broadcast on comment contents by different sound parameters through combining context contents and user attributes, the comment contents are presented in a conversation mode, and a user can distinguish different publishers of different comment information through sound, so that complete context contents are established, the comment information is better understood, and user experience of the user when browsing the comment contents is enriched.
With continued reference to fig. 2, fig. 2 shows a flow 200 of another embodiment of a voice announcement method according to the present application. The voice broadcasting method comprises the following steps:
In this embodiment, the execution subject of the voice broadcast method may obtain comment information of a currently browsed page of a user from the server.
In this embodiment, the execution subject may analyze the comment information acquired in step 201 to obtain time information of each piece of comment information, and user information corresponding to each piece of comment information, where the user information includes a user ID. The user ID is used for representing the identity of the user, one user corresponds to one user ID, each piece of comment information corresponds to one user ID, one user ID can issue a plurality of pieces of comment information, namely when the user IDs of a plurality of comments are the same, the comments are issued by the same user. The time information indicates the specific time of the comment information. Each piece of comment information corresponds to one piece of time information and one corresponding user ID.
In some optional implementations of this embodiment, the user information may further include attribute information of the user, and the attribute information of the user may include a geographic location where the user is located, an age of the user, or a gender of the user. When the user information includes the attribute information of the user, the execution main body can adjust and optimize the sound parameters based on the attribute information of the user, so that better user experience is brought to the user.
And step 203, configuring corresponding sound parameters for the comment information based on the time information of the comment information and the user ID corresponding to each comment information.
In this embodiment, the execution subject may configure corresponding sound parameters for the comment information based on the time information of the comment information and the user ID corresponding to each comment information. And sequentially configuring sound parameters for each comment according to the time sequence of comment publishing, wherein a plurality of comments of the user ID publishing the plurality of comments are configured to be the same sound parameter, and the sound parameters of N continuous comment information are ensured to be different.
In some optional implementation manners of this embodiment, the execution subject may configure, based on the user attribute information corresponding to the comment information, a corresponding sound parameter for the comment information. When the user information includes the geographic location of the user, the age of the user, or the gender of the user, the execution subject may adjust and optimize the sound parameters based on the attribute information, thereby providing a better browsing experience for the user. For example, when the attribute information of the user includes that the geographical location of the user is Harbin and the gender of the user is male, the sound parameter may be adjusted to a male sound in the northeast dialect.
And step 204, converting the comment information into voice, and broadcasting the voice by using the sound parameters corresponding to the comment information.
In this embodiment, the execution subject may convert the comment information into a voice, and broadcast the voice using the sound parameters configured for the comment information in step 203.
The voice broadcasting method provided by the embodiment of the application comprises the steps of firstly obtaining comment information of a current browsing page of a user; then, analyzing the comment information to obtain time information of the comment information and user information corresponding to the comment information, wherein the user information comprises a user ID; configuring corresponding sound parameters for the comment information based on the time information of the comment information and the user ID corresponding to each comment information; and finally, converting the comment information into voice, and broadcasting the voice by using the sound parameters corresponding to the comment information. According to the voice broadcasting method provided by the embodiment of the application, the voice parameters are configured for the plurality of pieces of comment information in the page based on the user ID in the user information corresponding to the comment information, the comment content is presented in a conversation mode, and a user can distinguish different publishers of different comment information through voice, so that complete context content is established, the comment information is better understood, and user experience when the user browses the comment content is enriched.
With continued reference to fig. 3, fig. 3 shows a flow 300 of yet another embodiment of a voice announcement method according to the present application. The voice broadcasting method comprises the following steps:
In this embodiment, the execution subject of the voice broadcast method may obtain comment information of a currently browsed page of a user from the server.
In this embodiment, the execution subject may analyze the obtained comment information, so as to obtain time information of posting of each piece of comment information and a user ID of posting of the piece of comment information.
In this embodiment, the execution subject counts the number of comment information corresponding to each user ID and the number of different user IDs. Since in the present embodiment, each user can issue one or more comment messages, there may be one or more comment messages corresponding to each user ID. When configuring sound parameters for comment information, it is necessary to ensure that a plurality of pieces of comment information issued by the same user use the same sound parameters as much as possible, and then it is necessary to determine the number of maximum sound parameters based on the number of different user IDs. Therefore, in this step, the number of comment information and the number of different user IDs corresponding to each user need to be counted.
In this embodiment, the execution body may determine the number of the maximum sound parameters based on the number of the counted different user IDs and the threshold number of the sound parameters set in advance. The threshold number of the sound parameters is preset according to the performance and configuration of the system, so that the threshold number of the sound parameters of different systems may be different, and the application does not limit the threshold number. After counting the number of different user IDs, comparing the number with the preset threshold number of sound parameters, thereby determining the number of the maximum sound parameters.
In some optional implementations of the embodiment, in response to the number of different user IDs not being greater than the preset threshold number of sound parameters, the number of maximum sound parameters is set to the number of different user IDs. That is, when the number of different user IDs is less than or equal to the threshold number of preset sound parameters, the number of the maximum sound parameters is set as the number of different user IDs. For example, when the number of different user IDs is counted as 90 and the threshold number of preset sound parameters is 100, the number of the maximum sound parameters is set to 90 at this time. On the premise that the system has the best performance, a plurality of pieces of comment information are better presented in a dialogue mode, and therefore user experience is improved.
In some optional implementations of the embodiment, in response to the number of different user IDs being greater than the threshold number of preset sound parameters, the number of maximum sound parameters is set to the threshold number of preset sound parameters. That is, when the number of different user IDs is greater than the threshold number of preset sound parameters, the number of the maximum sound parameters is set to the threshold number of the preset sound parameters. For example, when the number of different user IDs is counted as 150 and the threshold number of preset sound parameters is 100, the number of the maximum sound parameters is set as 100 at this time. On the premise that the system has the best performance, a plurality of pieces of comment information are better presented in a dialogue mode, and therefore user experience is improved.
And 305, configuring corresponding sound parameters for the comment information based on the user ID corresponding to the comment information.
In this embodiment, the execution subject may configure corresponding sound parameters for the comment information according to the user ID of the comment information.
In some optional implementations of the present embodiment, in response to that the plurality of pieces of comment information correspond to the same user ID, the same sound parameter is configured for the plurality of pieces of comment information corresponding to the same user ID. The same sound parameter is configured for the plurality of pieces of comment information of the user who issues the plurality of pieces of comment information, so that the user can obtain continuous context content, and the content of the comment information can be better understood.
In some optional implementations of the present embodiment, in response to a predetermined number of consecutive pieces of comment information corresponding to a plurality of user IDs, a different sound parameter is configured for each of the plurality of user IDs. In the case that the number of different user IDs is greater than the maximum number of sound parameters, one sound parameter cannot be configured for each user ID, and in this case, when configuring sound parameters for comment information, it is to be ensured that N consecutive comments use different sound parameters, where N is a positive integer, and N can be set according to actual situations, which is not limited in this application. The method effectively avoids the situation that the user cannot obtain continuous context according to the broadcasted voice when the continuous N voices are broadcasted.
And step 306, converting the comment information into voice, and broadcasting the voice by using the sound parameters corresponding to the comment information.
In this embodiment, the execution subject may convert the comment information into a voice, and broadcast the voice using the sound parameters configured for the comment information in the above step.
The voice broadcasting method provided by the embodiment of the application comprises the steps of firstly obtaining comment information of a current browsing page of a user; analyzing the comment information to obtain time information of the comment information and a user ID corresponding to the comment information; then, counting the number of comment information corresponding to each user ID and the number of different user IDs; determining the number of the maximum sound parameters based on the number of different user IDs and the preset threshold number of the sound parameters; configuring corresponding sound parameters for the comment information based on the user ID corresponding to the comment information; and finally, converting the comment information into voice, and broadcasting the voice by using the sound parameters corresponding to the comment information. According to the voice broadcasting method provided by the embodiment of the application, on the premise that the system has the best performance, the voice parameters are configured for the plurality of pieces of comment information in the page, the comment content is presented in a conversation mode, and a user can distinguish different publishers of different comment information through voice, so that a complete context is established, the comment information is better understood, and the user experience of the user when browsing the comment content is enriched.
With continued reference to fig. 4, fig. 4 is a block diagram of an implementation of the voice broadcasting method of the present application. As shown in fig. 4, the obtained comment information of the current browsing page of the user is:
(1) the comment content of the user a is as follows: 1;
(2) the comment content of the user b is as follows: 2;
(3) the comment content of the user a is as follows: 3;
(4) the comment content of the user a is as follows: 4;
(5) the comment content of the user c is: 5;
(6) the comment content of the user d is: 6;
(7) the comment content of the user e is: 7;
(8) the comment content of the user f is: 8;
(9) the comment content of the user b is as follows: 9.
it can be seen that the above 9 pieces of comment content are published by 6 users, namely, user a, user b, user c, user d, user e, and user f, and that user a publishes 3 pieces of comments, user b publishes 2 pieces of comments, and user c, user d, user e, and user f all publish one piece of comment.
In this embodiment, assuming that the preset threshold number of the sound parameters is 10, the number of the maximum sound parameters at this time is 6, and the 6 sound parameters are respectively a sound parameter a, a sound parameter b, a sound parameter c, a sound parameter d, a sound parameter e, and a sound parameter f, and different sound parameters can be configured for the 6 different users to broadcast the comment information, where the broadcast scheme is as follows:
broadcasting the comment of the user a by using the sound parameter a, and broadcasting 1;
broadcasting the comments of the user b by using the sound parameters b, and broadcasting 2;
broadcasting the comments of the user a by using the sound parameter a, and broadcasting 3;
broadcasting the comments of the user a by using the sound parameter a, and broadcasting 4;
broadcasting the comments of the user c by using the sound parameters c, and broadcasting 5;
broadcasting the comments of the user d by using the sound parameters d, and broadcasting 6;
broadcasting the comments of the user e by using the sound parameter e, and broadcasting 7;
broadcasting the comments of the user f by using the sound parameters f 8;
and broadcasting the comment of the user b by using the sound parameter b, and broadcasting 9.
When the number of different users is not larger than the preset threshold number of the sound parameters, one sound parameter can be configured for each user, so that different sound parameters are used for broadcasting the comment contents issued by different users, but a plurality of comments of the same user are broadcasted by using the same sound parameter.
With continued reference to fig. 5, fig. 5 is another implementation block diagram of the voice broadcasting method of the present application. As shown in fig. 5, the obtained comment information of the current browsing page of the user is:
(1) the comment content of the user a is as follows: 1;
(2) the comment content of the user b is as follows: 2;
(3) the comment content of the user a is as follows: 3;
(4) the comment content of the user a is as follows: 4;
(5) the comment content of the user c is: 5;
(6) the comment content of the user d is: 6;
(7) the comment content of the user e is: 7;
(8) the comment content of the user f is: 8;
(9) the comment content of the user b is as follows: 9.
the comment content acquired in fig. 5 is the same as that in fig. 4, and is also 9 comments posted by 6 users, namely, user a, user b, user c, user d, user e, and user f, where user a posts 3 comments, user b posts 2 comments, and user c, user d, user e, and user f all posts one comment.
However, in this embodiment, assuming that the preset threshold number of the sound parameters is 5, the number of the maximum sound parameters is 5, the 5 sound parameters are the sound parameter a, the sound parameter b, the sound parameter c, the sound parameter d, and the sound parameter e, respectively, and different sound parameters cannot be configured for the 6 different users, and the scheme for configuring the sound parameters for the 6 different users is as follows:
broadcasting the comment of the user a by using the sound parameter a, and broadcasting 1;
broadcasting the comments of the user b by using the sound parameters b, and broadcasting 2;
broadcasting the comments of the user a by using the sound parameter a, and broadcasting 3;
broadcasting the comments of the user a by using the sound parameter a, and broadcasting 4;
broadcasting the comments of the user c by using the sound parameters c, and broadcasting 5;
broadcasting the comments of the user d by using the sound parameters d, and broadcasting 6;
broadcasting the comments of the user e by using the sound parameter e, and broadcasting 7;
broadcasting the comments of the user f by using the sound parameters c, and broadcasting 8;
and broadcasting the comment of the user b by using the sound parameter b, and broadcasting 9.
According to the scheme, when the number of different users is larger than the number of the maximum sound parameters, the comment content of the user f is broadcasted by using the sound parameter c, because the fact that continuous multiple comments use different sound parameters is guaranteed, on the premise, the comment content of the user f can be broadcasted by using the sound parameter a or the sound parameter c, but because the user a has 3 comments, the 3 comments are broadcasted by using the sound parameter a, the user c only has one comment, and only the comment is broadcasted by using the sound parameter c, the comment content of the user f is broadcasted by preferably using the sound parameter c.
With further reference to fig. 6, as an implementation of the methods shown in the above-mentioned figures, the present application provides an embodiment of a voice broadcast apparatus, where the embodiment of the apparatus corresponds to the embodiment of the method shown in fig. 1, and the apparatus may be specifically applied to various electronic devices.
As shown in fig. 6, the voice broadcasting device 600 of the present embodiment may include: the broadcast system comprises an acquisition module 601, a configuration module 602 and a broadcast module 603. The obtaining module 601 is configured to obtain comment information of a currently browsed page of a user; the configuration module 602 is configured to configure corresponding sound parameters for the comment information based on the user information corresponding to the comment information; and the broadcasting module 603 is configured to convert the comment information into voice, and broadcast the voice according to the sound parameters corresponding to the comment information.
In this embodiment, in the voice broadcasting apparatus 600: the specific processing of the obtaining module 601, the configuring module 602, and the broadcasting module 603 and the technical effects thereof can refer to the related description of step 101 and step 103 in the corresponding embodiment of fig. 1, which is not repeated herein.
In some optional implementations of this embodiment, the user information includes a user ID; and the configuration module is further configured to: the analysis submodule is configured to analyze the comment information to obtain time information of the comment information and a user ID corresponding to the comment information; and the configuration sub-module is configured to configure corresponding sound parameters for the comment information based on the time information of the comment information and the user ID corresponding to each comment information.
In some optional implementations of this embodiment, the configuration module is further configured to: the counting submodule is configured to count the number of the comment information corresponding to each user ID and the number of different user IDs; a determining sub-module configured to determine the number of the maximum sound parameters based on the number of the different user IDs and a threshold number of the preset sound parameters.
In some optional implementations of this embodiment, the determining sub-module is further configured to: setting the number of the maximum sound parameters as the number of different user IDs in response to the number of different user IDs not being larger than the preset threshold number of sound parameters; in response to the number of different user IDs being greater than the threshold number of preset sound parameters, the number of maximum sound parameters is set to the threshold number of preset sound parameters.
In some optional implementations of this embodiment, the configuration submodule is further configured to: and responding to the plurality of pieces of comment information corresponding to the same user ID, and configuring the same sound parameter for the plurality of pieces of comment information corresponding to the same user ID.
In some optional implementations of this embodiment, the configuration submodule is further configured to: in response to a predetermined number of consecutive pieces of comment information corresponding to a plurality of user IDs, different sound parameters are configured for each of the plurality of user IDs.
In some optional implementations of this embodiment, the user information further includes attribute information of the user; and the configuration module is further configured to: and configuring corresponding sound parameters for the comment information based on the user attribute information corresponding to the comment information.
In some optional implementations of this embodiment, the attribute information of the user includes: the geographic location where the user is located, the age of the user, or the gender of the user.
There is also provided, in accordance with an embodiment of the present application, an electronic device, a readable storage medium, and a computer program product.
FIG. 7 illustrates a schematic block diagram of an example electronic device 700 that can be used to implement embodiments of the present disclosure. Electronic devices are intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other appropriate computers. The electronic device may also represent various forms of mobile devices, such as personal digital processing, cellular phones, smart phones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions, are meant to be examples only, and are not meant to limit implementations of the disclosure described and/or claimed herein.
As shown in fig. 7, the device 700 comprises a computing unit 701, which may perform various suitable actions and processes according to a computer program stored in a Read Only Memory (ROM)702 or a computer program loaded from a storage unit 708 into a Random Access Memory (RAM) 703. In the RAM 703, various programs and data required for the operation of the device 700 can also be stored. The computing unit 701, the ROM 702, and the RAM 703 are connected to each other by a bus 704. An input/output (I/O) interface 705 is also connected to bus 704.
Various components in the device 700 are connected to the I/O interface 705, including: an input unit 706 such as a keyboard, a mouse, or the like; an output unit 707 such as various types of displays, speakers, and the like; a storage unit 808, such as a magnetic disk, optical disk, or the like; and a communication unit 709 such as a network card, modem, wireless communication transceiver, etc. The communication unit 709 allows the device 700 to exchange information/data with other devices via a computer network, such as the internet, and/or various telecommunication networks.
Various implementations of the systems and techniques described here above may be implemented in digital electronic circuitry, integrated circuitry, Field Programmable Gate Arrays (FPGAs), Application Specific Integrated Circuits (ASICs), Application Specific Standard Products (ASSPs), system on a chip (SOCs), load programmable logic devices (CPLDs), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include: implemented in one or more computer programs that are executable and/or interpretable on a programmable system including at least one programmable processor, which may be special or general purpose, receiving data and instructions from, and transmitting data and instructions to, a storage system, at least one input device, and at least one output device.
Program code for implementing the methods of the present disclosure may be written in any combination of one or more programming languages. These program codes may be provided to a processor or controller of a general purpose computer, special purpose computer, or other programmable data processing apparatus, such that the program codes, when executed by the processor or controller, cause the functions/operations specified in the flowchart and/or block diagram to be performed. The program code may execute entirely on the machine, partly on the machine, as a stand-alone software package partly on the machine and partly on a remote machine or entirely on the remote machine or server.
In the context of this disclosure, a machine-readable medium may be a tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. A machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a machine-readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
To provide for interaction with a user, the systems and techniques described here can be implemented on a computer having: a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to a user; and a keyboard and a pointing device (e.g., a mouse or a trackball) by which a user can provide input to the computer. Other kinds of devices may also be used to provide for interaction with a user; for example, feedback provided to the user can be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user may be received in any form, including acoustic, speech, or tactile input.
The systems and techniques described here can be implemented in a computing system that includes a back-end component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front-end component (e.g., a user computer having a graphical user interface or a web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such back-end, middleware, or front-end components. The components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include: local Area Networks (LANs), Wide Area Networks (WANs), and the Internet.
The computer system may include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
It should be understood that various forms of the flows shown above may be used, with steps reordered, added, or deleted. For example, the steps described in the present disclosure may be executed in parallel, sequentially, or in different orders, as long as the desired results of the technical solutions disclosed in the present disclosure can be achieved, and the present disclosure is not limited herein.
The above detailed description should not be construed as limiting the scope of the disclosure. It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and substitutions may be made in accordance with design requirements and other factors. Any modification, equivalent replacement, and improvement made within the spirit and principle of the present disclosure should be included in the scope of protection of the present disclosure.
Claims (19)
1. A voice broadcasting method includes:
obtaining comment information of a current browsing page of a user;
configuring corresponding sound parameters for the comment information based on the user information corresponding to the comment information;
and converting the comment information into voice, and broadcasting the voice by using the sound parameters corresponding to the comment information.
2. The method of claim 1, wherein the user information comprises a user ID; and
configuring corresponding sound parameters for the comment information based on the user information corresponding to the comment information, wherein the configuring comprises the following steps:
analyzing the comment information to obtain time information of the comment information and a user ID corresponding to the comment information;
and configuring corresponding sound parameters for the comment information based on the time information of the comment information and the user ID corresponding to each comment information.
3. The method of claim 2, wherein after the analyzing the comment information to obtain time information of the comment information and a user ID corresponding to the comment information, the method further comprises:
counting the number of comment information corresponding to each user ID and the number of different user IDs;
determining the number of the maximum sound parameters based on the number of the different user IDs and a threshold number of preset sound parameters.
4. The method of claim 3, wherein the determining a number of maximum sound parameters based on the number of different user IDs and a threshold number of pre-set sound parameters comprises:
setting the number of the maximum sound parameters as the number of the different user IDs in response to the number of the different user IDs not being greater than a preset threshold number of sound parameters;
setting the number of the maximum sound parameters to a threshold number of preset sound parameters in response to the number of the different user IDs being greater than the threshold number of preset sound parameters.
5. The method of claim 2, wherein the configuring, for the comment information, a corresponding sound parameter based on the time information of the comment information and the user ID corresponding to each comment information includes:
and responding to the condition that the plurality of pieces of comment information correspond to the same user ID, and configuring the same sound parameter for the plurality of pieces of comment information corresponding to the same user ID.
6. The method of claim 5, wherein the configuring the comment information with the corresponding sound parameters based on the time information of the comment information and the user ID corresponding to each comment information further comprises:
in response to a predetermined number of consecutive pieces of comment information corresponding to a plurality of user IDs, configuring different sound parameters for each of the plurality of user IDs.
7. The method of claim 1, wherein the user information further includes attribute information of a user; and
configuring corresponding sound parameters for the comment information based on the user information corresponding to the comment information, and further comprising:
and configuring corresponding sound parameters for the comment information based on the user attribute information corresponding to the comment information.
8. The method of claim 7, wherein the attribute information of the user comprises: the geographic location where the user is located, the age of the user, or the gender of the user.
9. A voice broadcasting device comprising:
the acquisition module is configured to acquire comment information of a current browsing page of a user;
the configuration module is configured to configure corresponding sound parameters for the comment information based on the user information corresponding to the comment information;
and the broadcasting module is configured to convert the comment information into voice and broadcast the voice according to the sound parameters corresponding to the comment information.
10. The apparatus of claim 9, wherein the user information comprises a user ID; and
the configuration module is further configured to:
the analysis submodule is configured to analyze the comment information to obtain time information of the comment information and a user ID corresponding to the comment information;
and the configuration sub-module is configured to configure corresponding sound parameters for the comment information based on the time information of the comment information and the user ID corresponding to each comment information.
11. The apparatus of claim 10, wherein the configuration module is further configured to:
the counting submodule is configured to count the number of the comment information corresponding to each user ID and the number of different user IDs;
a determining sub-module configured to determine a number of maximum sound parameters based on the number of different user IDs and a threshold number of preset sound parameters.
12. The apparatus of claim 11, wherein the determination submodule is further configured to:
setting the number of the maximum sound parameters as the number of the different user IDs in response to the number of the different user IDs not being greater than a preset threshold number of sound parameters;
setting the number of the maximum sound parameters to a threshold number of preset sound parameters in response to the number of the different user IDs being greater than the threshold number of preset sound parameters.
13. The apparatus of claim 10, wherein the configuration sub-module is further configured to:
and responding to the condition that the plurality of pieces of comment information correspond to the same user ID, and configuring the same sound parameter for the plurality of pieces of comment information corresponding to the same user ID.
14. The apparatus of claim 13, wherein the configuration sub-module is further configured to:
in response to a predetermined number of consecutive pieces of comment information corresponding to a plurality of user IDs, configuring different sound parameters for each of the plurality of user IDs.
15. The apparatus of claim 9, wherein the user information further includes attribute information of a user; and
the configuration module is further configured to:
and configuring corresponding sound parameters for the comment information based on the user attribute information corresponding to the comment information.
16. The apparatus of claim 15, wherein the attribute information of the user comprises: the geographic location where the user is located, the age of the user, or the gender of the user.
17. An electronic device, comprising:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method of any one of claims 1-8.
18. A non-transitory computer readable storage medium having stored thereon computer instructions for causing the computer to perform the method of any one of claims 1-8.
19. A computer program product comprising a computer program which, when executed by a processor, implements the method according to any one of claims 1-8.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011563561.4A CN112735375A (en) | 2020-12-25 | 2020-12-25 | Voice broadcasting method, device, equipment and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011563561.4A CN112735375A (en) | 2020-12-25 | 2020-12-25 | Voice broadcasting method, device, equipment and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112735375A true CN112735375A (en) | 2021-04-30 |
Family
ID=75616285
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011563561.4A Pending CN112735375A (en) | 2020-12-25 | 2020-12-25 | Voice broadcasting method, device, equipment and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112735375A (en) |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090319273A1 (en) * | 2006-06-30 | 2009-12-24 | Nec Corporation | Audio content generation system, information exchanging system, program, audio content generating method, and information exchanging method |
JP2014011509A (en) * | 2012-06-27 | 2014-01-20 | Sharp Corp | Voice output control device, voice output control method, program, and recording medium |
CN103533519A (en) * | 2012-07-06 | 2014-01-22 | 盛乐信息技术(上海)有限公司 | Short message broadcasting method and system |
CN104485100A (en) * | 2014-12-18 | 2015-04-01 | 天津讯飞信息科技有限公司 | Text-to-speech pronunciation person self-adaptive method and system |
CN105959271A (en) * | 2016-04-25 | 2016-09-21 | 乐视控股(北京)有限公司 | Text content information voice conversion method, playing method, and playing device |
CN107452400A (en) * | 2017-07-24 | 2017-12-08 | 珠海市魅族科技有限公司 | Voice broadcast method and device, computer installation and computer-readable recording medium |
CN108366302A (en) * | 2018-02-06 | 2018-08-03 | 南京创维信息技术研究院有限公司 | TTS broadcast commands optimization method, smart television, system and storage device |
CN110379406A (en) * | 2019-06-14 | 2019-10-25 | 北京字节跳动网络技术有限公司 | Voice remark conversion method, system, medium and electronic equipment |
CN110968673A (en) * | 2019-12-04 | 2020-04-07 | 北京明略软件系统有限公司 | Voice comment playing method and device, voice equipment and storage medium |
CN111276123A (en) * | 2018-11-16 | 2020-06-12 | 阿拉的(深圳)人工智能有限公司 | Method and device for voice broadcasting message, computer equipment and storage medium |
-
2020
- 2020-12-25 CN CN202011563561.4A patent/CN112735375A/en active Pending
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090319273A1 (en) * | 2006-06-30 | 2009-12-24 | Nec Corporation | Audio content generation system, information exchanging system, program, audio content generating method, and information exchanging method |
JP2014011509A (en) * | 2012-06-27 | 2014-01-20 | Sharp Corp | Voice output control device, voice output control method, program, and recording medium |
CN103533519A (en) * | 2012-07-06 | 2014-01-22 | 盛乐信息技术(上海)有限公司 | Short message broadcasting method and system |
CN104485100A (en) * | 2014-12-18 | 2015-04-01 | 天津讯飞信息科技有限公司 | Text-to-speech pronunciation person self-adaptive method and system |
CN105959271A (en) * | 2016-04-25 | 2016-09-21 | 乐视控股(北京)有限公司 | Text content information voice conversion method, playing method, and playing device |
CN107452400A (en) * | 2017-07-24 | 2017-12-08 | 珠海市魅族科技有限公司 | Voice broadcast method and device, computer installation and computer-readable recording medium |
CN108366302A (en) * | 2018-02-06 | 2018-08-03 | 南京创维信息技术研究院有限公司 | TTS broadcast commands optimization method, smart television, system and storage device |
CN111276123A (en) * | 2018-11-16 | 2020-06-12 | 阿拉的(深圳)人工智能有限公司 | Method and device for voice broadcasting message, computer equipment and storage medium |
CN110379406A (en) * | 2019-06-14 | 2019-10-25 | 北京字节跳动网络技术有限公司 | Voice remark conversion method, system, medium and electronic equipment |
CN110968673A (en) * | 2019-12-04 | 2020-04-07 | 北京明略软件系统有限公司 | Voice comment playing method and device, voice equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113239705A (en) | Pre-training method and device of semantic representation model, electronic equipment and storage medium | |
CN114492831B (en) | Method and device for generating federal learning model | |
CN113055751B (en) | Data processing method, device, electronic equipment and storage medium | |
CN112508115B (en) | Method, apparatus, device and computer storage medium for establishing node representation model | |
CN113360711A (en) | Model training and executing method, device, equipment and medium for video understanding task | |
CN113691864A (en) | Video clipping method, video clipping device, electronic equipment and readable storage medium | |
CN114861059A (en) | Resource recommendation method and device, electronic equipment and storage medium | |
CN113903329B (en) | Voice processing method and device, electronic equipment and storage medium | |
CN118170378A (en) | Page generation method, page generation device, electronic device, storage medium and program product | |
CN113904943A (en) | Account detection method and device, electronic equipment and storage medium | |
CN113869042A (en) | Text title generation method and device, electronic equipment and storage medium | |
CN116992000A (en) | Interactive information processing method, device, electronic equipment and computer readable medium | |
US20230076471A1 (en) | Training method, text translation method, electronic device, and storage medium | |
CN113590447B (en) | Buried point processing method and device | |
CN112735375A (en) | Voice broadcasting method, device, equipment and storage medium | |
CN113254578B (en) | Method, apparatus, device, medium and product for data clustering | |
CN114330718B (en) | Method and device for extracting causal relationship and electronic equipment | |
CN114399513B (en) | Method and device for training image segmentation model and image segmentation | |
CN113838450B (en) | Audio synthesis and corresponding model training method, device, equipment and storage medium | |
CN113901316A (en) | Information pushing method and device, electronic equipment and storage medium | |
CN114071192A (en) | Information acquisition method, terminal, server, electronic device, and storage medium | |
CN113553413A (en) | Dialog state generation method and device, electronic equipment and storage medium | |
CN114254028A (en) | Event attribute extraction method and device, electronic equipment and storage medium | |
CN114138358A (en) | Application program starting optimization method, device, equipment and storage medium | |
CN114399998A (en) | Voice processing method, device, equipment, storage medium and program product |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |