CN112291281B - Voice broadcasting and voice broadcasting content setting method and device - Google Patents
Voice broadcasting and voice broadcasting content setting method and device Download PDFInfo
- Publication number
- CN112291281B CN112291281B CN201910613779.7A CN201910613779A CN112291281B CN 112291281 B CN112291281 B CN 112291281B CN 201910613779 A CN201910613779 A CN 201910613779A CN 112291281 B CN112291281 B CN 112291281B
- Authority
- CN
- China
- Prior art keywords
- name
- target object
- replaced
- voice file
- original name
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 82
- 238000013507 mapping Methods 0.000 claims description 76
- 230000000977 initiatory effect Effects 0.000 claims description 5
- 230000005021 gait Effects 0.000 claims description 3
- 230000000875 corresponding effect Effects 0.000 description 172
- 241000123069 Ocyurus chrysurus Species 0.000 description 30
- 238000010586 diagram Methods 0.000 description 20
- 230000008569 process Effects 0.000 description 16
- 238000005516 engineering process Methods 0.000 description 9
- 230000003993 interaction Effects 0.000 description 8
- 238000004364 calculation method Methods 0.000 description 7
- 238000004590 computer program Methods 0.000 description 7
- 230000006870 function Effects 0.000 description 7
- 230000004044 response Effects 0.000 description 5
- 238000004458 analytical method Methods 0.000 description 4
- 230000008520 organization Effects 0.000 description 3
- 241000219000 Populus Species 0.000 description 2
- 230000005540 biological transmission Effects 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 230000008676 import Effects 0.000 description 2
- 230000002452 interceptive effect Effects 0.000 description 2
- OKTJSMMVPCPJKN-UHFFFAOYSA-N Carbon Chemical compound [C] OKTJSMMVPCPJKN-UHFFFAOYSA-N 0.000 description 1
- 238000012550 audit Methods 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 229910021389 graphene Inorganic materials 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 235000012054 meals Nutrition 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 229920001690 polydopamine Polymers 0.000 description 1
- 239000004984 smart glass Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/12—Protocols specially adapted for proprietary or special-purpose networking environments, e.g. medical networks, sensor networks, networks in vehicles or remote metering networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/3331—Query processing
- G06F16/334—Query execution
- G06F16/3343—Query execution using phonetics
-
- G—PHYSICS
- G07—CHECKING-DEVICES
- G07C—TIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
- G07C1/00—Registering, indicating or recording the time of events or elapsed time, e.g. time-recorders for work people
- G07C1/10—Registering, indicating or recording the time of events or elapsed time, e.g. time-recorders for work people together with the recording, indicating or registering of other data, e.g. of signs of identity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/06—Protocols specially adapted for file transfer, e.g. file transfer protocol [FTP]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L9/00—Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols
- H04L9/32—Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols including means for verifying the identity or authority of a user of the system or for message authentication, e.g. authorization, entity authentication, data integrity or data verification, non-repudiation, key authentication or verification of credentials
- H04L9/3226—Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols including means for verifying the identity or authority of a user of the system or for message authentication, e.g. authorization, entity authentication, data integrity or data verification, non-repudiation, key authentication or verification of credentials using a predetermined code, e.g. password, passphrase or PIN
- H04L9/3231—Biological data, e.g. fingerprint, voice or retina
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Security & Cryptography (AREA)
- Acoustics & Sound (AREA)
- Biomedical Technology (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- General Engineering & Computer Science (AREA)
- Biodiversity & Conservation Biology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computing Systems (AREA)
- Medical Informatics (AREA)
- Telephonic Communication Services (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
One or more embodiments of the present disclosure provide a method and an apparatus for setting a voice broadcast and a voice broadcast content, where the voice broadcast method may include: identity recognition is carried out on the target object; acquiring a voice file corresponding to the target object according to the identification result of the target object; when the original name of the target object contains a polyphone element, the voice file is generated by a replaced name, and the replaced name is obtained by replacing the polyphone element in the original name with a designated monophone element; and playing the acquired voice file.
Description
Technical Field
One or more embodiments of the present disclosure relate to the field of identity recognition technologies, and in particular, to a method and apparatus for voice broadcasting and setting content of voice broadcasting.
Background
In some scenarios, identification of related objects is required. For example, the visitor is identified by the access control device to determine whether to pass or not, or the enterprise employee is identified by the attendance checking device to record the attendance checking data and the like. After the identification is successful, the name of the related object can be broadcasted through voice so as to indicate that the identification is successful.
Disclosure of Invention
In view of this, one or more embodiments of the present disclosure provide a method and apparatus for setting a voice broadcast and a voice broadcast content.
In order to achieve the above object, one or more embodiments of the present disclosure provide the following technical solutions:
according to a first aspect of one or more embodiments of the present disclosure, a method for voice broadcasting is provided, including:
identity recognition is carried out on the target object;
acquiring a voice file corresponding to the target object according to the identification result of the target object; when the original name of the target object contains a polyphone element, the voice file is generated by a replaced name, and the replaced name is obtained by replacing the polyphone element in the original name with a designated monophone element;
and playing the acquired voice file.
According to a second aspect of one or more embodiments of the present disclosure, a method for setting a voice broadcast content is provided, including:
initiating an object adding request to a server, wherein the object adding request is used for adding a target object as an identifiable object of an identity recognition device, and the original name of the target object contains a polyphone element;
Acquiring a plurality of alternatives returned by the server, wherein the alternatives respectively correspond to the pronunciation of the multitone element;
determining a designated single-tone element corresponding to the selected alternative, wherein the designated single-tone element is used for generating a voice file corresponding to the target object so as to play the voice file to the target user after the target user is successfully identified by the identity identification equipment.
According to a third aspect of one or more embodiments of the present disclosure, a method for voice broadcasting is provided, including:
receiving an inquiry request from an identification device, wherein the inquiry request comprises an original name of a target object identified by the identification device, and the original name comprises a polyphone element;
returning a replaced name corresponding to the original name, so that the identification equipment generates a voice file according to the replaced name and plays the voice file; or, returning the voice file generated according to the replaced name to be played by the identity recognition equipment; wherein the replaced name is obtained by replacing the multitone element in the original name with a designated monophone element.
According to a fourth aspect of one or more embodiments of the present disclosure, a method for setting a voice broadcast content is provided, including:
receiving an object adding request initiated by an administrator, wherein the object adding request is used for adding a target object as an identifiable object of identity recognition equipment, and the original name of the target object contains a polyphone element;
returning a plurality of alternatives to the administrator, the plurality of alternatives corresponding to respective readings of the multitone element;
determining a designated single-tone element corresponding to the selected alternative, wherein the designated single-tone element is used for generating a voice file corresponding to the target object so as to play the voice file to the target user after the target user is successfully identified by the identity identification equipment.
According to a fifth aspect of one or more embodiments of the present specification, there is provided a method of adding a community member, comprising:
acquiring a joining request initiated by a user aiming at a group, wherein personal information of the user comprises an original name and a designated single-tone element corresponding to the original name, and the designated single-tone element and a multi-tone element in the original name have the same pronunciation;
When the administrator of the community agrees to the joining request, the user is added as a member of the community, so that the original name and the designated monophonic element are updated into member information corresponding to the user.
According to a sixth aspect of one or more embodiments of the present specification, there is provided a display method, including:
acquiring content to be displayed, wherein the content to be displayed comprises an original name of a target object and a designated single-tone element corresponding to the original name, and the designated single-tone element and a multi-tone element in the original name have the same pronunciation;
and displaying the content to be displayed.
According to a seventh aspect of one or more embodiments of the present disclosure, there is provided a voice broadcasting apparatus, including:
the identification unit is used for carrying out identity identification on the target object;
the acquisition unit acquires a voice file corresponding to the target object according to the identification result of the target object; when the original name of the target object contains a polyphone element, the voice file is generated by a replaced name, and the replaced name is obtained by replacing the polyphone element in the original name with a designated monophone element;
And the playing unit plays the acquired voice file.
According to an eighth aspect of one or more embodiments of the present disclosure, there is provided a setting device for voice broadcast content, including:
a request unit for initiating an object addition request to a server, wherein the object addition request is used for adding a target object as an identifiable object of the identity recognition device, and the original name of the target object contains a polyphonic element;
an obtaining unit, configured to obtain a plurality of alternatives returned by the server, where the plurality of alternatives respectively correspond to each pronunciation of the multitone element;
and the determining unit is used for determining a designated single-tone element corresponding to the selected alternative, wherein the designated single-tone element is used for generating a voice file corresponding to the target object so as to play the voice file to the target user after the target user is successfully identified by the identity identification equipment.
According to a ninth aspect of one or more embodiments of the present disclosure, there is provided a voice broadcasting apparatus, including:
a receiving unit for receiving an inquiry request from an identification device, wherein the inquiry request comprises an original name of a target object identified by the identification device, and the original name comprises a polyphone element;
The returning unit returns the replaced name corresponding to the original name, so that the identification equipment generates a voice file according to the replaced name and plays the voice file; or, returning the voice file generated according to the replaced name to be played by the identity recognition equipment; wherein the replaced name is obtained by replacing the multitone element in the original name with a designated monophone element.
According to a tenth aspect of one or more embodiments of the present disclosure, there is provided a setting device for voice broadcast content, including:
the receiving unit is used for receiving an object adding request initiated by an administrator, wherein the object adding request is used for adding a target object as an identifiable object of the identity recognition equipment, and the original name of the target object contains a polyphone element;
a return unit that returns a plurality of alternatives to the administrator, the plurality of alternatives corresponding to respective readings of the multitone element, respectively;
and the determining unit is used for determining a designated single-tone element corresponding to the selected alternative, wherein the designated single-tone element is used for generating a voice file corresponding to the target object so as to play the voice file to the target user after the target user is successfully identified by the identity identification equipment.
According to an eleventh aspect of one or more embodiments of the present specification, there is provided an adding device of a community member, including:
the personal information of the user comprises an original name and a designated single-tone element corresponding to the original name, wherein the designated single-tone element and a multi-tone element in the original name have the same pronunciation;
and an adding unit for adding the user as a member of the community when the administrator of the community agrees to the joining request, so that the original name and the designated monophonic element are updated to the member information corresponding to the user.
According to a twelfth aspect of one or more embodiments of the present specification, there is provided a display device including:
an acquisition unit that acquires content to be displayed, the content to be displayed including an original name of a target object and a designated single-tone element corresponding to the original name, the designated single-tone element having the same pronunciation as a multi-tone element in the original name;
and the display unit displays the content to be displayed.
According to a thirteenth aspect of one or more embodiments of the present specification, there is provided an electronic device, comprising:
A processor;
a memory for storing processor-executable instructions;
wherein the processor implements the method of the first aspect by executing the executable instructions.
According to a fourteenth aspect of one or more embodiments of the present description, there is provided a computer readable storage medium having stored thereon computer instructions which, when executed by a processor, implement the steps of the method as described in the first aspect.
According to a fifteenth aspect of one or more embodiments of the present specification, there is provided an electronic device, comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor implements the method of the second aspect by executing the executable instructions.
According to a sixteenth aspect of one or more embodiments of the present description, there is provided a computer readable storage medium having stored thereon computer instructions which, when executed by a processor, implement the steps of the method as described in the second aspect.
According to a seventeenth aspect of one or more embodiments of the present specification, there is provided an electronic device, comprising:
a processor;
a memory for storing processor-executable instructions;
Wherein the processor implements the method according to the third aspect by executing the executable instructions.
According to an eighteenth aspect of one or more embodiments of the present description, there is provided a computer readable storage medium having stored thereon computer instructions which, when executed by a processor, implement the steps of the method according to the third aspect.
According to a nineteenth aspect of one or more embodiments of the present specification, there is provided an electronic device, comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor implements the method of the fourth aspect by executing the executable instructions.
According to a twentieth aspect of one or more embodiments of the present description, there is provided a computer readable storage medium having stored thereon computer instructions which, when executed by a processor, implement the steps of the method as described in the fourth aspect.
According to a twenty-first aspect of one or more embodiments of the present specification, there is provided an electronic device comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor implements the method of the fifth aspect by executing the executable instructions.
According to a twenty-second aspect of one or more embodiments of the present description, there is provided a computer-readable storage medium having stored thereon computer instructions which, when executed by a processor, implement the steps of the method as described in the fifth aspect.
According to a twenty-third aspect of one or more embodiments of the present specification, there is provided an electronic device, comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor implements the method of the sixth aspect by executing the executable instructions.
According to a twenty-fourth aspect of one or more embodiments of the present description, there is provided a computer readable storage medium having stored thereon computer instructions which, when executed by a processor, implement the steps of the method as described in the sixth aspect.
Drawings
Fig. 1 is a schematic architecture diagram of an identification system according to an exemplary embodiment.
Fig. 2 is a flowchart of a voice broadcasting method according to an exemplary embodiment.
Fig. 3 is a flowchart of another voice broadcast method according to an exemplary embodiment.
Fig. 4 is a flowchart of a method for setting a voice broadcast content according to an exemplary embodiment.
Fig. 5 is a flowchart of another method for setting voice broadcast content according to an exemplary embodiment.
FIG. 6 is an interactive schematic diagram of adding identifiable objects to an attendance device and performing an attendance operation, in accordance with an illustrative embodiment.
Fig. 7 is a schematic diagram of a pronunciation selection interface provided by an exemplary embodiment.
Fig. 8 is a flow chart of a method of adding community members provided in an exemplary embodiment.
Fig. 9 is a flowchart of a display method provided by an exemplary embodiment.
Fig. 10 is a schematic structural view of an apparatus provided in one of the exemplary embodiments.
Fig. 11 is a block diagram of a voice broadcast device according to an exemplary embodiment.
Fig. 12 is a schematic diagram of an apparatus according to a second exemplary embodiment.
Fig. 13 is a block diagram of a setting device for voice broadcast content according to an exemplary embodiment.
Fig. 14 is a schematic view of an apparatus according to a third exemplary embodiment.
Fig. 15 is a block diagram of another voice broadcast device according to an exemplary embodiment.
Fig. 16 is a schematic view of an apparatus according to a fourth exemplary embodiment.
Fig. 17 is a block diagram of another setting apparatus for voice broadcast content according to an exemplary embodiment.
Fig. 18 is a schematic diagram of an apparatus according to a fifth exemplary embodiment.
Fig. 19 is a block diagram of an add-on device for community members provided by an exemplary embodiment.
Fig. 20 is a schematic view of an apparatus according to a sixth exemplary embodiment.
Fig. 21 is a block diagram of a display device provided by an exemplary embodiment.
Detailed Description
Reference will now be made in detail to exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, the same numbers in different drawings refer to the same or similar elements, unless otherwise indicated. The implementations described in the following exemplary embodiments do not represent all implementations consistent with one or more embodiments of the present specification. Rather, they are merely examples of apparatus and methods consistent with aspects of one or more embodiments of the present description as detailed in the accompanying claims.
It should be noted that: in other embodiments, the steps of the corresponding method are not necessarily performed in the order shown and described in this specification. In some other embodiments, the method may include more or fewer steps than described in this specification. Furthermore, individual steps described in this specification, in other embodiments, may be described as being split into multiple steps; while various steps described in this specification may be combined into a single step in other embodiments.
Fig. 1 is a schematic architecture diagram of an identification system according to an exemplary embodiment. As shown in fig. 1, the system may include a server 11, a network 12, a number of electronic devices, such as a cell phone 13, a cell phone 14, a cell phone 15, etc., and an identification device 16.
The server 11 may be a physical server comprising a separate host, or the server 11 may be a virtual server carried by a cluster of hosts. During the running process, the server 11 may run a program on the server side of a certain application to implement the relevant service functions of the application, for example, when the server 11 runs a program of the mobile community office platform, it may be implemented as a service end of the community office platform.
The handsets 13-15 are but one type of electronic device that may be used by a user. Indeed, it is obvious that the user may also use electronic devices of the type such as: tablet devices, notebook computers, palm top computers (PDAs, personal Digital Assistants), wearable devices (e.g., smart glasses, smart watches, etc.), etc., as one or more embodiments of the present description are not limited in this regard. In the running process, the electronic device may run a program on the client side of an application to implement the relevant service function of the application, for example, when the electronic device runs a program of a group office platform, the electronic device may be implemented as a client of the group office platform.
The identification device 16 may be configured to perform an identification function, such as, but not limited to, fingerprint recognition, face recognition, iris recognition, gait recognition, and other physiological characteristic recognition techniques. The identification device 16 may be used to implement a plurality of scenarios such as access control and attendance management by implementing identification, which is not limited in this specification. During operation, the identification device 16 may run a program on the client side of an application to implement the relevant business functions of the application, such as being implemented as a client of a community office platform when the identification device 16 runs a program of the community office platform.
The group office platform not only can realize the communication function, but also can be used as an integrated functional platform with a plurality of other functions, such as processing of approval events (such as examination and approval events of leave, office article claim, finance and the like), attendance events, task events, log events and other group internal events, and processing of group external events such as meal ordering, purchasing and the like, which is not limited by one or more embodiments of the present specification. Through the server side running on the server 11, the electronic devices such as the mobile phones 13-15 and the client side running on the identity recognition device 16, the data interaction among the electronic devices such as the server 11, the mobile phones 13-15 and the identity recognition device 16 can be realized. For example, the server 11 may maintain organization architecture information of a group, and the electronic devices such as the mobile phones 13 to 15 may be used to manage the organization architecture information; and, the information of the group members contained in the organization architecture can be imported into the identification device 16 from the server 11 in batches, so that the identification device 16 can identify the members in the group. The group office platform can be suitable for a mobile office scene, namely the group office platform can be a mobile group office platform.
And the network 12 for interaction between the handsets 13-15, the identification device 16 and the server 11 may comprise various types of wired or wireless networks.
Fig. 2 is a flowchart of a voice broadcasting method according to an exemplary embodiment. As shown in fig. 2, the method applied to an identification device (e.g., identification device 16 shown in fig. 1, etc.), may include the steps of:
step 202, identity recognition is performed on the target object.
In an embodiment, the identification device is configured to perform feature collection on a target object, and the collected feature information is used to identify the target object. The type of the feature information is related to the identity recognition technology used by the identity recognition device, for example, when the fingerprint recognition technology is used, the feature information may include fingerprint features, and when the face recognition technology is used, the feature information may include face features.
In an embodiment, the identification device may locally identify the target object according to the collected feature information, so as to obtain an identification result. Or the identity recognition equipment can upload the acquired characteristic information to a server, and the server returns an identity recognition result to the identity recognition equipment after carrying out identity recognition on the target object; the server can operate a server side of the group office platform, the identity recognition device can operate a client side of the group office platform, and the server side and the client side can transmit characteristic information, identity recognition results and the like based on the client side and the server side.
Step 204, obtaining a voice file corresponding to the target object according to the identification result of the target object; wherein when the original name of the target object contains a multitone element, the voice file is generated from a replaced name obtained by replacing the multitone element in the original name with a specified monophone element.
In one embodiment, the name is made up of several elements, the element type being related to the literal type employed by the name. For example, when chinese is used, the elements may be chinese characters, and when english is used, the elements may be words, which are not listed here. The polyphonic element refers to an element having a plurality of pronunciations; taking chinese as an example, the polyphonic elements are polyphones in chinese, for example, the pronunciation of the chinese character "madder" includes "qi a n" and "x ī". Similarly, a single sound element refers to an element having only one pronunciation; taking chinese as an example, the single-tone element is a single-tone word in chinese, for example, the pronunciation of the chinese character "western" is only "x ī".
In an embodiment, for the text type adopted by the original name, a multitone element library containing all multitone elements and a monophone element library containing all monophone elements may be generated in advance. Then, by respectively matching all elements contained in the original name with a multi-sound element library, the multi-sound elements contained in the original name and all pronunciations of the multi-sound elements can be determined; further, by matching each reading with the single-tone element library, the single-tone element corresponding to each reading can be selected. For example, when the original name of the target object is "Yang Qian", by matching the Chinese character "yangzhu" and the Chinese character "madder" with a polyphonic element library (i.e., polyphonic word library), respectively, it can be determined that the Chinese character "madder" is a polyphonic word, and the pronunciation "qi, n" and "x ī" exist; further, by matching the pronunciation "qi a n" and "x ī" with the list of single-tone elements (i.e. list of single-tone words), it is possible to determine the single-tone word "qia" with the pronunciation "qi a n" and the single-tone word "xi" with the pronunciation "x ī".
Although the original name contains the polyphonic element, the original name of the target object must only adopt a fixed pronunciation, so that the monophonic element adopting the pronunciation can be set as the specified monophonic element according to the pronunciation actually adopted by the target object for the polyphonic element contained in the original name, thereby generating the replaced name. For example, when the original name of the target object actually reads "y ng x ī", the multitone element "madder" may be replaced with the designated monophone element "west" of the reading "x ī", and the resulting replaced name is "Yang Xi". Specifically, the target object or the administrator may select to determine the actual pronunciation of the polyphonic element contained in the original name by providing the target object or the administrator with alternatives of the respective pronunciation corresponding to the polyphonic element.
In an embodiment, by generating a corresponding replaced name for the original name containing the polyphonic element and generating the voice file based on the replaced name, the voice file can be directly generated without adding a special mark (such as the mark [ qian4] for indicating that the pronunciation of "madder" is "qi a n", etc.) to the actual pronunciation of the polyphonic element, and compared with the case that different manufacturers may adopt different mark formats, the scheme has extremely high universality, and analysis of the mark is avoided, and complexity and processing calculation amount are avoided when the voice file is generated.
In one embodiment, the identification device may obtain a post-replacement name of the target object and generate the voice file locally according to the post-replacement name. Then, just because the technical scheme of the specification does not need to add marks, the identification equipment only needs relatively low processing performance, and can smoothly and timely generate the voice file, and the delay of voice broadcasting and even the normal operation of the identification equipment are not influenced due to insufficient processing performance.
In one embodiment, the mapping relationship between the original names and the replaced names of all the objects is stored on the identification device. When the administrator inputs or imports the identity information of all the objects on the identity recognition device, the identity recognition device can recognize the multi-tone elements contained in the original names based on the multi-tone element library, determine the corresponding tone elements of each pronunciation respectively based on the tone element library, provide the alternative options corresponding to each pronunciation or tone element respectively for the administrator or related objects, and then generate the corresponding replaced names according to the appointed tone elements corresponding to the selected alternative options, so as to form the mapping relation between the corresponding original names and the replaced names on the identity recognition device. Alternatively, the administrator may pre-establish a mapping relationship between the original name and the replaced name of each object, and then import the mapping relationship into the identification device. Or, the server may identify the polyphonic element contained in the original name based on the polyphonic element library, determine the corresponding polyphonic element of each pronunciation based on the polyphonic element library, provide the administrator or the related object with the alternative items corresponding to each pronunciation or the monophonic element respectively, and then generate the corresponding replaced name according to the designated monophonic element corresponding to the selected alternative item, thereby forming a mapping relationship between the corresponding original name and the replaced name, and send the mapping relationship to the identity recognition device. Correspondingly, the identity recognition device can determine the original name of the target object according to the identity recognition result of the target object, and then acquire the replaced name of the target object according to the mapping relation between the locally stored original name and the replaced name, so that a corresponding voice file is generated based on the replaced name.
In one embodiment, the mapping relationship between the original names and the replaced names of all objects is stored on the server. For example, the server may identify the polyphonic element included in the original name based on the polyphonic element library, determine the corresponding polyphonic element for each pronunciation based on the polyphonic element library, provide the administrator or the related object with the candidate corresponding to each pronunciation or the corresponding monophonic element, and then generate the corresponding replaced name according to the designated monophonic element corresponding to the selected candidate, so as to form and store the mapping relationship between the corresponding original name and the replaced name. Accordingly, the identification device may determine the original name of the target object according to the identification result of the target object, then send an inquiry request containing the original name to the server, and receive a replaced name returned by the server, where the replaced name is determined by the server according to the mapping relationship between the stored original name and the replaced name.
In an embodiment, the identification device may acquire a pre-generated voice file corresponding to the target object, without generating the voice file temporarily after identifying the identity information of the target object, which may avoid feedback delay caused by the process of generating the voice file, and is helpful for accelerating the response speed of the identification device.
In an embodiment, the identification device may generate a corresponding voice file in advance according to the replaced name of the target object, and store the voice file in the local or cloud storage space. Similarly, the identification device may generate in advance a voice file corresponding to each of all the objects, and store the voice file in a local or cloud storage space.
In an embodiment, the identification device may receive a voice file that is pre-generated by the server according to the replaced name of the target object; after confirming the identity information of the target object, the identity identification device may temporarily request the server to obtain the voice file, or obtain the voice file in advance and store the voice file in a local or cloud storage space, and then read the voice file from the local or cloud storage space. Similarly, the identification device may acquire, from the server, voice files corresponding to all objects generated in advance, respectively.
In an embodiment, for a pre-generated voice file, a mapping relationship may be directly established between the original name and the voice file, so that the corresponding voice file may be directly determined according to the original name of the target object. Or, a mapping relationship can be established between the replaced name and the voice file, and then the mapping relationship between the original name and the replaced name is combined, so that after the original name is determined, the corresponding voice file can be determined as well.
In an embodiment, the above-mentioned mapping relationship (the mapping relationship between the original name and the replaced name, or the mapping relationship between the original name and the voice file, or the mapping relationship between the replaced name and the voice file) is not applicable to all objects recognized by the identification device, but is bound to one or more set objects. For example, two users with original names of "Yang Qian" may have one user with "yang qi a n" and another user with "y a ng x ī", so that for each user with original names containing multi-tone elements, a corresponding mapping relationship should be generated respectively to ensure accuracy of voice broadcasting.
When the identity recognition equipment locally recognizes the identity information of the object, generates the mapping relation, matches the mapping relation and generates the voice file, the identity recognition equipment can independently realize the technical scheme of the specification without a server, and realize the correct broadcasting of the multitone elements. Or the identification equipment can be matched with the server, for example, one or more steps of generating the mapping relation, matching the mapping relation, generating the voice file and the like are realized on the server, so that the processing calculation pressure of the identification equipment can be reduced, and the performance requirement of the identification equipment is reduced.
And step 206, playing the acquired voice file.
In an embodiment, after the identity information of the target object is identified, the identity information of the target object is successfully identified by the identity identification device through playing the voice file by the identity identification device to correctly broadcast the name of the target object.
In an embodiment, in addition to playing the voice file, the identification device may also display the original name of the target object, which may help the target object determine that the identification device has successfully identified its own identity information. Similarly, other identity information of the target object may also be presented; for example, in an enterprise scenario, other identity information may include employee job number, affiliated departments, job positions, etc., and, for example, in a school scenario, other identity information may include affiliated classes, etc.
Fig. 3 is a flowchart of another voice broadcast method according to an exemplary embodiment. As shown in fig. 3, the method applied to a server (e.g., the server 11 shown in fig. 1) may include the steps of:
step 302, receiving an inquiry request from an identification device, wherein the inquiry request contains an original name of a target object identified by the identification device, and the original name contains a multitone element.
In an embodiment, the identification device is configured to perform feature collection on a target object, and the collected feature information is used to identify the target object. The type of the feature information is related to the identity recognition technology used by the identity recognition device, for example, when the fingerprint recognition technology is used, the feature information may include fingerprint features, and when the face recognition technology is used, the feature information may include face features.
In an embodiment, the identification device may locally identify the target object according to the collected feature information, so as to obtain an identification result. Or the identity recognition equipment can upload the acquired characteristic information to a server, and the server returns an identity recognition result to the identity recognition equipment after carrying out identity recognition on the target object; the server can operate a server side of the group office platform, the identity recognition device can operate a client side of the group office platform, and the server side and the client side can transmit characteristic information, identity recognition results and the like based on the client side and the server side.
Step 304, returning to the replaced name corresponding to the original name, so that the identification equipment generates a voice file according to the replaced name and plays the voice file; or, returning the voice file generated according to the replaced name to be played by the identity recognition equipment; wherein the replaced name is obtained by replacing the multitone element in the original name with a designated monophone element.
In one embodiment, the name is made up of several elements, the element type being related to the literal type employed by the name. For example, when chinese is used, the elements may be chinese characters, and when english is used, the elements may be words, which are not listed here. The polyphonic element refers to an element having a plurality of pronunciations; taking chinese as an example, the polyphonic elements are polyphones in chinese, for example, the pronunciation of the chinese character "madder" includes "qi a n" and "x ī". Similarly, a single sound element refers to an element having only one pronunciation; taking chinese as an example, the single-tone element is a single-tone word in chinese, for example, the pronunciation of the chinese character "western" is only "x ī".
In an embodiment, for the text type adopted by the original name, a multitone element library containing all multitone elements and a monophone element library containing all monophone elements may be generated in advance. Then, by respectively matching all elements contained in the original name with a multi-sound element library, the multi-sound elements contained in the original name and all pronunciations of the multi-sound elements can be determined; further, by matching each reading with the single-tone element library, the single-tone element corresponding to each reading can be selected. For example, when the original name of the target object is "Yang Qian", by matching the Chinese character "yangzhu" and the Chinese character "madder" with a polyphonic element library (i.e., polyphonic word library), respectively, it can be determined that the Chinese character "madder" is a polyphonic word, and the pronunciation "qi, n" and "x ī" exist; further, by matching the pronunciation "qi a n" and "x ī" with the list of single-tone elements (i.e. list of single-tone words), it is possible to determine the single-tone word "qia" with the pronunciation "qi a n" and the single-tone word "xi" with the pronunciation "x ī".
Although the original name contains the polyphonic element, the original name of the target object must only adopt a fixed pronunciation, so that the monophonic element adopting the pronunciation can be set as the specified monophonic element according to the pronunciation actually adopted by the target object for the polyphonic element contained in the original name, thereby generating the replaced name. For example, when the original name of the target object actually reads "y ng x ī", the multitone element "madder" may be replaced with the designated monophone element "west" of the reading "x ī", and the resulting replaced name is "Yang Xi". Specifically, the target object or the administrator may select to determine the actual pronunciation of the polyphonic element contained in the original name by providing the target object or the administrator with alternatives of the respective pronunciation corresponding to the polyphonic element.
In an embodiment, by generating a corresponding replaced name for the original name containing the polyphonic element and generating the voice file based on the replaced name, the voice file can be directly generated without adding a special mark (such as the mark [ qian4] for indicating that the pronunciation of "madder" is "qi a n", etc.) to the actual pronunciation of the polyphonic element, and compared with the case that different manufacturers may adopt different mark formats, the scheme has extremely high universality, and analysis of the mark is avoided, and complexity and processing calculation amount are avoided when the voice file is generated.
In one embodiment, the mapping relationship between the original name and the replaced name of all the objects (which refer to the objects whose original names contain polyphonic elements, and the following description) is stored on the server. For example, the server may identify the polyphonic element included in the original name based on the polyphonic element library, determine the corresponding polyphonic element for each pronunciation based on the polyphonic element library, provide the administrator or the related object with the candidate corresponding to each pronunciation or the corresponding monophonic element, and then generate the corresponding replaced name according to the designated monophonic element corresponding to the selected candidate, so as to form and store the mapping relationship between the corresponding original name and the replaced name. Accordingly, the identification device may determine the original name of the target object according to the identification result of the target object, and then send an inquiry request containing the original name to the server, and the server may return a corresponding replaced name based on the mapping relationship, so that the identification device may generate a corresponding voice file based on the replaced name.
In an embodiment, the server may generate the voice file corresponding to the target object in advance. Then, after confirming the identity information of the target object, the identity recognition device can send the original name of the target object to the server, so that the server returns a corresponding voice file to the identity recognition device. Because the voice file is not required to be generated temporarily, the response speed of the identity recognition equipment can be increased.
In an embodiment, for a pre-generated voice file, a correspondence may be directly established between the original name and the voice file, so that the corresponding voice file may be directly determined according to the original name of the target object. Or, a corresponding relation can be established between the replaced name and the voice file, and then the mapping relation between the original name and the replaced name is combined, so that after the original name is determined, the corresponding voice file can be determined as well.
In one embodiment, the above mapping is not applicable to all objects identified by the identification device, but is bound to one or more set objects. For example, two users with original names of "Yang Qian" may have one user with "yang qi a n" and another user with "y a ng x ī", so that for each user with original names containing multi-tone elements, a corresponding mapping relationship should be generated respectively to ensure accuracy of voice broadcasting.
In an embodiment, the identity recognition device cooperates with the server, for example, one or more steps of generating a mapping relationship, matching the mapping relationship, generating a voice file and the like are implemented on the server, so that the processing calculation pressure of the identity recognition device can be reduced, and the performance requirement of the identity recognition device can be reduced.
In an embodiment, the replaced name or the voice file is returned to the identification device, so that the identification device can play the voice file to correctly broadcast the name of the target object after identifying the identity information of the target object, and the target object can quickly determine that the identification device has successfully identified the identity information of the target object.
Fig. 4 is a flowchart of a method for setting a voice broadcast content according to an exemplary embodiment. As shown in fig. 4, the method is applied to an electronic device (such as the mobile phones 13-15 shown in fig. 1) or an identity recognition device (such as the identity recognition device 16 shown in fig. 1), and may include the following steps:
step 402, initiating an object addition request to a server, wherein the object addition request is used for adding a target object as an identifiable object of an identification device, and an original name of the target object contains a multitone element.
In an embodiment, an object addition request may be initiated by the electronic device to the server, where the object addition request may be to add a single target object as an identifiable object of the identification device, or to add multiple objects, including the target object, as identifiable objects of the identification device. For example, a server of the group office platform may be operated on the server, and a client of the group office platform may be operated on the electronic device and the identification device, so that data interaction between the server, the electronic device and the identification device may be realized based on the server and the client.
Specifically, the electronic device may initiate an object addition request to the server, and notify the server of identity information of the target object to be added according to the object addition request; then, the server may issue an acquisition instruction to the identification device, so that the identification device acquires standard feature information of the target object. The standard characteristic information can be bound with the identity information of the target object and stored on the identity recognition equipment, so that the identity recognition equipment can carry out identity recognition on the target object according to the standard characteristic information and generate an identity recognition result; or the identification equipment can upload the standard characteristic information to the server, the server binds and stores the standard characteristic information and the identification information of the target object, and when the identification equipment subsequently carries out the identification on the target object, the identification equipment needs to acquire the characteristic information of the target object and upload the characteristic information to the server, the server completes the identification, and the identification equipment is informed of the identification result.
In an embodiment, the identification device may initiate an object addition request directly to the server, where the object addition request may be to add a single target object as an identifiable object of the identification device, or to add multiple objects including the target object as identifiable objects of the identification device. For example, a server side of the group office platform can be operated on the server, and a client side of the group office platform can be operated on the identity recognition device, so that data interaction between the server and the identity recognition device can be realized based on the server side and the client side.
Specifically, the identification device may collect standard feature information of the target object. The standard characteristic information can be bound with the identity information of the target object and stored on the identity recognition equipment, so that the identity recognition equipment can carry out identity recognition on the target object according to the standard characteristic information and generate an identity recognition result; meanwhile, the identity recognition device can upload the identity information of the target object to the server, so that the server knows that the identity recognition device can recognize the identity of the target object. Or the identification equipment can upload the standard characteristic information to the server, the server binds and stores the standard characteristic information and the identification information of the target object, and when the identification equipment subsequently carries out the identification on the target object, the identification equipment needs to acquire the characteristic information of the target object and upload the characteristic information to the server, the server completes the identification, and the identification equipment is informed of the identification result.
Step 404, obtaining a plurality of alternatives returned by the server, wherein the alternatives respectively correspond to the pronunciation of the multitone element.
In one embodiment, the name is made up of several elements, the element type being related to the literal type employed by the name. For example, when chinese is used, the elements may be chinese characters, and when english is used, the elements may be words, which are not listed here. The polyphonic element refers to an element having a plurality of pronunciations; taking chinese as an example, the polyphonic elements are polyphones in chinese, for example, the pronunciation of the chinese character "madder" includes "qi a n" and "x ī". Similarly, a single sound element refers to an element having only one pronunciation; taking chinese as an example, the single-tone element is a single-tone word in chinese, for example, the pronunciation of the chinese character "western" is only "x ī".
In an embodiment, for the text type adopted by the original name, a multitone element library containing all multitone elements and a monophone element library containing all monophone elements may be generated in advance. Then, by respectively matching all elements contained in the original name with a multi-sound element library, the multi-sound elements contained in the original name and all pronunciations of the multi-sound elements can be determined; further, by matching each reading with the single-tone element library, the single-tone element corresponding to each reading can be selected. For example, when the original name of the target object is "Yang Qian", by matching the Chinese character "yangzhu" and the Chinese character "madder" with a polyphonic element library (i.e., polyphonic word library), respectively, it can be determined that the Chinese character "madder" is a polyphonic word, and the pronunciation "qi, n" and "x ī" exist; further, by matching the pronunciation "qi a n" and "x ī" with the list of single-tone elements (i.e. list of single-tone words), it is possible to determine the single-tone word "qia" with the pronunciation "qi a n" and the single-tone word "xi" with the pronunciation "x ī".
Wherein, although the original name contains the polyphonic element, the original name of the target object necessarily adopts only one fixed pronunciation, so the server can return a plurality of alternatives corresponding to each pronunciation of the polyphonic element for selection; correspondingly, according to the selected alternative options, the pronunciation actually adopted by the target object for the polyphonic elements contained in the original names can be determined. For example, when the original name of the target object is "Yang Qian", the server may return two alternatives corresponding to the pronunciation "qi a n" and "x ī" according to the pronunciation "qi a n" and the pronunciation "x ī" of the multitone element, such as an alternative corresponding to the pronunciation "qi a n", for selection.
Step 406, determining a designated single-tone element corresponding to the selected alternative, wherein the designated single-tone element is used for generating a voice file corresponding to the target object, so as to play the voice file to the target user after the target user is successfully identified by the identity identification device.
As described above, although the original name contains the polyphonic element, the original name of the target object necessarily adopts only one fixed pronunciation, and thus the monophonic element adopting the pronunciation can be set as the specified monophonic element according to the pronunciation actually adopted by the target object for the polyphonic element contained in the original name, thereby generating the replaced name. For example, when the original name of the target object actually reads "y ng x ī", the multitone element "madder" may be replaced with the designated monophone element "west" of the reading "x ī", and the resulting replaced name is "Yang Xi". Specifically, the target object or the administrator may select to determine the actual pronunciation of the polyphonic element contained in the original name by providing the target object or the administrator with alternatives of the respective pronunciation corresponding to the polyphonic element.
In an embodiment, by generating a corresponding replaced name for the original name containing the polyphonic element and generating the voice file based on the replaced name, the voice file can be directly generated without adding a special mark (such as the mark [ qian4] for indicating that the pronunciation of "madder" is "qi a n", etc.) to the actual pronunciation of the polyphonic element, and compared with the case that different manufacturers may adopt different mark formats, the scheme has extremely high universality, and analysis of the mark is avoided, and complexity and processing calculation amount are avoided when the voice file is generated.
In an embodiment, the server may generate and store a mapping relationship between the corresponding original name and the replaced name according to the determined replaced name; similarly, the server may store the mapping between the original names and the replaced names of all objects. In the subsequent process, the identification device can determine the original name of the target object according to the identification result of the target object, then send an inquiry request containing the original name to the server, and the server can return a corresponding replaced name based on the mapping relation, so that the identification device can generate a corresponding voice file based on the replaced name.
In an embodiment, the server may generate a corresponding voice file in advance according to the determined replaced name and store the voice file; similarly, the server may store voice files corresponding to the replaced names of all objects. In the subsequent process, the identification device can send the original name of the target object to the server after confirming the identity information of the target object, so that the server returns a corresponding voice file to the identification device. Because the voice file is not required to be generated temporarily, the response speed of the identity recognition equipment can be increased.
In an embodiment, for a pre-generated voice file, a mapping relationship may be directly established between the original name and the voice file, so that the corresponding voice file may be directly determined according to the original name of the target object. Or, a mapping relationship can be established between the replaced name and the voice file, and then the mapping relationship between the original name and the replaced name is combined, so that after the original name is determined, the corresponding voice file can be determined as well.
In an embodiment, the above-mentioned mapping relationship (the mapping relationship between the original name and the replaced name, or the mapping relationship between the original name and the voice file, or the mapping relationship between the replaced name and the voice file) is not applicable to all objects recognized by the identification device, but is bound to one or more set objects. For example, two users with original names of "Yang Qian" may have one user with "yang qi a n" and another user with "y a ng x ī", so that for each user with original names containing multi-tone elements, a corresponding mapping relationship should be generated respectively to ensure accuracy of voice broadcasting.
In an embodiment, the replaced name or the voice file is returned to the identification device, so that the identification device can play the voice file to correctly broadcast the name of the target object after identifying the identity information of the target object, and the target object can quickly determine that the identification device has successfully identified the identity information of the target object.
Fig. 5 is a flowchart of another method for setting voice broadcast content according to an exemplary embodiment. As shown in fig. 5, the method applied to a server (e.g., the server 11 shown in fig. 1, etc.), may include the steps of:
step 502, receiving an object addition request initiated by an administrator, wherein the object addition request is used for adding a target object as an identifiable object of an identity recognition device, and an original name of the target object contains a multitone element.
In an embodiment, an object addition request may be initiated by the electronic device to the server, the object addition request may be to add a separate target object as an identifiable object of the identification device, or to add a plurality of objects including the target object as identifiable objects of the identification device. For example, a server of the group office platform may be operated on the server, and a client of the group office platform may be operated on the electronic device and the identification device, so that data interaction between the server, the electronic device and the identification device may be realized based on the server and the client.
Specifically, the electronic device may initiate an object addition request to the server, and notify the server of identity information of the target object to be added according to the object addition request; then, the server may issue an acquisition instruction to the identification device, so that the identification device acquires standard feature information of the target object. The standard characteristic information can be bound with the identity information of the target object and stored on the identity recognition equipment, so that the identity recognition equipment can carry out identity recognition on the target object according to the standard characteristic information and generate an identity recognition result; or the identification equipment can upload the standard characteristic information to the server, the server binds and stores the standard characteristic information and the identification information of the target object, and when the identification equipment subsequently carries out the identification on the target object, the identification equipment needs to acquire the characteristic information of the target object and upload the characteristic information to the server, the server completes the identification, and the identification equipment is informed of the identification result.
In an embodiment, the object addition request may be initiated by the identification device directly to the server, and the object addition request may be an individual target object added as an identifiable object of the identification device, or a plurality of objects including the target object may be added as identifiable objects of the identification device. For example, a server side of the group office platform can be operated on the server, and a client side of the group office platform can be operated on the identity recognition device, so that data interaction between the server and the identity recognition device can be realized based on the server side and the client side.
Specifically, the identification device may collect standard feature information of the target object. The standard characteristic information can be bound with the identity information of the target object and stored on the identity recognition equipment, so that the identity recognition equipment can carry out identity recognition on the target object according to the standard characteristic information and generate an identity recognition result; meanwhile, the identity recognition device can upload the identity information of the target object to the server, so that the server knows that the identity recognition device can recognize the identity of the target object. Or the identification equipment can upload the standard characteristic information to the server, the server binds and stores the standard characteristic information and the identification information of the target object, and when the identification equipment subsequently carries out the identification on the target object, the identification equipment needs to acquire the characteristic information of the target object and upload the characteristic information to the server, the server completes the identification, and the identification equipment is informed of the identification result.
Step 504, returning a plurality of alternatives to the administrator, wherein the alternatives respectively correspond to the pronunciation of the multitone element.
In one embodiment, the name is made up of several elements, the element type being related to the literal type employed by the name. For example, when chinese is used, the elements may be chinese characters, and when english is used, the elements may be words, which are not listed here. The polyphonic element refers to an element having a plurality of pronunciations; taking chinese as an example, the polyphonic elements are polyphones in chinese, for example, the pronunciation of the chinese character "madder" includes "qi a n" and "x ī". Similarly, a single sound element refers to an element having only one pronunciation; taking chinese as an example, the single-tone element is a single-tone word in chinese, for example, the pronunciation of the chinese character "western" is only "x ī".
In an embodiment, for the text type adopted by the original name, a multitone element library containing all multitone elements and a monophone element library containing all monophone elements may be generated in advance. Then, by respectively matching all elements contained in the original name with a multi-sound element library, the multi-sound elements contained in the original name and all pronunciations of the multi-sound elements can be determined; further, by matching each reading with the single-tone element library, the single-tone element corresponding to each reading can be selected. For example, when the original name of the target object is "Yang Qian", by matching the Chinese character "yangzhu" and the Chinese character "madder" with a polyphonic element library (i.e., polyphonic word library), respectively, it can be determined that the Chinese character "madder" is a polyphonic word, and the pronunciation "qi, n" and "x ī" exist; further, by matching the pronunciation "qi a n" and "x ī" with the list of single-tone elements (i.e. list of single-tone words), it is possible to determine the single-tone word "qia" with the pronunciation "qi a n" and the single-tone word "xi" with the pronunciation "x ī".
Wherein, although the original name contains the polyphonic element, the original name of the target object necessarily adopts only one fixed pronunciation, so the server can return a plurality of alternatives corresponding to each pronunciation of the polyphonic element for selection; correspondingly, according to the selected alternative options, the pronunciation actually adopted by the target object for the polyphonic elements contained in the original names can be determined. For example, when the original name of the target object is "Yang Qian", the server may return two alternatives corresponding to the pronunciation "qi a n" and "x ī" according to the pronunciation "qi a n" and the pronunciation "x ī" of the multitone element, such as an alternative corresponding to the pronunciation "qi a n", for selection.
Step 506, determining a designated single-tone element corresponding to the selected alternative, wherein the designated single-tone element is used for generating a voice file corresponding to the target object, so as to play the voice file to the target user after the target user is successfully identified by the identity identification device.
As described above, although the original name contains the polyphonic element, the original name of the target object necessarily adopts only one fixed pronunciation, and thus the monophonic element adopting the pronunciation can be set as the specified monophonic element according to the pronunciation actually adopted by the target object for the polyphonic element contained in the original name, thereby generating the replaced name. For example, when the original name of the target object actually reads "y ng x ī", the multitone element "madder" may be replaced with the designated monophone element "west" of the reading "x ī", and the resulting replaced name is "Yang Xi". Specifically, the target object or the administrator may select to determine the actual pronunciation of the polyphonic element contained in the original name by providing the target object or the administrator with alternatives of the respective pronunciation corresponding to the polyphonic element.
In an embodiment, by generating a corresponding replaced name for the original name containing the polyphonic element and generating the voice file based on the replaced name, the voice file can be directly generated without adding a special mark (such as the mark [ qian4] for indicating that the pronunciation of "madder" is "qi a n", etc.) to the actual pronunciation of the polyphonic element, and compared with the case that different manufacturers may adopt different mark formats, the scheme has extremely high universality, and analysis of the mark is avoided, and complexity and processing calculation amount are avoided when the voice file is generated.
In an embodiment, the server may generate and store a mapping relationship between the corresponding original name and the replaced name according to the determined replaced name; similarly, the server may store the mapping between the original names and the replaced names of all objects. In the subsequent process, the identification device can determine the original name of the target object according to the identification result of the target object, then send an inquiry request containing the original name to the server, and the server can return a corresponding replaced name based on the mapping relation, so that the identification device can generate a corresponding voice file based on the replaced name.
In an embodiment, the server may generate a corresponding voice file in advance according to the determined replaced name and store the voice file; similarly, the server may store voice files corresponding to the replaced names of all objects. In the subsequent process, the identification device can send the original name of the target object to the server after confirming the identity information of the target object, so that the server returns a corresponding voice file to the identification device. Because the voice file is not required to be generated temporarily, the response speed of the identity recognition equipment can be increased.
In an embodiment, for a pre-generated voice file, a mapping relationship may be directly established between the original name and the voice file, so that the corresponding voice file may be directly determined according to the original name of the target object. Or, a mapping relationship can be established between the replaced name and the voice file, and then the mapping relationship between the original name and the replaced name is combined, so that after the original name is determined, the corresponding voice file can be determined as well.
In an embodiment, the above-mentioned mapping relationship (the mapping relationship between the original name and the replaced name, or the mapping relationship between the original name and the voice file, or the mapping relationship between the replaced name and the voice file) is not applicable to all objects recognized by the identification device, but is bound to one or more set objects. For example, two users with original names of "Yang Qian" may have one user with "yang qi a n" and another user with "y a ng x ī", so that for each user with original names containing multi-tone elements, a corresponding mapping relationship should be generated respectively to ensure accuracy of voice broadcasting.
In an embodiment, the replaced name or the voice file is returned to the identification device, so that the identification device can play the voice file to correctly broadcast the name of the target object after identifying the identity information of the target object, and the target object can quickly determine that the identification device has successfully identified the identity information of the target object.
The technical scheme of the specification is described below by taking an attendance scene in an enterprise as an example. An attendance checking device is arranged at the foreground of the enterprise C, and the attendance checking device performs identity recognition on the incoming enterprise staff in a face recognition mode so as to complete corresponding attendance checking operation. The staff of the enterprise C comprises an administrator Ua and staff Ub, and the administrator Ua can add the staff Ub as an object which can be identified by the attendance checking equipment. The mobile phone used by the administrator Ua can operate the client side C1 of the group office platform T, the client side C2 of the group office platform T on the attendance checking equipment and the server side of the group office platform T, so that the administrator Ua can complete the adding operation of the staff Ub based on the interaction among the client side C1, the client side C2 and the server side. For example, fig. 6 is an interactive schematic diagram of adding identifiable objects to an attendance device and performing an attendance operation, which is provided in an exemplary embodiment. As shown in fig. 6, the interaction process may include the steps of:
In step 601, the administrator Ua initiates an employee addition request to the server.
In an embodiment, the administrator Ua initiates an employee adding request to the server running on the server through the client C1 described above. The employee adding request includes identity information of the employee Ub, such as a user name, a contact way, a position, a department, and the like.
In step 602, the server identifies polyphones.
In one embodiment, a predefined polyphone library is maintained at the server, which contains all polyphones in the Chinese characters. The server matches the user name of the employee Ub with the polyphone library to determine the polyphones contained in the user name and all the pronunciations thereof. Of course, the polyphonic word library may also be maintained at other entities than the server, and the server may call the polyphonic word library or initiate a request to the entity, which identifies the polyphonic words contained in the user name and all their pronunciations.
For example, assuming that the user name of the employee Ub is "Yang Qian", by matching with the polyphone library, it can be determined that "madder" is a polyphone, and there are 2 readings of "qi a n" and "x ī", respectively, and thus the following steps 603 to 605 and the like need to be performed. If the user name of employee Ub does not have polyphones, a jump can be made directly to step 606.
In step 603, the server looks up the alternate single-word, constructs an alternative, and returns to the administrator Ua.
In one embodiment, the server maintains a predefined library of words containing all words in the Chinese character. The server matches all the pronunciations of the polyphones identified in step 602 with the library of monophonic words, respectively, to determine the monophonic words corresponding to each of the pronunciations, respectively. Of course, the word stock may also be maintained at other entities than the server, and the server may call the word stock or initiate a request to the entity, from which the words respectively corresponding to each of the above-mentioned pronunciations are determined.
For example, when 2 pronunciations of "radix Rubiae" in the user name are "qi-a-n" and "x ī" respectively, the corresponding word "qin" of the pronunciations "qi-a-n" and the corresponding word "west" of the pronunciations "x ī" can be determined by matching with the word library.
Further, the server may construct a corresponding alternative for each pronunciation of the polyphones, and return to the administrator Ua for selection by the administrator Ua. Meanwhile, according to the corresponding relation between each pronunciation and the single word, there is a one-to-one corresponding relation between the alternative and the single word, for example, the alternative 1 corresponds to the single word "qian" and the alternative 2 corresponds to the single word "xi". Correspondingly, when the alternative 1 is selected, the pronunciation of the user name indicating the employee Ub is "yang qi a n", and the following replaced name can be generated according to the corresponding single word "qin"; alternatively, when alternative 2 is selected, the pronunciation of the user name indicating employee Ub is "ycng x ī", and the following replaced name may be generated from the corresponding word "western".
Fig. 7 is a schematic diagram of a pronunciation selection interface provided by an exemplary embodiment. The client C1 may present the administrator Ua with a pronunciation selection interface 700 as shown in fig. 7, in which alternative items 701, 702 may be shown in the pronunciation selection interface 700. Wherein the alternative 701 contains the content "Yang Qian (yang qi a n)", the pronunciation corresponding to the polyphone "madder" being "qi a n"; alternative 702 contains the content "Yang Qian (yang x ī)", with the pronunciation corresponding to the polyphonic word "madder" being "x ī". In other embodiments, the presented alternatives may include a corresponding word for each reading, such as the words "qin" and "xi" described above, and may be specifically presented as, for example, "Yang Qian (Yang Qian)" and "Yang Qian (Yang Xi)". In fact, the present description is not limited to the form of the alternatives, as long as the actual pronunciation of the user name of employee Ua can be determined.
In step 604, the administrator Ua selects a specified option, and returns a corresponding selection result to the server.
Step 605, the server generates and stores the replaced name according to the selection result.
In an embodiment, the server may learn, according to the selection result, the candidate item selected by the administrator Ua, and may determine, according to the result, a word corresponding to the selected candidate item. For example, when the administrator Ua selects the alternative 702 shown in fig. 7, the user name indicating the employee Ub reads "yang x ī", and the corresponding word is "western". Accordingly, the server may replace the polyphone "madder" in the original user name "Yang Qian" with the aforementioned monophone "western", and reconstruct the replaced name "Yang Xi".
Thus, when employee Ub has a user name of "Yang Qian" and a corresponding post-replacement name of "Yang Xi", user name "Yang Qian" and post-replacement name "Yang Xi" are both associated with employee Ub and stored at the server.
In step 606, the server pushes identity information to the attendance checking device and issues a face acquisition instruction.
In step 607, the attendance checking device collects the face features and stores them in association.
In an embodiment, the server issues the identity information of the employee Ub to the attendance checking device for storage, for example, the identity information may include a user name "Yang Qian" of the employee Ub. Meanwhile, the server issues a face acquisition instruction to the attendance equipment, so that the attendance equipment can acquire face images of staff Ub and record standard characteristic information contained in the acquired face images. Then, the attendance checking equipment can store the received identity information and the standard characteristic information in a correlated mode so as to be used for carrying out identity recognition on staff Ub and completing attendance checking operation.
The steps 601 to 607 described above belong to related steps of the configuration phase executed in advance, and based on the preparation work of the configuration phase, related steps of the attendance phase described below of the steps 608 to 612 can be realized.
At step 608, the attendance checking device identifies the current user.
In an embodiment, the attendance checking device has a certain detection range, for example, a sector area with a camera as a center of a circle, a preset radius and a preset angle, so that when the current user is in the detection range, the attendance checking device can implement face recognition on the current user to determine the identity information of the current user.
For example, when the attendance checking device detects that the current user is the employee Ub, the user name of the employee Ub may be obtained from the locally stored identity information, for example, the user name may be "Yang Qian".
In step 609, the attendance device sends the user name to the server.
In step 610, the server looks up the replaced name and returns the replaced name to the attendance checking device.
In an embodiment, for all employees in the enterprise C whose user names include polyphones, the mapping relationships between the corresponding user names and the replaced names are stored in the server, for example, the mapping relationships may be stored in the user polyphone self-selection pronunciation library. Of course, the user polyphone self-selection pronunciation library may be stored in a main body other than the server, so long as the server can call the user polyphone self-selection pronunciation library, or may request the main body to match the user polyphone self-selection pronunciation library.
When the user name uploaded by the attendance checking device is received, the server can search the mapping relation corresponding to the user name in the multi-word self-selection pronunciation library of the user, so that the replaced name corresponding to the user name is determined. If the user name uploaded by the attendance equipment does not contain the polyphones, the server cannot find the mapping relation matched with the user name in the user polyphone self-selection pronunciation library, and the server returns the user name to the attendance equipment, so that the attendance equipment generates a voice file based on the user name. If the user name uploaded by the attendance checking device contains polyphones, the server can find the mapping relation corresponding to the user name, so that the replaced name corresponding to the user name is determined. For example, when the user name is "Yang Qian", the server may find the corresponding post-replacement name "Yang Xi" so that the server may return the post-replacement name "Yang Xi" to the attendance device.
It should be noted that: taking the user name "Yang Qian" as an example, multiple employees with user names "Yang Qian" may exist simultaneously in the enterprise C. Therefore, when the server records the mapping relationship between the user name and the replaced name, other information of the employee Ub, such as the user ID of the employee Ub, etc., may be added to the server, and the attendance checking device sends the user name to the server in step 609, and carries the user ID, so as to ensure that the server finds the mapping relationship corresponding to the employee Ub, but not other mapping relationships corresponding to "Yang Qian".
In step 611, the attendance checking device generates a voice file according to the replaced name.
In an embodiment, the attendance checking device may automatically generate the corresponding voice file according to the kanji contained in the replaced name. For example, when the postname is replaced with "Yang Xi", the kanji "poplar" and "western" may be obtained such that the voice file generated accordingly by the attendance checking device corresponds to the pronunciation "yang x ī".
Step 612, the attendance checking device displays the user name, plays the voice file, and performs the attendance checking operation.
In one embodiment, the attendance checking device displays the user name "Yang Qian" on the screen, and simultaneously plays the voice file, wherein the broadcast sound formed by the voice file is "yang x ī", and the pronunciation adopted by the polyphone "madder" in the user name is "x ī". And when the employee Ub sees the displayed user name Yang Qian or hears the broadcast sound yang x ī, the employee Ub can determine that the attendance checking device accurately identifies the employee himself, so that the attendance checking device can complete corresponding attendance checking operation aiming at the employee.
In summary, in the description, by identifying the polyphone element contained in the original name of the target object and replacing the polyphone element with the monophone element with the same pronunciation according to the pronunciation actually adopted by the polyphone element, the identity recognition device can realize correct broadcasting of the original name by playing the language file corresponding to the monophone element, without adding a corresponding mark to the pronunciation adopted by the polyphone element or analyzing the mark in the subsequent process, so that the processing calculation amount of the identity recognition device is reduced and the response speed is accelerated.
In addition to setting the polyphones in the names for the members in the community by the administrator through the above embodiment, the polyphones may be actively set by the members at the beginning of joining the community so as to be added to the member information of the members, and thereafter may be applied to all scenes in the community without repeated setting. For example, fig. 8 is a flow chart of a method of adding community members provided in an exemplary embodiment. As shown in fig. 8, the method applied to a server (e.g., the server 11 shown in fig. 1) may include the steps of:
step 802, obtaining a joining request initiated by a user aiming at a group, wherein personal information of the user comprises an original name and a designated single-tone element corresponding to the original name, and the designated single-tone element and a multi-tone element in the original name have the same pronunciation.
In one embodiment, the user may generate the request content, fill in personal information such as an original name in the process of generating the request content, determine a designated single-tone element corresponding to the multi-tone element in the original name, and then initiate the joining request based on the generated request content. For example, the electronic device may open a request content generation interface that contains an original name input option, may actively fill in the original name by the user, or may automatically fill in the original name contained in the profile of the logged-in account; then, the electronic device may upload the original name to the server, the server identifies the multi-tone element contained in the original name, and returns the alternative item corresponding to each pronunciation, and the user selects a certain alternative item according to the actual pronunciation of the multi-tone element in the original name, where the single-tone element corresponding to the alternative item is determined as the specified single-tone element; finally, the electronic device generates a request content containing personal information such as an original name, a designated tone element, and the like as a join request, and initiates the join request. Then, when the join request is sent to an administrator of the community, the administrator can see personal information contained in the join request, and the administrator can audit the join request.
In one embodiment, the user may first initiate a join request and then determine a corresponding designated monophonic element for the polyphonic element. For example, the electronic device initiated join request may contain a portion of personal information, and the portion of personal information may contain the original name but not the specified monophonic element described above. Then, when the join request is sent to the administrator of the community, the administrator does not view the designated monophonic element corresponding to the polyphonic element in the original name during the auditing process. The joining request is obtained from the user by the server and sent to the administrator, so that the server can obtain the original name contained in the joining request, identify the polyphonic element contained in the original name, and return the alternative options corresponding to each pronunciation to the user, and the user selects a certain alternative option according to the actual pronunciation of the polyphonic element in the original name, wherein the monophonic element corresponding to the alternative option is determined as the specified monophonic element so as to be associated with the polyphonic element in the original name of the user.
In one embodiment, the name is made up of several elements, the element type being related to the literal type employed by the name. For example, when chinese is used, the elements may be chinese characters, and when english is used, the elements may be words, which are not listed here. The polyphonic element refers to an element having a plurality of pronunciations; taking chinese as an example, the polyphonic elements are polyphones in chinese, for example, the pronunciation of the chinese character "madder" includes "qi a n" and "x ī". Similarly, a single sound element refers to an element having only one pronunciation; taking chinese as an example, the single-tone element is a single-tone word in chinese, for example, the pronunciation of the chinese character "western" is only "x ī".
In an embodiment, for the text type adopted by the original name, a multitone element library containing all multitone elements and a monophone element library containing all monophone elements may be generated in advance. Then, by respectively matching all elements contained in the original name with a multi-sound element library, the multi-sound elements contained in the original name and all pronunciations of the multi-sound elements can be determined; further, by matching each reading with the single-tone element library, the single-tone element corresponding to each reading can be selected. For example, when the original name of the user is "Yang Qian", by matching the Chinese characters "poplar" and "madder" with a polyphonic element library (i.e., polyphonic word library), respectively, it is possible to determine that the Chinese character "madder" is a polyphonic word, and that the pronunciation "qi, n" and "x ī" exist; further, by matching the pronunciation "qi a n" and "x ī" with the list of single-tone elements (i.e. list of single-tone words), it is possible to determine the single-tone word "qia" with the pronunciation "qi a n" and the single-tone word "xi" with the pronunciation "x ī".
Wherein, although the original name contains the polyphonic element, the original name of the user must only adopt a fixed pronunciation, so the server can return a plurality of alternatives corresponding to each pronunciation of the polyphonic element for selection; correspondingly, according to the selected alternative options, the pronunciation actually adopted by the user for the polyphonic elements contained in the original names can be determined. For example, when the original name of the user is "Yang Qian", the server may return two alternatives corresponding to the pronunciation "qi a n" and "x ī" according to the pronunciation "qi a n" and the pronunciation "x ī" of the multitone element, such as the alternative corresponding to the pronunciation "qi a n", for selection.
Step 804, when the administrator of the community agrees to the joining request, adding the user as a member of the community, so that the original name and the designated monophonic element are updated to the member information corresponding to the user.
In an embodiment, in an enterprise office scenario, the community may be an enterprise, and the join request may be an job-in application of the user. Of course, the present description may also be applied in other scenarios.
In an embodiment, the original name and the designated single-tone element are updated to the member information corresponding to the user, so that the member information corresponding to the user can be applied in the event related to the user, which occurs later in the group, and multiplexing can be realized under each scene by the original name and the designated single-tone element contained in the member information. For example, in the attendance scene, the original name may be issued to the attendance device, and the voice file generated according to the replaced name (the replaced name is obtained by replacing the multitone element in the original name with the designated monophone element) may be issued to the attendance device, so that after the attendance device recognizes the identity information of the user, the attendance operation may be performed, and the original name may be displayed and the voice file may be reported. For another example, in the conference check-in scenario, the original name and the designated single-tone element may be issued to the check-in device, so that after the check-in device recognizes the identity information of the user, the check-in operation may be performed while the original name is displayed, and the multi-tone element contained in the original name may be marked with the designated single-tone element.
Besides voice broadcasting, the technical scheme of the specification can be applied to a display scene and marks the multi-sound elements contained in the display content, so that any user can accurately know the correct pronunciation of the relevant multi-sound elements after looking over. For example, fig. 9 is a flowchart of a display method provided in an exemplary embodiment. As shown in fig. 9, the method applied to the display device may include the steps of:
step 902, obtaining content to be displayed, wherein the content to be displayed comprises an original name of a target object and a designated single-tone element corresponding to the original name, and the designated single-tone element and a multi-tone element in the original name have the same pronunciation.
In an embodiment, the content to be displayed may be any content including a multitone element, such as a user name, an object name, a pet name, a poster content, a landmark content, and the like, which is not limited in this specification.
In an embodiment, an arbitrary user may provide the original content to a server, such as uploading the original content to the server or instructing the server to obtain the original content from a certain channel, the server identifies a polyphonic element contained in the original content, and returns a plurality of alternatives corresponding to each pronunciation of the polyphonic element to the arbitrary user for selection; correspondingly, the server can determine the pronunciation actually adopted by the user for the polyphonic elements contained in the original content according to the selected alternative options, so that the corresponding appointed monophonic element is determined. For example, a multitone element library including all multitone elements, a monophone element library including all monophone elements may be generated in advance for the text type employed by the original content. Then, by respectively matching all elements contained in the original content with a multi-sound element library, multi-sound elements contained in the original content and all pronunciations of the multi-sound elements can be determined; further, by matching each reading with the single-tone element library, the single-tone element corresponding to each reading can be selected.
Further, by adding a designated single-tone element corresponding to the multi-tone element in the original content, the designated single-tone element marks the multi-tone element, and the element to be displayed can be generated. For example, assuming that the original content includes a name "… … young madder … …" of a certain user, and that a designated monophonic element corresponding to a polyphonic element "madder" is "western" is determined based on the above manner, a corresponding content to be displayed may be generated as "… … Yang Qian (western) … …" to achieve the tagging effect. Of course, it may be marked in other ways, which the present specification does not limit.
In one embodiment, the user presents the original content by providing the original content to the server such that the server pushes the original content to the display device. Then, the server can actively determine the multitone elements contained in the original content through the above process, and further determine the corresponding designated monophone elements, so that after the original content is generated into the content to be displayed, the content to be displayed is pushed to the display device for displaying.
And step 904, displaying the content to be displayed.
Since the multitone elements are marked by designating the multitone elements, the displayed content to be displayed is easy to view and read, and the situation of pronunciation errors does not occur. For example, in a group office scenario, the display device may include an electronic station card of a group member, and the content to be displayed may include information such as names of the group members, so that the scheme may mark the polyphonic elements in the member names to avoid misreading.
Fig. 10 is a schematic structural diagram of an apparatus provided in an exemplary embodiment. Referring to fig. 10, at the hardware level, the device includes a processor 1002, an internal bus 1004, a network interface 1006, a memory 1008, and a non-volatile memory 1010, although other hardware required by other services is possible. The processor 1002 reads a corresponding computer program from the nonvolatile memory 1010 into the memory 1008 and then runs to form a voice broadcast device on a logic level. Of course, in addition to software implementation, one or more embodiments of the present disclosure do not exclude other implementation manners, such as a logic device or a combination of software and hardware, etc., that is, the execution subject of the following processing flow is not limited to each logic unit, but may also be hardware or a logic device.
Referring to fig. 11, in a software embodiment, the voice broadcasting device may include:
an identification unit 1101 that performs identification for a target object;
an obtaining unit 1102, configured to obtain a voice file corresponding to the target object according to an identification result of the target object; when the original name of the target object contains a polyphone element, the voice file is generated by a replaced name, and the replaced name is obtained by replacing the polyphone element in the original name with a designated monophone element;
And a playing unit 1103 for playing the acquired voice file.
Optionally, the acquiring unit 1102 is specifically configured to:
acquiring a replaced name of the target object;
and generating the voice file according to the replaced name.
Optionally, the obtaining unit 1102 obtains the replaced name of the target object by:
determining the original name of the target object according to the identification result of the target object; acquiring the replaced name of the target object according to the mapping relation between the locally stored original name and the replaced name; or alternatively, the first and second heat exchangers may be,
determining the original name of the target object according to the identification result of the target object; and sending an inquiry request containing the original name to a server, and receiving a replaced name returned by the server, wherein the replaced name is determined by the server according to the mapping relation between the stored original name and the replaced name.
Optionally, the acquiring unit 1102 is specifically configured to:
and acquiring a pre-generated voice file corresponding to the target object.
Optionally, the method further comprises:
a generation unit 1104 that generates the voice file in advance from the replaced name of the target object; or alternatively, the first and second heat exchangers may be,
And a receiving unit 1105, wherein the receiving server generates the voice file in advance according to the replaced name of the target object.
Optionally, the method further comprises:
and a display unit 1106 for displaying the original name of the target object.
Optionally, the identifying unit 1101 is specifically configured to:
identifying the target object by at least one of the following modes: fingerprint recognition, face recognition, iris recognition and gait recognition.
Fig. 12 is a schematic block diagram of an apparatus provided in an exemplary embodiment. Referring to fig. 12, at a hardware level, the apparatus includes a processor 1202, an internal bus 1204, a network interface 1206, a memory 1208, and a nonvolatile memory 1210, although other hardware may be included as needed for other services. The processor 1202 reads the corresponding computer program from the nonvolatile memory 1210 into the memory 1208 and then runs the program to form a setting device of the voice broadcast content on a logic level. Of course, in addition to software implementation, one or more embodiments of the present disclosure do not exclude other implementation manners, such as a logic device or a combination of software and hardware, etc., that is, the execution subject of the following processing flow is not limited to each logic unit, but may also be hardware or a logic device.
Referring to fig. 13, in a software embodiment, the setting device for voice broadcast content may include:
a request unit 1301, configured to initiate an object addition request to a server, where the object addition request is used to add a target object as an identifiable object of an identity recognition device, and an original name of the target object includes a polyphonic element;
an obtaining unit 1302, configured to obtain a plurality of alternatives returned by the server, where the plurality of alternatives respectively correspond to respective pronunciations of the multitone element;
the determining unit 1303 determines a designated single-tone element corresponding to the selected alternative, where the designated single-tone element is used to generate a voice file corresponding to the target object, so as to play the voice file to the target user after the target user is successfully identified by the identification device.
Fig. 14 is a schematic structural diagram of an apparatus provided in an exemplary embodiment. Referring to fig. 14, at a hardware level, the device includes a processor 1402, an internal bus 1404, a network interface 1406, a memory 1408, and a nonvolatile storage 1410, although other hardware required by other services is also possible. The processor 1402 reads the corresponding computer program from the nonvolatile memory 1410 to the memory 1408 and then executes the computer program to form a voice broadcasting device on a logic level. Of course, in addition to software implementation, one or more embodiments of the present disclosure do not exclude other implementation manners, such as a logic device or a combination of software and hardware, etc., that is, the execution subject of the following processing flow is not limited to each logic unit, but may also be hardware or a logic device.
Referring to fig. 15, in a software embodiment, the voice broadcasting device may include:
a receiving unit 1501 that receives an inquiry request from an identification device, the inquiry request including an original name of a target object identified by the identification device, the original name including a polyphonic element;
a returning unit 1502, configured to return a replaced name corresponding to the original name, so that the identification device generates and plays a voice file according to the replaced name; or, returning the voice file generated according to the replaced name to be played by the identity recognition equipment; wherein the replaced name is obtained by replacing the multitone element in the original name with a designated monophone element.
Fig. 16 is a schematic block diagram of an apparatus provided in an exemplary embodiment. Referring to fig. 16, at the hardware level, the device includes a processor 1602, an internal bus 1604, a network interface 1606, a memory 1608, and a non-volatile storage 1610, although other hardware required for other services is also possible. The processor 1602 reads a corresponding computer program from the nonvolatile memory 1610 into the memory 1608 and then runs, forming a setting device of the voice broadcast content on a logic level. Of course, in addition to software implementation, one or more embodiments of the present disclosure do not exclude other implementation manners, such as a logic device or a combination of software and hardware, etc., that is, the execution subject of the following processing flow is not limited to each logic unit, but may also be hardware or a logic device.
Referring to fig. 17, in a software embodiment, the setting device for voice broadcast content may include:
a receiving unit 1701, configured to receive an object addition request initiated by an administrator, where the object addition request is used to add a target object as an identifiable object of an identity recognition device, and an original name of the target object includes a polyphone element;
a return unit 1702 that returns a plurality of alternatives to the administrator, the plurality of alternatives corresponding to respective readings of the multitone element, respectively;
a determining unit 1703 determines a designated single-tone element corresponding to the selected candidate, where the designated single-tone element is used to generate a voice file corresponding to the target object, so as to play the voice file to the target user after the target user is successfully identified by the identification device.
Optionally, the return unit 1702 is specifically configured to:
matching each component element contained in the original name with a predefined multi-sound element library respectively so as to identify multi-sound elements contained in the original name and all pronunciations thereof;
matching each pronunciation of the identified multi-pronunciation element with a predefined single-pronunciation element library respectively to determine the single-pronunciation element corresponding to each pronunciation;
And returning the determined alternative options corresponding to the single-tone elements to the administrator.
Optionally, the method further comprises:
the establishing unit 1704 establishes a mapping relation between the original name and the replaced name so as to return the replaced name or a voice file generated according to the replaced name when the identification equipment receives an inquiry request containing the original name; or, establishing a mapping relation between the original name and the voice file so as to return to the voice file when the identification equipment receives an inquiry request containing the original name;
wherein the replaced name is obtained by replacing the multitone element in the original name with the specified monophone element.
Fig. 18 is a schematic structural diagram of an apparatus provided in an exemplary embodiment. Referring to fig. 18, at the hardware level, the device includes a processor 1802, an internal bus 1804, a network interface 1806, a memory 1808, and a non-volatile storage 1810, although other hardware may be included as needed for other services. The processor 1802 reads the corresponding computer program from the non-volatile memory 1810 into the memory 1808 and then runs to form an add-on device for community members at a logical level. Of course, in addition to software implementation, one or more embodiments of the present disclosure do not exclude other implementation manners, such as a logic device or a combination of software and hardware, etc., that is, the execution subject of the following processing flow is not limited to each logic unit, but may also be hardware or a logic device.
Referring to fig. 19, in a software implementation, the adding device of the community member may include:
an acquisition unit 1901 that acquires a join request initiated by a user for a group, the personal information of the user including an original name and a specified single-tone element corresponding to the original name, the specified single-tone element having the same pronunciation as a multi-tone element in the original name;
and an adding unit 1902 for adding the user as a member of the community when the administrator of the community agrees to the joining request, so that the original name and the designated monophonic element are updated into the member information corresponding to the user.
Fig. 20 is a schematic structural diagram of an apparatus provided in an exemplary embodiment. Referring to fig. 20, at a hardware level, the device includes a processor 2002, an internal bus 2004, a network interface 2006, a memory 2008, and a nonvolatile memory 2010, and may include hardware required by other services. The processor 2002 reads a corresponding computer program from the nonvolatile memory 2010 into the memory 2008 and then executes the program, thereby forming a display device on a logic level. Of course, in addition to software implementation, one or more embodiments of the present disclosure do not exclude other implementation manners, such as a logic device or a combination of software and hardware, etc., that is, the execution subject of the following processing flow is not limited to each logic unit, but may also be hardware or a logic device.
Referring to fig. 21, in a software implementation, the display device may include:
an acquisition unit 2101 that acquires a content to be displayed, the content to be displayed including an original name of a target object and a specified single-tone element corresponding to the original name, the specified single-tone element having the same pronunciation as a multi-tone element in the original name;
and a display unit 2102 for displaying the content to be displayed.
The system, apparatus, module or unit set forth in the above embodiments may be implemented in particular by a computer chip or entity, or by a product having a certain function. A typical implementation device is a computer, which may be in the form of a personal computer, laptop computer, cellular telephone, camera phone, smart phone, personal digital assistant, media player, navigation device, email device, game console, tablet computer, wearable device, or a combination of any of these devices.
In a typical configuration, a computer includes one or more processors (CPUs), input/output interfaces, network interfaces, and memory.
The memory may include volatile memory in a computer-readable medium, random Access Memory (RAM) and/or nonvolatile memory, such as Read Only Memory (ROM) or flash memory (flash RAM). Memory is an example of computer-readable media.
Computer readable media, including both non-transitory and non-transitory, removable and non-removable media, may implement information storage by any method or technology. The information may be computer readable instructions, data structures, modules of a program, or other data. Examples of storage media for a computer include, but are not limited to, phase change memory (PRAM), static Random Access Memory (SRAM), dynamic Random Access Memory (DRAM), other types of Random Access Memory (RAM), read Only Memory (ROM), electrically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, read only compact disc read only memory (CD-ROM), digital Versatile Discs (DVD) or other optical storage, magnetic cassettes, magnetic disk storage, quantum memory, graphene-based storage or other magnetic storage devices, or any other non-transmission medium, which can be used to store information that can be accessed by the computing device. Computer-readable media, as defined herein, does not include transitory computer-readable media (transmission media), such as modulated data signals and carrier waves.
It should also be noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article or apparatus that comprises the element.
The foregoing describes specific embodiments of the present disclosure. Other embodiments are within the scope of the following claims. In some cases, the actions or steps recited in the claims can be performed in a different order than in the embodiments and still achieve desirable results. In addition, the processes depicted in the accompanying figures do not necessarily require the particular order shown, or sequential order, to achieve desirable results. In some embodiments, multitasking and parallel processing are also possible or may be advantageous.
The terminology used in the one or more embodiments of the specification is for the purpose of describing particular embodiments only and is not intended to be limiting of the one or more embodiments of the specification. As used in this specification, one or more embodiments and the appended claims, the singular forms "a," "an," and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It should also be understood that the term "and/or" as used herein refers to and encompasses any or all possible combinations of one or more of the associated listed items.
It should be understood that although the terms first, second, third, etc. may be used in one or more embodiments of the present description to describe various information, these information should not be limited to these terms. These terms are only used to distinguish one type of information from another. For example, first information may also be referred to as second information, and similarly, second information may also be referred to as first information, without departing from the scope of one or more embodiments of the present description. The word "if" as used herein may be interpreted as "at … …" or "at … …" or "responsive to a determination", depending on the context.
The foregoing description of the preferred embodiment(s) is (are) merely intended to illustrate the embodiment(s) of the present invention, and it is not intended to limit the embodiment(s) of the present invention to the particular embodiment(s) described.
Claims (17)
1. A voice broadcasting method, comprising:
identity recognition is carried out on the target object;
acquiring a voice file corresponding to the target object according to the identification result of the target object; when the original name of the target object contains a polyphone element, the voice file is generated by a replaced name, and the replaced name is obtained by replacing the polyphone element in the original name with a designated monophone element;
playing the acquired voice file;
the obtaining a voice file corresponding to the target object includes:
acquiring a replaced name of the target object;
generating the voice file according to the replaced name;
the obtaining the replaced name of the target object includes:
determining the original name of the target object according to the identification result of the target object; acquiring the replaced name of the target object according to the mapping relation between the locally stored original name and the replaced name; or alternatively, the first and second heat exchangers may be,
Determining the original name of the target object according to the identification result of the target object; and sending an inquiry request containing the original name to a server, and receiving a replaced name returned by the server, wherein the replaced name is determined by the server according to the mapping relation between the stored original name and the replaced name.
2. The method of claim 1, wherein the obtaining a voice file corresponding to the target object comprises:
and acquiring a pre-generated voice file corresponding to the target object.
3. The method as recited in claim 2, further comprising:
generating the voice file in advance according to the replaced name of the target object; or alternatively, the first and second heat exchangers may be,
and the receiving server pre-generates the voice file according to the replaced name of the target object.
4. The method as recited in claim 1, further comprising:
and displaying the original name of the target object.
5. The method of claim 1, wherein the identifying the target object comprises:
identifying the target object by at least one of the following modes: fingerprint recognition, face recognition, iris recognition and gait recognition.
6. A setting method of voice broadcast content, characterized by comprising:
initiating an object adding request to a server, wherein the object adding request is used for adding a target object as an identifiable object of an identity recognition device, and the original name of the target object contains a polyphone element;
acquiring a plurality of alternatives returned by the server, wherein the alternatives respectively correspond to the pronunciation of the multitone element;
determining a designated single-tone element corresponding to the selected alternative, wherein the designated single-tone element is used for generating a voice file corresponding to the target object so as to play the voice file to the target object after the target object is successfully identified by the identity identification equipment;
the generating a voice file corresponding to the target object includes:
acquiring a replaced name of the target object;
generating the voice file according to the replaced name;
the obtaining the replaced name of the target object includes:
determining the original name of the target object according to the identification result of the target object; acquiring the replaced name of the target object according to the mapping relation between the locally stored original name and the replaced name; or alternatively, the first and second heat exchangers may be,
Determining the original name of the target object according to the identification result of the target object; and sending an inquiry request containing the original name to a server, and receiving a replaced name returned by the server, wherein the replaced name is determined by the server according to the mapping relation between the stored original name and the replaced name.
7. A voice broadcasting method, comprising:
receiving an inquiry request from an identification device, wherein the inquiry request comprises an original name of a target object identified by the identification device, and the original name comprises a polyphone element;
returning a replaced name corresponding to the original name, so that the identification equipment generates a voice file according to the replaced name and plays the voice file; or, returning the voice file generated according to the replaced name to be played by the identity recognition equipment; wherein the replaced name is obtained by replacing the multitone element in the original name with a designated monophone element.
8. A setting method of voice broadcast content, characterized by comprising:
receiving an object adding request initiated by an administrator, wherein the object adding request is used for adding a target object as an identifiable object of identity recognition equipment, and the original name of the target object contains a polyphone element;
Returning a plurality of alternatives to the administrator, the plurality of alternatives corresponding to respective readings of the multitone element;
determining a designated single-tone element corresponding to the selected alternative, wherein the designated single-tone element is used for generating a voice file corresponding to the target object so as to play the voice file to the target object after the target object is successfully identified by the identity identification equipment;
establishing a mapping relation between the original name and the replaced name, and returning the replaced name or a voice file generated according to the replaced name when the identification equipment receives an inquiry request containing the original name; or alternatively, the first and second heat exchangers may be,
establishing a mapping relation between the original name and the voice file, so as to return to the voice file when the identification equipment receives an inquiry request containing the original name;
wherein the replaced name is obtained by replacing the multitone element in the original name with the specified monophone element.
9. The method of claim 8, wherein the returning to the administrator a plurality of alternatives comprises:
Matching each component element contained in the original name with a predefined multi-sound element library respectively so as to identify multi-sound elements contained in the original name and all pronunciations thereof;
matching each pronunciation of the identified multi-pronunciation element with a predefined single-pronunciation element library respectively to determine the single-pronunciation element corresponding to each pronunciation;
and returning the determined alternative options corresponding to the single-tone elements to the administrator.
10. A voice broadcasting device, comprising:
the identification unit is used for carrying out identity identification on the target object;
the acquisition unit acquires a voice file corresponding to the target object according to the identification result of the target object; when the original name of the target object contains a polyphone element, the voice file is generated by a replaced name, and the replaced name is obtained by replacing the polyphone element in the original name with a designated monophone element;
a playing unit for playing the acquired voice file;
the acquisition unit is specifically configured to:
acquiring a replaced name of the target object;
generating the voice file according to the replaced name;
the acquisition unit acquires the replaced name of the target object by:
Determining the original name of the target object according to the identification result of the target object; acquiring the replaced name of the target object according to the mapping relation between the locally stored original name and the replaced name; or alternatively, the first and second heat exchangers may be,
determining the original name of the target object according to the identification result of the target object; and sending an inquiry request containing the original name to a server, and receiving a replaced name returned by the server, wherein the replaced name is determined by the server according to the mapping relation between the stored original name and the replaced name.
11. A setting device for voice broadcast content, comprising:
a request unit for initiating an object addition request to a server, wherein the object addition request is used for adding a target object as an identifiable object of the identity recognition device, and the original name of the target object contains a polyphonic element;
an obtaining unit, configured to obtain a plurality of alternatives returned by the server, where the plurality of alternatives respectively correspond to each pronunciation of the multitone element;
a determining unit, configured to determine a designated single-tone element corresponding to the selected candidate, where the designated single-tone element is used to generate a voice file corresponding to the target object, so as to play the voice file to the target object after the target object is successfully identified by the identity identifying device;
The generating a voice file corresponding to the target object includes:
acquiring a replaced name of the target object;
generating the voice file according to the replaced name;
the obtaining the replaced name of the target object includes:
determining the original name of the target object according to the identification result of the target object; acquiring the replaced name of the target object according to the mapping relation between the locally stored original name and the replaced name; or alternatively, the first and second heat exchangers may be,
determining the original name of the target object according to the identification result of the target object; and sending an inquiry request containing the original name to a server, and receiving a replaced name returned by the server, wherein the replaced name is determined by the server according to the mapping relation between the stored original name and the replaced name.
12. A voice broadcasting device, comprising:
a receiving unit for receiving an inquiry request from an identification device, wherein the inquiry request comprises an original name of a target object identified by the identification device, and the original name comprises a polyphone element;
The returning unit returns the replaced name corresponding to the original name, so that the identification equipment generates a voice file according to the replaced name and plays the voice file; or, returning the voice file generated according to the replaced name to be played by the identity recognition equipment; wherein the replaced name is obtained by replacing the multitone element in the original name with a designated monophone element.
13. A setting device for voice broadcast content, comprising:
the receiving unit is used for receiving an object adding request initiated by an administrator, wherein the object adding request is used for adding a target object as an identifiable object of the identity recognition equipment, and the original name of the target object contains a polyphone element;
a return unit that returns a plurality of alternatives to the administrator, the plurality of alternatives corresponding to respective readings of the multitone element, respectively;
a determining unit, configured to determine a designated single-tone element corresponding to the selected candidate, where the designated single-tone element is used to generate a voice file corresponding to the target object, so as to play the voice file to the target object after the target object is successfully identified by the identity identifying device;
The processing unit establishes a mapping relation between the original name and the replaced name so as to return the replaced name or a voice file generated according to the replaced name when the identification equipment receives an inquiry request containing the original name; or alternatively, the first and second heat exchangers may be,
establishing a mapping relation between the original name and the voice file, so as to return to the voice file when the identification equipment receives an inquiry request containing the original name;
wherein the replaced name is obtained by replacing the multitone element in the original name with the specified monophone element.
14. An electronic device, comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor is configured to implement the method of any of claims 1-6 by executing the executable instructions.
15. A computer readable storage medium having stored thereon computer instructions which, when executed by a processor, implement the steps of the method according to any of claims 1-6.
16. An electronic device, comprising:
A processor;
a memory for storing processor-executable instructions;
wherein the processor is configured to implement the method of any of claims 7-9 by executing the executable instructions.
17. A computer readable storage medium having stored thereon computer instructions which, when executed by a processor, implement the steps of the method according to any of claims 7-9.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910613779.7A CN112291281B (en) | 2019-07-09 | 2019-07-09 | Voice broadcasting and voice broadcasting content setting method and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910613779.7A CN112291281B (en) | 2019-07-09 | 2019-07-09 | Voice broadcasting and voice broadcasting content setting method and device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112291281A CN112291281A (en) | 2021-01-29 |
CN112291281B true CN112291281B (en) | 2023-11-03 |
Family
ID=74418994
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910613779.7A Active CN112291281B (en) | 2019-07-09 | 2019-07-09 | Voice broadcasting and voice broadcasting content setting method and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112291281B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114781396B (en) * | 2022-04-27 | 2024-10-08 | 中国银行股份有限公司 | Method and device for processing polyphones during mobile phone bank screen reading |
Citations (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8510112B1 (en) * | 2006-08-31 | 2013-08-13 | At&T Intellectual Property Ii, L.P. | Method and system for enhancing a speech database |
CN103905294A (en) * | 2012-12-30 | 2014-07-02 | 北京神州泰岳软件股份有限公司 | Control method for adding instant messaging group members, server and client side equipment |
CN105786880A (en) * | 2014-12-24 | 2016-07-20 | 中兴通讯股份有限公司 | Voice recognition method, client and terminal device |
CN105844247A (en) * | 2016-03-25 | 2016-08-10 | 杭州信鸽金融信息服务股份有限公司 | Bi-camera cabinet machine and face recognition and second-generation ID card identification system |
US9548825B1 (en) * | 2013-12-19 | 2017-01-17 | Jelli, Inc. | Audio event triggers for broadcast |
CN106600767A (en) * | 2016-11-29 | 2017-04-26 | 深圳智乐信息科技有限公司 | Verification method and system based on identification card |
CN106815906A (en) * | 2016-11-29 | 2017-06-09 | 深圳智乐信息科技有限公司 | A kind of intelligent entrance guard management method and system based on optic communication |
CN106850794A (en) * | 2017-01-24 | 2017-06-13 | 北京百度网讯科技有限公司 | Method and apparatus, server and terminal for operating group |
CN107972028A (en) * | 2017-07-28 | 2018-05-01 | 北京物灵智能科技有限公司 | Man-machine interaction method, device and electronic equipment |
CN108363817A (en) * | 2018-03-21 | 2018-08-03 | 广东欧珀移动通信有限公司 | Document handling method, device, mobile terminal and computer readable storage medium |
CN108874356A (en) * | 2018-05-31 | 2018-11-23 | 珠海格力电器股份有限公司 | Voice broadcasting method and device, mobile terminal and storage medium |
CN108989341A (en) * | 2018-08-21 | 2018-12-11 | 平安科技(深圳)有限公司 | The autonomous register method of voice, device, computer equipment and storage medium |
CN109147831A (en) * | 2018-09-26 | 2019-01-04 | 深圳壹账通智能科技有限公司 | A kind of voice connection playback method, terminal device and computer readable storage medium |
CN109299471A (en) * | 2018-11-05 | 2019-02-01 | 广州百田信息科技有限公司 | A kind of method, apparatus and terminal of text matches |
CN109308892A (en) * | 2018-10-25 | 2019-02-05 | 百度在线网络技术(北京)有限公司 | Voice synthesized broadcast method, apparatus, equipment and computer-readable medium |
CN109325771A (en) * | 2018-09-20 | 2019-02-12 | 北京得意音通技术有限责任公司 | Auth method, device, computer program, storage medium and electronic equipment |
CN109461459A (en) * | 2018-12-07 | 2019-03-12 | 平安科技(深圳)有限公司 | Speech assessment method, apparatus, computer equipment and storage medium |
CN109545183A (en) * | 2018-11-23 | 2019-03-29 | 北京羽扇智信息科技有限公司 | Text handling method, device, electronic equipment and storage medium |
CN109616111A (en) * | 2018-12-24 | 2019-04-12 | 北京恒泰实达科技股份有限公司 | A kind of scene interactivity control method based on speech recognition |
CN109686361A (en) * | 2018-12-19 | 2019-04-26 | 深圳前海达闼云端智能科技有限公司 | A kind of method, apparatus of speech synthesis calculates equipment and computer storage medium |
CN109861904A (en) * | 2019-02-19 | 2019-06-07 | 天津字节跳动科技有限公司 | Nametags display methods and device |
CN109933215A (en) * | 2019-01-31 | 2019-06-25 | 平安科技(深圳)有限公司 | Phonetic transcriptions of Chinese characters conversion method, device, terminal and computer readable storage medium |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2048655B1 (en) * | 2007-10-08 | 2014-02-26 | Nuance Communications, Inc. | Context sensitive multi-stage speech recognition |
US8510103B2 (en) * | 2009-10-15 | 2013-08-13 | Paul Angott | System and method for voice recognition |
US10170112B2 (en) * | 2017-05-11 | 2019-01-01 | Google Llc | Detecting and suppressing voice queries |
-
2019
- 2019-07-09 CN CN201910613779.7A patent/CN112291281B/en active Active
Patent Citations (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8510112B1 (en) * | 2006-08-31 | 2013-08-13 | At&T Intellectual Property Ii, L.P. | Method and system for enhancing a speech database |
CN103905294A (en) * | 2012-12-30 | 2014-07-02 | 北京神州泰岳软件股份有限公司 | Control method for adding instant messaging group members, server and client side equipment |
US9548825B1 (en) * | 2013-12-19 | 2017-01-17 | Jelli, Inc. | Audio event triggers for broadcast |
CN105786880A (en) * | 2014-12-24 | 2016-07-20 | 中兴通讯股份有限公司 | Voice recognition method, client and terminal device |
CN105844247A (en) * | 2016-03-25 | 2016-08-10 | 杭州信鸽金融信息服务股份有限公司 | Bi-camera cabinet machine and face recognition and second-generation ID card identification system |
CN106600767A (en) * | 2016-11-29 | 2017-04-26 | 深圳智乐信息科技有限公司 | Verification method and system based on identification card |
CN106815906A (en) * | 2016-11-29 | 2017-06-09 | 深圳智乐信息科技有限公司 | A kind of intelligent entrance guard management method and system based on optic communication |
CN106850794A (en) * | 2017-01-24 | 2017-06-13 | 北京百度网讯科技有限公司 | Method and apparatus, server and terminal for operating group |
CN107972028A (en) * | 2017-07-28 | 2018-05-01 | 北京物灵智能科技有限公司 | Man-machine interaction method, device and electronic equipment |
CN108363817A (en) * | 2018-03-21 | 2018-08-03 | 广东欧珀移动通信有限公司 | Document handling method, device, mobile terminal and computer readable storage medium |
CN108874356A (en) * | 2018-05-31 | 2018-11-23 | 珠海格力电器股份有限公司 | Voice broadcasting method and device, mobile terminal and storage medium |
CN108989341A (en) * | 2018-08-21 | 2018-12-11 | 平安科技(深圳)有限公司 | The autonomous register method of voice, device, computer equipment and storage medium |
CN109325771A (en) * | 2018-09-20 | 2019-02-12 | 北京得意音通技术有限责任公司 | Auth method, device, computer program, storage medium and electronic equipment |
CN109147831A (en) * | 2018-09-26 | 2019-01-04 | 深圳壹账通智能科技有限公司 | A kind of voice connection playback method, terminal device and computer readable storage medium |
CN109308892A (en) * | 2018-10-25 | 2019-02-05 | 百度在线网络技术(北京)有限公司 | Voice synthesized broadcast method, apparatus, equipment and computer-readable medium |
CN109299471A (en) * | 2018-11-05 | 2019-02-01 | 广州百田信息科技有限公司 | A kind of method, apparatus and terminal of text matches |
CN109545183A (en) * | 2018-11-23 | 2019-03-29 | 北京羽扇智信息科技有限公司 | Text handling method, device, electronic equipment and storage medium |
CN109461459A (en) * | 2018-12-07 | 2019-03-12 | 平安科技(深圳)有限公司 | Speech assessment method, apparatus, computer equipment and storage medium |
CN109686361A (en) * | 2018-12-19 | 2019-04-26 | 深圳前海达闼云端智能科技有限公司 | A kind of method, apparatus of speech synthesis calculates equipment and computer storage medium |
CN109616111A (en) * | 2018-12-24 | 2019-04-12 | 北京恒泰实达科技股份有限公司 | A kind of scene interactivity control method based on speech recognition |
CN109933215A (en) * | 2019-01-31 | 2019-06-25 | 平安科技(深圳)有限公司 | Phonetic transcriptions of Chinese characters conversion method, device, terminal and computer readable storage medium |
CN109861904A (en) * | 2019-02-19 | 2019-06-07 | 天津字节跳动科技有限公司 | Nametags display methods and device |
Non-Patent Citations (4)
Title |
---|
Dynamically automated interactive voice response system for smart city surveillance;S S M Saqquaf等;《2016 IEEE International Conference on Recent Trends in Electronics ,Information & Communication Technology(RTEICT)》;全文 * |
周建国,臧亮,杨皓,周万江,晏蒲柳.基于Linux平台的电话语音邮件系统.武汉大学学报(理学版).2002,(01),全文. * |
基于HMM的单元挑选语音合成方法研究;何鑫;《中国优秀硕士学位论文全文数据库 信息科技辑》(第11期);全文 * |
张子荣,初敏.解决多音字字-音转换的一种统计学习方法.中文信息学报.2002,(03),全文. * |
Also Published As
Publication number | Publication date |
---|---|
CN112291281A (en) | 2021-01-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107430858B (en) | Communicating metadata identifying a current speaker | |
US20200326197A1 (en) | Method, apparatus, computer device and storage medium for determining poi alias | |
CN114787814B (en) | Reference resolution | |
US20190095973A1 (en) | Desirability of product attributes | |
US20230244730A1 (en) | Matchmaking Video Chatting Partners | |
JP2020528705A (en) | Moving video scenes using cognitive insights | |
CN110221959B (en) | Application program testing method, device and computer readable medium | |
US11106933B2 (en) | Method, device and system for processing image tagging information | |
CN109189879A (en) | E-book display methods and device | |
CN101884038A (en) | Associating metadata with media objects using time | |
CN106033418A (en) | A voice adding method and device, a voice play method and device, a picture classifying method and device, and a picture search method and device | |
CN111832579B (en) | Map interest point data processing method and device, electronic equipment and readable medium | |
CN110019644B (en) | Search method, apparatus and computer-readable storage medium in dialog implementation | |
CN112291281B (en) | Voice broadcasting and voice broadcasting content setting method and device | |
RU2608670C2 (en) | Method, device and terminal for prompt of sns system application | |
WO2017024684A1 (en) | User behavioral intent acquisition method, device and equipment, and computer storage medium | |
CN110874554A (en) | Action recognition method, terminal device, server, system and storage medium | |
CN117393091A (en) | Ward round method and device, mobile robot and storage medium | |
CN116186119A (en) | User behavior analysis method, device, equipment and storage medium | |
CN108268545B (en) | Method and device for establishing hierarchical user label library | |
US11755633B2 (en) | Entity search system | |
CN110781191B (en) | Processing method of layout data and server | |
CN110471708B (en) | Method and device for acquiring configuration items based on reusable components | |
CN110619086B (en) | Method and apparatus for processing information | |
CN116627789B (en) | Model detection method and device, electronic equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
REG | Reference to a national code |
Ref country code: HK Ref legal event code: DE Ref document number: 40045089 Country of ref document: HK |
|
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |