CN112291281A - Voice broadcast and voice broadcast content setting method and device - Google Patents

Voice broadcast and voice broadcast content setting method and device Download PDF

Info

Publication number
CN112291281A
CN112291281A CN201910613779.7A CN201910613779A CN112291281A CN 112291281 A CN112291281 A CN 112291281A CN 201910613779 A CN201910613779 A CN 201910613779A CN 112291281 A CN112291281 A CN 112291281A
Authority
CN
China
Prior art keywords
name
target object
polyphonic
original name
voice file
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910613779.7A
Other languages
Chinese (zh)
Other versions
CN112291281B (en
Inventor
何海斌
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nail Holding Cayman Co ltd
Original Assignee
Nail Holding Cayman Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nail Holding Cayman Co ltd filed Critical Nail Holding Cayman Co ltd
Priority to CN201910613779.7A priority Critical patent/CN112291281B/en
Publication of CN112291281A publication Critical patent/CN112291281A/en
Application granted granted Critical
Publication of CN112291281B publication Critical patent/CN112291281B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/12Protocols specially adapted for proprietary or special-purpose networking environments, e.g. medical networks, sensor networks, networks in vehicles or remote metering networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/3331Query processing
    • G06F16/334Query execution
    • G06F16/3343Query execution using phonetics
    • GPHYSICS
    • G07CHECKING-DEVICES
    • G07CTIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
    • G07C1/00Registering, indicating or recording the time of events or elapsed time, e.g. time-recorders for work people
    • G07C1/10Registering, indicating or recording the time of events or elapsed time, e.g. time-recorders for work people together with the recording, indicating or registering of other data, e.g. of signs of identity
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/06Protocols specially adapted for file transfer, e.g. file transfer protocol [FTP]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L9/00Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols
    • H04L9/32Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols including means for verifying the identity or authority of a user of the system or for message authentication, e.g. authorization, entity authentication, data integrity or data verification, non-repudiation, key authentication or verification of credentials
    • H04L9/3226Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols including means for verifying the identity or authority of a user of the system or for message authentication, e.g. authorization, entity authentication, data integrity or data verification, non-repudiation, key authentication or verification of credentials using a predetermined code, e.g. password, passphrase or PIN
    • H04L9/3231Biological data, e.g. fingerprint, voice or retina

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Security & Cryptography (AREA)
  • Acoustics & Sound (AREA)
  • Biomedical Technology (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Biodiversity & Conservation Biology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computing Systems (AREA)
  • Medical Informatics (AREA)
  • Telephonic Communication Services (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

One or more embodiments of the present specification provide a voice broadcast and a method and an apparatus for setting a voice broadcast content, where the voice broadcast method may include: performing identity recognition on the target object; acquiring a voice file corresponding to the target object according to the identification result of the target object; when the original name of the target object contains polyphonic elements, the voice file is generated by a replaced name, and the replaced name is obtained by replacing the polyphonic elements in the original name with specified single-tone elements; and playing the acquired voice file.

Description

Voice broadcast and voice broadcast content setting method and device
Technical Field
One or more embodiments of the present disclosure relate to the field of identity recognition technologies, and in particular, to a method and an apparatus for voice broadcast and setting voice broadcast content.
Background
In some scenarios, identification of the relevant objects is required. For example, the visitor is identified by the access control device to determine whether to permit, or the employee is identified by the attendance device to record attendance data, etc. After the identification is successful, the name of the related object can be broadcasted through voice so as to indicate that the identity identification is successful.
Disclosure of Invention
In view of this, one or more embodiments of the present disclosure provide a method and an apparatus for voice broadcast and setting of voice broadcast content.
To achieve the above object, one or more embodiments of the present disclosure provide the following technical solutions:
according to a first aspect of one or more embodiments of the present specification, there is provided a voice broadcast method including:
performing identity recognition on the target object;
acquiring a voice file corresponding to the target object according to the identification result of the target object; when the original name of the target object contains polyphonic elements, the voice file is generated by a replaced name, and the replaced name is obtained by replacing the polyphonic elements in the original name with specified single-tone elements;
and playing the acquired voice file.
According to a second aspect of one or more embodiments of the present specification, there is provided a method of setting a voice broadcast content, including:
initiating an object adding request to a server, wherein the object adding request is used for adding a target object as a recognizable object of an identity recognition device, and an original name of the target object contains a polyphonic element;
acquiring a plurality of alternative items returned by the server, wherein the alternative items respectively correspond to the pronunciations of the polyphonic elements;
determining a designated tone element corresponding to the selected alternative, wherein the designated tone element is used for generating a voice file corresponding to the target object, and playing the voice file to the target user after the target user is successfully recognized by the identity recognition device.
According to a third aspect of one or more embodiments of the present specification, there is provided a voice broadcast method including:
receiving an inquiry request from an identity recognition device, wherein the inquiry request comprises an original name of a target object recognized by the identity recognition device, and the original name comprises a polyphonic element;
returning the replaced name corresponding to the original name so that the identity recognition equipment generates a voice file according to the replaced name and plays the voice file; or returning the voice file generated according to the replaced name to be played by the identity recognition equipment; wherein the replaced name is obtained by replacing a polyphonic element in the original name with a designated monophonic element.
According to a fourth aspect of one or more embodiments of the present specification, there is provided a method of setting a voice broadcast content, including:
receiving an object adding request initiated by an administrator, wherein the object adding request is used for adding a target object as a recognizable object of an identity recognition device, and an original name of the target object contains a polyphonic element;
returning a plurality of alternatives to the administrator, the plurality of alternatives corresponding to respective readings of the polyphonic elements;
determining a designated tone element corresponding to the selected alternative, wherein the designated tone element is used for generating a voice file corresponding to the target object, and playing the voice file to the target user after the target user is successfully recognized by the identity recognition device.
According to a fifth aspect of one or more embodiments of the present specification, there is provided a community member adding method including:
acquiring a joining request initiated by a user aiming at a group, wherein personal information of the user comprises an original name and a specified single-tone element corresponding to the original name, and the specified single-tone element has the same pronunciation as a polyphonic element in the original name;
and when the administrator of the community agrees to the joining request, adding the user as a member of the community, so that the original name and the specified single-tone element are updated to the member information corresponding to the user.
According to a sixth aspect of one or more embodiments of the present specification, there is provided a display method including:
acquiring content to be displayed, wherein the content to be displayed comprises an original name of a target object and a specified single-tone element corresponding to the original name, and the specified single-tone element and a multi-tone element in the original name have the same pronunciation;
and displaying the content to be displayed.
According to a seventh aspect of one or more embodiments of the present specification, there is provided a voice broadcasting device including:
the identification unit identifies the target object;
the acquisition unit is used for acquiring a voice file corresponding to the target object according to the identification result of the target object; when the original name of the target object contains polyphonic elements, the voice file is generated by a replaced name, and the replaced name is obtained by replacing the polyphonic elements in the original name with specified single-tone elements;
and the playing unit plays the acquired voice file.
According to an eighth aspect of one or more embodiments of the present specification, there is provided a setting apparatus of a voice broadcast content, including:
the system comprises a request unit and a control unit, wherein the request unit initiates an object adding request to a server, the object adding request is used for adding a target object as a recognizable object of an identity recognition device, and the original name of the target object contains polyphonic elements;
the acquisition unit is used for acquiring a plurality of alternative items returned by the server, wherein the alternative items respectively correspond to the pronunciations of the polyphonic elements;
and the determining unit is used for determining a specified tone element corresponding to the selected alternative, wherein the specified tone element is used for generating a voice file corresponding to the target object so as to play the voice file to the target user after the target user is successfully identified by the identity recognition device.
According to a ninth aspect of one or more embodiments of the present specification, there is provided a voice broadcasting device including:
the system comprises a receiving unit, a processing unit and a processing unit, wherein the receiving unit is used for receiving an inquiry request from an identity recognition device, the inquiry request comprises an original name of a target object recognized by the identity recognition device, and the original name comprises a polyphonic element;
the return unit returns the replaced name corresponding to the original name so that the identity recognition equipment generates a voice file according to the replaced name and plays the voice file; or returning the voice file generated according to the replaced name to be played by the identity recognition equipment; wherein the replaced name is obtained by replacing a polyphonic element in the original name with a designated monophonic element.
According to a tenth aspect of one or more embodiments of the present specification, there is provided a setting apparatus of a voice broadcast content, including:
the system comprises a receiving unit, a judging unit and a processing unit, wherein the receiving unit is used for receiving an object adding request initiated by an administrator, the object adding request is used for adding a target object as a recognizable object of an identity recognition device, and the original name of the target object contains polyphonic elements;
a returning unit that returns a plurality of alternatives to the administrator, the plurality of alternatives corresponding to respective readings of the polyphonic elements, respectively;
and the determining unit is used for determining a specified tone element corresponding to the selected alternative, wherein the specified tone element is used for generating a voice file corresponding to the target object so as to play the voice file to the target user after the target user is successfully identified by the identity recognition device.
According to an eleventh aspect of one or more embodiments of the present specification, there is provided a community member adding apparatus including:
the system comprises an acquisition unit, a processing unit and a processing unit, wherein the acquisition unit is used for acquiring a joining request initiated by a user aiming at a group, the personal information of the user comprises an original name and a specified single-tone element corresponding to the original name, and the specified single-tone element has the same pronunciation with a polyphonic element in the original name;
and the adding unit is used for adding the user as a member of the community when the administrator of the community agrees to the joining request, so that the original name and the specified single-tone element are updated to the member information corresponding to the user.
According to a twelfth aspect of one or more embodiments of the present specification, there is provided a display device including:
an acquisition unit that acquires a content to be displayed, the content to be displayed including an original name of a target object and a specified monophonic element corresponding to the original name, the specified monophonic element having the same pronunciation as a polyphonic element in the original name;
and the display unit is used for displaying the content to be displayed.
According to a thirteenth aspect of one or more embodiments of the present specification, there is provided an electronic apparatus, including:
a processor;
a memory for storing processor-executable instructions;
wherein the processor implements the method of the first aspect by executing the executable instructions.
According to a fourteenth aspect of one or more embodiments of the present specification, a computer-readable storage medium is presented, having stored thereon computer instructions which, when executed by a processor, implement the steps of the method according to the first aspect.
According to a fifteenth aspect of one or more embodiments herein, there is provided an electronic device comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor implements the method according to the second aspect by executing the executable instructions.
According to a sixteenth aspect of one or more embodiments of the present description, there is provided a computer readable storage medium having stored thereon computer instructions which, when executed by a processor, implement the steps of the method according to the second aspect.
According to a seventeenth aspect of one or more embodiments of the present specification, there is provided an electronic device comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor implements the method according to the third aspect by executing the executable instructions.
According to an eighteenth aspect of one or more embodiments of the present specification, a computer-readable storage medium is presented, on which computer instructions are stored, which instructions, when executed by a processor, implement the steps of the method according to the third aspect.
According to a nineteenth aspect of one or more embodiments of the present specification, there is provided an electronic apparatus comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor implements the method of the fourth aspect by executing the executable instructions.
According to a twentieth aspect of one or more embodiments of the present specification, there is provided a computer-readable storage medium having stored thereon computer instructions which, when executed by a processor, implement the steps of the method according to the fourth aspect.
According to a twenty-first aspect of one or more embodiments of the present specification, there is provided an electronic device comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor implements the method according to the fifth aspect by executing the executable instructions.
According to a twenty-second aspect of one or more embodiments of the present description, there is provided a computer-readable storage medium having stored thereon computer instructions which, when executed by a processor, implement the steps of the method according to the fifth aspect.
According to a twenty-third aspect of one or more embodiments of the present specification, there is provided an electronic apparatus comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor implements the method of the sixth aspect by executing the executable instructions.
According to a twenty-fourth aspect of one or more embodiments of the present specification, a computer-readable storage medium is presented, having stored thereon computer instructions which, when executed by a processor, carry out the steps of the method according to the sixth aspect.
Drawings
Fig. 1 is a schematic diagram of an architecture of an identification system according to an exemplary embodiment.
Fig. 2 is a flowchart of a voice broadcast method according to an exemplary embodiment.
Fig. 3 is a flowchart of another voice broadcasting method according to an exemplary embodiment.
Fig. 4 is a flowchart of a method for setting a voice broadcast content according to an exemplary embodiment.
Fig. 5 is a flowchart of another method for setting a voice broadcast content according to an exemplary embodiment.
Fig. 6 is an interaction diagram for adding recognizable objects to the attendance checking device and performing attendance checking operations according to an exemplary embodiment.
FIG. 7 is a schematic diagram of a pronunciation selection interface provided by an exemplary embodiment.
Fig. 8 is a flowchart of a community member adding method according to an exemplary embodiment.
FIG. 9 is a flow chart of a display method provided by an exemplary embodiment.
Fig. 10 is a schematic diagram of an apparatus according to an exemplary embodiment.
Fig. 11 is a block diagram of a voice broadcast apparatus according to an exemplary embodiment.
Fig. 12 is a schematic structural diagram of an apparatus according to a second exemplary embodiment.
Fig. 13 is a block diagram of a device for setting a content of a voice broadcast according to an exemplary embodiment.
Fig. 14 is a schematic structural diagram of an apparatus according to another exemplary embodiment.
Fig. 15 is a block diagram of another voice announcement device according to an exemplary embodiment.
Fig. 16 is a schematic structural diagram of an apparatus according to the fourth exemplary embodiment.
Fig. 17 is a block diagram of another apparatus for setting a content of a voice broadcast according to an exemplary embodiment.
Fig. 18 is a schematic structural diagram of an apparatus according to the fifth exemplary embodiment.
Fig. 19 is a block diagram of an adding apparatus of a community member according to an exemplary embodiment.
Fig. 20 is a schematic structural diagram of an apparatus according to a sixth exemplary embodiment.
Fig. 21 is a block diagram of a display device according to an exemplary embodiment.
Detailed Description
Reference will now be made in detail to the exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, like numbers in different drawings represent the same or similar elements unless otherwise indicated. The implementations described in the following exemplary embodiments do not represent all implementations consistent with one or more embodiments of the present specification. Rather, they are merely examples of apparatus and methods consistent with certain aspects of one or more embodiments of the specification, as detailed in the claims which follow.
It should be noted that: in other embodiments, the steps of the corresponding methods are not necessarily performed in the order shown and described herein. In some other embodiments, the method may include more or fewer steps than those described herein. Moreover, a single step described in this specification may be broken down into multiple steps for description in other embodiments; multiple steps described in this specification may be combined into a single step in other embodiments.
Fig. 1 is a schematic diagram of an architecture of an identification system according to an exemplary embodiment. As shown in fig. 1, the system may include a server 11, a network 12, a number of electronic devices such as a cell phone 13, a cell phone 14, a cell phone 15, and the like, and an identification device 16.
The server 11 may be a physical server comprising a separate host, or the server 11 may be a virtual server carried by a cluster of hosts. In the operation process, the server 11 may operate a server-side program of a certain application to implement a related service function of the application, for example, when the server 11 operates a program of a mobile group office platform, the server may be implemented as a server of the group office platform.
The handsets 13-15 are just one type of electronic device that a user may use. In fact, it is obvious that the user can also use electronic devices of the type such as: tablet devices, notebook computers, Personal Digital Assistants (PDAs), wearable devices (e.g., smart glasses, smart watches, etc.), etc., which are not limited by one or more embodiments of the present disclosure. During the operation process, the electronic device may operate a client-side program of an application to implement a relevant service function of the application, for example, when the electronic device operates a program of a group office platform, the electronic device may be implemented as a client of the group office platform.
The identification device 16 may be used to implement identification functions, such as fingerprint recognition, face recognition, iris recognition, gait recognition, and other one or more types of physiological feature recognition technologies, which is not limited in this specification. The identification device 16 can be used for implementing various scenes such as access control and attendance management by implementing identification, which is not limited in the specification. During operation, the identification device 16 may run a client-side program of an application to implement relevant business functions of the application, for example, when the identification device 16 runs a program of a group office platform, the application may be implemented as a client of the group office platform.
The group office platform may not only implement a communication function, but also be an integrated functional platform with many other functions, such as processing internal events of a group, such as an approval event (e.g., an approval event such as leave request, office article application, finance, etc.), an attendance event, a task event, a log event, etc., and processing external events of a group, such as ordering, purchasing, etc., which are not limited in one or more embodiments of the present specification. Through the server running on the server 11, and the client running on the electronic equipment such as the mobile phones 13 to 15 and the identity recognition equipment 16, data interaction can be realized among the electronic equipment such as the server 11, the mobile phones 13 to 15 and the identity recognition equipment 16. For example, the server 11 may maintain organization structure information of a group, and electronic devices such as mobile phones 13 to 15 may be used to manage the organization structure information; and, information of the community members included in the organization architecture can be imported from the server 11 into the identification device 16 in a batch, so that the identification device 16 can identify the members in the community. The group office platform can be suitable for a mobile office scene, namely the group office platform can be a mobile group office platform.
The network 12 for interaction between the handsets 13-15, the identification device 16 and the server 11 may include various types of wired or wireless networks.
Fig. 2 is a flowchart of a voice broadcast method according to an exemplary embodiment. As shown in fig. 2, the method applied to an identification device (e.g., the identification device 16 shown in fig. 1, etc.) may include the following steps:
step 202, performing identity recognition on the target object.
In an embodiment, the identity recognition device is configured to perform feature acquisition on a target object, and the acquired feature information is used to perform identity recognition on the target object. The type of feature information is related to the identification technology used by the identification device, for example, when fingerprint identification technology is used, the feature information may include fingerprint features, and when face identification technology is used, the feature information may include face features.
In an embodiment, the identity recognition device may perform identity recognition on the target object locally according to the collected feature information to obtain an identity recognition result. Or the identity recognition equipment can upload the collected characteristic information to the server, and the server performs identity recognition on the target object and then returns an identity recognition result to the identity recognition equipment; the server can operate the server side of the group office platform, the identity recognition device can operate the client side of the group office platform, and the server side and the client side can transmit characteristic information, identity recognition results and the like based on the client side and the server side.
Step 204, acquiring a voice file corresponding to the target object according to the identification result of the target object; wherein, when the original name of the target object contains polyphonic elements, the voice file is generated from a replaced name obtained by replacing the polyphonic elements in the original name with designated monophonic elements.
In one embodiment, the name is composed of several elements, the element type being related to the type of text the name takes. For example, when chinese is used, the elements may be chinese characters, and when english is used, the elements may be words, which are not listed one by one. Polyphonic elements refer to elements having multiple pronunciations; taking the chinese language as an example, polyphonic elements are polyphonic characters in the chinese language, for example, the pronunciation of the Chinese character "madder" includes "qi a n" and "x ī". Similarly, a monophonic element refers to an element that has only one pronunciation; taking Chinese as an example, the single tone element is a single tone character in Chinese, for example, the pronunciation of the Chinese character "West" is only "x ī".
In one embodiment, a multi-tone element library containing all multi-tone elements and a single-tone element library containing all single-tone elements can be generated in advance for the type of text used by the original name. Then, by respectively matching all elements contained in the original name with a polyphonic element library, the polyphonic elements contained in the original name and all the pronunciations of the polyphonic elements can be determined; further, by matching each pronunciation with the phone element library, the phone element corresponding to each pronunciation can be selected. For example, when the original name of the target object is "yang aliza", by matching the Chinese character "yang" and the Chinese character "aliza" with a polyphonic element library (i.e., polyphonic character library), respectively, it can be determined that the Chinese character "aliza" is a polyphonic character, and there are pronunciations "qi" n "and" x ī "; further, by matching the readings "qi a n" and "x ī" with the monophonic element library (i.e., the monophonic character library), respectively, the monophonic character "qian" with the reading "qi a n" and the monophonic character "west" with the reading "x ī" can be determined.
Although the original name contains polyphonic elements, the original name of the target object only needs to adopt a fixed pronunciation, so that the above-mentioned replaced name can be generated by setting the monophonic element adopting the pronunciation as the above-mentioned specified monophonic element according to the pronunciation actually adopted by the target object for the polyphonic elements contained in the original name. For example, when the original name of the target object is actually read as "y-ng x ī", the polyphonic element "mada" may be replaced with the designated monophonic element "west" of the reading "x ī", with the resulting replacement named "yang west". Specifically, the target object or the administrator may select the actual pronunciation of the polyphonic element included in the original name by providing the target object or the administrator with alternatives respectively corresponding to the respective pronunciations of the polyphonic element.
In one embodiment, by generating a corresponding replaced name for an original name containing a polyphonic element and generating a voice file based on the replaced name, the voice file can be directly generated without adding a special mark for the actual reading of the polyphonic element (for example, the mark [ qian4] indicates that the reading of "madder" is "qi" or the like), which has extremely high versatility compared to different mark formats that may be adopted by different manufacturers, and avoids parsing the mark, increasing complexity and processing computation when generating the voice file.
In one embodiment, the identification device may obtain the replaced name of the target object, and generate the voice file locally according to the replaced name. Therefore, the technical scheme of the specification does not need to add a mark, so that the voice file can be smoothly and timely generated by the identity recognition device only with relatively low processing performance, and the voice broadcasting delay and even the normal operation of the identity recognition device cannot be influenced due to insufficient processing performance.
In one embodiment, the original names and the replaced names of all the objects are stored on the identity recognition device. When the administrator enters or imports the identity information of all objects on the identity recognition device, the identity recognition device may recognize the polyphone elements included in the original name based on the polyphone element library, determine the monophone element corresponding to each reading respectively based on the monophone element library, provide the administrator or the related object with the alternatives corresponding to each reading or monophone element respectively, and then generate the corresponding replaced name according to the designated monophone element corresponding to the selected alternative, thereby forming the mapping relationship between the corresponding original name and the replaced name on the identity recognition device. Alternatively, the administrator may set up a mapping relationship between the original name and the replaced name of each object in advance, and then import the mapping relationship into the identification device. Or, the server may identify polyphonic elements included in the original name based on the polyphonic element library, determine a monophonic element corresponding to each reading based on the monophonic element library, provide alternatives corresponding to each reading or monophonic element to an administrator or a related object, generate a corresponding replaced name according to a designated monophonic element corresponding to the selected alternatives, thereby forming a mapping relationship between the corresponding original name and the replaced name, and send the mapping relationship to the identification device. Accordingly, the identification device may determine the original name of the target object according to the identification result of the target object, and then obtain the replaced name of the target object according to the mapping relationship between the locally stored original name and the replaced name, so as to generate a corresponding voice file based on the replaced name.
In one embodiment, the server stores the mapping relationship between the original name and the replaced name of all the objects. For example, the server may identify polyphone elements included in the original name based on the polyphone element library, determine a phone element corresponding to each reading respectively based on the phone element library, provide an option corresponding to each reading or phone element to an administrator or a related object, and generate a corresponding replaced name according to a designated phone element corresponding to the selected option, so as to form and store a mapping relationship between the corresponding original name and the replaced name. Accordingly, the identification device may determine an original name of the target object according to the identification result of the target object, then send an inquiry request including the original name to the server, and receive a replaced name returned by the server, where the replaced name is determined by the server according to the stored mapping relationship between the original name and the replaced name.
In an embodiment, the identity recognition device may obtain a pre-generated voice file corresponding to the target object, and need not generate the voice file temporarily after recognizing the identity information of the target object, so as to avoid feedback delay caused by the process of generating the voice file, and facilitate acceleration of the response speed of the identity recognition device.
In an embodiment, the identity recognition device may generate a corresponding voice file in advance according to the replaced name of the target object, and store the voice file in a local or cloud storage space. Similarly, the identification device may generate, in advance, the voice files corresponding to all the objects, and store the voice files in the local or cloud storage space.
In an embodiment, the identity recognition device may receive a voice file generated by the server in advance according to the replaced name of the target object; the identity recognition device can temporarily request the server to acquire the voice file after confirming the identity information of the target object, or acquire the voice file in advance and store the voice file in a local or cloud storage space, and then read the voice file from the local or cloud storage space. Similarly, the identification device may obtain, from the server, the voice files respectively corresponding to all the objects generated in advance.
In an embodiment, for a pre-generated voice file, a mapping relationship can be directly established between an original name and the voice file, so that the corresponding voice file can be directly determined according to the original name of a target object. Or, a mapping relationship may be established between the replaced name and the voice file, and then, in combination with the mapping relationship between the original name and the replaced name, after the original name is determined, the corresponding voice file may also be determined.
In one embodiment, the mapping relationship (the mapping relationship between the original name and the replaced name, or the mapping relationship between the original name and the voice file, or the mapping relationship between the replaced name and the voice file) is not applicable to all objects recognized by the identification device, but is bound to one or more set objects. For example, for a user whose original name is "yangqi" and whose other user is "y-ng x ī", the sound of one user is "yang qi" and the sound of the other user is "y-ng x ī", so that for each user whose original name contains polyphonic elements, a corresponding mapping should be generated separately to ensure the correctness of the voice broadcast.
When the identity recognition device locally recognizes the identity information of the object, generates the mapping relation, matches the mapping relation and generates the voice file, the identity recognition device can independently realize the technical scheme of the specification under the condition of no need of a server, and the correct broadcast of the multi-tone element is realized. Alternatively, the identification device may cooperate with the server, such as by implementing one or more steps of generating a mapping relationship, matching a mapping relationship, generating a voice file, etc. on the server, which may reduce the processing and computing pressure of the identification device and reduce the performance requirements of the identification device.
And step 206, playing the acquired voice file.
In one embodiment, after the identity information of the target object is recognized, the name of the target object is correctly broadcasted by playing a voice file through the identity recognition device, so that the target object can quickly determine that the identity recognition device has successfully recognized the identity information of the target object.
In one embodiment, in addition to playing the voice file, the identification device may also display the original name of the target object, which may help the target object determine that the identification device has successfully identified its own identity information. Similarly, other identity information of the target object may also be presented; for example, in an enterprise scenario, the other identity information may include employee job numbers, affiliated departments, positions, and the like, and for example, in a school scenario, the other identity information may include affiliated classes, and the like.
Fig. 3 is a flowchart of another voice broadcasting method according to an exemplary embodiment. As shown in fig. 3, the method applied to a server (e.g., the server 11 shown in fig. 1) may include the following steps:
step 302, receiving an inquiry request from an identification device, where the inquiry request includes an original name of a target object identified by the identification device, and the original name includes a polyphonic element.
In an embodiment, the identity recognition device is configured to perform feature acquisition on a target object, and the acquired feature information is used to perform identity recognition on the target object. The type of feature information is related to the identification technology used by the identification device, for example, when fingerprint identification technology is used, the feature information may include fingerprint features, and when face identification technology is used, the feature information may include face features.
In an embodiment, the identity recognition device may perform identity recognition on the target object locally according to the collected feature information to obtain an identity recognition result. Or the identity recognition equipment can upload the collected characteristic information to the server, and the server performs identity recognition on the target object and then returns an identity recognition result to the identity recognition equipment; the server can operate the server side of the group office platform, the identity recognition device can operate the client side of the group office platform, and the server side and the client side can transmit characteristic information, identity recognition results and the like based on the client side and the server side.
Step 304, returning the replaced name corresponding to the original name, so that the identity recognition device generates a voice file according to the replaced name and plays the voice file; or returning the voice file generated according to the replaced name to be played by the identity recognition equipment; wherein the replaced name is obtained by replacing a polyphonic element in the original name with a designated monophonic element.
In one embodiment, the name is composed of several elements, the element type being related to the type of text the name takes. For example, when chinese is used, the elements may be chinese characters, and when english is used, the elements may be words, which are not listed one by one. Polyphonic elements refer to elements having multiple pronunciations; taking the chinese language as an example, polyphonic elements are polyphonic characters in the chinese language, for example, the pronunciation of the Chinese character "madder" includes "qi a n" and "x ī". Similarly, a monophonic element refers to an element that has only one pronunciation; taking Chinese as an example, the single tone element is a single tone character in Chinese, for example, the pronunciation of the Chinese character "West" is only "x ī".
In one embodiment, a multi-tone element library containing all multi-tone elements and a single-tone element library containing all single-tone elements can be generated in advance for the type of text used by the original name. Then, by respectively matching all elements contained in the original name with a polyphonic element library, the polyphonic elements contained in the original name and all the pronunciations of the polyphonic elements can be determined; further, by matching each pronunciation with the phone element library, the phone element corresponding to each pronunciation can be selected. For example, when the original name of the target object is "yang aliza", by matching the Chinese character "yang" and the Chinese character "aliza" with a polyphonic element library (i.e., polyphonic character library), respectively, it can be determined that the Chinese character "aliza" is a polyphonic character, and there are pronunciations "qi" n "and" x ī "; further, by matching the readings "qi a n" and "x ī" with the monophonic element library (i.e., the monophonic character library), respectively, the monophonic character "qian" with the reading "qi a n" and the monophonic character "west" with the reading "x ī" can be determined.
Although the original name contains polyphonic elements, the original name of the target object only needs to adopt a fixed pronunciation, so that the above-mentioned replaced name can be generated by setting the monophonic element adopting the pronunciation as the above-mentioned specified monophonic element according to the pronunciation actually adopted by the target object for the polyphonic elements contained in the original name. For example, when the original name of the target object is actually read as "y-ng x ī", the polyphonic element "mada" may be replaced with the designated monophonic element "west" of the reading "x ī", with the resulting replacement named "yang west". Specifically, the target object or the administrator may select the actual pronunciation of the polyphonic element included in the original name by providing the target object or the administrator with alternatives respectively corresponding to the respective pronunciations of the polyphonic element.
In one embodiment, by generating a corresponding replaced name for an original name containing a polyphonic element and generating a voice file based on the replaced name, the voice file can be directly generated without adding a special mark for the actual reading of the polyphonic element (for example, the mark [ qian4] indicates that the reading of "madder" is "qi" or the like), which has extremely high versatility compared to different mark formats that may be adopted by different manufacturers, and avoids parsing the mark, increasing complexity and processing computation when generating the voice file.
In one embodiment, the server stores a mapping relationship between the original name and the replaced name of all objects (i.e., objects whose original names contain polyphonic elements, the same below). For example, the server may identify polyphone elements included in the original name based on the polyphone element library, determine a phone element corresponding to each reading respectively based on the phone element library, provide an option corresponding to each reading or phone element to an administrator or a related object, and generate a corresponding replaced name according to a designated phone element corresponding to the selected option, so as to form and store a mapping relationship between the corresponding original name and the replaced name. Accordingly, the identification device may determine an original name of the target object according to the identification result of the target object, and then send an inquiry request including the original name to the server, and the server may return a corresponding replaced name based on the mapping relationship, so that the identification device may generate a corresponding voice file based on the replaced name.
In an embodiment, the server may generate a voice file corresponding to the target object in advance. Then, after the identity recognition device confirms the identity information of the target object, the original name of the target object can be sent to the server, so that the server returns a corresponding voice file to the identity recognition device. Because the voice file is not required to be temporarily generated, the response speed of the identity recognition equipment can be increased.
In an embodiment, for a pre-generated voice file, a corresponding relationship can be directly established between an original name and the voice file, so that the corresponding voice file can be directly determined according to the original name of the target object. Or, a corresponding relationship may be established between the replaced name and the voice file, and then, by combining the mapping relationship between the original name and the replaced name, after the original name is determined, the corresponding voice file may also be determined.
In one embodiment, the mapping relationship is not applicable to all objects identified by the identification device, but is bound to one or more set objects. For example, for a user whose original name is "yangqi" and whose other user is "y-ng x ī", the sound of one user is "yang qi" and the sound of the other user is "y-ng x ī", so that for each user whose original name contains polyphonic elements, a corresponding mapping should be generated separately to ensure the correctness of the voice broadcast.
In one embodiment, the identity recognition device cooperates with the server, such as by implementing one or more steps of generating a mapping relationship, matching a mapping relationship, generating a voice file, etc. on the server, the processing and computing pressure of the identity recognition device can be reduced, and the performance requirements of the identity recognition device can be reduced.
In an embodiment, the replaced name or the voice file is returned to the identity recognition device, so that after the identity recognition device recognizes the identity information of the target object, the voice file can be played to correctly broadcast the name of the target object, and the target object can quickly determine that the identity recognition device successfully recognizes the identity information of the target object.
Fig. 4 is a flowchart of a method for setting a voice broadcast content according to an exemplary embodiment. As shown in fig. 4, the method is applied to an electronic device (e.g., the mobile phones 13 to 15 shown in fig. 1) or an identification device (e.g., the identification device 16 shown in fig. 1), and may include the following steps:
step 402, initiating an object adding request to a server, wherein the object adding request is used for adding a target object as a recognizable object of an identity recognition device, and an original name of the target object contains a polyphonic element.
In an embodiment, an object addition request may be initiated to the server by the electronic device, and the object addition request may be added to the server as a recognizable object of the identification device by the target object alone or by adding a plurality of objects including the target object as recognizable objects of the identification device. For example, a server of the group office platform may be run on the server, and a client of the group office platform may be run on the electronic device and the identification device, so that data interaction may be implemented between the server, the electronic device, and the identification device based on the server and the client.
Specifically, the electronic device may initiate an object addition request to the server, and accordingly notify the server of the identity information of the target object to be added; then, the server can issue a collection instruction to the identification device, so that the identification device collects the standard characteristic information of the target object. The standard characteristic information and the identity information of the target object can be bound and stored on the identity recognition equipment, so that the identity recognition equipment can perform identity recognition on the target object according to the standard characteristic information and the identity information to generate an identity recognition result; or, the identity recognition device may upload the standard characteristic information to the server, and the server binds and stores the standard characteristic information and the identity information of the target object, and when the identity recognition device performs identity recognition on the target object in the following, the identity recognition device needs to acquire the characteristic information of the target object and upload the characteristic information to the server, and the server completes identity recognition and informs the identity recognition device of an identity recognition result.
In an embodiment, the identification device may directly initiate an object addition request to the server, where the object addition request may be added to the identification device as a recognizable object of the identification device by the target object alone, or may be added to each of a plurality of objects including the target object as recognizable objects of the identification device. For example, a server of the group office platform can be operated on the server, and a client of the group office platform can be operated on the identification device, so that data interaction between the server and the identification device can be realized based on the server and the client.
Specifically, the identification device may collect standard characteristic information of the target object. The standard characteristic information and the identity information of the target object can be bound and stored on the identity recognition equipment, so that the identity recognition equipment can perform identity recognition on the target object according to the standard characteristic information and the identity information to generate an identity recognition result; meanwhile, the identity recognition device can upload the identity information of the target object to the server, so that the server knows that the identity recognition device can recognize the identity of the target object. Or, the identity recognition device may upload the standard characteristic information to the server, and the server binds and stores the standard characteristic information and the identity information of the target object, and when the identity recognition device performs identity recognition on the target object in the following, the identity recognition device needs to acquire the characteristic information of the target object and upload the characteristic information to the server, and the server completes identity recognition and informs the identity recognition device of an identity recognition result.
Step 404, obtaining a plurality of candidates returned by the server, where the plurality of candidates respectively correspond to the pronunciations of the polyphonic elements.
In one embodiment, the name is composed of several elements, the element type being related to the type of text the name takes. For example, when chinese is used, the elements may be chinese characters, and when english is used, the elements may be words, which are not listed one by one. Polyphonic elements refer to elements having multiple pronunciations; taking the chinese language as an example, polyphonic elements are polyphonic characters in the chinese language, for example, the pronunciation of the Chinese character "madder" includes "qi a n" and "x ī". Similarly, a monophonic element refers to an element that has only one pronunciation; taking Chinese as an example, the single tone element is a single tone character in Chinese, for example, the pronunciation of the Chinese character "West" is only "x ī".
In one embodiment, a multi-tone element library containing all multi-tone elements and a single-tone element library containing all single-tone elements can be generated in advance for the type of text used by the original name. Then, by respectively matching all elements contained in the original name with a polyphonic element library, the polyphonic elements contained in the original name and all the pronunciations of the polyphonic elements can be determined; further, by matching each pronunciation with the phone element library, the phone element corresponding to each pronunciation can be selected. For example, when the original name of the target object is "yang aliza", by matching the Chinese character "yang" and the Chinese character "aliza" with a polyphonic element library (i.e., polyphonic character library), respectively, it can be determined that the Chinese character "aliza" is a polyphonic character, and there are pronunciations "qi" n "and" x ī "; further, by matching the readings "qi a n" and "x ī" with the monophonic element library (i.e., the monophonic character library), respectively, the monophonic character "qian" with the reading "qi a n" and the monophonic character "west" with the reading "x ī" can be determined.
Although the original name contains polyphonic elements, the original name of the target object only needs to adopt a fixed pronunciation, so the server can return a plurality of alternative items corresponding to the pronunciations of the polyphonic elements for selection; correspondingly, the pronunciation actually adopted by the target object for the polyphonic elements contained in the original name can be determined according to the selected alternative items. For example, when the original name of the target object is "poplar alize", according to the readings "qi a n" and "x ī" of the polyphonic element "alize", the server may return the corresponding two alternatives, such as the alternative corresponding to the reading "qi a n", the alternative corresponding to the reading "x ī", for selection.
And 406, determining a designated tone element corresponding to the selected alternative, wherein the designated tone element is used for generating a voice file corresponding to the target object, so as to play the voice file to the target user after the target user is successfully identified by the identity recognition device.
As described above, although the original name includes polyphonic elements, the original name of the target object necessarily employs only one fixed pronunciation, and thus the above-described replaced name can be generated by setting the monophonic element employing the pronunciation as the above-described specified monophonic element according to the pronunciation actually employed by the target object for the polyphonic elements included in the original name. For example, when the original name of the target object is actually read as "y-ng x ī", the polyphonic element "mada" may be replaced with the designated monophonic element "west" of the reading "x ī", with the resulting replacement named "yang west". Specifically, the target object or the administrator may select the actual pronunciation of the polyphonic element included in the original name by providing the target object or the administrator with alternatives respectively corresponding to the respective pronunciations of the polyphonic element.
In one embodiment, by generating a corresponding replaced name for an original name containing a polyphonic element and generating a voice file based on the replaced name, the voice file can be directly generated without adding a special mark for the actual reading of the polyphonic element (for example, the mark [ qian4] indicates that the reading of "madder" is "qi" or the like), which has extremely high versatility compared to different mark formats that may be adopted by different manufacturers, and avoids parsing the mark, increasing complexity and processing computation when generating the voice file.
In an embodiment, the server may generate and store a mapping relationship between the corresponding original name and the replaced name according to the determined replaced name; similarly, the server may store the mapping between the original names and the replaced names of all objects. In the subsequent process, the identification device may determine the original name of the target object according to the identification result of the target object, and then send an inquiry request including the original name to the server, and the server may return a corresponding replaced name based on the mapping relationship, so that the identification device may generate a corresponding voice file based on the replaced name.
In an embodiment, the server may generate and store a corresponding voice file in advance according to the determined name after replacement; similarly, the server may store the voice files corresponding to the replaced names of all objects. In the subsequent process, the identity recognition device may send the original name of the target object to the server after confirming the identity information of the target object, so that the server returns a corresponding voice file to the identity recognition device. Because the voice file is not required to be temporarily generated, the response speed of the identity recognition equipment can be increased.
In an embodiment, for a pre-generated voice file, a mapping relationship can be directly established between an original name and the voice file, so that the corresponding voice file can be directly determined according to the original name of a target object. Or, a mapping relationship may be established between the replaced name and the voice file, and then, in combination with the mapping relationship between the original name and the replaced name, after the original name is determined, the corresponding voice file may also be determined.
In one embodiment, the mapping relationship (the mapping relationship between the original name and the replaced name, or the mapping relationship between the original name and the voice file, or the mapping relationship between the replaced name and the voice file) is not applicable to all objects recognized by the identification device, but is bound to one or more set objects. For example, for a user whose original name is "yangqi" and whose other user is "y-ng x ī", the sound of one user is "yang qi" and the sound of the other user is "y-ng x ī", so that for each user whose original name contains polyphonic elements, a corresponding mapping should be generated separately to ensure the correctness of the voice broadcast.
In an embodiment, the replaced name or the voice file is returned to the identity recognition device, so that after the identity recognition device recognizes the identity information of the target object, the voice file can be played to correctly broadcast the name of the target object, and the target object can quickly determine that the identity recognition device successfully recognizes the identity information of the target object.
Fig. 5 is a flowchart of another method for setting a voice broadcast content according to an exemplary embodiment. As shown in fig. 5, the method is applied to a server (e.g., the server 11 shown in fig. 1, etc.), and may include the following steps:
step 502, receiving an object adding request initiated by an administrator, where the object adding request is used to add a target object as a recognizable object of an identity recognition device, where an original name of the target object includes a polyphonic element.
In one embodiment, an object addition request may be initiated by the electronic device to the server, where the object addition request may be added individually to the target object as a recognizable object of the identification device, or may be added to each of a plurality of objects including the target object as recognizable objects of the identification device. For example, a server of the group office platform may be run on the server, and a client of the group office platform may be run on the electronic device and the identification device, so that data interaction may be implemented between the server, the electronic device, and the identification device based on the server and the client.
Specifically, the electronic device may initiate an object addition request to the server, and accordingly notify the server of the identity information of the target object to be added; then, the server can issue a collection instruction to the identification device, so that the identification device collects the standard characteristic information of the target object. The standard characteristic information and the identity information of the target object can be bound and stored on the identity recognition equipment, so that the identity recognition equipment can perform identity recognition on the target object according to the standard characteristic information and the identity information to generate an identity recognition result; or, the identity recognition device may upload the standard characteristic information to the server, and the server binds and stores the standard characteristic information and the identity information of the target object, and when the identity recognition device performs identity recognition on the target object in the following, the identity recognition device needs to acquire the characteristic information of the target object and upload the characteristic information to the server, and the server completes identity recognition and informs the identity recognition device of an identity recognition result.
In one embodiment, the object addition request may be initiated by the identification device directly to the server, and the object addition request may be added to the identification device as a recognizable object of the identification device separately or as a recognizable object of the identification device by adding a plurality of objects including the target object. For example, a server of the group office platform can be operated on the server, and a client of the group office platform can be operated on the identification device, so that data interaction between the server and the identification device can be realized based on the server and the client.
Specifically, the identification device may collect standard characteristic information of the target object. The standard characteristic information and the identity information of the target object can be bound and stored on the identity recognition equipment, so that the identity recognition equipment can perform identity recognition on the target object according to the standard characteristic information and the identity information to generate an identity recognition result; meanwhile, the identity recognition device can upload the identity information of the target object to the server, so that the server knows that the identity recognition device can recognize the identity of the target object. Or, the identity recognition device may upload the standard characteristic information to the server, and the server binds and stores the standard characteristic information and the identity information of the target object, and when the identity recognition device performs identity recognition on the target object in the following, the identity recognition device needs to acquire the characteristic information of the target object and upload the characteristic information to the server, and the server completes identity recognition and informs the identity recognition device of an identity recognition result.
Step 504, returning a plurality of options to the administrator, wherein the options respectively correspond to the pronunciations of the polyphonic elements.
In one embodiment, the name is composed of several elements, the element type being related to the type of text the name takes. For example, when chinese is used, the elements may be chinese characters, and when english is used, the elements may be words, which are not listed one by one. Polyphonic elements refer to elements having multiple pronunciations; taking the chinese language as an example, polyphonic elements are polyphonic characters in the chinese language, for example, the pronunciation of the Chinese character "madder" includes "qi a n" and "x ī". Similarly, a monophonic element refers to an element that has only one pronunciation; taking Chinese as an example, the single tone element is a single tone character in Chinese, for example, the pronunciation of the Chinese character "West" is only "x ī".
In one embodiment, a multi-tone element library containing all multi-tone elements and a single-tone element library containing all single-tone elements can be generated in advance for the type of text used by the original name. Then, by respectively matching all elements contained in the original name with a polyphonic element library, the polyphonic elements contained in the original name and all the pronunciations of the polyphonic elements can be determined; further, by matching each pronunciation with the phone element library, the phone element corresponding to each pronunciation can be selected. For example, when the original name of the target object is "yang aliza", by matching the Chinese character "yang" and the Chinese character "aliza" with a polyphonic element library (i.e., polyphonic character library), respectively, it can be determined that the Chinese character "aliza" is a polyphonic character, and there are pronunciations "qi" n "and" x ī "; further, by matching the readings "qi a n" and "x ī" with the monophonic element library (i.e., the monophonic character library), respectively, the monophonic character "qian" with the reading "qi a n" and the monophonic character "west" with the reading "x ī" can be determined.
Although the original name contains polyphonic elements, the original name of the target object only needs to adopt a fixed pronunciation, so the server can return a plurality of alternative items corresponding to the pronunciations of the polyphonic elements for selection; correspondingly, the pronunciation actually adopted by the target object for the polyphonic elements contained in the original name can be determined according to the selected alternative items. For example, when the original name of the target object is "poplar alize", according to the readings "qi a n" and "x ī" of the polyphonic element "alize", the server may return the corresponding two alternatives, such as the alternative corresponding to the reading "qi a n", the alternative corresponding to the reading "x ī", for selection.
Step 506, determining a designated tone element corresponding to the selected option, where the designated tone element is used to generate a voice file corresponding to the target object, so as to play the voice file to the target user after the target user is successfully recognized by the identification device.
As described above, although the original name includes polyphonic elements, the original name of the target object necessarily employs only one fixed pronunciation, and thus the above-described replaced name can be generated by setting the monophonic element employing the pronunciation as the above-described specified monophonic element according to the pronunciation actually employed by the target object for the polyphonic elements included in the original name. For example, when the original name of the target object is actually read as "y-ng x ī", the polyphonic element "mada" may be replaced with the designated monophonic element "west" of the reading "x ī", with the resulting replacement named "yang west". Specifically, the target object or the administrator may select the actual pronunciation of the polyphonic element included in the original name by providing the target object or the administrator with alternatives respectively corresponding to the respective pronunciations of the polyphonic element.
In one embodiment, by generating a corresponding replaced name for an original name containing a polyphonic element and generating a voice file based on the replaced name, the voice file can be directly generated without adding a special mark for the actual reading of the polyphonic element (for example, the mark [ qian4] indicates that the reading of "madder" is "qi" or the like), which has extremely high versatility compared to different mark formats that may be adopted by different manufacturers, and avoids parsing the mark, increasing complexity and processing computation when generating the voice file.
In an embodiment, the server may generate and store a mapping relationship between the corresponding original name and the replaced name according to the determined replaced name; similarly, the server may store the mapping between the original names and the replaced names of all objects. In the subsequent process, the identification device may determine the original name of the target object according to the identification result of the target object, and then send an inquiry request including the original name to the server, and the server may return a corresponding replaced name based on the mapping relationship, so that the identification device may generate a corresponding voice file based on the replaced name.
In an embodiment, the server may generate and store a corresponding voice file in advance according to the determined name after replacement; similarly, the server may store the voice files corresponding to the replaced names of all objects. In the subsequent process, the identity recognition device may send the original name of the target object to the server after confirming the identity information of the target object, so that the server returns a corresponding voice file to the identity recognition device. Because the voice file is not required to be temporarily generated, the response speed of the identity recognition equipment can be increased.
In an embodiment, for a pre-generated voice file, a mapping relationship can be directly established between an original name and the voice file, so that the corresponding voice file can be directly determined according to the original name of a target object. Or, a mapping relationship may be established between the replaced name and the voice file, and then, in combination with the mapping relationship between the original name and the replaced name, after the original name is determined, the corresponding voice file may also be determined.
In one embodiment, the mapping relationship (the mapping relationship between the original name and the replaced name, or the mapping relationship between the original name and the voice file, or the mapping relationship between the replaced name and the voice file) is not applicable to all objects recognized by the identification device, but is bound to one or more set objects. For example, for a user whose original name is "yangqi" and whose other user is "y-ng x ī", the sound of one user is "yang qi" and the sound of the other user is "y-ng x ī", so that for each user whose original name contains polyphonic elements, a corresponding mapping should be generated separately to ensure the correctness of the voice broadcast.
In an embodiment, the replaced name or the voice file is returned to the identity recognition device, so that after the identity recognition device recognizes the identity information of the target object, the voice file can be played to correctly broadcast the name of the target object, and the target object can quickly determine that the identity recognition device successfully recognizes the identity information of the target object.
The technical solution of the present specification will be described below by taking an attendance scene in an enterprise as an example. Suppose that the front desk of the enterprise C is provided with attendance equipment, and the attendance equipment identifies the identity of the enterprise staff who come and go in a face recognition mode so as to complete corresponding attendance operation. The staff of the enterprise C includes an administrator Ua and a staff Ub, and the administrator Ua can add the staff Ub to an object that can be recognized by the attendance checking device. The mobile phone used by the administrator Ua can operate the client C1 of the group office platform T, the attendance checking device can operate the client C2 of the group office platform T, and the server of the group office platform T, so that the administrator Ua can complete the adding operation of the employee Ub based on the interaction between the client C1, the client C2, and the server. For example, fig. 6 is an interaction diagram for adding recognizable objects to a attendance device and performing attendance operations according to an exemplary embodiment. As shown in fig. 6, the interactive process may include the following steps:
in step 601, the administrator Ua initiates an employee addition request to the server.
In one embodiment, administrator Ua initiates an employee addition request to a server running on the server through client C1 described above. The employee addition request includes identity information of the employee Ub, such as a user name, a contact address, a position, a department, and the like.
In step 602, the server identifies polyphones.
In one embodiment, a predefined polyphonic library is maintained at the server that contains all of the polyphonic characters in the Chinese characters. The server matches the user name of the employee Ub with the polyphone database to determine the polyphones and all the pronunciations contained in the user name. Of course, the polyphonic library may be maintained at a body other than the server, and the server may call or initiate a request to the body identifying the polyphonic words and all of their pronunciations contained in the user name.
For example, assuming that the user name of the employee Ub is called "yankee", by matching with the polyphonic character library, the "madder" can be determined to be polyphonic, and there are 2 readings "qi" n "and" x ī ", respectively, so that the following steps 603 to 605 and the like need to be performed. If the user name of employee Ub does not have polyphones, the process may jump directly to step 606.
Step 603, the server looks up the alternative monophonic word, constructs the alternate and returns to the administrator Ua.
In one embodiment, the server maintains a predefined monophonic character library that contains all of the monophonic characters in a Chinese character. The server matches all the pronunciations of the polyphonic characters identified in step 602 with the monophonic character library, respectively, to determine the monophonic character corresponding to each pronunciation, respectively. Of course, the monophonic character library may be maintained at a subject other than the server, and the server may call the monophonic character library or initiate a request to the subject, which determines the monophonic characters respectively corresponding to each of the above-mentioned readings.
For example, when 2 readings of "qi a n" and "x ī" exist in the "madder" in the user name, the tone word "qi" corresponding to the reading "qi a n" and the tone word "xi" corresponding to the reading "x ī" can be determined by matching with the above tone word library.
Further, the server may construct corresponding alternatives for each pronunciation of the polyphone, and return the alternatives to the administrator Ua to be selected by the administrator Ua. Meanwhile, according to the corresponding relationship between each pronunciation and the single-tone word, a one-to-one corresponding relationship also exists between the alternative and the single-tone word, for example, alternative 1 corresponds to the single-tone word "qian", and alternative 2 corresponds to the single-tone word "west". Accordingly, when the alternative 1 is selected, the pronunciation indicating the user name of the employee Ub is "yang qi a n", and the following alternative name can be generated from the corresponding monophonic word "qian"; alternatively, when alternate 2 is selected, indicating that the reading of employee Ub's username is "y ng x ī", the following alternative name may be generated from the corresponding single tone word "west".
FIG. 7 is a schematic diagram of a pronunciation selection interface provided by an exemplary embodiment. The client C1 may present the administrator Ua with a sound selection interface 700 as shown in fig. 7, in which sound selection interface 700 the options 701, 702 may be shown. The alternative 701 contains the content "yang qi a n" corresponding to the polyphone "alizarin" qi a n "; the alternate 702 contains the content "Yang x ī" and the reading corresponding to the polyphone "madder" is "x ī". In other embodiments, the displayed alternatives may include a single-tone word corresponding to each pronunciation, such as the single-tone words "qian" and "xi" described above, which may be specifically displayed as "yankee (beautiful yankee)" and "yankee (western yankee)". In fact, the description does not limit the form of the alternatives, as long as the actual pronunciation of the user name of the employee Ua can be determined.
In step 604, the administrator Ua selects the designated option and returns a corresponding selection result to the server.
And step 605, the server generates and stores the replaced name according to the selection result.
In an embodiment, the server may obtain the alternative selected by the administrator Ua according to the selection result, and may determine the single-tone word corresponding to the selected alternative accordingly. For example, when administrator Ua selects alternative 702 as shown in fig. 7, it indicates that the user name of employee Ub reads "yang x ī" and the corresponding monophonic character is "west". Accordingly, the server may replace the polyphonic character "madder" in the original user name "yankee" with the single-tone character "west" described above, and recompose the replaced name "yankee".
Therefore, the user name of the employee Ub is called "yang aliza" and the corresponding post-replacement name is "yang xi", that is, both the user name "yang aliza" and the post-replacement name "yang xi" are associated with the employee Ub and stored at the server.
And 606, the server pushes the identity information to the attendance checking equipment and issues a face acquisition instruction.
And step 607, the attendance checking equipment collects the human face characteristics and stores the human face characteristics in an associated manner.
In an embodiment, the server sends the identity information of the employee Ub to the attendance checking device for storage, for example, the identity information may include the user name "yankee" of the employee Ub. Meanwhile, the server issues a face acquisition instruction to the attendance checking equipment, so that the attendance checking equipment can acquire the face image of the employee Ub and record standard characteristic information contained in the acquired face image. Then, the attendance device can store the received identity information and the standard characteristic information in an associated manner, so as to be used for subsequently carrying out identity identification on the staff Ub and finishing attendance operation.
The steps 601 to 607 are related to the configuration phase executed in advance, and based on the preparation work of the configuration phase, the related steps of the attendance checking phase in the following steps 608 to 612 can be realized.
At step 608, the attendance device identifies the current user.
In an embodiment, the attendance checking device has a certain detection range, for example, a sector area with the camera as a circle center, a preset radius and a preset angle, so that when the current user is in the detection range, the attendance checking device can perform face recognition on the current user to determine the identity information of the current user.
For example, when the attendance checking device detects that the current user is the aforementioned employee Ub, the user name of the employee Ub may be obtained from the locally stored identity information, for example, the user name may be "yankee".
Step 609, the attendance device sends the user name to the server.
Step 610, the server searches for the replaced name and returns the replaced name to the attendance checking device.
In an embodiment, for all employees in the enterprise C whose user names include polyphones, mapping relationships between the corresponding user names and the replaced names are stored in the server, for example, the mapping relationships may be stored in the user polyphone self-selection reading library. Of course, the user polyphone self-selection reading library may be stored in other main bodies besides the server, as long as the server can call the user polyphone self-selection reading library, or may request the main body to match the user polyphone self-selection reading library.
When the user name uploaded by the attendance checking equipment is received, the server can search the mapping relation corresponding to the user name in the user polyphone self-selection reading library so as to determine the replaced name corresponding to the user name. If the user name uploaded by the attendance checking equipment does not contain polyphones, the server cannot find the mapping relation matched with the user name in the user polyphone self-selection reading library, and the server returns the user name to the attendance checking equipment, so that the attendance checking equipment generates a voice file based on the user name. If the user name uploaded by the attendance checking equipment contains polyphones, the server can find the mapping relation corresponding to the user name, so that the replaced name corresponding to the user name is determined. For example, when the user name is called "yankee," the server may find the corresponding replaced name as "yankee," so that the server may return the replaced name "yankee" to the attendance device.
It should be noted that: taking the user name "yang qian" as an example, there may be a plurality of employees called "yang qian" with user names in the enterprise C. Therefore, when the server records the mapping relationship between the user name and the replaced name, other information of the employee Ub, such as the user ID of the employee Ub, may also be added, and the attendance checking device sends the user name to the server in step 609 and also carries the user ID, so as to ensure that the server finds the mapping relationship corresponding to the employee Ub, but not the mapping relationship corresponding to the other "yankee".
Step 611, the attendance checking device generates a voice file according to the replaced name.
In one embodiment, the attendance checking equipment can automatically generate a corresponding voice file according to the Chinese characters contained in the replaced name. For example, when the latter name "yang xi" is replaced, the characters "yang" and "xi" can be obtained so that the voice file generated by the attendance device accordingly corresponds to the pronunciation "yang x ī".
Step 612, the attendance checking equipment displays the user name, plays the voice file, and executes attendance checking operation.
In one embodiment, the attendance checking device displays the user name "Yangqi" on the screen, and simultaneously plays the voice file, wherein the voice file forms a broadcast sound of "yang x ī", and the reading sound of the polyphonic character "xi" in the user name is "x ī". When the employee Ub sees the displayed user name "yangxi ī" or hears the broadcast sound "yang x ī", the attendance device can be determined to accurately identify the employee, so that the employee can complete corresponding attendance operation for the employee.
In summary, in this specification, by identifying the polyphonic elements included in the original name of the target object and replacing the polyphonic elements with the monophonic elements having the same pronunciation according to the actually adopted pronunciation of the polyphonic elements, the identity recognition device can correctly broadcast the original name by playing the language file corresponding to the monophonic element without adding corresponding marks to the pronunciation adopted by the polyphonic elements or analyzing the marks in the subsequent process, so that the processing calculation amount of the identity recognition device is reduced and the response speed is increased.
In addition to the setting of the polyphones in the name for the members within the community by the administrator through the above-described embodiment, the polyphones may be actively set by the members at the beginning of joining the community so as to be added to the member information of the members, and may be applied to all scenes within the community thereafter without repeated setting. For example, fig. 8 is a flowchart of a community member adding method provided by an exemplary embodiment. As shown in fig. 8, the method applied to a server (e.g., the server 11 shown in fig. 1) may include the following steps:
step 802, acquiring a join request initiated by a user for a group, wherein personal information of the user comprises an original name and a specified single-tone element corresponding to the original name, and the specified single-tone element has the same pronunciation as a polyphonic element in the original name.
In one embodiment, a user may generate a request content, fill in personal information such as an original name during the generation of the request content, determine a designated monophonic element corresponding to a polyphonic element in the original name, and then initiate a join request based on the generated request content. For example, the electronic device may open a request content generation interface that includes an original name entry option, may actively fill in the original name by the user, or may automatically fill in the original name included in the profile of the logged-in account; then, the electronic device can upload the original name to the server, the server identifies the polyphonic elements contained in the original name and returns the alternatives corresponding to each pronunciation, the user selects a certain alternative according to the actual pronunciation of the polyphonic elements in the original name, and the single-tone element corresponding to the alternative is determined as the appointed single-tone element; finally, the electronic device generates a request content containing personal information such as an original name, a specified tone element, and the like as a join request, and initiates the join request. Then, when the join request is sent to the administrator of the community, the administrator can see the personal information contained in the join request, and the administrator can audit the join request.
In one embodiment, the user may first initiate a join request and then determine the corresponding designated tone element for the multi-tone element. For example, a join request initiated by an electronic device may contain a portion of personal information, and the portion of personal information may contain the original name but not the designated tone element described above. Then, when the join request is sent to the administrator of the community, the administrator does not see the designated tone element corresponding to the polyphonic element in the original name during the auditing process. The joining request is obtained from the user by the server and sent to the administrator, so that the server can obtain an original name contained in the joining request, identify polyphonic elements contained in the original name, and return alternatives corresponding to all the pronunciations to the user, and the user selects a certain alternative according to the actual pronunciations of the polyphonic elements in the original name, and the monophonic element corresponding to the alternative is determined to be the appointed monophonic element so as to be related to the polyphonic elements in the original name of the user.
In one embodiment, the name is composed of several elements, the element type being related to the type of text the name takes. For example, when chinese is used, the elements may be chinese characters, and when english is used, the elements may be words, which are not listed one by one. Polyphonic elements refer to elements having multiple pronunciations; taking the chinese language as an example, polyphonic elements are polyphonic characters in the chinese language, for example, the pronunciation of the Chinese character "madder" includes "qi a n" and "x ī". Similarly, a monophonic element refers to an element that has only one pronunciation; taking Chinese as an example, the single tone element is a single tone character in Chinese, for example, the pronunciation of the Chinese character "West" is only "x ī".
In one embodiment, a multi-tone element library containing all multi-tone elements and a single-tone element library containing all single-tone elements can be generated in advance for the type of text used by the original name. Then, by respectively matching all elements contained in the original name with a polyphonic element library, the polyphonic elements contained in the original name and all the pronunciations of the polyphonic elements can be determined; further, by matching each pronunciation with the phone element library, the phone element corresponding to each pronunciation can be selected. For example, when the original name of the user is "yang aliza", by matching the Chinese character "yang" and the Chinese character "aliza" with a polyphonic element library (i.e., polyphonic character library), respectively, the Chinese character "aliza" can be determined to be a polyphonic character, and there are pronunciations "qi a n" and "x ī"; further, by matching the readings "qi a n" and "x ī" with the monophonic element library (i.e., the monophonic character library), respectively, the monophonic character "qian" with the reading "qi a n" and the monophonic character "west" with the reading "x ī" can be determined.
Although the original name contains polyphonic elements, the original name of the user only needs to adopt a fixed pronunciation, so the server can return a plurality of alternative items corresponding to the pronunciations of the polyphonic elements for selection; correspondingly, the pronunciation actually adopted by the user for the polyphonic elements contained in the original name can be determined according to the selected alternative items. For example, when the original name of the user is "poplar alize", according to the readings "qi a n" and "x ī" of the polyphonic element "alize", the server may return the corresponding two alternatives, such as the alternative corresponding to the reading "qi a n", the alternative corresponding to the reading "x ī", for selection.
Step 804, when the administrator of the community agrees to the join request, adding the user as a member of the community, so that the original name and the designated tone element are updated to the member information corresponding to the user.
In an embodiment, in an enterprise office scenario, the group may be an enterprise, and the join request may be an enrollment application of the user. Of course, the present description may also be applied in other scenarios.
In an embodiment, the original name and the designated tone element are updated to the member information corresponding to the user, so that the member information corresponding to the user can be applied to the events related to the user, which occur in the community later, and the original name and the designated tone element contained in the member information can be reused in various scenes. For example, in an attendance scenario, an original name may be issued to an attendance device, and a voice file generated according to a replaced name (the replaced name is obtained by replacing a polyphonic element in the original name with a specified single-tone element) may be issued to the attendance device, so that the attendance device may display the original name and broadcast the voice file while performing an attendance operation after recognizing the identity information of a user. For another example, in a conference check-in scenario, the original name and the designated tone element may be issued to the check-in device, so that after the check-in device identifies the identity information of the user, the original name may be displayed while performing the check-in operation, and the designated tone element is used to mark the polyphonic elements included in the original name.
Besides voice broadcast, the technical scheme of the specification can be applied to a display scene, and marks are carried out on polyphonic elements contained in display content, so that any user can accurately know the correct pronunciation of the related polyphonic elements after checking. For example, fig. 9 is a flowchart of a display method provided by an exemplary embodiment. As shown in fig. 9, the method applied to the display device may include the steps of:
step 902, obtaining a content to be displayed, where the content to be displayed includes an original name of a target object and a specified single-tone element corresponding to the original name, and the specified single-tone element and a polyphonic element in the original name have the same pronunciation.
In an embodiment, the content to be displayed may be any content containing polyphonic elements, such as a user name, an object name, a pet name, poster content, road sign content, and the like, which is not limited in this specification.
In one embodiment, any user may provide the original content to the server, such as uploading to the server or instructing the server to obtain the original content from a certain channel, identify polyphonic elements contained in the original content by the server, and return a plurality of alternatives respectively corresponding to the respective pronunciations of the polyphonic elements to the any user for selection; correspondingly, the server can determine the actually adopted pronunciation of the polyphonic elements contained in the original content by the user according to the selected alternative items, so as to determine the corresponding appointed polyphonic elements. For example, a multi-tone element library containing all multi-tone elements, and a monophonic element library containing all monophonic elements may be generated in advance for the type of text employed for the original content. Then, by respectively matching all elements contained in the original content with a polyphonic element library, polyphonic elements contained in the original content and all pronunciations of the polyphonic elements can be determined; further, by matching each pronunciation with the phone element library, the phone element corresponding to each pronunciation can be selected.
Furthermore, the element to be displayed can be generated by adding a specified single-tone element corresponding to the polyphonic element in the original content, so that the specified single-tone element marks the polyphonic element. For example, assuming that the original content includes a name "… … yang aliza … …" of a certain user, and it is determined that the designated tone element corresponding to the polyphonic element "aliza" is "west" based on the above-mentioned manner, the corresponding content to be displayed may be generated as "… … yang aliza (west) … …" to implement the tagging action. Of course, the marking can be performed in other ways, and the description is not limited thereto.
In one embodiment, a user causes a server to push original content onto a display device for presentation by providing the original content to the server. Then, the server may actively determine the polyphonic elements included in the original content through the above process in the process, and further determine the corresponding designated monophonic elements, so that after the original content is generated into the content to be displayed, the content to be displayed is pushed to the display device for display.
Step 904, displaying the content to be displayed.
Since the polyphonic elements have been marked by specifying the monophonic element, the contents to be displayed that are presented are easy to view and read, and a situation in which a reading error occurs does not occur. For example, in a group office scene, the display device may include electronic position cards of group members, and the content to be displayed may include information such as names of the respective group members, and the above solution may mark polyphonic elements in the names of the group members to avoid the situation of being misread.
FIG. 10 is a schematic block diagram of an apparatus provided in an exemplary embodiment. Referring to fig. 10, at the hardware level, the apparatus includes a processor 1002, an internal bus 1004, a network interface 1006, a memory 1008, and a non-volatile memory 1010, although it may also include hardware required for other services. The processor 1002 reads a corresponding computer program from the nonvolatile memory 1010 into the memory 1008 and then runs the computer program to form a voice broadcasting device on a logical level. Of course, besides software implementation, the one or more embodiments in this specification do not exclude other implementations, such as logic devices or combinations of software and hardware, and so on, that is, the execution subject of the following processing flow is not limited to each logic unit, and may also be hardware or logic devices.
Referring to fig. 11, in a software implementation, the voice broadcasting device may include:
a recognition unit 1101 that performs identification for the target object;
an obtaining unit 1102, configured to obtain a voice file corresponding to the target object according to an identity recognition result of the target object; when the original name of the target object contains polyphonic elements, the voice file is generated by a replaced name, and the replaced name is obtained by replacing the polyphonic elements in the original name with specified single-tone elements;
a playing unit 1103 plays the acquired voice file.
Optionally, the obtaining unit 1102 is specifically configured to:
acquiring the replaced name of the target object;
and generating the voice file according to the replaced name.
Optionally, the obtaining unit 1102 obtains the replaced name of the target object by:
determining the original name of the target object according to the identification result of the target object; acquiring a replaced name of the target object according to a mapping relation between a locally stored original name and the replaced name; or the like, or, alternatively,
determining the original name of the target object according to the identification result of the target object; and sending an inquiry request containing the original name to a server, and receiving a replaced name returned by the server, wherein the replaced name is determined by the server according to the stored mapping relation between the original name and the replaced name.
Optionally, the obtaining unit 1102 is specifically configured to:
and acquiring a pre-generated voice file corresponding to the target object.
Optionally, the method further includes:
a generating unit 1104 that generates the voice file in advance from the replaced name of the target object; or the like, or, alternatively,
receiving unit 1105, receiving server generates the voice file in advance according to the replaced name of the target object.
Optionally, the method further includes:
a presentation unit 1106 presents the original name of the target object.
Optionally, the identifying unit 1101 is specifically configured to:
performing identity recognition on the target object by at least one of the following ways: fingerprint identification, face identification, iris identification and gait identification.
FIG. 12 is a schematic block diagram of an apparatus provided in an exemplary embodiment. Referring to fig. 12, at the hardware level, the apparatus includes a processor 1202, an internal bus 1204, a network interface 1206, a memory 1208, and a non-volatile memory 1210, although other hardware required for services may be included. The processor 1202 reads a corresponding computer program from the nonvolatile memory 1210 into the memory 1208 and then runs the computer program, thereby forming a setting device of the voice broadcast content on a logical level. Of course, besides software implementation, the one or more embodiments in this specification do not exclude other implementations, such as logic devices or combinations of software and hardware, and so on, that is, the execution subject of the following processing flow is not limited to each logic unit, and may also be hardware or logic devices.
Referring to fig. 13, in a software implementation, the apparatus for setting the content of the voice broadcast may include:
a requesting unit 1301, which initiates an object adding request to a server, where the object adding request is used to add a target object as a recognizable object of an identity recognition device, where an original name of the target object includes a polyphonic element;
an obtaining unit 1302, configured to obtain a plurality of candidates returned by the server, where the plurality of candidates respectively correspond to the readings of the polyphonic elements;
and a determining unit 1303 configured to determine a specified single-tone element corresponding to the selected option, where the specified single-tone element is used to generate a voice file corresponding to the target object, so as to play the voice file to the target user after the target user is successfully identified by the identity recognition device.
FIG. 14 is a schematic block diagram of an apparatus provided in an exemplary embodiment. Referring to FIG. 14, at the hardware level, the device includes a processor 1402, an internal bus 1404, a network interface 1406, a memory 1408, and a non-volatile storage 1410, although other hardware required for service may be included. The processor 1402 reads a corresponding computer program from the nonvolatile memory 1410 into the memory 1408 and then runs, forming a voice broadcast device on a logical level. Of course, besides software implementation, the one or more embodiments in this specification do not exclude other implementations, such as logic devices or combinations of software and hardware, and so on, that is, the execution subject of the following processing flow is not limited to each logic unit, and may also be hardware or logic devices.
Referring to fig. 15, in a software implementation, the voice broadcasting device may include:
a receiving unit 1501, configured to receive an inquiry request from an identification device, where the inquiry request includes an original name of a target object identified by the identification device, and the original name includes a polyphonic element;
a returning unit 1502, returning the replaced name corresponding to the original name, so that the identity recognition device generates a voice file according to the replaced name and plays the voice file; or returning the voice file generated according to the replaced name to be played by the identity recognition equipment; wherein the replaced name is obtained by replacing a polyphonic element in the original name with a designated monophonic element.
Fig. 16 is a schematic block diagram of an apparatus provided in an exemplary embodiment. Referring to fig. 16, at the hardware level, the apparatus includes a processor 1602, an internal bus 1604, a network interface 1606, a memory 1608, and a nonvolatile memory 1610, but may also include hardware required for other services. The processor 1602 reads a corresponding computer program from the nonvolatile memory 1610 to the memory 1608 and runs the computer program to form a setting device of a voice broadcast content on a logical level. Of course, besides software implementation, the one or more embodiments in this specification do not exclude other implementations, such as logic devices or combinations of software and hardware, and so on, that is, the execution subject of the following processing flow is not limited to each logic unit, and may also be hardware or logic devices.
Referring to fig. 17, in a software implementation, the apparatus for setting the content of the voice broadcast may include:
a receiving unit 1701 that receives an object adding request initiated by an administrator, the object adding request being used for adding a target object as a recognizable object of an identification device, wherein an original name of the target object contains a polyphonic element;
a returning unit 1702 that returns a plurality of alternatives to the administrator, the plurality of alternatives respectively corresponding to the respective readings of the polyphonic elements;
a determining unit 1703, determining a designated tone element corresponding to the selected option, where the designated tone element is used to generate a voice file corresponding to the target object, so as to play the voice file to the target user after the target user is successfully recognized by the identity recognition device.
Optionally, the return unit 1702 is specifically configured to:
matching each component element contained in the original name with a predefined polyphonic element library respectively to identify polyphonic elements contained in the original name and all pronunciations thereof;
matching each pronunciation of the identified polyphonic elements with a predefined monophonic element library respectively to determine the monophonic elements corresponding to each pronunciation respectively;
and returning the determined options corresponding to the single-tone elements to the administrator.
Optionally, the method further includes:
an establishing unit 1704, configured to establish a mapping relationship between the original name and the replaced name, so as to return the replaced name or a voice file generated according to the replaced name when receiving an inquiry request including the original name, which is initiated by the identity recognition device; or, establishing a mapping relation between the original name and the voice file, so as to return the voice file when receiving an inquiry request containing the original name initiated by the identity recognition equipment;
wherein the replaced name is obtained by replacing a polyphonic element in the original name with the specified monophonic element.
FIG. 18 is a schematic block diagram of an apparatus provided in an exemplary embodiment. Referring to fig. 18, at the hardware level, the apparatus includes a processor 1802, an internal bus 1804, a network interface 1806, a memory 1808, and a nonvolatile memory 1810, but may also include hardware required for other services. The processor 1802 reads a corresponding computer program from the non-volatile storage 1810 into the memory 1808 and runs the computer program, thereby forming a community member adding device on a logical level. Of course, besides software implementation, the one or more embodiments in this specification do not exclude other implementations, such as logic devices or combinations of software and hardware, and so on, that is, the execution subject of the following processing flow is not limited to each logic unit, and may also be hardware or logic devices.
Referring to fig. 19, in a software implementation, the adding means of the community members may include:
an obtaining unit 1901, configured to obtain a join request initiated by a user for a community, where personal information of the user includes an original name and a specified tone element corresponding to the original name, and the specified tone element has the same pronunciation as a polyphonic element in the original name;
an adding unit 1902, configured to add the user as a member of the community when the administrator of the community approves the join request, so that the original name and the designated tone element are updated to the member information corresponding to the user.
Fig. 20 is a schematic block diagram of an apparatus provided in an exemplary embodiment. Referring to fig. 20, at the hardware level, the device includes a processor 2002, an internal bus 2004, a network interface 2006, a memory 2008, and a non-volatile storage 2010, but may also include hardware required for other services. The processor 2002 reads the corresponding computer program from the non-volatile storage 2010 into the memory 2008 and then runs the computer program to form the display device on a logical level. Of course, besides software implementation, the one or more embodiments in this specification do not exclude other implementations, such as logic devices or combinations of software and hardware, and so on, that is, the execution subject of the following processing flow is not limited to each logic unit, and may also be hardware or logic devices.
Referring to fig. 21, in a software implementation, the display device may include:
an acquiring unit 2101 configured to acquire a content to be displayed, the content to be displayed including an original name of a target object and a specified monophonic element corresponding to the original name, the specified monophonic element having a same pronunciation as a polyphonic element in the original name;
a display unit 2102 which displays the content to be displayed.
The systems, devices, modules or units illustrated in the above embodiments may be implemented by a computer chip or an entity, or by a product with certain functions. A typical implementation device is a computer, which may take the form of a personal computer, laptop computer, cellular telephone, camera phone, smart phone, personal digital assistant, media player, navigation device, email messaging device, game console, tablet computer, wearable device, or a combination of any of these devices.
In a typical configuration, a computer includes one or more processors (CPUs), input/output interfaces, network interfaces, and memory.
The memory may include forms of volatile memory in a computer readable medium, Random Access Memory (RAM) and/or non-volatile memory, such as Read Only Memory (ROM) or flash memory (flash RAM). Memory is an example of a computer-readable medium.
Computer-readable media, including both non-transitory and non-transitory, removable and non-removable media, may implement information storage by any method or technology. The information may be computer readable instructions, data structures, modules of a program, or other data. Examples of computer storage media include, but are not limited to, phase change memory (PRAM), Static Random Access Memory (SRAM), Dynamic Random Access Memory (DRAM), other types of Random Access Memory (RAM), Read Only Memory (ROM), Electrically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, compact disc read only memory (CD-ROM), Digital Versatile Discs (DVD) or other optical storage, magnetic cassettes, magnetic disk storage, quantum memory, graphene-based storage media or other magnetic storage devices, or any other non-transmission medium that can be used to store information that can be accessed by a computing device. As defined herein, a computer readable medium does not include a transitory computer readable medium such as a modulated data signal and a carrier wave.
It should also be noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
The foregoing description has been directed to specific embodiments of this disclosure. Other embodiments are within the scope of the following claims. In some cases, the actions or steps recited in the claims may be performed in a different order than in the embodiments and still achieve desirable results. In addition, the processes depicted in the accompanying figures do not necessarily require the particular order shown, or sequential order, to achieve desirable results. In some embodiments, multitasking and parallel processing may also be possible or may be advantageous.
The terminology used in the description of the one or more embodiments is for the purpose of describing the particular embodiments only and is not intended to be limiting of the description of the one or more embodiments. As used in one or more embodiments of the present specification and the appended claims, the singular forms "a," "an," and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It should also be understood that the term "and/or" as used herein refers to and encompasses any and all possible combinations of one or more of the associated listed items.
It should be understood that although the terms first, second, third, etc. may be used in one or more embodiments of the present description to describe various information, such information should not be limited to these terms. These terms are only used to distinguish one type of information from another. For example, first information may also be referred to as second information, and similarly, second information may also be referred to as first information, without departing from the scope of one or more embodiments herein. The word "if" as used herein may be interpreted as "at … …" or "when … …" or "in response to a determination", depending on the context.
The above description is only for the purpose of illustrating the preferred embodiments of the one or more embodiments of the present disclosure, and is not intended to limit the scope of the one or more embodiments of the present disclosure, and any modifications, equivalent substitutions, improvements, etc. made within the spirit and principle of the one or more embodiments of the present disclosure should be included in the scope of the one or more embodiments of the present disclosure.

Claims (24)

1. A voice broadcast method, comprising:
performing identity recognition on the target object;
acquiring a voice file corresponding to the target object according to the identification result of the target object; when the original name of the target object contains polyphonic elements, the voice file is generated by a replaced name, and the replaced name is obtained by replacing the polyphonic elements in the original name with specified single-tone elements;
and playing the acquired voice file.
2. The method of claim 1, wherein obtaining the voice file corresponding to the target object comprises:
acquiring the replaced name of the target object;
and generating the voice file according to the replaced name.
3. The method of claim 2, wherein the obtaining the replaced name of the target object comprises:
determining the original name of the target object according to the identification result of the target object; acquiring a replaced name of the target object according to a mapping relation between a locally stored original name and the replaced name; or the like, or, alternatively,
determining the original name of the target object according to the identification result of the target object; and sending an inquiry request containing the original name to a server, and receiving a replaced name returned by the server, wherein the replaced name is determined by the server according to the stored mapping relation between the original name and the replaced name.
4. The method of claim 1, wherein obtaining the voice file corresponding to the target object comprises:
and acquiring a pre-generated voice file corresponding to the target object.
5. The method of claim 4, further comprising:
generating the voice file in advance according to the replaced name of the target object; or the like, or, alternatively,
and the receiving server generates the voice file in advance according to the replaced name of the target object.
6. The method of claim 1, further comprising:
and displaying the original name of the target object.
7. The method of claim 1, wherein the identifying for the target object comprises:
performing identity recognition on the target object by at least one of the following ways: fingerprint identification, face identification, iris identification and gait identification.
8. A method for setting voice broadcast content is characterized by comprising the following steps:
initiating an object adding request to a server, wherein the object adding request is used for adding a target object as a recognizable object of an identity recognition device, and an original name of the target object contains a polyphonic element;
acquiring a plurality of alternative items returned by the server, wherein the alternative items respectively correspond to the pronunciations of the polyphonic elements;
determining a designated tone element corresponding to the selected alternative, wherein the designated tone element is used for generating a voice file corresponding to the target object, and playing the voice file to the target user after the target user is successfully recognized by the identity recognition device.
9. A voice broadcast method, comprising:
receiving an inquiry request from an identity recognition device, wherein the inquiry request comprises an original name of a target object recognized by the identity recognition device, and the original name comprises a polyphonic element;
returning the replaced name corresponding to the original name so that the identity recognition equipment generates a voice file according to the replaced name and plays the voice file; or returning the voice file generated according to the replaced name to be played by the identity recognition equipment; wherein the replaced name is obtained by replacing a polyphonic element in the original name with a designated monophonic element.
10. A method for setting voice broadcast content is characterized by comprising the following steps:
receiving an object adding request initiated by an administrator, wherein the object adding request is used for adding a target object as a recognizable object of an identity recognition device, and an original name of the target object contains a polyphonic element;
returning a plurality of alternatives to the administrator, the plurality of alternatives corresponding to respective readings of the polyphonic elements;
determining a designated tone element corresponding to the selected alternative, wherein the designated tone element is used for generating a voice file corresponding to the target object, and playing the voice file to the target user after the target user is successfully recognized by the identity recognition device.
11. The method of claim 10, wherein returning a plurality of alternatives to the administrator comprises:
matching each component element contained in the original name with a predefined polyphonic element library respectively to identify polyphonic elements contained in the original name and all pronunciations thereof;
matching each pronunciation of the identified polyphonic elements with a predefined monophonic element library respectively to determine the monophonic elements corresponding to each pronunciation respectively;
and returning the determined options corresponding to the single-tone elements to the administrator.
12. The method of claim 10, further comprising:
establishing a mapping relation between the original name and the replaced name so as to return the replaced name or a voice file generated according to the replaced name when receiving an inquiry request containing the original name initiated by the identity recognition equipment; or the like, or, alternatively,
establishing a mapping relation between the original name and the voice file so as to return the voice file when receiving an inquiry request containing the original name initiated by the identity recognition equipment;
wherein the replaced name is obtained by replacing a polyphonic element in the original name with the specified monophonic element.
13. A community member adding method, comprising:
acquiring a joining request initiated by a user aiming at a group, wherein personal information of the user comprises an original name and a specified single-tone element corresponding to the original name, and the specified single-tone element has the same pronunciation as a polyphonic element in the original name;
and when the administrator of the community agrees to the joining request, adding the user as a member of the community, so that the original name and the specified single-tone element are updated to the member information corresponding to the user.
14. A display method, comprising:
acquiring content to be displayed, wherein the content to be displayed comprises an original name of a target object and a specified single-tone element corresponding to the original name, and the specified single-tone element and a multi-tone element in the original name have the same pronunciation;
and displaying the content to be displayed.
15. A voice broadcast device, comprising:
the identification unit identifies the target object;
the acquisition unit is used for acquiring a voice file corresponding to the target object according to the identification result of the target object; when the original name of the target object contains polyphonic elements, the voice file is generated by a replaced name, and the replaced name is obtained by replacing the polyphonic elements in the original name with specified single-tone elements;
and the playing unit plays the acquired voice file.
16. The utility model provides a setting device of voice broadcast content which characterized in that includes:
the system comprises a request unit and a control unit, wherein the request unit initiates an object adding request to a server, the object adding request is used for adding a target object as a recognizable object of an identity recognition device, and the original name of the target object contains polyphonic elements;
the acquisition unit is used for acquiring a plurality of alternative items returned by the server, wherein the alternative items respectively correspond to the pronunciations of the polyphonic elements;
and the determining unit is used for determining a specified tone element corresponding to the selected alternative, wherein the specified tone element is used for generating a voice file corresponding to the target object so as to play the voice file to the target user after the target user is successfully identified by the identity recognition device.
17. A voice broadcast device, comprising:
the system comprises a receiving unit, a processing unit and a processing unit, wherein the receiving unit is used for receiving an inquiry request from an identity recognition device, the inquiry request comprises an original name of a target object recognized by the identity recognition device, and the original name comprises a polyphonic element;
the return unit returns the replaced name corresponding to the original name so that the identity recognition equipment generates a voice file according to the replaced name and plays the voice file; or returning the voice file generated according to the replaced name to be played by the identity recognition equipment; wherein the replaced name is obtained by replacing a polyphonic element in the original name with a designated monophonic element.
18. The utility model provides a setting device of voice broadcast content which characterized in that includes:
the system comprises a receiving unit, a judging unit and a processing unit, wherein the receiving unit is used for receiving an object adding request initiated by an administrator, the object adding request is used for adding a target object as a recognizable object of an identity recognition device, and the original name of the target object contains polyphonic elements;
a returning unit that returns a plurality of alternatives to the administrator, the plurality of alternatives corresponding to respective readings of the polyphonic elements, respectively;
and the determining unit is used for determining a specified tone element corresponding to the selected alternative, wherein the specified tone element is used for generating a voice file corresponding to the target object so as to play the voice file to the target user after the target user is successfully identified by the identity recognition device.
19. An adding apparatus of community members, comprising:
the system comprises an acquisition unit, a processing unit and a processing unit, wherein the acquisition unit is used for acquiring a joining request initiated by a user aiming at a group, the personal information of the user comprises an original name and a specified single-tone element corresponding to the original name, and the specified single-tone element has the same pronunciation with a polyphonic element in the original name;
and the adding unit is used for adding the user as a member of the community when the administrator of the community agrees to the joining request, so that the original name and the specified single-tone element are updated to the member information corresponding to the user.
20. A display device, comprising:
an acquisition unit that acquires a content to be displayed, the content to be displayed including an original name of a target object and a specified monophonic element corresponding to the original name, the specified monophonic element having the same pronunciation as a polyphonic element in the original name;
and the display unit is used for displaying the content to be displayed.
21. An electronic device, comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor implements the method of any one of claims 1-8 by executing the executable instructions.
22. A computer-readable storage medium having stored thereon computer instructions, which, when executed by a processor, carry out the steps of the method according to any one of claims 1 to 8.
23. An electronic device, comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor implements the method of any one of claims 9-12 by executing the executable instructions.
24. A computer-readable storage medium having stored thereon computer instructions, which, when executed by a processor, carry out the steps of the method according to any one of claims 9-12.
CN201910613779.7A 2019-07-09 2019-07-09 Voice broadcasting and voice broadcasting content setting method and device Active CN112291281B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910613779.7A CN112291281B (en) 2019-07-09 2019-07-09 Voice broadcasting and voice broadcasting content setting method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910613779.7A CN112291281B (en) 2019-07-09 2019-07-09 Voice broadcasting and voice broadcasting content setting method and device

Publications (2)

Publication Number Publication Date
CN112291281A true CN112291281A (en) 2021-01-29
CN112291281B CN112291281B (en) 2023-11-03

Family

ID=74418994

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910613779.7A Active CN112291281B (en) 2019-07-09 2019-07-09 Voice broadcasting and voice broadcasting content setting method and device

Country Status (1)

Country Link
CN (1) CN112291281B (en)

Citations (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090182559A1 (en) * 2007-10-08 2009-07-16 Franz Gerl Context sensitive multi-stage speech recognition
US20110093261A1 (en) * 2009-10-15 2011-04-21 Paul Angott System and method for voice recognition
US8510112B1 (en) * 2006-08-31 2013-08-13 At&T Intellectual Property Ii, L.P. Method and system for enhancing a speech database
CN103905294A (en) * 2012-12-30 2014-07-02 北京神州泰岳软件股份有限公司 Control method for adding instant messaging group members, server and client side equipment
CN105786880A (en) * 2014-12-24 2016-07-20 中兴通讯股份有限公司 Voice recognition method, client and terminal device
CN105844247A (en) * 2016-03-25 2016-08-10 杭州信鸽金融信息服务股份有限公司 Bi-camera cabinet machine and face recognition and second-generation ID card identification system
US9548825B1 (en) * 2013-12-19 2017-01-17 Jelli, Inc. Audio event triggers for broadcast
CN106600767A (en) * 2016-11-29 2017-04-26 深圳智乐信息科技有限公司 Verification method and system based on identification card
CN106815906A (en) * 2016-11-29 2017-06-09 深圳智乐信息科技有限公司 A kind of intelligent entrance guard management method and system based on optic communication
CN106850794A (en) * 2017-01-24 2017-06-13 北京百度网讯科技有限公司 Method and apparatus, server and terminal for operating group
CN107972028A (en) * 2017-07-28 2018-05-01 北京物灵智能科技有限公司 Man-machine interaction method, device and electronic equipment
CN108363817A (en) * 2018-03-21 2018-08-03 广东欧珀移动通信有限公司 Document handling method, device, mobile terminal and computer readable storage medium
US20180330728A1 (en) * 2017-05-11 2018-11-15 Google Inc. Detecting and suppressing voice queries
CN108874356A (en) * 2018-05-31 2018-11-23 珠海格力电器股份有限公司 voice broadcast method, device, mobile terminal and storage medium
CN108989341A (en) * 2018-08-21 2018-12-11 平安科技(深圳)有限公司 The autonomous register method of voice, device, computer equipment and storage medium
CN109147831A (en) * 2018-09-26 2019-01-04 深圳壹账通智能科技有限公司 A kind of voice connection playback method, terminal device and computer readable storage medium
CN109299471A (en) * 2018-11-05 2019-02-01 广州百田信息科技有限公司 A kind of method, apparatus and terminal of text matches
CN109308892A (en) * 2018-10-25 2019-02-05 百度在线网络技术(北京)有限公司 Voice synthesized broadcast method, apparatus, equipment and computer-readable medium
CN109325771A (en) * 2018-09-20 2019-02-12 北京得意音通技术有限责任公司 Auth method, device, computer program, storage medium and electronic equipment
CN109461459A (en) * 2018-12-07 2019-03-12 平安科技(深圳)有限公司 Speech assessment method, apparatus, computer equipment and storage medium
CN109545183A (en) * 2018-11-23 2019-03-29 北京羽扇智信息科技有限公司 Text handling method, device, electronic equipment and storage medium
CN109616111A (en) * 2018-12-24 2019-04-12 北京恒泰实达科技股份有限公司 A kind of scene interactivity control method based on speech recognition
CN109686361A (en) * 2018-12-19 2019-04-26 深圳前海达闼云端智能科技有限公司 A kind of method, apparatus of speech synthesis calculates equipment and computer storage medium
CN109861904A (en) * 2019-02-19 2019-06-07 天津字节跳动科技有限公司 Nametags display methods and device
CN109933215A (en) * 2019-01-31 2019-06-25 平安科技(深圳)有限公司 Phonetic transcriptions of Chinese characters conversion method, device, terminal and computer readable storage medium

Patent Citations (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8510112B1 (en) * 2006-08-31 2013-08-13 At&T Intellectual Property Ii, L.P. Method and system for enhancing a speech database
US20090182559A1 (en) * 2007-10-08 2009-07-16 Franz Gerl Context sensitive multi-stage speech recognition
US20110093261A1 (en) * 2009-10-15 2011-04-21 Paul Angott System and method for voice recognition
CN103905294A (en) * 2012-12-30 2014-07-02 北京神州泰岳软件股份有限公司 Control method for adding instant messaging group members, server and client side equipment
US9548825B1 (en) * 2013-12-19 2017-01-17 Jelli, Inc. Audio event triggers for broadcast
CN105786880A (en) * 2014-12-24 2016-07-20 中兴通讯股份有限公司 Voice recognition method, client and terminal device
CN105844247A (en) * 2016-03-25 2016-08-10 杭州信鸽金融信息服务股份有限公司 Bi-camera cabinet machine and face recognition and second-generation ID card identification system
CN106600767A (en) * 2016-11-29 2017-04-26 深圳智乐信息科技有限公司 Verification method and system based on identification card
CN106815906A (en) * 2016-11-29 2017-06-09 深圳智乐信息科技有限公司 A kind of intelligent entrance guard management method and system based on optic communication
CN106850794A (en) * 2017-01-24 2017-06-13 北京百度网讯科技有限公司 Method and apparatus, server and terminal for operating group
US20180330728A1 (en) * 2017-05-11 2018-11-15 Google Inc. Detecting and suppressing voice queries
CN107972028A (en) * 2017-07-28 2018-05-01 北京物灵智能科技有限公司 Man-machine interaction method, device and electronic equipment
CN108363817A (en) * 2018-03-21 2018-08-03 广东欧珀移动通信有限公司 Document handling method, device, mobile terminal and computer readable storage medium
CN108874356A (en) * 2018-05-31 2018-11-23 珠海格力电器股份有限公司 voice broadcast method, device, mobile terminal and storage medium
CN108989341A (en) * 2018-08-21 2018-12-11 平安科技(深圳)有限公司 The autonomous register method of voice, device, computer equipment and storage medium
CN109325771A (en) * 2018-09-20 2019-02-12 北京得意音通技术有限责任公司 Auth method, device, computer program, storage medium and electronic equipment
CN109147831A (en) * 2018-09-26 2019-01-04 深圳壹账通智能科技有限公司 A kind of voice connection playback method, terminal device and computer readable storage medium
CN109308892A (en) * 2018-10-25 2019-02-05 百度在线网络技术(北京)有限公司 Voice synthesized broadcast method, apparatus, equipment and computer-readable medium
CN109299471A (en) * 2018-11-05 2019-02-01 广州百田信息科技有限公司 A kind of method, apparatus and terminal of text matches
CN109545183A (en) * 2018-11-23 2019-03-29 北京羽扇智信息科技有限公司 Text handling method, device, electronic equipment and storage medium
CN109461459A (en) * 2018-12-07 2019-03-12 平安科技(深圳)有限公司 Speech assessment method, apparatus, computer equipment and storage medium
CN109686361A (en) * 2018-12-19 2019-04-26 深圳前海达闼云端智能科技有限公司 A kind of method, apparatus of speech synthesis calculates equipment and computer storage medium
CN109616111A (en) * 2018-12-24 2019-04-12 北京恒泰实达科技股份有限公司 A kind of scene interactivity control method based on speech recognition
CN109933215A (en) * 2019-01-31 2019-06-25 平安科技(深圳)有限公司 Phonetic transcriptions of Chinese characters conversion method, device, terminal and computer readable storage medium
CN109861904A (en) * 2019-02-19 2019-06-07 天津字节跳动科技有限公司 Nametags display methods and device

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
S S M SAQQUAF等: "Dynamically automated interactive voice response system for smart city surveillance", 《2016 IEEE INTERNATIONAL CONFERENCE ON RECENT TRENDS IN ELECTRONICS ,INFORMATION & COMMUNICATION TECHNOLOGY(RTEICT)》 *
何鑫: "基于HMM的单元挑选语音合成方法研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》, no. 11 *
周建国,臧亮,杨皓,周万江,晏蒲柳: "基于Linux平台的电话语音邮件系统", no. 01 *
张子荣,初敏: "解决多音字字-音转换的一种统计学习方法", no. 03 *

Also Published As

Publication number Publication date
CN112291281B (en) 2023-11-03

Similar Documents

Publication Publication Date Title
CN107430858B (en) Communicating metadata identifying a current speaker
US11698261B2 (en) Method, apparatus, computer device and storage medium for determining POI alias
US20220269529A1 (en) Task completion through inter-application communication
TWI730524B (en) Identity authentication method, device and system
CN110020009B (en) Online question and answer method, device and system
US11907990B2 (en) Desirability of product attributes
CN103377652B (en) A kind of method, device and equipment for carrying out speech recognition
CN104794122A (en) Position information recommending method, device and system
CN101884038A (en) Associating metadata with media objects using time
CN114787814A (en) Reference resolution
CN112351350B (en) Content display method, device, system, equipment and storage medium
CN110392068A (en) A kind of data transmission method, device and its equipment
CN106033418A (en) A voice adding method and device, a voice play method and device, a picture classifying method and device, and a picture search method and device
US20150278907A1 (en) User Inactivity Aware Recommendation System
CN114490975B (en) User question labeling method and device
WO2017024684A1 (en) User behavioral intent acquisition method, device and equipment, and computer storage medium
CN108595141A (en) Pronunciation inputting method and device, computer installation and computer readable storage medium
US11257029B2 (en) Pickup article cognitive fitment
CN112291281B (en) Voice broadcasting and voice broadcasting content setting method and device
CN110781191B (en) Processing method of layout data and server
US11755633B2 (en) Entity search system
CN110619086B (en) Method and apparatus for processing information
CN111695936B (en) Information binding method, device and equipment
CN111724187A (en) DMP audience data real-time processing method and device and computer readable storage medium
CN112948219A (en) Link tracking method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
REG Reference to a national code

Ref country code: HK

Ref legal event code: DE

Ref document number: 40045089

Country of ref document: HK

SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant