CN113407850B - Method and device for determining and acquiring virtual image and electronic equipment - Google Patents

Method and device for determining and acquiring virtual image and electronic equipment Download PDF

Info

Publication number
CN113407850B
CN113407850B CN202110801937.9A CN202110801937A CN113407850B CN 113407850 B CN113407850 B CN 113407850B CN 202110801937 A CN202110801937 A CN 202110801937A CN 113407850 B CN113407850 B CN 113407850B
Authority
CN
China
Prior art keywords
target
keyword
avatar
virtual image
determining
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110801937.9A
Other languages
Chinese (zh)
Other versions
CN113407850A (en
Inventor
彭昊天
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baidu Netcom Science and Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN202110801937.9A priority Critical patent/CN113407850B/en
Publication of CN113407850A publication Critical patent/CN113407850A/en
Application granted granted Critical
Publication of CN113407850B publication Critical patent/CN113407850B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9535Search customisation based on user profiles and personalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9536Search customisation based on social or collaborative filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/01Social networking

Landscapes

  • Engineering & Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Resources & Organizations (AREA)
  • Health & Medical Sciences (AREA)
  • Economics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Marketing (AREA)
  • Primary Health Care (AREA)
  • Strategic Management (AREA)
  • Tourism & Hospitality (AREA)
  • General Business, Economics & Management (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The disclosure provides a method and a device for determining an avatar, a method and a device for acquiring the avatar, an electronic device, a readable storage medium and a computer program product, and relates to the field of artificial intelligence such as augmented reality, computer vision and deep learning. The specific implementation scheme is as follows: acquiring virtual image acquisition request information sent by a user side; determining at least one target keyword corresponding to the avatar acquisition request information, wherein the at least one target keyword comprises a keyword for describing an avatar characteristic of the avatar; and determining a target virtual image corresponding to the at least one target keyword. According to the scheme, the target virtual image corresponding to at least one target keyword can be determined according to the at least one target keyword for describing the image characteristics of the virtual image, and the virtual image is not required to be selected through a human-computer interaction interface. The complexity of determining the virtual image is reduced, and the user experience can be improved.

Description

Method and device for determining and acquiring virtual image and electronic equipment
Technical Field
The present disclosure relates to the field of artificial intelligence, and more particularly to computer vision and image processing techniques, which can be used in the fields of artificial intelligence such as augmented reality, computer vision, and deep learning.
Background
The virtual image is widely applied to scenes such as social contact, live broadcast and games. In the existing technology for determining an Avatar, in order to generate a personalized Avatar according to the requirements of a user, the user is often required to perform an Avatar selection operation through a human-machine interaction interface by using a PTA (Photo-to-Avatar) technology.
However, the process of determining the avatar is often complicated by the selection operation of the avatar through the human-computer interface, thereby reducing the user experience of the user.
Disclosure of Invention
The present disclosure provides a determination method of an avatar, a determination apparatus of an avatar, an acquisition method of an avatar, an acquisition apparatus of an avatar, an electronic device, a readable storage medium, and a computer program product to reduce complexity of determining an avatar.
According to an aspect of the present disclosure, there is provided a method of determining an avatar, the method may include the steps of:
acquiring virtual image acquisition request information sent by a user side;
determining at least one target keyword corresponding to the avatar acquisition request information, wherein the at least one target keyword comprises a keyword for describing an avatar characteristic of the avatar;
and determining a target virtual image corresponding to the at least one target keyword.
According to another aspect of the present disclosure, there is provided an avatar acquisition method, which may include the steps of:
sending virtual image acquisition request information to a server;
acquiring a target virtual image sent by the server aiming at the virtual image acquisition request information; the target virtual image is determined by the server side according to at least one target keyword corresponding to the virtual image acquisition request information; the at least one target keyword is a keyword used for describing the image characteristics of the virtual image in the voice information.
According to a third aspect of the present disclosure, there is provided an avatar determination apparatus, which may include:
the request information acquisition module is used for acquiring the virtual image acquisition request information sent by the user side;
a target keyword determining module, configured to determine at least one target keyword corresponding to the avatar acquisition request message, where the at least one target keyword includes a keyword for describing an avatar feature of the avatar;
and the target virtual image determining module is used for determining the target virtual image corresponding to the at least one target keyword.
According to a fourth aspect of the present disclosure, there is provided an avatar acquisition apparatus, which may include:
the request information sending module is used for sending the virtual image acquisition request information to the server;
the target virtual image acquisition module is used for acquiring a target virtual image sent by the server aiming at the virtual image acquisition request information; the target virtual image is determined by the server side according to at least one target keyword corresponding to the virtual image acquisition request information; the at least one target keyword is a keyword used for describing the image characteristics of the virtual image in the voice information.
According to another aspect of the present disclosure, there is provided an electronic device including:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein the content of the first and second substances,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform a method according to any one of the embodiments of the present disclosure.
According to another aspect of the present disclosure, there is provided a non-transitory computer readable storage medium having stored thereon computer instructions for causing a computer to perform a method in any of the embodiments of the present disclosure.
According to another aspect of the present disclosure, a computer program product is provided, comprising a computer program which, when executed by a processor, implements the method in any of the embodiments of the present disclosure.
According to the technology disclosed by the invention, the target virtual image corresponding to at least one target keyword can be determined according to the at least one target keyword for describing the image characteristics of the virtual image, and the virtual image is not required to be selected through a human-computer interaction interface. The complexity of determining the virtual image is reduced, and the user experience of the user can be improved.
It should be understood that the statements in this section are not intended to identify key or critical features of the embodiments of the present disclosure, nor are they intended to limit the scope of the present disclosure. Other features of the present disclosure will become apparent from the following description.
Drawings
The drawings are included to provide a better understanding of the present solution and are not to be construed as limiting the present disclosure. Wherein:
fig. 1 is a flowchart of a method for determining an avatar according to an embodiment of the present disclosure;
fig. 2 is a flowchart of a method for determining a target avatar provided in an embodiment of the present disclosure;
fig. 3 is a flowchart of a method for determining a name of a target person provided in an embodiment of the present disclosure;
fig. 4 is a schematic diagram of a second correspondence list provided in an embodiment of the disclosure;
FIG. 5 is a flow chart of another method of determining a target avatar provided in embodiments of the present disclosure;
FIG. 6 is a flowchart of a method for obtaining an avatar according to an embodiment of the present disclosure;
fig. 7 is a flowchart of another avatar acquisition method provided in an embodiment of the present disclosure;
FIG. 8 is a schematic diagram of an avatar determination apparatus according to an embodiment of the present disclosure;
fig. 9 is a schematic diagram of an avatar acquisition device according to an embodiment of the present disclosure;
fig. 10 is a schematic diagram of an electronic device according to an embodiment of the disclosure.
Detailed Description
Exemplary embodiments of the present disclosure are described below with reference to the accompanying drawings, in which various details of the embodiments of the disclosure are included to assist understanding, and which are to be considered as merely exemplary. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the present disclosure. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
The present disclosure provides a method for determining an avatar, specifically referring to fig. 1, which is a flowchart of a method for determining an avatar according to an embodiment of the present disclosure. The method may comprise the steps of:
step S101: acquiring the virtual image acquisition request information sent by the user side.
Step S102: determining at least one target keyword corresponding to the virtual image acquisition request information, wherein the at least one target keyword comprises a keyword for describing image characteristics of the virtual image;
step S103: and determining a target virtual image corresponding to at least one target keyword.
The method for determining the virtual image provided in the embodiment of the disclosure is a server side in which an execution main body provides services such as data processing and storage for a user side. The specific implementation manner of the server is generally a server or a server cluster.
The user terminal may specifically be a user terminal running the target program or application, and the user terminal is implemented as an electronic device installed with the target program or application.
Electronic devices include, but are not limited to, cell phones, tablet computers, notebook computers, and desktop computers.
The so-called target program or application may be a program or application capable of implementing the avatar acquisition function. Including, but not limited to, application programs, computer applications, computer programs, and web application or software.
The method for determining the avatar provided in the embodiments of the present disclosure can determine the target avatar corresponding to at least one target keyword according to the at least one target keyword for describing the avatar characteristics of the avatar, without performing an avatar selection operation through a human-computer interaction interface. The complexity of determining the virtual image is reduced, and the user experience of the user can be improved.
The avatar acquisition request information may be request information including at least one target keyword. That is, the user side sends the avatar acquisition request message carrying at least one target keyword.
The at least one target keyword is used to describe at least one keyword in the target voice information of the avatar or at least one keyword in the target text information of the avatar.
The target voice information is generally voice information collected by a user terminal through an audio collecting device. The target text information is generally text information input by a user through a human-computer interaction interface at a user terminal.
Under the condition that the virtual image acquisition request information comprises at least one target keyword, the specific implementation mode of determining the at least one target keyword corresponding to the virtual image acquisition request information is as follows: and analyzing the request information acquired by the virtual image to obtain at least one target keyword.
At this time, the server only needs to analyze the avatar acquisition request information to obtain at least one target keyword. The determination process of the target avatar can be made easier.
In order to reduce the complexity of data processing of the user terminal and improve the operation performance of the user terminal. In an embodiment of the present disclosure, the avatar acquisition request message sent by the user side may be a request message including target voice information for describing the avatar, the voice information including at least one target keyword.
Under the condition that the virtual character acquisition request information comprises target voice information used for describing the virtual character, the specific implementation mode of determining at least one target keyword corresponding to the virtual character acquisition request information is as follows: firstly, acquiring request information aiming at the virtual image, and determining voice information; and then, extracting keywords aiming at the voice information to obtain at least one target keyword.
The process of determining the voice information for acquiring the request information of the avatar is as follows: and analyzing the information of the virtual image acquisition request information to acquire the voice information carried in the virtual image acquisition request information.
The process of extracting keywords from the speech information to obtain at least one target keyword generally includes: firstly, converting voice information into text information; and then, extracting keywords aiming at the text information to obtain at least one target keyword.
The specific implementation mode can be as follows: and inputting the voice information into a pre-trained keyword extraction model to obtain at least one target keyword. The pre-trained keyword extraction model is obtained by pre-training the keyword extraction model according to the sample voice information and the corresponding labeled keyword extraction result.
By avatar is generally meant a three-dimensional avatar such as: a virtual player image in a game, a virtual anchor image in a news live broadcast, or a virtual teacher image in online education, etc.
The character feature generally refers to a feature such as an appearance, a figure, or a character name of an avatar. The target keywords include, but are not limited to, keywords describing the character names of the avatar and keywords describing shape appearance features, patterns.
The keywords describing the character names of the avatar are: zhang San, Li Si, etc.; the key words for describing shape appearance features and patterns, such as handsome, big eyes, high nose bridge or cherry small mouth, etc.
In the embodiment of the present disclosure, an implementation process of determining a target avatar is shown in fig. 2, which is a flowchart of a method for determining a target avatar provided in the embodiment of the present disclosure.
Step S201: and determining the target person name corresponding to at least one target keyword.
Step S202: and determining the target virtual image in the virtual image library by utilizing the first corresponding relation list according to the name of the target person.
The first correspondence list includes the correspondence between the name of the target person and the target avatar. The first correspondence list may be preconfigured
In the first correspondence list, different avatars corresponding to different person names respectively, that is, each person name in the first correspondence list corresponds to an avatar.
In the embodiment of the disclosure, a target character name corresponding to at least one target keyword is determined, and then a corresponding target avatar is determined according to the target character name. The target avatar can be ensured to conform to the image of a specific character, thereby enabling the target avatar to have strong pertinence.
The avatar library may be generated in advance, and the generated avatar library may be generated by the following steps:
first, a search engine searches for face images corresponding to a plurality of character name keywords by designating the character name keywords in advance.
Secondly, after obtaining the authorization permission of the related personnel, respectively generating an avatar corresponding to each of the plurality of character name keywords by utilizing the PTA technology based on the face images corresponding to the plurality of character name keywords;
and thirdly, generating a pre-generated virtual image library based on the virtual image corresponding to each character name keyword.
And respectively generating a virtual image corresponding to each of the plurality of character name keywords by utilizing a PTA technology based on the face images corresponding to the plurality of character name keywords. The design cost of the avatar can be reduced and an avatar library having rich avatars can be generated.
In order to ensure that the target person name is determined simply and directly, when the target keyword includes the person name keyword, a specific implementation manner of determining the target person name corresponding to at least one target keyword is as follows: firstly, selecting a character name keyword in at least one target keyword; then, the person name keyword is set as the target person name.
It should be noted that, the avatar in the embodiment of the present disclosure is not an avatar for a specific user, and cannot reflect personal information of a specific user. The construction of the virtual image is performed after the authorization of the user, and the construction process conforms to the relevant laws and regulations.
In addition, the face image in the embodiment of the present disclosure may be from a public data set, or the face image may be obtained by authorization of a user corresponding to the face image.
At this time, the target keyword describes a keyword of the character name of the avatar.
Specifically, if there is a character name keyword "zhang san" in the target keyword, the "zhang san" is used as the target character name.
When the target keywords do not include the character name keywords, the target character name is accurately determined so as to ensure the accuracy of the target avatar. In the embodiment of the present disclosure, the method of determining the name of the target person as shown in fig. 3 may also be adopted to determine the name of the target person. Fig. 3 is a flowchart of a method for determining a name of a target person according to an embodiment of the disclosure.
Step S301: determining at least one character name corresponding to each target keyword by utilizing the second corresponding relation list according to each target keyword in the at least one target keyword; the second corresponding relation list comprises corresponding relations between each target keyword and at least one corresponding character name.
Step S302: and determining the person name with the largest occurrence frequency in at least one person name corresponding to each target keyword.
Step S303: the person name with the largest number of occurrences is taken as the target person name.
At this time, the target keyword is described as a keyword describing the shape appearance feature, a pattern.
Please refer to fig. 4, which is a diagram illustrating a second mapping relationship list according to an embodiment of the disclosure. Each target keyword in fig. 4 corresponds to at least one character name, and if "zhang san" and "lie si" both have a high nose bridge, if one of the target keywords is "high nose bridge", it is determined that the character name corresponding to the target keyword is "zhang san" and "lie si"; the following steps are repeated: if all the three-piece keywords and the five-piece keywords have big eyes, when one of the target keywords is the big eye keyword, the character names corresponding to the target keywords are determined to be the three-piece keywords and the five-piece keywords.
The second correspondence list may be a preconfigured list.
The implementation manner of determining the name with the largest occurrence frequency in at least one name corresponding to each target keyword is as follows: if at least one target keyword is "high nose bridge" and "big eye", then at least one character name corresponding to each target keyword has "zhang san", "lie si", "zhang san", and "wang wu". Wherein, the occurrence frequency of Zhang three is twice, each of Li four and Wang five is once.
The implementation manner of using the person name with the largest occurrence number as the target person name is as follows: "Zhang three" is taken as the name of the target person.
In the embodiment of the present disclosure, an implementation process for determining a target avatar is shown in fig. 4, which is a flowchart of another method for determining a target avatar provided in the embodiment of the present disclosure.
In the embodiment of the present disclosure, one implementation process for determining a target avatar is shown in fig. 5, which is a flowchart of another determination method for a target avatar provided in the embodiment of the present disclosure.
Step S501: determining at least one virtual image corresponding to each target keyword by utilizing the third corresponding relation list according to each target keyword in the at least one target keyword; the third corresponding relation list comprises the corresponding relation between each target keyword and at least one corresponding virtual image.
Step S502: and determining the avatar with the maximum occurrence frequency in at least one avatar corresponding to each target keyword.
Step S503: and taking the avatar with the largest occurrence number as the target avatar.
In the embodiment of the disclosure, the target virtual image is determined according to each target keyword in the at least one target keyword, so that the target virtual image can be determined more directly and simply.
At this time, the target keyword is described as a keyword describing the shape appearance feature, a pattern.
The third correspondence list may be configured in advance, where the third correspondence list includes a correspondence between each target keyword and at least one corresponding avatar, and means that each target keyword in the third correspondence list corresponds to and exceeds an avatar.
The following at least one keyword comprises three keywords which are respectively: "big eye", "handsome" and "big mouth", the "big eye" in the third correspondence list corresponds to "first avatar", "second avatar" and "third avatar" three avatars, the "handsome" in the third correspondence list corresponds to "first avatar", "second avatar" two avatars, the "big mouth" in the third correspondence list corresponds to "first avatar" an avatar as an example, explain another target avatar's determination method that the embodiment of this disclosure provides:
at this time, the specific implementation manner of determining at least one avatar corresponding to each target keyword is as follows: determining at least one virtual image corresponding to the 'big eye' as a 'first virtual image', a 'second virtual image' and a 'third virtual image', and determining at least one virtual image corresponding to the 'handsome' as: the first virtual image and the second virtual image determine that at least one virtual image corresponding to the big mouth is as follows: "first avatar".
Correspondingly, the process of determining the avatar with the most occurrence times in at least one avatar corresponding to each target keyword is as follows: determining the first avatar, the second avatar and the third avatar as: and determining the first avatar to be the avatar with the largest occurrence number for 3 times, 2 times and 1 time.
Correspondingly, the mode of taking the avatar with the largest occurrence number as the target avatar is as follows: the "first avatar" is taken as the target avatar.
In order to enable the user end to obtain the target avatar for the user to use, the present disclosure further provides an avatar determination method, after determining the target avatar, further comprising: and sending the target virtual image to the user side.
The present disclosure further provides a method for obtaining an avatar, specifically referring to fig. 6, which is a flowchart of a method for obtaining an avatar according to an embodiment of the present disclosure. The method may comprise the steps of:
step S601: and sending the virtual image acquisition request information to the server.
Step S602: acquiring a target virtual image sent by a server aiming at virtual image acquisition request information; the target virtual image is determined by the server according to at least one target keyword corresponding to the virtual image acquisition request information; at least one target keyword is a keyword for describing the character feature of the avatar in the voice message.
According to the method for acquiring the virtual image, the target virtual image corresponding to at least one target keyword can be determined according to the at least one target keyword for describing the image characteristics of the virtual image, and the virtual image is not required to be selected through a human-computer interaction interface. The complexity of determining the virtual image is reduced, and the user experience of the user can be improved.
In the method for acquiring an avatar provided in the embodiments of the present disclosure, the execution main body is a user side for running a preset program or application, and the user side is an electronic device installed with the preset program or application.
Electronic devices include, but are not limited to, cell phones, tablet computers, notebook computers, and desktop computers.
The function of the acquisition method of the avatar provided in the embodiment of the present disclosure may be implemented by a so-called preset program or application. Including, but not limited to, application programs, computer applications, computer programs, and web application or software.
The server provides services such as data processing and storage for the user side. The specific implementation manner of the server is generally a server or a server cluster.
The avatar acquisition request information may be request information including at least one target keyword. That is, the user side sends the avatar acquisition request message carrying at least one target keyword.
The at least one target keyword is used to describe at least one keyword in the target voice information of the avatar or at least one keyword in the target text information of the avatar.
The target voice information is generally voice information collected by a user terminal through an audio collecting device. The target text message is generally a text message input by a user through a human-computer interaction interface at a user terminal.
Under the condition that the virtual image acquisition request information comprises at least one target keyword, the specific implementation mode of determining the at least one target keyword corresponding to the virtual image acquisition request information is as follows: and analyzing the request information acquired by the virtual image to obtain at least one target keyword.
At this time, the server only needs to analyze the avatar acquisition request information to obtain at least one target keyword. The determination process of the target avatar can be made easier.
In order to reduce the complexity of data processing of the user terminal and improve the operation performance of the user terminal. In an embodiment of the present disclosure, the avatar acquisition request message sent by the user side may be a request message including target voice information for describing the avatar, the voice information including at least one target keyword.
Under the condition that the virtual image acquisition request information comprises target voice information used for describing the virtual image, the specific implementation mode of determining at least one target keyword corresponding to the virtual image acquisition request information is as follows: firstly, acquiring request information aiming at the virtual image, and determining voice information; and then, extracting keywords aiming at the voice information to obtain at least one target keyword.
In the embodiment of the present disclosure, a complete flow of the acquisition of the avatar is shown in fig. 7, which is a flow chart of another method for acquiring the avatar provided in the embodiment of the present disclosure.
The user side performs steps S701 to S702, and the server side performs steps S703 to S704. Wherein:
step S701: and collecting target voice information of the user.
Step S702: and sending the virtual image acquisition request information to the server.
Step S703: a target avatar is determined.
Step S704: and providing the target virtual image to the user terminal.
Before executing step S703, the server needs to execute the following steps in advance:
step S703-1: and searching the face images corresponding to the plurality of character name keywords by using the plurality of character name keywords.
Step S703-2: and respectively generating an avatar corresponding to each character name keyword in the plurality of character name keywords by utilizing an avatar generation technology.
Step S703-3: and generating a pre-generated virtual image library based on the virtual image corresponding to each character name keyword.
At this time, the user side is a user side that runs a target program or application, which may be a program or application capable of implementing the avatar acquisition function. The server side is used for providing services such as data processing, data storage and the like for the user side.
As shown in fig. 8, the present disclosure provides an avatar determination apparatus, including:
a request information obtaining module 801, configured to obtain an avatar obtaining request information sent by a user side;
a target keyword determining module 802, configured to determine at least one target keyword corresponding to the avatar acquisition request message, where the at least one target keyword includes a keyword for describing an avatar characteristic of the avatar;
a target avatar determination module 803, configured to determine a target avatar corresponding to the at least one target keyword.
In one embodiment, the target keyword determination module 802 may further include:
the target character name determining submodule is used for determining a target character name corresponding to the at least one target keyword;
the first target virtual image determining submodule is used for determining the target virtual image in a virtual image library by utilizing a first corresponding relation list according to the name of the target person;
the first corresponding relation list comprises the corresponding relation between the name of the target person and the target virtual image.
In one embodiment, the target person name determination sub-module may further include:
a character name keyword selection submodule for selecting a character name keyword from the at least one target keyword;
a first target person name determining sub-module configured to use the person name keyword as the target person name.
In one embodiment, the target person name determination sub-module may further include:
the first person name determining submodule is used for determining at least one person name corresponding to each target keyword by utilizing a second corresponding relation list according to each target keyword in the at least one target keyword; the second corresponding relation list comprises corresponding relations between each target keyword and at least one corresponding character name;
the second person name determining submodule is used for determining the person name with the largest occurrence frequency in at least one person name corresponding to each target keyword;
and the second target person name determining submodule is used for taking the person name with the largest occurrence frequency as the target person name.
In one embodiment, the target avatar determination module 803 may further include:
the first avatar determination sub-module is used for determining at least one avatar corresponding to each target keyword by utilizing a third corresponding relation list according to each target keyword in the at least one target keyword; the third corresponding relation list comprises the corresponding relation between each target keyword and at least one corresponding virtual image;
the second virtual character determining submodule is used for determining the virtual character with the largest occurrence frequency in at least one virtual character corresponding to each target keyword;
and the second target virtual image determining submodule is used for taking the virtual image with the maximum occurrence frequency as the target virtual image.
In an embodiment, in a case that the avatar acquisition request information includes the at least one target keyword, the request information acquisition module 801 is specifically configured to parse the avatar acquisition request information to obtain the at least one target keyword.
In one embodiment, in a case where the avatar acquisition request information includes target voice information for describing an avatar, the request information acquisition module 801 may further include:
the voice information determining submodule is used for obtaining request information aiming at the virtual image and determining the voice information;
and the target keyword determining submodule is used for extracting keywords aiming at the voice information to obtain at least one target keyword.
In one embodiment, the apparatus further comprises: and the target virtual image sending module is used for sending the target virtual image to the user side.
As shown in fig. 9, the present disclosure provides an avatar acquisition apparatus, including:
a request information sending module 901, configured to send an avatar acquisition request information to a server;
a target avatar acquiring module 902, configured to acquire a target avatar sent by the server in response to the avatar acquisition request message; the target virtual image is determined by the server side according to at least one target keyword corresponding to the virtual image acquisition request information; the at least one target keyword is a keyword used for describing the image characteristics of the virtual image in the voice information.
In one embodiment, the avatar acquisition request information includes the at least one target keyword.
In one embodiment, the avatar acquisition request information includes target voice information for describing the avatar, the voice information including the at least one target keyword.
In the technical scheme of the disclosure, the acquisition, storage, application and the like of the personal information of the related user all accord with the regulations of related laws and regulations, and do not violate the good customs of the public order. The present disclosure also provides an electronic device, a readable storage medium, and a computer program product according to embodiments of the present disclosure.
FIG. 10 shows a schematic block diagram of an electronic device 1000 that may be used to implement embodiments of the present disclosure. Electronic devices are intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other appropriate computers. The electronic device may also represent various forms of mobile devices, such as personal digital processing, cellular phones, smart phones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions, are meant to be examples only, and are not intended to limit implementations of the disclosure described and/or claimed herein.
As shown in fig. 10, the electronic device 1000 includes a computing unit 1010 that can perform various appropriate actions and processes according to a computer program stored in a Read Only Memory (ROM)1020 or a computer program loaded from a storage unit 10100 into a Random Access Memory (RAM) 1030. In the RAM 1030, various programs and data required for the operation of the device 1000 can also be stored. The calculation unit 1010, the ROM 1020, and the RAM 1030 are connected to each other by a bus 1040. An input/output (I/O) interface 1050 is also connected to bus 1040.
A number of components in the electronic device 1000 are connected to the I/O interface 1050, including: an input unit 1060 such as a keyboard, a mouse, or the like; an output unit 1070 such as various types of displays, speakers, and the like; a storage unit 1070, such as a magnetic disk, optical disk, or the like; and a communication unit 1090 such as a network card, modem, wireless communication transceiver, or the like. The communication unit 1090 allows the electronic device 1000 to exchange information/data with other devices via a computer network such as the internet and/or various telecommunication networks.
Computing unit 1010 may be a variety of general purpose and/or special purpose processing components with processing and computing capabilities. Some examples of the computing unit 1010 include, but are not limited to, a Central Processing Unit (CPU), a Graphics Processing Unit (GPU), various dedicated Artificial Intelligence (AI) computing chips, various computing units running machine learning model algorithms, a digital signal processor (D step SP), and any suitable processor, controller, microcontroller, and the like. The calculation unit 1010 executes each method and processing described above, such as the determination method of the avatar and/or the acquisition method of the avatar. For example, in some embodiments, the avatar determination method and/or the avatar acquisition method may be implemented as a computer software program tangibly embodied in a machine-readable medium, such as storage unit 1080. In some embodiments, part or all of the computer program may be loaded and/or installed onto the electronic device 1000 via the ROM 1020 and/or the communication unit 1090. When the computer program is loaded into the RAM 1030 and executed by the computing unit 1010, one or more steps of the avatar determination method and/or the avatar acquisition method described above may be performed. Alternatively, in other embodiments, the computing unit 1010 may be configured to perform the avatar determination method and/or the avatar acquisition method in any other suitable manner (e.g., by means of firmware).
Various implementations of the systems and techniques described here above may be implemented in digital electronic circuitry, integrated circuitry, Field Programmable Gate Arrays (FPGAs), application specific integrated circuits (step SIC), application specific standard products (step SP), systems of system on a chip (step SOC), load programmable logic devices (CPLDs), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include: implemented in one or more computer programs that are executable and/or interpretable on a programmable system including at least one programmable processor, which may be special or general purpose, receiving data and instructions from, and transmitting data and instructions to, a storage system, at least one input device, and at least one output device.
Program code for implementing the methods of the present disclosure may be written in any combination of one or more programming languages. These program code may be provided to a processor or controller of a general purpose computer, special purpose computer, or other programmable data processing apparatus, such that the program code, when executed by the processor or controller, causes the functions/acts specified in the flowchart and/or block diagram to be performed. The program code may execute entirely on the machine, partly on the machine, as a stand-alone software package partly on the machine and partly on a remote machine or entirely on the remote machine or server.
In the context of this disclosure, a machine-readable medium may be a tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. A machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a machine-readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
To provide for interaction with a user, the systems and techniques described here can be implemented on a computer having: a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to a user; and a keyboard and a pointing device (e.g., a mouse or a trackball) by which a user can provide input to the computer. Other kinds of devices may also be used to provide for interaction with a user; for example, feedback provided to the user can be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user may be received in any form, including acoustic, speech, or tactile input.
The systems and techniques described here can be implemented in a computing system that includes a back-end component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front-end component (e.g., a user computer having a graphical user interface or a web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such back-end, middleware, or front-end components. The components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include: local Area Networks (LANs), Wide Area Networks (WANs), and the Internet.
The computer system may include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other. The server may be a cloud server, a server of a distributed system, or a server with a combined blockchain.
It should be understood that various forms of the flows shown above may be used, with steps reordered, added, or deleted. For example, the steps described in the present disclosure may be executed in parallel, sequentially, or in different orders, as long as the desired results of the technical solutions disclosed in the present disclosure can be achieved, and the present disclosure is not limited herein.
The above detailed description should not be construed as limiting the scope of the disclosure. It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and substitutions may be made in accordance with design requirements and other factors. Any modification, equivalent replacement, and improvement made within the spirit and principle of the present disclosure should be included in the scope of protection of the present disclosure.

Claims (20)

1. A method for determining an avatar, comprising:
acquiring virtual image acquisition request information sent by a user side;
determining at least one target keyword corresponding to the avatar acquisition request information, wherein the at least one target keyword comprises a keyword for describing an avatar characteristic of the avatar;
determining a target virtual image corresponding to the at least one target keyword; wherein the determining the target avatar corresponding to the at least one target keyword comprises:
determining a target character name corresponding to the at least one target keyword;
determining the target virtual image in an virtual image library by utilizing a first corresponding relation list according to the name of the target person;
the first corresponding relation list comprises the corresponding relation between the name of the target person and the target virtual image;
in a case that the target keyword does not include a person name keyword, the determining a target person name corresponding to the at least one target keyword includes:
determining at least one character name corresponding to each target keyword by utilizing a second corresponding relation list according to each target keyword used for describing shape appearance characteristics and patterns in the at least one target keyword; the second corresponding relation list comprises corresponding relations between each target keyword and at least one corresponding character name;
determining the name with the largest occurrence frequency in at least one character name corresponding to each target keyword;
and taking the person name with the largest occurrence number as the target person name.
2. The method of claim 1, wherein in the case that the target keyword comprises a person name keyword, the determining a target person name corresponding to the at least one target keyword further comprises:
selecting a character name keyword in the at least one target keyword;
and taking the character name keyword as the target character name.
3. The method of claim 1, wherein determining the target avatar corresponding to the at least one target keyword further comprises:
determining at least one virtual image corresponding to each target keyword by utilizing a third corresponding relation list according to each target keyword in the at least one target keyword; the third corresponding relation list comprises the corresponding relation between each target keyword and at least one corresponding virtual image;
determining the avatar with the most occurrence times in at least one avatar corresponding to each target keyword;
and taking the avatar with the maximum occurrence number as the target avatar.
4. The method of claim 1, wherein in case that the avatar acquisition request information includes the at least one target keyword, the determining of the at least one target keyword to which the avatar acquisition request information corresponds includes:
and analyzing the virtual image acquisition request information to obtain the at least one target keyword.
5. The method according to claim 1, wherein in the case where the avatar acquisition request information includes target voice information for describing an avatar, the determining at least one target keyword to which the avatar acquisition request information corresponds includes:
determining the voice information aiming at the request information obtained by the virtual image;
and extracting keywords from the voice information to obtain the at least one target keyword.
6. The method of claim 1, further comprising: and sending the target virtual image to the user side.
7. An avatar acquisition method, comprising:
sending virtual image acquisition request information to a server;
acquiring a target virtual image sent by the server aiming at the virtual image acquisition request information; the target virtual image is determined by the server side according to at least one target keyword corresponding to the virtual image acquisition request information; the at least one target keyword includes a keyword for describing an avatar characteristic of the avatar;
the method for determining the target virtual image by the server side according to the at least one target keyword comprises the following steps:
determining a target character name corresponding to the at least one target keyword;
determining the target virtual image in an virtual image library by utilizing a first corresponding relation list according to the name of the target person;
the first corresponding relation list comprises the corresponding relation between the name of the target person and the target virtual image;
the method for the server side to determine the name of the target person comprises the following steps:
under the condition that the target keywords do not comprise character name keywords, determining at least one character name corresponding to each target keyword by utilizing a second corresponding relation list according to each target keyword used for describing shape appearance characteristics and patterns in the at least one target keyword; the second corresponding relation list comprises corresponding relations between each target keyword and at least one corresponding character name;
determining the name with the largest occurrence frequency in at least one character name corresponding to each target keyword;
and taking the person name with the largest occurrence number as the target person name.
8. The method of claim 7, wherein the avatar acquisition request message includes the at least one target keyword.
9. The method of claim 7, wherein the avatar acquisition request message includes target voice information for describing the avatar, the voice information including the at least one target keyword.
10. An avatar determination apparatus, comprising:
the request information acquisition module is used for acquiring the virtual image acquisition request information sent by the user side;
a target keyword determining module, configured to determine at least one target keyword corresponding to the avatar acquisition request message, where the at least one target keyword includes a keyword for describing an avatar feature of the avatar;
the target virtual image determining module is used for determining a target virtual image corresponding to the at least one target keyword;
wherein, the target keyword determination module comprises:
the target character name determining submodule is used for determining a target character name corresponding to the at least one target keyword;
the first target virtual image determining submodule is used for determining the target virtual image in a virtual image library by utilizing a first corresponding relation list according to the name of the target person;
the first corresponding relation list comprises the corresponding relation between the name of the target person and the target virtual image;
the target person name determination submodule includes:
a first person name determining sub-module, configured to, when the target keywords do not include a person name keyword, determine, according to each target keyword used for describing shape and appearance features and patterns in the at least one target keyword, at least one person name corresponding to each target keyword by using a second correspondence list; the second corresponding relation list comprises corresponding relations between each target keyword and at least one corresponding character name;
the second person name determining submodule is used for determining the person name with the largest occurrence frequency in at least one person name corresponding to each target keyword;
and the second target person name determining submodule is used for taking the person name with the largest occurrence frequency as the target person name.
11. The apparatus of claim 10, wherein the target person name determination sub-module further comprises:
a character name keyword selection submodule for selecting a character name keyword from the at least one target keyword;
a first target person name determining sub-module configured to take the person name keyword as the target person name.
12. The apparatus of claim 10, wherein the target avatar determination module further comprises:
the first virtual image determining sub-module is used for determining at least one virtual image corresponding to each target keyword by utilizing a third corresponding relation list according to each target keyword in the at least one target keyword; the third corresponding relation list comprises the corresponding relation between each target keyword and at least one corresponding virtual image;
the second virtual character determining submodule is used for determining the virtual character with the largest occurrence frequency in at least one virtual character corresponding to each target keyword;
and the second target virtual character determining submodule is used for taking the virtual character with the largest occurrence frequency as the target virtual character.
13. The apparatus according to claim 10, wherein in a case that the avatar acquisition request information includes the at least one target keyword, the request information acquisition module is specifically configured to parse the avatar acquisition request information to obtain the at least one target keyword.
14. The apparatus of claim 10, wherein in the case where the avatar acquisition request information includes target voice information for describing the avatar, the request information acquisition module includes:
the voice information determining submodule is used for obtaining request information aiming at the virtual image and determining the voice information;
and the target keyword determining submodule is used for extracting keywords aiming at the voice information to obtain at least one target keyword.
15. The apparatus of claim 10, further comprising: and the target virtual image sending module is used for sending the target virtual image to the user side.
16. An avatar acquisition apparatus, comprising:
the request information sending module is used for sending the virtual image acquisition request information to the server;
the target virtual image acquisition module is used for acquiring a target virtual image sent by the server aiming at the virtual image acquisition request information; the target virtual image is determined by the server side according to at least one target keyword corresponding to the virtual image acquisition request information; the at least one target keyword includes a keyword for describing an avatar characteristic of the avatar;
the method for determining the target virtual image by the server side according to the at least one target keyword comprises the following steps:
determining a target character name corresponding to the at least one target keyword;
determining the target virtual image in an virtual image library by utilizing a first corresponding relation list according to the name of the target person;
the first corresponding relation list comprises the corresponding relation between the name of the target person and the target virtual image;
the method for the server side to determine the name of the target person comprises the following steps:
under the condition that the target keywords do not comprise character name keywords, determining at least one character name corresponding to each target keyword by utilizing a second corresponding relation list according to each target keyword used for describing shape appearance characteristics and patterns in the at least one target keyword; the second corresponding relation list comprises corresponding relations between each target keyword and at least one corresponding character name;
determining the name with the largest occurrence frequency in at least one character name corresponding to each target keyword;
and taking the character name with the largest occurrence number as the target character name.
17. The apparatus of claim 16, wherein the avatar acquisition request information includes the at least one target keyword.
18. The apparatus of claim 16, wherein the avatar acquisition request information includes target voice information for describing the avatar, the voice information including the at least one target keyword.
19. An electronic device, comprising:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein the content of the first and second substances,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method of any one of claims 1 to 9.
20. A non-transitory computer readable storage medium storing computer instructions for causing a computer to perform the method of any one of claims 1 to 9.
CN202110801937.9A 2021-07-15 2021-07-15 Method and device for determining and acquiring virtual image and electronic equipment Active CN113407850B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110801937.9A CN113407850B (en) 2021-07-15 2021-07-15 Method and device for determining and acquiring virtual image and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110801937.9A CN113407850B (en) 2021-07-15 2021-07-15 Method and device for determining and acquiring virtual image and electronic equipment

Publications (2)

Publication Number Publication Date
CN113407850A CN113407850A (en) 2021-09-17
CN113407850B true CN113407850B (en) 2022-08-26

Family

ID=77686561

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110801937.9A Active CN113407850B (en) 2021-07-15 2021-07-15 Method and device for determining and acquiring virtual image and electronic equipment

Country Status (1)

Country Link
CN (1) CN113407850B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114187429B (en) * 2021-11-09 2023-03-24 北京百度网讯科技有限公司 Virtual image switching method and device, electronic equipment and storage medium
CN114239241B (en) * 2021-11-30 2023-02-28 北京百度网讯科技有限公司 Card generation method and device and electronic equipment
CN114187394B (en) * 2021-12-13 2023-05-05 北京百度网讯科技有限公司 Avatar generation method, apparatus, electronic device, and storage medium
CN114693294A (en) * 2022-03-04 2022-07-01 支付宝(杭州)信息技术有限公司 Interaction method and device based on electronic certificate and electronic equipment

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006235671A (en) * 2005-02-22 2006-09-07 Norinaga Tsukiji Conversation device and computer readable record medium
CN107728895A (en) * 2017-10-25 2018-02-23 中国移动通信集团公司 A kind of processing method of virtual objects, device and storage medium
CN112866762A (en) * 2020-12-31 2021-05-28 北京达佳互联信息技术有限公司 Processing method and device for acquiring video associated information, electronic equipment and server

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060141426A1 (en) * 2004-12-23 2006-06-29 Inventec Corporation Virtual scene interactive language learning system and method
US9285951B2 (en) * 2013-02-14 2016-03-15 Disney Enterprises, Inc. Avatar personalization in a virtual environment
CN109448737B (en) * 2018-08-30 2020-09-01 百度在线网络技术(北京)有限公司 Method and device for creating virtual image, electronic equipment and storage medium
WO2020056691A1 (en) * 2018-09-20 2020-03-26 太平洋未来科技(深圳)有限公司 Method for generating interactive object, device, and electronic apparatus
CN112152901A (en) * 2019-06-26 2020-12-29 腾讯科技(深圳)有限公司 Virtual image control method and device and electronic equipment
CN110662083B (en) * 2019-09-30 2022-04-22 北京达佳互联信息技术有限公司 Data processing method and device, electronic equipment and storage medium
CN112598785B (en) * 2020-12-25 2022-03-25 游艺星际(北京)科技有限公司 Method, device and equipment for generating three-dimensional model of virtual image and storage medium
CN113050795A (en) * 2021-03-24 2021-06-29 北京百度网讯科技有限公司 Virtual image generation method and device
CN113099298B (en) * 2021-04-08 2022-07-12 广州华多网络科技有限公司 Method and device for changing virtual image and terminal equipment

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006235671A (en) * 2005-02-22 2006-09-07 Norinaga Tsukiji Conversation device and computer readable record medium
CN107728895A (en) * 2017-10-25 2018-02-23 中国移动通信集团公司 A kind of processing method of virtual objects, device and storage medium
CN112866762A (en) * 2020-12-31 2021-05-28 北京达佳互联信息技术有限公司 Processing method and device for acquiring video associated information, electronic equipment and server

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
B2C网站虚拟人物应用模型研究;王超;《中国优秀硕士学位论文全文数据库 信息科技辑》;20100815;I139-148 *
Text to visual synthesis with appearance models;J. Melenchon等;《Proceedings 2003 International Conference on Image Processing》;20031124;第237-240页 *
互联网环境下品牌虚拟形象设计与推广研究;李常隆;《中国优秀硕士学位论文全文数据库 哲学与人文科学辑》;20170215;F088-896 *
增强现实环境中的虚拟人物助手研究与实现;管煜祥;《中国优秀硕士学位论文全文数据库 信息科技辑》;20200115;I138-2301 *

Also Published As

Publication number Publication date
CN113407850A (en) 2021-09-17

Similar Documents

Publication Publication Date Title
CN113407850B (en) Method and device for determining and acquiring virtual image and electronic equipment
CN113836333A (en) Training method of image-text matching model, method and device for realizing image-text retrieval
CN112487173B (en) Man-machine conversation method, device and storage medium
CN113963110B (en) Texture map generation method and device, electronic equipment and storage medium
CN114787814A (en) Reference resolution
CN112084366A (en) Method, apparatus, device and storage medium for retrieving image
CN113378855A (en) Method for processing multitask, related device and computer program product
CN113836268A (en) Document understanding method and device, electronic equipment and medium
CN114861889A (en) Deep learning model training method, target object detection method and device
CN113627536A (en) Model training method, video classification method, device, equipment and storage medium
CN112580666A (en) Image feature extraction method, training method, device, electronic equipment and medium
CN112529180A (en) Method and apparatus for model distillation
JP2023002690A (en) Semantics recognition method, apparatus, electronic device, and storage medium
CN114861059A (en) Resource recommendation method and device, electronic equipment and storage medium
CN112182255A (en) Method and apparatus for storing media files and for retrieving media files
CN113468857B (en) Training method and device for style conversion model, electronic equipment and storage medium
CN114494747A (en) Model training method, image processing method, device, electronic device and medium
CN113792876A (en) Backbone network generation method, device, equipment and storage medium
CN112784102B (en) Video retrieval method and device and electronic equipment
CN114266937A (en) Model training method, image processing method, device, equipment and storage medium
CN114187405A (en) Method, apparatus, device, medium and product for determining an avatar
CN113360683A (en) Method for training cross-modal retrieval model and cross-modal retrieval method and device
CN112860995A (en) Interaction method, device, client, server and storage medium
CN114186039A (en) Visual question answering method and device and electronic equipment
CN114547252A (en) Text recognition method and device, electronic equipment and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant