WO2023139961A1 - Information processing device - Google Patents

Information processing device Download PDF

Info

Publication number
WO2023139961A1
WO2023139961A1 PCT/JP2022/045381 JP2022045381W WO2023139961A1 WO 2023139961 A1 WO2023139961 A1 WO 2023139961A1 JP 2022045381 W JP2022045381 W JP 2022045381W WO 2023139961 A1 WO2023139961 A1 WO 2023139961A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
avatar
head
user
face
Prior art date
Application number
PCT/JP2022/045381
Other languages
French (fr)
Japanese (ja)
Inventor
晃平 大山
Original Assignee
株式会社Nttドコモ
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社Nttドコモ filed Critical 株式会社Nttドコモ
Publication of WO2023139961A1 publication Critical patent/WO2023139961A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/20Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts

Definitions

  • the present invention relates to an information processing device.
  • avatars are sometimes used as users' alter ego.
  • 3D scanning it has become possible to use avatars, which are 3D images of users, in a 3D virtual space.
  • Patent Document 1 discloses a technique in which a server device selects one avatar from multiple avatars.
  • the user's communication device requests the server device to display a predetermined page including the avatar specified by the identification information.
  • the server device selects which of the first avatar and the second avatar having different scales to be displayed based on the presence or absence of the display area of the first avatar and the display area of the second avatar on the predetermined page.
  • the server device generates image data of the first avatar or the second avatar according to the selection result.
  • the server device transmits the generated image data of the first avatar or the second avatar to the communication device of the user.
  • the server device automatically selects one of two avatars with different scales, so even if either avatar was selected, the user himself could not select the avatar with the style of painting that he liked. As a result, the user sometimes had to use an avatar with a drawing style that is not to his liking.
  • an object of the present invention is to provide an information processing apparatus that enables a user to use an avatar with a style that suits the user's taste from among a plurality of avatars that have different styles of painting.
  • An information processing apparatus includes: a face image generation unit that generates a plurality of face images with different styles; a face image acquisition unit that acquires a first face image selected by a user from the plurality of face images; a head generation unit that generates a three-dimensional image showing the head of an avatar based on the first face image; and an avatar generation unit that generates a three-dimensional image showing the overall appearance of the avatar using the three-dimensional image showing the head of the avatar and the three-dimensional image showing the body of the avatar.
  • the present invention it is possible for a user to use an avatar with a painting style that suits the user's taste from among a plurality of avatars with different painting styles.
  • FIG. 2 is a block diagram showing a configuration example of a terminal device 20; FIG. The figure which shows the production
  • 2 is a block diagram showing a configuration example of the server 10;
  • FIG. 3 is a functional block diagram of an acquisition unit 111;
  • FIG. 10 is a diagram showing an example of a plurality of face images FP1 to FP4 with different styles;
  • 4 is a flowchart showing operations of the server 10 according to the first embodiment;
  • FIG. 4 is a diagram showing an operation example of a face image acquisition unit 111E and a head generation unit 113A;
  • FIG. 2 is a block diagram showing a configuration example of a server 10B;
  • FIG. 10 is a diagram showing an example of a head/body number table HT;
  • 11 is a flowchart showing operations of a server 10B according to the third embodiment; The figure which shows the whole structure of 1 C of information processing systems which concern on the modification 1.
  • FIG. 10 is a diagram showing an example of a head/body number table HT;
  • 11 is a flowchart showing operations of a server 10B according to the third embodiment; The figure which shows the whole structure of 1 C of information processing systems which concern on the modification 1.
  • FIG. 1 First Embodiment
  • a server 10 as an information processing apparatus according to a first embodiment of the present invention
  • FIG. 1 is a diagram showing the overall configuration of an information processing system 1 according to the first embodiment of the present invention.
  • the information processing system 1 displays an avatar A1 corresponding to the user U1 and an avatar A2 corresponding to the user U2 on the terminal devices 20 used by the users U1 and U2.
  • the information processing system 1 includes a server 10 and terminal devices 20 .
  • the server 10 is an example of an information processing device.
  • the server 10 and the terminal device 20 are communicably connected to each other via a communication network NET.
  • NET communication network
  • the suffix "-X" is used for the code.
  • X is an arbitrary integer of 1 or more.
  • FIG. 1 shows two terminal devices 20, ie, a terminal device 20-1 and a terminal device 20-2. However, this number is merely an example, and the information processing system 1 may include any number of terminal devices 20 .
  • FIG. 1 it is assumed that user U1 uses terminal device 20-1 and user U2 uses terminal device 20-2.
  • the server 10 provides various data and cloud services to the terminal device 20 via the communication network NET.
  • the server 10 provides the terminal device 20 with various data for displaying the avatar A1 corresponding to the user U1 and the avatar A2 corresponding to the user U2 on the terminal device 20 .
  • the server 10 provides the terminal device 20-1 with various data for displaying the avatar A2 on the display 24-1 of the terminal device 20-1 used by the user U1.
  • the server 10 also provides the terminal device 20-2 with various data for displaying the avatar A1 on the display 24-2 provided in the terminal device 20-2 used by the user U2.
  • the terminal device 20-1 and the terminal device 20-2 are preferably portable terminal devices such as smartphones and tablets, for example.
  • FIG. 2 is a block diagram showing a configuration example of the terminal device 20.
  • the terminal device 20 includes a processing device 21 , a storage device 22 , a communication device 23 , a display 24 , an input device 25 and an imaging device 26 .
  • Each element of the terminal device 20 is interconnected by one or more buses for communicating information.
  • the processing device 21 is a processor that controls the terminal device 20 as a whole. Also, the processing device 21 is configured using, for example, a single chip or a plurality of chips. The processing unit 21 is configured using, for example, a central processing unit (CPU) including interfaces with peripheral devices, arithmetic units, registers, and the like. A part or all of the functions of the processing device 21 may be realized by hardware such as DSP, ASIC, PLD, and FPGA. The processing device 21 executes various processes in parallel or sequentially.
  • CPU central processing unit
  • the storage device 22 is a recording medium that can be read and written by the processing device 21 .
  • the storage device 22 also stores a plurality of programs including the control program PR2 executed by the processing device 21 .
  • the communication device 23 is hardware as a transmission/reception device for communicating with other devices.
  • the communication device 23 is also called a network device, a network controller, a network card, a communication module, or the like, for example.
  • the communication device 23 may include a connector for wired connection and an interface circuit corresponding to the connector. Further, the communication device 23 may have a wireless communication interface. Products conforming to wired LAN, IEEE1394, and USB are examples of connectors and interface circuits for wired connection. Also, as a wireless communication interface, there are products conforming to wireless LAN, Bluetooth (registered trademark), and the like.
  • the display 24 is a device that displays images and character information.
  • the display 24 displays various images under the control of the processing device 21 .
  • various display panels such as a liquid crystal display panel and an organic EL (Electro Luminescence) display panel are preferably used as the display 24 .
  • the display 24 displays an image showing Avatar A in this embodiment. More specifically, when terminal device 20 is terminal device 20-1 used by user U1, display 24-1 mainly displays an image showing avatar A2 corresponding to user U2. On the other hand, if terminal device 20 is terminal device 20-2 used by user U2, display 24-2 mainly displays an image showing avatar A1 corresponding to user U1.
  • the input device 25 accepts operations from the user U1.
  • the input device 25 includes a pointing device such as a keyboard, touch pad, touch panel, or mouse.
  • the input device 25 may also serve as the display 24 .
  • the user U1 uploads an input image IP showing the front part of the user U1's face from the terminal device 20 to the server 10 for the purpose of generating a three-dimensional avatar.
  • the input image IP is typically a two-dimensional image generated based on the photograph of the face of the user U1.
  • the input image IP is not limited to a two-dimensional image generated based on the face photograph of the user U1.
  • the input device 25 is used by the user U1 to input the input image IP to the terminal device 20 .
  • the input image IP may be obtained by capturing an image of the user U1 with the imaging device 26 described later, or may be obtained from an external device using the communication device 23 described above.
  • the imaging device 26 outputs imaging information obtained by imaging the outside world.
  • the imaging device 26 includes, for example, a lens, an imaging element, an amplifier, and an AD converter.
  • the light condensed through the lens is converted into an image pickup signal, which is an analog signal, by the image pickup device.
  • the amplifier amplifies the imaging signal and outputs it to the AD converter.
  • the AD converter converts the amplified imaging signal, which is an analog signal, into imaging information, which is a digital signal.
  • the converted imaging information is output to the processing device 21 .
  • the imaging information output to the processing device 21 is output to the server 10 via the communication device 23 .
  • the processing device 21 functions as an acquisition unit 211, an output unit 212, and a display control unit 213 by reading the control program PR2 from the storage device 22 and executing it.
  • the acquisition unit 211 acquires image information representing the input image IP representing the front portion of the face of the user U1.
  • the server 10 generates a plurality of face images FP1 to FPn having different styles based on the input image IP.
  • n is an integer of 2 or more.
  • the server 10 outputs the plurality of generated face images FP1 to FPn to the terminal device 20.
  • FIG. The user U1 uses the input device 25 to select one face image FPk from a plurality of face images FP1 to FPn displayed on the display 24 of the terminal device 20.
  • the obtaining unit 211 obtains the selection result k of the one face image FPk.
  • "k” is an integer of 1 or more and n or less.
  • the user U1 also uses the input device 25 to input a number HB representing the head size of the avatar A1 to be generated, specifically, a numerical value indicating how many heads the avatar A1 is to be generated as.
  • the acquisition unit 211 acquires the head/body count HB.
  • the acquisition unit 211 acquires image information indicating an image displayed on the display 24 from the server 10 by using the communication device 23 .
  • the output unit 212 outputs to the server 10 the image information indicating the input image IP indicating the front part of the face of the user U1, the selection result k of one face image FPk, and the head-to-body number HB obtained by the obtaining unit 211.
  • FIG. 3 shows the flow of generating a three-dimensional image WP showing the overall appearance of avatar A1, which is the avatar corresponding to user U1.
  • a face image generation unit 112 provided in the server 10, which will be described later, generates a plurality of face images FP1 to FPn based on the input image IP.
  • Image information representing the plurality of face images FP1 to FPn is output from the server 10 to the terminal device 20.
  • FIG. As described above, the plurality of face images FP1 to FPn are displayed on the display 24 of the terminal device 20.
  • FIG. The user U1 uses the input device 25 to select one face image FPk from the plurality of face images FP1 to FPn.
  • a selection result k by the user U1 is output to the server 10 from the output unit 212 provided in the terminal device 20 .
  • the face image FPk selected by the user U1 is used by the server 10 to generate a three-dimensional image HP representing the head.
  • the head/body count HB is output to the server 10 from the output unit 212 provided in the terminal device 20 .
  • a three-dimensional image WP showing the overall appearance of the avatar A1 is generated based on the three-dimensional image HP showing the head, the number HB of the head and body, and the three-dimensional image BP showing the body parts other than the head of the avatar A1.
  • the display control unit 213 causes the acquisition unit 211 to display the image indicated by the image information acquired from the server 10 on the display 24 .
  • the image information is image information indicating an image of avatar A.
  • FIG. That is, when the terminal device 20 is the terminal device 20-1, the display control unit 213 mainly displays the avatar A2 corresponding to the user U2 on the display 24-1. On the other hand, when the terminal device 20 is the terminal device 20-2, the display control unit 213 mainly displays the avatar A1 corresponding to the user U1 on the display 24-2.
  • FIG. 4 is a block diagram showing a configuration example of the server 10. As shown in FIG.
  • the server 10 comprises a processing device 11 , a storage device 12 , a communication device 13 , a display 14 and an input device 15 .
  • Each element of server 10 is interconnected by one or more buses for communicating information.
  • the processing device 11 is a processor that controls the server 10 as a whole. Also, the processing device 11 is configured using, for example, a single chip or a plurality of chips. The processing unit 11 is configured using, for example, a central processing unit (CPU) including interfaces with peripheral devices, arithmetic units, registers, and the like. A part or all of the functions of the processing device 11 may be implemented by hardware such as DSP, ASIC, PLD, and FPGA. The processing device 11 executes various processes in parallel or sequentially.
  • CPU central processing unit
  • the storage device 12 is a recording medium that can be read and written by the processing device 11.
  • the storage device 12 also stores a plurality of programs including the control program PR1 executed by the processing device 11 .
  • the storage device 12 also stores avatar information AI.
  • Avatar information AI includes image information indicating face images FP1 to FPn generated by face image generating section 112, which will be described later.
  • the avatar information AI also includes information used by the later-described body part generation unit 114 when generating image information representing a three-dimensional image BP representing the body part of the avatar A1.
  • the communication device 13 is hardware as a transmission/reception device for communicating with other devices.
  • the communication device 13 is also called a network device, a network controller, a network card, a communication module, or the like, for example.
  • the communication device 13 may include a connector for wired connection and an interface circuit corresponding to the connector. Further, the communication device 13 may have a wireless communication interface. Products conforming to wired LAN, IEEE1394, and USB are examples of connectors and interface circuits for wired connection. Also, as a wireless communication interface, there are products conforming to wireless LAN, Bluetooth (registered trademark), and the like.
  • the display 14 is a device that displays images and character information.
  • the display 14 displays various images under the control of the processing device 11 .
  • various display panels such as a liquid crystal display panel and an organic EL display panel are preferably used as the display 14 .
  • the input device 15 is a device that receives operations from the administrator of the information processing system 1 .
  • the input device 15 includes a pointing device such as a keyboard, touch pad, touch panel, or mouse.
  • the input device 15 may also serve as the display 14 .
  • the processing device 11 functions as an acquisition unit 111, a face image generation unit 112, a head generation unit 113, a body generation unit 114, an avatar generation unit 115, and an output unit 116 by reading and executing the control program PR1 from the storage device 12.
  • the acquisition unit 111 acquires various types of information from the terminal device 20, including image information indicating the input image IP indicating the front portion of the face of the user U1, the selection result k of one face image FPk, the number of heads and bodies HB, line-of-sight information, position information, movement information, and imaging information.
  • FIG. 5 is a functional block diagram of the acquisition unit 111.
  • the acquisition unit 111 includes an input image acquisition unit 111A, a face image acquisition unit 111B, and a head/body acquisition unit 111C.
  • the input image acquisition unit 111A acquires, from the terminal device 20-1, image information representing the input image IP representing the front portion of the face of the user U1.
  • the facial image acquisition unit 111B acquires the selection result k of the first facial image FPk from the terminal device 20-1. Further, the facial image acquisition unit 111B acquires the facial image FPk from the storage device 12 based on the selection result k.
  • the head/body acquisition unit 111C acquires the head/body number HB from the terminal device 20-1.
  • the face image generation unit 112 generates a plurality of face images FP1 to FPn with different styles based on the input image IP representing the front part of the face of the user U1 acquired by the input image acquisition unit 111A.
  • the input image IP is typically a two-dimensional image generated based on the facial photograph of the user U1. That is, the facial image generation unit 112 typically generates a plurality of facial images FP1 to FPn based on an image representing a facial photograph of the user U1.
  • FIG. 6 is an example of a plurality of face images FP1 to FP4 with different styles.
  • the face images FP1 to FP4 shown in FIG. 6 are all generated from the same input image IP, but have different styles.
  • the face image generation unit 112 includes a plurality of generation engines from a first generation engine that generates the face image FP1 to an n-th generation engine that generates the face image FPn. After that, the face image generation unit 112 inputs the input image IP to each engine from the first generation engine to the n-th generation engine.
  • the face image generating unit 112 generates the face images FP1 to FPn by outputting the face images FP1 to FPn from the first generation engine to the nth generation engine.
  • the face image generation unit 112 stores the generated face images FP1 to FPn in the storage device 12.
  • the head generation unit 113 generates a three-dimensional image HP representing the head of the avatar A1 based on the first face image FPk acquired by the face image acquisition unit 111B. More specifically, the head generation unit 113 generates the image information representing the three-dimensional image HP, so that the overall style of the three-dimensional image HP representing the head of the avatar A1 becomes the same style as that represented by one face image FPk.
  • the body generation unit 114 uses the head/body number HB acquired by the head/body acquisition unit 111C and the avatar information AI stored in the storage device 12 to generate image information representing a three-dimensional image BP of the body of the avatar A1 other than the head.
  • the body generator 114 uses information included in the avatar information AI to once generate a temporary three-dimensional image SP as a three-dimensional image of the body of the avatar A1.
  • the body part generating unit 114 adjusts the size of the provisional three-dimensional image SP to set the size ratio between the three-dimensional image HP showing the head of the avatar A1 and the provisional three-dimensional image SP as the ratio indicated by the number of heads and bodies HB.
  • the body generation unit 114 sets the temporary 3D image SP after adjusting the size as the 3D image BP of the body of the avatar A1, and then generates image information indicating the 3D image BP.
  • the body generation unit 114 may set the overall style of the three-dimensional image BP to be the same style as that represented by the face image FPk.
  • the body generation unit 114 may set the overall style of the 3D image BP to be different from the style represented by the face image FPk.
  • the avatar generation unit 115 uses image information representing the three-dimensional image HP representing the head of the avatar A1 and image information representing the three-dimensional image BP representing the body of the avatar A1 to generate image information representing the three-dimensional image WP representing the overall appearance of the avatar A1.
  • the output unit 116 uses the communication device 13 to transmit image information indicating an image displayed on the display 24 of the terminal device 20 and image information indicating the plurality of face images FP1 to FPn to the terminal device 20.
  • the output unit 116 also uses the communication device 13 to transmit image information representing a three-dimensional image WP representing the overall appearance of the avatar A1 generated by the avatar generation unit 115 to the terminal device 20 .
  • the server 10 can create avatars A1 in various styles, and create avatars A1 in styles that suit the tastes of user U1.
  • FIG. 7 is a flow chart showing the operation of the server 10 according to the first embodiment. The operation of the server 10 will be described below with reference to FIG.
  • step S1 the processing device 11 functions as an input image acquisition section 111A.
  • the processing device 11 acquires, from the terminal device 20-1, image information representing the input image IP representing the front portion of the face of the user U1.
  • step S2 the processing device 11 functions as the face image generator 112.
  • the processing device 11 generates a plurality of face images FP1 to FPn with different styles based on the input image IP representing the front portion of the face of the user U1 acquired in step S1. Further, the processing device 11 stores the generated face images FP1 to FPn in the storage device 12.
  • step S3 the processing device 11 functions as the face image acquisition section 111B.
  • the processing device 11 acquires the selection result k of the first face image FPk from the terminal device 20-1, and acquires the face image FPk from the storage device 12 based on the selection result k.
  • step S4 the processing device 11 functions as the head generation unit 113.
  • the processing device 11 generates a three-dimensional image HP representing the head of the avatar A1 based on one face image FPk acquired in step S3.
  • step S5 the processing device 11 functions as the head/body acquisition unit 111C.
  • the processing device 11 acquires the head and body number HB from the terminal device 20-1.
  • step S6 the processing device 11 functions as the body generation unit 114.
  • the processing device 11 uses the head/body count HB acquired in step S5 and the avatar information AI stored in the storage device 12 to generate image information representing the three-dimensional image BP of the body of the avatar A1.
  • step S7 the processing device 11 functions as the avatar generation unit 115.
  • the processing device 11 uses image information representing a three-dimensional image HP representing the head of the avatar A1 and image information representing a three-dimensional image BP representing the body of the avatar A1 to generate image information representing a three-dimensional image WP representing the overall appearance of the avatar A1.
  • step S8 the processing device 11 functions as the output unit 116.
  • the processing device 11 uses the communication device 13 to transmit the image information representing the three-dimensional image WP representing the overall appearance of the avatar A1 generated in step S7 to the terminal device 20 . After that, the processing device 11 ends all the processes shown in FIG.
  • the server 10 as an information processing device includes a face image generation unit 112, a face image acquisition unit 111B, a head generation unit 113, a head/body acquisition unit 111C, a body generation unit 114, and an avatar generation unit 115.
  • the face image generator 112 generates a plurality of face images FP1 to FPn with different styles.
  • the face image obtaining unit 111B obtains the first face image FPk selected by the user U1 from the plurality of face images FP1 to FPn.
  • the head generation unit 113 generates a three-dimensional image HP representing the head of the avatar A1 based on one face image FPk.
  • the head and body acquisition unit 111C acquires a head and body number HB representing how many heads and bodies the avatar A1 has.
  • the body part generation unit 114 generates a three-dimensional image BP showing the body parts other than the head of the avatar A1 based on the head-to-body number HB.
  • the avatar generation unit 115 generates a 3D image WP representing the overall appearance of the avatar A1 using the 3D image HP representing the head of the avatar A1 and the 3D image BP representing the body of the avatar A1.
  • the user U1 can use the avatar A1 with a style that suits the user U1's taste from among a plurality of avatars A with different styles. Further, in the present embodiment, the server 10 separates and generates a three-dimensional image HP representing the head of the avatar A1 and a three-dimensional image BP representing the body of the avatar A1. More specifically, the user U1 selects a face image FPk with a style that matches his/her taste from among a plurality of face images FP1 to FPn with different styles. The head generation unit 113 generates a three-dimensional image HP representing the head of the avatar A1 based on the selected face image FPk.
  • the characteristics of the avatar A1 are likely to appear in the head of the avatar A1. That is, the user U1 can use the overall appearance of the avatar A1 having features that match his/her preferences by a simple means of selecting one face image FPk from the plurality of face images FP1 to FPn.
  • the body part generation unit 114 does not necessarily generate the three-dimensional image BP showing the body parts of the plurality of avatars A1 in different styles. That is, the processing load on the server device 10 can be kept lower when generating a plurality of face images FP1 to FPn in different styles than when generating the overall appearance of a plurality of avatars A1 in different styles. Therefore, the server 10 according to the present embodiment can generate the avatar A1 having characteristics that match the preferences of the user U1 while reducing the processing load on the server 10 .
  • the face image generation unit 112 generates a plurality of face images FP1 to FPn with different styles based on the input image IP as the image representing the face photograph of the user U1.
  • the server 10 has the above-described configuration, when the user U1 uses the avatar A1 with a drawing style that suits the user U1's preference from among a plurality of avatars A having different drawing styles, the face of the user U1 can be reflected more realistically as a three-dimensional image HP showing the head of the avatar A1. Further, since the user U1 does not need to draw the input image IP from the beginning, the user U1 can more easily use the three-dimensional image HP representing the head of the avatar A1. Furthermore, since the image representing the photograph of the face of user U1 is a two-dimensional image, the server 10 can reduce the processing load on the server 10 compared to the case of using a three-dimensional image. On the other hand, since the photograph of the face accurately represents the features of the user U1, it is possible to create an avatar A1 that resembles the user U1.
  • An information processing system 1A according to the second embodiment of the present invention differs from the information processing system 1 according to the first embodiment in that a server 10A is provided instead of the server 10.
  • FIG. Otherwise, the overall configuration of the information processing system 1A is the same as the overall configuration of the information processing system 1 according to the first embodiment shown in FIG. 1, so illustration and description thereof will be omitted.
  • the server 10A is provided with a processing device 11A instead of the processing device 11 and a storage device 12A instead of the storage device 12.
  • the processing device 11A includes an acquisition unit 111D instead of the acquisition unit 111 and a head generation unit 113A instead of the head generation unit 113 .
  • the acquisition unit 111D includes a face image acquisition unit 111E instead of the face image acquisition unit 111B.
  • the configuration of the server 10A is the same as the configuration of the server 10 according to the first embodiment shown in FIGS. 4 and 5, so illustration and description thereof will be omitted.
  • the facial image acquisition unit 111E extracts an elemental image EP representing facial elements of the user U1 from the first facial image FPk selected by the user U1.
  • the head generation unit 113A also generates a three-dimensional image HP representing the head of the avatar A1 by superimposing the element image EP extracted by the face image acquisition unit 111E on the outline image OP prepared in advance.
  • the head generation unit 113A generates a three-dimensional image HP representing the head of the avatar A1 by superimposing one element image EP included in one face image FPk selected from a plurality of face images FP1 to FPn having different styles on the outline image OP.
  • FIG. 8 is a diagram showing an operation example of the face image acquisition unit 111E and the head generation unit 113A.
  • the facial image acquisition unit 111E extracts an element image EP representing facial elements of the user U1 from the facial image FPk of the user U1.
  • the facial image acquisition unit 111E extracts elemental images EP representing eyebrows, eyes, nose, and mouth as facial elements from the facial image FPk.
  • the face image acquisition unit 111E may extract only some of these eyebrows, eyes, nose, and mouth as the element image EP.
  • the face image acquisition unit 111E may extract additional elements such as eyelashes and moles in addition to these eyebrows, eyes, nose, and mouth as elemental images EP.
  • the head generation unit 113A generates a three-dimensional image HP representing the head of the avatar A1 by superimposing the extracted element image EP on the outline image OP prepared in advance.
  • the outline image OP may be a two-dimensional image or a three-dimensional image.
  • the head generating unit 113A may generate a three-dimensional image HP representing the head of the avatar A1 by superimposing the element image EP on the external image OP as a two-dimensional image, and then three-dimensionalizing the external image OP superimposed with the element image EP.
  • the head generation unit 113A may generate a three-dimensional image HP representing the head of the avatar A1 by superimposing the element image EP on the outline image OP as a three-dimensional image.
  • the server 10A can create avatars A1 in various styles while using outline images OP prepared in advance as templates.
  • the server 10A can create an avatar A1 with a style that matches the taste of the user U1, using an outline image OP prepared in advance as a template.
  • step S3 the processing device 11A functions as the facial image acquiring section 111E. After obtaining the first facial image FPk, the processing device 11A extracts an elemental image EP representing facial elements of the user U1 from the facial image FPk. Further, in step S4, the processing device 11A functions as the head generating section 113A. The processing device 11A superimposes the extracted elemental image EP on the outline image OP prepared in advance to generate a three-dimensional image HP representing the head of the avatar A1.
  • the facial image acquisition unit 111E extracts the element image EP representing the facial elements of the user U1 from the selected first facial image FPk.
  • the head generation unit 113A generates a three-dimensional image HP representing the head of the avatar A1 by superimposing the extracted element image EP on the outline image OP prepared in advance.
  • the user U1 can use an avatar A1 with a style that suits his/her taste from among a plurality of avatars A with different styles while using the outline image OP prepared in advance as a template.
  • the server 10A commonly uses outline images OP prepared in advance as templates when generating three-dimensional images HP showing the heads of a plurality of avatars A that are different from each other.
  • the server 10A allows the plurality of avatars A to have a sense of unity.
  • the server 10A can make each of the avatars A have a style that matches the preferences of each user U while giving the plurality of avatars A a sense of unity.
  • FIG. 3 Third Embodiment
  • the configuration of an information processing system 1B including a server 10B as an information processing apparatus according to a third embodiment of the present invention will be described with reference to FIGS. 9 to 11.
  • FIG. 1B for the purpose of simplifying the description, among the components provided in the information processing system 1B according to the third embodiment, the same components as those of the information processing system 1 according to the first embodiment are denoted by the same reference numerals, and the description thereof may be omitted.
  • 3-1 Configuration of the third embodiment 3-1-1: Overall Configuration
  • An information processing system 1B according to the third embodiment of the present invention differs from the information processing system 1 according to the first embodiment in that a server 10B is provided instead of the server 10.
  • FIG. Otherwise, the overall configuration of the information processing system 1B is the same as the overall configuration of the information processing system 1 according to the first embodiment shown in FIG. 1, so illustration and description thereof will be omitted.
  • the storage device 12B stores the control program PR1B instead of the control program PR1.
  • the storage device 12B also stores a learning model LM and a head/height number table HT.
  • the learning model LM is a learning model for the age estimating unit 117 to estimate the age of the user U1 based on the input image IP representing the facial photograph of the user U1 acquired by the input image acquiring unit 111A.
  • the learning model LM is generated by learning teacher data in the learning phase.
  • the teacher data used to generate the learning model LM has a plurality of one-to-one pairs of the feature information extracted from the input image IP showing the facial photograph of one person, which is acquired by the input image acquisition unit 111A, and the age of the person.
  • the learning model LM is generated outside the server 10B.
  • the learning model LM is preferably generated in a second server (not shown).
  • the server 10B acquires the learning model LM from a second server (not shown) via the communication network NET.
  • the head and body count table HT is a table for defining the correspondence relationship between the age estimated by the age estimation unit 117 described later and the head and body count HB.
  • FIG. 10 shows an example of the head/body number table HT.
  • the avatar A1 has 2 heads and bodies.
  • the processing device 11B includes an acquisition unit 111F instead of the acquisition unit 111.
  • the acquisition unit 111F includes a head/body acquisition unit 111G instead of the head/body acquisition unit 111C.
  • the configuration of the acquisition unit 111F is the same as the configuration of the acquisition unit 111 according to the first embodiment shown in FIG. 5, so illustration and description thereof will be omitted.
  • the processing device 11B includes an age estimation unit 117 and a head/body number generation unit 118. FIG.
  • the age estimation unit 117 estimates the age of the user U1 based on the input image IP showing the photograph of the face of the user U1. More specifically, the age estimation unit 117 inputs the input image IP acquired by the input image acquisition unit 111A to the learning model LM. After that, the age estimation unit 117 generates the estimated age of the user U1 by outputting the estimated age from the learning model LM. Age estimation section 117 also outputs the generated estimated age to head/body number generation section 118 .
  • the head/body number generation unit 118 Based on the age estimated by the age estimation unit 117, the head/body number generation unit 118 generates a head/body number HB that indicates the size of the avatar A1. More specifically, the head-to-body number generation unit 118 generates the head-to-body number HB of the avatar A1 by checking the age acquired from the age estimation unit 117 against the head-to-body number table HT stored in the storage device 12B.
  • the head and body acquisition unit 111G acquires the head and body number HB of the avatar A1 generated by the head and body generation unit 118.
  • the server 10B can generate the avatar A1 with the head size HB that matches the impression given by the estimated age of the user U1.
  • FIG. 11 is a flow chart showing the operation of the server 10B according to the third embodiment. The operation of the server 10B will be described below with reference to FIG.
  • step S11 the processing device 11B functions as an input image acquisition section 111A.
  • the processing device 11B acquires, from the terminal device 20-1, image information representing the input image IP representing the front portion of the face of the user U1.
  • the processing device 11B functions as the face image generator 112.
  • the processing device 11B generates a plurality of face images FP1 to FPn having different styles based on the input image IP representing the front portion of the face of the user U1 acquired in step S11. Further, the processing device 11B stores the generated face images FP1 to FPn in the storage device 12B.
  • the processing device 11B functions as the face image acquiring section 111B.
  • the processing device 11B acquires the selection result k of one face image FPk from the terminal device 20 .
  • the processing device 11B functions as a facial image acquisition section 111B.
  • the processing device 11B acquires the face image FPk from the storage device 12B based on the selection result k.
  • step S14 the processing device 11B functions as the head generation unit 113.
  • the processing device 11B generates a three-dimensional image HP representing the head of the avatar A1 based on one face image FPk acquired in step S13.
  • step S15 the processing device 11B functions as the age estimation unit 117.
  • the processing device 11B estimates the age of the user U1 based on the input image IP acquired in step S11.
  • step S16 the processing device 11B functions as the head/body generation unit 118. Accordingly, the processing device 11B generates the head-to-body count HB of the avatar A1 based on the age of the user U1 estimated in step S15.
  • step S17 the processing device 11B functions as the head/body acquisition unit 111G.
  • the processing device 11B acquires the head/body count HB generated in step S16.
  • step S18 the processing device 11B functions as the body generation unit 114.
  • the processing device 11B generates image information representing a three-dimensional image BP of the body of the avatar A1 other than the head, using the head/body count HB obtained in step S17 and the avatar information AI stored in the storage device 12B.
  • step S19 the processing device 11B functions as the avatar generation unit 115.
  • the processing device 11B uses the image information representing the three-dimensional image HP representing the head of the avatar A1 and the image information representing the three-dimensional image BP representing the body of the avatar A1 to generate the image information representing the three-dimensional image WP representing the overall appearance of the avatar A1.
  • step S20 the processing device 11B functions as the output unit 116.
  • the processing device 11B transmits image information representing a three-dimensional image WP representing the overall appearance of the avatar A1 generated in step S19 to the terminal device 20 via the communication device 13 . After that, the processing device 11B ends all the processes shown in FIG.
  • the server 10B as an information processing device includes the age estimating section 117 and the head/body number generating section 118 .
  • the age estimation unit 117 estimates the age of the user U1 based on the image showing the photograph of the face of the user U1.
  • the head-to-body generation unit 118 generates the head-to-body number HB of the avatar A1 based on the estimated age.
  • the server 10B has the above configuration, when the user U1 uses the avatar A1 with the style that suits the taste of the user U1 from among the plurality of avatars A having different styles, the avatar A1 with the head size HB that matches the impression given by the estimated age of the user U1 can be used. Moreover, since the head/body generation unit 118 generates the head/body number HB based on the age estimated by the age estimation unit 117, the user U1 does not need to input the head/body number HB. That is, the user U1 can use the avatar A1 that matches the impression given by the estimated age of the user U1 by a simple method.
  • the servers 10 to 10B display the avatar A on the display 24 of the terminal device 20.
  • servers 10 to 10B may display avatar A on XR glasses instead of display 24 .
  • FIG. 12 is a diagram showing the overall configuration of an information processing system 1C according to this modified example.
  • the information processing system 1C uses XR technology to provide a virtual space to users U1 and U2 wearing the XR glasses 30 .
  • the information processing system 1C causes the XR glasses 30 to display an avatar A1 corresponding to the user U1 and an avatar A2 corresponding to the user U2.
  • XR technology is a general term for VR (Virtual Reality) technology, AR (Augmented Reality) technology, and MR (Mixed Reality) technology.
  • VR technology is technology for displaying a digital virtual space on a device such as VR glasses or an HMD (Head Mounted Display) employing VR technology.
  • AR technology is technology that adds information indicated by digital content to the real world in an augmented reality space displayed on a device such as AR glasses or an HMD that employs AR technology.
  • MR technology is a technology that precisely superimposes a digital virtual space on a real space using MR glasses or a device such as an HMD (Head Mounted Display) employing MR technology.
  • the information processing system 1C includes a server 10, a terminal device 20, and XR glasses 30.
  • the server 10 and the terminal device 20 are communicably connected to each other via a communication network NET.
  • the terminal device 20 and the XR glasses 30 are connected so as to be able to communicate with each other.
  • the suffix "-X" is used for the reference numerals.
  • the same is true for each component of the XR glass 30 .
  • two pairs are shown as pairs of the terminal device 20 and the XR glasses 30: the pair of the terminal device 20-1 and the XR glasses 30-1 and the pair of the terminal device 20-2 and the XR glasses 30-2.
  • the number of sets is merely an example, and the information processing system 1C can include any number of sets of the terminal device 20 and the XR glasses 30 .
  • FIG. 12 it is assumed that user U1 uses a set of terminal device 20-1 and XR glasses 30-1, and user U2 uses a set of terminal device 20-2 and XR glasses 30-2.
  • the server 10 provides various data and cloud services to the terminal device 20 via the communication network NET.
  • the server 10 provides the terminal device 20 with various data for displaying the avatar A1 corresponding to the user U1 and the avatar A2 corresponding to the user U2 on the XR glasses 30 connected to the terminal device 20 .
  • the server 10 provides the terminal device 20-1 with various data for displaying the avatar A2 on the display 38-1 of the XR glasses 30-1 used by the user U1.
  • the server 10 also provides the terminal device 20-2 with various data for displaying the avatar A1 on the display 38-2 of the XR glasses 30-2 used by the user U2.
  • the terminal device 20-1 causes the XR glasses 30-1 worn on the head by the user U1 to display virtual objects arranged in the virtual space. Further, the terminal device 20-2 causes the XR glasses 30-2 worn on the head of the user U2 to display a virtual object arranged in the virtual space.
  • the virtual space is, for example, a celestial space.
  • the virtual objects are, for example, virtual objects representing data such as still images, moving images, 3DCG models, HTML files, and text files, and virtual objects representing applications. Examples of text files include memos, source codes, diaries, and recipes. Examples of applications include browsers, applications for using SNS, and applications for generating document files.
  • the terminal device 20-1 and the terminal device 20-2 are preferably portable terminal devices such as smartphones and tablets, for example.
  • the terminal device 20-1 causes the XR glasses 30-1 to display a virtual object mainly corresponding to the avatar A2. Also, the terminal device 20-2 displays a virtual object mainly corresponding to the avatar A1 on the XR glasses 30-2.
  • the XR glasses 30 are display devices worn on the heads of users U1 and U2. More specifically, the XR glasses 30-1 are display devices worn on the head of the user U1. Also, the XR glasses 30-2 are a display device worn on the head of the user U2. The XR glasses 30 are, for example, a see-through wearable display. The XR glasses 30 are controlled by the terminal device 20 to display a virtual object on the display panel provided corresponding to each of the binocular lenses.
  • the user U1 and the user U2 can observe the avatars A1 and A2 displayed on the display 38. More specifically, the user U1 wearing the XR glasses 30-1 can observe the avatar A2 displayed on the display 38-1. On the other hand, the user U2 wearing the XR glasses 30-2 can observe the avatar A1 displayed on the display 38-2.
  • the terminal device 20 and the XR glasses 30 are implemented separately.
  • the method of realizing the terminal device 20 and the XR glasses 30 in this modified example is not limited to this.
  • the terminal device 20 and the XR glasses 30 may be implemented in a single housing by providing the XR glasses 30 with the same functions as the terminal device 20 .
  • the information processing system 1C may include a device such as an HMD that employs any one of VR technology, AR technology, and MR technology instead of the XR glasses 30 .
  • the terminal device 20-1 outputs a selection result k in which one face image FPk is selected from the plurality of face images FP1 to FPn to the servers 10 to 10B.
  • the face image FPk itself may be output from the terminal device 20-1 to the servers 10 to 10B.
  • the servers 10 to 10B acquire from the terminal device 20 the input image IP representing the front portion of the face of the user U1.
  • the servers 10 to 10B may acquire the input image IP from a device other than the terminal device 20.
  • the storage devices 12 to 12B and the storage device 22 were ROM and RAM, etc., but flexible disks, magneto-optical disks (e.g., compact discs, digital versatile discs, Blu-ray (registered trademark) discs), smart cards, flash memory devices (e.g., cards, sticks, key drives), CD-ROMs (Compact Disc-ROM), registers, removable disks, hard disks, floppy (registered trademark) disks, magnetic strips, databases, servers, and other suitable A storage medium.
  • the program may be transmitted from a network via an electric communication line.
  • the program may be transmitted from the communication network NET via an electric communication line.
  • the information, signals, etc. described may be represented using any of a variety of different technologies.
  • the data, instructions, commands, information, signals, bits, symbols, chips, etc. that may be referred to throughout the above description may be represented by voltages, currents, electromagnetic waves, magnetic fields or magnetic particles, optical fields or photons, or any combination thereof.
  • input/output information and the like may be stored in a specific location (for example, memory), or may be managed using a management table. Input/output information and the like can be overwritten, updated, or appended. The output information and the like may be deleted. The entered information and the like may be transmitted to another device.
  • the determination may be made by a value (0 or 1) represented using 1 bit, by a boolean value (Boolean: true or false), or by numerical comparison (for example, comparison with a predetermined value).
  • each function illustrated in FIGS. 1 to 12 is realized by any combination of at least one of hardware and software.
  • the method of realizing each functional block is not particularly limited. That is, each functional block may be implemented using one device that is physically or logically coupled, or two or more devices that are physically or logically separated may be directly or indirectly (e.g., wired, wireless, etc.) connected and implemented using these multiple devices.
  • a functional block may be implemented by combining software in the one device or the plurality of devices.
  • the programs illustrated in the above embodiments should be construed broadly to mean instructions, instruction sets, code, code segments, program code, programs, subprograms, software modules, applications, software applications, software packages, routines, subroutines, objects, executables, threads of execution, procedures, functions, etc., whether referred to as software, firmware, middleware, microcode, hardware description language, or by any other name.
  • software, instructions, information, etc. may be transmitted and received via a transmission medium.
  • a transmission medium For example, if the software is transmitted from a website, server, or other remote source using wired technologies (coaxial cable, fiber optic cable, twisted pair, Digital Subscriber Line (DSL), etc.) and/or wireless technologies (infrared, microwave, etc.), then these wired and/or wireless technologies are included within the definition of transmission medium.
  • wired technologies coaxial cable, fiber optic cable, twisted pair, Digital Subscriber Line (DSL), etc.
  • wireless technologies infrared, microwave, etc.
  • system and “network” are used interchangeably.
  • Information, parameters, etc. described in the present disclosure may be expressed using absolute values, may be expressed using relative values from a predetermined value, or may be expressed using corresponding separate information.
  • the servers 10 to 10B and the terminal device 20 may be mobile stations (MS).
  • a mobile station may also be referred to by those skilled in the art as a subscriber station, mobile unit, subscriber unit, wireless unit, remote unit, mobile device, wireless device, wireless communication device, remote device, mobile subscriber station, access terminal, mobile terminal, wireless terminal, remote terminal, handset, user agent, mobile client, client, or some other suitable term.
  • terms such as “mobile station”, “user terminal”, “user equipment (UE)", “terminal”, etc. may be used interchangeably.
  • connection means any direct or indirect connection or coupling between two or more elements, including the presence of one or more intermediate elements between two elements that are “connected” or “coupled” to each other.
  • Couplings or connections between elements may be physical couplings or connections, logical couplings or connections, or a combination thereof.
  • connection may be replaced with "access.”
  • two elements are considered to be “connected” or “coupled” to each other using at least one of one or more wires, cables, and printed electrical connections, and using electromagnetic energy having wavelengths in the radio frequency, microwave, and light (both visible and invisible) regions, as some non-limiting and non-exhaustive examples.
  • the phrase “based on” does not mean “based only on,” unless expressly specified otherwise. In other words, the phrase “based on” means both “based only on” and “based at least on.”
  • determining and “determining” as used in this disclosure may encompass a wide variety of actions.
  • “Judgement”, “determining” can include, for example, judging, calculating, computing, processing, deriving, investigating, looking up, searching, inquiring (e.g., searching in a table, database, or other data structure), ascertaining as “judging", “determining", etc.
  • “determining” and “determining” include receiving (e.g., receiving information), transmitting (e.g., transmitting information), input, output, and accessing (e.g., accessing data in memory).
  • determining or “determining” may include resolving, selecting, choosing, establishing, comparing, etc., to be regarded as “determining” or “determining.”
  • judgment and “decision” may include considering that some action is “judgment” and “decision”.
  • judgment (decision) may be replaced by “assuming”, “expecting”, “considering”, and the like.
  • the term "A and B are different” may mean “A and B are different from each other.” The term may also mean that "A and B are different from C”. Terms such as “separate,” “coupled,” etc. may also be interpreted in the same manner as “different.”
  • notification of predetermined information is not limited to explicit notification, and may be performed implicitly (e.g., not notifying the predetermined information).

Landscapes

  • Engineering & Computer Science (AREA)
  • Architecture (AREA)
  • Computer Graphics (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Processing Or Creating Images (AREA)

Abstract

An information processing device comprises a face image generation unit that generates a plurality of face images having different styles, a face image acquisition unit that acquires a first face image selected by a user from the plurality of face images, a head generation unit that generates a three-dimensional image of the head of an avatar on the basis of the first face image, and an avatar generation unit that generates a three-dimensional image of the outer appearance of the entire avatar on the basis of the three-dimensional image of the head of the avatar and a three-dimensional image of a body which is the rest of the avatar other than the head.

Description

情報処理装置Information processing equipment
 本発明は、情報処理装置に関する。 The present invention relates to an information processing device.
 インターネットにおいて、アバター(avatar)と呼称される、ユーザの分身となるキャラクターが用いられることがある。近年、3Dスキャン等の技術を用いることにより、三次元の仮想空間において、ユーザの姿を三次元画像としたアバターを利用することが可能となった。 On the Internet, characters called avatars are sometimes used as users' alter ego. In recent years, by using techniques such as 3D scanning, it has become possible to use avatars, which are 3D images of users, in a 3D virtual space.
 例えば、特許文献1は、サーバ装置が複数のアバターから一つのアバターを選択する技術を開示している。具体的には、特許文献1に係る技術において、ユーザの通信装置は、サーバ装置に対し、識別情報で特定されるアバターを含む所定ページの表示を要求する。当該サーバ装置は、表示を要求されると、互いに縮尺の異なる第1のアバターと第2のアバターのうちいずれのアバターを表示させるかを、上記の所定ページにおける、第1のアバターの表示領域、及び第2のアバターの表示領域の有無に基づいて選択する。また、サーバ装置は、上記の選択結果に応じて、第1のアバター又は第2のアバターの画像データを生成する。更に、サーバ装置は、生成された第1のアバター又は第2のアバターの画像データを、ユーザの通信装置に送信する。 For example, Patent Document 1 discloses a technique in which a server device selects one avatar from multiple avatars. Specifically, in the technique disclosed in Patent Literature 1, the user's communication device requests the server device to display a predetermined page including the avatar specified by the identification information. When the display is requested, the server device selects which of the first avatar and the second avatar having different scales to be displayed based on the presence or absence of the display area of the first avatar and the display area of the second avatar on the predetermined page. Also, the server device generates image data of the first avatar or the second avatar according to the selection result. Furthermore, the server device transmits the generated image data of the first avatar or the second avatar to the communication device of the user.
特開2013-029951号公報JP 2013-029951 A
 しかし、従来の技術では、サーバ装置が、縮尺が互いに異なる2つのアバターのうち一方のアバターを自動的に選択するので、いずれのアバターが選択されたとしても、ユーザ自らが好みの画風のアバターを選択することはできなかった。その結果、ユーザは自身の好みでは無い画風のアバターを利用しなければならないことがあった。 However, in the conventional technology, the server device automatically selects one of two avatars with different scales, so even if either avatar was selected, the user himself could not select the avatar with the style of painting that he liked. As a result, the user sometimes had to use an avatar with a drawing style that is not to his liking.
 そこで、本発明は、互いに異なる画風を有する複数のアバターの中から、ユーザの好みに合った画風のアバターをユーザが利用可能にする情報処理装置を提供することを目的とする。 Therefore, an object of the present invention is to provide an information processing apparatus that enables a user to use an avatar with a style that suits the user's taste from among a plurality of avatars that have different styles of painting.
 本発明の好適な態様に係る情報処理装置は、画風が互いに異なる複数の顔画像を生成する顔画像生成部と、前記複数の顔画像から、ユーザによって選択された第1の顔画像を取得する顔画像取得部と、前記第1の顔画像に基づいて、アバターの頭部を示す三次元画像を生成する頭部生成部と、前記アバターが何頭身であるかを表す頭身数を取得する頭身取得部と、前記頭身数に基づいて、前記アバターの前記頭部以外の部分である体部を示す三次元画像を生成する体部生成部と、前記アバターの前記頭部を示す前記三次元画像、及び前記アバターの前記体部を示す前記三次元画像を用いて、前記アバターの全体の外観を示す三次元画像を生成するアバター生成部と、を備える情報処理装置である。 An information processing apparatus according to a preferred aspect of the present invention includes: a face image generation unit that generates a plurality of face images with different styles; a face image acquisition unit that acquires a first face image selected by a user from the plurality of face images; a head generation unit that generates a three-dimensional image showing the head of an avatar based on the first face image; and an avatar generation unit that generates a three-dimensional image showing the overall appearance of the avatar using the three-dimensional image showing the head of the avatar and the three-dimensional image showing the body of the avatar.
 本発明によれば、互いに異なる画風を有する複数のアバターの中から、ユーザの好みに合った画風のアバターをユーザが利用することが可能となる。 According to the present invention, it is possible for a user to use an avatar with a painting style that suits the user's taste from among a plurality of avatars with different painting styles.
第1実施形態に係る情報処理システム1の全体構成を示す図。The figure which shows the whole structure of the information processing system 1 which concerns on 1st Embodiment. 端末装置20の構成例を示すブロック図。2 is a block diagram showing a configuration example of a terminal device 20; FIG. アバターA1の全体の外観を示す三次元画像WPの生成フローを示す図。The figure which shows the production|generation flow of the three-dimensional image WP which shows the appearance of the whole avatar A1. サーバ10の構成例を示すブロック図。2 is a block diagram showing a configuration example of the server 10; FIG. 取得部111の機能ブロック図。3 is a functional block diagram of an acquisition unit 111; FIG. 画風が互いに異なる複数の顔画像FP1~FP4の例を示す図。FIG. 10 is a diagram showing an example of a plurality of face images FP1 to FP4 with different styles; 第1実施形態に係るサーバ10の動作を示すフローチャート。4 is a flowchart showing operations of the server 10 according to the first embodiment; 顔画像取得部111E及び頭部生成部113Aの動作例を示す図。FIG. 4 is a diagram showing an operation example of a face image acquisition unit 111E and a head generation unit 113A; サーバ10Bの構成例を示すブロック図。FIG. 2 is a block diagram showing a configuration example of a server 10B; 頭身数テーブルHTの例を示す図。FIG. 10 is a diagram showing an example of a head/body number table HT; 第3実施形態に係るサーバ10Bの動作を示すフローチャート。11 is a flowchart showing operations of a server 10B according to the third embodiment; 変形例1に係る情報処理システム1Cの全体構成を示す図。The figure which shows the whole structure of 1 C of information processing systems which concern on the modification 1. FIG.
1:第1実施形態
 以下、図1~図7を参照しつつ、本発明の第1実施形態に係る情報処理装置としてのサーバ10を含む情報処理システム1の構成について説明する。
1: First Embodiment Hereinafter, the configuration of an information processing system 1 including a server 10 as an information processing apparatus according to a first embodiment of the present invention will be described with reference to FIGS. 1 to 7. FIG.
1-1:第1実施形態の構成
1-1-1:全体構成
 図1は、本発明の第1実施形態に係る情報処理システム1の全体構成を示す図である。情報処理システム1は、ユーザU1に対応するアバターA1と、ユーザU2に対応するアバターA2とを、ユーザU1及びユーザU2が使用する端末装置20に表示させる。
1-1: Configuration of the first embodiment
1-1-1: Overall Configuration FIG. 1 is a diagram showing the overall configuration of an information processing system 1 according to the first embodiment of the present invention. The information processing system 1 displays an avatar A1 corresponding to the user U1 and an avatar A2 corresponding to the user U2 on the terminal devices 20 used by the users U1 and U2.
 情報処理システム1は、サーバ10、及び端末装置20を備える。サーバ10は、情報処理装置の一例である。情報処理システム1において、サーバ10と端末装置20とは、通信網NETを介して互いに通信可能に接続される。以下の説明において、ユーザごとに使用される端末装置20を区別する場合は、符号に添え字「-X」を用いる。Xは1以上の任意の整数である。また、端末装置20各々の構成要素についても同様である。なお、図1において、端末装置20-1と端末装置20-2との2つの端末装置20が記載される。しかし、当該個数はあくまで一例であって、情報処理システム1は、任意の数の端末装置20を備え得る。図1においては、ユーザU1が端末装置20-1を使用し、ユーザU2が端末装置20-2を使用することを前提とする。 The information processing system 1 includes a server 10 and terminal devices 20 . The server 10 is an example of an information processing device. In the information processing system 1, the server 10 and the terminal device 20 are communicably connected to each other via a communication network NET. In the following description, when distinguishing the terminal device 20 used for each user, the suffix "-X" is used for the code. X is an arbitrary integer of 1 or more. In addition, the same applies to the constituent elements of each terminal device 20 . Note that FIG. 1 shows two terminal devices 20, ie, a terminal device 20-1 and a terminal device 20-2. However, this number is merely an example, and the information processing system 1 may include any number of terminal devices 20 . In FIG. 1, it is assumed that user U1 uses terminal device 20-1 and user U2 uses terminal device 20-2.
 サーバ10は、通信網NETを介して、端末装置20に対して各種データ及びクラウドサービスを提供する。とりわけ、サーバ10は、端末装置20に、ユーザU1に対応するアバターA1、及びユーザU2に対応するアバターA2を表示させるための各種データを、端末装置20に対して提供する。より詳細には、サーバ10は、ユーザU1が使用する端末装置20-1に備わるディスプレイ24-1に、アバターA2を表示させるための各種データを、端末装置20-1に対して提供する。また、サーバ10は、ユーザU2が使用する端末装置20-2に備わるディスプレイ24-2に、アバターA1を表示させるための各種データを、端末装置20-2に対して提供する。なお、端末装置20-1及び端末装置20-2は、例として、スマートフォン、及びタブレット等の携帯端末装置であることが好適である。 The server 10 provides various data and cloud services to the terminal device 20 via the communication network NET. In particular, the server 10 provides the terminal device 20 with various data for displaying the avatar A1 corresponding to the user U1 and the avatar A2 corresponding to the user U2 on the terminal device 20 . More specifically, the server 10 provides the terminal device 20-1 with various data for displaying the avatar A2 on the display 24-1 of the terminal device 20-1 used by the user U1. The server 10 also provides the terminal device 20-2 with various data for displaying the avatar A1 on the display 24-2 provided in the terminal device 20-2 used by the user U2. The terminal device 20-1 and the terminal device 20-2 are preferably portable terminal devices such as smartphones and tablets, for example.
1-1-2:端末装置の構成
 図2は、端末装置20の構成例を示すブロック図である。端末装置20は、処理装置21、記憶装置22、通信装置23、ディスプレイ24、入力装置25、及び撮像装置26を備える。端末装置20が有する各要素は、情報を通信するための単体又は複数のバスによって相互に接続される。
1-1-2: Configuration of Terminal Device FIG. 2 is a block diagram showing a configuration example of the terminal device 20. As shown in FIG. The terminal device 20 includes a processing device 21 , a storage device 22 , a communication device 23 , a display 24 , an input device 25 and an imaging device 26 . Each element of the terminal device 20 is interconnected by one or more buses for communicating information.
 処理装置21は、端末装置20の全体を制御するプロセッサである。また、処理装置21は、例えば、単数又は複数のチップを用いて構成される。処理装置21は、例えば、周辺装置とのインタフェース、演算装置及びレジスタ等を含む中央処理装置(CPU)を用いて構成される。なお、処理装置21が有する機能の一部又は全部を、DSP、ASIC、PLD、及びFPGA等のハードウェアによって実現してもよい。処理装置21は、各種の処理を並列的又は逐次的に実行する。 The processing device 21 is a processor that controls the terminal device 20 as a whole. Also, the processing device 21 is configured using, for example, a single chip or a plurality of chips. The processing unit 21 is configured using, for example, a central processing unit (CPU) including interfaces with peripheral devices, arithmetic units, registers, and the like. A part or all of the functions of the processing device 21 may be realized by hardware such as DSP, ASIC, PLD, and FPGA. The processing device 21 executes various processes in parallel or sequentially.
 記憶装置22は、処理装置21による読取及び書込が可能な記録媒体である。また、記憶装置22は、処理装置21が実行する制御プログラムPR2を含む複数のプログラムを記憶する。 The storage device 22 is a recording medium that can be read and written by the processing device 21 . The storage device 22 also stores a plurality of programs including the control program PR2 executed by the processing device 21 .
 通信装置23は、他の装置と通信を行うための、送受信デバイスとしてのハードウェアである。通信装置23は、例えば、ネットワークデバイス、ネットワークコントローラ、ネットワークカード、及び通信モジュール等とも呼ばれる。通信装置23は、有線接続用のコネクターを備え、上記コネクターに対応するインタフェース回路を備えていてもよい。また、通信装置23は、無線通信インタフェースを備えていてもよい。有線接続用のコネクター及びインタフェース回路としては有線LAN、IEEE1394、及びUSBに準拠した製品が挙げられる。また、無線通信インタフェースとしては無線LAN及びBluetooth(登録商標)等に準拠した製品が挙げられる。 The communication device 23 is hardware as a transmission/reception device for communicating with other devices. The communication device 23 is also called a network device, a network controller, a network card, a communication module, or the like, for example. The communication device 23 may include a connector for wired connection and an interface circuit corresponding to the connector. Further, the communication device 23 may have a wireless communication interface. Products conforming to wired LAN, IEEE1394, and USB are examples of connectors and interface circuits for wired connection. Also, as a wireless communication interface, there are products conforming to wireless LAN, Bluetooth (registered trademark), and the like.
 ディスプレイ24は、画像及び文字情報を表示するデバイスである。ディスプレイ24は、処理装置21の制御のもとで各種の画像を表示する。例えば、液晶表示パネル及び有機EL(Electro Luminescence)表示パネル等の各種の表示パネルがディスプレイ24として好適に利用される。とりわけ、本実施形態において、ディスプレイ24は、アバターAを示す画像を表示する。より詳細には、端末装置20がユーザU1によって使用される端末装置20-1である場合、ディスプレイ24-1は、主として、ユーザU2に対応するアバターA2を示す画像を表示する。一方で、端末装置20がユーザU2によって使用される端末装置20-2である場合、ディスプレイ24-2は、主として、ユーザU1に対応するアバターA1を示す画像を表示する。 The display 24 is a device that displays images and character information. The display 24 displays various images under the control of the processing device 21 . For example, various display panels such as a liquid crystal display panel and an organic EL (Electro Luminescence) display panel are preferably used as the display 24 . Among other things, the display 24 displays an image showing Avatar A in this embodiment. More specifically, when terminal device 20 is terminal device 20-1 used by user U1, display 24-1 mainly displays an image showing avatar A2 corresponding to user U2. On the other hand, if terminal device 20 is terminal device 20-2 used by user U2, display 24-2 mainly displays an image showing avatar A1 corresponding to user U1.
 入力装置25は、ユーザU1からの操作を受け付ける。例えば、入力装置25は、キーボード、タッチパッド、タッチパネル又はマウス等のポインティングデバイスを含んで構成される。ここで、入力装置25は、タッチパネルを含んで構成される場合、ディスプレイ24を兼ねてもよい。 The input device 25 accepts operations from the user U1. For example, the input device 25 includes a pointing device such as a keyboard, touch pad, touch panel, or mouse. Here, when the input device 25 includes a touch panel, the input device 25 may also serve as the display 24 .
 本実施形態において、ユーザU1は、三次元のアバターを生成することを目的に、ユーザU1の顔の正面部を示す入力画像IPを、端末装置20からサーバ10にアップロードする。入力画像IPは、典型的には、ユーザU1の顔写真に基づいて生成される二次元画像である。しかし、入力画像IPは、ユーザU1の顔写真に基づいて生成される二次元画像に限定されない。アップロード時において、入力装置25は、ユーザU1が上記の入力画像IPを端末装置20に入力するために用いられる。入力画像IPは、後述の撮像装置26によって、ユーザU1を撮像することによって得られてもよく、上記の通信装置23を用いることにより、外部機器から取得されてもよい。 In this embodiment, the user U1 uploads an input image IP showing the front part of the user U1's face from the terminal device 20 to the server 10 for the purpose of generating a three-dimensional avatar. The input image IP is typically a two-dimensional image generated based on the photograph of the face of the user U1. However, the input image IP is not limited to a two-dimensional image generated based on the face photograph of the user U1. At the time of uploading, the input device 25 is used by the user U1 to input the input image IP to the terminal device 20 . The input image IP may be obtained by capturing an image of the user U1 with the imaging device 26 described later, or may be obtained from an external device using the communication device 23 described above.
 撮像装置26は、外界を撮像して得られた撮像情報を出力する。また、撮像装置26は、例えば、レンズ、撮像素子、増幅器、及びAD変換器を備える。レンズを介して集光された光は、撮像素子によってアナログ信号である撮像信号に変換される。増幅器は撮像信号を増幅した上でAD変換器に出力する。AD変換器はアナログ信号である増幅された撮像信号をデジタル信号である撮像情報に変換する。変換された撮像情報は、処理装置21に出力される。処理装置21に出力された撮像情報は、通信装置23を介して、サーバ10に出力される。 The imaging device 26 outputs imaging information obtained by imaging the outside world. Also, the imaging device 26 includes, for example, a lens, an imaging element, an amplifier, and an AD converter. The light condensed through the lens is converted into an image pickup signal, which is an analog signal, by the image pickup device. The amplifier amplifies the imaging signal and outputs it to the AD converter. The AD converter converts the amplified imaging signal, which is an analog signal, into imaging information, which is a digital signal. The converted imaging information is output to the processing device 21 . The imaging information output to the processing device 21 is output to the server 10 via the communication device 23 .
 処理装置21は、記憶装置22から制御プログラムPR2を読み出して実行することによって、取得部211、出力部212、及び表示制御部213として機能する。 The processing device 21 functions as an acquisition unit 211, an output unit 212, and a display control unit 213 by reading the control program PR2 from the storage device 22 and executing it.
 取得部211は、ユーザU1の顔の正面部を示す入力画像IPを示す画像情報を取得する。 The acquisition unit 211 acquires image information representing the input image IP representing the front portion of the face of the user U1.
 後述する図3に示されるように、サーバ10は、入力画像IPに基づいて画風が互いに異なる複数の顔画像FP1~FPnを生成する。なお、「n」は2以上の整数である。
また、サーバ10は、生成した複数の顔画像FP1~FPnを端末装置20に出力する。ユーザU1は、端末装置20に備わるディスプレイ24に表示された複数の顔画像FP1~FPnから、入力装置25を用いて一つの顔画像FPkを選択する。取得部211は、当該一つの顔画像FPkの選択結果kを取得する。なお、「k」は、1以上n以下の整数である。
As shown in FIG. 3, which will be described later, the server 10 generates a plurality of face images FP1 to FPn having different styles based on the input image IP. Note that "n" is an integer of 2 or more.
Further, the server 10 outputs the plurality of generated face images FP1 to FPn to the terminal device 20. FIG. The user U1 uses the input device 25 to select one face image FPk from a plurality of face images FP1 to FPn displayed on the display 24 of the terminal device 20. FIG. The obtaining unit 211 obtains the selection result k of the one face image FPk. In addition, "k" is an integer of 1 or more and n or less.
 また、ユーザU1は、生成するアバターA1が何頭身であるかを表す頭身数HB、具体的にはアバターA1を何頭身のアバターとして生成するかを示す数値を、入力装置25を用いて入力する。取得部211は、当該頭身数HBを取得する。 The user U1 also uses the input device 25 to input a number HB representing the head size of the avatar A1 to be generated, specifically, a numerical value indicating how many heads the avatar A1 is to be generated as. The acquisition unit 211 acquires the head/body count HB.
 また、取得部211は、通信装置23を用いることにより、サーバ10から、ディスプレイ24に表示される画像を示す画像情報を取得する。 Also, the acquisition unit 211 acquires image information indicating an image displayed on the display 24 from the server 10 by using the communication device 23 .
 出力部212は、取得部211によって取得された、ユーザU1の顔の正面部を示す入力画像IPを示す画像情報、一つの顔画像FPkの選択結果k、及び頭身数HBを、サーバ10に出力する。 The output unit 212 outputs to the server 10 the image information indicating the input image IP indicating the front part of the face of the user U1, the selection result k of one face image FPk, and the head-to-body number HB obtained by the obtaining unit 211.
 図3は、ユーザU1に対応するアバターである、アバターA1の全体の外観を示す三次元画像WPの生成フローを示す。図3に示されるように、サーバ10に備わる後述の顔画像生成部112は、入力画像IPに基づいて、複数の顔画像FP1~FPnを生成する。複数の顔画像FP1~FPnを示す画像情報は、サーバ10から端末装置20に出力される。上記のように、当該複数の顔画像FP1~FPnは、端末装置20のディスプレイ24に表示される。ユーザU1は、入力装置25を用いることによって、複数の顔画像FP1~FPnから一つの顔画像FPkを選択する。ユーザU1による選択結果kは、端末装置20に備わる出力部212からサーバ10に出力される。ユーザU1によって選択された顔画像FPkは、サーバ10において、頭部を示す三次元画像HPを生成するために用いられる。また、頭身数HBが端末装置20に備わる出力部212からサーバ10に出力される。サーバ10において、頭部を示す三次元画像HP、頭身数HB、及びアバターA1の頭部以外の部分である体部を示す三次元画像BPに基づいて、アバターA1の全体の外観を示す三次元画像WPが生成される。 FIG. 3 shows the flow of generating a three-dimensional image WP showing the overall appearance of avatar A1, which is the avatar corresponding to user U1. As shown in FIG. 3, a face image generation unit 112 provided in the server 10, which will be described later, generates a plurality of face images FP1 to FPn based on the input image IP. Image information representing the plurality of face images FP1 to FPn is output from the server 10 to the terminal device 20. FIG. As described above, the plurality of face images FP1 to FPn are displayed on the display 24 of the terminal device 20. FIG. The user U1 uses the input device 25 to select one face image FPk from the plurality of face images FP1 to FPn. A selection result k by the user U1 is output to the server 10 from the output unit 212 provided in the terminal device 20 . The face image FPk selected by the user U1 is used by the server 10 to generate a three-dimensional image HP representing the head. Also, the head/body count HB is output to the server 10 from the output unit 212 provided in the terminal device 20 . In the server 10, a three-dimensional image WP showing the overall appearance of the avatar A1 is generated based on the three-dimensional image HP showing the head, the number HB of the head and body, and the three-dimensional image BP showing the body parts other than the head of the avatar A1.
 説明を図2に戻すと、表示制御部213は、取得部211によって、サーバ10から取得された画像情報によって示される画像をディスプレイ24に表示させる。とりわけ本実施形態において、当該画像情報は、アバターAの画像を示す画像情報である。すなわち、端末装置20が端末装置20-1である場合、表示制御部213は、主としてユーザU2に対応するアバターA2をディスプレイ24-1に表示させる。一方で、端末装置20が端末装置20-2である場合、表示制御部213は、主としてユーザU1に対応するアバターA1をディスプレイ24-2に表示させる。 Returning the description to FIG. 2, the display control unit 213 causes the acquisition unit 211 to display the image indicated by the image information acquired from the server 10 on the display 24 . Particularly in this embodiment, the image information is image information indicating an image of avatar A. FIG. That is, when the terminal device 20 is the terminal device 20-1, the display control unit 213 mainly displays the avatar A2 corresponding to the user U2 on the display 24-1. On the other hand, when the terminal device 20 is the terminal device 20-2, the display control unit 213 mainly displays the avatar A1 corresponding to the user U1 on the display 24-2.
1-1-3:サーバの構成
 図4は、サーバ10の構成例を示すブロック図である。サーバ10は、処理装置11、記憶装置12、通信装置13、ディスプレイ14、及び入力装置15を備える。サーバ10が有する各要素は、情報を通信するための単体又は複数のバスによって相互に接続される。
1-1-3: Server Configuration FIG. 4 is a block diagram showing a configuration example of the server 10. As shown in FIG. The server 10 comprises a processing device 11 , a storage device 12 , a communication device 13 , a display 14 and an input device 15 . Each element of server 10 is interconnected by one or more buses for communicating information.
 処理装置11は、サーバ10の全体を制御するプロセッサである。また、処理装置11は、例えば、単数又は複数のチップを用いて構成される。処理装置11は、例えば、周辺装置とのインタフェース、演算装置及びレジスタ等を含む中央処理装置(CPU)を用いて構成される。なお、処理装置11の機能の一部又は全部を、DSP、ASIC、PLD、及びFPGA等のハードウェアによって実現してもよい。処理装置11は、各種の処理を並列的又は逐次的に実行する。 The processing device 11 is a processor that controls the server 10 as a whole. Also, the processing device 11 is configured using, for example, a single chip or a plurality of chips. The processing unit 11 is configured using, for example, a central processing unit (CPU) including interfaces with peripheral devices, arithmetic units, registers, and the like. A part or all of the functions of the processing device 11 may be implemented by hardware such as DSP, ASIC, PLD, and FPGA. The processing device 11 executes various processes in parallel or sequentially.
 記憶装置12は、処理装置11による読取及び書込が可能な記録媒体である。また、記憶装置12は、処理装置11が実行する制御プログラムPR1を含む複数のプログラムを記憶する。また、記憶装置12は、アバター情報AIを記憶する。アバター情報AIは、後述の顔画像生成部112によって生成される顔画像FP1~FPnを示す画像情報を含む。また、アバター情報AIは、後述の体部生成部114が、アバターA1の体部を示す三次元画像BPを示す画像情報の生成時に用いる情報を含む。 The storage device 12 is a recording medium that can be read and written by the processing device 11. The storage device 12 also stores a plurality of programs including the control program PR1 executed by the processing device 11 . The storage device 12 also stores avatar information AI. Avatar information AI includes image information indicating face images FP1 to FPn generated by face image generating section 112, which will be described later. The avatar information AI also includes information used by the later-described body part generation unit 114 when generating image information representing a three-dimensional image BP representing the body part of the avatar A1.
 通信装置13は、他の装置と通信を行うための、送受信デバイスとしてのハードウェアである。通信装置13は、例えば、ネットワークデバイス、ネットワークコントローラ、ネットワークカード、及び通信モジュール等とも呼ばれる。通信装置13は、有線接続用のコネクターを備え、上記コネクターに対応するインタフェース回路を備えていてもよい。また、通信装置13は、無線通信インタフェースを備えていてもよい。有線接続用のコネクター及びインタフェース回路としては有線LAN、IEEE1394、及びUSBに準拠した製品が挙げられる。また、無線通信インタフェースとしては無線LAN及びBluetooth(登録商標)等に準拠した製品が挙げられる。 The communication device 13 is hardware as a transmission/reception device for communicating with other devices. The communication device 13 is also called a network device, a network controller, a network card, a communication module, or the like, for example. The communication device 13 may include a connector for wired connection and an interface circuit corresponding to the connector. Further, the communication device 13 may have a wireless communication interface. Products conforming to wired LAN, IEEE1394, and USB are examples of connectors and interface circuits for wired connection. Also, as a wireless communication interface, there are products conforming to wireless LAN, Bluetooth (registered trademark), and the like.
 ディスプレイ14は、画像及び文字情報を表示するデバイスである。ディスプレイ14は、処理装置11による制御のもとで各種の画像を表示する。例えば、液晶表示パネル及び有機EL表示パネル等の各種の表示パネルがディスプレイ14として好適に利用される。 The display 14 is a device that displays images and character information. The display 14 displays various images under the control of the processing device 11 . For example, various display panels such as a liquid crystal display panel and an organic EL display panel are preferably used as the display 14 .
 入力装置15は、情報処理システム1の管理者からの操作を受け付ける機器である。例えば、入力装置15は、キーボード、タッチパッド、タッチパネル又はマウス等のポインティングデバイスを含んで構成される。ここで、入力装置15は、タッチパネルを含んで構成される場合、ディスプレイ14を兼ねてもよい。 The input device 15 is a device that receives operations from the administrator of the information processing system 1 . For example, the input device 15 includes a pointing device such as a keyboard, touch pad, touch panel, or mouse. Here, when the input device 15 includes a touch panel, the input device 15 may also serve as the display 14 .
 処理装置11は、例えば、記憶装置12から制御プログラムPR1を読み出して実行することによって、取得部111、顔画像生成部112、頭部生成部113、体部生成部114、アバター生成部115、及び出力部116として機能する。 For example, the processing device 11 functions as an acquisition unit 111, a face image generation unit 112, a head generation unit 113, a body generation unit 114, an avatar generation unit 115, and an output unit 116 by reading and executing the control program PR1 from the storage device 12.
 取得部111は、端末装置20から、ユーザU1の顔の正面部を示す入力画像IPを示す画像情報、一つの顔画像FPkの選択結果k、頭身数HB、視線情報、位置情報、動き情報、及び撮像情報を含む各種情報を取得する。 The acquisition unit 111 acquires various types of information from the terminal device 20, including image information indicating the input image IP indicating the front portion of the face of the user U1, the selection result k of one face image FPk, the number of heads and bodies HB, line-of-sight information, position information, movement information, and imaging information.
 図5は、取得部111の機能ブロック図である。図5に示されるように、取得部111は、入力画像取得部111A、顔画像取得部111B、及び頭身取得部111Cを備える。 FIG. 5 is a functional block diagram of the acquisition unit 111. FIG. As shown in FIG. 5, the acquisition unit 111 includes an input image acquisition unit 111A, a face image acquisition unit 111B, and a head/body acquisition unit 111C.
 入力画像取得部111Aは、端末装置20-1から、ユーザU1の顔の正面部を示す入力画像IPを示す画像情報を取得する。 The input image acquisition unit 111A acquires, from the terminal device 20-1, image information representing the input image IP representing the front portion of the face of the user U1.
 顔画像取得部111Bは、端末装置20-1から、第1の顔画像FPkの選択結果kを取得する。また、顔画像取得部111Bは、当該選択結果kに基づいて、記憶装置12から顔画像FPkを取得する。 The facial image acquisition unit 111B acquires the selection result k of the first facial image FPk from the terminal device 20-1. Further, the facial image acquisition unit 111B acquires the facial image FPk from the storage device 12 based on the selection result k.
 頭身取得部111Cは、端末装置20-1から、頭身数HBを取得する。 The head/body acquisition unit 111C acquires the head/body number HB from the terminal device 20-1.
 説明を図4に戻すと、顔画像生成部112は、入力画像取得部111Aによって取得されたユーザU1の顔の正面部を示す入力画像IPに基づいて、画風が互いに異なる複数の顔画像FP1~FPnを生成する。上記のように、入力画像IPは、典型的には、ユーザU1の顔写真に基づいて生成される二次元画像である。すなわち、顔画像生成部112は、典型的には、ユーザU1の顔写真を示す画像に基づいて、複数の顔画像FP1~FPnを生成する。 Returning to FIG. 4, the face image generation unit 112 generates a plurality of face images FP1 to FPn with different styles based on the input image IP representing the front part of the face of the user U1 acquired by the input image acquisition unit 111A. As described above, the input image IP is typically a two-dimensional image generated based on the facial photograph of the user U1. That is, the facial image generation unit 112 typically generates a plurality of facial images FP1 to FPn based on an image representing a facial photograph of the user U1.
 図6は、画風が互いに異なる複数の顔画像FP1~FP4の例である。図6に示される顔画像FP1~FP4は、すべて同一の入力画像IPから生成されながら、互いに画風が異なる。顔画像生成部112は、顔画像FP1を生成する第1生成エンジンから、顔画像FPnを生成する第n生成エンジンまでの、複数の生成エンジンを備える。その上で、顔画像生成部112は、入力画像IPを第1生成エンジン~第n生成エンジンまでの各エンジンに入力する。顔画像生成部112は、当該第1生成エンジン~第n生成エンジンから、顔画像FP1~FPnを出力させることにより、顔画像FP1~FPnを生成する。 FIG. 6 is an example of a plurality of face images FP1 to FP4 with different styles. The face images FP1 to FP4 shown in FIG. 6 are all generated from the same input image IP, but have different styles. The face image generation unit 112 includes a plurality of generation engines from a first generation engine that generates the face image FP1 to an n-th generation engine that generates the face image FPn. After that, the face image generation unit 112 inputs the input image IP to each engine from the first generation engine to the n-th generation engine. The face image generating unit 112 generates the face images FP1 to FPn by outputting the face images FP1 to FPn from the first generation engine to the nth generation engine.
 また、顔画像生成部112は、生成した顔画像FP1~FPnを、記憶装置12に格納する。 In addition, the face image generation unit 112 stores the generated face images FP1 to FPn in the storage device 12.
 頭部生成部113は、顔画像取得部111Bによって取得された第1の顔画像FPkに基づいて、アバターA1の頭部を示す三次元画像HPを生成する。より詳細には、頭部生成部113が当該三次元画像HPを示す画像情報を生成することにより、アバターA1の頭部を示す三次元画像HPの全体の画風が、一つの顔画像FPkによって示される画風と同一の画風となる。 The head generation unit 113 generates a three-dimensional image HP representing the head of the avatar A1 based on the first face image FPk acquired by the face image acquisition unit 111B. More specifically, the head generation unit 113 generates the image information representing the three-dimensional image HP, so that the overall style of the three-dimensional image HP representing the head of the avatar A1 becomes the same style as that represented by one face image FPk.
 体部生成部114は、頭身取得部111Cによって取得された頭身数HBと、記憶装置12に記憶されるアバター情報AIとを用いて、アバターA1の頭部以外の部分である体部の三次元画像BPを示す画像情報を生成する。例えば、体部生成部114は、アバター情報AIに含まれる情報を用いて、いったんアバターA1の体部の三次元画像として、仮の三次元画像SPを生成する。その後、体部生成部114は、当該仮の三次元画像SPの大きさを調整することにより、アバターA1の頭部を示す三次元画像HPと、当該仮の三次元画像SPとの大きさの比率を、頭身数HBによって示される比率とする。体部生成部114は、大きさを調整した後の、仮の三次元画像SPを、アバターA1の体部の三次元画像BPとした上で、当該三次元画像BPを示す画像情報を生成する。なお、体部生成部114は、当該三次元画像BPを示す画像情報の生成時において、当該三次元画像BPの全体の画風を、顔画像FPkによって示される画風と同一の画風としてもよい。あるいは、体部生成部114は、当該三次元画像BPを示す画像情報の生成時において、当該三次元画像BPの全体の画風を、顔画像FPkによって示される画風とは異なる画風としてもよい。 The body generation unit 114 uses the head/body number HB acquired by the head/body acquisition unit 111C and the avatar information AI stored in the storage device 12 to generate image information representing a three-dimensional image BP of the body of the avatar A1 other than the head. For example, the body generator 114 uses information included in the avatar information AI to once generate a temporary three-dimensional image SP as a three-dimensional image of the body of the avatar A1. After that, the body part generating unit 114 adjusts the size of the provisional three-dimensional image SP to set the size ratio between the three-dimensional image HP showing the head of the avatar A1 and the provisional three-dimensional image SP as the ratio indicated by the number of heads and bodies HB. The body generation unit 114 sets the temporary 3D image SP after adjusting the size as the 3D image BP of the body of the avatar A1, and then generates image information indicating the 3D image BP. When generating image information representing the three-dimensional image BP, the body generation unit 114 may set the overall style of the three-dimensional image BP to be the same style as that represented by the face image FPk. Alternatively, when generating the image information representing the 3D image BP, the body generation unit 114 may set the overall style of the 3D image BP to be different from the style represented by the face image FPk.
 アバター生成部115は、アバターA1の頭部を示す三次元画像HPを示す画像情報、及びアバターA1の体部を示す三次元画像BPを示す画像情報を用いて、アバターA1の全体の外観を示す三次元画像WPを示す画像情報を生成する。 The avatar generation unit 115 uses image information representing the three-dimensional image HP representing the head of the avatar A1 and image information representing the three-dimensional image BP representing the body of the avatar A1 to generate image information representing the three-dimensional image WP representing the overall appearance of the avatar A1.
 出力部116は、端末装置20に備わるディスプレイ24に表示される画像を示す画像情報、及び複数の顔画像FP1~FPnを示す画像情報を、通信装置13を用いて端末装置20に送信する。 The output unit 116 uses the communication device 13 to transmit image information indicating an image displayed on the display 24 of the terminal device 20 and image information indicating the plurality of face images FP1 to FPn to the terminal device 20.
 また出力部116は、アバター生成部115によって生成されたアバターA1の全体の外観を示す三次元画像WPを示す画像情報を、通信装置13を用いて端末装置20に送信する。 The output unit 116 also uses the communication device 13 to transmit image information representing a three-dimensional image WP representing the overall appearance of the avatar A1 generated by the avatar generation unit 115 to the terminal device 20 .
 この結果、サーバ10は、様々な画風のアバターA1を作成したり、ユーザU1の好みに合った画風のアバターA1を作成したりできる。 As a result, the server 10 can create avatars A1 in various styles, and create avatars A1 in styles that suit the tastes of user U1.
1-2:第1実施形態の動作
 図7は、第1実施形態に係るサーバ10の動作を示すフローチャートである。以下、図7を参照することにより、サーバ10の動作について説明する。
1-2: Operation of the First Embodiment FIG. 7 is a flow chart showing the operation of the server 10 according to the first embodiment. The operation of the server 10 will be described below with reference to FIG.
 ステップS1において、処理装置11は、入力画像取得部111Aとして機能する。処理装置11は、端末装置20-1から、ユーザU1の顔の正面部を示す入力画像IPを示す画像情報を取得する。 In step S1, the processing device 11 functions as an input image acquisition section 111A. The processing device 11 acquires, from the terminal device 20-1, image information representing the input image IP representing the front portion of the face of the user U1.
 ステップS2において、処理装置11は、顔画像生成部112として機能する。処理装置11は、ステップS1において取得されたユーザU1の顔の正面部を示す入力画像IPに基づいて、画風が互いに異なる複数の顔画像FP1~FPnを生成する。また、処理装置11は、生成した顔画像FP1~FPnを、記憶装置12に格納する。 In step S2, the processing device 11 functions as the face image generator 112. The processing device 11 generates a plurality of face images FP1 to FPn with different styles based on the input image IP representing the front portion of the face of the user U1 acquired in step S1. Further, the processing device 11 stores the generated face images FP1 to FPn in the storage device 12. FIG.
 ステップS3において、処理装置11は、顔画像取得部111Bとして機能する。処理装置11は、端末装置20-1から、第1の顔画像FPkの選択結果kを取得し、当該選択結果kに基づいて、記憶装置12から顔画像FPkを取得する。 In step S3, the processing device 11 functions as the face image acquisition section 111B. The processing device 11 acquires the selection result k of the first face image FPk from the terminal device 20-1, and acquires the face image FPk from the storage device 12 based on the selection result k.
 ステップS4において、処理装置11は、頭部生成部113として機能する。処理装置11は、ステップS3において取得された一つの顔画像FPkに基づいて、アバターA1の頭部を示す三次元画像HPを生成する。 In step S4, the processing device 11 functions as the head generation unit 113. The processing device 11 generates a three-dimensional image HP representing the head of the avatar A1 based on one face image FPk acquired in step S3.
 ステップS5において、処理装置11は、頭身取得部111Cとして機能する。処理装置11は、端末装置20-1から、頭身数HBを取得する。 In step S5, the processing device 11 functions as the head/body acquisition unit 111C. The processing device 11 acquires the head and body number HB from the terminal device 20-1.
 ステップS6において、処理装置11は、体部生成部114として機能する。処理装置11は、ステップS5において取得された頭身数HBと、記憶装置12に記憶されるアバター情報AIとを用いて、アバターA1の体部の三次元画像BPを示す画像情報を生成する。 In step S6, the processing device 11 functions as the body generation unit 114. The processing device 11 uses the head/body count HB acquired in step S5 and the avatar information AI stored in the storage device 12 to generate image information representing the three-dimensional image BP of the body of the avatar A1.
 ステップS7において、処理装置11は、アバター生成部115として機能する。処理装置11は、アバターA1の頭部を示す三次元画像HPを示す画像情報、及びアバターA1の体部を示す三次元画像BPを示す画像情報を用いて、アバターA1の全体の外観を示す三次元画像WPを示す画像情報を生成する。 In step S7, the processing device 11 functions as the avatar generation unit 115. The processing device 11 uses image information representing a three-dimensional image HP representing the head of the avatar A1 and image information representing a three-dimensional image BP representing the body of the avatar A1 to generate image information representing a three-dimensional image WP representing the overall appearance of the avatar A1.
 ステップS8において、処理装置11は、出力部116として機能する。処理装置11は、ステップS7において生成されたアバターA1の全体の外観を示す三次元画像WPを示す画像情報を、通信装置13を用いて端末装置20に送信する。その後、処理装置11は、図7に示される全ての処理を終了する。 In step S8, the processing device 11 functions as the output unit 116. The processing device 11 uses the communication device 13 to transmit the image information representing the three-dimensional image WP representing the overall appearance of the avatar A1 generated in step S7 to the terminal device 20 . After that, the processing device 11 ends all the processes shown in FIG.
1-3:第1実施形態が奏する効果
 以上の説明によれば、情報処理装置としてのサーバ10は、顔画像生成部112、顔画像取得部111B、頭部生成部113、頭身取得部111C、体部生成部114、及びアバター生成部115を備える。顔画像生成部112は、画風が互いに異なる複数の顔画像FP1~FPnを生成する。顔画像取得部111Bは、複数の顔画像FP1~FPnから、ユーザU1によって選択された第1の顔画像FPkを取得する。頭部生成部113は、一つの顔画像FPkに基づいて、アバターA1の頭部を示す三次元画像HPを生成する。頭身取得部111Cは、アバターA1が何頭身であるかを表す頭身数HBを取得する。体部生成部114は、頭身数HBに基づいて、アバターA1の頭部以外の部分である体部を示す三次元画像BPを生成する。アバター生成部115は、アバターA1の頭部を示す三次元画像HP、及びアバターA1の体部を示す三次元画像BPを用いて、アバターA1の全体の外観を示す三次元画像WPを生成する。
1-3: Effects of the First Embodiment According to the above description, the server 10 as an information processing device includes a face image generation unit 112, a face image acquisition unit 111B, a head generation unit 113, a head/body acquisition unit 111C, a body generation unit 114, and an avatar generation unit 115. The face image generator 112 generates a plurality of face images FP1 to FPn with different styles. The face image obtaining unit 111B obtains the first face image FPk selected by the user U1 from the plurality of face images FP1 to FPn. The head generation unit 113 generates a three-dimensional image HP representing the head of the avatar A1 based on one face image FPk. The head and body acquisition unit 111C acquires a head and body number HB representing how many heads and bodies the avatar A1 has. The body part generation unit 114 generates a three-dimensional image BP showing the body parts other than the head of the avatar A1 based on the head-to-body number HB. The avatar generation unit 115 generates a 3D image WP representing the overall appearance of the avatar A1 using the 3D image HP representing the head of the avatar A1 and the 3D image BP representing the body of the avatar A1.
 サーバ10は、上記の構成を備えるので、互いに異なる画風を有する複数のアバターAの中から、ユーザU1の好みに合った画風のアバターA1を、ユーザU1が利用できる。また、本実施形態において、サーバ10は、アバターA1の頭部を示す三次元画像HPと、アバターA1の体部を示す三次元画像BPとを分離して生成する。より詳細には、ユーザU1は、互いに画風の異なる複数の顔画像FP1~FPnの中から、自身の好みに合致した画風の顔画像FPkを選択する。頭部生成部113は、選択された顔画像FPkに基づいて、アバターA1の頭部を示す三次元画像HPを生成する。ここで、アバターA1の特徴は、アバターA1の頭部に表れやすい。すなわちユーザU1は、複数の顔画像FP1~FPnの中から、一つの顔画像FPkを選択するという簡便な手段によって、自身の好みに合致した特徴を有するアバターA1の全体の外観を利用できる。また、体部生成部114は、必ずしも、互いに異なる画風の、複数のアバターA1の体部を示す三次元画像BPを生成しない。すなわち、互いに異なる画風の複数のアバターA1の全体の外観を生成する場合に比較して、互いに異なる画風の複数の顔画像FP1~FPnを生成する場合の方が、サーバ装置10における処理負荷は低く抑えられる。従って、本実施形態に係るサーバ10は、当該サーバ10における処理負荷を軽減しながら、ユーザU1の好みに合致した特徴を有するアバターA1を生成できる。 Since the server 10 has the above configuration, the user U1 can use the avatar A1 with a style that suits the user U1's taste from among a plurality of avatars A with different styles. Further, in the present embodiment, the server 10 separates and generates a three-dimensional image HP representing the head of the avatar A1 and a three-dimensional image BP representing the body of the avatar A1. More specifically, the user U1 selects a face image FPk with a style that matches his/her taste from among a plurality of face images FP1 to FPn with different styles. The head generation unit 113 generates a three-dimensional image HP representing the head of the avatar A1 based on the selected face image FPk. Here, the characteristics of the avatar A1 are likely to appear in the head of the avatar A1. That is, the user U1 can use the overall appearance of the avatar A1 having features that match his/her preferences by a simple means of selecting one face image FPk from the plurality of face images FP1 to FPn. Also, the body part generation unit 114 does not necessarily generate the three-dimensional image BP showing the body parts of the plurality of avatars A1 in different styles. That is, the processing load on the server device 10 can be kept lower when generating a plurality of face images FP1 to FPn in different styles than when generating the overall appearance of a plurality of avatars A1 in different styles. Therefore, the server 10 according to the present embodiment can generate the avatar A1 having characteristics that match the preferences of the user U1 while reducing the processing load on the server 10 .
 また以上の説明によれば、顔画像生成部112は、ユーザU1の顔写真を示す画像としての入力画像IPに基づいて、画風が互いに異なる複数の顔画像FP1~FPnを生成する。 Also, according to the above description, the face image generation unit 112 generates a plurality of face images FP1 to FPn with different styles based on the input image IP as the image representing the face photograph of the user U1.
 サーバ10は、上記の構成を備えるので、互いに異なる画風を有する複数のアバターAの中から、ユーザU1の好みに合った画風のアバターA1をユーザU1が利用する上で、アバターA1の頭部を示す三次元画像HPとして、ユーザU1の顔をよりリアルに反映できる。また、ユーザU1は、入力画像IPを初めから作画する必要がないため、より簡便に、アバターA1の頭部を示す三次元画像HPを利用できる。更に、ユーザU1の顔写真を示す画像は二次元画像であるので、三次元画像を用いる場合に比較して、サーバ10は、当該サーバ10における処理負荷を低減できる。その一方で、顔写真はユーザU1の特徴を的確に表しているので、ユーザU1に似せたアバターA1を作成することができる。 Since the server 10 has the above-described configuration, when the user U1 uses the avatar A1 with a drawing style that suits the user U1's preference from among a plurality of avatars A having different drawing styles, the face of the user U1 can be reflected more realistically as a three-dimensional image HP showing the head of the avatar A1. Further, since the user U1 does not need to draw the input image IP from the beginning, the user U1 can more easily use the three-dimensional image HP representing the head of the avatar A1. Furthermore, since the image representing the photograph of the face of user U1 is a two-dimensional image, the server 10 can reduce the processing load on the server 10 compared to the case of using a three-dimensional image. On the other hand, since the photograph of the face accurately represents the features of the user U1, it is possible to create an avatar A1 that resembles the user U1.
2:第2実施形態
 以下、図8を参照しつつ、本発明の第2実施形態に係る情報処理装置としてのサーバ10Aを含む情報処理システム1Aの構成について説明する。なお、以下の説明では、説明の簡略化を目的に、第2実施形態に係る情報処理システム1Aが備える構成要素のうち、第1実施形態に係る情報処理システム1と同一の構成要素については、同一の符号を用いると共に、その説明を省略することがある。
2: Second Embodiment Hereinafter, the configuration of an information processing system 1A including a server 10A as an information processing apparatus according to a second embodiment of the present invention will be described with reference to FIG. In the following description, for the purpose of simplification of the description, among the components included in the information processing system 1A according to the second embodiment, the same components as those of the information processing system 1 according to the first embodiment are denoted by the same reference numerals, and the description thereof may be omitted.
2-1:第2実施形態の構成
2-1-1:全体構成
 本発明の第2実施形態に係る情報処理システム1Aは、第1実施形態に係る情報処理システム1に比較して、サーバ10の代わりにサーバ10Aを備える点で異なる。それ以外の点では、情報処理システム1Aの全体構成は、図1に示される第1実施形態に係る情報処理システム1の全体構成と同一であるので、その図示と説明を省略する。
2-1: Configuration of the second embodiment
2-1-1: Overall Configuration An information processing system 1A according to the second embodiment of the present invention differs from the information processing system 1 according to the first embodiment in that a server 10A is provided instead of the server 10. FIG. Otherwise, the overall configuration of the information processing system 1A is the same as the overall configuration of the information processing system 1 according to the first embodiment shown in FIG. 1, so illustration and description thereof will be omitted.
2-1-2:サーバの構成
 サーバ10Aはサーバ10と異なり、処理装置11の代わりに処理装置11Aを、記憶装置12の代わりに記憶装置12Aを備える。記憶装置12Aは、記憶装置12と異なり、制御プログラムPR1の代わりに制御プログラムPR1Aを記憶する。処理装置11Aは、処理装置11と異なり、取得部111の代わりに取得部111Dを、頭部生成部113の代わりに頭部生成部113Aを備える。また、取得部111Dは、取得部111と異なり、顔画像取得部111Bの代わりに顔画像取得部111Eを備える。それ以外の点では、サーバ10Aの構成は、図4及び図5に示される第1実施形態に係るサーバ10の構成と同一であるため、その図示と説明を省略する。
2-1-2: Configuration of Server Unlike the server 10, the server 10A is provided with a processing device 11A instead of the processing device 11 and a storage device 12A instead of the storage device 12. FIG. Unlike the storage device 12, the storage device 12A stores a control program PR1A instead of the control program PR1. Unlike the processing device 11 , the processing device 11A includes an acquisition unit 111D instead of the acquisition unit 111 and a head generation unit 113A instead of the head generation unit 113 . Further, unlike the acquisition unit 111, the acquisition unit 111D includes a face image acquisition unit 111E instead of the face image acquisition unit 111B. Otherwise, the configuration of the server 10A is the same as the configuration of the server 10 according to the first embodiment shown in FIGS. 4 and 5, so illustration and description thereof will be omitted.
 顔画像取得部111Eは、ユーザU1によって選択された第1の顔画像FPkから、ユーザU1の顔の要素を示す要素画像EPを抽出する。 The facial image acquisition unit 111E extracts an elemental image EP representing facial elements of the user U1 from the first facial image FPk selected by the user U1.
 また、頭部生成部113Aは、顔画像取得部111Eによって抽出された要素画像EPを、予め用意された外形画像OPに重畳することにより、アバターA1の頭部を示す三次元画像HPを生成する。とりわけ、頭部生成部113Aは、画風が互いに異なる複数の顔画像FP1~FPnから選択された一つの顔画像FPkに含まれる一つの要素画像EPを、外形画像OPに重畳することにより、アバターA1の頭部を示す三次元画像HPを生成する。 The head generation unit 113A also generates a three-dimensional image HP representing the head of the avatar A1 by superimposing the element image EP extracted by the face image acquisition unit 111E on the outline image OP prepared in advance. In particular, the head generation unit 113A generates a three-dimensional image HP representing the head of the avatar A1 by superimposing one element image EP included in one face image FPk selected from a plurality of face images FP1 to FPn having different styles on the outline image OP.
 図8は、顔画像取得部111E及び頭部生成部113Aの動作例を示す図である。図8に示されるように、顔画像取得部111Eは、ユーザU1の顔画像FPkから、ユーザU1の顔の要素を示す要素画像EPを抽出する。図8に示される例において、顔画像取得部111Eは、顔画像FPkから、顔の要素として、眉毛、目、鼻、及び口を示す要素画像EPを抽出する。しかし、顔画像取得部111Eは、これら眉毛、目、鼻、及び口のうち一部の要素のみを、要素画像EPとして抽出してもよい。あるいは、顔画像取得部111Eは、これら眉毛、目、鼻、及び口に加えて、例えば睫毛、及び黒子等の更なる要素も、要素画像EPとして抽出してもよい。 FIG. 8 is a diagram showing an operation example of the face image acquisition unit 111E and the head generation unit 113A. As shown in FIG. 8, the facial image acquisition unit 111E extracts an element image EP representing facial elements of the user U1 from the facial image FPk of the user U1. In the example shown in FIG. 8, the facial image acquisition unit 111E extracts elemental images EP representing eyebrows, eyes, nose, and mouth as facial elements from the facial image FPk. However, the face image acquisition unit 111E may extract only some of these eyebrows, eyes, nose, and mouth as the element image EP. Alternatively, the face image acquisition unit 111E may extract additional elements such as eyelashes and moles in addition to these eyebrows, eyes, nose, and mouth as elemental images EP.
 また、図8に示されるように、頭部生成部113Aは、抽出した要素画像EPを、予め用意された外形画像OPに重畳することにより、アバターA1の頭部を示す三次元画像HPを生成する。外形画像OPは、二次元画像であってもよく、三次元画像であってもよい。具体的には、頭部生成部113Aは、要素画像EPを二次元画像としての外形画像OPに重畳した後、要素画像EPが重畳された外形画像OPを三次元化することにより、アバターA1の頭部を示す三次元画像HPを生成してもよい。あるいは、頭部生成部113Aは、要素画像EPを、三次元画像としての外形画像OPに重畳することにより、アバターA1の頭部を示す三次元画像HPを生成してもよい。 Also, as shown in FIG. 8, the head generation unit 113A generates a three-dimensional image HP representing the head of the avatar A1 by superimposing the extracted element image EP on the outline image OP prepared in advance. The outline image OP may be a two-dimensional image or a three-dimensional image. Specifically, the head generating unit 113A may generate a three-dimensional image HP representing the head of the avatar A1 by superimposing the element image EP on the external image OP as a two-dimensional image, and then three-dimensionalizing the external image OP superimposed with the element image EP. Alternatively, the head generation unit 113A may generate a three-dimensional image HP representing the head of the avatar A1 by superimposing the element image EP on the outline image OP as a three-dimensional image.
 この結果、サーバ10Aは、テンプレートとして予め用意された外形画像OPを用いながら、様々な画風のアバターA1を作成できる。また、サーバ10Aは、テンプレートとして予め用意された外形画像OPを用いながら、ユーザU1の好みに合った画風のアバターA1を作成できる。 As a result, the server 10A can create avatars A1 in various styles while using outline images OP prepared in advance as templates. In addition, the server 10A can create an avatar A1 with a style that matches the taste of the user U1, using an outline image OP prepared in advance as a template.
2-2:第2実施形態の動作
 第2実施形態に係るサーバ10Aの動作は、基本的に図7によって示される第1実施形態に係るサーバ10の動作と同一であるため、その図示と詳細な説明は省略する。サーバ10Aの動作では、ステップS3において、処理装置11Aは、顔画像取得部111Eとして機能する。処理装置11Aは、第1の顔画像FPkを取得した後、当該顔画像FPkから、ユーザU1の顔の要素を示す要素画像EPを抽出する。また、ステップS4において、処理装置11Aは、頭部生成部113Aとして機能する。処理装置11Aは、抽出した要素画像EPを、予め用意された外形画像OPに重畳して、アバターA1の頭部を示す三次元画像HPを生成する。
2-2: Operation of Second Embodiment Since the operation of the server 10A according to the second embodiment is basically the same as the operation of the server 10 according to the first embodiment shown in FIG. 7, its illustration and detailed description are omitted. In the operation of the server 10A, in step S3, the processing device 11A functions as the facial image acquiring section 111E. After obtaining the first facial image FPk, the processing device 11A extracts an elemental image EP representing facial elements of the user U1 from the facial image FPk. Further, in step S4, the processing device 11A functions as the head generating section 113A. The processing device 11A superimposes the extracted elemental image EP on the outline image OP prepared in advance to generate a three-dimensional image HP representing the head of the avatar A1.
2-3:第2実施形態が奏する効果
 以上の説明によれば、情報処理装置としてのサーバ10Aにおいて、顔画像取得部111Eは、選択された第1の顔画像FPkから、ユーザU1の顔の要素を示す要素画像EPを抽出する。頭部生成部113Aは、抽出された要素画像EPを、予め用意された外形画像OPに重畳することにより、アバターA1の頭部を示す三次元画像HPを生成する。
2-3: Effects of the Second Embodiment According to the above description, in the server 10A as an information processing device, the facial image acquisition unit 111E extracts the element image EP representing the facial elements of the user U1 from the selected first facial image FPk. The head generation unit 113A generates a three-dimensional image HP representing the head of the avatar A1 by superimposing the extracted element image EP on the outline image OP prepared in advance.
 サーバ10Aは、上記の構成を備えるので、テンプレートとして予め用意された外形画像OPを用いながら、ユーザU1は、互いに異なる画風を有する複数のアバターAの中から、自身の好みに合った画風のアバターA1を利用できる。とりわけ、サーバ10Aは、互いに異なる複数のアバターAの頭部を示す三次元画像HPの生成時に、テンプレートとして予め用意された外形画像OPを共通して用いる。その結果、サーバ10Aは、当該複数のアバターAに統一感を持たせられる。更には、サーバ10Aは、当該複数のアバターAに統一感を持たせながら、各々のアバターAを、ユーザU毎に異なる好みに合わせた画風とすることができる。 Since the server 10A has the above configuration, the user U1 can use an avatar A1 with a style that suits his/her taste from among a plurality of avatars A with different styles while using the outline image OP prepared in advance as a template. In particular, the server 10A commonly uses outline images OP prepared in advance as templates when generating three-dimensional images HP showing the heads of a plurality of avatars A that are different from each other. As a result, the server 10A allows the plurality of avatars A to have a sense of unity. Furthermore, the server 10A can make each of the avatars A have a style that matches the preferences of each user U while giving the plurality of avatars A a sense of unity.
3:第3実施形態
 以下、図9~図11を参照しつつ、本発明の第3実施形態に係る情報処理装置としてのサーバ10Bを含む情報処理システム1Bの構成について説明する。なお、以下の説明では、説明の簡略化を目的に、第3実施形態に係る情報処理システム1Bが備える構成要素のうち、第1実施形態に係る情報処理システム1と同一の構成要素については、同一の符号を用いると共に、その説明を省略することがある。
3: Third Embodiment Hereinafter, the configuration of an information processing system 1B including a server 10B as an information processing apparatus according to a third embodiment of the present invention will be described with reference to FIGS. 9 to 11. FIG. In the following description, for the purpose of simplifying the description, among the components provided in the information processing system 1B according to the third embodiment, the same components as those of the information processing system 1 according to the first embodiment are denoted by the same reference numerals, and the description thereof may be omitted.
3-1:第3実施形態の構成
3-1-1:全体構成
 本発明の第3実施形態に係る情報処理システム1Bは、第1実施形態に係る情報処理システム1に比較して、サーバ10の代わりにサーバ10Bを備える点で異なる。それ以外の点では、情報処理システム1Bの全体構成は、図1に示される第1実施形態に係る情報処理システム1の全体構成と同一であるので、その図示と説明を省略する。
3-1: Configuration of the third embodiment
3-1-1: Overall Configuration An information processing system 1B according to the third embodiment of the present invention differs from the information processing system 1 according to the first embodiment in that a server 10B is provided instead of the server 10. FIG. Otherwise, the overall configuration of the information processing system 1B is the same as the overall configuration of the information processing system 1 according to the first embodiment shown in FIG. 1, so illustration and description thereof will be omitted.
3-1-2:サーバの構成
 図9は、サーバ10Bの構成例を示すブロック図である。サーバ10Bはサーバ10と異なり、処理装置11の代わりに処理装置11Bを、記憶装置12の代わりに記憶装置12Bを備える。
3-1-2: Server Configuration FIG. 9 is a block diagram showing a configuration example of the server 10B. Unlike the server 10, the server 10B includes a processing device 11B instead of the processing device 11 and a storage device 12B instead of the storage device 12. FIG.
 記憶装置12Bは、記憶装置12と異なり、制御プログラムPR1の代わりに制御プログラムPR1Bを記憶する。また、記憶装置12Bは、記憶装置12によって記憶される構成要素に加えて、学習モデルLM及び頭身数テーブルHTを更に記憶する。 Unlike the storage device 12, the storage device 12B stores the control program PR1B instead of the control program PR1. In addition to the components stored by the storage device 12, the storage device 12B also stores a learning model LM and a head/height number table HT.
 学習モデルLMは、入力画像取得部111Aによって取得された、ユーザU1の顔写真を示す入力画像IPに基づいて、年齢推定部117がユーザU1の年齢を推定するための学習モデルである。 The learning model LM is a learning model for the age estimating unit 117 to estimate the age of the user U1 based on the input image IP representing the facial photograph of the user U1 acquired by the input image acquiring unit 111A.
 学習モデルLMは、学習フェーズにおいて、教師データを学習することによって生成される。学習モデルLMを生成するために用いられる教師データは、入力画像取得部111Aによって取得された、1人の人物の顔写真を示す入力画像IPから抽出される特徴情報と、当該人物の年齢との1対1の組を複数有する。 The learning model LM is generated by learning teacher data in the learning phase. The teacher data used to generate the learning model LM has a plurality of one-to-one pairs of the feature information extracted from the input image IP showing the facial photograph of one person, which is acquired by the input image acquisition unit 111A, and the age of the person.
 また、学習モデルLMは、サーバ10Bの外部において生成される。とりわけ学習モデルLMは、図示しない第2のサーバにおいて生成されることが好適である。この場合、サーバ10Bは、通信網NETを介して図示しない第2のサーバから学習モデルLMを取得する。 Also, the learning model LM is generated outside the server 10B. Especially, the learning model LM is preferably generated in a second server (not shown). In this case, the server 10B acquires the learning model LM from a second server (not shown) via the communication network NET.
 頭身数テーブルHTは、後述の年齢推定部117によって推定された年齢と、頭身数HBとの対応関係を定義するための表である。図10は、頭身数テーブルHTの例を示す。図10に示される頭身数テーブルHTにおいては、一例として、年齢推定部117によって推定された年齢が0歳以上3歳未満の場合には、アバターA1が2頭身となることが定義される。 The head and body count table HT is a table for defining the correspondence relationship between the age estimated by the age estimation unit 117 described later and the head and body count HB. FIG. 10 shows an example of the head/body number table HT. As an example, in the head-to-body number table HT shown in FIG. 10, when the age estimated by the age estimation unit 117 is between 0 and 3 years old, it is defined that the avatar A1 has 2 heads and bodies.
 処理装置11Bは、処理装置11と異なり、取得部111の代わりに取得部111Fを備える。取得部111Fは、取得部111と異なり、頭身取得部111Cの代わりに頭身取得部111Gを備える。それ以外の点では、取得部111Fの構成は、図5に示される第1実施形態に係る取得部111の構成と同一であるので、その図示と説明を省略する。また、処理装置11Bは、処理装置11に備わる構成要素に加えて、年齢推定部117及び頭身数生成部118を備える。 Unlike the processing device 11, the processing device 11B includes an acquisition unit 111F instead of the acquisition unit 111. Unlike the acquisition unit 111, the acquisition unit 111F includes a head/body acquisition unit 111G instead of the head/body acquisition unit 111C. Other than that, the configuration of the acquisition unit 111F is the same as the configuration of the acquisition unit 111 according to the first embodiment shown in FIG. 5, so illustration and description thereof will be omitted. In addition to the components of the processing device 11, the processing device 11B includes an age estimation unit 117 and a head/body number generation unit 118. FIG.
 年齢推定部117は、ユーザU1の顔写真を示す入力画像IPに基づいて、ユーザU1の年齢を推定する。より詳細には、年齢推定部117は、入力画像取得部111Aによって取得された入力画像IPを、学習モデルLMに入力する。その後、年齢推定部117は、学習モデルLMから推定年齢を出力させることにより、ユーザU1の推定年齢を生成する。また、年齢推定部117は、生成された推定年齢を頭身数生成部118に出力する。 The age estimation unit 117 estimates the age of the user U1 based on the input image IP showing the photograph of the face of the user U1. More specifically, the age estimation unit 117 inputs the input image IP acquired by the input image acquisition unit 111A to the learning model LM. After that, the age estimation unit 117 generates the estimated age of the user U1 by outputting the estimated age from the learning model LM. Age estimation section 117 also outputs the generated estimated age to head/body number generation section 118 .
 頭身数生成部118は、年齢推定部117によって推定された年齢に基づいてアバターA1が何頭身であるかを表す頭身数HBを生成する。より詳細には、頭身数生成部118は、年齢推定部117から取得した年齢を、記憶装置12Bに格納される頭身数テーブルHTに照合することにより、アバターA1の頭身数HBを生成する。 Based on the age estimated by the age estimation unit 117, the head/body number generation unit 118 generates a head/body number HB that indicates the size of the avatar A1. More specifically, the head-to-body number generation unit 118 generates the head-to-body number HB of the avatar A1 by checking the age acquired from the age estimation unit 117 against the head-to-body number table HT stored in the storage device 12B.
 頭身取得部111Gは、頭身生成部118によって生成されたアバターA1の頭身数HBを取得する。 The head and body acquisition unit 111G acquires the head and body number HB of the avatar A1 generated by the head and body generation unit 118.
 この結果、サーバ10Bは、ユーザU1の推定年齢によってもたらされる印象と合致した頭身数HBで、アバターA1を生成できる。 As a result, the server 10B can generate the avatar A1 with the head size HB that matches the impression given by the estimated age of the user U1.
3-2:第3実施形態の動作
 図11は、第3実施形態に係るサーバ10Bの動作を示すフローチャートである。以下、図11を参照することにより、サーバ10Bの動作について説明する。
3-2: Operation of the Third Embodiment FIG. 11 is a flow chart showing the operation of the server 10B according to the third embodiment. The operation of the server 10B will be described below with reference to FIG.
 ステップS11において、処理装置11Bは、入力画像取得部111Aとして機能する。処理装置11Bは、端末装置20-1から、ユーザU1の顔の正面部を示す入力画像IPを示す画像情報を取得する。 In step S11, the processing device 11B functions as an input image acquisition section 111A. The processing device 11B acquires, from the terminal device 20-1, image information representing the input image IP representing the front portion of the face of the user U1.
 ステップS12において、処理装置11Bは、顔画像生成部112として機能する。処理装置11Bは、ステップS11において取得されたユーザU1の顔の正面部を示す入力画像IPに基づいて、画風が互いに異なる複数の顔画像FP1~FPnを生成する。また、処理装置11Bは、生成した顔画像FP1~FPnを、記憶装置12Bに格納する。 In step S12, the processing device 11B functions as the face image generator 112. The processing device 11B generates a plurality of face images FP1 to FPn having different styles based on the input image IP representing the front portion of the face of the user U1 acquired in step S11. Further, the processing device 11B stores the generated face images FP1 to FPn in the storage device 12B.
 ステップS13において、処理装置11Bは、顔画像取得部111Bとして機能する。処理装置11Bは、端末装置20から、一つの顔画像FPkの選択結果kを取得する。また、処理装置11Bは、顔画像取得部111Bとして機能する。処理装置11Bは、当該選択結果kに基づいて、記憶装置12Bから顔画像FPkを取得する。 In step S13, the processing device 11B functions as the face image acquiring section 111B. The processing device 11B acquires the selection result k of one face image FPk from the terminal device 20 . In addition, the processing device 11B functions as a facial image acquisition section 111B. The processing device 11B acquires the face image FPk from the storage device 12B based on the selection result k.
 ステップS14において、処理装置11Bは、頭部生成部113として機能する。処理装置11Bは、ステップS13において取得された一つの顔画像FPkに基づいて、アバターA1の頭部を示す三次元画像HPを生成する。 In step S14, the processing device 11B functions as the head generation unit 113. The processing device 11B generates a three-dimensional image HP representing the head of the avatar A1 based on one face image FPk acquired in step S13.
 ステップS15において、処理装置11Bは、年齢推定部117として機能する。処理装置11Bは、ステップS11において取得された入力画像IPに基づいて、ユーザU1の年齢を推定する。 In step S15, the processing device 11B functions as the age estimation unit 117. The processing device 11B estimates the age of the user U1 based on the input image IP acquired in step S11.
 ステップS16において、処理装置11Bは、頭身生成部118として機能する。ことにより、処理装置11Bは、ステップS15において推定されたユーザU1の年齢に基づいて、アバターA1の頭身数HBを生成する。 In step S16, the processing device 11B functions as the head/body generation unit 118. Accordingly, the processing device 11B generates the head-to-body count HB of the avatar A1 based on the age of the user U1 estimated in step S15.
 ステップS17において、処理装置11Bは、頭身取得部111Gとして機能する。処理装置11Bは、ステップS16において生成された頭身数HBを取得する。 In step S17, the processing device 11B functions as the head/body acquisition unit 111G. The processing device 11B acquires the head/body count HB generated in step S16.
 ステップS18において、処理装置11Bは、体部生成部114として機能する。処理装置11Bは、ステップS17において取得された頭身数HBと、記憶装置12Bに記憶されるアバター情報AIとを用いて、アバターA1の頭部以外の部分である体部の三次元画像BPを示す画像情報を生成する。 In step S18, the processing device 11B functions as the body generation unit 114. The processing device 11B generates image information representing a three-dimensional image BP of the body of the avatar A1 other than the head, using the head/body count HB obtained in step S17 and the avatar information AI stored in the storage device 12B.
 ステップS19において、処理装置11Bは、アバター生成部115として機能する。処理装置11Bは、アバターA1の頭部を示す三次元画像HPを示す画像情報、及びアバターA1の体部を示す三次元画像BPを示す画像情報を用いて、アバターA1の全体の外観を示す三次元画像WPを示す画像情報を生成する。 In step S19, the processing device 11B functions as the avatar generation unit 115. The processing device 11B uses the image information representing the three-dimensional image HP representing the head of the avatar A1 and the image information representing the three-dimensional image BP representing the body of the avatar A1 to generate the image information representing the three-dimensional image WP representing the overall appearance of the avatar A1.
 ステップS20において、処理装置11Bは、出力部116として機能する。処理装置11Bは、ステップS19において生成されたアバターA1の全体の外観を示す三次元画像WPを示す画像情報を、通信装置13を介して端末装置20に送信する。その後、処理装置11Bは、図11に示される全ての処理を終了する。 In step S20, the processing device 11B functions as the output unit 116. The processing device 11B transmits image information representing a three-dimensional image WP representing the overall appearance of the avatar A1 generated in step S19 to the terminal device 20 via the communication device 13 . After that, the processing device 11B ends all the processes shown in FIG.
3-3:第3実施形態が奏する効果
 以上の説明によれば、情報処理装置としてのサーバ10Bは、年齢推定部117と頭身数生成部118を備える。年齢推定部117は、ユーザU1の顔写真を示す画像に基づいて、ユーザU1の年齢を推定する。頭身生成部118は、推定された年齢に基づいて、アバターA1の頭身数HBを生成する。
3-3: Effect of the Third Embodiment According to the above description, the server 10B as an information processing device includes the age estimating section 117 and the head/body number generating section 118 . The age estimation unit 117 estimates the age of the user U1 based on the image showing the photograph of the face of the user U1. The head-to-body generation unit 118 generates the head-to-body number HB of the avatar A1 based on the estimated age.
 サーバ10Bは、上記の構成を備えるので、互いに異なる画風を有する複数のアバターAの中から、ユーザU1の好みに合った画風のアバターA1をユーザU1が利用する上で、ユーザU1の推定年齢によってもたらされる印象と合致した頭身数HBのアバターA1を利用できる。また、頭身生成部118は、年齢推定部117によって推定された年齢に基づいて頭身数HBを生成するため、ユーザU1自身は頭身数HBを入力する必要がない。すなわち、ユーザU1は、簡便な方法で、ユーザU1の推定年齢によってもたらされる印象と合致したアバターA1を利用できる。 Since the server 10B has the above configuration, when the user U1 uses the avatar A1 with the style that suits the taste of the user U1 from among the plurality of avatars A having different styles, the avatar A1 with the head size HB that matches the impression given by the estimated age of the user U1 can be used. Moreover, since the head/body generation unit 118 generates the head/body number HB based on the age estimated by the age estimation unit 117, the user U1 does not need to input the head/body number HB. That is, the user U1 can use the avatar A1 that matches the impression given by the estimated age of the user U1 by a simple method.
4:変形例
 本開示は、以上に例示した実施形態に限定されない。具体的な変形の態様を以下に例示する。以下の例示から任意に選択された2以上の態様を併合してもよい。
4: Modifications The present disclosure is not limited to the embodiments illustrated above. Specific modification modes are exemplified below. Two or more aspects arbitrarily selected from the following examples may be combined.
4-1:変形例1
 上記の実施形態に係る情報処理システム1~1Bにおいて、サーバ10~10Bは、端末装置20に備わるディスプレイ24にアバターAを表示させていた。しかし、サーバ10~サーバ10Bは、ディスプレイ24の代わりにXRグラスにアバターAを表示させてもよい。
4-1: Modification 1
In the information processing systems 1 to 1B according to the above embodiments, the servers 10 to 10B display the avatar A on the display 24 of the terminal device 20. FIG. However, servers 10 to 10B may display avatar A on XR glasses instead of display 24 .
 図12は、本変形例に係る情報処理システム1Cの全体構成を示す図である。情報処理システム1Cは、XRグラス30を装着したユーザU1及びユーザU2に対して、XR技術を用いて仮想空間を提供する。とりわけ本実施形態において、情報処理システム1Cは、ユーザU1に対応するアバターA1と、ユーザU2に対応するアバターA2とを、XRグラス30に表示させる。なお、XR技術とは、VR(Virtual Reality)技術、AR(Augmented Reality)技術、及びMR(Mixed Reality)技術の総称である。VR技術とは、VRグラス、又はVR技術が採用されたHMD(Head Mounted Diplay)等の装置に対して、デジタルの仮想空間を表示させる技術である。AR技術とは、ARグラス、又はAR技術が採用されたHMD等の装置に表示される拡張現実空間において、現実世界にデジタルコンテンツによって示される情報を付加する技術である。MR技術とは、MRグラス、又はMR技術が採用されたHMD(Head Mounted Display)等の装置を用いて、現実空間に対してデジタルの仮想空間を精緻に重ね合わせる技術のことである。 FIG. 12 is a diagram showing the overall configuration of an information processing system 1C according to this modified example. The information processing system 1C uses XR technology to provide a virtual space to users U1 and U2 wearing the XR glasses 30 . Especially in this embodiment, the information processing system 1C causes the XR glasses 30 to display an avatar A1 corresponding to the user U1 and an avatar A2 corresponding to the user U2. Note that XR technology is a general term for VR (Virtual Reality) technology, AR (Augmented Reality) technology, and MR (Mixed Reality) technology. VR technology is technology for displaying a digital virtual space on a device such as VR glasses or an HMD (Head Mounted Display) employing VR technology. AR technology is technology that adds information indicated by digital content to the real world in an augmented reality space displayed on a device such as AR glasses or an HMD that employs AR technology. MR technology is a technology that precisely superimposes a digital virtual space on a real space using MR glasses or a device such as an HMD (Head Mounted Display) employing MR technology.
 図12に示されるように、情報処理システム1Cは、サーバ10、端末装置20、及びXRグラス30を備える。情報処理システム1Cにおいて、サーバ10と端末装置20とは、通信網NETを介して互いに通信可能に接続される。また、端末装置20とXRグラス30とは互いに通信可能に接続される。以下の説明において、ユーザごとに使用されるXRグラス30を区別する場合は、符号に添え字「-X」を用いる。また、XRグラス30各々の構成要素についても同様である。なお、図12において、端末装置20とXRグラス30との組として、端末装置20-1とXRグラス30-1との組、及び端末装置20-2とXRグラス30-2との組の合計2組が記載される。しかし、当該組数はあくまで一例であって、情報処理システム1Cは、任意の数の端末装置20とXRグラス30との組を備え得る。図12においては、ユーザU1が端末装置20-1とXRグラス30-1の組を使用し、ユーザU2が端末装置20-2とXRグラス30-2の組を使用することを前提とする。 As shown in FIG. 12, the information processing system 1C includes a server 10, a terminal device 20, and XR glasses 30. In the information processing system 1C, the server 10 and the terminal device 20 are communicably connected to each other via a communication network NET. Also, the terminal device 20 and the XR glasses 30 are connected so as to be able to communicate with each other. In the following description, when distinguishing the XR glasses 30 used for each user, the suffix "-X" is used for the reference numerals. In addition, the same is true for each component of the XR glass 30 . In FIG. 12, two pairs are shown as pairs of the terminal device 20 and the XR glasses 30: the pair of the terminal device 20-1 and the XR glasses 30-1 and the pair of the terminal device 20-2 and the XR glasses 30-2. However, the number of sets is merely an example, and the information processing system 1C can include any number of sets of the terminal device 20 and the XR glasses 30 . In FIG. 12, it is assumed that user U1 uses a set of terminal device 20-1 and XR glasses 30-1, and user U2 uses a set of terminal device 20-2 and XR glasses 30-2.
 サーバ10は、通信網NETを介して、端末装置20に対して各種データ及びクラウドサービスを提供する。とりわけ、サーバ10は、端末装置20に接続されるXRグラス30に、ユーザU1に対応するアバターA1、及びユーザU2に対応するアバターA2を表示させるための各種データを、端末装置20に対して提供する。より詳細には、サーバ10は、ユーザU1が使用するXRグラス30-1に備わるディスプレイ38-1に、アバターA2を表示させるための各種データを、端末装置20-1に対して提供する。また、サーバ10は、ユーザU2が使用するXRグラス30-2に備わるディスプレイ38-2に、アバターA1を表示させるための各種データを、端末装置20-2に対して提供する。 The server 10 provides various data and cloud services to the terminal device 20 via the communication network NET. In particular, the server 10 provides the terminal device 20 with various data for displaying the avatar A1 corresponding to the user U1 and the avatar A2 corresponding to the user U2 on the XR glasses 30 connected to the terminal device 20 . More specifically, the server 10 provides the terminal device 20-1 with various data for displaying the avatar A2 on the display 38-1 of the XR glasses 30-1 used by the user U1. The server 10 also provides the terminal device 20-2 with various data for displaying the avatar A1 on the display 38-2 of the XR glasses 30-2 used by the user U2.
 端末装置20-1は、ユーザU1が頭部に装着するXRグラス30-1に対して、仮想空間に配置される仮想オブジェクトを表示させる。また、端末装置20-2は、ユーザU2が頭部に装着するXRグラス30-2に対して、仮想空間に配置される仮想オブジェクトを表示させる。当該仮想空間は、一例として、天球型の空間である。また、仮想オブジェクトは、例として、静止画像、動画、3DCGモデル、HTMLファイル、及びテキストファイル等のデータを示す仮想オブジェクト、及びアプリケーションを示す仮想オブジェクトである。ここで、テキストファイルとしては、例として、メモ、ソースコード、日記、及びレシピが挙げられる。また、アプリケーションとしては、例として、ブラウザ、SNSを用いるためのアプリケーション、及びドキュメントファイルを生成するためのアプリケーションが挙げられる。なお、端末装置20-1及び端末装置20-2は、例として、スマートフォン、及びタブレット等の携帯端末装置であることが好適である。 The terminal device 20-1 causes the XR glasses 30-1 worn on the head by the user U1 to display virtual objects arranged in the virtual space. Further, the terminal device 20-2 causes the XR glasses 30-2 worn on the head of the user U2 to display a virtual object arranged in the virtual space. The virtual space is, for example, a celestial space. The virtual objects are, for example, virtual objects representing data such as still images, moving images, 3DCG models, HTML files, and text files, and virtual objects representing applications. Examples of text files include memos, source codes, diaries, and recipes. Examples of applications include browsers, applications for using SNS, and applications for generating document files. The terminal device 20-1 and the terminal device 20-2 are preferably portable terminal devices such as smartphones and tablets, for example.
 とりわけ本実施形態において、端末装置20-1は、XRグラス30-1に対して、主としてアバターA2に対応する仮想オブジェクトを表示させる。また、端末装置20-2は、XRグラス30-2に対して、主としてアバターA1に対応する仮想オブジェクトを表示させる。 Especially in this embodiment, the terminal device 20-1 causes the XR glasses 30-1 to display a virtual object mainly corresponding to the avatar A2. Also, the terminal device 20-2 displays a virtual object mainly corresponding to the avatar A1 on the XR glasses 30-2.
 XRグラス30は、ユーザU1及びユーザU2の頭部に装着する表示装置である。より詳細には、XRグラス30-1は、ユーザU1の頭部に装着する表示装置である。また、XRグラス30-2は、ユーザU2の頭部に装着する表示装置である。XRグラス30は、一例として、シースルー型のウェアラブルディスプレイである。XRグラス30は、端末装置20が制御することによって、両眼用のレンズの各々に対応して設けられた表示パネルに仮想オブジェクトを表示させる。 The XR glasses 30 are display devices worn on the heads of users U1 and U2. More specifically, the XR glasses 30-1 are display devices worn on the head of the user U1. Also, the XR glasses 30-2 are a display device worn on the head of the user U2. The XR glasses 30 are, for example, a see-through wearable display. The XR glasses 30 are controlled by the terminal device 20 to display a virtual object on the display panel provided corresponding to each of the binocular lenses.
 以上の構成において、ユーザU1及びユーザU2は、ディスプレイ38に表示されるアバターA1及びA2を観察できる。より詳細には、XRグラス30-1を頭部に装着したユーザU1は、ディスプレイ38-1に表示されるアバターA2を観察できる。一方で、XRグラス30-2を頭部に装着したユーザU2は、ディスプレイ38-2に表示されるアバターA1を観察できる。 With the above configuration, the user U1 and the user U2 can observe the avatars A1 and A2 displayed on the display 38. More specifically, the user U1 wearing the XR glasses 30-1 can observe the avatar A2 displayed on the display 38-1. On the other hand, the user U2 wearing the XR glasses 30-2 can observe the avatar A1 displayed on the display 38-2.
 なお、情報処理システム1Cにおいて、端末装置20とXRグラス30とは別体として実現されている。しかし、本変形例における、端末装置20とXRグラス30の実現方法は、これには限定されない。例えば、XRグラス30が、端末装置20と同一の機能を備えることにより、端末装置20とXRグラス30とが単一の筐体内において実現されてもよい。 It should be noted that in the information processing system 1C, the terminal device 20 and the XR glasses 30 are implemented separately. However, the method of realizing the terminal device 20 and the XR glasses 30 in this modified example is not limited to this. For example, the terminal device 20 and the XR glasses 30 may be implemented in a single housing by providing the XR glasses 30 with the same functions as the terminal device 20 .
 また、情報処理システム1Cは、XRグラス30の代わりに、VR技術、AR技術、及びMR技術のうちいずれか1つの技術が採用されたHMD等の装置を備えてもよい。 Also, the information processing system 1C may include a device such as an HMD that employs any one of VR technology, AR technology, and MR technology instead of the XR glasses 30 .
4-2:変形例2
 上記の実施形態に係る情報処理システム1~1Cにおいて、端末装置20-1からサーバ10~10Bに対して、複数の顔画像FP1~FPnから一つの顔画像FPkが選択された選択結果kが出力される。しかし、端末装置20-1からサーバ10~10Bに対して、選択結果kではなく、顔画像FPk自体が出力されてもよい。
4-2: Modification 2
In the information processing systems 1 to 1C according to the above embodiments, the terminal device 20-1 outputs a selection result k in which one face image FPk is selected from the plurality of face images FP1 to FPn to the servers 10 to 10B. However, instead of the selection result k, the face image FPk itself may be output from the terminal device 20-1 to the servers 10 to 10B.
4-3:変形例3
 上記の実施形態に係る情報処理システム1~1Cにおいて、サーバ10~10Bは、端末装置20から、ユーザU1の顔の正面部を示す入力画像IPを取得する。しかし、サーバ10~10Bは、端末装置20以外の装置から、入力画像IPを取得してもよい。
4-3: Modification 3
In the information processing systems 1 to 1C according to the above embodiments, the servers 10 to 10B acquire from the terminal device 20 the input image IP representing the front portion of the face of the user U1. However, the servers 10 to 10B may acquire the input image IP from a device other than the terminal device 20. FIG.
5:その他
(1)上述した実施形態では、記憶装置12~12B、記憶装置22は、ROM及びRAM等を例示したが、フレキシブルディスク、光磁気ディスク(例えば、コンパクトディスク、デジタル多用途ディスク、Blu-ray(登録商標)ディスク)、スマートカード、フラッシュメモリデバイス(例えば、カード、スティック、キードライブ)、CD-ROM(Compact Disc-ROM)、レジスタ、リムーバブルディスク、ハードディスク、フロッピー(登録商標)ディスク、磁気ストリップ、データベース、サーバその他の適切な記憶媒体である。また、プログラムは、電気通信回線を介してネットワークから送信されてもよい。また、プログラムは、電気通信回線を介して通信網NETから送信されてもよい。
5: Others (1) In the above-described embodiments, the storage devices 12 to 12B and the storage device 22 were ROM and RAM, etc., but flexible disks, magneto-optical disks (e.g., compact discs, digital versatile discs, Blu-ray (registered trademark) discs), smart cards, flash memory devices (e.g., cards, sticks, key drives), CD-ROMs (Compact Disc-ROM), registers, removable disks, hard disks, floppy (registered trademark) disks, magnetic strips, databases, servers, and other suitable A storage medium. Also, the program may be transmitted from a network via an electric communication line. Also, the program may be transmitted from the communication network NET via an electric communication line.
(2)上述した実施形態において、説明した情報、信号等は、様々な異なる技術のいずれかを使用して表されてもよい。例えば、上記の説明全体に渡って言及され得るデータ、命令、コマンド、情報、信号、ビット、シンボル、チップ等は、電圧、電流、電磁波、磁界若しくは磁性粒子、光場若しくは光子、又はこれらの任意の組み合わせによって表されてもよい。 (2) In the embodiments described above, the information, signals, etc. described may be represented using any of a variety of different technologies. For example, the data, instructions, commands, information, signals, bits, symbols, chips, etc. that may be referred to throughout the above description may be represented by voltages, currents, electromagnetic waves, magnetic fields or magnetic particles, optical fields or photons, or any combination thereof.
(3)上述した実施形態において、入出力された情報等は特定の場所(例えば、メモリ)に保存されてもよいし、管理テーブルを用いて管理してもよい。入出力される情報等は、上書き、更新、又は追記され得る。出力された情報等は削除されてもよい。入力された情報等は他の装置へ送信されてもよい。 (3) In the above-described embodiments, input/output information and the like may be stored in a specific location (for example, memory), or may be managed using a management table. Input/output information and the like can be overwritten, updated, or appended. The output information and the like may be deleted. The entered information and the like may be transmitted to another device.
(4)上述した実施形態において、判定は、1ビットを用いて表される値(0か1か)によって行われてもよいし、真偽値(Boolean:true又はfalse)によって行われてもよいし、数値の比較(例えば、所定の値との比較)によって行われてもよい。 (4) In the above-described embodiments, the determination may be made by a value (0 or 1) represented using 1 bit, by a boolean value (Boolean: true or false), or by numerical comparison (for example, comparison with a predetermined value).
(5)上述した実施形態において例示した処理手順、シーケンス、フローチャート等は、矛盾の無い限り、順序を入れ替えてもよい。例えば、本開示において説明した方法については、例示的な順序を用いて様々なステップの要素を提示しており、提示した特定の順序に限定されない。 (5) As long as there is no contradiction, the order of the processing procedures, sequences, flowcharts, and the like exemplified in the above-described embodiments may be changed. For example, the methods described in this disclosure present elements of the various steps using a sample order, and are not limited to the specific order presented.
(6)図1~図12に例示された各機能は、ハードウェア及びソフトウェアの少なくとも一方の任意の組み合わせによって実現される。また、各機能ブロックの実現方法は特に限定されない。すなわち、各機能ブロックは、物理的又は論理的に結合した1つの装置を用いて実現されてもよいし、物理的又は論理的に分離した2つ以上の装置を直接的又は間接的に(例えば、有線、無線等を用いて)接続し、これら複数の装置を用いて実現されてもよい。機能ブロックは、上記1つの装置又は上記複数の装置にソフトウェアを組み合わせて実現されてもよい。 (6) Each function illustrated in FIGS. 1 to 12 is realized by any combination of at least one of hardware and software. Also, the method of realizing each functional block is not particularly limited. That is, each functional block may be implemented using one device that is physically or logically coupled, or two or more devices that are physically or logically separated may be directly or indirectly (e.g., wired, wireless, etc.) connected and implemented using these multiple devices. A functional block may be implemented by combining software in the one device or the plurality of devices.
(7)上述した実施形態において例示したプログラムは、ソフトウェア、ファームウェア、ミドルウェア、マイクロコード、ハードウェア記述言語と呼ばれるか、他の名称を用いて呼ばれるかを問わず、命令、命令セット、コード、コードセグメント、プログラムコード、プログラム、サブプログラム、ソフトウェアモジュール、アプリケーション、ソフトウェアアプリケーション、ソフトウェアパッケージ、ルーチン、サブルーチン、オブジェクト、実行可能ファイル、実行スレッド、手順、機能等を意味するよう広く解釈されるべきである。 (7) The programs illustrated in the above embodiments should be construed broadly to mean instructions, instruction sets, code, code segments, program code, programs, subprograms, software modules, applications, software applications, software packages, routines, subroutines, objects, executables, threads of execution, procedures, functions, etc., whether referred to as software, firmware, middleware, microcode, hardware description language, or by any other name.
 また、ソフトウェア、命令、情報等は、伝送媒体を介して送受信されてもよい。例えば、ソフトウェアが、有線技術(同軸ケーブル、光ファイバケーブル、ツイストペア、デジタル加入者回線(DSL:Digital Subscriber Line)等)及び無線技術(赤外線、マイクロ波等)の少なくとも一方を使用してウェブサイト、サーバ、又は他のリモートソースから送信される場合、これらの有線技術及び無線技術の少なくとも一方は、伝送媒体の定義内に含まれる。 In addition, software, instructions, information, etc. may be transmitted and received via a transmission medium. For example, if the software is transmitted from a website, server, or other remote source using wired technologies (coaxial cable, fiber optic cable, twisted pair, Digital Subscriber Line (DSL), etc.) and/or wireless technologies (infrared, microwave, etc.), then these wired and/or wireless technologies are included within the definition of transmission medium.
(8)前述の各形態において、「システム」及び「ネットワーク」という用語は、互換的に使用される。 (8) In each of the above aspects, the terms "system" and "network" are used interchangeably.
(9)本開示において説明した情報、パラメータ等は、絶対値を用いて表されてもよいし、所定の値からの相対値を用いて表されてもよいし、対応する別の情報を用いて表されてもよい。 (9) Information, parameters, etc. described in the present disclosure may be expressed using absolute values, may be expressed using relative values from a predetermined value, or may be expressed using corresponding separate information.
(10)上述した実施形態において、サーバ10~10B、及び端末装置20は、移動局(MS:Mobile Station)である場合が含まれる。移動局は、当業者によって、加入者局、モバイルユニット、加入者ユニット、ワイヤレスユニット、リモートユニット、モバイルデバイス、ワイヤレスデバイス、ワイヤレス通信デバイス、リモートデバイス、モバイル加入者局、アクセス端末、モバイル端末、ワイヤレス端末、リモート端末、ハンドセット、ユーザエージェント、モバイルクライアント、クライアント、又はいくつかの他の適切な用語によって呼ばれる場合もある。また、本開示においては、「移動局」、「ユーザ端末(user terminal)」、「ユーザ装置(UE:User Equipment)」、「端末」等の用語は、互換的に使用され得る。 (10) In the above-described embodiments, the servers 10 to 10B and the terminal device 20 may be mobile stations (MS). A mobile station may also be referred to by those skilled in the art as a subscriber station, mobile unit, subscriber unit, wireless unit, remote unit, mobile device, wireless device, wireless communication device, remote device, mobile subscriber station, access terminal, mobile terminal, wireless terminal, remote terminal, handset, user agent, mobile client, client, or some other suitable term. Also, in the present disclosure, terms such as "mobile station", "user terminal", "user equipment (UE)", "terminal", etc. may be used interchangeably.
(11)上述した実施形態において、「接続された(connected)」、「結合された(coupled)」という用語、又はこれらのあらゆる変形は、2又はそれ以上の要素間の直接的又は間接的なあらゆる接続又は結合を意味し、互いに「接続」又は「結合」された2つの要素間に1又はそれ以上の中間要素が存在することを含められる。要素間の結合又は接続は、物理的な結合又は接続であっても、論理的な結合又は接続であっても、或いはこれらの組み合わせであってもよい。例えば、「接続」は「アクセス」を用いて読み替えられてもよい。本開示において使用する場合、2つの要素は、1又はそれ以上の電線、ケーブル及びプリント電気接続の少なくとも一つを用いて、並びにいくつかの非限定的かつ非包括的な例として、無線周波数領域、マイクロ波領域及び光(可視及び不可視の両方)領域の波長を有する電磁エネルギー等を用いて、互いに「接続」又は「結合」されると考えられる。 (11) In the above-described embodiments, the terms "connected", "coupled", or any variation thereof, mean any direct or indirect connection or coupling between two or more elements, including the presence of one or more intermediate elements between two elements that are "connected" or "coupled" to each other. Couplings or connections between elements may be physical couplings or connections, logical couplings or connections, or a combination thereof. For example, "connection" may be replaced with "access." As used in this disclosure, two elements are considered to be “connected” or “coupled” to each other using at least one of one or more wires, cables, and printed electrical connections, and using electromagnetic energy having wavelengths in the radio frequency, microwave, and light (both visible and invisible) regions, as some non-limiting and non-exhaustive examples.
(12)上述した実施形態において、「に基づいて」という記載は、別段に明記されていない限り、「のみに基づいて」を意味しない。言い換えれば、「に基づいて」という記載は、「のみに基づいて」と「に少なくとも基づいて」の両方を意味する。 (12) In the above-described embodiments, the phrase "based on" does not mean "based only on," unless expressly specified otherwise. In other words, the phrase "based on" means both "based only on" and "based at least on."
(13)本開示において使用される「判断(determining)」、「決定(determining)」という用語は、多種多様な動作を包含する場合がある。「判断」、「決定」は、例えば、判定(judging)、計算(calculating)、算出(computing)、処理(processing)、導出(deriving)、調査(investigating)、探索(looking up、search、inquiry)(例えば、テーブル、データベース又は別のデータ構造での探索)、確認(ascertaining)した事を「判断」「決定」したとみなす事等を含み得る。また、「判断」、「決定」は、受信(receiving)(例えば、情報を受信すること)、送信(transmitting)(例えば、情報を送信すること)、入力(input)、出力(output)、アクセス(accessing)(例えば、メモリ中のデータにアクセスすること)した事を「判断」「決定」したとみなす事等を含み得る。また、「判断」、「決定」は、解決(resolving)、選択(selecting)、選定(choosing)、確立(establishing)、比較(comparing)等した事を「判断」「決定」したとみなす事を含み得る。つまり、「判断」「決定」は、何らかの動作を「判断」「決定」したとみなす事を含み得る。また、「判断(決定)」は、「想定する(assuming)」、「期待する(expecting)」、「みなす(considering)」等によって読み替えられてもよい。 (13) The terms "determining" and "determining" as used in this disclosure may encompass a wide variety of actions. "Judgement", "determining" can include, for example, judging, calculating, computing, processing, deriving, investigating, looking up, searching, inquiring (e.g., searching in a table, database, or other data structure), ascertaining as "judging", "determining", etc. In addition, "determining" and "determining" include receiving (e.g., receiving information), transmitting (e.g., transmitting information), input, output, and accessing (e.g., accessing data in memory). Also, "determining" or "determining" may include resolving, selecting, choosing, establishing, comparing, etc., to be regarded as "determining" or "determining." In other words, "judgment" and "decision" may include considering that some action is "judgment" and "decision". Also, "judgment (decision)" may be replaced by "assuming", "expecting", "considering", and the like.
(14)上述した実施形態において、「含む(include)」、「含んでいる(including)」及びそれらの変形が使用されている場合、これらの用語は、用語「備える(comprising)」と同様に、包括的であることが意図される。更に、本開示において使用されている用語「又は(or)」は、排他的論理和ではないことが意図される。 (14) Where “include,” “including,” and variations thereof are used in the above-described embodiments, these terms, like the term “comprising,” are intended to be inclusive. Furthermore, the term "or" as used in this disclosure is not intended to be an exclusive OR.
(15)本開示において、例えば、英語でのa, an及びtheのように、翻訳により冠詞が追加された場合、本開示は、これらの冠詞の後に続く名詞が複数形であることを含んでもよい。 (15) In this disclosure, where articles have been added by translation, such as a, an, and the in English, the disclosure may include the nouns following these articles being plural.
(16)本開示において、「AとBが異なる」という用語は、「AとBが互いに異なる」ことを意味してもよい。なお、当該用語は、「AとBがそれぞれCと異なる」ことを意味してもよい。「離れる」、「結合される」等の用語も、「異なる」と同様に解釈されてもよい。 (16) In the present disclosure, the term "A and B are different" may mean "A and B are different from each other." The term may also mean that "A and B are different from C". Terms such as "separate," "coupled," etc. may also be interpreted in the same manner as "different."
(17)本開示において説明した各態様/実施形態は単独で用いてもよいし、組み合わせて用いてもよいし、実行に伴って切り替えて用いてもよい。また、所定の情報の通知(例えば、「Xであること」の通知)は、明示的に行う通知に限られず、暗黙的(例えば、当該所定の情報の通知を行わない)ことによって行われてもよい。 (17) Each aspect/embodiment described in the present disclosure may be used alone, may be used in combination, or may be used by switching according to execution. Further, notification of predetermined information (e.g., notification of “being X”) is not limited to explicit notification, and may be performed implicitly (e.g., not notifying the predetermined information).
 以上、本開示について詳細に説明したが、当業者にとっては、本開示が本開示中に説明した実施形態に限定されないということは明らかである。本開示は、請求の範囲の記載により定まる本開示の趣旨及び範囲を逸脱することなく修正及び変更態様として実施できる。従って、本開示の記載は、例示説明を目的とし、本開示に対して何ら制限的な意味を有さない。 Although the present disclosure has been described in detail above, it is clear to those skilled in the art that the present disclosure is not limited to the embodiments described in this disclosure. The present disclosure can be practiced with modifications and variations without departing from the spirit and scope of the present disclosure as defined by the claims. Accordingly, the description of the present disclosure is for the purpose of illustration and description and is not meant to be limiting in any way on the present disclosure.
1~1C…情報処理システム、10~10B…サーバ、11~11B…処理装置、12~12B…記憶装置、13…通信装置、14…ディスプレイ、15…入力装置、20…端末装置、21…処理装置、22…記憶装置、23…通信装置、24…ディスプレイ、25…入力装置、26…撮像装置、30…XRグラス、32…記憶装置、38…ディスプレイ、111…取得部、111A…入力画像取得部、111B…顔画像取得部、111C…頭身取得部、111D…取得部、111E…顔画像取得部、111F…取得部、111G…頭身取得部、112…顔画像生成部、113、113A…頭部生成部、114…体部生成部、115…アバター生成部、116…出力部、117…年齢推定部、118…頭身数生成部、211…取得部、212…出力部、213…表示制御部、A1~A2…アバター、PR1~PR3B…制御プログラム、U1~U2…ユーザ Reference Signs List 1 to 1C information processing system 10 to 10B server 11 to 11B processing device 12 to 12B storage device 13 communication device 14 display 15 input device 20 terminal device 21 processing device 22 storage device 23 communication device 24 display 25 input device 26 imaging device 30 XR glass 32 storage device 38 display 111 acquisition unit 1 11A... input image acquisition unit 111B... face image acquisition unit 111C... head/body acquisition unit 111D... acquisition unit 111E... face image acquisition unit 111F... acquisition unit 111G... head/body acquisition unit 112... face image generation unit 113, 113A... head generation unit 114... body generation unit 115... avatar generation unit 116... output unit 117... age estimation unit 118... Head and body number generation unit 211 Acquisition unit 212 Output unit 213 Display control unit A1 to A2 Avatar PR1 to PR3B Control program U1 to U2 User

Claims (4)

  1.  画風が互いに異なる複数の顔画像を生成する顔画像生成部と、
     前記複数の顔画像から、ユーザによって選択された第1の顔画像を取得する顔画像取得部と、
     前記第1の顔画像に基づいて、アバターの頭部を示す三次元画像を生成する頭部生成部と、
     前記アバターが何頭身であるかを表す頭身数を取得する頭身取得部と、
     前記頭身数に基づいて、前記アバターの前記頭部以外の部分である体部を示す三次元画像を生成する体部生成部と、
     前記アバターの前記頭部を示す前記三次元画像、及び前記アバターの前記体部を示す前記三次元画像を用いて、前記アバターの全体の外観を示す三次元画像を生成するアバター生成部と、
    を備える情報処理装置。
    a face image generation unit that generates a plurality of face images with different styles;
    a facial image acquisition unit that acquires a first facial image selected by a user from the plurality of facial images;
    a head generation unit that generates a three-dimensional image showing the head of the avatar based on the first face image;
    a head and body acquisition unit that acquires a head and body number indicating how many heads and bodies the avatar has;
    a body part generation unit that generates a three-dimensional image showing a body part other than the head part of the avatar based on the number of head and body parts;
    an avatar generation unit that generates a three-dimensional image showing the overall appearance of the avatar using the three-dimensional image showing the head of the avatar and the three-dimensional image showing the body of the avatar;
    Information processing device.
  2.  前記顔画像生成部は、前記ユーザの顔写真を示す画像に基づいて、画風が互いに異なる複数の顔画像を生成する、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the facial image generation unit generates a plurality of facial images having different styles based on an image representing the facial photograph of the user.
  3.  前記顔画像取得部は、前記選択された第1の顔画像から、前記ユーザの顔の要素を示す要素画像を抽出し、
     前記頭部生成部は、抽出された前記要素画像を、予め用意された外形画像に重畳することにより、前記アバターの頭部を示す三次元画像を生成する、請求項1に記載の情報処理装置。
    The face image acquisition unit extracts an element image representing an element of the user's face from the selected first face image,
    2. The information processing apparatus according to claim 1, wherein the head generation unit generates a three-dimensional image representing the head of the avatar by superimposing the extracted element image on a previously prepared outline image.
  4.  前記ユーザの顔写真を示す画像に基づいて、前記ユーザの年齢を推定する年齢推定部と、
     推定された前記年齢に基づいて、前記アバターの頭身数を生成する頭身生成部と、
    を更に備える、請求項1から請求項3のいずれか1項に記載の情報処理装置。
    an age estimating unit for estimating the age of the user based on an image showing the face photograph of the user;
    a head-to-body generation unit that generates the head-to-body number of the avatar based on the estimated age;
    The information processing apparatus according to any one of claims 1 to 3, further comprising:
PCT/JP2022/045381 2022-01-21 2022-12-08 Information processing device WO2023139961A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2022-007762 2022-01-21
JP2022007762 2022-01-21

Publications (1)

Publication Number Publication Date
WO2023139961A1 true WO2023139961A1 (en) 2023-07-27

Family

ID=87348078

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2022/045381 WO2023139961A1 (en) 2022-01-21 2022-12-08 Information processing device

Country Status (1)

Country Link
WO (1) WO2023139961A1 (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2010090259A1 (en) * 2009-02-05 2010-08-12 Sairyo Kazuo Portrait illustration creation system, character creation system, and created portrait illustration display system
JP2010267241A (en) * 2008-10-17 2010-11-25 Square Enix Co Ltd Three-dimensional model display system
JP2014153747A (en) * 2013-02-05 2014-08-25 Kddi Corp Program, information equipment and method for controlling character display on basis of image complexity
JP2019145108A (en) * 2018-02-23 2019-08-29 三星電子株式会社Samsung Electronics Co.,Ltd. Electronic device for generating image including 3d avatar with facial movements reflected thereon, using 3d avatar for face
CN113204282A (en) * 2021-04-12 2021-08-03 领悦数字信息技术有限公司 Interactive apparatus, interactive method, computer-readable storage medium, and computer program product

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010267241A (en) * 2008-10-17 2010-11-25 Square Enix Co Ltd Three-dimensional model display system
WO2010090259A1 (en) * 2009-02-05 2010-08-12 Sairyo Kazuo Portrait illustration creation system, character creation system, and created portrait illustration display system
JP2014153747A (en) * 2013-02-05 2014-08-25 Kddi Corp Program, information equipment and method for controlling character display on basis of image complexity
JP2019145108A (en) * 2018-02-23 2019-08-29 三星電子株式会社Samsung Electronics Co.,Ltd. Electronic device for generating image including 3d avatar with facial movements reflected thereon, using 3d avatar for face
CN113204282A (en) * 2021-04-12 2021-08-03 领悦数字信息技术有限公司 Interactive apparatus, interactive method, computer-readable storage medium, and computer program product

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
ANONYMOUS: "Create Western-style portraits from photos! AI Artist", MR. PC, vol. 13, no. 8, 24 June 2022 (2022-06-24), pages 73, XP009548011 *
WATANABE SHINYA, TOZAWA, MASAHIRO, NAKANO, KEISUKE, ONO, SATOSHI.: "A Proposal for an Avatar Creation Support System Based on Cooperative Evolution by User and System", IPSJ JOURNAL, vol. 7, no. 1, 28 April 2014 (2014-04-28), pages 22 - 33, XP093079563, ISSN: 1882-7780 *

Similar Documents

Publication Publication Date Title
US11887231B2 (en) Avatar animation system
US9058056B2 (en) System and method of dynamically generating a frequency pattern to realize the sense of touch in a computing device
WO2013169237A1 (en) Eye tracking based selective accentuation of portions of a display
US11380134B2 (en) Method and device for determining parameter for gaze tracking device
CN107210830B (en) Object presenting and recommending method and device based on biological characteristics
CN113407850B (en) Method and device for determining and acquiring virtual image and electronic equipment
CN110288705A (en) The method and apparatus for generating threedimensional model
JP2022544240A (en) Systems and methods for virtual and augmented reality
JP7110738B2 (en) Information processing device, program and information processing system
WO2023139961A1 (en) Information processing device
CN111580679A (en) Space capsule display method and device, electronic equipment and storage medium
CN110832525A (en) Augmented reality advertising on objects
CN107544660B (en) Information processing method and electronic equipment
CN112464009A (en) Method and device for generating pairing image, electronic equipment and storage medium
CN111274489A (en) Information processing method, device, equipment and storage medium
JP2018088604A (en) Image display device, image display method, and system
CN116069159A (en) Method, apparatus and medium for displaying avatar
WO2023162499A1 (en) Display control device
JP2003077001A (en) Face image communication device and program
WO2023120472A1 (en) Avatar generation system
WO2023145892A1 (en) Display control device, and server
JP7398854B1 (en) Web page viewing analysis system, web page viewing analysis method, and web page viewing analysis program
WO2023145890A1 (en) Terminal device
WO2023176317A1 (en) Display control device
Guo et al. Synchronous mixed reality (SMR): A personalized virtual‐real fusion framework with high immersion and effective interaction

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22922112

Country of ref document: EP

Kind code of ref document: A1