WO2023189601A1 - Information processing device, recording medium, and information processing method - Google Patents

Information processing device, recording medium, and information processing method Download PDF

Info

Publication number
WO2023189601A1
WO2023189601A1 PCT/JP2023/010101 JP2023010101W WO2023189601A1 WO 2023189601 A1 WO2023189601 A1 WO 2023189601A1 JP 2023010101 W JP2023010101 W JP 2023010101W WO 2023189601 A1 WO2023189601 A1 WO 2023189601A1
Authority
WO
WIPO (PCT)
Prior art keywords
parts
character
information processing
processing device
model
Prior art date
Application number
PCT/JP2023/010101
Other languages
French (fr)
Japanese (ja)
Inventor
康次 佐藤
Original Assignee
ソニーグループ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニーグループ株式会社 filed Critical ソニーグループ株式会社
Publication of WO2023189601A1 publication Critical patent/WO2023189601A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/20Finite element generation, e.g. wire-frame surface description, tesselation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics

Definitions

  • the present disclosure relates to an information processing device, a recording medium, and an information processing method.
  • a method is known in which a 3D model is created by combining multiple parts.
  • a system for managing parts of a 3D model for example, a system is known that extracts data on parts and searches for parts using the data.
  • the system described above extracts part data from a 3D model that has already been divided into parts. In this way, the above-described system assumes that parts have already been separated from the 3D model, and does not consider how to separate the parts from the 3D model.
  • 3D data such as 3D models requires complex processing and a large amount of calculation, making it highly difficult. Analysis of 3D data is used to separate the 3D model into parts. For this reason, it has been difficult to easily (for example, in a short time or with high precision) perform a process of separating a 3D model into parts. Therefore, it is desired that 3D models can be analyzed more easily.
  • the present disclosure provides a mechanism that allows 3D models to be analyzed more easily.
  • the information processing device of the present disclosure includes a control unit.
  • the control unit obtains a three-dimensional model of the character.
  • the control unit performs image recognition processing on the image seen from the virtual viewpoint of the three-dimensional model to estimate the positions of the parts of the character.
  • the control unit estimates a part area in the three-dimensional model of the character based on the three-dimensional model and the part position.
  • FIG. 2 is a diagram for explaining an overview of a 3D model analysis process according to an embodiment of the present disclosure.
  • FIG. 1 is a block diagram illustrating a configuration example of an information processing device according to an embodiment of the present disclosure.
  • FIG. 3 is a diagram for explaining an example of rendering processing by a rendering unit according to an embodiment of the present disclosure. It is a figure showing an example of part position estimation by a position estimating part concerning an embodiment of this indication.
  • FIG. 7 is a diagram illustrating another example of part position estimation by the position estimation unit according to the embodiment of the present disclosure.
  • FIG. 3 is a diagram for explaining an example of a parts region estimation process by a region estimating unit according to an embodiment of the present disclosure.
  • FIG. 1 is a block diagram illustrating a configuration example of an information processing device according to an embodiment of the present disclosure.
  • FIG. 3 is a diagram for explaining an example of rendering processing by a rendering unit according to an embodiment of the present disclosure. It is a figure showing an example of part position estimation by
  • FIG. 3 is a diagram for explaining an example of a correction process performed by a region estimation unit according to an embodiment of the present disclosure.
  • FIG. 7 is a diagram for explaining an example of rattling detection performed by a region estimation unit according to an embodiment of the present disclosure.
  • FIG. 2 is a diagram for explaining an example of a search process according to an embodiment of the present disclosure.
  • FIG. 2 is a diagram illustrating an example of a search image according to an embodiment of the present disclosure.
  • FIG. 3 is a diagram for explaining an example of extraction of feature amount information by a search processing unit according to an embodiment of the present disclosure.
  • FIG. 2 is a diagram for explaining an example of a search in a latent space by a search processing unit according to an embodiment of the present disclosure.
  • FIG. 3 is a diagram illustrating an example of a search result image according to an embodiment of the present disclosure.
  • FIG. 3 is a diagram for explaining an example of a search range changing process in a search processing unit according to an embodiment of the present disclosure.
  • FIG. 7 is a diagram illustrating another example of a search result image according to an embodiment of the present disclosure. It is a flow chart which shows an example of a flow of the 1st parts separation processing concerning an embodiment of this indication. It is a flow chart which shows an example of the flow of the 2nd parts separation processing concerning an embodiment of this indication.
  • FIG. 3 is a diagram for explaining an example of correction of recognition results according to an embodiment of the present disclosure.
  • FIG. 2 is a diagram for explaining an example of a UI image showing an estimation result according to an embodiment of the present disclosure.
  • FIG. 7 is a diagram for explaining another example of a UI image showing an estimation result according to an embodiment of the present disclosure.
  • FIG. 7 is a diagram for explaining another example of a UI image showing an estimation result according to an embodiment of the present disclosure. It is a flowchart which shows an example of the flow of the 3rd parts separation process concerning an embodiment of this indication.
  • 1 is a block diagram illustrating an example of a hardware configuration of an information processing device according to an embodiment.
  • One or more embodiments (including examples and modifications) described below can each be implemented independently. On the other hand, at least a portion of the plurality of embodiments described below may be implemented in combination with at least a portion of other embodiments as appropriate. These multiple embodiments may include novel features that are different from each other. Therefore, these multiple embodiments may contribute to solving mutually different objectives or problems, and may produce mutually different effects.
  • One method to simplify the production of 3D models is, for example, to store 3D models for each part in a database (DB), combine them to create the basic body of the 3D model, and finally perform finishing processing.
  • DB database
  • a system can be considered that creates a 3D model of a character by combining parts such as the character's body parts such as the head, eyes, nose, ears, mouth, and body, as well as the character's costumes and accessories such as hats, glasses, and clothes. .
  • 3D models of characters are complex and do not have unified specifications, but for example, the system analyzes the shape of the 3D model and classifies each part of the character, allowing the user to combine each part to create the desired 3D model. You can easily create a character's body.
  • the system can analyze metadata for each part and store the parts and metadata in association with each other.
  • the system can create a database that stores metadata obtained from character features and parts features in association with the parts, allowing the user to more easily find the desired part by searching this database. be able to obtain it.
  • the system analyzes the 3D model of the character, separates the parts, and extracts metadata, allowing the user to more easily create the 3D model of the character.
  • the 3D model when a 3D model is expressed as mesh data, the 3D model is expressed by a plurality of vertices, edges connecting the vertices, and a surface formed by the vertices and edges.
  • the system separates a nose as a part from a 3D model of a character's face. In this case, it is not easy for the system to determine which vertex data of the face corresponds to the nose.
  • 3D models have a higher degree of freedom than 2D images, and there are no restrictions such as resolution. Therefore, the higher the quality of the 3D model, the larger the amount of data (for example, the number of vertices). Therefore, the analysis process of the 3D model becomes more complicated and the calculation load increases.
  • the system analyzes the 3D model in order to separate specific parts from the character's 3D model information (for example, the mesh data mentioned above) and extract metadata that is characteristic of this part. It wasn't easy to do.
  • the information processing device performs image recognition processing using a rendered image of a 3D model of a character, and uses the result of the image recognition processing to narrow down the 3D models to be analyzed.
  • the information processing device performs analysis on the narrowed down 3D models. Thereby, the information processing device can more easily analyze the 3D model and can more easily separate parts from the 3D model.
  • FIG. 1 is a diagram for explaining an overview of a 3D model analysis process according to an embodiment of the present disclosure.
  • the analysis process in FIG. 1 is executed by the information processing device 100, for example.
  • the information processing device 100 first obtains 3D model information (hereinafter also referred to as 3D model) of a character (step S1). For example, the information processing device 100 acquires a 3D model of a character from a database.
  • the 3D model includes, for example, the mesh data described above.
  • the information processing device 100 performs rendering (drawing) of the character based on the acquired 3D model, and generates an image of the character viewed from a virtual viewpoint (step S2).
  • the information processing device 100 performs part image recognition processing on the generated image (step S3). Thereby, the information processing device 100 estimates the position of the part in the image. Note that the position of a part estimated by the information processing apparatus 100 based on image recognition processing is also referred to as an image recognition position. For example, in FIG. 1, the information processing device 100 performs right eye image recognition processing on the image, and estimates an area including the right eye as the image recognition position.
  • the information processing device 100 estimates the image recognition position in the 3D model based on the image recognition process (step S4). For example, the information processing device 100 estimates the position of the 3D model corresponding to the image recognition position in the image as the image recognition position in the 3D model.
  • the information processing device 100 performs part analysis of the 3D model based on the image recognition position in the 3D model, and estimates the region of the part in the 3D model (step S5).
  • the information processing device 100 estimates data corresponding to a part from among the mesh data of the 3D model as a 3D model of the parts area. For example, in FIG. 1, the information processing apparatus 100 estimates the vertex data group corresponding to the right eye as a 3D model of the right eye part.
  • the information processing device 100 extracts metadata of the parts (step S6). For example, the information processing device 100 extracts metadata based on a 3D model of a character, an image, and a 3D model of parts.
  • the information processing device 100 associates and stores parts and metadata (step S7).
  • the information processing device 100 associates the 3D model of the part whose region was estimated in step S5 with the metadata of the part extracted in step S6, and stores the 3D model in the database.
  • the information processing apparatus 100 stores a 3D model of a part in a parts DB (Data Base), and stores metadata of the part in the metadata DB.
  • the information processing device 100 acquires a 3D model of a character (an example of 3D model information).
  • the information processing device 100 performs image recognition processing on an image drawn based on a 3D model, in which the character is viewed from a virtual viewpoint, to estimate the positions of the character's parts.
  • the information processing device 100 estimates the part area in the 3D model of the character based on the 3D model and the part position.
  • the information processing apparatus 100 can narrow down the 3D models to be analyzed from among the 3D models, and can more easily analyze the 3D models. Therefore, the information processing apparatus 100 can further reduce the processing load of separating the 3D model of the character into parts. Furthermore, the information processing apparatus 100 can separate parts of a 3D model of a character with higher precision.
  • the information processing device 100 can efficiently create a character parts DB and a metadata DB used for searching parts by analyzing the 3D shape of the character through analysis processing. By using the information processing device 100, a user can more efficiently create a 3D model body of a character.
  • FIG. 2 is a block diagram illustrating a configuration example of the information processing device 100 according to the embodiment of the present disclosure.
  • the information processing apparatus 100 according to the embodiment of the present disclosure narrows the search space on the 3D model by using image processing on rendered images to analyze the shape of a 3D character that has a high degree of freedom and is difficult to process. Perform feature analysis. Thereby, the information processing apparatus 100 can perform part area estimation and metadata extraction of a 3D model with less processing load through simpler processing.
  • the information processing device 100 shown in FIG. 2 includes a communication section 110, an input/output section 120, a storage section 130, and a control section 140.
  • the information processing device 100 may be a terminal device used by a user, such as a personal computer or a tablet terminal, or may be a server device placed on a network (for example, a cloud server device or a local server device). good.
  • the information processing device 100 includes both a control unit 140 that executes applications such as the analysis processing described above, and a storage unit 130 that includes a parts DB 133 and a metadata DB 134 and functions as storage.
  • a storage unit 130 that includes a parts DB 133 and a metadata DB 134 and functions as storage.
  • some functions, such as the storage function of the storage unit 130 may be realized by an information processing device (for example, a server device) different from the information processing device 100 in FIG. 2.
  • the information processing device 100 in FIG. 2 has both an acquisition function that analyzes a 3D model of a character and acquires parts, and a search function that searches for parts.
  • the search function may be realized by an information processing device different from the information processing device 100 having the acquisition function.
  • Communication unit 110 is a communication interface for communicating with other devices.
  • the communication unit 110 is a LAN (Local Area Network) interface such as a NIC (Network Interface Card).
  • Communication unit 110 may be a wired interface or a wireless interface.
  • the communication unit 110 communicates with other devices under the control of the control unit 140.
  • the input/output unit 120 is a user interface for exchanging information with the user.
  • the input/output unit 120 is an operating device, such as a keyboard, a mouse, an operation key, a touch panel, etc., for the user to perform various operations.
  • the input/output unit 120 is a display device such as a liquid crystal display (Liquid Crystal Display) or an organic EL display (Organic Electroluminescence Display).
  • the input/output unit 120 may be an audio device such as a speaker or a buzzer.
  • the input/output unit 120 may be a lighting device such as an LED (Light Emitting Diode) lamp.
  • the storage unit 130 is realized by, for example, a semiconductor memory element such as a RAM (Random Access Memory), a ROM (Read Only Memory), or a flash memory, or a storage device such as a hard disk or an optical disk.
  • a semiconductor memory element such as a RAM (Random Access Memory), a ROM (Read Only Memory), or a flash memory
  • a storage device such as a hard disk or an optical disk.
  • the storage unit 130 in FIG. 2 includes a 3D model DB 131, a log file DB 132, a parts DB 133, and a metadata DB 134.
  • the 3D model DB 131 is a database that stores 3D models of characters on which the information processing device 100 performs 3D shape analysis.
  • the log file DB 132 is a database that stores log files that hold analysis results of 3D shape analysis performed by the information processing device 100.
  • the parts DB 133 is a database that stores 3D models of character parts regions obtained by the information processing device 100 performing 3D shape analysis.
  • the metadata DB 134 is a database that stores metadata corresponding to parts areas.
  • the storage unit 130 stores the 3D model and metadata of the parts area in association with each other.
  • the control unit 140 is a controller that controls each unit of the information processing device 100.
  • the control unit 140 is realized by, for example, a processor such as a CPU (Central Processing Unit), an MPU (Micro Processing Unit), or a GPU (Graphics Processing Unit).
  • the control unit 140 is realized by a processor executing various programs stored in a storage device inside the information processing device 100 using a RAM or the like as a work area.
  • the control unit 140 may be realized by an integrated circuit such as an ASIC (Application Specific Integrated Circuit) or an FPGA (Field Programmable Gate Array).
  • ASIC Application Specific Integrated Circuit
  • FPGA Field Programmable Gate Array
  • the control unit 140 includes a model acquisition unit 141, a rendering unit 142, an image recognition unit 143, a position estimation unit 144, a region estimation unit 145, an extraction unit 146, a search processing unit 147, and a UI control unit 148. , is provided.
  • the control unit 140 realizes an acquisition function (application function) in which the model acquisition unit 141 to the extraction unit 146 analyze the 3D model of the character described above and acquire parts. Further, the control unit 140 realizes a search function (application function) for searching for parts using the search processing unit 147.
  • Each block (model acquisition unit 141 to UI control unit 148) constituting the control unit 140 is a functional block indicating a function of the control unit 140.
  • These functional blocks may be software blocks or hardware blocks.
  • each of the above functional blocks may be one software module realized by software (including a microprogram), or one circuit block on a semiconductor chip (die).
  • each functional block may be one processor or one integrated circuit.
  • the control unit 140 may be configured in functional units different from the above-mentioned functional blocks.
  • the functional blocks can be configured in any way.
  • control unit 140 may be configured in a functional unit different from the above-mentioned functional blocks. Further, some or all of the blocks (model acquisition unit 141 to UI control unit 148) constituting the control unit 140 may be operated by another device.
  • Model acquisition unit 141 The model acquisition unit 141 acquires a 3D model of the character by reading the 3D model from the 3D model DB 131. Note that the model acquisition unit 141 can acquire a 3D model of a character from another device via the communication unit 110. The model acquisition unit 141 outputs the acquired 3D model to the rendering unit 142, the position estimation unit 144, and the area estimation unit 145.
  • FIG. 3 is a diagram for explaining an example of rendering processing by the rendering unit 142 according to the embodiment of the present disclosure.
  • the rendering unit 142 generates a character image (2D image) based on the 3D model by executing rendering processing.
  • the rendering unit 142 generates an image of the character viewed from the virtual viewpoint of the virtual camera C. At this time, the rendering unit 142 can generate a plurality of images with different virtual viewpoints.
  • the rendering unit 142 generates images P_1 to P_N in which the character is viewed from each of a plurality of virtual cameras C_1 to C_N arranged around the character at a predetermined angle (for example, 30 degrees or 45 degrees).
  • Images P_1 to P_N include characters with different orientations.
  • the rendering unit 142 outputs the generated images P_1 to P_N to the image recognition unit 143. At this time, the rendering unit 142 may output information regarding the virtual cameras C_1 to C_N corresponding to the images P_1 to P_N to the image recognition unit 143.
  • Image recognition unit 143 performs image recognition processing on images P_1 to P_N to estimate parts positions (image recognition positions) included in images P_1 to P_N.
  • the image recognition unit 143 estimates the position of a pre-designated part in images P_1 to P_N.
  • the part (parts region) estimated by the image recognition unit 143 includes, for example, a region including a body part of the character.
  • the part areas include, for example, the character's eye area, nose area, mouth area, and ear area.
  • the parts area includes an area including body parts such as fingers, palms, and feet, an area including accessories such as glasses, watches, earrings, and necklaces, and costumes such as clothes and hats.
  • the parts estimated by the information processing device 100 are not limited to the examples described here.
  • the information processing device 100 can estimate any part, such as a part specified by a user or a part specialized for a character.
  • the image recognition unit 143 performs image recognition processing on the image P using pattern recognition technology or semantic segmentation technology, and determines the pixel coordinates of parts within the image P.
  • the image recognition unit 143 estimates, for example, a rectangular or polygonal area within the image P as the image recognition position of the part.
  • the image recognition unit 143 outputs information regarding the image recognition position in the image P to the position estimation unit 144.
  • the image recognition unit 143 also outputs the result of the image recognition process to the extraction unit 146. At this time, the image recognition unit 143 can output information regarding the virtual camera C corresponding to the image P to the position estimation unit 144.
  • the image recognition unit 143 may associate the parts that could not be estimated with the image P and write them in a log file, for example.
  • the position estimation unit 144 estimates the part position (image recognition position) in the 3D model based on the image recognition result by the image recognition unit 143.
  • the image recognition position in the 3D model estimated by the position estimation unit 144 is a position (area) more roughly than the part area actually estimated by the information processing device 100.
  • FIG. 4 is a diagram illustrating an example of part position estimation by the position estimation unit 144 according to the embodiment of the present disclosure.
  • the position estimating unit 144 estimates the approximate position of the part in the 3D model by calculating backwards from the settings of the virtual camera C used to render the 3D model.
  • the image recognition unit 143 estimates the image recognition position Rp in the image P_1 using the right eye as a part.
  • the image P_1 is an image obtained by rendering a 3D model viewed from the virtual camera C_1.
  • the position estimating unit 144 calculates the image recognition position in the 3D model, as shown in the middle diagram of FIG. Estimate Rm. For example, the position estimation unit 144 estimates the image recognition position Rm in the 3D model by projecting the image recognition position Rp in the image P_1 into the 3D space of the 3D model.
  • the lower diagram in FIG. 4 is an enlarged diagram of the image recognition position Rm in the 3D model.
  • the image recognition position Rm estimated by the position estimating unit 144 represents the approximate position (area) of a part (here, the "eye") in the 3D model. Therefore, the image recognition position Rm does not necessarily match the mesh of the 3D model. That is, the contour line of the image recognition position Rm in the 3D model does not necessarily match the edge of the mesh in the 3D model.
  • the position estimation unit 144 estimates a rectangular area as the image recognition position Rm in the 3D model. In this way, the position estimating unit 144 roughly estimates the position of the parts in the 3D model. Therefore, the estimated position (area) may differ from the actual part area (for example, a surface of a 3D model).
  • FIG. 5 is a diagram illustrating another example of part position estimation by the position estimation unit 144 according to the embodiment of the present disclosure.
  • the image recognition unit 143 can estimate the position of parts in the image P using semantic segmentation technology in addition to pattern-based image recognition.
  • FIG. 5 shows part position estimation by the position estimation unit 144 when the image recognition unit 143 estimates the part position by semantic segmentation.
  • the image recognition unit 143 performs image recognition using semantic segmentation on the image P shown in the upper diagram of FIG. Estimate. By using semantic segmentation, the image recognition unit 143 can estimate a more detailed image recognition position Rp1 than the image recognition process shown in FIG. 4 .
  • the position estimation unit 144 calculates the image recognition position Rm1 in the 3D model based on the arrangement of the image recognition position Rp1 in the image P and the virtual viewpoint and angle of view of the virtual camera C in the 3D space, as shown in the lower diagram of FIG. Estimate.
  • the method of estimating the image recognition position Rm1 in the 3D model by the position estimation unit 144 is the same as in the case of FIG. 4, but the image recognition unit 143 estimates the image recognition position Rp1 more finely than the image recognition position Rp of FIG. 4. . Therefore, the image recognition position Rm1 in the 3D model estimated by the position estimation unit 144 is finer than the image recognition position Rm in FIG. 4.
  • the estimation of the image recognition position Rm1 (part position) in the 3D model by the position estimation unit 144 is to estimate a rough position (area). Therefore, the image recognition position Rm1 and the mesh of the 3D model do not necessarily match. Further, depending on the estimation result of the image recognition position Rp1 by the image recognition unit 143, there is a possibility that the contour of the image recognition position Rm1 in the 3D model may become uneven (jagged).
  • the position estimating unit 144 outputs information regarding the image recognition position (part position) in the estimated 3D model to the area estimating unit 145.
  • the position estimation unit 144 writes a notification to that effect in a log file, for example. For example, based on the recognition result of the image recognition unit 143, the position estimation unit 144 determines that estimation of the image recognition position has failed when the image recognition position estimated in the 3D space is not on the 3D model. If estimation of the image recognition position fails in this way, the position estimating unit 144 may associate the 3D model of the character with the parts and write them in a log file, for example.
  • the area estimation unit 145 estimates the part area in the 3D model of the character based on the 3D model and the part position Rm (image recognition position Rm) estimated by the position estimation unit 144.
  • the region estimating unit 145 estimates and extracts a 3D model (for example, mesh data) of a character's parts as a parts region. Thereby, the area estimation unit 145 separates the parts from the 3D model of the character.
  • the part position estimated by the position estimation unit 144 is the rough position (area) of the part in the 3D model of the character. Therefore, this part position may not match the area defined by the mesh of the actual 3D model, or may be a jagged area with unevenness. In this way, the parts positions estimated by the position estimation unit 144 cannot be said to have sufficient accuracy to separate the parts from the 3D model.
  • the region estimating section 145 performs analysis of the 3D model (3D analysis) focusing on the parts position Rm estimated by the position estimating section 144, and estimates the parts region in the 3D model of the character. do.
  • FIG. 6 is a diagram for explaining an example of part region estimation processing by the region estimating unit 145 according to the embodiment of the present disclosure.
  • the area estimating unit 145 analyzes the 3D model within the image recognition position Rm estimated by the position estimating unit 144, and estimates a part area Rr along the mesh of the 3D model.
  • the region estimating unit 145 can estimate the part region with higher accuracy by, for example, performing an analysis based on the characteristics of the 3D shape according to the part to be extracted (for example, eyes, etc.). can.
  • the region estimating unit 145 performs analysis by taking into account the characteristics of the 3D structure and the characteristics of the part attributes as the characteristics of each part.
  • Features of the 3D structure include, for example, curvature, gradient, and Lavrasian in the 3D model of the part.
  • Features of part attributes include volume ratio (for example, the ratio of the volume of the part to the volume of the entire character), and aspect ratio (for example, the aspect ratio of the part).
  • the region estimating unit 145 may perform analysis not on the entire 3D model of the character, but on the parts positions estimated by the position estimating unit 144. Thereby, the region estimating unit 145 can further reduce the processing load of analysis.
  • the region estimating unit 145 may separate the parts region Rr estimated by 3D analysis as a 3D model of the character's parts, but may also correct the parts region Rr as a corrected parts region Rc.
  • FIG. 7 is a diagram for explaining an example of a correction process performed by the region estimation unit 145 according to the embodiment of the present disclosure.
  • FIG. 8 is a diagram for explaining an example of rattling detection performed by the area estimating unit 145 according to an embodiment of the present disclosure.
  • the left diagram in FIG. 7 shows the parts region Rr estimated by the region estimation unit 145.
  • the outline of the part region Rr estimated by the region estimation unit 145 may be jittery due to the mesh structure of the 3D model.
  • the region estimating unit 145 detects wobbling in the part region Rr.
  • the region estimating unit 145 performs wobbling detection using the normal information of the part region Rr. For example, the region estimation unit 145 detects the normal vector (arrow in FIG. 8) of the outline of the part region Rr.
  • the region estimation unit 145 checks the direction of the normal vector of the contour line along the contour of the part region Rr.
  • the region estimating unit 145 detects a location where the direction of the normal vector is substantially the same as a location where the contour is jittery (wobbly location).
  • the region estimating unit 145 generates a corrected part region Rc by correcting the detected wobbling portion. For example, as shown in FIG. 8, the region estimation unit 145 corrects the wobbling of the region A and generates a corrected parts region Rc. For example, the region estimating unit 145 generates a corrected part region Rc without wobbling in the outline by creating a new edge at the wobbling location.
  • the region estimation unit 145 corrects the shape of the parts region Rr according to the wobbling of the outline of the parts region Rr.
  • the region estimation unit 145 separates the 3D model (3D model information, for example, mesh data) of the corrected parts region Rc generated by the correction as a 3D model of the character's parts.
  • the region estimation unit 145 separates the character into parts by, for example, generating a 3D model of the corrected parts region Rc from the 3D model of the character.
  • the area estimation unit 145 outputs information regarding the 3D model of the separated part to the extraction unit 146.
  • the region estimating unit 145 writes a notification to that effect in a log file, for example. For example, if the region estimating unit 145 determines that there is no parts region Rr as a result of the 3D analysis, it determines that estimation of the parts region Rr has failed. If estimation of the parts region Rr fails in this way, the region estimating unit 145 may associate the 3D model of the character with the parts and write them in a log file, for example.
  • the region estimation unit 145 detects wobbling according to a change in the direction of the normal vector of the parts region Rr and corrects the shape of the parts region Rr. Correction of the parts region Rr is not limited to this.
  • the region estimation unit 145 may correct the parts region Rr using machine learning.
  • the region estimating unit 145 can correct the parts region Rr using a learned correction model that receives the parts region Rr as an input and outputs the corrected parts region Rc.
  • extraction unit 146 acquires character information regarding the character as metadata based on at least one of the image recognition processing result, the estimation result of the parts position (image recognition position Rm), and the parts region Rr. .
  • the metadata acquired by the extraction unit 146 includes metadata obtained by image recognition by the image recognition unit 143 and metadata obtained based on the parts region Rr estimated by the region estimation unit 145.
  • the extraction unit 146 extracts, for example, at least one of classification information, feature amount information, and relative information as metadata.
  • the classification information includes, for example, information regarding class classification in which the parts region Rr is classified into classes.
  • the feature amount information includes information regarding the feature amount vector (in other words, the latent space indicating the feature amount) of the parts region Rr.
  • the relative information includes information about the relative sizes and positions (relative positions) between a plurality of parts, information about the relative sizes and positions (relative positions) between a character and the parts, and the like.
  • the extraction unit 146 extracts, for example, classification information and feature amount information based on the image recognition result.
  • the extraction unit 146 extracts classification information from the image P using, for example, a deep learning class classification task.
  • the following classes are classified by the extraction unit 146 based on the image recognition results. Note that the following is an example, and the extraction unit 146 may classify parts into classes other than the following classes.
  • ⁇ Color e.g. the color of eyes, hair, clothes, etc.
  • shape for example, for eyes, sagging or slanted eyes; for hair, short or long, etc.
  • Gene for example, accessories such as glasses and necklaces; costumes such as hats and jackets
  • the extraction unit 146 extracts feature information by clustering on a latent space created by Variational Auto Encoder, Generative Adversarial Network, etc., based on the image recognition results.
  • the extraction unit 146 measures the degree of similarity according to the character of the part by extracting the feature amount information.
  • the information processing device 100 can estimate the degree of similarity between the same parts (for example, faces, eyes, hairstyles, etc.) of different characters (for example, characters #1 and #2, not shown).
  • the extraction unit 146 extracts, for example, classification information, feature amount information, and relative information based on the 3D model of the part region Rr, in other words, the analysis result of the 3D shape of the 3D model.
  • the extraction unit 146 extracts classification information and feature amount information from the 3D analysis results, for example, in the same manner as the image recognition results. At this time, the extraction unit 146 can extract classification information and feature amount information, for example, limited to the parts region Rr.
  • the following classes are classified by the extraction unit 146 based on the 3D analysis results. Note that the following is an example, and the extraction unit 146 may classify parts into classes other than the following classes. ⁇ Texture (rugged, wrinkled, smooth, etc.) ⁇ Level of detail (mesh data with many vertices (high poly), few (low poly), etc.)
  • the extraction unit 146 extracting metadata using the 3D model of the parts region Rr, the extraction accuracy can be further improved and the processing load can be further reduced.
  • the extraction unit 146 extracts metadata (for example, classification information and feature amount information) using the image recognition results.
  • the extraction unit 146 can extract metadata with higher precision by extracting metadata based on the 3D analysis results together with the metadata extracted from the image recognition results.
  • the extraction unit 146 extracts metadata (classification information) indicating that the character is a "muscular male” and that the 3D model "has a hand region” from the image recognition results.
  • the extraction unit 146 classifies the parts, including this metadata.
  • the extraction unit 146 extracts the classification information of the parts by inputting the 3D model (mesh data) of the parts region Rr and the metadata acquired from the image recognition results into a neural network.
  • the extraction unit 146 extracting metadata based on the image recognition result and the 3D analysis result, the extraction unit 146 can extract the metadata of the part with higher accuracy.
  • the extraction unit 146 extracts, for example, relative information based on the 3D model of the character, in other words, the analysis result of the 3D shape of the 3D model.
  • the extraction unit 146 extracts, for example, the relative positional relationship and relative size of multiple parts (for example, right eye and left eye, or face and eyes) of a specific character as relative information.
  • the extraction unit 146 measures the position and size of the character's eyes relative to the head, and extracts the measurement results as metadata.
  • the extraction unit 146 stores the parts and the extracted metadata in the parts DB 133 and metadata DB 134 in association with each other. Note that the region estimating unit 145 may store the parts in the parts DB 133.
  • the extraction unit 146 extracts metadata of parts by image recognition. Further, the extraction unit 146 uses the image recognition results to narrow down the metadata extraction range, and then extracts the metadata of the part by 3D shape analysis. Thereby, the extraction unit 146 can further reduce the load of extraction processing. Further, the extraction unit 146 can extract metadata with higher accuracy.
  • the metadata extracted by the extraction unit 146 is used, for example, when a user searches for parts. In this manner, the extraction unit 146 extracts metadata, associates it with parts, and stores it in the metadata DB 134, allowing the user to search for desired parts faster and more easily.
  • the search processing unit 147 presents the user with parts (an example of parts information) corresponding to the metadata according to the search conditions specified by the user.
  • the search processing unit 147 searches the metadata DB 134 according to search conditions specified by the user, and presents parts corresponding to the search results to the user.
  • the search processing unit 147 presents the user with a 2D rendered image of the part, for example, in accordance with the 3D model of the part.
  • FIG. 9 is a diagram for explaining an example of search processing according to the embodiment of the present disclosure.
  • the search processing unit 147 receives parts search conditions from the user.
  • the search processing unit 147 receives search conditions via the input/output unit 120, for example.
  • the search processing unit 147 searches the metadata DB 134 by specifying metadata according to the search conditions accepted from the user.
  • the metadata DB 134 specifies parts corresponding to the metadata specified by the search processing unit 147 to the parts DB 133.
  • the parts DB 133 notifies the search processing unit 147 of the parts specified from the metadata DB 134.
  • the search processing unit 147 presents the parts acquired from the parts DB 133 to the user as a search result.
  • the search processing unit 147 searching for parts using the metadata stored in the metadata DB 134, the user can search more easily and in a shorter time.
  • the search processing unit 147 presents a search UI image and accepts search conditions from the user.
  • FIG. 10 is a diagram illustrating an example of a search UI image according to an embodiment of the present disclosure.
  • the search UI image shown in FIG. 10 is generated by the UI control unit 148 based on an instruction from the search processing unit 147, for example.
  • the search processing unit 147 when searching for parts based on classification information or relative information, the search processing unit 147 presents the search UI image shown in FIG. 10 to the user. For example, when the part is an "eye", the search processing unit 147 further narrows down the parts using classification information (corresponding to words and keywords in FIG. 10) and relative information (corresponding to size specification in FIG. 10). Search information is received from the user using the search UI image shown in FIG.
  • the user can specify the class of the classification information, for example, by inputting a free word or selecting a tag (such as "anime” or "girl” in FIG. 10). Furthermore, the user can specify the relative positions and relative sizes of parts by adjusting numerical values using sliders. For example, in the example of FIG. 10, by adjusting the slider, the user can specify the relative size of the eyes to the size of the face. Note that the user may specify the relative information by adjusting the notification using a slider, or may specify the relative information by directly specifying a numerical value.
  • the search processing unit 147 can obtain search conditions by the user directly specifying metadata.
  • the search processing unit 147 may acquire search conditions by the user specifying an image.
  • the search processing unit 147 extracts feature information from the image specified by the user, and performs a parts search based on the feature information.
  • FIG. 11 is a diagram for explaining an example of feature information extraction by the search processing unit 147 according to the embodiment of the present disclosure. In FIG. 11, it is assumed that the user has specified image S_0 as a search condition.
  • the search processing unit 147 inputs, for example, image S_0 to the encoder.
  • the encoder for example, extracts a feature amount vector (latent space indicating a feature amount) from an image.
  • the search processing unit 147 extracts the feature vector V_0 corresponding to the image S_0 by inputting the image S_0 to an encoder.
  • the search processing unit 147 searches for a character (or part) close to the image S_0 in the latent space using the extracted feature vector V_0.
  • FIG. 12 is a diagram for explaining an example of a search in the latent space by the search processing unit 147 according to the embodiment of the present disclosure.
  • FIG. 12 shows a two-dimensional latent space to simplify the illustration, the actual latent space is a multidimensional space with two or more dimensions.
  • the search processing unit 147 maps the feature vector V_0 extracted from the image S_0 to the latent space.
  • the search processing unit 147 selects a representative feature vector from among the feature vectors located within the search range SR_0 that includes the feature vector V_0 in the latent space as a search result vector.
  • the search processing unit 147 can select a search result vector depending on the distance and direction in the latent space. For example, the search processing unit 147 may randomly select a search result vector from within the search range SR_0. In the example of FIG. 12, the search processing unit 147 selects feature vectors Vc_01 to Vc_04 as search result vectors.
  • the search processing unit 147 specifies the search result vector and searches the metadata DB 134, thereby acquiring the parts corresponding to the search result vector from the parts DB 133 as the search results.
  • the search processing unit 147 presents the acquired parts to the user as a search result.
  • FIG. 13 is a diagram illustrating an example of a search result image according to the embodiment of the present disclosure.
  • the search result image shown in FIG. 13 is generated by the UI control unit 148 based on an instruction from the search processing unit 147, for example.
  • the search processing unit 147 displays the image S_0 specified by the user in the center of the search result images. Furthermore, the search processing unit 147 displays 2D images Sc_01 to Sc_04 of parts (parts showing the upper body in the example of FIG. 13) that are search results around the image S_0.
  • the search processing unit 147 searches for parts similar to the image specified by the user (within a predetermined search range in the latent space) by searching using the latent space (feature vector). can do.
  • the search processing unit 147 can accept changes in the search range from the user using icon I in FIG. 13 .
  • FIG. 14 is a diagram for explaining an example of search range changing processing in the search processing unit 147 according to the embodiment of the present disclosure.
  • the search processing unit 147 receives a movement of the icon I as an instruction to change the search range from the user.
  • the search processing unit 147 that has received such a movement of the icon I changes the search range according to the movement of the icon I and performs a parts search process. For example, as shown in the lower diagram of FIG. 14, the search processing unit 147 searches for parts by changing the search range SR_0 to the search range SR_1.
  • the search range SR_0 is a search range according to the feature vector V_0 corresponding to the image S_0 specified by the user, and is, for example, a range centered on the feature vector V_0.
  • the search range SR_1 is a search range according to the feature vector V_1 according to the movement of the icon I, and is, for example, a range centered on the feature vector V_1.
  • the feature vector V_1 is a vector obtained by moving the feature vector V_0 in the direction of the feature vector Vc_02 according to the amount of movement of the icon I (the length of the arrow in the upper diagram of FIG. 14).
  • the search processing unit 147 moves the feature vector V_0 in the direction of the feature vector Vc_02 by the ratio of the amount of movement of the icon I to the distance between the icon I_0 and the 2D image Sc02. Calculate V_1.
  • the search processing unit 147 selects a search result vector from the changed search range SR_1.
  • the search processing unit 147 selects a search result vector from within the search range SR_1 in the same manner as the method for selecting a search result vector from within the search range SR_0.
  • the search processing unit 147 selects feature vectors Vc_11 to Vc_14 as search result vectors.
  • FIG. 15 is a diagram illustrating another example of a search result image according to the embodiment of the present disclosure.
  • an image showing the results of the search performed by the search processing unit 147 in the search range SR_1 is shown.
  • the search processing unit 147 displays 2D images Sc_11 to Sc_14 of parts corresponding to the feature vectors Vc_11 to Vc_14, which are search result vectors, around the icon I. Further, the search processing unit 147 may display the image S_0 specified by the user in addition to the 2D images Sc_11 to Sc_14.
  • the search processing unit 147 receives a change in the search range within the latent space from the user using the icon I, for example. Thereby, the user can more intuitively change the search range in the latent space, and can more easily search for a desired part.
  • the search processing unit 147 accepts changes in the search range using icons, more specifically, according to the movement of the icon, but the method for changing the search range is not limited to this.
  • the search processing unit 147 may change the search range when the user clicks on the 2D images Sc_01 to Sc_04. In this case, the search processing unit 147 selects a search result vector within the search range SR according to the feature vector Vc corresponding to the clicked 2D image Sc, for example.
  • the search processing unit 147 may accept a change in the search range from the user using a tool such as a slider to adjust numerical values.
  • the search processing unit 147 changes the search range SR according to a numerical value specified by the user using a slider, for example, and searches for parts.
  • search processing unit 147 searches in the latent space using the image S_0 specified by the user, but the search in the latent space by the search processing unit 147 is not limited to this.
  • the search processing unit 147 may randomly select a feature vector corresponding to the search range SR. For example, when the user specifies a part, the search processing unit 147 randomly selects one feature vector of the specified part. The search processing unit 147 sets a search range corresponding to the selected feature amount vector in the latent space, and selects a search result vector within the set search range.
  • the search processing unit 147 can randomly search for parts and present them to the user.
  • the search processing unit 147 selects four feature vectors as search result vectors, but the number of feature vectors selected by the search processing unit 147 is not limited to four.
  • the search processing unit 147 may select three or less feature vectors as the search result vector, or may select five or more feature vectors.
  • the search processing unit 147 presents to the user 2D images corresponding to all the feature vectors selected as search result vectors, but the 2D images presented to the user are not limited to this.
  • the search processing unit 147 may present to the user a part of the 2D image corresponding to the feature vector selected as the search result vector. For example, as shown in FIG. 14, even if the search processing unit 147 selects four feature vectors Vc_11 to Vc_14, the search processing unit 147 can present three or less 2D images to the user.
  • the UI control unit 148 generates a screen (UI) and accepts operations on the UI.
  • the UI control unit 148 generates a search UI image or a search result image according to an instruction from the search processing unit 147, for example, and presents them to the user via the input/output unit 120. Further, the UI control unit 148 receives input of search conditions and changes in the search range from the user via the input/output unit 120, for example.
  • the UI control unit 148 notifies, for example, the search processing unit 147 of the input results from the user.
  • FIG. 16 is a flowchart illustrating an example of the flow of the first parts separation process according to the embodiment of the present disclosure.
  • the first parts separation process shown in FIG. 16 is executed by the information processing device 100, for example.
  • the information processing device 100 executes the first parts separation process, for example, in accordance with an instruction from a user.
  • the information processing device 100 first obtains a 3D model of a character (step S101).
  • the information processing device 100 acquires a 3D model from the 3D model DB 131, for example.
  • the information processing apparatus 100 may acquire the 3D model of the character from a range specified by the user.
  • the information processing device 100 classifies the acquired 3D model into major parts (step S102).
  • the major classification parts are parts that are larger than the parts that the information processing apparatus 100 separates in the first parts separation process.
  • the major classification parts include, for example, a head region and a body region.
  • the major classification parts may include a head region, an upper body region, and a lower body region. In this way, the information processing apparatus 100 divides the 3D model into large classification parts that are larger than the parts (for example, eyes, nose, etc.) to be separated in the first parts separation process.
  • the major classification parts are larger than the parts to be separated in the first part separation process. Therefore, the process in which the information processing apparatus 100 separates a 3D model into major parts requires less processing load than the process in which the 3D model is separated into parts (for example, eyes, nose, etc.).
  • the information processing device 100 selects one major classification part from among the divided major classification parts, renders the selected major classification part, and generates an image P (step S103).
  • the information processing device 100 performs part image recognition on the image P generated in step S103 (step S104). For example, the information processing apparatus 100 selects one part to be separated from a plurality of parts, and executes image recognition processing to estimate the position of the selected part with respect to the image P.
  • the information processing device 100 generates the image P by rendering the major classification parts. Therefore, the information processing apparatus 100 can recognize the image P with higher accuracy than when performing image recognition of an image in which the entire character is rendered.
  • the information processing device 100 determines whether or not recognition of the image P has been successful (step S105). For example, the information processing device 100 determines whether or not the recognition of the image P is successful depending on whether or not the parts can be recognized and whether or not the recognition accuracy of the parts is equal to or higher than a threshold value.
  • step S105 If it is determined that recognition of the image P has failed (step S105; No), that is, if the part could not be recognized or the recognition accuracy is less than the threshold, the information processing device 100 proceeds to step S110.
  • step S105 When it is determined that the image P has been successfully recognized (step S105; Yes), that is, when the parts can be recognized or when the recognition accuracy is equal to or higher than the threshold, the information processing device 100 sets the image recognition position Rm in the 3D model. is estimated (step S106). The information processing device 100 estimates the image recognition position Rm in the 3D model according to the image recognition position Rp obtained from the recognition result of the image P and the setting information of the virtual camera C.
  • the information processing device 100 estimates the parts region Rr in the 3D model based on the image recognition position Rm (step S107). For example, the information processing device 100 estimates the parts region Rr according to the characteristics of the parts to be separated.
  • the information processing device 100 extracts metadata corresponding to the parts region Rr based on the image recognition results of the parts region Rr and the image P (step S108). The information processing device 100 separates the 3D model of the parts region Rr into parts.
  • the information processing device 100 stores the parts and metadata (step S109).
  • the information processing device 100 associates parts and metadata and stores them in a parts DB 133 and a metadata DB 134, respectively.
  • the information processing device 100 determines whether all parts have been separated in the major classification parts selected in step S103 (step S110). If there are parts that have not been separated (step S110; No), the information processing device 100 returns to step S104 and executes separation processing for the parts that have not been separated yet.
  • step S110 determines whether all parts have been separated. That is, the information processing device 100 determines whether all major classification parts have been separated (step S111). That is, the information processing device 100 determines whether all parts have been extracted in the 3D model of the character.
  • step S111 If there are major classification parts that have not been separated yet (step S111; No), the information processing device 100 returns to step S103 and performs the process of separating parts for the major classification parts that have not been separated. conduct.
  • step S111 determines whether the parts of all the 3D models have been separated. That is, the information processing device 100 determines whether all parts have been extracted for all characters.
  • step S112 If there is a 3D model whose parts have not been separated yet (step S112; No), the information processing device 100 returns to step S101 and obtains a 3D model of the character whose parts have not been separated.
  • step S112 if parts have been separated in all 3D models (step S112; Yes), the information processing device 100 ends the first parts separation process.
  • the information processing apparatus 100 divides the 3D model into major classification parts and then renders the major classification parts.
  • the information processing device 100 may render the 3D model and then divide it into major parts.
  • the information processing device 100 renders the entire 3D model and generates an image including the entire character.
  • the information processing device 100 performs image recognition processing on the image including the entire image of the character, cuts out a region including the major classification parts, and generates the image P.
  • the information processing device 100 may generate the image P by estimating a region including the major classification parts through image recognition processing and re-rendering a 3D model corresponding to the estimated region.
  • the information processing apparatus 100 leaves a log in a log file and performs next part or next 3D Analysis (parts separation) of the model can be performed.
  • the information processing device 100 generates a 2D image by rendering the 3D model information of the character.
  • the information processing device 100 performs image recognition processing to recognize parts to be separated on the generated 2D image, and estimates the part position Rp in the 2D image.
  • the information processing device 100 estimates a rough part position Rm in the 3D space (3D model) based on the part position Rp in the 2D image.
  • the information processing device 100 analyzes the 3D model according to the rough part position Rm in the 3D model and the characteristics of the part, and estimates the part region Rr in the 3D model.
  • the information processing device 100 separates (generates) 3D model information (for example, mesh data) of the parts region Rr from the 3D model information of the character. Thereby, the information processing device 100 separates the parts from the character.
  • 3D model information for example, mesh data
  • the information processing device 100 extracts metadata corresponding to the part using the parts region Rr and the results of the image recognition process.
  • the information processing device 100 stores parts and metadata in association with each other.
  • the information processing device 100 can separate parts from a character with higher accuracy while further reducing the processing load of the process of separating parts from a character. Furthermore, the information processing device 100 associates and holds parts and metadata, allowing the user to search for desired parts more accurately and in a shorter time.
  • the information processing apparatus 100 automatically performs the process of separating parts from the character, but the user may perform part of the process. That is, the information processing device 100 may perform a process of separating parts from a character (second parts separation process) while interacting with the user.
  • FIG. 17 is a flowchart showing an example of the flow of the second parts separation process according to the embodiment of the present disclosure.
  • the second parts separation process shown in FIG. 17 is executed by the information processing device 100, for example.
  • the information processing apparatus 100 executes the second parts separation process shown in FIG. 17, for example, in accordance with an instruction from a user. Note that among the second parts separation processing shown in FIG. 17, the same processes as the first parts separation processing shown in FIG.
  • the information processing device 100 that performed image recognition of the part in step S104 presents the recognition result to the user (step S201).
  • the information processing device 100 receives a modification (change) of the recognition result from the user by presenting the recognition result to the user.
  • FIG. 18 is a diagram for explaining an example of correction of recognition results according to the embodiment of the present disclosure.
  • the information processing apparatus 100 presents the user with a UI image PU_1 in which the part position Rp, which is the recognition result, is superimposed on the image P, which is the recognition target.
  • the image recognition unit 143 of the information processing device 100 instructs the UI control unit 148 to generate the UI image PU_1.
  • the UI control unit 148 displays the part position Rp using, for example, a primitive figure such as an ellipse or a rectangle.
  • the user can check whether the part recognition by the information processing device 100 is correct using the UI image PU_1. For example, if the information processing device 100 recognizes the part incorrectly, such as when the part position Rp deviates from the actual character part position, the user corrects the part position Rp. The user corrects the part position Rp by, for example, performing a GUI operation such as drag and drop. Thereby, as shown in the lower diagram of FIG. 18, the user can instruct the information processing apparatus 100 about the correct position of the parts.
  • the UI control unit 148 is unable to generate the UI image PU_1 in which a figure indicating the part position Rp is superimposed on the image P, such as when the information processing device 100 is unable to recognize any parts.
  • the UI control unit 148 presents the user with, for example, a UI image that includes the image P but does not include the figure indicating the part position Rp.
  • the user instructs the information processing apparatus 100 about the correct position of the part by drawing a figure indicating the part position Rp on the image P.
  • the UI control unit 148 may present to the user a UI image in which a figure indicating the part position Rp is drawn at a predefined position (default position) such as a corner of the image P, for example.
  • the user instructs the information processing apparatus 100 about the correct position of the part by, for example, correcting the part position Rp by performing a GUI operation such as drag and drop.
  • the information processing device 100 estimates the image recognition position Rm in the 3D model (step S106) and the parts region Rr (step S107) based on the part position Rp specified by the user.
  • the information processing device 100 presents the estimation result of the parts region Rr to the user (step S202).
  • the information processing device 100 receives a modification (change) of the estimation result from the user by presenting the estimation result of the parts region Rr to the user.
  • FIG. 19 is a diagram for explaining an example of a UI image showing estimation results according to the embodiment of the present disclosure.
  • the information processing apparatus 100 presents the user with a UI image PU_2 showing the parts region Rr.
  • the area estimation unit 145 of the information processing device 100 instructs the UI control unit 148 to generate the UI image PU_2.
  • the UI control unit 148 generates a rendered image of the 3D model including the parts region Rr as the UI image PU_2.
  • the UI image PU_2 shown in FIG. 19 is a rendered image of the parts region Rr viewed from the front.
  • the UI control unit 148 generates the UI image PU_2 by superimposing information regarding the mesh of the 3D model (for example, information indicating vertices and edges). In this way, by the information processing device 100 presenting the parts region Rr including the information regarding the mesh to the user, the user can more easily confirm the parts region Rr in the 3D model.
  • the UI control unit 148 may highlight the parts region Rr in the UI image PU_2, for example by brightly highlighting the parts region Rr.
  • the UI control unit 148 may display the area other than the parts area Rr in a display color different from that of the parts area Rr, such as by making the area other than the parts area Rr darker in the UI image PU_2.
  • the information processing device 100 may present the parts region Rr viewed from a plurality of viewpoints to the user.
  • 20 and 21 are diagrams for explaining other examples of UI images showing estimation results according to the embodiment of the present disclosure.
  • the information processing device 100 generates a UI image PU_3 that is a rendered 3D model of the parts region Rr from a different viewpoint than the UI image PU_2, and presents it to the user.
  • the information processing apparatus 100 generates a UI image PU_4, which is a 3D model of the parts region Rr rendered from a different viewpoint than the UI images PU_2 and PU_3, and presents it to the user.
  • the information processing device 100 may present the UI images PU_2 to PU_3 side by side to the user.
  • the information processing apparatus 100 may generate UI images PU_3 and PU_4 with changed viewpoints in response to instructions from the user, and present them to the user.
  • the information processing device 100 accepts corrections to the parts region Rr from the user.
  • the user modifies the parts area Rr by performing GUI operations. For example, the user modifies the parts region Rr by selecting a surface to be added to the parts region Rr by a click operation or the like. Alternatively, the user can modify the part region Rr by selecting multiple faces using a range selection tool such as a rectangle or a lasso (a freehand drawn figure).
  • the UI control unit 148 is unable to generate the parts region Rr as the UI image PU_2, such as when the information processing device 100 is unable to estimate the parts region Rr at all.
  • the UI control unit 148 presents the user with a UI image in which a 3D model of the character is rendered, for example.
  • the UI control unit 148 may present the user with a UI image in which the major classification parts are rendered.
  • the user instructs the information processing apparatus 100 about the correct part region Rr by selecting a surface included in the UI image using a click operation or a range selection tool.
  • the information processing device 100 extracts metadata based on the parts region Rr specified by the user (step S108).
  • the subsequent processing is the same as the first parts separation processing in FIG. 16.
  • the information processing device 100 can separate parts from the character with higher accuracy by accepting at least one of the correction of the parts position Rp and the correction of the parts region Rr by the user. I can do it.
  • the information processing apparatus 100 processed all parts while interacting with the user, that is, while confirming the estimation results of the part position Rp and the part region Rr with the user.
  • the information processing apparatus 100 accepts at least one of the correction of the parts position Rp and the correction of the parts region Rr by the user. Good too. Note that if the information processing device 100 fails to estimate the parts position Rp or the parts position Rm, it does not estimate the parts region Rr. Therefore, when the information processing apparatus 100 fails to estimate the part position Rp or the part position Rm, it means that the information processing apparatus 100 fails to estimate the parts region Rr.
  • FIG. 22 is a flowchart showing an example of the flow of the third parts separation process according to the embodiment of the present disclosure.
  • the third parts separation process shown in FIG. 22 is executed by the information processing device 100, for example.
  • the information processing apparatus 100 executes the third parts separation process shown in FIG. 22, for example, in accordance with instructions from the user.
  • the information processing device 100 executes a first parts separation process (step S301). At this time, the information processing apparatus 100 writes information regarding the 3D model for which estimation of the parts position Rp or parts region Rr has failed in the log file.
  • the information processing device 100 obtains a log file from the log file DB 132 (step S302).
  • the information processing device 100 performs the second parts separation process on the 3D model for which the estimation of the part position Rp or the estimation of the parts region Rr has failed (step S303).
  • the information processing device 100 performs the second parts separation process on the 3D model whose parts have failed to be separated, and by accepting corrections from the user, separates parts from the character with higher precision. be able to.
  • the information processing apparatus 100 does not accept corrections from the user for all 3D models, but only accepts corrections from the user for 3D models in which parts separation has failed. Thereby, the information processing apparatus 100 can separate parts with higher precision while suppressing an increase in the burden on the user.
  • FIG. 23 is a block diagram showing an example of the hardware configuration of the information processing device 100 according to this embodiment. Note that the information processing device 800 shown in FIG. 23 can realize the information processing device 100, for example. Information processing by the information processing apparatus 100 according to the present embodiment is realized by cooperation between software and hardware described below.
  • the information processing device 800 includes, for example, a CPU 871, a ROM 872, a RAM 873, a host bus 874, a bridge 875, an external bus 876, and an interface 877.
  • the information processing device 800 also includes an input device 878, an output device 879, a storage 880, a drive 881, a connection port 882, and a communication device 883.
  • the hardware configuration shown here is an example, and some of the components may be omitted. In addition, components other than those shown here may be further included.
  • the CPU 871 functions, for example, as an arithmetic processing device or a control device, and controls the overall operation of each component or a portion thereof based on various programs recorded in the ROM 872, RAM 873, storage 880, or removable recording medium 901.
  • the CPU 871 implements operational processing within the information processing device 100.
  • the ROM 872 is a means for storing programs read into the CPU 871, data used for calculations, and the like.
  • the RAM 873 temporarily or permanently stores, for example, programs read into the CPU 871 and various parameters that change as appropriate when executing the programs.
  • the CPU 871, ROM 872, and RAM 873 are interconnected, for example, via a host bus 874 capable of high-speed data transmission.
  • the host bus 874 is connected, for example, via a bridge 875 to an external bus 876 whose data transmission speed is relatively low.
  • the external bus 876 is connected to various components via an interface 877.
  • the input device 878 includes, for example, a mouse, a keyboard, a touch panel, a button, a switch, a lever, and the like. Furthermore, as the input device 878, a remote controller (hereinafter referred to as remote control) that can transmit control signals using infrared rays or other radio waves may be used. Furthermore, the input device 878 includes an audio input device such as a microphone.
  • the output device 879 is, for example, a display device such as a CRT (Cathode Ray Tube), LCD, or organic EL, an audio output device such as a speaker or headphone, a printer, a mobile phone, or a facsimile, etc., for transmitting the acquired information to the user. This is a device that can notify visually or audibly. Further, the output device 879 according to the present disclosure includes various vibration devices capable of outputting tactile stimulation.
  • Storage 880 is a device for storing various data.
  • a magnetic storage device such as a hard disk drive (HDD), a semiconductor storage device, an optical storage device, a magneto-optical storage device, or the like is used.
  • the drive 881 is a device that reads information recorded on a removable recording medium 901 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory, or writes information to the removable recording medium 901, for example.
  • a removable recording medium 901 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory
  • the removable recording medium 901 is, for example, DVD media, Blu-ray (registered trademark) media, HD DVD media, various semiconductor storage media, and the like.
  • the removable recording medium 901 may be, for example, an IC card equipped with a non-contact IC chip, an electronic device, or the like.
  • connection port 882 is, for example, a port for connecting an external connection device 902 such as a USB (Universal Serial Bus) port, an IEEE1394 port, a SCSI (Small Computer System Interface), an RS-232C port, or an optical audio terminal.
  • an external connection device 902 such as a USB (Universal Serial Bus) port, an IEEE1394 port, a SCSI (Small Computer System Interface), an RS-232C port, or an optical audio terminal.
  • the external connection device 902 is, for example, a printer, a portable music player, a digital camera, a digital video camera, or an IC recorder.
  • the communication device 883 is a communication device for connecting to a network, and includes, for example, a communication card for wired or wireless LAN, Wi-Fi (registered trademark), Bluetooth (registered trademark), or WUSB (Wireless USB), optical communication.
  • a computer program for causing hardware such as a CPU, ROM, and RAM built into the information processing device 100 described above to exhibit the functions of the information processing device 100.
  • a computer-readable storage medium (recording medium) storing the computer program is also provided.
  • the present technology can also have the following configuration.
  • (1) Obtain a 3D model of the character, performing image recognition processing on an image seen from a virtual viewpoint of the three-dimensional model to estimate the position of the character's parts;
  • An information processing device comprising: a control unit that estimates a part area in the three-dimensional model of the character based on the three-dimensional model and the part position.
  • (2) The information processing device according to (1), wherein the parts area includes an area corresponding to a body part of the character.
  • (3) The information processing device according to (1) or (2), wherein the parts area includes at least one of an eye area, a nose area, a mouth area, and an ear area of the character.
  • the information processing device according to any one of (1) to (3), wherein the control unit generates a three-dimensional model of the parts area from the three-dimensional model of the character based on the parts area. (5) (1) to (4), wherein the control unit obtains character information regarding the character based on at least one of a result of the image recognition process, a result of estimating the part position, and a result of estimating the part area.
  • the information processing device according to any one of the above.
  • the character information includes at least classification information regarding class classification in which the parts area is classified into classes, feature information regarding a feature vector of the parts area, and relative information regarding the relative relationship of the parts area to the character.
  • the information processing device including one.
  • control unit stores the character information in association with parts information regarding the parts area.
  • control unit presents the user with the parts information corresponding to the character information according to conditions specified by the user.
  • the control unit estimates the part position in the three-dimensional model based on the part position of the character in the image estimated by the image recognition process. information processing equipment.
  • the control unit estimates the part position in the three-dimensional model based on the angle of view of the virtual viewpoint corresponding to the image and the part position of the character in the image. Processing equipment.
  • the information processing device according to any one of (1) to (10), wherein the control unit estimates the part area in mesh data included in the three-dimensional model. (12) The information processing device according to any one of (1) to (11), wherein the control unit corrects the shape of the part area in the three-dimensional model according to wobbling in the outline of the part area. . (13) The information processing device according to any one of (1) to (12), wherein the control unit receives a change in at least one of the part position and the part area from a user. (14) The information processing according to (13), wherein the control unit receives from the user the change of at least one of the part position and the part area of the character whose image recognition has failed as a result of the image recognition process of the image. Device.
  • Information processing device 110 Communication unit 120 Input/output unit 130 Storage unit 131 3D model DB 132 Log file DB 133 Parts DB 134 Metadata DB 140 Control unit 141 Model acquisition unit 142 Rendering unit 143 Image recognition unit 144 Position estimation unit 145 Area estimation unit 146 Extraction unit 147 Search processing unit 148 UI control unit

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Graphics (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Geometry (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Processing Or Creating Images (AREA)

Abstract

This information processing device (100) comprises a control unit (140). The control unit (140) acquires a three-dimensional model of a character. The control unit (140) performs image recognition processing on an image viewed from a virtual viewpoint of the three-dimensional model, and estimates part positions of the character. The control unit (140) estimates a part area in the three-dimensional model of the character on the basis of the three-dimensional model and the part positions.

Description

情報処理装置、記録媒体及び情報処理方法Information processing device, recording medium and information processing method
 本開示は、情報処理装置、記録媒体及び情報処理方法に関する。 The present disclosure relates to an information processing device, a recording medium, and an information processing method.
 従来、コンピュータグラフィックスの分野において、仮想空間上の3次元モデル(以下、3Dモデルとも記載する)を制作する場合、複数のパーツを組み合わせて3Dモデルを作成する手法が知られている。また、3Dモデルのパーツを管理するシステムとして、例えば、パーツのデータを抽出し、当該データを用いてパーツの検索を行うシステムが知られている。 Conventionally, in the field of computer graphics, when creating a three-dimensional model (hereinafter also referred to as a 3D model) in virtual space, a method is known in which a 3D model is created by combining multiple parts. Further, as a system for managing parts of a 3D model, for example, a system is known that extracts data on parts and searches for parts using the data.
特開2007-109221号公報Japanese Patent Application Publication No. 2007-109221
 上述したシステムは、既にパーツに分けられた3Dモデルからパーツのデータを抽出する。このように、上述したシステムでは、既に3Dモデルからパーツが分けられていることが前提となっており、3Dモデルからどのようにパーツを分離するかについて考慮されていない。 The system described above extracts part data from a 3D model that has already been divided into parts. In this way, the above-described system assumes that parts have already been separated from the 3D model, and does not consider how to separate the parts from the 3D model.
 3Dモデルなど3Dデータの分析は、処理が複雑で計算量が多く、難易度が高い。3Dモデルをパーツに分離する処理には、3Dデータの分析が用いられる。このことから、3Dモデルをパーツに分離する処理を容易(例えば、短時間又は高精度)に行うことは難しかった。そのため、より容易に3Dモデルの分析が行えるようになることが望まれる。 Analyzing 3D data such as 3D models requires complex processing and a large amount of calculation, making it highly difficult. Analysis of 3D data is used to separate the 3D model into parts. For this reason, it has been difficult to easily (for example, in a short time or with high precision) perform a process of separating a 3D model into parts. Therefore, it is desired that 3D models can be analyzed more easily.
 そこで、本開示では、より容易に3Dモデルを分析することができる仕組みを提供する。 Therefore, the present disclosure provides a mechanism that allows 3D models to be analyzed more easily.
 なお、上記課題又は目的は、本明細書に開示される複数の実施形態が解決し得、又は達成し得る複数の課題又は目的の1つに過ぎない。 Note that the above-mentioned problem or object is only one of the plurality of problems or objects that can be solved or achieved by the plurality of embodiments disclosed in this specification.
 本開示の情報処理装置は、制御部を備える。制御部は、キャラクタの3次元モデルを取得する。制御部は、前記3次元モデルの仮想視点から見た画像に対して画像認識処理を行って、前記キャラクタのパーツ位置を推定する。制御部は、前記3次元モデル、及び、前記パーツ位置に基づき、前記キャラクタの前記3次元モデルにおけるパーツ領域を推定する。 The information processing device of the present disclosure includes a control unit. The control unit obtains a three-dimensional model of the character. The control unit performs image recognition processing on the image seen from the virtual viewpoint of the three-dimensional model to estimate the positions of the parts of the character. The control unit estimates a part area in the three-dimensional model of the character based on the three-dimensional model and the part position.
本開示の実施形態に係る3Dモデルの分析処理の概要を説明するための図である。FIG. 2 is a diagram for explaining an overview of a 3D model analysis process according to an embodiment of the present disclosure. 本開示の実施形態に係る情報処理装置の構成例を示すブロック図である。FIG. 1 is a block diagram illustrating a configuration example of an information processing device according to an embodiment of the present disclosure. 本開示の実施形態に係るレンダリング部によるレンダリング処理の一例を説明するための図である。FIG. 3 is a diagram for explaining an example of rendering processing by a rendering unit according to an embodiment of the present disclosure. 本開示の実施形態に係る位置推定部によるパーツ位置推定の一例を示す図である。It is a figure showing an example of part position estimation by a position estimating part concerning an embodiment of this indication. 本開示の実施形態に係る位置推定部によるパーツ位置推定の他の例を示す図である。FIG. 7 is a diagram illustrating another example of part position estimation by the position estimation unit according to the embodiment of the present disclosure. 本開示の実施形態に係る領域推定部によるパーツ領域の推定処理の一例について説明するための図である。FIG. 3 is a diagram for explaining an example of a parts region estimation process by a region estimating unit according to an embodiment of the present disclosure. 本開示の実施形態に係る領域推定部が行う補正処理の一例を説明するための図である。FIG. 3 is a diagram for explaining an example of a correction process performed by a region estimation unit according to an embodiment of the present disclosure. 本開示の実施異形態に係る領域推定部が行うがたつき検出の一例を説明するための図である。FIG. 7 is a diagram for explaining an example of rattling detection performed by a region estimation unit according to an embodiment of the present disclosure. 本開示の実施形態に係る検索処理の一例を説明するための図である。FIG. 2 is a diagram for explaining an example of a search process according to an embodiment of the present disclosure. 本開示の実施形態に係る検索画像の一例を示す図である。FIG. 2 is a diagram illustrating an example of a search image according to an embodiment of the present disclosure. 本開示の実施形態に係る検索処理部による特徴量情報の抽出例を説明するための図である。FIG. 3 is a diagram for explaining an example of extraction of feature amount information by a search processing unit according to an embodiment of the present disclosure. 本開示の実施形態に係る検索処理部による潜在空間内での検索の一例を説明するための図である。FIG. 2 is a diagram for explaining an example of a search in a latent space by a search processing unit according to an embodiment of the present disclosure. 本開示の実施形態に係る検索結果の画像の一例を示す図である。FIG. 3 is a diagram illustrating an example of a search result image according to an embodiment of the present disclosure. 本開示の実施形態に係る検索処理部における検索範囲の変更処理の一例を説明するための図である。FIG. 3 is a diagram for explaining an example of a search range changing process in a search processing unit according to an embodiment of the present disclosure. 本開示の実施形態に係る検索結果の画像の他の例を示す図である。FIG. 7 is a diagram illustrating another example of a search result image according to an embodiment of the present disclosure. 本開示の実施形態に係る第1パーツ分離処理の一例の流れを示すフローチャートである。It is a flow chart which shows an example of a flow of the 1st parts separation processing concerning an embodiment of this indication. 本開示の実施形態に係る第2パーツ分離処理の一例の流れを示すフローチャートである。It is a flow chart which shows an example of the flow of the 2nd parts separation processing concerning an embodiment of this indication. 本開示の実施形態に係る認識結果の修正例について説明するための図である。FIG. 3 is a diagram for explaining an example of correction of recognition results according to an embodiment of the present disclosure. 本開示の実施形態に係る推定結果を示すUI画像の一例について説明するための図である。FIG. 2 is a diagram for explaining an example of a UI image showing an estimation result according to an embodiment of the present disclosure. 本開示の実施形態に係る推定結果を示すUI画像の他の例について説明するための図である。FIG. 7 is a diagram for explaining another example of a UI image showing an estimation result according to an embodiment of the present disclosure. 本開示の実施形態に係る推定結果を示すUI画像の他の例について説明するための図である。FIG. 7 is a diagram for explaining another example of a UI image showing an estimation result according to an embodiment of the present disclosure. 本開示の実施形態に係る第3パーツ分離処理の一例の流れを示すフローチャートである。It is a flowchart which shows an example of the flow of the 3rd parts separation process concerning an embodiment of this indication. 本実施形態に係る情報処理装置のハードウェア構成の一例を示すブロック図である。1 is a block diagram illustrating an example of a hardware configuration of an information processing device according to an embodiment. FIG.
 以下に添付図面を参照しながら、本開示の実施形態について詳細に説明する。なお、本明細書及び図面において、実質的に同一の機能構成を有する構成要素については、同一の符号を付することにより重複説明を省略する。 Embodiments of the present disclosure will be described in detail below with reference to the accompanying drawings. Note that, in this specification and the drawings, components having substantially the same functional configurations are designated by the same reference numerals and redundant explanation will be omitted.
 また、本明細書及び図面において、実施形態の類似する構成要素については、同一の符号の後に異なるアルファベット又は数字を付して区別する場合がある。ただし、類似する構成要素の各々を特に区別する必要がない場合、同一符号のみを付する。 Furthermore, in this specification and the drawings, similar components of the embodiments may be distinguished by using different alphabets or numbers after the same reference numerals. However, if there is no particular need to distinguish between similar components, only the same reference numerals are given.
 以下に説明される1又は複数の実施形態(実施例、変形例を含む)は、各々が独立に実施されることが可能である。一方で、以下に説明される複数の実施形態は少なくとも一部が他の実施形態の少なくとも一部と適宜組み合わせて実施されてもよい。これら複数の実施形態は、互いに異なる新規な特徴を含み得る。したがって、これら複数の実施形態は、互いに異なる目的又は課題を解決することに寄与し得、互いに異なる効果を奏し得る。 One or more embodiments (including examples and modifications) described below can each be implemented independently. On the other hand, at least a portion of the plurality of embodiments described below may be implemented in combination with at least a portion of other embodiments as appropriate. These multiple embodiments may include novel features that are different from each other. Therefore, these multiple embodiments may contribute to solving mutually different objectives or problems, and may produce mutually different effects.
<<1.はじめに>>
<1.1.課題>
 近年、メタバース(Metaverse)への期待から、3Dキャラクタモデルの需要増加が見込まれる。3Dキャラクタモデルは、メタバース以外でも、例えば映画などの映像制作などでも使用される。
<<1. Introduction >>
<1.1. Challenge>
In recent years, demand for 3D character models is expected to increase due to expectations for the Metaverse. 3D character models are used not only in the Metaverse but also in video production such as movies.
 メタバースや映像制作では、より高品質な3Dモデルが求められる。しかしながら、高品質な3Dモデルの制作はコストがかかる。そのため、より低コストに3Dモデルを制作できるよう、制作の簡略化が求められている。 Higher quality 3D models are required in the Metaverse and video production. However, producing high-quality 3D models is costly. Therefore, there is a need to simplify production so that 3D models can be produced at lower costs.
 3Dモデルの制作を簡略化する手法の一つとして、例えば、3Dモデルをパーツごとにデータベース(DB)に保持し、それらを組み合わせて3Dモデルの素体を作り、最後に仕上げ処理を行う手法が挙げられる。例えば、頭、目、鼻、耳、口、身体など、キャラクタの身体部位や、帽子、メガネ、服など、キャラクタの衣装、アクセサリなどのパーツを組み合わせてキャラクタの3Dモデルを作成するシステムが考えられる。 One method to simplify the production of 3D models is, for example, to store 3D models for each part in a database (DB), combine them to create the basic body of the 3D model, and finally perform finishing processing. Can be mentioned. For example, a system can be considered that creates a 3D model of a character by combining parts such as the character's body parts such as the head, eyes, nose, ears, mouth, and body, as well as the character's costumes and accessories such as hats, glasses, and clothes. .
 このように、パーツを組み合わせて3Dモデルを作成するために、システムは、キャラクタの3Dモデルをパーツごとに分類して記憶しておくことが望ましい。キャラクタの3Dモデルは、複雑で統一された仕様がないが、例えば、システムが、3Dモデルの形状分析を行い、キャラクタのパーツごとに分類することで、ユーザは、各パーツを組み合わせて所望の3Dキャラクタの素体を手軽に作成できるようになる。 In order to create a 3D model by combining parts in this way, it is desirable for the system to classify and store the 3D model of the character by parts. 3D models of characters are complex and do not have unified specifications, but for example, the system analyzes the shape of the 3D model and classifies each part of the character, allowing the user to combine each part to create the desired 3D model. You can easily create a character's body.
 また、ユーザが所望のパーツを選択(検索)できるようにするためには、システムは、パーツごとにメタデータを分析し、パーツとメタデータとを対応付けて記憶しておくことが望ましい。例えば、システムは、キャラクタの特徴やパーツの特徴から得られるメタデータを、パーツと対応付けて格納したデータベースを作成することで、ユーザは、このデータベースを検索することでより容易に所望のパーツを取得することができるようになる。 Furthermore, in order to enable the user to select (search) for a desired part, it is desirable for the system to analyze metadata for each part and store the parts and metadata in association with each other. For example, the system can create a database that stores metadata obtained from character features and parts features in association with the parts, allowing the user to more easily find the desired part by searching this database. be able to obtain it.
 このように、システムが、キャラクタの3Dモデルの分析を行い、パーツ分離やメタデータの抽出を行うことで、ユーザがより容易にキャラクタの3Dモデルを作成することができるようになる。 In this way, the system analyzes the 3D model of the character, separates the parts, and extracts metadata, allowing the user to more easily create the 3D model of the character.
 しかしながら、3Dモデルを分析し、パーツごとに分類したりメタデータを抽出したりする処理は複雑であり、計算量が大きく、難易度が高い。これは、3Dモデルが、(x,y,z)の3次元のデータであるためであり、さらに、3Dモデルでは、その構成要素にルールがなく、データ構造の自由度が非常に高いためである。 However, the process of analyzing a 3D model, classifying it into parts, and extracting metadata is complex, requires a large amount of calculation, and is highly difficult. This is because a 3D model is three-dimensional data (x, y, z), and also because there are no rules for its constituent elements and the data structure has a very high degree of freedom. be.
 例えば、3Dモデルがメッシュデータとして表現される場合、3Dモデルは、複数の頂点、この頂点を結ぶ辺、及び、この頂点と辺とで構成される面などによって表現される。 For example, when a 3D model is expressed as mesh data, the 3D model is expressed by a plurality of vertices, edges connecting the vertices, and a surface formed by the vertices and edges.
 しかしながら、3Dモデルの表現には決まった仕様がない。例えば、どの面や頂点がどのパーツに対応するか、や、パーツを構成する面や頂点の数など、明確なルールが存在しない。そのため、システムが、3Dモデルのメッシュデータを分析し、パーツを抽出することは容易ではない。 However, there are no fixed specifications for the representation of 3D models. For example, there are no clear rules regarding which faces and vertices correspond to which parts, or the number of faces and vertices that make up a part. Therefore, it is not easy for the system to analyze the mesh data of the 3D model and extract parts.
 例えば、システムがキャラクタの顔の3Dモデルから、パーツとして鼻を分離するとする。この場合、顔の頂点データのうち、どの頂点データが鼻に該当するかをシステムが判定することは容易ではない。 For example, suppose that the system separates a nose as a part from a 3D model of a character's face. In this case, it is not easy for the system to determine which vertex data of the face corresponds to the nose.
 また、3Dモデルは、2Dの画像と比較して自由度が高く、解像度のような制約がない。そのため、3Dモデルの品質が高くなるほど、データ量(例えば、頂点数など)が大きくなる。そのため、3Dモデルの分析処理がより複雑になり、計算負荷が高くなる。 Additionally, 3D models have a higher degree of freedom than 2D images, and there are no restrictions such as resolution. Therefore, the higher the quality of the 3D model, the larger the amount of data (for example, the number of vertices). Therefore, the analysis process of the 3D model becomes more complicated and the calculation load increases.
 このように、システムが、キャラクタの3Dモデル情報(例えば、上述したメッシュデータ)から、特定のパーツを分離したり、このパーツの特徴となるメタデータを抽出したりするために、3Dモデルの分析を行うことは容易ではなかった。 In this way, the system analyzes the 3D model in order to separate specific parts from the character's 3D model information (for example, the mesh data mentioned above) and extract metadata that is characteristic of this part. It wasn't easy to do.
<1.2.提案技術の概要>
 そこで、本開示の提案技術に係る情報処理装置は、キャラクタの3Dモデルをレンダリングした画像を用いて画像認識処理を行い、画像認識処理の結果を用いて、分析を行う3Dモデルの絞り込みを行う。情報処理装置は、絞り込んだ3Dモデルに対して分析を行う。これにより、情報処理装置は、3Dモデルの分析をより容易に行うことができ、より容易に3Dモデルからパーツを分離することができる。
<1.2. Overview of proposed technology>
Therefore, the information processing device according to the proposed technology of the present disclosure performs image recognition processing using a rendered image of a 3D model of a character, and uses the result of the image recognition processing to narrow down the 3D models to be analyzed. The information processing device performs analysis on the narrowed down 3D models. Thereby, the information processing device can more easily analyze the 3D model and can more easily separate parts from the 3D model.
 図1は、本開示の実施形態に係る3Dモデルの分析処理の概要を説明するための図である。図1の分析処理は、例えば、情報処理装置100によって実行される。 FIG. 1 is a diagram for explaining an overview of a 3D model analysis process according to an embodiment of the present disclosure. The analysis process in FIG. 1 is executed by the information processing device 100, for example.
 情報処理装置100は、まず、キャラクタの3Dモデル情報(以下、3Dモデルとも記載する)を取得する(ステップS1)。情報処理装置100は、例えば、データベースからキャラクタの3Dモデルを取得する。3Dモデルは、例えば、上述したメッシュデータを含む。 The information processing device 100 first obtains 3D model information (hereinafter also referred to as 3D model) of a character (step S1). For example, the information processing device 100 acquires a 3D model of a character from a database. The 3D model includes, for example, the mesh data described above.
 次に、情報処理装置100は、取得した3Dモデルに基づいて、キャラクタのレンダリング(描画)を行い、キャラクタを仮想視点から見た画像を生成する(ステップS2)。 Next, the information processing device 100 performs rendering (drawing) of the character based on the acquired 3D model, and generates an image of the character viewed from a virtual viewpoint (step S2).
 情報処理装置100は、生成した画像に対してパーツの画像認識処理を行う(ステップS3)。これにより、情報処理装置100は、画像におけるパーツの位置を推定する。なお、情報処理装置100が画像認識処理に基づいて推定したパーツの位置を画像認識位置とも記載する。例えば、図1では、情報処理装置100は、画像に対して右目の画像認識処理を行い、右目を含む領域を画像認識位置として推定する。 The information processing device 100 performs part image recognition processing on the generated image (step S3). Thereby, the information processing device 100 estimates the position of the part in the image. Note that the position of a part estimated by the information processing apparatus 100 based on image recognition processing is also referred to as an image recognition position. For example, in FIG. 1, the information processing device 100 performs right eye image recognition processing on the image, and estimates an area including the right eye as the image recognition position.
 情報処理装置100は、画像認識処理に基づき、3Dモデルにおける画像認識位置を推定する(ステップS4)。例えば、情報処理装置100は、画像における画像認識位置に対応する3Dモデルの位置を、3Dモデルにおける画像認識位置として推定する。 The information processing device 100 estimates the image recognition position in the 3D model based on the image recognition process (step S4). For example, the information processing device 100 estimates the position of the 3D model corresponding to the image recognition position in the image as the image recognition position in the 3D model.
 情報処理装置100は、3Dモデルにおける画像認識位置に基づき、3Dモデルのパーツ分析を行い、3Dモデルにおけるパーツの領域を推定する(ステップS5)。情報処理装置100は、3Dモデルのメッシュデータのうち、パーツに該当するデータをパーツ領域の3Dモデルとして推定する。例えば、図1では、情報処理装置100は、右目に該当する頂点データ群を右目パーツの3Dモデルとして推定する。 The information processing device 100 performs part analysis of the 3D model based on the image recognition position in the 3D model, and estimates the region of the part in the 3D model (step S5). The information processing device 100 estimates data corresponding to a part from among the mesh data of the 3D model as a 3D model of the parts area. For example, in FIG. 1, the information processing apparatus 100 estimates the vertex data group corresponding to the right eye as a 3D model of the right eye part.
 情報処理装置100は、パーツのメタデータを抽出する(ステップS6)。例えば、情報処理装置100は、キャラクタの3Dモデル、画像、及び、パーツの3Dモデルに基づき、メタデータを抽出する。 The information processing device 100 extracts metadata of the parts (step S6). For example, the information processing device 100 extracts metadata based on a 3D model of a character, an image, and a 3D model of parts.
 情報処理装置100は、パーツ及びメタデータを対応付けて記憶する(ステップS7)。例えば、情報処理装置100は、ステップS5で領域を推定したパーツの3Dモデルと、ステップS6で抽出した当該パーツのメタデータと、を対応付けてデータベースに保存する。図1の例では、情報処理装置100は、パーツの3DモデルをパーツDB(Data Base)に保存し、パーツのメタデータをメタデータDBに保存する。 The information processing device 100 associates and stores parts and metadata (step S7). For example, the information processing device 100 associates the 3D model of the part whose region was estimated in step S5 with the metadata of the part extracted in step S6, and stores the 3D model in the database. In the example of FIG. 1, the information processing apparatus 100 stores a 3D model of a part in a parts DB (Data Base), and stores metadata of the part in the metadata DB.
 このように、情報処理装置100は、キャラクタの3Dモデル(3次元モデル情報の一例)を取得する。情報処理装置100は、3Dモデルに基づいて描画される画像であって、キャラクタを仮想視点から見た画像に対して画像認識処理を行って、キャラクタのパーツ位置を推定する。情報処理装置100は、3Dモデル及びパーツ位置に基づき、キャラクタの3Dモデルにおけるパーツ領域を推定する。 In this way, the information processing device 100 acquires a 3D model of a character (an example of 3D model information). The information processing device 100 performs image recognition processing on an image drawn based on a 3D model, in which the character is viewed from a virtual viewpoint, to estimate the positions of the character's parts. The information processing device 100 estimates the part area in the 3D model of the character based on the 3D model and the part position.
 これにより、情報処理装置100は、3Dモデルのうち、分析を行う3Dモデルを絞り込むことができ、3Dモデルの分析をより容易に行うことができるようになる。そのため、情報処理装置100は、キャラクタの3Dモデルをパーツごとに分離する処理の処理負荷をより低減することができる。また、情報処理装置100は、キャラクタの3Dモデルのパーツをより高精度に分離することができる。 Thereby, the information processing apparatus 100 can narrow down the 3D models to be analyzed from among the 3D models, and can more easily analyze the 3D models. Therefore, the information processing apparatus 100 can further reduce the processing load of separating the 3D model of the character into parts. Furthermore, the information processing apparatus 100 can separate parts of a 3D model of a character with higher precision.
 また、情報処理装置100は、分析処理によってキャラクタの3D形状分析を行うことで、キャラクタのパーツDB、及び、パーツの検索に使用するメタデータDBを効率よく作成することができる。ユーザは、情報処理装置100を利用することで、キャラクタの3Dモデルの素体をより効率的に作成することができるようになる。 Additionally, the information processing device 100 can efficiently create a character parts DB and a metadata DB used for searching parts by analyzing the 3D shape of the character through analysis processing. By using the information processing device 100, a user can more efficiently create a 3D model body of a character.
<<2.情報処理装置の構成例>>
 図2は、本開示の実施形態に係る情報処理装置100の構成例を示すブロック図である。本開示の実施形態に係る情報処理装置100は、自由度が高く処理が難しい3Dキャラクタの形状分析に対して、レンダリングした画像への画像処理を用いることで3Dモデル上の探索空間を狭めて3D特徴量分析を行う。これにより、情報処理装置100は、より簡単な処理で、処理負荷の少ない3Dモデルのパーツ領域推定やメタデータ抽出を行うことができる。
<<2. Configuration example of information processing device >>
FIG. 2 is a block diagram illustrating a configuration example of the information processing device 100 according to the embodiment of the present disclosure. The information processing apparatus 100 according to the embodiment of the present disclosure narrows the search space on the 3D model by using image processing on rendered images to analyze the shape of a 3D character that has a high degree of freedom and is difficult to process. Perform feature analysis. Thereby, the information processing apparatus 100 can perform part area estimation and metadata extraction of a 3D model with less processing load through simpler processing.
 図2に示す情報処理装置100は、通信部110と、入出力部120と、記憶部130と、制御部140と、を備える。 The information processing device 100 shown in FIG. 2 includes a communication section 110, an input/output section 120, a storage section 130, and a control section 140.
 情報処理装置100は、パーソナルコンピュータやタブレット端末のように、ユーザが使用する端末装置であってもよく、ネットワーク上に配置されるサーバ装置(例えば、クラウドサーバ装置又はローカルサーバ装置)であってもよい。 The information processing device 100 may be a terminal device used by a user, such as a personal computer or a tablet terminal, or may be a server device placed on a network (for example, a cloud server device or a local server device). good.
 図2では、情報処理装置100が、上述した分析処理等のアプリケーションを実行する制御部140と、パーツDB133やメタデータDB134を有し、ストレージとして機能する記憶部130と、の両方を有する。あるいは、記憶部130のストレージとしての機能など、一部の機能は、図2の情報処理装置100とは異なる情報処理装置(例えば、サーバ装置)によって実現されてもよい。 In FIG. 2, the information processing device 100 includes both a control unit 140 that executes applications such as the analysis processing described above, and a storage unit 130 that includes a parts DB 133 and a metadata DB 134 and functions as storage. Alternatively, some functions, such as the storage function of the storage unit 130, may be realized by an information processing device (for example, a server device) different from the information processing device 100 in FIG. 2.
 また、後述するように、図2の情報処理装置100は、キャラクタの3Dモデルの分析を行ってパーツを取得する取得機能と、パーツの検索を行う検索機能と、の両方の機能を有する。あるいは、検索機能は、取得機能を有する情報処理装置100とは異なる情報処理装置によって実現されてもよい。 Furthermore, as will be described later, the information processing device 100 in FIG. 2 has both an acquisition function that analyzes a 3D model of a character and acquires parts, and a search function that searches for parts. Alternatively, the search function may be realized by an information processing device different from the information processing device 100 having the acquisition function.
(通信部110)
 通信部110は、他の装置と通信するための通信インタフェースである。例えば、通信部110は、NIC(Network Interface Card)等のLAN(Local Area Network)インタフェースである。通信部110は、有線インタフェースであってもよいし、無線インタフェースであってもよい。通信部110は、制御部140の制御に従って他の装置と通信する。
(Communication Department 110)
Communication unit 110 is a communication interface for communicating with other devices. For example, the communication unit 110 is a LAN (Local Area Network) interface such as a NIC (Network Interface Card). Communication unit 110 may be a wired interface or a wireless interface. The communication unit 110 communicates with other devices under the control of the control unit 140.
(入出力部120)
 入出力部120は、ユーザと情報をやりとりするためのユーザインタフェースである。例えば、入出力部120は、キーボード、マウス、操作キー、タッチパネル等、ユーザが各種操作を行うための操作装置である。又は、入出力部120は、液晶ディスプレイ(Liquid Crystal Display)、有機ELディスプレイ(Organic Electroluminescence Display)等の表示装置である。入出力部120は、スピーカー、ブザー等の音響装置であってもよい。また、入出力部120は、LED(Light Emitting Diode)ランプ等の点灯装置であってもよい。
(Input/output unit 120)
The input/output unit 120 is a user interface for exchanging information with the user. For example, the input/output unit 120 is an operating device, such as a keyboard, a mouse, an operation key, a touch panel, etc., for the user to perform various operations. Alternatively, the input/output unit 120 is a display device such as a liquid crystal display (Liquid Crystal Display) or an organic EL display (Organic Electroluminescence Display). The input/output unit 120 may be an audio device such as a speaker or a buzzer. Further, the input/output unit 120 may be a lighting device such as an LED (Light Emitting Diode) lamp.
(記憶部130)
 記憶部130は、例えば、RAM(Random Access Memory)、ROM(Read Only Memory)、フラッシュメモリ(Flash Memory)等の半導体メモリ素子、または、ハードディスク、光ディスク等の記憶装置によって実現される。
(Storage unit 130)
The storage unit 130 is realized by, for example, a semiconductor memory element such as a RAM (Random Access Memory), a ROM (Read Only Memory), or a flash memory, or a storage device such as a hard disk or an optical disk.
 図2の記憶部130は、3DモデルDB131と、ログファイルDB132と、パーツDB133と、メタデータDB134と、を有する。 The storage unit 130 in FIG. 2 includes a 3D model DB 131, a log file DB 132, a parts DB 133, and a metadata DB 134.
 3DモデルDB131は、情報処理装置100が3D形状分析を行うキャラクタの3Dモデルを記憶するデータベースである。ログファイルDB132は、情報処理装置100による3D形状分析の分析結果を保持するログファイルを記憶するデータベースである。 The 3D model DB 131 is a database that stores 3D models of characters on which the information processing device 100 performs 3D shape analysis. The log file DB 132 is a database that stores log files that hold analysis results of 3D shape analysis performed by the information processing device 100.
 パーツDB133は、情報処理装置100が3D形状分析を行うことで取得したキャラクタのパーツ領域の3Dモデルを記憶するデータベースである。メタデータDB134は、パーツ領域に対応するメタデータを記憶するデータベースである。記憶部130は、パーツ領域の3Dモデル及びメタデータを対応付けて記憶する。 The parts DB 133 is a database that stores 3D models of character parts regions obtained by the information processing device 100 performing 3D shape analysis. The metadata DB 134 is a database that stores metadata corresponding to parts areas. The storage unit 130 stores the 3D model and metadata of the parts area in association with each other.
(制御部140)
 制御部140は、情報処理装置100の各部を制御するコントローラ(controller)である。制御部140は、例えば、CPU(Central Processing Unit)、MPU(Micro Processing Unit)、GPU(Graphics Processing Unit)等のプロセッサにより実現される。例えば、制御部140は、情報処理装置100内部の記憶装置に記憶されている各種プログラムを、プロセッサがRAM等を作業領域として実行することにより実現される。なお、制御部140は、ASIC(Application Specific Integrated Circuit)やFPGA(Field Programmable Gate Array)等の集積回路により実現されてもよい。CPU、MPU、GPU、ASIC、及びFPGAは何れもコントローラとみなすことができる。
(Control unit 140)
The control unit 140 is a controller that controls each unit of the information processing device 100. The control unit 140 is realized by, for example, a processor such as a CPU (Central Processing Unit), an MPU (Micro Processing Unit), or a GPU (Graphics Processing Unit). For example, the control unit 140 is realized by a processor executing various programs stored in a storage device inside the information processing device 100 using a RAM or the like as a work area. Note that the control unit 140 may be realized by an integrated circuit such as an ASIC (Application Specific Integrated Circuit) or an FPGA (Field Programmable Gate Array). CPUs, MPUs, GPUs, ASICs, and FPGAs can all be considered controllers.
 制御部140は、モデル取得部141と、レンダリング部142と、画像認識部143と、位置推定部144と、領域推定部145と、抽出部146と、検索処理部147と、UI制御部148と、を備える。制御部140は、モデル取得部141~抽出部146によって上述したキャラクタの3Dモデルの分析を行ってパーツを取得する取得機能(アプリケーション機能)を実現する。また、制御部140は、検索処理部147によってパーツの検索を行う検索機能(アプリケーション機能)を実現する。 The control unit 140 includes a model acquisition unit 141, a rendering unit 142, an image recognition unit 143, a position estimation unit 144, a region estimation unit 145, an extraction unit 146, a search processing unit 147, and a UI control unit 148. , is provided. The control unit 140 realizes an acquisition function (application function) in which the model acquisition unit 141 to the extraction unit 146 analyze the 3D model of the character described above and acquire parts. Further, the control unit 140 realizes a search function (application function) for searching for parts using the search processing unit 147.
 制御部140を構成する各ブロック(モデル取得部141~UI制御部148)はそれぞれ制御部140の機能を示す機能ブロックである。これら機能ブロックはソフトウェアブロックであってもよいし、ハードウェアブロックであってもよい。例えば、上述の機能ブロックが、それぞれ、ソフトウェア(マイクロプログラムを含む。)で実現される1つのソフトウェアモジュールであってもよいし、半導体チップ(ダイ)上の1つの回路ブロックであってもよい。勿論、各機能ブロックがそれぞれ1つのプロセッサ又は1つの集積回路であってもよい。制御部140は上述の機能ブロックとは異なる機能単位で構成されていてもよい。機能ブロックの構成方法は任意である。 Each block (model acquisition unit 141 to UI control unit 148) constituting the control unit 140 is a functional block indicating a function of the control unit 140. These functional blocks may be software blocks or hardware blocks. For example, each of the above functional blocks may be one software module realized by software (including a microprogram), or one circuit block on a semiconductor chip (die). Of course, each functional block may be one processor or one integrated circuit. The control unit 140 may be configured in functional units different from the above-mentioned functional blocks. The functional blocks can be configured in any way.
 なお、制御部140は上述の機能ブロックとは異なる機能単位で構成されていてもよい。また、制御部140を構成する各ブロック(モデル取得部141~UI制御部148)の一部又は全部の動作を、他の装置が行ってもよい。 Note that the control unit 140 may be configured in a functional unit different from the above-mentioned functional blocks. Further, some or all of the blocks (model acquisition unit 141 to UI control unit 148) constituting the control unit 140 may be operated by another device.
(モデル取得部141)
 モデル取得部141は、3DモデルDB131から3Dモデルを読み込むことで、キャラクタの3Dモデルを取得する。なお、モデル取得部141は、通信部110を介して他の装置からキャラクタの3Dモデルを取得し得る。モデル取得部141は、取得した3Dモデルをレンダリング部142、位置推定部144、及び、領域推定部145に出力する。
(Model acquisition unit 141)
The model acquisition unit 141 acquires a 3D model of the character by reading the 3D model from the 3D model DB 131. Note that the model acquisition unit 141 can acquire a 3D model of a character from another device via the communication unit 110. The model acquisition unit 141 outputs the acquired 3D model to the rendering unit 142, the position estimation unit 144, and the area estimation unit 145.
(レンダリング部142)
 図3は、本開示の実施形態に係るレンダリング部142によるレンダリング処理の一例を説明するための図である。レンダリング部142は、レンダリング処理を実行することで、3Dモデルに基づいたキャラクタの画像(2D画像)を生成する。
(Rendering unit 142)
FIG. 3 is a diagram for explaining an example of rendering processing by the rendering unit 142 according to the embodiment of the present disclosure. The rendering unit 142 generates a character image (2D image) based on the 3D model by executing rendering processing.
 図3に示すように、レンダリング部142は、仮想カメラCの仮想視点からキャラクタを見た画像を生成する。このとき、レンダリング部142は、仮想視点が異なる画像を複数生成し得る。 As shown in FIG. 3, the rendering unit 142 generates an image of the character viewed from the virtual viewpoint of the virtual camera C. At this time, the rendering unit 142 can generate a plurality of images with different virtual viewpoints.
 例えば、レンダリング部142は、キャラクタの周囲を所定の角度(例えば、30度又は45度など)ずつずらして配置した複数の仮想カメラC_1~C_Nそれぞれからキャラクタを見た画像P_1~P_Nを生成する。画像P_1~P_Nには、それぞれ向きが異なるキャラクタが含まれる。 For example, the rendering unit 142 generates images P_1 to P_N in which the character is viewed from each of a plurality of virtual cameras C_1 to C_N arranged around the character at a predetermined angle (for example, 30 degrees or 45 degrees). Images P_1 to P_N include characters with different orientations.
 レンダリング部142は、生成した画像P_1~P_Nを画像認識部143に出力する。このとき、レンダリング部142は、画像P_1~P_Nに対応する仮想カメラC_1~C_Nに関する情報を画像認識部143に出力するようにしてもよい。 The rendering unit 142 outputs the generated images P_1 to P_N to the image recognition unit 143. At this time, the rendering unit 142 may output information regarding the virtual cameras C_1 to C_N corresponding to the images P_1 to P_N to the image recognition unit 143.
(画像認識部143)
 図2に戻り、画像認識部143は、画像P_1~P_Nに対して画像認識処理を行い、画像P_1~P_Nに含まれるパーツ位置(画像認識位置)を推定する。
(Image recognition unit 143)
Returning to FIG. 2, the image recognition unit 143 performs image recognition processing on images P_1 to P_N to estimate parts positions (image recognition positions) included in images P_1 to P_N.
 画像認識部143は、例えば、予め指定されるパーツの、画像P_1~P_Nにおける位置を推定する。ここで、画像認識部143が推定するパーツ(パーツ領域)として、例えば、キャラクタの身体部位を含む領域が挙げられる。具体的に、このパーツ領域として、例えば、キャラクタの目領域、鼻領域、口領域、及び、耳領域が挙げられる。 For example, the image recognition unit 143 estimates the position of a pre-designated part in images P_1 to P_N. Here, the part (parts region) estimated by the image recognition unit 143 includes, for example, a region including a body part of the character. Specifically, the part areas include, for example, the character's eye area, nose area, mouth area, and ear area.
 また、パーツ領域として、指、掌、足などの身体部位を含む領域や、メガネや腕時計、イヤリング、ネックレスなどのアクセサリ、服や帽子などいった衣装などを含む領域が挙げられる。 In addition, the parts area includes an area including body parts such as fingers, palms, and feet, an area including accessories such as glasses, watches, earrings, and necklaces, and costumes such as clothes and hats.
 なお、情報処理装置100が推定するパーツは、ここで記載した例に限定されない。情報処理装置100は、例えば、ユーザが指定したパーツやキャラクタに特化したパーツなど、任意のパーツを推定し得る。 Note that the parts estimated by the information processing device 100 are not limited to the examples described here. The information processing device 100 can estimate any part, such as a part specified by a user or a part specialized for a character.
 画像認識部143は、画像Pに対して、パターン認識技術やセマンティックセグメンテーション技術を用いて画像認識処理を行い、画像P内でのパーツのピクセル座標を求める。画像認識部143は、例えば、画像P内での矩形状や多角形状の領域をパーツの画像認識位置として推定する。 The image recognition unit 143 performs image recognition processing on the image P using pattern recognition technology or semantic segmentation technology, and determines the pixel coordinates of parts within the image P. The image recognition unit 143 estimates, for example, a rectangular or polygonal area within the image P as the image recognition position of the part.
 画像認識部143は、画像Pにおける画像認識位置に関する情報を位置推定部144に出力する。また、画像認識部143は、画像認識処理の結果を抽出部146に出力する。このとき、画像認識部143は、画像Pに対応する仮想カメラCに関する情報を位置推定部144に出力し得る。 The image recognition unit 143 outputs information regarding the image recognition position in the image P to the position estimation unit 144. The image recognition unit 143 also outputs the result of the image recognition process to the extraction unit 146. At this time, the image recognition unit 143 can output information regarding the virtual camera C corresponding to the image P to the position estimation unit 144.
 また、画像認識部143は、画像Pからパーツを推定できなかった場合、画像認識に失敗した旨をログファイルに書き出す。このとき、画像認識部143は、例えば、推定できなかったパーツと画像Pとを対応付けてログファイルに書き出し得る。 Further, if the image recognition unit 143 cannot estimate a part from the image P, it writes a message to the effect that image recognition has failed in a log file. At this time, the image recognition unit 143 may associate the parts that could not be estimated with the image P and write them in a log file, for example.
(位置推定部144)
 位置推定部144は、画像認識部143による画像認識結果に基づき、3Dモデルでのパーツ位置(画像認識位置)を推定する。位置推定部144が推定する3Dモデルにおける画像認識位置は、実際に情報処理装置100が推定するパーツ領域よりも大まかな位置(領域)となる。
(Position estimation unit 144)
The position estimation unit 144 estimates the part position (image recognition position) in the 3D model based on the image recognition result by the image recognition unit 143. The image recognition position in the 3D model estimated by the position estimation unit 144 is a position (area) more roughly than the part area actually estimated by the information processing device 100.
 図4は、本開示の実施形態に係る位置推定部144によるパーツ位置推定の一例を示す図である。位置推定部144は、3Dモデルのレンダリングに使用した仮想カメラCの設定から逆算することで、3Dモデルにおけるパーツのおおよその位置を推定する。 FIG. 4 is a diagram illustrating an example of part position estimation by the position estimation unit 144 according to the embodiment of the present disclosure. The position estimating unit 144 estimates the approximate position of the part in the 3D model by calculating backwards from the settings of the virtual camera C used to render the 3D model.
 図4の上図に示すように、例えば、画像認識部143が、画像P_1において、右目をパーツとして、画像認識位置Rpを推定したとする。ここで、画像P_1は、仮想カメラC_1から見た3Dモデルをレンダリングした画像である。 As shown in the upper diagram of FIG. 4, for example, it is assumed that the image recognition unit 143 estimates the image recognition position Rp in the image P_1 using the right eye as a part. Here, the image P_1 is an image obtained by rendering a 3D model viewed from the virtual camera C_1.
 位置推定部144は、画像P_1における画像認識位置Rpの配置と、3D空間における仮想カメラC_1の仮想視点及び画角と、に基づき、図4の中図に示すように、3Dモデルにおける画像認識位置Rmを推定する。例えば、位置推定部144は、画像P_1における画像認識位置Rpを、3Dモデルの3D空間内に投影することで、3Dモデルにおける画像認識位置Rmを推定する。 The position estimating unit 144 calculates the image recognition position in the 3D model, as shown in the middle diagram of FIG. Estimate Rm. For example, the position estimation unit 144 estimates the image recognition position Rm in the 3D model by projecting the image recognition position Rp in the image P_1 into the 3D space of the 3D model.
 ここで、図4の下図は、3Dモデルにおける画像認識位置Rmの拡大図である。図4の下図に示すように、位置推定部144が推定する画像認識位置Rmは、3Dモデルにおいてパーツ(ここでは「目」)の大まかな位置(領域)を表している。そのため、画像認識位置Rmは、3Dモデルのメッシュと必ずしも一致しない。すなわち、3Dモデルにおける画像認識位置Rmの輪郭線は、3Dモデルにおけるメッシュのエッジと必ずしも一致しない。 Here, the lower diagram in FIG. 4 is an enlarged diagram of the image recognition position Rm in the 3D model. As shown in the lower diagram of FIG. 4, the image recognition position Rm estimated by the position estimating unit 144 represents the approximate position (area) of a part (here, the "eye") in the 3D model. Therefore, the image recognition position Rm does not necessarily match the mesh of the 3D model. That is, the contour line of the image recognition position Rm in the 3D model does not necessarily match the edge of the mesh in the 3D model.
 例えば、図4では、位置推定部144は、3Dモデルにおける画像認識位置Rmとして矩形領域を推定する。このように、位置推定部144は、3Dモデルにおけるパーツ位置を大まかに推定する。そのため、推定した位置(領域)は、実際のパーツ領域(例えば3Dモデルの面)と異なる場合がある。 For example, in FIG. 4, the position estimation unit 144 estimates a rectangular area as the image recognition position Rm in the 3D model. In this way, the position estimating unit 144 roughly estimates the position of the parts in the 3D model. Therefore, the estimated position (area) may differ from the actual part area (for example, a surface of a 3D model).
 図5は、本開示の実施形態に係る位置推定部144によるパーツ位置推定の他の例を示す図である。上述したように、画像認識部143は、パターンによる画像認識以外にも、セマンティックセグメンテーション技術を用いて、画像Pにおけるパーツ位置を推定し得る。図5では、画像認識部143がセマンティックセグメンテーションによってパーツ位置を推定した場合の、位置推定部144によるパーツ位置推定が示される。 FIG. 5 is a diagram illustrating another example of part position estimation by the position estimation unit 144 according to the embodiment of the present disclosure. As described above, the image recognition unit 143 can estimate the position of parts in the image P using semantic segmentation technology in addition to pattern-based image recognition. FIG. 5 shows part position estimation by the position estimation unit 144 when the image recognition unit 143 estimates the part position by semantic segmentation.
 例えば、画像認識部143は、図5の上図に示す画像Pにおいて、セマンティックセグメンテーションを用いた画像認識を行うことで、図5の中図に示すように、画像Pにおける右目の画像認識位置Rp1を推定する。画像認識部143は、セマンティックセグメンテーションを用いることで、図4の画像認識処理より、より細かい画像認識位置Rp1を推定することができる。 For example, the image recognition unit 143 performs image recognition using semantic segmentation on the image P shown in the upper diagram of FIG. Estimate. By using semantic segmentation, the image recognition unit 143 can estimate a more detailed image recognition position Rp1 than the image recognition process shown in FIG. 4 .
 位置推定部144は、画像Pにおける画像認識位置Rp1の配置と、3D空間における仮想カメラCの仮想視点及び画角と、に基づき、図5の下図に示すように、3Dモデルにおける画像認識位置Rm1を推定する。 The position estimation unit 144 calculates the image recognition position Rm1 in the 3D model based on the arrangement of the image recognition position Rp1 in the image P and the virtual viewpoint and angle of view of the virtual camera C in the 3D space, as shown in the lower diagram of FIG. Estimate.
 位置推定部144による3Dモデルにおける画像認識位置Rm1の推定方法は、図4の場合と同じであるが、画像認識部143が、画像認識位置Rp1を図4の画像認識位置Rpよりも細かく推定する。そのため、位置推定部144が推定する3Dモデルにおける画像認識位置Rm1が、図4の画像認識位置Rmよりも細かくなっている。 The method of estimating the image recognition position Rm1 in the 3D model by the position estimation unit 144 is the same as in the case of FIG. 4, but the image recognition unit 143 estimates the image recognition position Rp1 more finely than the image recognition position Rp of FIG. 4. . Therefore, the image recognition position Rm1 in the 3D model estimated by the position estimation unit 144 is finer than the image recognition position Rm in FIG. 4.
 ただし、上述したように、位置推定部144による3Dモデルにおける画像認識位置Rm1(パーツ位置)の推定は、大まかな位置(領域)を推定するものである。そのため、画像認識位置Rm1と3Dモデルのメッシュとは必ずしも一致するわけではない。また、画像認識部143による画像認識位置Rp1の推定結果によっては、3Dモデルにおける画像認識位置Rm1の輪郭が凸凹(ギザギザ)にがたつく恐れがある。 However, as described above, the estimation of the image recognition position Rm1 (part position) in the 3D model by the position estimation unit 144 is to estimate a rough position (area). Therefore, the image recognition position Rm1 and the mesh of the 3D model do not necessarily match. Further, depending on the estimation result of the image recognition position Rp1 by the image recognition unit 143, there is a possibility that the contour of the image recognition position Rm1 in the 3D model may become uneven (jagged).
 位置推定部144は、推定した3Dモデルにおける画像認識位置(パーツ位置)に関する情報を領域推定部145に出力する。 The position estimating unit 144 outputs information regarding the image recognition position (part position) in the estimated 3D model to the area estimating unit 145.
 また、位置推定部144が画像認識位置の推定に失敗した場合、位置推定部144は、例えば、その旨をログファイルに書き出す。例えば、位置推定部144は、画像認識部143の認識結果に基づき、3D空間内で推定した画像認識位置が3Dモデル上にない場合、画像認識位置の推定に失敗したと判定する。このように画像認識位置の推定に失敗した場合、位置推定部144は、例えば、キャラクタの3Dモデルと、パーツと、を対応付けてログファイルに書き出し得る。 Furthermore, if the position estimation unit 144 fails in estimating the image recognition position, the position estimation unit 144 writes a notification to that effect in a log file, for example. For example, based on the recognition result of the image recognition unit 143, the position estimation unit 144 determines that estimation of the image recognition position has failed when the image recognition position estimated in the 3D space is not on the 3D model. If estimation of the image recognition position fails in this way, the position estimating unit 144 may associate the 3D model of the character with the parts and write them in a log file, for example.
(領域推定部145)
 図2に戻り、領域推定部145は、3Dモデル、及び、位置推定部144が推定したパーツ位置Rm(画像認識位置Rm)に基づき、キャラクタの3Dモデルにおけるパーツ領域を推定する。領域推定部145は、パーツ領域として、キャラクタのパーツの3Dモデル(例えば、メッシュデータ)を推定し、抽出する。これにより、領域推定部145は、キャラクタの3Dモデルからパーツを分離する。
(Area estimation unit 145)
Returning to FIG. 2, the area estimation unit 145 estimates the part area in the 3D model of the character based on the 3D model and the part position Rm (image recognition position Rm) estimated by the position estimation unit 144. The region estimating unit 145 estimates and extracts a 3D model (for example, mesh data) of a character's parts as a parts region. Thereby, the area estimation unit 145 separates the parts from the 3D model of the character.
 上述したように、位置推定部144が推定するパーツ位置は、キャラクタの3Dモデルにおけるパーツの大まかな位置(領域)である。そのため、このパーツ位置は、実際の3Dモデルのメッシュで規定される領域と一致しなかったり、凹凸のあるぎざぎざした領域であったりする。このように、位置推定部144が推定するパーツ位置は、3Dモデルからパーツを分離するための十分な精度を有しているとは言えない。 As described above, the part position estimated by the position estimation unit 144 is the rough position (area) of the part in the 3D model of the character. Therefore, this part position may not match the area defined by the mesh of the actual 3D model, or may be a jagged area with unevenness. In this way, the parts positions estimated by the position estimation unit 144 cannot be said to have sufficient accuracy to separate the parts from the 3D model.
 そこで、本開示の実施形態に係る領域推定部145は、位置推定部144が推定したパーツ位置Rmに絞って、3Dモデルの分析(3D分析)を行って、キャラクタの3Dモデルにおけるパーツ領域を推定する。 Therefore, the region estimating section 145 according to the embodiment of the present disclosure performs analysis of the 3D model (3D analysis) focusing on the parts position Rm estimated by the position estimating section 144, and estimates the parts region in the 3D model of the character. do.
 図6は、本開示の実施形態に係る領域推定部145によるパーツ領域の推定処理の一例について説明するための図である。図6に示すように、領域推定部145は、位置推定部144が推定した画像認識位置Rm内の3Dモデルを分析し、3Dモデルのメッシュに沿ったパーツ領域Rrを推定する。 FIG. 6 is a diagram for explaining an example of part region estimation processing by the region estimating unit 145 according to the embodiment of the present disclosure. As shown in FIG. 6, the area estimating unit 145 analyzes the 3D model within the image recognition position Rm estimated by the position estimating unit 144, and estimates a part area Rr along the mesh of the 3D model.
 このとき、領域推定部145は、例えば、抽出対象とするパーツ(例えば、目など)に応じた3D形状の特徴を踏まえて分析を行うことで、パーツ領域の推定をより高精度に行うことができる。 At this time, the region estimating unit 145 can estimate the part region with higher accuracy by, for example, performing an analysis based on the characteristics of the 3D shape according to the part to be extracted (for example, eyes, etc.). can.
 例えば、領域推定部145は、パーツごとの特徴として、3D構造の特徴やパーツ属性の特徴を考慮して分析を行う。3D構造の特徴として、例えば、パーツの3Dモデルにおける曲率やグラディエン、ラブラシアンなどが挙げられる。パーツ属性の特徴として、体積比率(例えば、キャラクタ全体の体積に対するパーツの体積の比率)や、縦横比率(例えば、パーツの縦横比率)などが挙げられる。 For example, the region estimating unit 145 performs analysis by taking into account the characteristics of the 3D structure and the characteristics of the part attributes as the characteristics of each part. Features of the 3D structure include, for example, curvature, gradient, and Lavrasian in the 3D model of the part. Features of part attributes include volume ratio (for example, the ratio of the volume of the part to the volume of the entire character), and aspect ratio (for example, the aspect ratio of the part).
 また、領域推定部145は、キャラクタの3Dモデル全てを対象として分析を行うのではなく、位置推定部144が推定したパーツ位置を対象として分析を行い得る。これにより、領域推定部145は、分析の処理負荷をより低減することができる。 Furthermore, the region estimating unit 145 may perform analysis not on the entire 3D model of the character, but on the parts positions estimated by the position estimating unit 144. Thereby, the region estimating unit 145 can further reduce the processing load of analysis.
 領域推定部145は、3D分析により推定したパーツ領域Rrを、キャラクタのパーツの3Dモデルとして分離してもよいが、パーツ領域Rrを補正した補正パーツ領域Rcとしてもよい。 The region estimating unit 145 may separate the parts region Rr estimated by 3D analysis as a 3D model of the character's parts, but may also correct the parts region Rr as a corrected parts region Rc.
 図7は、本開示の実施形態に係る領域推定部145が行う補正処理の一例を説明するための図である。図8は、本開示の実施異形態に係る領域推定部145が行うがたつき検出の一例を説明するための図である。 FIG. 7 is a diagram for explaining an example of a correction process performed by the region estimation unit 145 according to the embodiment of the present disclosure. FIG. 8 is a diagram for explaining an example of rattling detection performed by the area estimating unit 145 according to an embodiment of the present disclosure.
 図7の左図には、領域推定部145が推定するパーツ領域Rrが示される。図7の左図の領域Aで示すように、領域推定部145が推定するパーツ領域Rrは、3Dモデルのメッシュ構造によって輪郭線ががたつく場合がある。 The left diagram in FIG. 7 shows the parts region Rr estimated by the region estimation unit 145. As shown by region A in the left diagram of FIG. 7, the outline of the part region Rr estimated by the region estimation unit 145 may be jittery due to the mesh structure of the 3D model.
 そこで、領域推定部145は、パーツ領域Rrのがたつき検出を行う。領域推定部145は、パーツ領域Rrの法線情報を用いてがたつき検出を行う。例えば、領域推定部145は、パーツ領域Rrの輪郭線の法線ベクトル(図8の矢印)を検出する。 Therefore, the region estimating unit 145 detects wobbling in the part region Rr. The region estimating unit 145 performs wobbling detection using the normal information of the part region Rr. For example, the region estimation unit 145 detects the normal vector (arrow in FIG. 8) of the outline of the part region Rr.
 図8の左図に示すように、パーツ領域Rrにがたつきがない場合、法線ベクトルは、パーツ領域Rrの輪郭に沿って、同じ方向に回転する。一方、図8の右図に示すように、パーツ領域Rrにがたつきがある場合、パーツ領域Rrの輪郭に沿って、法線ベクトルの向きが逆向きになる箇所がある。 As shown in the left diagram of FIG. 8, when there is no wobbling in the part region Rr, the normal vector rotates in the same direction along the outline of the part region Rr. On the other hand, as shown in the right diagram of FIG. 8, when there is wobbling in the parts region Rr, there are places along the outline of the parts region Rr where the direction of the normal vector is reversed.
 領域推定部145は、パーツ領域Rrの輪郭に沿って、輪郭線の法線ベクトルの向きを確認する。領域推定部145は、法線ベクトルの向きが略向きになる箇所を輪郭ががたつく箇所(がたつき箇所)として検出する。 The region estimation unit 145 checks the direction of the normal vector of the contour line along the contour of the part region Rr. The region estimating unit 145 detects a location where the direction of the normal vector is substantially the same as a location where the contour is jittery (wobbly location).
 領域推定部145は、検出したがたつき箇所を補正することで、補正パーツ領域Rcを生成する。例えば、領域推定部145は、図8に示すように、領域Aのがたつきを補正し、補正パーツ領域Rcを生成する。例えば、領域推定部145は、がたつき箇所に新たにエッジを作成することで、輪郭にがたつきのない補正パーツ領域Rcを生成する。 The region estimating unit 145 generates a corrected part region Rc by correcting the detected wobbling portion. For example, as shown in FIG. 8, the region estimation unit 145 corrects the wobbling of the region A and generates a corrected parts region Rc. For example, the region estimating unit 145 generates a corrected part region Rc without wobbling in the outline by creating a new edge at the wobbling location.
 このように、領域推定部145は、パーツ領域Rrの形状を、パーツ領域Rrの輪郭のがたつきに応じて補正する。 In this way, the region estimation unit 145 corrects the shape of the parts region Rr according to the wobbling of the outline of the parts region Rr.
 領域推定部145は、補正によって生成した補正パーツ領域Rcの3Dモデル(3Dモデル情報、例えば、メッシュデータ)を、キャラクタのパーツの3Dモデルとして分離する。領域推定部145は、例えば、キャラクタの3Dモデルから補正パーツ領域Rcの3Dモデルを生成することで、キャラクタをパーツに分離する。領域推定部145は、分離したパーツの3Dモデルに関する情報を抽出部146に出力する。 The region estimation unit 145 separates the 3D model (3D model information, for example, mesh data) of the corrected parts region Rc generated by the correction as a 3D model of the character's parts. The region estimation unit 145 separates the character into parts by, for example, generating a 3D model of the corrected parts region Rc from the 3D model of the character. The area estimation unit 145 outputs information regarding the 3D model of the separated part to the extraction unit 146.
 また、領域推定部145がパーツ領域Rrの推定に失敗した場合、領域推定部145は、例えば、その旨をログファイルに書き出す。例えば、領域推定部145は、3D分析を行った結果、パーツ領域Rrがないと判定した場合、パーツ領域Rrの推定に失敗したと判定する。このようにパーツ領域Rrの推定に失敗した場合、領域推定部145は、例えば、キャラクタの3Dモデルと、パーツと、を対応付けてログファイルに書き出し得る。 Furthermore, if the region estimating unit 145 fails in estimating the parts region Rr, the region estimating unit 145 writes a notification to that effect in a log file, for example. For example, if the region estimating unit 145 determines that there is no parts region Rr as a result of the 3D analysis, it determines that estimation of the parts region Rr has failed. If estimation of the parts region Rr fails in this way, the region estimating unit 145 may associate the 3D model of the character with the parts and write them in a log file, for example.
 なお、ここでは、領域推定部145が、パーツ領域Rrの法線ベクトルの方向の変化に応じてがたつきを検出し、パーツ領域Rrの形状の補正を行うとしたが、領域推定部145によるパーツ領域Rrの補正はこれに限定されない。 Note that here, it is assumed that the region estimation unit 145 detects wobbling according to a change in the direction of the normal vector of the parts region Rr and corrects the shape of the parts region Rr. Correction of the parts region Rr is not limited to this.
 例えば、領域推定部145が、機械学習を用いてパーツ領域Rrの補正を行うようにしてもよい。例えば、領域推定部145は、パーツ領域Rrを入力とし、補正パーツ領域Rcを出力とする学習済みの補正モデルを用いて、パーツ領域Rrの補正を行い得る。 For example, the region estimation unit 145 may correct the parts region Rr using machine learning. For example, the region estimating unit 145 can correct the parts region Rr using a learned correction model that receives the parts region Rr as an input and outputs the corrected parts region Rc.
(抽出部146)
 図2に戻り、抽出部146は、画像認識処理の結果、パーツ位置(画像認識位置Rm)の推定結果、及び、パーツ領域Rrの少なくとも1つに基づき、キャラクタに関するキャラクタ情報をメタデータとして取得する。
(Extraction unit 146)
Returning to FIG. 2, the extraction unit 146 acquires character information regarding the character as metadata based on at least one of the image recognition processing result, the estimation result of the parts position (image recognition position Rm), and the parts region Rr. .
 抽出部146が取得するメタデータには、画像認識部143による画像認識によって得られるメタデータ、及び、領域推定部145によって推定されたパーツ領域Rrに基づいて得られるメタデータが含まれる。 The metadata acquired by the extraction unit 146 includes metadata obtained by image recognition by the image recognition unit 143 and metadata obtained based on the parts region Rr estimated by the region estimation unit 145.
 抽出部146は、例えば、メタデータとして、分類情報、特徴量情報、及び、相対情報の少なくとも1つを抽出する。 The extraction unit 146 extracts, for example, at least one of classification information, feature amount information, and relative information as metadata.
 分類情報は、例えば、パーツ領域Rrをクラスに分類したクラス分類に関する情報を含む。特徴量情報は、パーツ領域Rrの特徴量ベクトル(言い換えると、特徴量を示す潜在空間)に関する情報を含む。相対情報は、複数のパーツ間の相対的な大きさや位置(相対位置)に関する情報、キャラクタとパーツとの相対的な大きさや位置(相対位置)などを示す情報を含む。 The classification information includes, for example, information regarding class classification in which the parts region Rr is classified into classes. The feature amount information includes information regarding the feature amount vector (in other words, the latent space indicating the feature amount) of the parts region Rr. The relative information includes information about the relative sizes and positions (relative positions) between a plurality of parts, information about the relative sizes and positions (relative positions) between a character and the parts, and the like.
 抽出部146は、画像認識結果に基づき、例えば、分類情報及び特徴量情報を抽出する。抽出部146は、例えばディープラーニングのクラス分類タスク等を用いて、画像Pから分類情報を抽出する。 The extraction unit 146 extracts, for example, classification information and feature amount information based on the image recognition result. The extraction unit 146 extracts classification information from the image P using, for example, a deep learning class classification task.
 例えば、抽出部146が画像認識結果に基づいて分類するクラスとして以下のクラスが挙げられる。なお、以下は一例であり、抽出部146が以下のクラス以外のクラスにパーツを分類するようにしてもよい。 For example, the following classes are classified by the extraction unit 146 based on the image recognition results. Note that the following is an example, and the extraction unit 146 may classify parts into classes other than the following classes.
 例えば、キャラクタの特徴として以下のクラスが挙げられる。
 ・Photo Real(PR)であるかNon Photo Real(NPR)であるか
 ・NPRの場合、描画がリアルか、デフォルメか
 ・NPRの場合、年代
 ・年齢、性別など
 ・性格
For example, the following classes can be cited as character characteristics.
・Whether it is Photo Real (PR) or Non Photo Real (NPR) ・For NPR, whether the drawing is realistic or deformed ・For NPR, age ・Age, gender, etc. ・Personality
 例えば、パーツの特徴として以下のクラスが挙げられる。
 ・色(例えば、目や髪、服などの色)
 ・形状(例えば、目の場合はたれ目やつり目など。髪の場合はショートやロングなど)
 ・ジャンル(例えば、アクセサリの場合、メガネやネックレスなど。衣装の場合、帽子やジャケットなど)
For example, the following classes can be cited as part characteristics.
・Color (e.g. the color of eyes, hair, clothes, etc.)
・Shape (for example, for eyes, sagging or slanted eyes; for hair, short or long, etc.)
・Genre (for example, accessories such as glasses and necklaces; costumes such as hats and jackets)
 また、抽出部146は、画像認識結果に基づき、Variational Auto EncoderやGenerative Adversarial Networkなどで作られる潜在空間上でのクラスタリングによって特徴量情報を抽出する。抽出部146は、特徴量情報を抽出することで、パーツのキャラクタに応じた類似度を測定する。特徴量情報を用いることで、情報処理装置100は、異なるキャラクタ(例えば、キャラクタ#1、#2。図示省略)の同一パーツ(例えば、顔や目、髪型など)の類似度を推定し得る。 Furthermore, the extraction unit 146 extracts feature information by clustering on a latent space created by Variational Auto Encoder, Generative Adversarial Network, etc., based on the image recognition results. The extraction unit 146 measures the degree of similarity according to the character of the part by extracting the feature amount information. By using the feature amount information, the information processing device 100 can estimate the degree of similarity between the same parts (for example, faces, eyes, hairstyles, etc.) of different characters (for example, characters #1 and #2, not shown).
 抽出部146は、パーツ領域Rrの3Dモデル、換言すると、3Dモデルの3D形状の分析結果に基づき、例えば、分類情報、特徴量情報、及び、相対情報を抽出する。 The extraction unit 146 extracts, for example, classification information, feature amount information, and relative information based on the 3D model of the part region Rr, in other words, the analysis result of the 3D shape of the 3D model.
 抽出部146は、例えば、画像認識結果と同様にして、3D分析結果から分類情報及び特徴量情報を抽出する。このとき、抽出部146が、例えばパーツ領域Rrに限定して分類情報及び特徴量情報を抽出し得る。 The extraction unit 146 extracts classification information and feature amount information from the 3D analysis results, for example, in the same manner as the image recognition results. At this time, the extraction unit 146 can extract classification information and feature amount information, for example, limited to the parts region Rr.
 例えば、抽出部146が3D分析結果に基づいて分類するクラスとして以下のクラスが挙げられる。なお、以下は一例であり、抽出部146が以下のクラス以外のクラスにパーツを分類するようにしてもよい。
 ・質感(ごつごつ、しわしわ、つるつるなど)
 ・詳細度(メッシュデータの頂点数が多い(ハイポリ)少ない(ローポリ)など)
For example, the following classes are classified by the extraction unit 146 based on the 3D analysis results. Note that the following is an example, and the extraction unit 146 may classify parts into classes other than the following classes.
・Texture (rugged, wrinkled, smooth, etc.)
・Level of detail (mesh data with many vertices (high poly), few (low poly), etc.)
 このように、抽出部146が、パーツ領域Rrの3Dモデルを用いてメタデータを抽出することで、抽出精度をより向上させることができ、かつ、処理負荷をより低減することができる。 In this way, by the extraction unit 146 extracting metadata using the 3D model of the parts region Rr, the extraction accuracy can be further improved and the processing load can be further reduced.
 また、上述したように、抽出部146は、画像認識結果を用いてメタデータ(例えば、分類情報及び特徴量情報)を抽出する。抽出部146は、画像認識結果から抽出したメタデータとあわせて、3D分析結果に基づくメタデータの抽出を行うことで、メタデータをより高精度に抽出することができる。 Furthermore, as described above, the extraction unit 146 extracts metadata (for example, classification information and feature amount information) using the image recognition results. The extraction unit 146 can extract metadata with higher precision by extracting metadata based on the 3D analysis results together with the metadata extracted from the image recognition results.
 例えば、抽出部146は、画像認識結果からキャラクタが「筋肉質な男性」であり、3Dモデルに「手の領域がある」とするメタデータ(分類情報)を抽出したとする。この場合、抽出部146は、このメタデータも含めて、パーツのクラス分類を行う。具体的には、例えば、抽出部146は、パーツ領域Rrの3Dモデル(メッシュデータ)、及び、画像認識結果から取得したメタデータをニューラルネットワークに入力することでパーツの分類情報を抽出する。 For example, assume that the extraction unit 146 extracts metadata (classification information) indicating that the character is a "muscular male" and that the 3D model "has a hand region" from the image recognition results. In this case, the extraction unit 146 classifies the parts, including this metadata. Specifically, for example, the extraction unit 146 extracts the classification information of the parts by inputting the 3D model (mesh data) of the parts region Rr and the metadata acquired from the image recognition results into a neural network.
 このように、抽出部146が、画像認識結果及び3D分析結果に基づいてメタデータを抽出することで、抽出部146は、より精度よくパーツのメタデータを抽出することができる。 In this way, by the extraction unit 146 extracting metadata based on the image recognition result and the 3D analysis result, the extraction unit 146 can extract the metadata of the part with higher accuracy.
 抽出部146は、キャラクタの3Dモデル、換言すると、3Dモデルの3D形状の分析結果に基づき、例えば、相対情報を抽出する。抽出部146は、例えば、特定のキャラクタの複数パーツ(例えば、右目及び左目、又は、顔及び目)の相対的な位置関係や相対的な大きさを相対情報として抽出する。例えば、抽出部146は、キャラクタの目が、頭に対して相対的にどの位置に、どのくらいのサイズで配置されるかを測定し、測定結果をメタデータとして抽出する。 The extraction unit 146 extracts, for example, relative information based on the 3D model of the character, in other words, the analysis result of the 3D shape of the 3D model. The extraction unit 146 extracts, for example, the relative positional relationship and relative size of multiple parts (for example, right eye and left eye, or face and eyes) of a specific character as relative information. For example, the extraction unit 146 measures the position and size of the character's eyes relative to the head, and extracts the measurement results as metadata.
 抽出部146は、パーツ及び抽出したメタデータをそれぞれ対応付けてパーツDB133及びメタデータDB134に保存する。なお、パーツは、領域推定部145がパーツDB133に保存するようにしてもよい。 The extraction unit 146 stores the parts and the extracted metadata in the parts DB 133 and metadata DB 134 in association with each other. Note that the region estimating unit 145 may store the parts in the parts DB 133.
 上述したように、抽出部146は、画像認識によってパーツのメタデータを抽出する。また、抽出部146は、画像認識結果を用いてメタデータの抽出範囲を絞り込んだ上で、3D形状分析によるパーツのメタデータ抽出を行う。これにより、抽出部146は、抽出処理の負荷をより低減することができる。また、抽出部146は、より高精度にメタデータを抽出することができる。 As described above, the extraction unit 146 extracts metadata of parts by image recognition. Further, the extraction unit 146 uses the image recognition results to narrow down the metadata extraction range, and then extracts the metadata of the part by 3D shape analysis. Thereby, the extraction unit 146 can further reduce the load of extraction processing. Further, the extraction unit 146 can extract metadata with higher accuracy.
 抽出部146が抽出したメタデータは、例えば、ユーザによるパーツの検索時に使用される。このように、抽出部146がメタデータを抽出し、パーツと対応付けてメタデータDB134に保存することで、ユーザは、所望のパーツをより高速により手軽に検索することができる。 The metadata extracted by the extraction unit 146 is used, for example, when a user searches for parts. In this manner, the extraction unit 146 extracts metadata, associates it with parts, and stores it in the metadata DB 134, allowing the user to search for desired parts faster and more easily.
(検索処理部147)
 検索処理部147は、ユーザが指示する検索条件に応じたメタデータに対応するパーツ(パーツ情報の一例)をユーザに提示する。検索処理部147は、ユーザが指示する検索条件に応じてメタデータDB134を検索し、検索結果に対応するパーツをユーザに提示する。このとき、検索処理部147は、例えば、パーツの3Dモデルに応じて、このパーツをレンダリングした2Dの画像をユーザに提示する。
(Search processing unit 147)
The search processing unit 147 presents the user with parts (an example of parts information) corresponding to the metadata according to the search conditions specified by the user. The search processing unit 147 searches the metadata DB 134 according to search conditions specified by the user, and presents parts corresponding to the search results to the user. At this time, the search processing unit 147 presents the user with a 2D rendered image of the part, for example, in accordance with the 3D model of the part.
 図9は、本開示の実施形態に係る検索処理の一例を説明するための図である。図9に示すように、検索処理部147は、ユーザからパーツの検索条件を受け付ける。検索処理部147は、例えば、入出力部120を介して検索条件を受け付ける。 FIG. 9 is a diagram for explaining an example of search processing according to the embodiment of the present disclosure. As shown in FIG. 9, the search processing unit 147 receives parts search conditions from the user. The search processing unit 147 receives search conditions via the input/output unit 120, for example.
 検索処理部147は、ユーザから受け付けた検索条件に応じたメタデータを指定してメタデータDB134を検索する。メタデータDB134は、パーツDB133に対して、検索処理部147から指定されたメタデータに対応するパーツを指定する。 The search processing unit 147 searches the metadata DB 134 by specifying metadata according to the search conditions accepted from the user. The metadata DB 134 specifies parts corresponding to the metadata specified by the search processing unit 147 to the parts DB 133.
 パーツDB133は、メタデータDB134から指定されたパーツを検索処理部147に通知する。検索処理部147は、パーツDB133から取得したパーツを検索結果としてユーザに提示する。 The parts DB 133 notifies the search processing unit 147 of the parts specified from the metadata DB 134. The search processing unit 147 presents the parts acquired from the parts DB 133 to the user as a search result.
 このように、検索処理部147が、メタデータDB134に保存されるメタデータを用いてパーツを検索することで、ユーザはより短時間でより手軽に検索することができる。 In this way, by the search processing unit 147 searching for parts using the metadata stored in the metadata DB 134, the user can search more easily and in a shorter time.
 検索処理部147は、検索UI画像を提示し、ユーザから検索条件を受け付ける。図10は、本開示の実施形態に係る検索UI画像の一例を示す図である。図10に示す検索UI画像は、例えば、検索処理部147の指示に基づき、UI制御部148によって生成される。 The search processing unit 147 presents a search UI image and accepts search conditions from the user. FIG. 10 is a diagram illustrating an example of a search UI image according to an embodiment of the present disclosure. The search UI image shown in FIG. 10 is generated by the UI control unit 148 based on an instruction from the search processing unit 147, for example.
 検索処理部147は、例えば、分類情報や相対情報に基づいてパーツを検索する場合、図10に示す検索UI画像をユーザに提示する。検索処理部147は、例えば、パーツが「目」である場合に、さらに分類情報(図10のワードやキーワードに相当)や相対情報(図10の大きさ指定に相当)を用いてパーツの絞り込む検索情報を、図10の検索UI画像を用いてユーザから受け付ける。 For example, when searching for parts based on classification information or relative information, the search processing unit 147 presents the search UI image shown in FIG. 10 to the user. For example, when the part is an "eye", the search processing unit 147 further narrows down the parts using classification information (corresponding to words and keywords in FIG. 10) and relative information (corresponding to size specification in FIG. 10). Search information is received from the user using the search UI image shown in FIG.
 ユーザは、例えばフリーワードの入力やタグ(図10では「アニメ」や「少女」等)の選択によって分類情報のクラスを指定し得る。また、ユーザは、スライダーによる数値調整でパーツの相対位置や相対的な大きさを指定し得る。例えば、図10の例では、スライダーを調整することで、ユーザは、顔の大きさに対する目の相対的な大きさを指定し得る。なお、ユーザは、相対情報をスライダーによる通知調整で指定してもよく、あるいは、直接数値を指示することで指定してもよい。 The user can specify the class of the classification information, for example, by inputting a free word or selecting a tag (such as "anime" or "girl" in FIG. 10). Furthermore, the user can specify the relative positions and relative sizes of parts by adjusting numerical values using sliders. For example, in the example of FIG. 10, by adjusting the slider, the user can specify the relative size of the eyes to the size of the face. Note that the user may specify the relative information by adjusting the notification using a slider, or may specify the relative information by directly specifying a numerical value.
 図10に示すように、ユーザが直接メタデータを指定することで、検索処理部147は検索条件を取得し得る。あるいは、ユーザが画像を指定することで、検索処理部147が検索条件を取得するようにしてもよい。この場合、検索処理部147は、ユーザが指定する画像から特徴量情報を抽出し、当該特徴量情報に基づいてパーツ検索を行う。 As shown in FIG. 10, the search processing unit 147 can obtain search conditions by the user directly specifying metadata. Alternatively, the search processing unit 147 may acquire search conditions by the user specifying an image. In this case, the search processing unit 147 extracts feature information from the image specified by the user, and performs a parts search based on the feature information.
 図11は、本開示の実施形態に係る検索処理部147による特徴量情報の抽出例を説明するための図である。図11では、ユーザが検索条件として画像S_0を指定したものとする。 FIG. 11 is a diagram for explaining an example of feature information extraction by the search processing unit 147 according to the embodiment of the present disclosure. In FIG. 11, it is assumed that the user has specified image S_0 as a search condition.
 この場合、検索処理部147は、例えば、画像S_0をエンコーダに入力する。エンコーダは、例えば、画像から特徴量ベクトル(特徴量を示す潜在空間)を抽出するものである。検索処理部147は、例えば、画像S_0をエンコーダに入力することで、画像S_0に対応する特徴量ベクトルV_0を抽出する。検索処理部147は、抽出した特徴量ベクトルV_0を用いて潜在空間内で、画像S_0に近いキャラクタ(又は、パーツ)の検索を行う。 In this case, the search processing unit 147 inputs, for example, image S_0 to the encoder. The encoder, for example, extracts a feature amount vector (latent space indicating a feature amount) from an image. For example, the search processing unit 147 extracts the feature vector V_0 corresponding to the image S_0 by inputting the image S_0 to an encoder. The search processing unit 147 searches for a character (or part) close to the image S_0 in the latent space using the extracted feature vector V_0.
 図12は、本開示の実施形態に係る検索処理部147による潜在空間内での検索の一例を説明するための図である。図12では、図示を簡略化するため、2次元の潜在空間を示しているが、実際の潜在空間は2次元以上の多次元空間である。 FIG. 12 is a diagram for explaining an example of a search in the latent space by the search processing unit 147 according to the embodiment of the present disclosure. Although FIG. 12 shows a two-dimensional latent space to simplify the illustration, the actual latent space is a multidimensional space with two or more dimensions.
 図12に示すように、検索処理部147は、画像S_0から抽出した特徴量ベクトルV_0を潜在空間にマッピングする。検索処理部147は、潜在空間内において、特徴量ベクトルV_0を含む検索範囲SR_0内に位置する特徴量ベクトルのうち代表的な特徴量ベクトルを、検索結果ベクトルとして選択する。 As shown in FIG. 12, the search processing unit 147 maps the feature vector V_0 extracted from the image S_0 to the latent space. The search processing unit 147 selects a representative feature vector from among the feature vectors located within the search range SR_0 that includes the feature vector V_0 in the latent space as a search result vector.
 例えば、検索処理部147は、潜在空間における距離や方向に応じて検索結果ベクトルを選択し得る。例えば、検索処理部147は、検索範囲SR_0内からランダムに検索結果ベクトルを選択し得る。図12の例では、検索処理部147は、特徴量ベクトルVc_01~Vc_04を検索結果ベクトルとして選択する。 For example, the search processing unit 147 can select a search result vector depending on the distance and direction in the latent space. For example, the search processing unit 147 may randomly select a search result vector from within the search range SR_0. In the example of FIG. 12, the search processing unit 147 selects feature vectors Vc_01 to Vc_04 as search result vectors.
 検索処理部147は、例えば、検索結果ベクトルを指定してメタデータDB134を検索することで、検索結果ベクトルに対応するパーツを検索結果としてパーツDB133から取得する。検索処理部147は、取得したパーツを検索結果としてユーザに提示する。 The search processing unit 147, for example, specifies the search result vector and searches the metadata DB 134, thereby acquiring the parts corresponding to the search result vector from the parts DB 133 as the search results. The search processing unit 147 presents the acquired parts to the user as a search result.
 図13は、本開示の実施形態に係る検索結果の画像の一例を示す図である。図13に示す検索結果の画像は、例えば、検索処理部147の指示に基づき、UI制御部148によって生成される。 FIG. 13 is a diagram illustrating an example of a search result image according to the embodiment of the present disclosure. The search result image shown in FIG. 13 is generated by the UI control unit 148 based on an instruction from the search processing unit 147, for example.
 図13の例では、検索処理部147は、検索結果の画像の中央にユーザが指定する画像S_0を表示する。また、検索処理部147は、画像S_0の周囲に検索結果であるパーツ(図13の例では上半身を示すパーツ)の2D画像Sc_01~Sc_04を表示する。 In the example of FIG. 13, the search processing unit 147 displays the image S_0 specified by the user in the center of the search result images. Furthermore, the search processing unit 147 displays 2D images Sc_01 to Sc_04 of parts (parts showing the upper body in the example of FIG. 13) that are search results around the image S_0.
 このように、検索処理部147は、潜在空間(特徴量ベクトル)を用いて検索を行うことで、例えばユーザが指定する画像と似た(潜在空間において所定の検索範囲内にある)パーツを検索することができる。 In this way, the search processing unit 147 searches for parts similar to the image specified by the user (within a predetermined search range in the latent space) by searching using the latent space (feature vector). can do.
 また、例えば、検索処理部147は、図13のアイコンIを用いて、ユーザから検索範囲の変更を受け付け得る。 Furthermore, for example, the search processing unit 147 can accept changes in the search range from the user using icon I in FIG. 13 .
 図14は、本開示の実施形態に係る検索処理部147における検索範囲の変更処理の一例を説明するための図である。検索処理部147は、ユーザからの検索範囲の変更指示として、アイコンIの移動を受け付ける。 FIG. 14 is a diagram for explaining an example of search range changing processing in the search processing unit 147 according to the embodiment of the present disclosure. The search processing unit 147 receives a movement of the icon I as an instruction to change the search range from the user.
 図14の上図に示すように、ユーザがアイコンIを矢印に示すように移動させたとする。図14の上図では、ユーザが移動させる前のアイコンIをアイコンI_0とし、移動後のアイコンIをアイコンI_1とする。 As shown in the upper diagram of FIG. 14, it is assumed that the user moves the icon I as shown by the arrow. In the upper diagram of FIG. 14, the icon I before being moved by the user is designated as icon I_0, and the icon I after being moved is designated as icon I_1.
 このようなアイコンIの移動を受け付けた検索処理部147は、アイコンIの移動に応じて検索範囲を変更してパーツの検索処理を行う。例えば、検索処理部147は、図14の下図に示すように、検索範囲SR_0から検索範囲SR_1に変更してパーツの検索を行う。 The search processing unit 147 that has received such a movement of the icon I changes the search range according to the movement of the icon I and performs a parts search process. For example, as shown in the lower diagram of FIG. 14, the search processing unit 147 searches for parts by changing the search range SR_0 to the search range SR_1.
 ここで、検索範囲SR_0は、ユーザが指定した画像S_0に対応する特徴量ベクトルV_0に応じた検索範囲であり、例えば、特徴量ベクトルV_0を中心とした範囲である。 Here, the search range SR_0 is a search range according to the feature vector V_0 corresponding to the image S_0 specified by the user, and is, for example, a range centered on the feature vector V_0.
 また、検索範囲SR_1は、アイコンIの移動に応じた特徴量ベクトルV_1に応じた検索範囲であり、例えば、特徴量ベクトルV_1を中心とした範囲である。特徴量ベクトルV_1は、特徴量ベクトルV_0を、アイコンIの移動量(図14上図の矢印の長さ)に応じて、特徴量ベクトルVc_02の方向に移動させたベクトルである。例えば、検索処理部147は、アイコンI_0と2D画像Sc02との間の距離に対するアイコンIの移動量の割合の分、特徴量ベクトルV_0を特徴量ベクトルVc_02の方向に移動させることで、特徴量ベクトルV_1を算出する。 Further, the search range SR_1 is a search range according to the feature vector V_1 according to the movement of the icon I, and is, for example, a range centered on the feature vector V_1. The feature vector V_1 is a vector obtained by moving the feature vector V_0 in the direction of the feature vector Vc_02 according to the amount of movement of the icon I (the length of the arrow in the upper diagram of FIG. 14). For example, the search processing unit 147 moves the feature vector V_0 in the direction of the feature vector Vc_02 by the ratio of the amount of movement of the icon I to the distance between the icon I_0 and the 2D image Sc02. Calculate V_1.
 検索処理部147は、変更後の検索範囲SR_1の中から検索結果ベクトルを選択する。検索処理部147は、検索範囲SR_0の中から検索結果ベクトルを選択する方法と同様にして、検索範囲SR_1の中から検索結果ベクトルを選択する。図14の例では、検索処理部147は、検索結果ベクトルとして、特徴量ベクトルVc_11~Vc_14を選択する。 The search processing unit 147 selects a search result vector from the changed search range SR_1. The search processing unit 147 selects a search result vector from within the search range SR_1 in the same manner as the method for selecting a search result vector from within the search range SR_0. In the example of FIG. 14, the search processing unit 147 selects feature vectors Vc_11 to Vc_14 as search result vectors.
 図15は、本開示の実施形態に係る検索結果の画像の他の例を示す図である。図15では、検索処理部147が検索範囲SR_1で検索を行った結果を示す画像が示される。 FIG. 15 is a diagram illustrating another example of a search result image according to the embodiment of the present disclosure. In FIG. 15, an image showing the results of the search performed by the search processing unit 147 in the search range SR_1 is shown.
 図15に示すように、検索処理部147は、検索結果ベクトルである特徴量ベクトルVc_11~Vc_14に対応するパーツの2D画像Sc_11~Sc_14をアイコンIの周辺に表示する。また、検索処理部147は、2D画像Sc_11~Sc_14に加え、ユーザが指定する画像S_0を表示するようにしてもよい。 As shown in FIG. 15, the search processing unit 147 displays 2D images Sc_11 to Sc_14 of parts corresponding to the feature vectors Vc_11 to Vc_14, which are search result vectors, around the icon I. Further, the search processing unit 147 may display the image S_0 specified by the user in addition to the 2D images Sc_11 to Sc_14.
 このように、検索処理部147は、例えば、アイコンIを用いてユーザから潜在空間内の検索範囲の変更を受け付ける。これにより、ユーザは、より直感的に潜在空間内の検索範囲を変更することができ、所望のパーツをより容易に検索することができる。 In this way, the search processing unit 147 receives a change in the search range within the latent space from the user using the icon I, for example. Thereby, the user can more intuitively change the search range in the latent space, and can more easily search for a desired part.
 なお、ここでは、検索処理部147が、アイコンを用いて、より具体的には、アイコンの移動に応じて、検索範囲の変更を受け付けるとしたが、検索範囲の変更方法はこれに限定されない。 Here, the search processing unit 147 accepts changes in the search range using icons, more specifically, according to the movement of the icon, but the method for changing the search range is not limited to this.
 例えば、ユーザが2D画像Sc_01~Sc_04をクリックすることで、検索処理部147が検索範囲を変更するようにしてもよい。この場合、検索処理部147は、例えば、クリックされた2D画像Scに対応する特徴量ベクトルVcに応じた検索範囲SR内で検索結果ベクトルを選択する。 For example, the search processing unit 147 may change the search range when the user clicks on the 2D images Sc_01 to Sc_04. In this case, the search processing unit 147 selects a search result vector within the search range SR according to the feature vector Vc corresponding to the clicked 2D image Sc, for example.
 あるいは、検索処理部147は、スライダーなどの数値を調整するツールを用いてユーザからの検索範囲の変更を受け付け得る。この場合、検索処理部147は、例えば、ユーザがスライダーを用いて指定する数値に応じて検索範囲SRを変更し、パーツの検索を行う。 Alternatively, the search processing unit 147 may accept a change in the search range from the user using a tool such as a slider to adjust numerical values. In this case, the search processing unit 147 changes the search range SR according to a numerical value specified by the user using a slider, for example, and searches for parts.
 また、ここでは、検索処理部147が、ユーザが指定する画像S_0を用いて潜在空間内での検索を行うとしたが、検索処理部147による潜在空間内での検索は、これに限定されない。 Furthermore, here, it is assumed that the search processing unit 147 searches in the latent space using the image S_0 specified by the user, but the search in the latent space by the search processing unit 147 is not limited to this.
 例えば、検索処理部147は、検索範囲SRに対応する特徴量ベクトルをランダムに選択し得る。例えば、ユーザがパーツを指定すると、検索処理部147は、指定されたパーツの特徴量ベクトルをランダムに1つ選択する。検索処理部147は、選択した特徴量ベクトルに対応する検索範囲を潜在空間内に設定し、設定した検索範囲内で検索結果ベクトルを選択する。 For example, the search processing unit 147 may randomly select a feature vector corresponding to the search range SR. For example, when the user specifies a part, the search processing unit 147 randomly selects one feature vector of the specified part. The search processing unit 147 sets a search range corresponding to the selected feature amount vector in the latent space, and selects a search result vector within the set search range.
 このように、検索処理部147は、ランダムにパーツを検索し、ユーザに提示し得る。 In this way, the search processing unit 147 can randomly search for parts and present them to the user.
 また、ここでは、検索処理部147が、検索結果ベクトルとして4つの特徴量ベクトルを選択するとしたが、検索処理部147が選択する特徴量ベクトルの数は4つに限定されない。例えば、検索処理部147は、検索結果ベクトルとして3つ以下の特徴量ベクトルを選択してもよく、また、5つ以上の特徴量ベクトルを選択してもよい。 Further, here, it is assumed that the search processing unit 147 selects four feature vectors as search result vectors, but the number of feature vectors selected by the search processing unit 147 is not limited to four. For example, the search processing unit 147 may select three or less feature vectors as the search result vector, or may select five or more feature vectors.
 また、ここでは、検索処理部147が、検索結果ベクトルとして選択した全ての特徴量ベクトルに対応する2D画像をユーザに提示するとしたが、ユーザに提示する2D画像はこれに限定されない。例えば、検索処理部147は、検索結果ベクトルとして選択した特徴量ベクトルに対応する2D画像の一部をユーザに提示するようにしてもよい。例えば、図14に示すように、検索処理部147が4つの特徴量ベクトルVc_11~Vc_14を選択した場合であっても、検索処理部147は3つ以下の2D画像をユーザに提示し得る。 Furthermore, here, it is assumed that the search processing unit 147 presents to the user 2D images corresponding to all the feature vectors selected as search result vectors, but the 2D images presented to the user are not limited to this. For example, the search processing unit 147 may present to the user a part of the 2D image corresponding to the feature vector selected as the search result vector. For example, as shown in FIG. 14, even if the search processing unit 147 selects four feature vectors Vc_11 to Vc_14, the search processing unit 147 can present three or less 2D images to the user.
(UI制御部148)
 図2に戻り、UI制御部148は、画面(UI)を生成し、UIへの操作を受け付ける。UI制御部148は、例えば、検索処理部147からの指示に従って検索UI画像や検索結果の画像を生成し、入出力部120を介してユーザに提示する。また、UI制御部148は、例えば、ユーザから入出力部120を介して検索条件の入力や検索範囲の変更を受け付ける。UI制御部148は、ユーザからの入力結果を例えば検索処理部147に通知する。
(UI control unit 148)
Returning to FIG. 2, the UI control unit 148 generates a screen (UI) and accepts operations on the UI. The UI control unit 148 generates a search UI image or a search result image according to an instruction from the search processing unit 147, for example, and presents them to the user via the input/output unit 120. Further, the UI control unit 148 receives input of search conditions and changes in the search range from the user via the input/output unit 120, for example. The UI control unit 148 notifies, for example, the search processing unit 147 of the input results from the user.
<<3.パーツ分離処理>>
(第1パーツ分離処理)
 図16は、本開示の実施形態に係る第1パーツ分離処理の一例の流れを示すフローチャートである。図16に示す第1パーツ分離処理は、例えば情報処理装置100によって実行される。情報処理装置100は、例えばユーザからの指示に従って第1パーツ分離処理を実行する。
<<3. Parts separation process >>
(First parts separation process)
FIG. 16 is a flowchart illustrating an example of the flow of the first parts separation process according to the embodiment of the present disclosure. The first parts separation process shown in FIG. 16 is executed by the information processing device 100, for example. The information processing device 100 executes the first parts separation process, for example, in accordance with an instruction from a user.
 図16に示すように、情報処理装置100は、まず、キャラクタの3Dモデルを取得する(ステップS101)。情報処理装置100は、例えば、3DモデルDB131から3Dモデルを取得する。なお、情報処理装置100は、ユーザが指定する範囲からキャラクタの3Dモデルを取得するようにしてもよい。 As shown in FIG. 16, the information processing device 100 first obtains a 3D model of a character (step S101). The information processing device 100 acquires a 3D model from the 3D model DB 131, for example. Note that the information processing apparatus 100 may acquire the 3D model of the character from a range specified by the user.
 情報処理装置100は、取得した3Dモデルを大分類パーツに分類する(ステップS102)。ここで、大分類パーツは、情報処理装置100が第1パーツ分離処理で分離するパーツよりも大きいパーツである。大分類パーツは、例えば、頭部領域や体領域を含む。あるいは、大分類パーツが、頭部領域、上半身領域及び下半身領域を含んでいてもよい。このように、情報処理装置100は、第1パーツ分離処理で分離するパーツ(例えば、目や鼻など)よりも大きい大分類パーツに3Dモデルを分割する。 The information processing device 100 classifies the acquired 3D model into major parts (step S102). Here, the major classification parts are parts that are larger than the parts that the information processing apparatus 100 separates in the first parts separation process. The major classification parts include, for example, a head region and a body region. Alternatively, the major classification parts may include a head region, an upper body region, and a lower body region. In this way, the information processing apparatus 100 divides the 3D model into large classification parts that are larger than the parts (for example, eyes, nose, etc.) to be separated in the first parts separation process.
 なお、大分類パーツは、第1パーツ分離処理で分離するパーツよりも大きい。したがって、情報処理装置100が3Dモデルを大分類パーツに分離する処理は、3Dモデルをパーツ(例えば、目や鼻など)に分離する処理と比較して処理負荷が小さくて済む。 Note that the major classification parts are larger than the parts to be separated in the first part separation process. Therefore, the process in which the information processing apparatus 100 separates a 3D model into major parts requires less processing load than the process in which the 3D model is separated into parts (for example, eyes, nose, etc.).
 次に、情報処理装置100は、分割した大分類パーツのうち、1つの大分類パーツを選択し、選択した大分類パーツをレンダリングして、画像Pを生成する(ステップS103)。 Next, the information processing device 100 selects one major classification part from among the divided major classification parts, renders the selected major classification part, and generates an image P (step S103).
 情報処理装置100は、ステップS103で生成した画像Pに対して、パーツの画像認識を行う(ステップS104)。情報処理装置100は、例えば、複数のパーツのうち、分離するパーツを1つ選択し、画像Pに対して選択したパーツの位置を推定する画像認識処理を実行する。 The information processing device 100 performs part image recognition on the image P generated in step S103 (step S104). For example, the information processing apparatus 100 selects one part to be separated from a plurality of parts, and executes image recognition processing to estimate the position of the selected part with respect to the image P.
 上述したように、情報処理装置100は、大分類パーツをレンダリングして画像Pを生成する。そのため、キャラクタの全体をレンダリングした画像の画像認識を行う場合と比較して、情報処理装置100は、より高精度に画像Pの認識を行うことができる。 As described above, the information processing device 100 generates the image P by rendering the major classification parts. Therefore, the information processing apparatus 100 can recognize the image P with higher accuracy than when performing image recognition of an image in which the entire character is rendered.
 情報処理装置100は、画像Pの認識に成功したか否かを判定する(ステップS105)。例えば、情報処理装置100は、パーツを認識できたか否かや、パーツの認識精度が閾値以上であるか否かに応じて画像Pの認識に成功したか否かを判定する。 The information processing device 100 determines whether or not recognition of the image P has been successful (step S105). For example, the information processing device 100 determines whether or not the recognition of the image P is successful depending on whether or not the parts can be recognized and whether or not the recognition accuracy of the parts is equal to or higher than a threshold value.
 画像Pの認識に失敗したと判定した場合(ステップS105;No)、すなわち、パーツを認識できなかった場合や認識精度が閾値未満であった場合、情報処理装置100は、ステップS110に進む。 If it is determined that recognition of the image P has failed (step S105; No), that is, if the part could not be recognized or the recognition accuracy is less than the threshold, the information processing device 100 proceeds to step S110.
 画像Pの認識に成功したと判定した場合(ステップS105;Yes)、すなわち、パーツを認識できた場合や認識精度が閾値以上であった場合、情報処理装置100は、3Dモデルにおける画像認識位置Rmを推定する(ステップS106)。情報処理装置100は、画像Pの認識結果から得られる画像認識位置Rp、及び、仮想カメラCの設定情報に応じて、3Dモデルにおける画像認識位置Rmを推定する。 When it is determined that the image P has been successfully recognized (step S105; Yes), that is, when the parts can be recognized or when the recognition accuracy is equal to or higher than the threshold, the information processing device 100 sets the image recognition position Rm in the 3D model. is estimated (step S106). The information processing device 100 estimates the image recognition position Rm in the 3D model according to the image recognition position Rp obtained from the recognition result of the image P and the setting information of the virtual camera C.
 情報処理装置100は、画像認識位置Rmに基づき、3Dモデルにおけるパーツ領域Rrを推定する(ステップS107)。情報処理装置100は、例えば、分離するパーツの特徴に応じてパーツ領域Rrを推定する。 The information processing device 100 estimates the parts region Rr in the 3D model based on the image recognition position Rm (step S107). For example, the information processing device 100 estimates the parts region Rr according to the characteristics of the parts to be separated.
 情報処理装置100は、パーツ領域Rr及び画像Pの画像認識結果に基づき、パーツ領域Rrに対応するメタデータを抽出する(ステップS108)。情報処理装置100は、パーツ領域Rrの3Dモデルを分離するパーツとする。 The information processing device 100 extracts metadata corresponding to the parts region Rr based on the image recognition results of the parts region Rr and the image P (step S108). The information processing device 100 separates the 3D model of the parts region Rr into parts.
 情報処理装置100は、パーツ及びメタデータを保存する(ステップS109)。情報処理装置100は、パーツ及びメタデータを対応付けてパーツDB133及びメタデータDB134にそれぞれ保存する。 The information processing device 100 stores the parts and metadata (step S109). The information processing device 100 associates parts and metadata and stores them in a parts DB 133 and a metadata DB 134, respectively.
 情報処理装置100は、ステップS103で選択した大分類パーツにおいて、全てのパーツを分離したか否かを判定する(ステップS110)。分離していないパーツが存在する場合(ステップS110;No)、情報処理装置100は、ステップS104に戻り、まだ分離していないパーツの分離処理を実行する。 The information processing device 100 determines whether all parts have been separated in the major classification parts selected in step S103 (step S110). If there are parts that have not been separated (step S110; No), the information processing device 100 returns to step S104 and executes separation processing for the parts that have not been separated yet.
 一方、全てのパーツを分離した場合(ステップS110;Yes)、情報処理装置100は、全ての大分類パーツでパーツを分離したか否かを判定する(ステップS111)。すなわち、情報処理装置100は、キャラクタの3Dモデルにおいて、全てのパーツの抽出を行ったか否かを判定する。 On the other hand, if all parts have been separated (step S110; Yes), the information processing device 100 determines whether all major classification parts have been separated (step S111). That is, the information processing device 100 determines whether all parts have been extracted in the 3D model of the character.
 まだパーツの分離を行っていない大分類パーツがある場合(ステップS111;No)、情報処理装置100は、ステップS103に戻り、パーツの分離を行っていない大分類パーツにおいて、パーツを分離する処理を行う。 If there are major classification parts that have not been separated yet (step S111; No), the information processing device 100 returns to step S103 and performs the process of separating parts for the major classification parts that have not been separated. conduct.
 一方、全ての大分類パーツにおいてパーツを分離した場合(ステップS111;Yes)、情報処理装置100は、全ての3Dモデルのパーツを分離したか否かを判定する(ステップS112)。すなわち、情報処理装置100は、全てのキャラクタにおいて、全てのパーツの抽出を行ったか否かを判定する。 On the other hand, if the parts have been separated in all major classification parts (step S111; Yes), the information processing device 100 determines whether the parts of all the 3D models have been separated (step S112). That is, the information processing device 100 determines whether all parts have been extracted for all characters.
 まだパーツの分離を行っていない3Dモデルがある場合(ステップS112;No)、情報処理装置100は、ステップS101に戻り、パーツの分離を行っていないキャラクタの3Dモデルを取得する。 If there is a 3D model whose parts have not been separated yet (step S112; No), the information processing device 100 returns to step S101 and obtains a 3D model of the character whose parts have not been separated.
 一方、全ての3Dモデルにおいてパーツを分離した場合(ステップS112;Yes)、情報処理装置100は、第1パーツ分離処理を終了する。 On the other hand, if parts have been separated in all 3D models (step S112; Yes), the information processing device 100 ends the first parts separation process.
 なお、ここでは、情報処理装置100が、3Dモデルを大分類パーツに分割してから、大分類パーツのレンダリングを行うとした。あるいは、情報処理装置100が、3Dモデルのレンダリングを行ってから、大分類パーツに分割するようにしてもよい。 Here, it is assumed that the information processing apparatus 100 divides the 3D model into major classification parts and then renders the major classification parts. Alternatively, the information processing device 100 may render the 3D model and then divide it into major parts.
 例えば、情報処理装置100は、3Dモデル全体のレンダリングを行い、キャラクタの全体像を含む画像を生成する。情報処理装置100は、例えば、このキャラクタの全体像を含む画像に対して、画像認識処理を行い、大分類パーツを含む領域を切り出して画像Pを生成する。あるいは、情報処理装置100は、画像認識処理によって大分類パーツを含む領域を推定し、推定した領域に対応する3Dモデルを再度レンダリングすることで画像Pを生成するようにしてもよい。 For example, the information processing device 100 renders the entire 3D model and generates an image including the entire character. For example, the information processing device 100 performs image recognition processing on the image including the entire image of the character, cuts out a region including the major classification parts, and generates the image P. Alternatively, the information processing device 100 may generate the image P by estimating a region including the major classification parts through image recognition processing and re-rendering a 3D model corresponding to the estimated region.
 また、情報処理装置100は、画像の認識、パーツ位置Rp、Rmの推定、及び、パーツ領域Rrの推定の少なくとも1つに失敗した場合、ログファイルにログを残し、次のパーツ又は次の3Dモデルの分析(パーツ分離)を実行し得る。 Furthermore, if at least one of image recognition, estimation of part positions Rp and Rm, and estimation of part region Rr fails, the information processing apparatus 100 leaves a log in a log file and performs next part or next 3D Analysis (parts separation) of the model can be performed.
 以上のように、情報処理装置100は、キャラクタの3Dモデル情報をレンダリングして2Dの画像を生成する。情報処理装置100は、生成した2Dの画像に対して分離対象であるパーツを認識する画像認識処理を行い、2Dの画像におけるパーツ位置Rpを推定する。 As described above, the information processing device 100 generates a 2D image by rendering the 3D model information of the character. The information processing device 100 performs image recognition processing to recognize parts to be separated on the generated 2D image, and estimates the part position Rp in the 2D image.
 情報処理装置100は、2Dの画像におけるパーツ位置Rpに基づき、3D空間(3Dモデル)における大まかなパーツ位置Rmを推定する。情報処理装置100は、3Dモデルにおける大まかなパーツ位置Rm、及び、パーツの特徴に応じて、3Dモデルの分析を行い、3Dモデルにおけるパーツ領域Rrを推定する。 The information processing device 100 estimates a rough part position Rm in the 3D space (3D model) based on the part position Rp in the 2D image. The information processing device 100 analyzes the 3D model according to the rough part position Rm in the 3D model and the characteristics of the part, and estimates the part region Rr in the 3D model.
 情報処理装置100は、例えば、キャラクタの3Dモデル情報からパーツ領域Rrの3Dモデル情報(例えば、メッシュデータ)を分離(生成)する。これにより、情報処理装置100は、キャラクタからパーツを分離する。 For example, the information processing device 100 separates (generates) 3D model information (for example, mesh data) of the parts region Rr from the 3D model information of the character. Thereby, the information processing device 100 separates the parts from the character.
 また、情報処理装置100は、パーツ領域Rr及び画像認識処理の結果を用いてパーツに対応するメタデータを抽出する。情報処理装置100は、パーツ及びメタデータを対応付けて保持する。 Additionally, the information processing device 100 extracts metadata corresponding to the part using the parts region Rr and the results of the image recognition process. The information processing device 100 stores parts and metadata in association with each other.
 これにより、情報処理装置100は、キャラクタからパーツを分離する処理の処理負荷をより低減しつつ、より高精度にキャラクタからパーツを分離することができる。また、情報処理装置100がパーツとメタデータとを紐付けて保持することで、ユーザは、より短時間でより精度よく所望のパーツを検索することができる。 Thereby, the information processing device 100 can separate parts from a character with higher accuracy while further reducing the processing load of the process of separating parts from a character. Furthermore, the information processing device 100 associates and holds parts and metadata, allowing the user to search for desired parts more accurately and in a shorter time.
(第2パーツ分離処理)
 上述した第1パーツ分離処理では、情報処理装置100が、全て自動でキャラクタからパーツを分離する処理を実行するとしたが、ユーザが処理の一部を実行するようにしてもよい。すなわち、情報処理装置100が、ユーザと対話しながらキャラクタからパーツを分離する処理(第2パーツ分離処理)を実行するようにしてもよい。
(Second parts separation process)
In the first parts separation process described above, the information processing apparatus 100 automatically performs the process of separating parts from the character, but the user may perform part of the process. That is, the information processing device 100 may perform a process of separating parts from a character (second parts separation process) while interacting with the user.
 図17は、本開示の実施形態に係る第2パーツ分離処理の一例の流れを示すフローチャートである。図17に示す第2パーツ分離処理は、例えば情報処理装置100によって実行される。情報処理装置100は、例えばユーザからの指示に従って、図17に示す第2パーツ分離処理を実行する。なお、図17に示す第2パーツ分離処理のうち、図16に示す第1パーツ分離処理と同じ処理については同一符号を付し説明を省略する。 FIG. 17 is a flowchart showing an example of the flow of the second parts separation process according to the embodiment of the present disclosure. The second parts separation process shown in FIG. 17 is executed by the information processing device 100, for example. The information processing apparatus 100 executes the second parts separation process shown in FIG. 17, for example, in accordance with an instruction from a user. Note that among the second parts separation processing shown in FIG. 17, the same processes as the first parts separation processing shown in FIG.
 図17に示すように、ステップS104でパーツの画像認識を行った情報処理装置100は、認識結果をユーザに提示する(ステップS201)。情報処理装置100は、認識結果をユーザに提示することで、認識結果の修正(変更)をユーザから受け付ける。 As shown in FIG. 17, the information processing device 100 that performed image recognition of the part in step S104 presents the recognition result to the user (step S201). The information processing device 100 receives a modification (change) of the recognition result from the user by presenting the recognition result to the user.
 図18は、本開示の実施形態に係る認識結果の修正例について説明するための図である。図18の上図に示すように情報処理装置100は、認識対象である画像Pに、認識結果であるパーツ位置Rpを重畳したUI画像PU_1をユーザに提示する。例えば、情報処理装置100の画像認識部143が、UI制御部148にUI画像PU_1を生成するよう指示する。このとき、UI制御部148は、パーツ位置Rpを、例えば、楕円や矩形などプリミティブな図形を用いて表示する。 FIG. 18 is a diagram for explaining an example of correction of recognition results according to the embodiment of the present disclosure. As shown in the upper diagram of FIG. 18, the information processing apparatus 100 presents the user with a UI image PU_1 in which the part position Rp, which is the recognition result, is superimposed on the image P, which is the recognition target. For example, the image recognition unit 143 of the information processing device 100 instructs the UI control unit 148 to generate the UI image PU_1. At this time, the UI control unit 148 displays the part position Rp using, for example, a primitive figure such as an ellipse or a rectangle.
 ユーザは、UI画像PU_1によって、情報処理装置100によるパーツの認識が正しいか否かを確認することができる。例えば、パーツ位置Rpが実際のキャラクタのパーツ位置とずれているなど、情報処理装置100によるパーツの認識が間違っている場合、ユーザは、パーツ位置Rpを修正する。ユーザは、例えば、ドラッグ&ドロップのようなGUI操作を行うことでパーツ位置Rpの修正を行う。これにより、図18の下図に示すように、ユーザは、情報処理装置100に対してパーツの正しい位置を指示することができる。 The user can check whether the part recognition by the information processing device 100 is correct using the UI image PU_1. For example, if the information processing device 100 recognizes the part incorrectly, such as when the part position Rp deviates from the actual character part position, the user corrects the part position Rp. The user corrects the part position Rp by, for example, performing a GUI operation such as drag and drop. Thereby, as shown in the lower diagram of FIG. 18, the user can instruct the information processing apparatus 100 about the correct position of the parts.
 なお、情報処理装置100が全くパーツを認識できなかった場合など、UI制御部148が、画像Pにパーツ位置Rpを示す図形を重畳したUI画像PU_1を生成できなかったとする。この場合、UI制御部148は、例えば、画像Pを含みパーツ位置Rpを示す図形を含まないUI画像をユーザに提示する。ユーザは、画像Pにパーツ位置Rpを示す図形を描画することで、情報処理装置100に対してパーツの正しい位置を指示する。 It is assumed that the UI control unit 148 is unable to generate the UI image PU_1 in which a figure indicating the part position Rp is superimposed on the image P, such as when the information processing device 100 is unable to recognize any parts. In this case, the UI control unit 148 presents the user with, for example, a UI image that includes the image P but does not include the figure indicating the part position Rp. The user instructs the information processing apparatus 100 about the correct position of the part by drawing a figure indicating the part position Rp on the image P.
 あるいは、UI制御部148が、例えば、画像Pの角など予め規定された位置(デフォルトの位置)にパーツ位置Rpを示す図形を描画したUI画像をユーザに提示するようにしてもよい。ユーザは、例えば、ドラッグ&ドロップのようなGUI操作を行うことでパーツ位置Rpの修正を行うことで、情報処理装置100に対してパーツの正しい位置を指示する。 Alternatively, the UI control unit 148 may present to the user a UI image in which a figure indicating the part position Rp is drawn at a predefined position (default position) such as a corner of the image P, for example. The user instructs the information processing apparatus 100 about the correct position of the part by, for example, correcting the part position Rp by performing a GUI operation such as drag and drop.
 図17に戻り、情報処理装置100は、ユーザが指示したパーツ位置Rpに基づき、3Dモデルにおける画像認識位置Rmの推定(ステップS106)及びパーツ領域Rrの推定(ステップS107)を行う。 Returning to FIG. 17, the information processing device 100 estimates the image recognition position Rm in the 3D model (step S106) and the parts region Rr (step S107) based on the part position Rp specified by the user.
 情報処理装置100は、パーツ領域Rrの推定結果をユーザに提示する(ステップS202)。情報処理装置100は、パーツ領域Rrの推定結果をユーザに提示することで、推定結果の修正(変更)をユーザから受け付ける。 The information processing device 100 presents the estimation result of the parts region Rr to the user (step S202). The information processing device 100 receives a modification (change) of the estimation result from the user by presenting the estimation result of the parts region Rr to the user.
 図19は、本開示の実施形態に係る推定結果を示すUI画像の一例について説明するための図である。図19に示すように情報処理装置100は、パーツ領域Rrを示すUI画像PU_2をユーザに提示する。例えば、情報処理装置100の領域推定部145が、UI制御部148にUI画像PU_2を生成するよう指示する。 FIG. 19 is a diagram for explaining an example of a UI image showing estimation results according to the embodiment of the present disclosure. As shown in FIG. 19, the information processing apparatus 100 presents the user with a UI image PU_2 showing the parts region Rr. For example, the area estimation unit 145 of the information processing device 100 instructs the UI control unit 148 to generate the UI image PU_2.
 図19に示すように、UI制御部148は、パーツ領域Rrを含む3Dモデルのレンダリング画像をUI画像PU_2として生成する。例えば、図19に示すUI画像PU_2は、パーツ領域Rrを正面から見たレンダリング画像である。 As shown in FIG. 19, the UI control unit 148 generates a rendered image of the 3D model including the parts region Rr as the UI image PU_2. For example, the UI image PU_2 shown in FIG. 19 is a rendered image of the parts region Rr viewed from the front.
 このとき、UI制御部148は、3Dモデルのメッシュに関する情報(例えば、頂点や辺を示す情報)を重畳してUI画像PU_2を生成する。このように、情報処理装置100がメッシュに関する情報を含めてパーツ領域Rrをユーザに提示することで、ユーザは3Dモデルにおけるパーツ領域Rrをより容易に確認することができる。 At this time, the UI control unit 148 generates the UI image PU_2 by superimposing information regarding the mesh of the 3D model (for example, information indicating vertices and edges). In this way, by the information processing device 100 presenting the parts region Rr including the information regarding the mesh to the user, the user can more easily confirm the parts region Rr in the 3D model.
 UI制御部148は、例えば、UI画像PU_2において、パーツ領域Rrを明るくハイライト表示するなど、パーツ領域Rrを強調表示し得る。UI制御部148は、UI画像PU_2においてパーツ領域Rr以外の領域を例えば暗くするなど、パーツ領域Rr以外の領域をパーツ領域Rrとは異なる表示色で表示するようにしてもよい。 The UI control unit 148 may highlight the parts region Rr in the UI image PU_2, for example by brightly highlighting the parts region Rr. The UI control unit 148 may display the area other than the parts area Rr in a display color different from that of the parts area Rr, such as by making the area other than the parts area Rr darker in the UI image PU_2.
 また、情報処理装置100は、複数の視点から見たパーツ領域Rrをユーザに提示するようにしてもよい。図20及び図21は、本開示の実施形態に係る推定結果を示すUI画像の他の例について説明するための図である。 Additionally, the information processing device 100 may present the parts region Rr viewed from a plurality of viewpoints to the user. 20 and 21 are diagrams for explaining other examples of UI images showing estimation results according to the embodiment of the present disclosure.
 情報処理装置100は、図20に示すように、UI画像PU_2とは異なる視点からパーツ領域Rrの3DモデルをレンダリングしたUI画像PU_3を生成し、ユーザに提示する。情報処理装置100は、図21に示すように、UI画像PU_2、PU_3とは異なる視点からパーツ領域Rrの3DモデルをレンダリングしたUI画像PU_4を生成し、ユーザに提示する。 As shown in FIG. 20, the information processing device 100 generates a UI image PU_3 that is a rendered 3D model of the parts region Rr from a different viewpoint than the UI image PU_2, and presents it to the user. As shown in FIG. 21, the information processing apparatus 100 generates a UI image PU_4, which is a 3D model of the parts region Rr rendered from a different viewpoint than the UI images PU_2 and PU_3, and presents it to the user.
 情報処理装置100は、UI画像PU_2~PU_3を並べてユーザに提示するようにしてもよい。あるいは、情報処理装置100は、ユーザからの指示に応じて、視点を変更したUI画像PU_3、PU_4を生成し、ユーザに提示し得る。 The information processing device 100 may present the UI images PU_2 to PU_3 side by side to the user. Alternatively, the information processing apparatus 100 may generate UI images PU_3 and PU_4 with changed viewpoints in response to instructions from the user, and present them to the user.
 なお、UI画像PU_3、PU_4の生成方法は、UI画像PU_2と同じである。 Note that the generation method of UI images PU_3 and PU_4 is the same as that of UI image PU_2.
 情報処理装置100は、ユーザからのパーツ領域Rrの修正を受け付ける。ユーザは、GUI操作を行うことでパーツ領域Rrの修正を行う。例えば、ユーザは、パーツ領域Rrに追加する面をクリック操作等により選択することで、パーツ領域Rrの修正を行う。あるいは、ユーザは、矩形や投げ縄(フリーハンドで描いた図形)などの範囲選択ツールにより複数の面を選択することで、パーツ領域Rrの修正を行い得る。 The information processing device 100 accepts corrections to the parts region Rr from the user. The user modifies the parts area Rr by performing GUI operations. For example, the user modifies the parts region Rr by selecting a surface to be added to the parts region Rr by a click operation or the like. Alternatively, the user can modify the part region Rr by selecting multiple faces using a range selection tool such as a rectangle or a lasso (a freehand drawn figure).
 なお、情報処理装置100が全くパーツ領域Rrを推定できなかった場合など、UI制御部148が、パーツ領域RrをUI画像PU_2として生成できなかったとする。この場合、UI制御部148は、例えば、キャラクタの3DモデルをレンダリングしたUI画像をユーザに提示する。UI制御部148は、大分類パーツをレンダリングしたUI画像をユーザに提示してもよい。ユーザは、UI画像に含まれる面をクリック操作や範囲選択ツールを用いて選択することで、情報処理装置100に対して正しいパーツ領域Rrを指示する。 Note that it is assumed that the UI control unit 148 is unable to generate the parts region Rr as the UI image PU_2, such as when the information processing device 100 is unable to estimate the parts region Rr at all. In this case, the UI control unit 148 presents the user with a UI image in which a 3D model of the character is rendered, for example. The UI control unit 148 may present the user with a UI image in which the major classification parts are rendered. The user instructs the information processing apparatus 100 about the correct part region Rr by selecting a surface included in the UI image using a click operation or a range selection tool.
 図17に戻り、情報処理装置100は、ユーザが指示したパーツ領域Rrに基づき、メタデータの抽出を行う(ステップS108)。以降の処理は、図16の第1パーツ分離処理と同じである。 Returning to FIG. 17, the information processing device 100 extracts metadata based on the parts region Rr specified by the user (step S108). The subsequent processing is the same as the first parts separation processing in FIG. 16.
 このように、情報処理装置100が、ユーザによるパーツ位置Rpの修正、及び、パーツ領域Rrの修正の少なくとも一方を受け付けることで、情報処理装置100は、より高精度にキャラクタからパーツを分離することができる。 In this way, the information processing device 100 can separate parts from the character with higher accuracy by accepting at least one of the correction of the parts position Rp and the correction of the parts region Rr by the user. I can do it.
(第3パーツ分離処理)
 上述した第2パーツ分離処理では、情報処理装置100は、全てのパーツにおいて、ユーザと対話しながら、すなわち、パーツ位置Rp及びパーツ領域Rrの推定結果をユーザに確認しながら処理を行った。これに対し、例えば、情報処理装置100は、パーツ位置Rp又はパーツ領域Rrの推定に失敗した場合に、ユーザによるパーツ位置Rpの修正、及び、パーツ領域Rrの修正の少なくとも一方を受け付けるようにしてもよい。なお、情報処理装置100は、パーツ位置Rp又はパーツ位置Rmの推定に失敗すると、パーツ領域Rrの推定を行わない。したがって、情報処理装置100が、パーツ位置Rp又はパーツ位置Rmの推定を失敗することは、パーツ領域Rrの推定を失敗することを意味する。
(Third parts separation process)
In the second parts separation process described above, the information processing apparatus 100 processed all parts while interacting with the user, that is, while confirming the estimation results of the part position Rp and the part region Rr with the user. On the other hand, for example, when the estimation of the parts position Rp or the parts region Rr fails, the information processing apparatus 100 accepts at least one of the correction of the parts position Rp and the correction of the parts region Rr by the user. Good too. Note that if the information processing device 100 fails to estimate the parts position Rp or the parts position Rm, it does not estimate the parts region Rr. Therefore, when the information processing apparatus 100 fails to estimate the part position Rp or the part position Rm, it means that the information processing apparatus 100 fails to estimate the parts region Rr.
 図22は、本開示の実施形態に係る第3パーツ分離処理の一例の流れを示すフローチャートである。図22に示す第3パーツ分離処理は、例えば情報処理装置100によって実行される。情報処理装置100は、例えばユーザからの指示に従って、図22に示す第3パーツ分離処理を実行する。 FIG. 22 is a flowchart showing an example of the flow of the third parts separation process according to the embodiment of the present disclosure. The third parts separation process shown in FIG. 22 is executed by the information processing device 100, for example. The information processing apparatus 100 executes the third parts separation process shown in FIG. 22, for example, in accordance with instructions from the user.
 図22に示すように、情報処理装置100は、第1パーツ分離処理を実行する(ステップS301)。このとき、情報処理装置100は、パーツ位置Rpの推定又はパーツ領域Rrの推定に失敗した3Dモデルに関する情報をログファイルに書き込む。 As shown in FIG. 22, the information processing device 100 executes a first parts separation process (step S301). At this time, the information processing apparatus 100 writes information regarding the 3D model for which estimation of the parts position Rp or parts region Rr has failed in the log file.
 次に、情報処理装置100は、ログファイルDB132からログファイルを取得する(ステップS302)。情報処理装置100は、パーツ位置Rpの推定又はパーツ領域Rrの推定に失敗した3Dモデルに対して第2パーツ分離処理を実行する(ステップS303)。 Next, the information processing device 100 obtains a log file from the log file DB 132 (step S302). The information processing device 100 performs the second parts separation process on the 3D model for which the estimation of the part position Rp or the estimation of the parts region Rr has failed (step S303).
 このように、情報処理装置100は、パーツの分離に失敗した3Dモデルに対して、第2パーツ分離処理を実行し、ユーザからの修正を受け付けることで、より高精度にキャラクタからパーツを分離することができる。 In this way, the information processing device 100 performs the second parts separation process on the 3D model whose parts have failed to be separated, and by accepting corrections from the user, separates parts from the character with higher precision. be able to.
 また、情報処理装置100は、全ての3Dモデルにおいてユーザからの修正を受け付けるのではなく、パーツの分離に失敗した3Dモデルに限定してユーザからの修正を受け付ける。これにより、情報処理装置100は、ユーザの負担増加を抑制しつつ、より高精度にパーツを分離することができる。 Furthermore, the information processing apparatus 100 does not accept corrections from the user for all 3D models, but only accepts corrections from the user for 3D models in which parts separation has failed. Thereby, the information processing apparatus 100 can separate parts with higher precision while suppressing an increase in the burden on the user.
<<4.ハードウェア構成>>
 図23は、本実施形態に係る情報処理装置100のハードウェア構成の一例を示すブロック図である。なお、図23に示す情報処理装置800は、例えば、情報処理装置100を実現し得る。本実施形態に係る情報処理装置100による情報処理は、ソフトウェアと、以下に説明するハードウェアとの協働により実現される。
<<4. Hardware configuration >>
FIG. 23 is a block diagram showing an example of the hardware configuration of the information processing device 100 according to this embodiment. Note that the information processing device 800 shown in FIG. 23 can realize the information processing device 100, for example. Information processing by the information processing apparatus 100 according to the present embodiment is realized by cooperation between software and hardware described below.
 図11に示すように、情報処理装置800は、例えば、CPU871と、ROM872と、RAM873と、ホストバス874と、ブリッジ875と、外部バス876と、インタフェース877と、を有する。また、情報処理装置800は、入力装置878と、出力装置879と、ストレージ880と、ドライブ881と、接続ポート882と、通信装置883と、を有する。なお、ここで示すハードウェア構成は一例であり、構成要素の一部が省略されてもよい。また、ここで示される構成要素以外の構成要素をさらに含んでもよい。 As shown in FIG. 11, the information processing device 800 includes, for example, a CPU 871, a ROM 872, a RAM 873, a host bus 874, a bridge 875, an external bus 876, and an interface 877. The information processing device 800 also includes an input device 878, an output device 879, a storage 880, a drive 881, a connection port 882, and a communication device 883. Note that the hardware configuration shown here is an example, and some of the components may be omitted. In addition, components other than those shown here may be further included.
(CPU871)
 CPU871は、例えば、演算処理装置又は制御装置として機能し、ROM872、RAM873、ストレージ880、又はリムーバブル記録媒体901に記録された各種プログラムに基づいて各構成要素の動作全般又はその一部を制御する。
(CPU871)
The CPU 871 functions, for example, as an arithmetic processing device or a control device, and controls the overall operation of each component or a portion thereof based on various programs recorded in the ROM 872, RAM 873, storage 880, or removable recording medium 901.
 具体的には、CPU871は、情報処理装置100内の動作処理を実現する。 Specifically, the CPU 871 implements operational processing within the information processing device 100.
(ROM872、RAM873)
 ROM872は、CPU871に読み込まれるプログラムや演算に用いるデータ等を格納する手段である。RAM873には、例えば、CPU871に読み込まれるプログラムや、そのプログラムを実行する際に適宜変化する各種パラメータ等が一時的又は永続的に格納される。
(ROM872, RAM873)
The ROM 872 is a means for storing programs read into the CPU 871, data used for calculations, and the like. The RAM 873 temporarily or permanently stores, for example, programs read into the CPU 871 and various parameters that change as appropriate when executing the programs.
(ホストバス874、ブリッジ875、外部バス876、インタフェース877)
 CPU871、ROM872、RAM873は、例えば、高速なデータ伝送が可能なホストバス874を介して相互に接続される。一方、ホストバス874は、例えば、ブリッジ875を介して比較的データ伝送速度が低速な外部バス876に接続される。また、外部バス876は、インタフェース877を介して種々の構成要素と接続される。
(Host bus 874, bridge 875, external bus 876, interface 877)
The CPU 871, ROM 872, and RAM 873 are interconnected, for example, via a host bus 874 capable of high-speed data transmission. On the other hand, the host bus 874 is connected, for example, via a bridge 875 to an external bus 876 whose data transmission speed is relatively low. Further, the external bus 876 is connected to various components via an interface 877.
(入力装置878)
 入力装置878には、例えば、マウス、キーボード、タッチパネル、ボタン、スイッチ、及びレバー等が用いられる。さらに、入力装置878としては、赤外線やその他の電波を利用して制御信号を送信することが可能なリモートコントローラ(以下、リモコン)が用いられることもある。また、入力装置878には、マイクロフォンなどの音声入力装置が含まれる。
(Input device 878)
The input device 878 includes, for example, a mouse, a keyboard, a touch panel, a button, a switch, a lever, and the like. Furthermore, as the input device 878, a remote controller (hereinafter referred to as remote control) that can transmit control signals using infrared rays or other radio waves may be used. Furthermore, the input device 878 includes an audio input device such as a microphone.
(出力装置879)
 出力装置879は、例えば、CRT(Cathode Ray Tube)、LCD、又は有機EL等のディスプレイ装置、スピーカー、ヘッドホン等のオーディオ出力装置、プリンタ、携帯電話、又はファクシミリ等、取得した情報を利用者に対して視覚的又は聴覚的に通知することが可能な装置である。また、本開示に係る出力装置879は、触覚刺激を出力することが可能な種々の振動デバイスを含む。
(Output device 879)
The output device 879 is, for example, a display device such as a CRT (Cathode Ray Tube), LCD, or organic EL, an audio output device such as a speaker or headphone, a printer, a mobile phone, or a facsimile, etc., for transmitting the acquired information to the user. This is a device that can notify visually or audibly. Further, the output device 879 according to the present disclosure includes various vibration devices capable of outputting tactile stimulation.
(ストレージ880)
 ストレージ880は、各種のデータを格納するための装置である。ストレージ880としては、例えば、ハードディスクドライブ(HDD)等の磁気記憶デバイス、半導体記憶デバイス、光記憶デバイス、又は光磁気記憶デバイス等が用いられる。
(Storage 880)
Storage 880 is a device for storing various data. As the storage 880, for example, a magnetic storage device such as a hard disk drive (HDD), a semiconductor storage device, an optical storage device, a magneto-optical storage device, or the like is used.
(ドライブ881)
 ドライブ881は、例えば、磁気ディスク、光ディスク、光磁気ディスク、又は半導体メモリ等のリムーバブル記録媒体901に記録された情報を読み出し、又はリムーバブル記録媒体901に情報を書き込む装置である。
(drive 881)
The drive 881 is a device that reads information recorded on a removable recording medium 901 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory, or writes information to the removable recording medium 901, for example.
(リムーバブル記録媒体901)
リムーバブル記録媒体901は、例えば、DVDメディア、Blu-ray(登録商標)メディア、HD DVDメディア、各種の半導体記憶メディア等である。もちろん、リムーバブル記録媒体901は、例えば、非接触型ICチップを搭載したICカード、又は電子機器等であってもよい。
(Removable recording medium 901)
The removable recording medium 901 is, for example, DVD media, Blu-ray (registered trademark) media, HD DVD media, various semiconductor storage media, and the like. Of course, the removable recording medium 901 may be, for example, an IC card equipped with a non-contact IC chip, an electronic device, or the like.
(接続ポート882)
 接続ポート882は、例えば、USB(Universal Serial Bus)ポート、IEEE1394ポート、SCSI(Small Computer System Interface)、RS-232Cポート、又は光オーディオ端子等のような外部接続機器902を接続するためのポートである。
(Connection port 882)
The connection port 882 is, for example, a port for connecting an external connection device 902 such as a USB (Universal Serial Bus) port, an IEEE1394 port, a SCSI (Small Computer System Interface), an RS-232C port, or an optical audio terminal. be.
(外部接続機器902)
 外部接続機器902は、例えば、プリンタ、携帯音楽プレーヤ、デジタルカメラ、デジタルビデオカメラ、又はICレコーダ等である。
(External connection device 902)
The external connection device 902 is, for example, a printer, a portable music player, a digital camera, a digital video camera, or an IC recorder.
(通信装置883)
 通信装置883は、ネットワークに接続するための通信デバイスであり、例えば、有線又は無線LAN、Wi-Fi(登録商標)、Bluetooth(登録商標)、又はWUSB(Wireless USB)用の通信カード、光通信用のルータ、ADSL(Asymmetric Digital Subscriber Line)用のルータ、又は各種通信用のモデム等である。
(Communication device 883)
The communication device 883 is a communication device for connecting to a network, and includes, for example, a communication card for wired or wireless LAN, Wi-Fi (registered trademark), Bluetooth (registered trademark), or WUSB (Wireless USB), optical communication. A router for ADSL (Asymmetric Digital Subscriber Line), a modem for various communications, etc.
<<5.まとめ>>
 以上、添付図面を参照しながら本開示の好適な実施形態について詳細に説明したが、本技術はかかる例に限定されない。本開示の技術分野における通常の知識を有する者であれば、請求の範囲に記載された技術的思想の範疇内において、各種の変更例または修正例に想到し得ることは明らかであり、これらについても、当然に本開示の技術的範囲に属するものと了解される。
<<5. Summary >>
Although preferred embodiments of the present disclosure have been described above in detail with reference to the accompanying drawings, the present technology is not limited to such examples. It is clear that a person with ordinary knowledge in the technical field of the present disclosure can come up with various changes or modifications within the scope of the technical idea described in the claims, and It is understood that these also naturally fall within the technical scope of the present disclosure.
 例えば、上述した情報処理装置100に内蔵されるCPU、ROM、およびRAM等のハードウェアに、情報処理装置100の機能を発揮させるためのコンピュータプログラムも作成可能である。また、当該コンピュータプログラムを記憶させたコンピュータ読み取り可能な記憶媒体(記録媒体)も提供される。 For example, it is also possible to create a computer program for causing hardware such as a CPU, ROM, and RAM built into the information processing device 100 described above to exhibit the functions of the information processing device 100. Furthermore, a computer-readable storage medium (recording medium) storing the computer program is also provided.
 また、本明細書に記載された効果は、あくまで説明的または例示的なものであって限定的ではない。つまり、本開示に係る技術は、上記の効果とともに、または上記の効果に代えて、本明細書の記載から当業者には明らかな他の効果を奏しうる。 Furthermore, the effects described in this specification are merely explanatory or illustrative, and are not limiting. In other words, the technology according to the present disclosure can have other effects that are obvious to those skilled in the art from the description of this specification, in addition to or in place of the above effects.
 なお、本技術は以下のような構成も取ることができる。
(1)
 キャラクタの3次元モデルを取得し、
 前記3次元モデルの仮想視点から見た画像に対して画像認識処理を行って、前記キャラクタのパーツ位置を推定し、
 前記3次元モデル、及び、前記パーツ位置に基づき、前記キャラクタの前記3次元モデルにおけるパーツ領域を推定する、制御部
 を備える情報処理装置。
(2)
 前記パーツ領域は、前記キャラクタの身体部位に応じた領域を含む(1)に記載の情報処理装置。
(3)
 前記パーツ領域は、前記キャラクタの目領域、鼻領域、口領域、及び、耳領域の少なくとも1つを含む、(1)又は(2)に記載の情報処理装置。
(4)
 前記制御部は、前記パーツ領域に基づき、前記キャラクタの前記3次元モデルから前記パーツ領域の3次元モデルを生成する、(1)~(3)のいずれか1つに記載の情報処理装置。
(5)
 前記制御部は、前記画像認識処理の結果、前記パーツ位置の推定結果、及び、前記パーツ領域の推定結果の少なくとも1つに基づき、前記キャラクタに関するキャラクタ情報を取得する、(1)~(4)のいずれか1つに記載の情報処理装置。
(6)
 前記キャラクタ情報は、前記パーツ領域をクラスに分類したクラス分類に関する分類情報、前記パーツ領域の特徴量ベクトルに関する特徴量情報、及び、前記キャラクタに対する前記パーツ領域の相対的な関係に関する相対情報、の少なくとも1つを含む、(5)に記載の情報処理装置。
(7)
 前記制御部は、前記キャラクタ情報を前記パーツ領域に関するパーツ情報と対応付けて記憶する、(5)又は(6)に記載の情報処理装置。
(8)
 前記制御部は、ユーザが指示する条件に応じた前記キャラクタ情報に対応する前記パーツ情報を、前記ユーザに提示する、(7)に記載の情報処理装置。
(9)
 前記制御部は、前記画像認識処理によって推定した前記画像における前記キャラクタの前記パーツ位置に基づき、前記3次元モデルにおける前記パーツ位置を推定する、(1)~(8)のいずれか1つに記載の情報処理装置。
(10)
 前記制御部は、前記画像に対応する前記仮想視点の画角、及び、前記画像における前記キャラクタの前記パーツ位置に基づき、前記3次元モデルにおける前記パーツ位置を推定する、(9)に記載の情報処理装置。
(11)
 前記制御部は、前記3次元モデルに含まれるメッシュデータにおける前記パーツ領域を推定する、(1)~(10)のいずれか1つに記載の情報処理装置。
(12)
 前記制御部は、前記3次元モデルにおける前記パーツ領域の形状を、前記パーツ領域の輪郭のがたつきに応じて補正する、(1)~(11)のいずれか1つに記載の情報処理装置。
(13)
 前記制御部は、ユーザから前記パーツ位置及び前記パーツ領域の少なくとも一方の変更を受け付ける、(1)~(12)のいずれか1つに記載の情報処理装置。
(14)
 前記制御部は、前記画像の前記画像認識処理の結果、画像認識に失敗した前記キャラクタにおける前記パーツ位置及び前記パーツ領域の少なくとも一方の前記変更を前記ユーザから受け付ける、(13)に記載の情報処理装置。
(15)
 コンピュータに
 キャラクタの3次元モデルを取得することと、
 前記3次元モデルの仮想視点から見た画像に対して画像認識処理を行って、前記キャラクタのパーツ位置を推定することと、
 前記3次元モデル、及び、前記パーツ位置に基づき、前記キャラクタの前記3次元モデルにおけるパーツ領域を推定することと、
 を実行させるためのプログラムを記録したコンピュータ読み取り可能な記録媒体。
(16)
 キャラクタの3次元モデルを取得することと、
 前記3次元モデルの仮想視点から見た画像に対して画像認識処理を行って、前記キャラクタのパーツ位置を推定することと、
 前記3次元モデル、及び、前記パーツ位置に基づき、前記キャラクタの前記3次元モデルにおけるパーツ領域を推定することと、
 を含む情報処理方法。
Note that the present technology can also have the following configuration.
(1)
Obtain a 3D model of the character,
performing image recognition processing on an image seen from a virtual viewpoint of the three-dimensional model to estimate the position of the character's parts;
An information processing device comprising: a control unit that estimates a part area in the three-dimensional model of the character based on the three-dimensional model and the part position.
(2)
The information processing device according to (1), wherein the parts area includes an area corresponding to a body part of the character.
(3)
The information processing device according to (1) or (2), wherein the parts area includes at least one of an eye area, a nose area, a mouth area, and an ear area of the character.
(4)
The information processing device according to any one of (1) to (3), wherein the control unit generates a three-dimensional model of the parts area from the three-dimensional model of the character based on the parts area.
(5)
(1) to (4), wherein the control unit obtains character information regarding the character based on at least one of a result of the image recognition process, a result of estimating the part position, and a result of estimating the part area. The information processing device according to any one of the above.
(6)
The character information includes at least classification information regarding class classification in which the parts area is classified into classes, feature information regarding a feature vector of the parts area, and relative information regarding the relative relationship of the parts area to the character. The information processing device according to (5), including one.
(7)
The information processing device according to (5) or (6), wherein the control unit stores the character information in association with parts information regarding the parts area.
(8)
The information processing device according to (7), wherein the control unit presents the user with the parts information corresponding to the character information according to conditions specified by the user.
(9)
According to any one of (1) to (8), the control unit estimates the part position in the three-dimensional model based on the part position of the character in the image estimated by the image recognition process. information processing equipment.
(10)
The information according to (9), wherein the control unit estimates the part position in the three-dimensional model based on the angle of view of the virtual viewpoint corresponding to the image and the part position of the character in the image. Processing equipment.
(11)
The information processing device according to any one of (1) to (10), wherein the control unit estimates the part area in mesh data included in the three-dimensional model.
(12)
The information processing device according to any one of (1) to (11), wherein the control unit corrects the shape of the part area in the three-dimensional model according to wobbling in the outline of the part area. .
(13)
The information processing device according to any one of (1) to (12), wherein the control unit receives a change in at least one of the part position and the part area from a user.
(14)
The information processing according to (13), wherein the control unit receives from the user the change of at least one of the part position and the part area of the character whose image recognition has failed as a result of the image recognition process of the image. Device.
(15)
Obtaining a 3D model of the character on the computer,
performing image recognition processing on an image seen from a virtual viewpoint of the three-dimensional model to estimate the positions of the parts of the character;
estimating a part area in the three-dimensional model of the character based on the three-dimensional model and the part position;
A computer-readable recording medium that records a program for executing.
(16)
Obtaining a three-dimensional model of the character;
performing image recognition processing on an image seen from a virtual viewpoint of the three-dimensional model to estimate the positions of the parts of the character;
estimating a part area in the three-dimensional model of the character based on the three-dimensional model and the part position;
Information processing methods including.
 100 情報処理装置
 110 通信部
 120 入出力部
 130 記憶部
 131 3DモデルDB
 132 ログファイルDB
 133 パーツDB
 134 メタデータDB
 140 制御部
 141 モデル取得部
 142 レンダリング部
 143 画像認識部
 144 位置推定部
 145 領域推定部
 146 抽出部
 147 検索処理部
 148 UI制御部
100 Information processing device 110 Communication unit 120 Input/output unit 130 Storage unit 131 3D model DB
132 Log file DB
133 Parts DB
134 Metadata DB
140 Control unit 141 Model acquisition unit 142 Rendering unit 143 Image recognition unit 144 Position estimation unit 145 Area estimation unit 146 Extraction unit 147 Search processing unit 148 UI control unit

Claims (16)

  1.  キャラクタの3次元モデルを取得し、
     前記3次元モデルの仮想視点から見た画像に対して画像認識処理を行って、前記キャラクタのパーツ位置を推定し、
     前記3次元モデル、及び、前記パーツ位置に基づき、前記キャラクタの前記3次元モデルにおけるパーツ領域を推定する、制御部
     を備える情報処理装置。
    Obtain a 3D model of the character,
    performing image recognition processing on an image seen from a virtual viewpoint of the three-dimensional model to estimate the position of the character's parts;
    An information processing device comprising: a control unit that estimates a part area in the three-dimensional model of the character based on the three-dimensional model and the part position.
  2.  前記パーツ領域は、前記キャラクタの身体部位に応じた領域を含む請求項1に記載の情報処理装置。 The information processing device according to claim 1, wherein the parts area includes an area corresponding to a body part of the character.
  3.  前記パーツ領域は、前記キャラクタの目領域、鼻領域、口領域、及び、耳領域の少なくとも1つを含む、請求項1に記載の情報処理装置。 The information processing device according to claim 1, wherein the parts area includes at least one of an eye area, a nose area, a mouth area, and an ear area of the character.
  4.  前記制御部は、前記パーツ領域に基づき、前記キャラクタの前記3次元モデルから前記パーツ領域の3次元モデルを生成する、請求項1に記載の情報処理装置。 The information processing device according to claim 1, wherein the control unit generates a three-dimensional model of the parts area from the three-dimensional model of the character based on the parts area.
  5.  前記制御部は、前記画像認識処理の結果、前記パーツ位置の推定結果、及び、前記パーツ領域の推定結果の少なくとも1つに基づき、前記キャラクタに関するキャラクタ情報を取得する、請求項1に記載の情報処理装置。 The information according to claim 1, wherein the control unit acquires character information regarding the character based on at least one of a result of the image recognition process, a result of estimating the part position, and a result of estimating the part area. Processing equipment.
  6.  前記キャラクタ情報は、前記パーツ領域をクラスに分類したクラス分類に関する分類情報、前記パーツ領域の特徴量ベクトルに関する特徴量情報、及び、前記キャラクタに対する前記パーツ領域の相対的な関係に関する相対情報、の少なくとも1つを含む、請求項5に記載の情報処理装置。 The character information includes at least classification information regarding class classification in which the parts area is classified into classes, feature information regarding a feature vector of the parts area, and relative information regarding the relative relationship of the parts area to the character. The information processing device according to claim 5, comprising one.
  7.  前記制御部は、前記キャラクタ情報を前記パーツ領域に関するパーツ情報と対応付けて記憶する、請求項5に記載の情報処理装置。 The information processing device according to claim 5, wherein the control unit stores the character information in association with parts information regarding the parts area.
  8.  前記制御部は、ユーザが指示する条件に応じた前記キャラクタ情報に対応する前記パーツ情報を、前記ユーザに提示する、請求項7に記載の情報処理装置。 The information processing device according to claim 7, wherein the control unit presents the part information corresponding to the character information to the user according to a condition instructed by the user.
  9.  前記制御部は、前記画像認識処理によって推定した前記画像における前記キャラクタの前記パーツ位置に基づき、前記3次元モデルにおける前記パーツ位置を推定する、請求項1に記載の情報処理装置。 The information processing device according to claim 1, wherein the control unit estimates the part position in the three-dimensional model based on the part position of the character in the image estimated by the image recognition process.
  10.  前記制御部は、前記画像に対応する前記仮想視点の画角、及び、前記画像における前記キャラクタの前記パーツ位置に基づき、前記3次元モデルにおける前記パーツ位置を推定する、請求項9に記載の情報処理装置。 The information according to claim 9, wherein the control unit estimates the part position in the three-dimensional model based on the angle of view of the virtual viewpoint corresponding to the image and the part position of the character in the image. Processing equipment.
  11.  前記制御部は、前記3次元モデルに含まれるメッシュデータにおける前記パーツ領域を推定する、請求項1に記載の情報処理装置。 The information processing device according to claim 1, wherein the control unit estimates the part area in mesh data included in the three-dimensional model.
  12.  前記制御部は、前記3次元モデルにおける前記パーツ領域の形状を、前記パーツ領域の輪郭のがたつきに応じて補正する、請求項1に記載の情報処理装置。 The information processing device according to claim 1, wherein the control unit corrects the shape of the part area in the three-dimensional model according to wobbling in the outline of the part area.
  13.  前記制御部は、ユーザから前記パーツ位置及び前記パーツ領域の少なくとも一方の変更を受け付ける、請求項1に記載の情報処理装置。 The information processing device according to claim 1, wherein the control unit receives a change in at least one of the part position and the part area from a user.
  14.  前記制御部は、前記パーツ領域の推定に失敗した前記キャラクタの前記パーツ位置及び前記パーツ領域の少なくとも一方の前記変更を前記ユーザから受け付ける、請求項13に記載の情報処理装置。 The information processing device according to claim 13, wherein the control unit receives from the user the change of at least one of the part position and the part area of the character whose part area has failed to be estimated.
  15.  コンピュータに
     キャラクタの3次元モデルを取得することと、
     前記3次元モデルの仮想視点から見た画像に対して画像認識処理を行って、前記キャラクタのパーツ位置を推定することと、
     前記3次元モデル、及び、前記パーツ位置に基づき、前記キャラクタの前記3次元モデルにおけるパーツ領域を推定することと、
     を実行させるためのプログラムを記録したコンピュータ読み取り可能な記録媒体。
    Obtaining a 3D model of the character on the computer,
    performing image recognition processing on an image seen from a virtual viewpoint of the three-dimensional model to estimate the positions of the parts of the character;
    estimating a part area in the three-dimensional model of the character based on the three-dimensional model and the part position;
    A computer-readable recording medium that records a program for executing.
  16.  キャラクタの3次元モデルを取得することと、
     前記3次元モデルの仮想視点から見た画像に対して画像認識処理を行って、前記キャラクタのパーツ位置を推定することと、
     前記3次元モデル、及び、前記パーツ位置に基づき、前記キャラクタの前記3次元モデルにおけるパーツ領域を推定することと、
     を含む情報処理方法。
    Obtaining a three-dimensional model of the character;
    performing image recognition processing on an image seen from a virtual viewpoint of the three-dimensional model to estimate the positions of the parts of the character;
    estimating a part area in the three-dimensional model of the character based on the three-dimensional model and the part position;
    Information processing methods including.
PCT/JP2023/010101 2022-03-29 2023-03-15 Information processing device, recording medium, and information processing method WO2023189601A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2022-052947 2022-03-29
JP2022052947 2022-03-29

Publications (1)

Publication Number Publication Date
WO2023189601A1 true WO2023189601A1 (en) 2023-10-05

Family

ID=88201569

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2023/010101 WO2023189601A1 (en) 2022-03-29 2023-03-15 Information processing device, recording medium, and information processing method

Country Status (1)

Country Link
WO (1) WO2023189601A1 (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2019053369A (en) * 2017-09-13 2019-04-04 ファナック株式会社 Three-dimensional model forming device
JP2019168251A (en) * 2018-03-22 2019-10-03 株式会社Jvcケンウッド Shape measuring apparatus, shape measuring method, and program

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2019053369A (en) * 2017-09-13 2019-04-04 ファナック株式会社 Three-dimensional model forming device
JP2019168251A (en) * 2018-03-22 2019-10-03 株式会社Jvcケンウッド Shape measuring apparatus, shape measuring method, and program

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
MIKA INOMAKI, AND 2 OTHERS: "Automatic setting of ability values considering physical characteristics in automatic generation of game characters", JAPAN DIGITAL GAME SOCIETY 9TH ANNUAL CONFERENCE PROCEEDINGS; 2019/03/03-04, DIGRA, JP, 4 March 2019 (2019-03-04) - 4 March 2019 (2019-03-04), JP, pages 146 - 149, XP009550245 *

Similar Documents

Publication Publication Date Title
US11354825B2 (en) Method, apparatus for generating special effect based on face, and electronic device
US20210073953A1 (en) Method for applying bokeh effect to image and recording medium
US8976182B2 (en) Facial sketch creation device, configuration information generation device, configuration information generation method, and storage medium
US11574424B2 (en) Augmented reality map curation
US20110292051A1 (en) Automatic Avatar Creation
JP2011090466A (en) Information processing apparatus, method, and program
US10713471B2 (en) System and method for simulating facial expression of virtual facial model
US9495789B2 (en) Information processing apparatus, information processing method and computer program
WO2020019664A1 (en) Deformed image generation method and apparatus based on human face
EP3591580A1 (en) Method and device for recognizing descriptive attributes of appearance feature
KR102546016B1 (en) Systems and methods for providing personalized video
KR20220063256A (en) Method and device for controlling the cabin environment
CN111428671A (en) Face structured information identification method, system, device and storage medium
US20230237777A1 (en) Information processing apparatus, learning apparatus, image recognition apparatus, information processing method, learning method, image recognition method, and non-transitory-computer-readable storage medium
US11341675B2 (en) Image processing apparatus, image processing method, and image processing program
JP6914724B2 (en) Information processing equipment, information processing methods and programs
JP7238963B2 (en) Information processing device, search method, and program
WO2023189601A1 (en) Information processing device, recording medium, and information processing method
CN112150347A (en) Image modification patterns learned from a limited set of modified images
JP2009009296A (en) Face image deformation method and face image deformation device implementing this method
US20220284701A1 (en) Side by side image detection method and electronic apparatus using the same
US11308150B2 (en) Mobile device event control with topographical analysis of digital images inventors
JP7385416B2 (en) Image processing device, image processing system, image processing method, and image processing program
CN111368674B (en) Image recognition method and device
JP6419560B2 (en) Search device, method and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23779619

Country of ref document: EP

Kind code of ref document: A1