WO2022083389A1 - Virtual image generation method and apparatus - Google Patents

Virtual image generation method and apparatus Download PDF

Info

Publication number
WO2022083389A1
WO2022083389A1 PCT/CN2021/119751 CN2021119751W WO2022083389A1 WO 2022083389 A1 WO2022083389 A1 WO 2022083389A1 CN 2021119751 W CN2021119751 W CN 2021119751W WO 2022083389 A1 WO2022083389 A1 WO 2022083389A1
Authority
WO
WIPO (PCT)
Prior art keywords
shoe
image
leg
dimensional
area
Prior art date
Application number
PCT/CN2021/119751
Other languages
French (fr)
Chinese (zh)
Inventor
车广富
郭景昊
张夏杰
安山
Original Assignee
北京沃东天骏信息技术有限公司
北京京东世纪贸易有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京沃东天骏信息技术有限公司, 北京京东世纪贸易有限公司 filed Critical 北京沃东天骏信息技术有限公司
Publication of WO2022083389A1 publication Critical patent/WO2022083389A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/005General purpose rendering architectures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/10Geometric effects
    • G06T15/20Perspective computation
    • G06T15/205Image-based rendering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality

Definitions

  • the present disclosure relates to the field of computer technologies, and in particular, to a method for generating a virtual image, a device for generating a virtual image, and a non-volatile computer-readable storage medium.
  • VR Virtual Reality, Virtual Reality
  • AR Augmented Reality, Augmented Reality
  • the virtual shoe try-on technology realized by the combination of AR augmented reality technology and smartphone camera can help users see the effect of shoes on their feet.
  • the virtual shoe trial technology needs to perform virtual occlusion on the three-dimensional shoe model to replace the real shoes of human feet.
  • a virtual leg model is modeled outside the shoe mouth area of the 3D shoe model, and the virtual leg model and the 3D shoe model are rendered on the screen to generate a virtual image of a human foot wearing shoes.
  • a method for generating a virtual image including: acquiring a leg area in an image to be processed including legs and feet; Internal parameters, rendering the 3D shoe model as a 2D shoe image corresponding to the image to be processed; according to the overlapping part of the leg area and the shoe area in the 2D shoe image, determine the part of the shoe area that is occluded by the leg; The partially occluded part is rendered, and the composite image of the to-be-processed image and the two-dimensional shoe image is rendered to generate a virtual image with a leg occlusion effect.
  • determining the portion of the shoe region occluded by the leg according to the overlapping portion of the leg region and the shoe region in the two-dimensional shoe image includes: determining the two-dimensional shoe image according to the position of the shoe body region in the two-dimensional shoe image For the outer contour of the middle shoe, the inner contour of the shoe in the two-dimensional shoe image is determined according to the position of the shoe mouth area in the two-dimensional image; according to the intersection of the contour of the leg area and the outer contour and the inner contour, it is determined that the shoe area is blocked by the leg. part.
  • determining the portion of the shoe region that is occluded by the leg according to the intersection of the contour of the leg region and the outer contour and the inner contour includes: on the inner contour, determining a point closest to the intersection; The closest point that determines the portion of the shoe area that is occluded by the leg.
  • rendering the three-dimensional shoe model into a two-dimensional shoe image corresponding to the image to be processed includes: performing transparency processing on the shoe opening area in the three-dimensional shoe model; rendering the transparently processed three-dimensional shoe model as Two-dimensional shoe image; according to the binary image of the two-dimensional shoe image, the shoe body area of the two-dimensional shoe image and the shoe mouth area of the two-dimensional shoe image are determined.
  • the transparent processing of the shoe opening area in the three-dimensional shoe model includes: detecting the shoe opening area of the three-dimensional shoe model, and using a closed mesh to cover the shoe opening area of the three-dimensional shoe model; covering part of the closed mesh Transparency is performed.
  • acquiring the leg region in the to-be-processed image including the legs and feet includes: inputting the to-be-processed image into a machine learning model to determine the leg region in the to-be-processed image.
  • the machine learning model includes a convolutional neural network module and a spatial pyramid pooling module connected in sequence.
  • the convolutional neural network module is set according to the Fast-SCNN (Fast Segmentation Convolutional Neural Network, fast segmentation convolutional neural network) model.
  • Fast-SCNN Fast Segmentation Convolutional Neural Network, fast segmentation convolutional neural network
  • the image to be processed is each frame of images in the video
  • generating a virtual image with a leg occlusion effect includes: generating a virtual image with a leg occlusion effect corresponding to each frame of image; the generating method further includes: according to each frame.
  • the virtual image corresponding to the frame image generates a video with a leg blocking effect.
  • an apparatus for generating a virtual image comprising: a determining unit, configured to acquire a leg area in an image to be processed including legs and feet, according to the relationship between the leg area and the two-dimensional shoe The overlapping part of the shoe area in the image determines the part of the shoe area that is occluded by the leg; the processing unit is used to render the three-dimensional shoe model as the image to be processed according to the posture parameters of the foot in the image to be processed and the internal parameters of the camera For the corresponding two-dimensional shoe image, a composite image of the to-be-processed image and the two-dimensional shoe image is rendered according to the part blocked by the leg, so as to generate a virtual image with a leg blocking effect.
  • the determining unit determines the outer contour of the shoe in the two-dimensional shoe image according to the position of the shoe body region in the two-dimensional shoe image, and determines the inner contour of the shoe in the two-dimensional shoe image according to the position of the shoe mouth region in the two-dimensional image ; According to the intersection of the contour of the leg area with the outer contour and the inner contour, determine the part of the shoe area that is occluded by the leg.
  • the determining unit determines the point on the inner contour that is closest to the intersection point; and determines the portion of the shoe area that is occluded by the leg according to the intersection point and the closest point.
  • the processing unit performs transparency processing on the shoe opening area in the three-dimensional shoe model; and renders the three-dimensional shoe model after the transparent processing into the two-dimensional shoe image.
  • the determining unit determines the shoe body area of the two-dimensional shoe image and the shoe mouth area of the two-dimensional shoe image according to the binary image of the two-dimensional shoe image.
  • the processing unit detects the shoe opening area of the three-dimensional shoe model, and uses a closed mesh to cover the shoe opening area of the three-dimensional shoe model; and performs transparency processing on the covering part of the closed mesh.
  • the determining unit determines the position where the closed mesh covers the shoe mouth area of the three-dimensional shoe model according to the size of the part of the shoe area that is not covered by the leg.
  • the determining unit inputs the image to be processed into the machine learning model, and determines the leg region in the image to be processed.
  • the machine learning model includes a convolutional neural network module and a spatial pyramid pooling module connected in sequence.
  • the convolutional neural network module is set up according to the Fast-SCNN model.
  • the image to be processed is each frame image in the video; the processing unit generates a virtual image with a leg blocking effect corresponding to each frame image, and generates a leg blocking effect according to the virtual image corresponding to each frame image. video.
  • an apparatus for generating a virtual image comprising: a memory; and a processor coupled to the memory, the processor being configured to execute any one of the above implementations based on instructions stored in the memory apparatus The generation method of the virtual image in the example.
  • a non-volatile computer-readable storage medium on which a computer program is stored, and when the program is executed by a processor, implements the method for generating a virtual image in any one of the foregoing embodiments.
  • FIG. 1 shows a flowchart of some embodiments of the method for generating a virtual image of the present disclosure
  • FIG. 2 shows a flowchart of some embodiments of step 130 in FIG. 1;
  • 3a-3c show schematic diagrams of some embodiments of the method for generating a virtual image of the present disclosure
  • FIG. 4 shows a block diagram of some embodiments of the apparatus for generating virtual images of the present disclosure
  • FIG. 5 shows a block diagram of other embodiments of the virtual image generating apparatus of the present disclosure
  • FIG. 6 shows a block diagram of further embodiments of the apparatus for generating virtual images of the present disclosure.
  • the inventor of the present disclosure found that the above-mentioned related art has the following problems: the virtual image of human feet wearing shoes is only generated by approximate modeling, lacking real three-dimensional structure information, resulting in poor effect of generating the virtual image.
  • the present disclosure proposes a technical solution for generating a virtual image, which can synthesize a virtual image by using a real leg image and a three-dimensional shoe model to improve the effect of the virtual image.
  • the position of the virtual human leg model relative to the shoe model is fixed and cannot reflect the real three-dimensional structure of the leg (such as the user's actual trousers position, shape, posture, leg position, etc.). This will cause the effect of the virtual image to decrease.
  • the visual clues of the legs can be extracted in the real scene (for example, using a neural network model), and combined with the outline of the 3D model of the shoe after the shoe opening area is transparent, the area that needs to be covered in the virtual image can be accurately divided.
  • the technical solutions of the present disclosure can be implemented through the following embodiments.
  • FIG. 1 shows a flowchart of some embodiments of the method for generating a virtual image of the present disclosure.
  • the production method includes: step 110 , acquiring a leg area; step 120 , rendering a two-dimensional shoe image; step 130 , determining the part occluded by the leg; and step 140 , generating a virtual image.
  • step 110 the leg region in the to-be-processed image including the legs and feet is acquired.
  • the image to be processed is input into a machine learning model that determines the region of the leg in the image to be processed.
  • the machine learning model can be a convolutional neural network module (as set up according to the Fast-SCNN model).
  • a skeleton network for extracting leg regions is set up based on the lightweight model Fast-SCNN.
  • a simple and efficient composite coefficient can be used to build a more structured network structure, thereby compressing the amount of parameters of the model and improving the training speed.
  • Fast-SCNN can reduce the amount of floating-point operations of the model, thereby improving computing performance.
  • the machine learning model includes a convolutional neural network module and an SPP (Spatial Pyramid Pooling) module connected in sequence.
  • the SPP module includes a convolution processing module, an upsampling module, and a concatenation (Concat) module that are connected in sequence.
  • the machine learning model includes a convolutional neural network module, a first SPP module, and a second SPP module.
  • the convolutional neural network module is connected with the convolution processing module and the connection module of the first SPP module, and the connection module of the second SPP module is connected; the first SPP module is connected with the second SPP module.
  • the SPP module can well preserve the complete inter-context information and avoid misclassification in image processing. Moreover, the SPP module is more robust to small-sized, insignificant object recognition and can pay attention to different sub-regions containing insignificant objects, thereby improving the accuracy of leg region recognition.
  • the aforementioned machine learning model can be trained using the SoftMax Loss setting loss function.
  • the three-dimensional shoe model is rendered into a two-dimensional shoe image corresponding to the to-be-processed image according to the posture parameters of the foot in the to-be-processed image and the camera's internal parameters.
  • the internal parameters of the camera are parameters related to the characteristics of the camera itself, such as the focal length and pixel size of the camera.
  • a PnP (Perspective-n-Point, multi-point perspective) algorithm can be used to determine the pose parameters
  • the three-dimensional shoe model can be rendered into a two-dimensional shoe image using a rendering tool such as OpenGL.
  • the shoe opening area of the three-dimensional shoe model is detected, and a closed mesh (mesh) is used as a baffle to cover the shoe opening area; the covering part of the closed mesh is transparentized.
  • a transparent mesh of baffles can be equipped on the 3D model of the shoe and placed on the inner side of the shoe opening area to achieve transparency of the shoe opening area.
  • the position of the closed mesh covering the shoe opening area is determined according to the size of the part of the shoe area that is not covered by the leg. For example, the position of the baffle can be moved a preset distance toward the sole, so that the edge thickness of the shoe opening area exceeds a threshold value.
  • the uncropped part of the shoe opening area has a certain thickness, which can increase the spatial layering of the virtual picture and enhance the real effect.
  • step 130 according to the overlapping portion of the leg region and the shoe region in the two-dimensional shoe image, the portion of the shoe region that is occluded by the leg is determined.
  • only the leg area can be accurately segmented by using the deep learning model, so as to determine the boundary area between the leg model and the shoe model, and then the occluded part of the shoe opening area can be accurately obtained as the cropping area.
  • the visual effect of virtual occlusion will be generated.
  • step 130 may be implemented according to the embodiment in FIG. 2 .
  • FIG. 2 shows a flowchart of some embodiments of step 130 in FIG. 1 .
  • step 130 includes: step 1310 , determining the inner and outer contours of the shoe; and step 1320 , determining the portion occluded by the leg.
  • step 1310 the outer contour of the shoe is determined according to the position of the shoe body region in the two-dimensional shoe image, and the inner contour of the shoe is determined according to the position of the shoe mouth region in the two-dimensional image.
  • the shoe body area and the shoe mouth area may be determined for use in determining the inner and outer contours using the embodiment of Figure 3a.
  • Figure 3a shows a schematic diagram of some embodiments of the method for generating a virtual image of the present disclosure.
  • the outer contour and the inner contour can be determined, and then the occlusion part can be determined through the remaining steps in FIG. 2 .
  • step 1320 the part of the shoe area that is blocked by the leg is determined according to the position of the intersection of the contour of the leg area and the outer contour, and the inner contour.
  • the shoe body area and the shoe mouth area may be determined for use in determining the inner and outer contours through the embodiments in Figures 3b and 3c.
  • Figure 3b shows a schematic diagram of some embodiments of the method for generating a virtual image of the present disclosure.
  • the segmentation mask binary image of the leg region 30 can be inferred.
  • the intersection of the contour of the leg region 30 and the contour of the shoe body region 31 can be determined.
  • Figure 3c shows a schematic diagram of some embodiments of the method for generating a virtual image of the present disclosure.
  • the outline of the shoe body region 31 is an outer outline 311
  • the outline of the shoe mouth region 32 is an inner outline 321 .
  • the intersection points of the contour of the leg region 30 and the outer contour 311 are 3a, 3b, the point closest to the intersection 3a on the inner contour 321 is 3d, and the closest point to the intersection 3b is 3c.
  • connection points 3a, 3b, 3c, 3d form a closed area (located on the left side of the shoe body), which is defined as the part of the shoe area that is obscured by the legs.
  • the intersection of the contour of the leg area 30 and the inner contour 321 and the closed area formed by 3a, 3b determine the part of the shoe area that is obscured by the leg.
  • a virtual image can be generated through step 140 in FIG. 1 .
  • step 140 a composite image of the to-be-processed image and the two-dimensional shoe image is rendered according to the part blocked by the leg, so as to generate a virtual image with a leg blocking effect.
  • the images to be processed are frames of images in the video.
  • a video with a leg blocking effect can be generated according to the virtual image corresponding to each generated frame of image.
  • the virtual occlusion portion is accurately determined according to the obtained visual clues of the real leg and the position of the shoe in the combined two-dimensional shoe image.
  • a virtual image can be synthesized by using the real leg image and the three-dimensional shoe model to improve the effect of the virtual image.
  • FIG. 4 shows a block diagram of some embodiments of an apparatus for generating a virtual image of the present disclosure.
  • the virtual image generating apparatus 4 includes a determination unit 41 and a processing unit 42 .
  • the determination unit 41 acquires the leg area in the to-be-processed image including the legs and feet; and determines the part of the shoe area covered by the leg according to the overlapping part of the leg area and the shoe area in the two-dimensional shoe image.
  • the determining unit 41 determines the outer contour of the shoe according to the position of the shoe body region in the two-dimensional shoe image, and determines the inner contour of the shoe according to the position of the shoe mouth region in the two-dimensional image; The intersection of the contours and the inner contour, determine the portion of the shoe area that is occluded by the leg.
  • the determining unit 41 determines the point closest to the intersection on the inner contour; and determines the part of the shoe area that is occluded by the leg according to the intersection and the closest point.
  • the determining unit 41 inputs the image to be processed into the machine learning model, and determines the leg region in the image to be processed.
  • the machine learning model includes a convolutional neural network module and a spatial pyramid pooling module connected in sequence.
  • the convolutional neural network module is set up according to the Fast-SCNN model.
  • the processing unit 42 renders the three-dimensional shoe model as a two-dimensional shoe image corresponding to the to-be-processed image according to the posture parameters of the foot in the image to be processed and the internal parameters of the camera; A composite image of dimensional shoe images to generate a virtual image with a leg occlusion effect.
  • the processing unit 42 performs transparency processing on the shoe mouth area in the three-dimensional shoe model; and renders the three-dimensional shoe model after the transparent processing as the two-dimensional shoe image.
  • the determining unit 41 determines the shoe body area and the shoe opening area according to the binary image of the two-dimensional shoe image.
  • the processing unit 42 detects the shoe opening area of the three-dimensional shoe model, and covers the shoe opening area with a closed mesh; and performs transparency processing on the covering part of the closed mesh.
  • the determining unit 41 determines the position where the closed mesh covers the shoe opening area according to the size of the part of the shoe area that is not covered by the legs.
  • the image to be processed is each frame of image in the video; the processing unit 42 generates a video with a leg blocking effect according to the generated virtual image corresponding to each frame of image.
  • FIG. 5 shows a block diagram of other embodiments of the virtual image generating apparatus of the present disclosure.
  • the virtual image generating apparatus 5 of this embodiment includes: a memory 51 and a processor 52 coupled to the memory 51 , and the processor 52 is configured to execute the present disclosure based on instructions stored in the memory 51 The method for generating a virtual image in any one of the embodiments.
  • the memory 51 may include, for example, a system memory, a fixed non-volatile storage medium, and the like.
  • the system memory stores, for example, an operating system, an application program, a boot loader Boot Loader, a database, and other programs.
  • FIG. 6 shows a block diagram of further embodiments of the apparatus for generating virtual images of the present disclosure.
  • the virtual image generating apparatus 6 of this embodiment includes: a memory 610 and a processor 620 coupled to the memory 610 , and the processor 620 is configured to execute any of the foregoing based on the instructions stored in the memory 610 .
  • a method for generating a virtual image in one embodiment.
  • Memory 610 may include, for example, system memory, fixed non-volatile storage media, and the like.
  • the system memory stores, for example, an operating system, an application program, a boot loader, and other programs.
  • the virtual image generating apparatus 6 may further include an input/output interface 630, a network interface 640, a storage interface 650, and the like. These interfaces 630 , 640 , 650 and the memory 610 and the processor 620 may be connected, for example, through a bus 660 .
  • the input and output interface 630 provides a connection interface for input and output devices such as a display, a mouse, a keyboard, a touch screen, a microphone, and a speaker.
  • Network interface 640 provides a connection interface for various networked devices.
  • the storage interface 650 provides a connection interface for external storage devices such as SD cards and U disks.
  • embodiments of the present disclosure may be provided as a method, system, or computer program product. Accordingly, the present disclosure may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present disclosure may take the form of a computer program product embodied on one or more computer-usable non-transitory storage media having computer-usable program code embodied therein, including but not limited to disk storage, CD-ROM, optical storage, and the like.
  • the methods and systems of the present disclosure may be implemented in many ways.
  • the methods and systems of the present disclosure may be implemented in software, hardware, firmware, or any combination of software, hardware, and firmware.
  • the above order of steps for the method is for illustration only, and the steps of the method of the present disclosure are not limited to the order specifically described above unless specifically stated otherwise.
  • the present disclosure can also be implemented as programs recorded in a recording medium, the programs including machine-readable instructions for implementing methods according to the present disclosure.
  • the present disclosure also covers a recording medium storing a program for executing the method according to the present disclosure.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Graphics (AREA)
  • Software Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Geometry (AREA)
  • Computer Hardware Design (AREA)
  • Processing Or Creating Images (AREA)

Abstract

Provided are a virtual image generation method and apparatus (4, 5, 6), relating to the technical field of computers. The method comprises: obtaining a leg region in an image to be processed containing a leg and a foot; according to posture parameters of the foot in the image to be processed and the internal parameters of a camera, rendering a three-dimensional shoe model into a two-dimensional shoe image corresponding to the image to be processed; according to the overlapping part of the leg with the shoe region in the two-dimensional shoe image, determining a part of the shoe region that is obscured by the leg; according to the part obscured by the leg, rendering a composite image of the image to be processed and the two-dimensional shoe image, so as to generate a virtual image having a leg obscuring effect.

Description

虚拟图像的生成方法和装置Method and device for generating virtual image
相关申请的交叉引用CROSS-REFERENCE TO RELATED APPLICATIONS
本申请是以CN申请号为202011134938.4,申请日为2020年10月21日的申请为基础,并主张其优先权,该CN申请的公开内容在此作为整体引入本申请中。This application is based on the CN application number 202011134938.4 and the filing date is October 21, 2020, and claims its priority. The disclosure of the CN application is hereby incorporated into this application as a whole.
技术领域technical field
本公开涉及计算机技术领域,特别涉及一种虚拟图像的生成方法、虚拟图像的生成装置和非易失性计算机可读存储介质。The present disclosure relates to the field of computer technologies, and in particular, to a method for generating a virtual image, a device for generating a virtual image, and a non-volatile computer-readable storage medium.
背景技术Background technique
随着VR(Virtual Reality,虚拟现实)和AR(Augmented Reality,增强现实)技术的进步,通过虚拟试穿方式形成导购转化的功能越来越受到大众欢迎。通过AR增强现实技术与智能手机相机的结合实现的虚拟试鞋技术,可以帮助用户看到鞋款穿在自己脚上的效果。With the advancement of VR (Virtual Reality, Virtual Reality) and AR (Augmented Reality, Augmented Reality) technology, the function of forming shopping guide conversion through virtual try-on is more and more popular. The virtual shoe try-on technology realized by the combination of AR augmented reality technology and smartphone camera can help users see the effect of shoes on their feet.
为了实现人脚穿鞋视觉效果,虚拟试鞋技术需对三维鞋模型进行虚拟遮挡,以替换人脚的真实鞋子。In order to realize the visual effect of human feet wearing shoes, the virtual shoe trial technology needs to perform virtual occlusion on the three-dimensional shoe model to replace the real shoes of human feet.
在相关技术中,在三维鞋模型的鞋口区域外建模一条虚拟的腿部模型,将虚拟的腿部模型和三维鞋模型渲染到屏幕上以生成人脚穿鞋的虚拟图像。In the related art, a virtual leg model is modeled outside the shoe mouth area of the 3D shoe model, and the virtual leg model and the 3D shoe model are rendered on the screen to generate a virtual image of a human foot wearing shoes.
发明内容SUMMARY OF THE INVENTION
根据本公开的一些实施例,提供了一种虚拟图像的生成方法,包括:获取包含腿部和脚部的待处理图像中的腿部区域;根据待处理图像中脚部的姿态参数和相机的内部参数,将三维鞋模型渲染为与待处理图像对应的二维鞋图像;根据腿部区域与二维鞋图像中鞋区域的重叠部分,确定鞋区域中被腿部遮挡的部分;根据被腿部遮挡的部分,渲染待处理图像和二维鞋图像的合成图像,以生成具有腿部遮挡效果的虚拟图像。According to some embodiments of the present disclosure, a method for generating a virtual image is provided, including: acquiring a leg area in an image to be processed including legs and feet; Internal parameters, rendering the 3D shoe model as a 2D shoe image corresponding to the image to be processed; according to the overlapping part of the leg area and the shoe area in the 2D shoe image, determine the part of the shoe area that is occluded by the leg; The partially occluded part is rendered, and the composite image of the to-be-processed image and the two-dimensional shoe image is rendered to generate a virtual image with a leg occlusion effect.
在一些实施例中,根据腿部区域与二维鞋图像中鞋区域的重叠部分,确定鞋区域中被腿部遮挡的部分包括:根据二维鞋图像中鞋体区域的位置确定二维鞋图像中鞋的外轮廓,根据二维图像中鞋口区域的位置确定二维鞋图像中鞋的内轮廓;根据腿部区域的轮廓与外轮廓的交点和内轮廓,确定鞋区域中被腿部遮挡的部分。In some embodiments, determining the portion of the shoe region occluded by the leg according to the overlapping portion of the leg region and the shoe region in the two-dimensional shoe image includes: determining the two-dimensional shoe image according to the position of the shoe body region in the two-dimensional shoe image For the outer contour of the middle shoe, the inner contour of the shoe in the two-dimensional shoe image is determined according to the position of the shoe mouth area in the two-dimensional image; according to the intersection of the contour of the leg area and the outer contour and the inner contour, it is determined that the shoe area is blocked by the leg. part.
在一些实施例中,根据腿部区域的轮廓与外轮廓的交点和内轮廓,确定鞋区域中被腿部遮挡的部分包括:在内轮廓上,确定与交点距离最近的点;根据交点、距离最近的点,确定鞋区域中被腿部遮挡的部分。In some embodiments, determining the portion of the shoe region that is occluded by the leg according to the intersection of the contour of the leg region and the outer contour and the inner contour includes: on the inner contour, determining a point closest to the intersection; The closest point that determines the portion of the shoe area that is occluded by the leg.
在一些实施例中,将三维鞋模型渲染为与待处理图像对应的二维鞋图像包括:对三维鞋模型中的鞋口区域进行透明化处理;将透明化处理后的三维鞋模型,渲染为二维鞋图像;根据二维鞋图像的二值图像,确定二维鞋图像的鞋体区域和二维鞋图像的鞋口区域。In some embodiments, rendering the three-dimensional shoe model into a two-dimensional shoe image corresponding to the image to be processed includes: performing transparency processing on the shoe opening area in the three-dimensional shoe model; rendering the transparently processed three-dimensional shoe model as Two-dimensional shoe image; according to the binary image of the two-dimensional shoe image, the shoe body area of the two-dimensional shoe image and the shoe mouth area of the two-dimensional shoe image are determined.
在一些实施例中,对三维鞋模型中的鞋口区域进行透明化处理包括:检测三维鞋模型的鞋口区域,并利用封闭网格覆盖三维鞋模型的鞋口区域;对封闭网格覆盖部分进行透明化处理。In some embodiments, the transparent processing of the shoe opening area in the three-dimensional shoe model includes: detecting the shoe opening area of the three-dimensional shoe model, and using a closed mesh to cover the shoe opening area of the three-dimensional shoe model; covering part of the closed mesh Transparency is performed.
在一些实施例中,获取包含腿部和脚部的待处理图像中的腿部区域包括:将待处理图像输入机器学习模型,确定待处理图像中的腿部区域。In some embodiments, acquiring the leg region in the to-be-processed image including the legs and feet includes: inputting the to-be-processed image into a machine learning model to determine the leg region in the to-be-processed image.
在一些实施例中,机器学习模型包括依次连接的卷积神经网络模块和空间金字塔池化模块。In some embodiments, the machine learning model includes a convolutional neural network module and a spatial pyramid pooling module connected in sequence.
在一些实施例中,卷积神经网络模块根据Fast-SCNN(Fast Segmentation Convolutional Neural Network,快速分割卷积神经网络)模型设置。In some embodiments, the convolutional neural network module is set according to the Fast-SCNN (Fast Segmentation Convolutional Neural Network, fast segmentation convolutional neural network) model.
在一些实施例中,待处理图像为视频中的各帧图像,生成具有腿部遮挡效果的虚拟图像包括:生成各帧图像对应的具有腿部遮挡效果的虚拟图像;生成方法还包括:根据各帧图像对应的虚拟图像,生成具有腿部遮效果的视频。In some embodiments, the image to be processed is each frame of images in the video, and generating a virtual image with a leg occlusion effect includes: generating a virtual image with a leg occlusion effect corresponding to each frame of image; the generating method further includes: according to each frame. The virtual image corresponding to the frame image generates a video with a leg blocking effect.
根据本公开的另一些实施例,提供一种虚拟图像的生成装置,包括:确定单元,用于获取包含腿部和脚部的待处理图像中的腿部区域,根据腿部区域与二维鞋图像中鞋区域的重叠部分,确定鞋区域中被腿部遮挡的部分;处理单元,用于根据待处理图像中脚部的姿态参数和相机的内部参数,将三维鞋模型渲染为与待处理图像对应的二维鞋图像,根据被腿部遮挡的部分,渲染待处理图像和二维鞋图像的合成图像,以生成具有腿部遮挡效果的虚拟图像。According to other embodiments of the present disclosure, there is provided an apparatus for generating a virtual image, comprising: a determining unit, configured to acquire a leg area in an image to be processed including legs and feet, according to the relationship between the leg area and the two-dimensional shoe The overlapping part of the shoe area in the image determines the part of the shoe area that is occluded by the leg; the processing unit is used to render the three-dimensional shoe model as the image to be processed according to the posture parameters of the foot in the image to be processed and the internal parameters of the camera For the corresponding two-dimensional shoe image, a composite image of the to-be-processed image and the two-dimensional shoe image is rendered according to the part blocked by the leg, so as to generate a virtual image with a leg blocking effect.
在一些实施例中,确定单元根据二维鞋图像中鞋体区域的位置确定二维鞋图像中鞋的外轮廓,根据二维图像中鞋口区域的位置确定二维鞋图像中鞋的内轮廓;根据腿部区域的轮廓与外轮廓的交点和内轮廓,确定鞋区域中被腿部遮挡的部分。In some embodiments, the determining unit determines the outer contour of the shoe in the two-dimensional shoe image according to the position of the shoe body region in the two-dimensional shoe image, and determines the inner contour of the shoe in the two-dimensional shoe image according to the position of the shoe mouth region in the two-dimensional image ; According to the intersection of the contour of the leg area with the outer contour and the inner contour, determine the part of the shoe area that is occluded by the leg.
在一些实施例中,确定单元在内轮廓上,确定与交点距离最近的点;根据交点、距离最近的点,确定鞋区域中被腿部遮挡的部分。In some embodiments, the determining unit determines the point on the inner contour that is closest to the intersection point; and determines the portion of the shoe area that is occluded by the leg according to the intersection point and the closest point.
在一些实施例中,处理单元对三维鞋模型中的鞋口区域进行透明化处理;将透明化处理后的三维鞋模型,渲染为所述二维鞋图像。确定单元根据二维鞋图像的二值图像,确定二维鞋图像的鞋体区域和二维鞋图像的鞋口区域。In some embodiments, the processing unit performs transparency processing on the shoe opening area in the three-dimensional shoe model; and renders the three-dimensional shoe model after the transparent processing into the two-dimensional shoe image. The determining unit determines the shoe body area of the two-dimensional shoe image and the shoe mouth area of the two-dimensional shoe image according to the binary image of the two-dimensional shoe image.
在一些实施例中,处理单元检测三维鞋模型的鞋口区域,并利用封闭网格覆盖三维鞋模型的鞋口区域;对封闭网格覆盖部分进行透明化处理。In some embodiments, the processing unit detects the shoe opening area of the three-dimensional shoe model, and uses a closed mesh to cover the shoe opening area of the three-dimensional shoe model; and performs transparency processing on the covering part of the closed mesh.
在一些实施例中确定单元根据预设的鞋区域中未被腿部遮挡的部分的大小,确定封闭网格覆盖三维鞋模型的鞋口区域的位置。In some embodiments, the determining unit determines the position where the closed mesh covers the shoe mouth area of the three-dimensional shoe model according to the size of the part of the shoe area that is not covered by the leg.
在一些实施例中,确定单元将待处理图像输入机器学习模型,确定待处理图像中的腿部区域。In some embodiments, the determining unit inputs the image to be processed into the machine learning model, and determines the leg region in the image to be processed.
在一些实施例中,机器学习模型包括依次连接的卷积神经网络模块和空间金字塔池化模块。In some embodiments, the machine learning model includes a convolutional neural network module and a spatial pyramid pooling module connected in sequence.
在一些实施例中,卷积神经网络模块根据Fast-SCNN模型设置。In some embodiments, the convolutional neural network module is set up according to the Fast-SCNN model.
在一些实施例中,待处理图像为视频中的各帧图像;处理单元生成各帧图像对应的具有腿部遮挡效果的虚拟图像,根据各帧图像对应的虚拟图像,生成具有腿部遮效果的视频。In some embodiments, the image to be processed is each frame image in the video; the processing unit generates a virtual image with a leg blocking effect corresponding to each frame image, and generates a leg blocking effect according to the virtual image corresponding to each frame image. video.
根据本公开的又一些实施例,提供一种虚拟图像的生成装置,包括:存储器;和耦接至存储器的处理器,处理器被配置为基于存储在存储器装置中的指令,执行上述任一个实施例中的虚拟图像的生成方法。According to further embodiments of the present disclosure, there is provided an apparatus for generating a virtual image, comprising: a memory; and a processor coupled to the memory, the processor being configured to execute any one of the above implementations based on instructions stored in the memory apparatus The generation method of the virtual image in the example.
根据本公开的再一些实施例,提供一种非易失性计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现上述任一个实施例中的虚拟图像的生成方法。According to still other embodiments of the present disclosure, there is provided a non-volatile computer-readable storage medium on which a computer program is stored, and when the program is executed by a processor, implements the method for generating a virtual image in any one of the foregoing embodiments.
附图说明Description of drawings
构成说明书的一部分的附图描述了本公开的实施例,并且连同说明书一起用于解释本公开的原理。The accompanying drawings, which form a part of the specification, illustrate embodiments of the present disclosure and together with the description serve to explain the principles of the present disclosure.
参照附图,根据下面的详细描述,可以更加清楚地理解本公开:The present disclosure may be more clearly understood from the following detailed description with reference to the accompanying drawings:
图1示出本公开的虚拟图像的生成方法的一些实施例的流程图;FIG. 1 shows a flowchart of some embodiments of the method for generating a virtual image of the present disclosure;
图2示出图1中步骤130的一些实施例的流程图;FIG. 2 shows a flowchart of some embodiments of step 130 in FIG. 1;
图3a~3c示出本公开的虚拟图像的生成方法的一些实施例的示意图;3a-3c show schematic diagrams of some embodiments of the method for generating a virtual image of the present disclosure;
图4示出本公开的虚拟图像的生成装置的一些实施例的框图;4 shows a block diagram of some embodiments of the apparatus for generating virtual images of the present disclosure;
图5示出本公开的虚拟图像的生成装置的另一些实施例的框图;FIG. 5 shows a block diagram of other embodiments of the virtual image generating apparatus of the present disclosure;
图6示出本公开的虚拟图像的生成装置的又一些实施例的框图。FIG. 6 shows a block diagram of further embodiments of the apparatus for generating virtual images of the present disclosure.
具体实施方式Detailed ways
现在将参照附图来详细描述本公开的各种示例性实施例。应注意到:除非另外具体说明,否则在这些实施例中阐述的部件和步骤的相对布置、数字表达式和数值不限制本公开的范围。Various exemplary embodiments of the present disclosure will now be described in detail with reference to the accompanying drawings. It should be noted that the relative arrangement of the components and steps, the numerical expressions and numerical values set forth in these embodiments do not limit the scope of the present disclosure unless specifically stated otherwise.
同时,应当明白,为了便于描述,附图中所示出的各个部分的尺寸并不是按照实际的比例关系绘制的。Meanwhile, it should be understood that, for the convenience of description, the dimensions of various parts shown in the accompanying drawings are not drawn in an actual proportional relationship.
以下对至少一个示例性实施例的描述实际上仅仅是说明性的,决不作为对本公开及其应用或使用的任何限制。The following description of at least one exemplary embodiment is merely illustrative in nature and is in no way intended to limit the disclosure, its application or uses in any way.
对于相关领域普通技术人员已知的技术、方法和设备可能不作详细讨论,但在适当情况下,技术、方法和设备应当被视为授权说明书的一部分。Techniques, methods, and apparatus known to those of ordinary skill in the relevant art may not be discussed in detail, but where appropriate, techniques, methods, and apparatus should be considered part of the authorized description.
在这里示出和讨论的所有示例中,任何具体值应被解释为仅仅是示例性的,而不是作为限制。因此,示例性实施例的其它示例可以具有不同的值。In all examples shown and discussed herein, any specific value should be construed as illustrative only and not as limiting. Accordingly, other examples of exemplary embodiments may have different values.
应注意到:相似的标号和字母在下面的附图中表示类似项,因此,一旦某一项在一个附图中被定义,则在随后的附图中不需要对其进行进一步讨论。It should be noted that like numerals and letters refer to like items in the following figures, so once an item is defined in one figure, it does not require further discussion in subsequent figures.
本公开的发明人发现上述相关技术中存在如下问题:仅通过近似建模生成人脚穿鞋的虚拟图像,缺乏真实的三维构造信息,导致生成虚拟图像的效果差。The inventor of the present disclosure found that the above-mentioned related art has the following problems: the virtual image of human feet wearing shoes is only generated by approximate modeling, lacking real three-dimensional structure information, resulting in poor effect of generating the virtual image.
鉴于此,本公开提出了一种虚拟图像的生成技术方案,能够利用真实的腿部图像与三维鞋模型合成虚拟图像,提高虚拟图像的效果。In view of this, the present disclosure proposes a technical solution for generating a virtual image, which can synthesize a virtual image by using a real leg image and a three-dimensional shoe model to improve the effect of the virtual image.
如前所述,虚拟的人腿模型相对于鞋模型的位是固定的,无法反映真实的腿部三维构造(如用户实际的裤子位置、形状、姿态、腿部位置等)。这样会导致虚拟图像的效果下降。As mentioned above, the position of the virtual human leg model relative to the shoe model is fixed and cannot reflect the real three-dimensional structure of the leg (such as the user's actual trousers position, shape, posture, leg position, etc.). This will cause the effect of the virtual image to decrease.
针对上述技术问题,可以在真实场景中提取腿部的视觉线索(如利用神经网络模型提取),并结合鞋口区域透明化后的鞋三维模型的轮廓,准确划分出虚拟图像中需要遮挡的区域。例如,可以通过如下的实施例实现本公开的技术方案。In view of the above technical problems, the visual clues of the legs can be extracted in the real scene (for example, using a neural network model), and combined with the outline of the 3D model of the shoe after the shoe opening area is transparent, the area that needs to be covered in the virtual image can be accurately divided. . For example, the technical solutions of the present disclosure can be implemented through the following embodiments.
图1示出本公开的虚拟图像的生成方法的一些实施例的流程图。FIG. 1 shows a flowchart of some embodiments of the method for generating a virtual image of the present disclosure.
如图1所示,该生产方法包括:步骤110,获取腿部区域;步骤120,渲染二维鞋图像;步骤130,确定被腿部遮挡的部分;和步骤140,生成虚拟图像。As shown in FIG. 1 , the production method includes: step 110 , acquiring a leg area; step 120 , rendering a two-dimensional shoe image; step 130 , determining the part occluded by the leg; and step 140 , generating a virtual image.
在步骤110中,获取包含腿部和脚部的待处理图像中的腿部区域。In step 110, the leg region in the to-be-processed image including the legs and feet is acquired.
在一些实施例中,将待处理图像输入机器学习模型,确定待处理图像中的腿部区域。例如,机器学习模型可以为卷积神经网络模块(如根据Fast-SCNN模型设置)。In some embodiments, the image to be processed is input into a machine learning model that determines the region of the leg in the image to be processed. For example, the machine learning model can be a convolutional neural network module (as set up according to the Fast-SCNN model).
这样,基于轻量模型Fast-SCNN设置用于提取腿部区域的骨架网络。可以利用一个简单而高效的复合系数,更结构化地构建网络结构,从而压缩模型的参数量,提高训练速度。而且,Fast-SCNN能够减少模型浮点运算量,从而提升计算性能。In this way, a skeleton network for extracting leg regions is set up based on the lightweight model Fast-SCNN. A simple and efficient composite coefficient can be used to build a more structured network structure, thereby compressing the amount of parameters of the model and improving the training speed. Moreover, Fast-SCNN can reduce the amount of floating-point operations of the model, thereby improving computing performance.
在一些实施例中,机器学习模型包括依次连接的卷积神经网络模块和SPP(Spatial Pyramid Pooling,空间金字塔池化)模块。例如,SPP模块包括依次连接的卷积处理模块、上采样模块和连接(Concat)模块。In some embodiments, the machine learning model includes a convolutional neural network module and an SPP (Spatial Pyramid Pooling) module connected in sequence. For example, the SPP module includes a convolution processing module, an upsampling module, and a concatenation (Concat) module that are connected in sequence.
在一些实施例中,机器学习模型包括卷积神经网络模块、第一SPP模块、第二SPP模块。卷积神经网络模块与第一SPP模块的卷积处理模块和连接模块连接,第二SPP模块的连接模块连接;第一SPP模块与第二SPP模块连接。In some embodiments, the machine learning model includes a convolutional neural network module, a first SPP module, and a second SPP module. The convolutional neural network module is connected with the convolution processing module and the connection module of the first SPP module, and the connection module of the second SPP module is connected; the first SPP module is connected with the second SPP module.
这样,SPP模块能够很好地保持完整的上下文间信息,避免图像处理中的错误分类。而且,SPP模块对小尺寸、不显著的目标识别具有更好的鲁棒性,能够注意包含不显著物体的不同子区域,从而提高腿部区域识别的准确性。In this way, the SPP module can well preserve the complete inter-context information and avoid misclassification in image processing. Moreover, the SPP module is more robust to small-sized, insignificant object recognition and can pay attention to different sub-regions containing insignificant objects, thereby improving the accuracy of leg region recognition.
在一些实施例中,可以使用SoftMax Loss设置损失函数训练上述机器学习模型。In some embodiments, the aforementioned machine learning model can be trained using the SoftMax Loss setting loss function.
在步骤120中,根据待处理图像中脚部的姿态参数和相机的内部参数,将三维鞋模型渲染为与待处理图像对应的二维鞋图像。例如,相机的内部参数为与相机自身特性相关的参数,如相机的焦距、像素大小等。In step 120, the three-dimensional shoe model is rendered into a two-dimensional shoe image corresponding to the to-be-processed image according to the posture parameters of the foot in the to-be-processed image and the camera's internal parameters. For example, the internal parameters of the camera are parameters related to the characteristics of the camera itself, such as the focal length and pixel size of the camera.
在一些实施例中,可以采用PnP(Perspective-n-Point,多点透视)算法确定姿态参数In some embodiments, a PnP (Perspective-n-Point, multi-point perspective) algorithm can be used to determine the pose parameters
在一些实施例中,可以利用OpenGL等渲染工具将三维鞋模型渲染为二维鞋图像。In some embodiments, the three-dimensional shoe model can be rendered into a two-dimensional shoe image using a rendering tool such as OpenGL.
在一些实施例中,检测三维鞋模型的鞋口区域,并利用封闭网格(mesh)作为挡片覆盖鞋口区域;对封闭网络的覆盖部分进行透明化处理。例如,可以在鞋三维模型上配备挡片的透明网格,并放在鞋口区域的内侧以实现鞋口区域的透明化。In some embodiments, the shoe opening area of the three-dimensional shoe model is detected, and a closed mesh (mesh) is used as a baffle to cover the shoe opening area; the covering part of the closed mesh is transparentized. For example, a transparent mesh of baffles can be equipped on the 3D model of the shoe and placed on the inner side of the shoe opening area to achieve transparency of the shoe opening area.
在一些实施例中,根据预设的鞋区域中未被腿部遮挡的部分的大小,确定封闭网格覆盖鞋口区域的位置。例如,可以将挡片的位置向鞋底方向移动预设距离,使得鞋口区域的边缘厚度超过阈值。In some embodiments, the position of the closed mesh covering the shoe opening area is determined according to the size of the part of the shoe area that is not covered by the leg. For example, the position of the baffle can be moved a preset distance toward the sole, so that the edge thickness of the shoe opening area exceeds a threshold value.
这样,根据挡片位置的深度不同,鞋口区域未被裁剪的部分具有一定厚度,可以增加虚拟图片的空间层次感,增强真实效果。In this way, according to the depth of the position of the blocking piece, the uncropped part of the shoe opening area has a certain thickness, which can increase the spatial layering of the virtual picture and enhance the real effect.
在步骤130中,根据腿部区域与二维鞋图像中鞋区域的重叠部分,确定鞋区域中被腿部遮挡的部分。In step 130, according to the overlapping portion of the leg region and the shoe region in the two-dimensional shoe image, the portion of the shoe region that is occluded by the leg is determined.
在一些实施例中,利用深度学习模型可以对腿部区域仅准确分割,从而确定腿部模型与鞋模型的交界区域,进而可以准确获取鞋口区域被遮挡部分作为裁剪区域。经过渲染处理上屏显示,就会产生虚拟遮挡的视觉效果。例如,可以根据图2中的实施例实现步骤130。In some embodiments, only the leg area can be accurately segmented by using the deep learning model, so as to determine the boundary area between the leg model and the shoe model, and then the occluded part of the shoe opening area can be accurately obtained as the cropping area. After the rendering process is displayed on the screen, the visual effect of virtual occlusion will be generated. For example, step 130 may be implemented according to the embodiment in FIG. 2 .
图2示出图1中步骤130的一些实施例的流程图。FIG. 2 shows a flowchart of some embodiments of step 130 in FIG. 1 .
如图2所示,步骤130包括:步骤1310,确定鞋的内轮廓和外轮廓;和步骤1320,确定被腿部遮挡部分。As shown in FIG. 2 , step 130 includes: step 1310 , determining the inner and outer contours of the shoe; and step 1320 , determining the portion occluded by the leg.
在步骤1310中,根据二维鞋图像中鞋体区域的位置确定鞋的外轮廓,根据二维图像中鞋口区域的位置确定鞋的内轮廓。In step 1310, the outer contour of the shoe is determined according to the position of the shoe body region in the two-dimensional shoe image, and the inner contour of the shoe is determined according to the position of the shoe mouth region in the two-dimensional image.
在一些实施例中,可以通过图3a中的实施例,确定鞋体区域和鞋口区域用于确定内轮廓和外轮廓。In some embodiments, the shoe body area and the shoe mouth area may be determined for use in determining the inner and outer contours using the embodiment of Figure 3a.
图3a示出本公开的虚拟图像的生成方法的一些实施例的示意图。Figure 3a shows a schematic diagram of some embodiments of the method for generating a virtual image of the present disclosure.
如图3a所示,对三维鞋模型中的鞋口区域进行透明化处理后,渲染为二维鞋图像。根据二维鞋图像的二值图像,确定鞋体区域31、鞋口区域32。As shown in Figure 3a, after transparent processing is performed on the shoe mouth area in the 3D shoe model, it is rendered into a 2D shoe image. According to the binary image of the two-dimensional shoe image, the shoe body area 31 and the shoe mouth area 32 are determined.
在确定了鞋体区域31和鞋口区域32后,即可确定外轮廓和内轮廓,进而通过图2中的其余步骤确定遮挡部分。After the shoe body area 31 and the shoe mouth area 32 are determined, the outer contour and the inner contour can be determined, and then the occlusion part can be determined through the remaining steps in FIG. 2 .
在步骤1320中,根据腿部区域的轮廓与外轮廓的交点位置、内轮廓,确定鞋区域中被腿部遮挡的部分。In step 1320, the part of the shoe area that is blocked by the leg is determined according to the position of the intersection of the contour of the leg area and the outer contour, and the inner contour.
在一些实施例中,可以通过图3b、3c中的实施例,确定鞋体区域和鞋口区域用于确定内轮廓和外轮廓。In some embodiments, the shoe body area and the shoe mouth area may be determined for use in determining the inner and outer contours through the embodiments in Figures 3b and 3c.
图3b示出本公开的虚拟图像的生成方法的一些实施例的示意图。Figure 3b shows a schematic diagram of some embodiments of the method for generating a virtual image of the present disclosure.
如图3b所示,利用神经网络模块,能够推理出腿部区域30的分割mask二值图像。结合图3a中鞋的二值图像和图3b中的腿部的二值图像,可以确定腿部区域30的轮廓与鞋体区域31的轮廓的交点。As shown in Figure 3b, using the neural network module, the segmentation mask binary image of the leg region 30 can be inferred. Combining the binary image of the shoe in FIG. 3 a and the binary image of the leg in FIG. 3 b , the intersection of the contour of the leg region 30 and the contour of the shoe body region 31 can be determined.
图3c示出本公开的虚拟图像的生成方法的一些实施例的示意图。Figure 3c shows a schematic diagram of some embodiments of the method for generating a virtual image of the present disclosure.
如图3c所示,鞋体区域31的轮廓为外轮廓311,鞋口区域32的轮廓为内轮廓321。腿部区域30的轮廓与外轮廓311的交点为3a、3b,内轮廓321上与交点3a最近的点为3d,与交点3b最近的点为3c。As shown in FIG. 3 c , the outline of the shoe body region 31 is an outer outline 311 , and the outline of the shoe mouth region 32 is an inner outline 321 . The intersection points of the contour of the leg region 30 and the outer contour 311 are 3a, 3b, the point closest to the intersection 3a on the inner contour 321 is 3d, and the closest point to the intersection 3b is 3c.
连接点3a、3b、3c、3d形成闭合区域(位于鞋体左侧),将该闭合区域确定为鞋区域中被腿部遮挡的部分。The connection points 3a, 3b, 3c, 3d form a closed area (located on the left side of the shoe body), which is defined as the part of the shoe area that is obscured by the legs.
在一些实施例中,也可以根据。腿部区域30的轮廓与内轮廓321的交点和3a、3b形成的闭合区域,确定鞋区域中被腿部遮挡的部分。In some embodiments, can also be based on. The intersection of the contour of the leg area 30 and the inner contour 321 and the closed area formed by 3a, 3b determine the part of the shoe area that is obscured by the leg.
确定了鞋区域中被腿部遮挡的部分,可以通过图1中的步骤140生成虚拟图像。After determining the part of the shoe area that is occluded by the leg, a virtual image can be generated through step 140 in FIG. 1 .
在步骤140中,根据被腿部遮挡的部分,渲染待处理图像和二维鞋图像的合成图像,生成具有腿部遮挡效果的虚拟图像。In step 140, a composite image of the to-be-processed image and the two-dimensional shoe image is rendered according to the part blocked by the leg, so as to generate a virtual image with a leg blocking effect.
在一些实施例中,待处理图像为视频中的各帧图像。可以根据生成的各帧图像对应的虚拟图像,生成具有腿部遮效果的视频。In some embodiments, the images to be processed are frames of images in the video. A video with a leg blocking effect can be generated according to the virtual image corresponding to each generated frame of image.
在上述实施例中,根据获取的真实腿部的视觉线索,结合的二维鞋图像中鞋的位置,准确地确定虚拟遮挡部分。这样,可以利用真实的腿部图像与三维鞋模型合成虚拟图像,提高虚拟图像的效果。In the above-mentioned embodiment, the virtual occlusion portion is accurately determined according to the obtained visual clues of the real leg and the position of the shoe in the combined two-dimensional shoe image. In this way, a virtual image can be synthesized by using the real leg image and the three-dimensional shoe model to improve the effect of the virtual image.
图4示出本公开的虚拟图像的生成装置的一些实施例的框图。FIG. 4 shows a block diagram of some embodiments of an apparatus for generating a virtual image of the present disclosure.
如图4所示,虚拟图像的生成装置4包括确定单元41、处理单元42。As shown in FIG. 4 , the virtual image generating apparatus 4 includes a determination unit 41 and a processing unit 42 .
确定单元41获取包含腿部和脚部的待处理图像中的腿部区域;根据腿部区域与二维鞋图像中鞋区域的重叠部分,确定鞋区域中被腿部遮挡的部分。The determination unit 41 acquires the leg area in the to-be-processed image including the legs and feet; and determines the part of the shoe area covered by the leg according to the overlapping part of the leg area and the shoe area in the two-dimensional shoe image.
在一些实施例中,确定单元41根据二维鞋图像中鞋体区域的位置确定鞋的外轮廓,根据二维图像中鞋口区域的位置确定鞋的内轮廓;根据腿部区域的轮廓与外轮廓的交点和内轮廓,确定鞋区域中被腿部遮挡的部分。In some embodiments, the determining unit 41 determines the outer contour of the shoe according to the position of the shoe body region in the two-dimensional shoe image, and determines the inner contour of the shoe according to the position of the shoe mouth region in the two-dimensional image; The intersection of the contours and the inner contour, determine the portion of the shoe area that is occluded by the leg.
在一些实施例中,确定单元41在内轮廓上,确定与交点距离最近的点;根据交点、距离最近的点,确定鞋区域中被腿部遮挡的部分。In some embodiments, the determining unit 41 determines the point closest to the intersection on the inner contour; and determines the part of the shoe area that is occluded by the leg according to the intersection and the closest point.
在一些实施例中,确定单元41将待处理图像输入机器学习模型,确定待处理图像中的腿部区域。In some embodiments, the determining unit 41 inputs the image to be processed into the machine learning model, and determines the leg region in the image to be processed.
在一些实施例中,机器学习模型包括依次连接的卷积神经网络模块和空间金字塔池化模块。In some embodiments, the machine learning model includes a convolutional neural network module and a spatial pyramid pooling module connected in sequence.
在一些实施例中,卷积神经网络模块根据Fast-SCNN模型设置。In some embodiments, the convolutional neural network module is set up according to the Fast-SCNN model.
处理单元42根据待处理图像中脚部的姿态参数和相机的内部参数,将三维鞋模型渲染为与待处理图像对应的二维鞋图像;根据被腿部遮挡的部分,渲染待处理图像和二维鞋图像的合成图像,生成具有腿部遮挡效果的虚拟图像。The processing unit 42 renders the three-dimensional shoe model as a two-dimensional shoe image corresponding to the to-be-processed image according to the posture parameters of the foot in the image to be processed and the internal parameters of the camera; A composite image of dimensional shoe images to generate a virtual image with a leg occlusion effect.
在一些实施例中,处理单元42对三维鞋模型中的鞋口区域进行透明化处理;将 透明化处理后的三维鞋模型,渲染为所述二维鞋图像。确定单元41根据二维鞋图像的二值图像,确定鞋体区域和鞋口区域。In some embodiments, the processing unit 42 performs transparency processing on the shoe mouth area in the three-dimensional shoe model; and renders the three-dimensional shoe model after the transparent processing as the two-dimensional shoe image. The determining unit 41 determines the shoe body area and the shoe opening area according to the binary image of the two-dimensional shoe image.
在一些实施例中,处理单元42检测三维鞋模型的鞋口区域,并利用封闭网格覆盖鞋口区域;对封闭网格覆盖部分进行透明化处理。In some embodiments, the processing unit 42 detects the shoe opening area of the three-dimensional shoe model, and covers the shoe opening area with a closed mesh; and performs transparency processing on the covering part of the closed mesh.
在一些实施例中,确定单元41根据预设的鞋区域中未被腿部遮挡的部分的大小,确定封闭网格覆盖鞋口区域的位置。In some embodiments, the determining unit 41 determines the position where the closed mesh covers the shoe opening area according to the size of the part of the shoe area that is not covered by the legs.
在一些实施例中,待处理图像为视频中的各帧图像;处理单元42根据生成的各帧图像对应的虚拟图像,生成具有腿部遮效果的视频。In some embodiments, the image to be processed is each frame of image in the video; the processing unit 42 generates a video with a leg blocking effect according to the generated virtual image corresponding to each frame of image.
图5示出本公开的虚拟图像的生成装置的另一些实施例的框图。FIG. 5 shows a block diagram of other embodiments of the virtual image generating apparatus of the present disclosure.
如图5所示,该实施例的虚拟图像的生成装置5包括:存储器51以及耦接至该存储器51的处理器52,处理器52被配置为基于存储在存储器51中的指令,执行本公开中任意一个实施例中的虚拟图像的生成方法。As shown in FIG. 5 , the virtual image generating apparatus 5 of this embodiment includes: a memory 51 and a processor 52 coupled to the memory 51 , and the processor 52 is configured to execute the present disclosure based on instructions stored in the memory 51 The method for generating a virtual image in any one of the embodiments.
其中,存储器51例如可以包括系统存储器、固定非易失性存储介质等。系统存储器例如存储有操作系统、应用程序、引导装载程序Boot Loader、数据库以及其他程序等。Wherein, the memory 51 may include, for example, a system memory, a fixed non-volatile storage medium, and the like. The system memory stores, for example, an operating system, an application program, a boot loader Boot Loader, a database, and other programs.
图6示出本公开的虚拟图像的生成装置的又一些实施例的框图。FIG. 6 shows a block diagram of further embodiments of the apparatus for generating virtual images of the present disclosure.
如图6所示,该实施例的虚拟图像的生成装置6包括:存储器610以及耦接至该存储器610的处理器620,处理器620被配置为基于存储在存储器610中的指令,执行前述任意一个实施例中的虚拟图像的生成方法。As shown in FIG. 6 , the virtual image generating apparatus 6 of this embodiment includes: a memory 610 and a processor 620 coupled to the memory 610 , and the processor 620 is configured to execute any of the foregoing based on the instructions stored in the memory 610 . A method for generating a virtual image in one embodiment.
存储器610例如可以包括系统存储器、固定非易失性存储介质等。系统存储器例如存储有操作系统、应用程序、引导装载程序Boot Loader以及其他程序等。 Memory 610 may include, for example, system memory, fixed non-volatile storage media, and the like. The system memory stores, for example, an operating system, an application program, a boot loader, and other programs.
虚拟图像的生成装置6还可以包括输入输出接口630、网络接口640、存储接口650等。这些接口630、640、650以及存储器610和处理器620之间例如可以通过总线660连接。其中,输入输出接口630为显示器、鼠标、键盘、触摸屏、麦克、音箱等输入输出设备提供连接接口。网络接口640为各种联网设备提供连接接口。存储接口650为SD卡、U盘等外置存储设备提供连接接口。The virtual image generating apparatus 6 may further include an input/output interface 630, a network interface 640, a storage interface 650, and the like. These interfaces 630 , 640 , 650 and the memory 610 and the processor 620 may be connected, for example, through a bus 660 . The input and output interface 630 provides a connection interface for input and output devices such as a display, a mouse, a keyboard, a touch screen, a microphone, and a speaker. Network interface 640 provides a connection interface for various networked devices. The storage interface 650 provides a connection interface for external storage devices such as SD cards and U disks.
本领域内的技术人员应当明白,本公开的实施例可提供为方法、系统、或计算机程序产品。因此,本公开可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本公开可采用在一个或多个其中包含有计算机可用程序代码的计算机可用非瞬时性存储介质包括但不限于磁盘存储器、CD-ROM、光学存 储器等上实施的计算机程序产品的形式。As will be appreciated by one skilled in the art, embodiments of the present disclosure may be provided as a method, system, or computer program product. Accordingly, the present disclosure may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present disclosure may take the form of a computer program product embodied on one or more computer-usable non-transitory storage media having computer-usable program code embodied therein, including but not limited to disk storage, CD-ROM, optical storage, and the like.
至此,已经详细描述了根据本公开的虚拟图像的生成方法、虚拟图像的生成装置和非易失性计算机可读存储介质。为了避免遮蔽本公开的构思,没有描述本领域所公知的一些细节。本领域技术人员根据上面的描述,完全可以明白如何实施这里公开的技术方案。So far, the method for generating a virtual image, the device for generating a virtual image, and the non-volatile computer-readable storage medium according to the present disclosure have been described in detail. Some details that are well known in the art are not described in order to avoid obscuring the concept of the present disclosure. Those skilled in the art can fully understand how to implement the technical solutions disclosed herein based on the above description.
可能以许多方式来实现本公开的方法和系统。例如,可通过软件、硬件、固件或者软件、硬件、固件的任何组合来实现本公开的方法和系统。用于方法的步骤的上述顺序仅是为了进行说明,本公开的方法的步骤不限于以上具体描述的顺序,除非以其它方式特别说明。此外,在一些实施例中,还可将本公开实施为记录在记录介质中的程序,这些程序包括用于实现根据本公开的方法的机器可读指令。因而,本公开还覆盖存储用于执行根据本公开的方法的程序的记录介质。The methods and systems of the present disclosure may be implemented in many ways. For example, the methods and systems of the present disclosure may be implemented in software, hardware, firmware, or any combination of software, hardware, and firmware. The above order of steps for the method is for illustration only, and the steps of the method of the present disclosure are not limited to the order specifically described above unless specifically stated otherwise. Furthermore, in some embodiments, the present disclosure can also be implemented as programs recorded in a recording medium, the programs including machine-readable instructions for implementing methods according to the present disclosure. Thus, the present disclosure also covers a recording medium storing a program for executing the method according to the present disclosure.
虽然已经通过示例对本公开的一些特定实施例进行了详细说明,但是本领域的技术人员应该理解,以上示例仅是为了进行说明,而不是为了限制本公开的范围。本领域的技术人员应该理解,可在不脱离本公开的范围和精神的情况下,对以上实施例进行修改。本公开的范围由所附权利要求来限定。While some specific embodiments of the present disclosure have been described in detail by way of examples, those skilled in the art will appreciate that the above examples are provided for illustration only, and are not intended to limit the scope of the present disclosure. Those skilled in the art will appreciate that modifications may be made to the above embodiments without departing from the scope and spirit of the present disclosure. The scope of the present disclosure is defined by the appended claims.

Claims (13)

  1. 一种虚拟图像的生成方法,包括:A method for generating a virtual image, comprising:
    获取包含腿部和脚部的待处理图像中的腿部区域;Get the leg region in the to-be-processed image that contains the legs and feet;
    根据所述待处理图像中脚部的姿态参数和相机的内部参数,将三维鞋模型渲染为与所述待处理图像对应的二维鞋图像;rendering the three-dimensional shoe model into a two-dimensional shoe image corresponding to the to-be-processed image according to the posture parameters of the foot in the to-be-processed image and the internal parameters of the camera;
    根据所述腿部区域与所述二维鞋图像中鞋区域的重叠部分,确定所述鞋区域中被腿部遮挡的部分;According to the overlapping portion of the leg region and the shoe region in the two-dimensional shoe image, determining the portion of the shoe region that is occluded by the leg;
    根据所述被腿部遮挡的部分,渲染所述待处理图像和所述二维鞋图像的合成图像,以生成具有腿部遮挡效果的虚拟图像。A composite image of the to-be-processed image and the two-dimensional shoe image is rendered according to the portion occluded by the leg, so as to generate a virtual image with a leg occlusion effect.
  2. 根据权利要求1所述的生成方法,其中,所述根据所述腿部区域与所述二维鞋图像中鞋区域的重叠部分,确定所述鞋区域中被腿部遮挡的部分包括:The generation method according to claim 1, wherein, according to the overlapping part of the leg region and the shoe region in the two-dimensional shoe image, determining the part of the shoe region that is occluded by the leg in the shoe region comprises:
    根据所述二维鞋图像中鞋体区域的位置确定所述二维鞋图像中鞋的外轮廓,根据所述二维图像中鞋口区域的位置确定所述二维鞋图像中鞋的内轮廓;The outer contour of the shoe in the two-dimensional shoe image is determined according to the position of the shoe body region in the two-dimensional shoe image, and the inner contour of the shoe in the two-dimensional shoe image is determined according to the position of the shoe mouth region in the two-dimensional image. ;
    根据所述腿部区域的轮廓与所述外轮廓的交点和所述内轮廓,确定所述鞋区域中被腿部遮挡的部分。According to the intersection of the contour of the leg area with the outer contour and the inner contour, the portion of the shoe area that is covered by the leg is determined.
  3. 根据权利要求2所述的生成方法,其中,所述根据所述腿部区域的轮廓与所述外轮廓的交点和所述内轮廓,确定所述鞋区域中被腿部遮挡的部分包括:The generation method according to claim 2, wherein, according to the intersection of the contour of the leg area and the outer contour and the inner contour, determining the part of the shoe area that is occluded by the leg includes:
    在所述内轮廓上,确定与所述交点距离最近的点;On the inner contour, determine the point closest to the intersection;
    根据所述交点、所述距离最近的点,确定所述鞋区域中被腿部遮挡的部分。According to the intersection point and the point with the closest distance, the part of the shoe area that is occluded by the leg is determined.
  4. 根据权利要求1所述的生成方法,其中,所述将三维鞋模型渲染为与所述待处理图像对应的二维鞋图像包括:The generation method according to claim 1, wherein the rendering of the three-dimensional shoe model into a two-dimensional shoe image corresponding to the to-be-processed image comprises:
    对所述三维鞋模型中的鞋口区域进行透明化处理;performing transparent processing on the shoe mouth area in the three-dimensional shoe model;
    将透明化处理后的三维鞋模型,渲染为所述二维鞋图像;Rendering the transparent three-dimensional shoe model into the two-dimensional shoe image;
    根据所述二维鞋图像的二值图像,确定所述二维鞋图像的鞋体区域和所述二维鞋图像的鞋口区域。According to the binary image of the two-dimensional shoe image, the shoe body area of the two-dimensional shoe image and the shoe mouth area of the two-dimensional shoe image are determined.
  5. 根据权利要求4所述的生成方法,其中,所述对所述三维鞋模型中的鞋口区域进行透明化处理包括:The generation method according to claim 4, wherein the transparent processing of the shoe mouth region in the three-dimensional shoe model comprises:
    检测所述三维鞋模型的鞋口区域,并利用封闭网格覆盖所述三维鞋模型的鞋口区域;Detecting the shoe mouth area of the three-dimensional shoe model, and covering the shoe mouth area of the three-dimensional shoe model with a closed mesh;
    对所述封闭网格覆盖部分进行透明化处理。Transparency processing is performed on the closed mesh covering portion.
  6. 根据权利要求5所述的生成方法,其中,所述检测所述三维鞋模型的鞋口区域,并利用封闭网格覆盖所述三维鞋模型的鞋口区域包括:The generation method according to claim 5, wherein the detecting the shoe opening area of the three-dimensional shoe model and covering the shoe opening area of the three-dimensional shoe model with a closed mesh comprises:
    根据预设的所述鞋区域中未被腿部遮挡的部分的大小,确定所述封闭网格覆盖所述三维鞋模型的鞋口区域的位置。According to the preset size of the portion of the shoe area that is not covered by the leg, the position where the closed mesh covers the shoe mouth area of the three-dimensional shoe model is determined.
  7. 根据权利要求1-6任一项所述的生成方法,其中,所述获取包含腿部和脚部的待处理图像中的腿部区域包括:The generation method according to any one of claims 1-6, wherein the acquiring the leg region in the to-be-processed image including the leg and the foot comprises:
    将所述待处理图像输入机器学习模型,确定所述待处理图像中的腿部区域。The to-be-processed image is input into a machine learning model, and the leg region in the to-be-processed image is determined.
  8. 根据权利要求7所述的生成方法,其中,The generating method according to claim 7, wherein,
    所述机器学习模型包括依次连接的卷积神经网络模块和空间金字塔池化模块。The machine learning model includes sequentially connected convolutional neural network modules and spatial pyramid pooling modules.
  9. 根据权利要求8所述的生成方法,其中,The generating method according to claim 8, wherein,
    所述卷积神经网络模块根据快速分割卷积神经网络Fast-SCNN模型设置。The convolutional neural network module is set according to the fast segmentation convolutional neural network Fast-SCNN model.
  10. 根据权利要求1-6任一项所述的生成方法,其中,所述待处理图像为视频中的各帧图像,The generation method according to any one of claims 1-6, wherein the to-be-processed image is each frame of image in the video,
    所述生成具有腿部遮挡效果的虚拟图像包括:The generating a virtual image with a leg occlusion effect includes:
    生成所述各帧图像对应的具有腿部遮挡效果的虚拟图像;generating a virtual image with a leg occlusion effect corresponding to each of the frame images;
    还包括:Also includes:
    根据所述各帧图像对应的虚拟图像,生成具有腿部遮效果的视频。According to the virtual image corresponding to each frame of image, a video with a leg blocking effect is generated.
  11. 一种虚拟图像的生成装置,包括:A device for generating virtual images, comprising:
    确定单元,用于获取包含腿部和脚部的待处理图像中的腿部区域,根据所述腿部区域与二维鞋图像中鞋区域的重叠部分,确定所述鞋区域中被腿部遮挡的部分;a determining unit, configured to acquire the leg area in the to-be-processed image including the leg and the foot, and determine, according to the overlapping part of the leg area and the shoe area in the two-dimensional shoe image, that the shoe area is occluded by the leg part;
    处理单元,用于根据所述待处理图像中脚部的姿态参数和相机的内部参数,将三维鞋模型渲染为与所述待处理图像对应的二维鞋图像,根据所述被腿部遮挡的部分,渲染所述待处理图像和所述二维鞋图像的合成图像,生成具有腿部遮挡效果的虚拟图像。The processing unit is configured to render the three-dimensional shoe model into a two-dimensional shoe image corresponding to the to-be-processed image according to the posture parameters of the foot in the to-be-processed image and the internal parameters of the camera, and according to the part, rendering a composite image of the to-be-processed image and the two-dimensional shoe image to generate a virtual image with a leg occlusion effect.
  12. 一种虚拟图像的生成装置,包括:A device for generating virtual images, comprising:
    存储器;和memory; and
    耦接至所述存储器的处理器,所述处理器被配置为基于存储在所述存储器中的指令,执行权利要求1-10任一项所述的虚拟图像的生成方法。A processor coupled to the memory, the processor configured to perform the method of generating a virtual image of any one of claims 1-10 based on instructions stored in the memory.
  13. 一种非易失性计算机可读存储介质,其上存储有计算机程序,该程序被处理 器执行时实现权利要求1-10任一项所述的虚拟图像的生成方法。A non-volatile computer-readable storage medium on which a computer program is stored, and when the program is executed by a processor, implements the method for generating a virtual image according to any one of claims 1-10.
PCT/CN2021/119751 2020-10-21 2021-09-23 Virtual image generation method and apparatus WO2022083389A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011134938.4A CN112330784A (en) 2020-10-21 2020-10-21 Virtual image generation method and device
CN202011134938.4 2020-10-21

Publications (1)

Publication Number Publication Date
WO2022083389A1 true WO2022083389A1 (en) 2022-04-28

Family

ID=74311276

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/119751 WO2022083389A1 (en) 2020-10-21 2021-09-23 Virtual image generation method and apparatus

Country Status (2)

Country Link
CN (1) CN112330784A (en)
WO (1) WO2022083389A1 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112330784A (en) * 2020-10-21 2021-02-05 北京沃东天骏信息技术有限公司 Virtual image generation method and device
CN113034655A (en) * 2021-03-11 2021-06-25 北京字跳网络技术有限公司 Shoe fitting method and device based on augmented reality and electronic equipment
JP7508145B1 (en) 2023-04-25 2024-07-01 株式会社wearcoord Coordination device, coordination system, coordination method, and coordination program

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2647305A1 (en) * 2010-12-03 2013-10-09 Alu Group, S.L. Method for virtually trying on footwear
CN110751716A (en) * 2019-05-08 2020-02-04 叠境数字科技(上海)有限公司 Virtual shoe fitting method based on single-view-angle RGBD sensor
CN111369686A (en) * 2020-03-03 2020-07-03 足购科技(杭州)有限公司 AR imaging virtual shoe fitting method and device capable of processing local shielding objects
CN111507806A (en) * 2020-04-23 2020-08-07 北京百度网讯科技有限公司 Virtual shoe fitting method, device, equipment and storage medium
CN112330784A (en) * 2020-10-21 2021-02-05 北京沃东天骏信息技术有限公司 Virtual image generation method and device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2647305A1 (en) * 2010-12-03 2013-10-09 Alu Group, S.L. Method for virtually trying on footwear
CN110751716A (en) * 2019-05-08 2020-02-04 叠境数字科技(上海)有限公司 Virtual shoe fitting method based on single-view-angle RGBD sensor
CN111369686A (en) * 2020-03-03 2020-07-03 足购科技(杭州)有限公司 AR imaging virtual shoe fitting method and device capable of processing local shielding objects
CN111507806A (en) * 2020-04-23 2020-08-07 北京百度网讯科技有限公司 Virtual shoe fitting method, device, equipment and storage medium
CN112330784A (en) * 2020-10-21 2021-02-05 北京沃东天骏信息技术有限公司 Virtual image generation method and device

Also Published As

Publication number Publication date
CN112330784A (en) 2021-02-05

Similar Documents

Publication Publication Date Title
WO2022083389A1 (en) Virtual image generation method and apparatus
JP7249390B2 (en) Method and system for real-time 3D capture and live feedback using a monocular camera
JP6742405B2 (en) Head-mounted display with facial expression detection function
JP5818773B2 (en) Image processing apparatus, image processing method, and program
JP4950787B2 (en) Image processing apparatus and method
KR101556992B1 (en) 3d scanning system using facial plastic surgery simulation
KR101339900B1 (en) Three dimensional montage generation system and method based on two dimensinal single image
US8933928B2 (en) Multiview face content creation
KR101560508B1 (en) Method and arrangement for 3-dimensional image model adaptation
CN104123749A (en) Picture processing method and system
JP2018530045A (en) Method for 3D reconstruction of objects from a series of images, computer-readable storage medium and apparatus configured to perform 3D reconstruction of objects from a series of images
WO2021078179A1 (en) Image display method and device
JP2011521357A (en) System, method and apparatus for motion capture using video images
JP2013235537A (en) Image creation device, image creation program and recording medium
US11138743B2 (en) Method and apparatus for a synchronous motion of a human body model
WO2012096907A1 (en) Mesh animation
CN115496863B (en) Short video generation method and system for scene interaction of movie and television intelligent creation
KR102570897B1 (en) Apparatus and method for generating 3-dimentional model
CN108734772A (en) High accuracy depth image acquisition methods based on Kinect fusion
KR101696007B1 (en) Method and device for creating 3d montage
WO2024088061A1 (en) Face reconstruction and occlusion region recognition method, apparatus and device, and storage medium
CN111783497A (en) Method, device and computer-readable storage medium for determining characteristics of target in video
JP7326965B2 (en) Image processing device, image processing program, and image processing method
CN107369209A (en) A kind of data processing method
WO2020183598A1 (en) Learning data generator, learning data generating method, and learning data generating program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21881808

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21881808

Country of ref document: EP

Kind code of ref document: A1