CN111369686A - AR imaging virtual shoe fitting method and device capable of processing local shielding objects - Google Patents

AR imaging virtual shoe fitting method and device capable of processing local shielding objects Download PDF

Info

Publication number
CN111369686A
CN111369686A CN202010138425.4A CN202010138425A CN111369686A CN 111369686 A CN111369686 A CN 111369686A CN 202010138425 A CN202010138425 A CN 202010138425A CN 111369686 A CN111369686 A CN 111369686A
Authority
CN
China
Prior art keywords
target
image
shoe
foot
try
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010138425.4A
Other languages
Chinese (zh)
Inventor
李汪洋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Huotui Technology Co ltd
Original Assignee
Zugou Technology Hangzhou Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zugou Technology Hangzhou Co ltd filed Critical Zugou Technology Hangzhou Co ltd
Priority to CN202010138425.4A priority Critical patent/CN111369686A/en
Publication of CN111369686A publication Critical patent/CN111369686A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/005General purpose rendering architectures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Graphics (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The invention discloses an AR imaging virtual shoe fitting method capable of processing local shelters and a device for realizing the method. The method mainly comprises the following steps: step 1, segmenting and identifying an ankle target, a instep target or a shoe target and a shelter target in a foot image and a shelter relation of the ankle target, the instep target or the shoe target and the shelter target through a Mask R-CNN neural network; calculating the predicted 6D pose of the instep or shoe target by using a PVnet algorithm; generating a try-on shoe image corresponding to the predicted 6D pose based on the 3D model of the try-on shoe, covering the try-on shoe image on the instep or the shoe target in the foot area image, and maintaining the shielding relation. This scheme can be to the discernment shoot in the image intentionally the foreign object get into the condition of sheltering from the foot and carry out effective processing, still can be complete accurately in shoot the image foot show the AR three-dimensional model of shoes, can restore the effect of sheltering from of these parts of trousers and ankle better simultaneously, improve the user and try on experience.

Description

AR imaging virtual shoe fitting method and device capable of processing local shielding objects
Technical Field
The invention relates to the technical field of AR (augmented reality), in particular to an AR imaging virtual shoe fitting method and device.
Background
At present, online shopping is a trend, consumers can not know information of various commodities and select favorite commodities, and payment is convenient and rapid. But has problems that: (1) the traditional shoe picture and video mode for the e-commerce cannot be used by taking the user as the center, and the display of the commodity is transferred by taking the attention point of the user as the transfer. (2) The user can not well judge whether the style of the shoes purchased on the network accords with the user, the goods return can be caused, and the conversion rate is reduced.
By applying the AR technology, a user can use the camera to shoot the steps in the shot image at any time and any place to increase the AR three-dimensional model of the shoes, try on various shoes, and can check the full appearance of the commodity at 360 degrees through moving and rotating the user and check the details of the commodity at the same time. However, the existing virtual shoe fitting technology needs to shoot an image of the whole foot to accurately identify the position of the foot. When there is the foot and is sheltered from by trousers, both feet are alternately, other people or other objects shelter from the foot, only expose less parts such as heel, the foot can't be discerned, when leading to rotatory foot to observe the whole effect of examination shoes, the shoes formation of image is not good, perhaps has the virtual examination shoes system can't normally work when shielding the thing beyond.
Meanwhile, when the user tries on the shoes, the user needs to consider the fitting condition of the shoes to other parts such as trousers, ankles and the like, the shielding effect of the parts is simply shielded, and the user trying experience is influenced. The best effect is that when the AR three-dimensional image of the try-on shoes is added into the camera image, the shielding effect of the parts of the trousers and the ankles can be well restored.
Disclosure of Invention
The invention aims to overcome the problem that the virtual shoe trying method and the virtual shoe trying system in the prior art cannot process the condition that a foot is shielded by a foreign object in a shot image, so that the practical operation is inconvenient, and provides an AR imaging virtual shoe trying method and an AR imaging virtual shoe trying device capable of processing a local shielding object.
In order to achieve the purpose, the invention adopts the following technical scheme:
the invention provides an AR imaging virtual shoe fitting method capable of processing local shelters, which comprises the following steps:
step 1, acquiring and calling a camera to capture a foot region image, segmenting and identifying an ankle target, a foot surface or a vamp target and a shielding object target in the foot image and a shielding relation of the ankle target, the foot surface or the vamp target and the shielding object target through a MaskR-CNN neural network;
step 2, calculating the predicted 6D pose of the instep or vamp target by using a PVnet algorithm;
and 3, generating a try-on shoe image corresponding to the predicted 6D pose based on the 3D model of the try-on shoe, covering the try-on shoe image on the foot surface or the vamp target in the foot area image, and maintaining the shielding relation, so that the rendered AR try-on effect image is displayed on the user terminal.
Preferably, the step 2 specifically comprises the following steps:
step 201, down-sampling the obtained foot area image through a ResNet network, up-sampling the down-sampled characteristics of the foot area image, and obtaining a segmentation graph of semantic segmentation of a foot surface or vamp target in the foot area image and a vector field pointing to a 2D key point in the foot surface or vamp target;
step 202, calculating the voting score of each pixel point to each 2D key point according to the vector from each pixel point to the 2D key point;
and 203, calculating the 6D pose of the instep or vamp target by using a PnP algorithm according to the mean value and covariance of the voting scores of all the 2D key points.
Preferably, the 2D keypoints are determined by using a farthest point sampling algorithm.
Preferably, in step 202, the vector calculation formula from each pixel to the 2D key point is as follows:
Figure BDA0002398150760000031
wherein p represents a pixel point, xK represents a 2D key point;
the voting score calculation formula of each pixel point to each key point is as follows:
Figure BDA0002398150760000032
wherein k and i are two pixel points for determining possible key, h (k, i) is a possible key point, and p is a pixel point.
Preferably, the step 203 specifically includes the following steps:
calculating the mean value of the voting scores of all the 2D key points, wherein the calculation formula is as follows:
Figure BDA0002398150760000033
where h (k, i) is a possible key point, Wk,iIs a voting score;
and (3) calculating the covariance of the voting scores of all the 2D key points, wherein the calculation formula is as follows:
Figure BDA0002398150760000034
and performing algorithm calculation on the 6D pose by using PnP, wherein the 6D pose is calculated by minimizing the Mahalanobis distance:
Figure BDA0002398150760000035
wherein XKIs the coordinates of the 3D key points,
Figure BDA0002398150760000036
is XKThe 2D mapping of (a) to (b),
and obtaining the mapping relation between the 2D key points in the foot image and the 3D key points of the preset foot object, and obtaining the 6D pose of the user foot in the foot image.
Preferably, the foot area image includes a picture of a current frame of the user foot area image acquired by the user camera.
Preferably, the step 3 specifically comprises the following steps:
step 301, loading a 3D model of a try-on shoe through a network or a local place;
step 302, acquiring a try-on shoe image when the 3D model of the try-on shoe is in the predicted 6D pose of the instep or vamp target;
step 303, covering the image of the try-on shoe at the corresponding position of the instep or vamp target in the image of the foot area, simultaneously shielding the ankle target and the shelter target at the corresponding position of the image of the try-on shoe according to the shielding relationship of the ankle target, the instep or vamp target and the shelter target, and displaying the image on the user terminal after rendering.
The invention provides an AR imaging virtual shoe fitting device capable of processing local shelters, which is used for realizing the method and comprises the following steps:
the segmentation module is used for acquiring and calling a camera to capture a foot region image, segmenting and identifying an ankle target, a foot surface or an upper target and a shielding object target in the foot image and a shielding relation of the ankle target, the foot surface or the upper target and the shielding object target through a MaskR-CNN neural network; (ii) a
The pose prediction module is used for calculating the predicted 6D pose of the instep or vamp target by utilizing a PVnet algorithm;
and the rendering module is used for generating a try-on shoe image corresponding to the predicted 6D pose based on the 3D model of the try-on shoe, covering the try-on shoe image on the foot surface or the vamp target in the foot area image, and maintaining the shielding relation, so that the rendered AR try-on effect image is displayed on the user terminal.
Preferably, the pose prediction module includes:
the sampling unit is used for carrying out down-sampling on the obtained foot area image through a ResNet network, carrying out up-sampling on the down-sampled characteristics of the foot area image, and obtaining a segmentation map for semantic segmentation of a foot surface or vamp target in the foot area image and a vector field pointing to a 2D key point in the foot surface or vamp target;
the voting unit is used for calculating the voting score of each pixel point to each 2D key point according to the vector from each pixel point to the 2D key point;
and the pose unit is used for calculating the 6D pose of the instep or vamp target by using a PnP algorithm according to the mean value and covariance of the voting scores of all the 2D key points.
Preferably, the rendering module includes:
the loading unit is used for loading the 3D model of the try-on shoe through a network or locally;
the image unit is used for acquiring a try-on shoe image when the 3D model of the try-on shoe is in the predicted 6D pose of the instep or vamp target;
and the rendering unit is used for covering the positions corresponding to the instep or vamp targets in the foot area image with the try-on shoe image, simultaneously shielding the ankle target and the shelter target at the corresponding positions of the try-on shoe image according to the shielding relation of the ankle target, the instep or vamp target and the shelter target, and displaying the image on the user terminal after rendering.
The method can eliminate the interference of other unrelated objects when the feet of the user are shielded by unexpected articles (such as trouser legs and the like). And detecting key points by using local information of the visible part of the instep or the vamp target. Then, each pixel predicts a direction vector pointing to the key point of the object, thereby judging the pose of the foot of the user. And superposing the 3D model image of the try-on shoe prestored in the server to the foot position on the image, thereby realizing the effect of virtual try-on. Simultaneously, according to the shielding relation between the ankle target, the instep or the vamp target and the shielding object target, the ankle target and the shielding object target are shielded at the corresponding positions of the images of the try-on shoes according to the shielding relation, and the shielding effects of the parts of the trousers and the ankles can be well restored.
Drawings
FIG. 1 is a first flowchart of an AR imaging virtual shoe fitting method of the present invention capable of handling local occlusions.
FIG. 2 is a second flowchart of the AR imaging virtual shoe fitting method of the present invention capable of handling local occlusions.
FIG. 3 is a third flowchart of the AR imaging virtual shoe fitting method of the present invention capable of handling local occlusions.
FIG. 4 is a functional block diagram of an AR imaging virtual shoe fitting apparatus of the present invention capable of handling local obstructions.
FIG. 5 is another functional block diagram of an AR imaging virtual shoe fitting apparatus of the present invention capable of handling local obstructions.
Detailed Description
The invention is further described with reference to the following figures and detailed description.
As shown in fig. 1, the present application provides an AR imaging virtual shoe fitting method capable of handling local obstructions, the method comprising the steps of:
step 1, acquiring and calling a camera to capture a foot region image, segmenting and identifying an ankle target, a foot surface or an upper target and a shielding object target in the foot image and a shielding relation of the ankle target, the foot surface or the upper target and the shielding object target through a MaskR-CNN neural network.
The foot area image comprises a still picture of the foot area of the user shot by the user camera or an acquired picture of the current frame of the image of the foot area of the user. Therefore, the method can perform AR virtual shoe fitting on the pictures and real-time images shot by the user. Dividing and identifying an ankle target by using a MaskR-CNN neural network, wherein the ankle target refers to the identified ankle of the user; identifying the instep or upper is targeted to the shoe in the case of distinguishing the user wearing the shoe so that the test shoe image can be rendered at the position of the shoe, or identifying the instep in the case of the user wearing the foot so that the test shoe image can be rendered at the position of the instep. And the shade target includes the bottom of a trouser leg or other accidental shade. Meanwhile, the maskR-CNN neural network can be used for identifying the shielding relation between an ankle target, a instep target or an upper target and a shielding object target, namely the shielding relation between the ankle, the upper, the trouser leg and an accidental shielding object from inside to outside.
And 2, calculating the predicted 6D pose of the instep or vamp target by utilizing a PVnet algorithm.
And 3, generating a try-on shoe image corresponding to the predicted 6D pose based on the 3D model of the try-on shoe, covering the try-on shoe image on the foot surface or the vamp target in the foot area image, and maintaining the shielding relation, so that the rendered AR try-on effect image is displayed on the user terminal.
As shown in fig. 2, preferably, the step 2 specifically includes the following steps:
step 201, down-sampling the obtained foot area image through a ResNet network, up-sampling the down-sampled characteristics of the foot area image, and obtaining a segmentation graph of semantic segmentation of a foot surface or vamp target in the foot area image and a vector field pointing to a 2D key point in the foot surface or vamp target.
The 2D key points are determined by adopting a farthest point sampling algorithm.
Step 202, calculating the voting score of each pixel point to each 2D key point according to the vector from each pixel point to the 2D key point.
In step 202, the vector calculation formula from each pixel to the 2D keypoint is:
Figure BDA0002398150760000071
wherein p represents a pixel point, XKRepresenting 2D keypoints;
the voting score calculation formula of each pixel point to each key point is as follows:
Figure BDA0002398150760000072
wherein k and i are two pixel points for determining possible key, h (k, i) is a possible key point, and p is a pixel point.
And 203, calculating the 6D pose of the instep or vamp target by using a PnP algorithm according to the mean value and covariance of the voting scores of all the 2D key points.
Said step 203 further comprises the steps of:
calculating the mean value of the voting scores of all the 2D key points, wherein the calculation formula is as follows:
Figure BDA0002398150760000073
where h (k, i) is a possible key point, Wk,iIs a voting score;
and (3) calculating the covariance of the voting scores of all the 2D key points, wherein the calculation formula is as follows:
Figure BDA0002398150760000074
and performing algorithm calculation on the 6D pose by using PnP, wherein the 6D pose is calculated by minimizing the Mahalanobis distance:
Figure BDA0002398150760000081
wherein XKIs the coordinates of the 3D key points,
Figure BDA0002398150760000082
is XKThe 2D mapping of (a) to (b),
and obtaining the mapping relation between the 2D key points in the foot image and the 3D key points of the preset foot object, and obtaining the 6D pose of the user foot in the foot image.
In the prior art, for identification and judgment of the 6D pose of an object in a space, an image of the object needs to be acquired completely, and the effect is poor when the object is partially shielded. However, when virtual try-on is performed, the vamp or the instep may be shielded by the bottoms of trousers or other objects, so that the pose identification is inaccurate, and the virtual try-on effect is deviated. According to the technical scheme of the step 2, when the target of the instep or the vamp is partially shielded by the trouser legs or other accidental shielding objects, the instep or the vamp can still be identified, and the 6D position of the instep or the vamp can be judged, so that an accurate basis can be provided for generating the image of the try-on shoe according to the 6D position, and the try-on effect can be faithfully rendered.
As shown in fig. 3, preferably, the step 3 specifically includes the following steps:
step 301, loading a 3D model of a try-on shoe through a network or a local place;
step 302, acquiring a try-on shoe image when the 3D model of the try-on shoe is in the predicted 6D pose of the instep or vamp target;
step 303, covering the image of the try-on shoe at the corresponding position of the instep or vamp target in the image of the foot area, simultaneously shielding the ankle target and the shelter target at the corresponding position of the image of the try-on shoe according to the shielding relationship of the ankle target, the instep or vamp target and the shelter target, and displaying the image on the user terminal after rendering.
The technical scheme of the step 3 is that by utilizing the shielding relation between the ankle target, the instep or the vamp target and the shielding object target identified in the step 1, when the trying-on shoe image is covered on the instep or inclined plane target for rendering, the shielding relation of the shielding objects such as the trouser legs and the like on the instep or the vamp target and the shielding relation of the vamp target for shielding the ankle target are maintained. Therefore, the effect of trying on the feet of the shoes can be simulated more truly, and the user can see the matching effect of the trying on shoes and own socks or trouser legs.
As shown in fig. 4, the present application provides an AR imaging virtual shoe fitting apparatus capable of handling local obstructions, the apparatus being used for implementing the above method, the apparatus comprising:
the segmentation module 1 is used for acquiring and calling a camera to capture a foot region image, segmenting and identifying an ankle target, a foot surface or an upper target and a shielding object target in the foot image and a shielding relation between the ankle target, the foot surface or the upper target and the shielding object target through a MaskR-CNN neural network.
And the pose prediction module 2 is used for calculating the predicted 6D pose of the instep or vamp target by utilizing a PVnet algorithm.
And the rendering module 3 is used for generating a try-on shoe image corresponding to the predicted 6D pose based on the 3D model of the try-on shoe, covering the try-on shoe image on a foot surface or a vamp target in the foot area image, and maintaining the shielding relation, so that the rendered AR try-on effect image is displayed on the user terminal.
As shown in fig. 5, preferably, the pose prediction module 2 includes:
the sampling unit 201 is configured to perform downsampling on the obtained foot area image through a ResNet network, perform upsampling on the downsampled features of the foot area image, and obtain a segmentation map of semantic segmentation of the instep or upper target in the foot area image and a vector field pointing to a 2D key point in the instep or upper target.
And the voting unit 202 is configured to calculate a voting score of each pixel point for each 2D key point according to the vector from each pixel point to the 2D key point.
And the pose unit 203 is used for calculating the 6D pose of the instep or vamp target by using a PnP algorithm according to the mean value and covariance of the voting scores of all the 2D key points.
The rendering module 3 comprises:
a loading unit 301, configured to load the 3D model of the try-on shoe through a network or locally.
An image unit 302, configured to obtain an image of the try-on shoe when the 3D model of the try-on shoe is at the predicted 6D pose of the instep or upper target.
And the rendering unit 303 is configured to cover the image of the try-on shoe at a position corresponding to the instep or upper target in the image of the foot area, and simultaneously, according to a shielding relationship between the ankle target, the instep or upper target and the shielding object target, shield the ankle target and the shielding object target at the corresponding position of the image of the try-on shoe according to the shielding relationship, and display the image on the user terminal after rendering.
The method can eliminate the interference of other unrelated objects when the feet of the user are shielded by unexpected articles (such as trouser legs and the like). And detecting key points by using local information of the visible part of the instep or the vamp target. Then, each pixel predicts a direction vector pointing to the key point of the object, thereby judging the pose of the foot of the user. And superposing the 3D model image of the try-on shoe prestored in the server to the foot position on the image, thereby realizing the effect of virtual try-on. Simultaneously, according to the shielding relation between the ankle target, the instep or the vamp target and the shielding object target, the ankle target and the shielding object target are shielded at the corresponding positions of the images of the try-on shoes according to the shielding relation, and the shielding effects of the parts of the trousers and the ankles can be well restored.
The method can eliminate the interference of other unrelated objects when the feet of the user are shielded by unexpected articles (such as trouser legs and the like). And detecting the key points by using the local information of the visible part of the object. Then, each pixel predicts a direction vector pointing to the key point of the object, thereby judging the pose of the foot of the user. And superposing the 3D model image of the try-on shoe prestored in the server to the foot position on the image, thereby realizing the effect of virtual try-on.
The new keypoint locating method has three advantages over the existing keypoint detection method. One is that the existing method only predicts key points once, but in the method, pixels of the visible part of an object have one prediction on the key points, so that the robustness of the model is greatly improved. The second advantage is that we use the properties of the user's foot to a large extent for the representation of the direction vector field of the object's key points. When the user's foot sees a part of the exposed object, the direction of the other part of the object can be estimated. Through the representation of the direction vector field, the network can be helped to learn the structural properties of the foot of the user. The third advantage is that the current method can only represent key points in the picture, and the direction vector field can detect key points outside the picture, so that the 6D pose of the foot in the trunk state can be detected.
Meanwhile, the MaskR-CNN neural network is applied to the scheme, the Mask-RCNN is an image segmentation model of the convolutional neural network based on deep learning, and can segment and identify ankle targets, instep targets or vamp targets and shielding object targets in foot images, wherein the shielding object targets comprise trousers and other accidental shielding objects, and the ankle targets and the shielding object targets are shielded at corresponding positions of the trying-on shoe images according to the shielding relation of the ankle targets, the instep targets or the vamp targets and the shielding object targets, so that the shielding effects of the trousers and the ankle can be well restored.

Claims (10)

1. An AR imaging virtual shoe fitting method capable of processing local shelters is characterized by comprising the following steps:
step 1, acquiring and calling a camera to capture a foot region image, segmenting and identifying an ankle target, a foot surface or a vamp target and a shielding object target in the foot image and a shielding relation of the ankle target, the foot surface or the vamp target and the shielding object target through a Mask R-CNN neural network;
step 2, calculating the predicted 6D pose of the instep or vamp target by using a PVnet algorithm;
and 3, generating a try-on shoe image corresponding to the predicted 6D pose based on the 3D model of the try-on shoe, covering the try-on shoe image on the foot surface or the vamp target in the foot area image, and maintaining the shielding relation, so that the rendered AR try-on effect image is displayed on the user terminal.
2. The AR imaging virtual shoe fitting method capable of processing local shelters according to claim 1, characterized in that the step 2 specifically comprises the following steps:
step 201, down-sampling the obtained foot area image through a ResNet network, up-sampling the down-sampled characteristics of the foot area image, and obtaining a segmentation graph of semantic segmentation of a foot surface or vamp target in the foot area image and a vector field pointing to a 2D key point in the foot surface or vamp target;
step 202, calculating the voting score of each pixel point to each 2D key point according to the vector from each pixel point to the 2D key point;
and 203, calculating the 6D pose of the instep or vamp target by using a PnP algorithm according to the mean value and covariance of the voting scores of all the 2D key points.
3. The AR imaging virtual shoe fitting method capable of handling local obstructions of claim 2, wherein the 2D key points are determined using a farthest point sampling algorithm.
4. The AR imaging virtual shoe fitting method capable of handling local obstructions of claim 2,
in step 202, the vector calculation formula from each pixel to the 2D keypoint is:
Figure FDA0002398150750000021
wherein p represents a pixel point, xK represents a 2D key point;
the voting score calculation formula of each pixel point to each key point is as follows:
Figure FDA0002398150750000022
wherein k and i are two pixel points for determining possible key, h (k, i) is a possible key point, and p is a pixel point.
5. The AR imaging virtual shoe fitting method capable of processing local shelters according to claim 2, characterized in that said step 203 comprises the following steps:
calculating the mean value of the voting scores of all the 2D key points, wherein the calculation formula is as follows:
Figure FDA0002398150750000023
where h (k, i) is a possible key point, Wk,iIs a voting score;
and (3) calculating the covariance of the voting scores of all the 2D key points, wherein the calculation formula is as follows:
Figure FDA0002398150750000024
and performing algorithm calculation on the 6D pose by using PnP, wherein the 6D pose is calculated by minimizing the Mahalanobis distance:
Figure FDA0002398150750000025
wherein XKIs the coordinates of the 3D key points,
Figure FDA0002398150750000026
is XKThe 2D mapping of (a) to (b),
and obtaining the mapping relation between the 2D key points in the foot image and the 3D key points of the preset foot object, and obtaining the 6D pose of the user foot in the foot image.
6. The AR imaging virtual shoe fitting method capable of handling local obstructions of claim 1, wherein the foot area image comprises a picture of a current frame of a user foot area image acquired by a user camera.
7. The AR imaging virtual shoe fitting method capable of processing local shelters according to claim 1, characterized in that said step 3 comprises the following steps:
step 301, loading a 3D model of a try-on shoe through a network or a local place;
step 302, acquiring a try-on shoe image when the 3D model of the try-on shoe is in the predicted 6D pose of the instep or vamp target;
step 303, covering the image of the try-on shoe at the corresponding position of the instep or vamp target in the image of the foot area, simultaneously shielding the ankle target and the shelter target at the corresponding position of the image of the try-on shoe according to the shielding relationship of the ankle target, the instep or vamp target and the shelter target, and displaying the image on the user terminal after rendering.
8. An AR imaging virtual shoe fitting apparatus capable of handling local obstructions, the apparatus comprising:
the segmentation module is used for acquiring and calling a camera to capture a foot region image, segmenting and identifying an ankle target, a foot surface or an upper target and a shielding object target in the foot image and a shielding relation of the ankle target, the foot surface or the upper target and the shielding object target through a Mask R-CNN neural network; (ii) a
The pose prediction module is used for calculating the predicted 6D pose of the instep or vamp target by utilizing a PVnet algorithm;
and the rendering module is used for generating a try-on shoe image corresponding to the predicted 6D pose based on the 3D model of the try-on shoe, covering the try-on shoe image on the foot surface or the vamp target in the foot area image, and maintaining the shielding relation, so that the rendered AR try-on effect image is displayed on the user terminal.
9. The AR imaging virtual shoe fitting apparatus capable of handling local obstructions of claim 8, wherein the pose prediction module comprises:
the sampling unit is used for carrying out down-sampling on the obtained foot area image through a ResNet network, carrying out up-sampling on the down-sampled characteristics of the foot area image, and obtaining a segmentation map for semantic segmentation of a foot surface or vamp target in the foot area image and a vector field pointing to a 2D key point in the foot surface or vamp target;
the voting unit is used for calculating the voting score of each pixel point to each 2D key point according to the vector from each pixel point to the 2D key point;
and the pose unit is used for calculating the 6D pose of the instep or vamp target by using a PnP algorithm according to the mean value and covariance of the voting scores of all the 2D key points.
10. The AR imaging virtual shoe fitting apparatus capable of handling local obstructions of claim 8, wherein the rendering module comprises:
the loading unit is used for loading the 3D model of the try-on shoe through a network or locally;
the image unit is used for acquiring a try-on shoe image when the 3D model of the try-on shoe is in the predicted 6D pose of the instep or vamp target;
and the rendering unit is used for covering the positions corresponding to the instep or vamp targets in the foot area image with the try-on shoe image, simultaneously shielding the ankle target and the shelter target at the corresponding positions of the try-on shoe image according to the shielding relation of the ankle target, the instep or vamp target and the shelter target, and displaying the image on the user terminal after rendering.
CN202010138425.4A 2020-03-03 2020-03-03 AR imaging virtual shoe fitting method and device capable of processing local shielding objects Pending CN111369686A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010138425.4A CN111369686A (en) 2020-03-03 2020-03-03 AR imaging virtual shoe fitting method and device capable of processing local shielding objects

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010138425.4A CN111369686A (en) 2020-03-03 2020-03-03 AR imaging virtual shoe fitting method and device capable of processing local shielding objects

Publications (1)

Publication Number Publication Date
CN111369686A true CN111369686A (en) 2020-07-03

Family

ID=71211674

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010138425.4A Pending CN111369686A (en) 2020-03-03 2020-03-03 AR imaging virtual shoe fitting method and device capable of processing local shielding objects

Country Status (1)

Country Link
CN (1) CN111369686A (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111882408A (en) * 2020-09-27 2020-11-03 北京达佳互联信息技术有限公司 Virtual trial method and device, electronic equipment and storage equipment
CN111985995A (en) * 2020-08-14 2020-11-24 足购科技(杭州)有限公司 WeChat applet-based shoe virtual fitting method and device
CN112330784A (en) * 2020-10-21 2021-02-05 北京沃东天骏信息技术有限公司 Virtual image generation method and device
CN112562063A (en) * 2020-12-08 2021-03-26 北京百度网讯科技有限公司 Method, device, equipment and storage medium for carrying out three-dimensional attempt on object
CN113763440A (en) * 2021-04-26 2021-12-07 腾讯科技(深圳)有限公司 Image processing method, device, equipment and storage medium
CN114445601A (en) * 2022-04-08 2022-05-06 北京大甜绵白糖科技有限公司 Image processing method, device, equipment and storage medium
WO2022171020A1 (en) * 2021-02-10 2022-08-18 北京字跳网络技术有限公司 Image display method and apparatus, device, and medium
WO2022188708A1 (en) * 2021-03-11 2022-09-15 北京字跳网络技术有限公司 Shoe try-on method and apparatus based on augmented reality, and electronic device
CN115174985A (en) * 2022-08-05 2022-10-11 北京字跳网络技术有限公司 Special effect display method, device, equipment and storage medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170076362A1 (en) * 2015-09-11 2017-03-16 Zachary Joyner Future Kicks
JP2017146909A (en) * 2016-02-19 2017-08-24 株式会社アスコン Try-on simulation system and program
US20170323374A1 (en) * 2016-05-06 2017-11-09 Seok Hyun Park Augmented reality image analysis methods for the virtual fashion items worn
US20190034996A1 (en) * 2017-07-27 2019-01-31 ELSE CORP S.r.I. Shoe last selection method, based on virtual fitting simulation and customer feedback
US20190180410A1 (en) * 2017-12-11 2019-06-13 Youspace, Inc. Object separation for scanned assets
US20200000180A1 (en) * 2017-02-18 2020-01-02 Digital Animal Interactive Inc. System, method, and apparatus for modelling feet and selecting footwear
US20200065991A1 (en) * 2018-08-21 2020-02-27 National Tsing Hua University Method and system of virtual footwear try-on with improved occlusion

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170076362A1 (en) * 2015-09-11 2017-03-16 Zachary Joyner Future Kicks
JP2017146909A (en) * 2016-02-19 2017-08-24 株式会社アスコン Try-on simulation system and program
US20170323374A1 (en) * 2016-05-06 2017-11-09 Seok Hyun Park Augmented reality image analysis methods for the virtual fashion items worn
US20200000180A1 (en) * 2017-02-18 2020-01-02 Digital Animal Interactive Inc. System, method, and apparatus for modelling feet and selecting footwear
US20190034996A1 (en) * 2017-07-27 2019-01-31 ELSE CORP S.r.I. Shoe last selection method, based on virtual fitting simulation and customer feedback
US20190180410A1 (en) * 2017-12-11 2019-06-13 Youspace, Inc. Object separation for scanned assets
US20200065991A1 (en) * 2018-08-21 2020-02-27 National Tsing Hua University Method and system of virtual footwear try-on with improved occlusion

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
CLAUDIA REDAELLI ET AL.: "Shoe customers\' behaviour with new technologies: the Magic Mirror case" *
P. EISERT ET AL.: "3-D Tracking of Shoes for Virtual Mirror Applications" *

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111985995A (en) * 2020-08-14 2020-11-24 足购科技(杭州)有限公司 WeChat applet-based shoe virtual fitting method and device
CN111882408A (en) * 2020-09-27 2020-11-03 北京达佳互联信息技术有限公司 Virtual trial method and device, electronic equipment and storage equipment
CN111882408B (en) * 2020-09-27 2021-07-09 北京达佳互联信息技术有限公司 Virtual trial method and device, electronic equipment and storage medium
CN112330784A (en) * 2020-10-21 2021-02-05 北京沃东天骏信息技术有限公司 Virtual image generation method and device
WO2022083389A1 (en) * 2020-10-21 2022-04-28 北京沃东天骏信息技术有限公司 Virtual image generation method and apparatus
CN112562063A (en) * 2020-12-08 2021-03-26 北京百度网讯科技有限公司 Method, device, equipment and storage medium for carrying out three-dimensional attempt on object
WO2022171020A1 (en) * 2021-02-10 2022-08-18 北京字跳网络技术有限公司 Image display method and apparatus, device, and medium
WO2022188708A1 (en) * 2021-03-11 2022-09-15 北京字跳网络技术有限公司 Shoe try-on method and apparatus based on augmented reality, and electronic device
CN113763440A (en) * 2021-04-26 2021-12-07 腾讯科技(深圳)有限公司 Image processing method, device, equipment and storage medium
CN114445601A (en) * 2022-04-08 2022-05-06 北京大甜绵白糖科技有限公司 Image processing method, device, equipment and storage medium
CN115174985A (en) * 2022-08-05 2022-10-11 北京字跳网络技术有限公司 Special effect display method, device, equipment and storage medium
CN115174985B (en) * 2022-08-05 2024-01-30 北京字跳网络技术有限公司 Special effect display method, device, equipment and storage medium

Similar Documents

Publication Publication Date Title
CN111369686A (en) AR imaging virtual shoe fitting method and device capable of processing local shielding objects
US10621779B1 (en) Artificial intelligence based generation and analysis of 3D models
US10757373B2 (en) Method and system for providing at least one image captured by a scene camera of a vehicle
US8036416B2 (en) Method and apparatus for augmenting a mirror with information related to the mirrored contents and motion
US6188777B1 (en) Method and apparatus for personnel detection and tracking
CN105404888B (en) The conspicuousness object detection method of color combining and depth information
JP2015181042A (en) detection and tracking of moving objects
JP5439787B2 (en) Camera device
JPH11259660A (en) Three-dimensional operation restoration system
US20220284492A1 (en) Methods of capturing images and making garments
JP7162750B2 (en) Image processing device, image processing method, and program
JP7499280B2 (en) Method and system for monocular depth estimation of a person - Patents.com
CN110751716B (en) Virtual shoe test method based on single-view RGBD sensor
Hayashi et al. Occlusion detection of real objects using contour based stereo matching
GB2509783A (en) System and method for foot tracking
Wientapper et al. Composing the feature map retrieval process for robust and ready-to-use monocular tracking
Rim et al. Real-time human pose estimation using RGB-D images and deep learning
Le et al. Overlay upper clothing textures to still images based on human pose estimation
CN115345927A (en) Exhibit guide method and related device, mobile terminal and storage medium
CN114445601A (en) Image processing method, device, equipment and storage medium
CN111597893B (en) Pedestrian image matching method and device, storage medium and terminal
EP4160533A1 (en) Estimation program, estimation method, and estimation device
Malik et al. Private Try-On Using Marker Based Augmented Reality
Sundari et al. Development of Apparel 360°-An AR based Virtual Trial Room
Ning et al. Markerless client-server augmented reality system with natural features

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20220831

Address after: Room 308-2, Building 1, Flying Bird Business Center, Wuchang Street, Yuhang District, Hangzhou City, Zhejiang Province, 310000

Applicant after: Hangzhou Huotui Technology Co.,Ltd.

Address before: 310000 Yuhang District, Hangzhou City, Zhejiang Province

Applicant before: Zugou Technology (Hangzhou) Co.,Ltd.

AD01 Patent right deemed abandoned
AD01 Patent right deemed abandoned

Effective date of abandoning: 20231229