US20170309075A1 - Image to item mapping - Google Patents
Image to item mapping Download PDFInfo
- Publication number
- US20170309075A1 US20170309075A1 US15/526,629 US201515526629A US2017309075A1 US 20170309075 A1 US20170309075 A1 US 20170309075A1 US 201515526629 A US201515526629 A US 201515526629A US 2017309075 A1 US2017309075 A1 US 2017309075A1
- Authority
- US
- United States
- Prior art keywords
- image
- digital image
- item
- manipulated
- reference item
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/003—Reconstruction from projections, e.g. tomography
- G06T11/006—Inverse problem, transformation from projection-space into object-space, e.g. transform methods, back-projection, algebraic methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2210/00—Indexing scheme for image generation or computer graphics
- G06T2210/16—Cloth
Definitions
- the invention relates to the display of a digital image on a representation of an object, and particularly but not exclusively to the display of a digital image representing a design of an item of clothing onto a person.
- a digital image of an item of clothing may be displayed onto an image of a person.
- a problem associated with displaying images onto representations of objects is to facilitate the display in such a way that the resulting displayed image looks realistic.
- a perfect representation of an image when displayed on a representation of an object, may need not to be perfect, to reflect the fact that the object “carries” the image in a particular way.
- the displayed image needs to reflect the fact that the person is wearing an item and the item is adjusted according to how they are wearing it.
- a method of displaying a digital image on a representation of an object comprising: defining reference point positions on a digital image template of a reference item; fitting the reference item to an object; capturing an image of the reference item fitted to the object; mapping positions on a digital image template of a non-reference item to the location positions of the reference points of the captured image to generate a manipulated non-reference digital image; and displaying the manipulated non-reference digital image on the reference item fitted to the object.
- the representation of the object may be the captured image of the object, and the manipulated non-reference digital image is displayed on the captured image.
- the representation of the object may be the object, and the manipulated non-reference digital image is projected on to the object.
- the mapping step may comprise detecting the reference point positions on the captured image and mapping the positions to location points, wherein corresponding points on the digital image template of a non-reference item are mapped to the associated location points to manipulate the non-reference image for display on the representation of the object.
- the non-reference digital image may be a catalogue digital image.
- Displaying the manipulated digital image on the reference item fitted to the representation of the object may further comprise combining the manipulated digital image with the captured image of the reference item fitted to the object.
- the manipulated digital image and the captured image may be combined by digitally merging them.
- the combined image may be displayed in a virtual reality mirror.
- the manipulated digital image and the captured image are combined by digitally merging them, the combined image may be displayed on a computer display.
- Displaying the manipulated digital image on the reference item fitted to the object may further comprise projecting the manipulated digital image onto the reference item fitted to the object.
- Capturing the image may include capturing information about the reference item.
- Manipulating the digital image then includes manipulating the digital image in dependence on the captured information about the reference item.
- the captured information may comprise the texture of the item of clothing.
- Capturing the image may include capturing information about the fitting of the reference item to the object.
- Manipulating the digital image may then include manipulating the digital image in dependence on the captured information about the fitting of the reference item to the object.
- the information may include lighting information.
- the information may include information of the lighting of the reference item fitted to the object. Where the reference item is an item of clothing the information may comprise lighting information and/or illumination information of the clothing on the object.
- Capturing the image may include capturing information about the context of the reference item.
- Manipulating the digital image then includes manipulating the digital image in dependence on the captured information about the context of the reference item to the object.
- the information includes luminance information.
- the manipulated image is lit appropriately.
- the method may comprise capturing a static image of the reference item fitted to the object.
- the object may be static.
- Displaying the manipulated digital image on the reference item fitted to the object may comprise displaying the manipulated data image on a static captured image.
- the method may comprise capturing a moving image of the reference item fitted to the object.
- the object may be moving.
- the image may be adapted in real-time in accordance with the movement of the object.
- Displaying the manipulated digital image on the reference item fitted to the object may comprise dynamically displaying a dynamically manipulated data image on a moving captured image.
- the image may be captured and adapted in real-time, or it may be pre-rendered.
- the method may comprise storing a captured image.
- Displaying the manipulated digital image on the reference item fitted to the object may comprise retrieving a stored captured image and displaying the manipulated captured image on the retrieved stored captured image.
- the object may be a person.
- the reference item may be an item of clothing.
- the digital image may be a digital image of an item of clothing.
- the reference item may be a clothing item of a type corresponding to the type of clothing item of the digital image. For example, a clothing item image comprising three materials may be mapped to a reference item which is made for three materials.
- the reference item may be made from two distinct materials such that the two different materials are distinct on a captured image of the reference item.
- the reference item may be made from materials of distinct colours with the same luminance.
- the captured image may be re-touched before combining with the displayed image.
- the reference item may be made from materials having distinct ultra-violet properties. The captured image may not need to be re-touched.
- a system is configured to provide the method features.
- the system may be a computer system.
- the computer program code may be provided for performing any described method.
- a computer program product for storing computer program code for performing any described method.
- the computer product may be a disk or memory device, or a hard drive of a computer system.
- the method may be implemented by computer program code of more than one computing device operating together.
- the method may be a computer-implemented method.
- FIG. 1 illustrates a reference matrix for a digital image
- FIG. 2 illustrates a reference item overlaid on the digital image template of FIG. 1 ;
- FIG. 3 illustrates a manufactured reference item based on the reference item digital image template of FIG. 2 ;
- FIG. 4 illustrates the reference item as fitted to a person
- FIG. 5 illustrates a location mapping of a digital photographic image
- FIG. 6 illustrates an image of the fitted reference item of FIG. 4 applied to a digital image template such as that of FIG. 5 ;
- FIG. 7 illustrates a catalogue item presented on a digital image template
- FIG. 8 illustrates the display of the catalogue item on the reference image of FIG. 6 ;
- FIG. 9 illustrates the manufactured catalogue item of FIG. 7 ;
- FIG. 10 illustrates an exemplary process flow
- FIG. 11 illustrates an exemplary computer system architecture.
- the invention is now described by way of example with reference in particular to the application of a technique for displaying a digital image on a clothing item fitted to a person.
- the invention is more broadly applicable, as will be apparent to one skilled in the art.
- FIG. 1 illustrates a digital image template 10 .
- the digital image template 10 is shown as having an array of pixels, p 1,1 to p n,m .
- the size of the array of pixels is implementation dependent.
- FIG. 2 illustrates a reference item applied to the digital image template of FIG. 2 .
- the reference item is an item of clothing.
- a pattern for the item of clothing is applied to the digital image template.
- Reference numeral 12 denotes the pattern for a front portion of the item of clothing, and reference numerals 14 and 16 denote the left and right sleeves of the reference item of clothing.
- each portion of the pattern is associated with a plurality of pixels.
- the number of pixels which coincide with the portions of the pattern will depend upon the density of the pixel array.
- pixels R 1 , R 2 , R 3 and R 4 are coincident with the front of the pattern of the clothing item 12 , being denoted as pixels R 1 , R 2 , R 3 and R 4 .
- Two pixels are coincident with the sleeve pattern 14 , being denoted as pixels R 5 and R 6 .
- Two pixels are coincident with a sleeve pattern 16 being denoted R 7 and R 8 .
- the reference points R 1 to R 8 may be specifically identified on the pattern.
- FIG. 3 illustrates an item of clothing made according to the pattern of FIG. 2 , and it can be seen that the reference points R 1 to R 8 are located on particular points of the item of clothing.
- the reference points may be illustrated on the item of clothing as shown in FIG. 3 by the use of different coloured materials to produce the item, so that the reference marks physically stand out.
- FIG. 4 there is shown the manufactured reference item of FIG. 3 fitted to a person, the person being an example of an object.
- the reference points R 1 to R 8 are visible on the reference item of clothing.
- FIG. 5 there is shown a photographic image template denoted by reference numeral 20 .
- An array of pixels is associated with the photographic image template, denoted by l 1,1 to l n,m .
- the calibration reference for the photographic image template may simply be one corner of the image, such as the top left corner being the pixel l 1,1 and the bottom right corner being the pixel l n,m .
- the pixel array as shown in FIG. 5 may therefore simply be obtained by providing a set of coordinates to the space of the photographic image.
- FIG. 6 thus shows the fitted reference item digitally captured to a photographic template.
- each of the reference points R 1 to R 8 may be identified as being located at a physical location of the image being one of the pixels l 1,1 to l n,m .
- reference points may be mapped to locations in space of the photographic digital image, for example as follows:
- Point R 1 maps to position l a,b
- Point R 2 maps to position l c,d
- Point R 3 maps to position l e,f
- Point R 4 maps to position l g,h
- Point R 5 maps to position l i,j
- Point R 6 maps to position l k,l
- Point R 7 maps to position l o,p
- Point R 8 maps to position l q,r
- a user may access a catalogue showing various clothing items, which may have different patterns applied to them. A user may then select a pattern from the catalogue.
- FIG. 7 illustrates a digital image template for a particular catalogue item, comprising a front portion 22 , and arm portions 24 a, 24 b. As shown in FIG. 7 , a pattern comprising a circle 26 is applied to the front portion 22 .
- a user accesses a catalogue item, they may simply be shown an image representing a pattern, such as an image representing the front portion 22 of the catalogue item with the pattern 26 applied.
- the user then may select the catalogue item such as shown in FIG. 7 , and the catalogue item is then displayed to the user by applying the digital image template of the catalogue item as shown in FIG. 7 to the digitally captured image of the fitted reference item as shown in FIG. 4 .
- the catalogue item is shown applied to the captured image of a person, but with the pattern 26 suitably adjusted to take into account the wearing of the item by the person, taking into account the location of the reference points of the reference object relevant to the location points of the digital image.
- the catalogue item is shown on a digital image template and therefore has all the reference points which are associated with the reference item.
- the digital image template of the catalogue item has pixel positions C 1 to C 8 corresponding to the pixel locations in the reference digital template of points R 1 to R 8 .
- the points C 1 to C 8 on the catalogue item associated with the reference points R 1 to R 8 can be displayed at the appropriate location points, to take into account the adjustment of the item when it is actually applied to a clothing item worn on a person.
- the digital image of the catalogue item is thus manipulated.
- the points of the catalogue item may be mapped to the captured reference item as follows:
- Point C 1 maps to point R 1 which maps to position l a,b
- Point C 2 maps to point R 2 which maps to position l c,d
- Point C 3 maps to point R 3 which maps to position l e,f
- Point C 4 maps to point R 4 which maps to position l g,h
- Point C 5 maps to point R 5 which maps to position l i,j
- Point C 6 maps to point R 6 which maps to position l k,l
- Point C 7 maps to point R 7 which maps to position l o,p
- Point C 8 maps to point R 8 which maps to position l q,r
- FIG. 9 there is illustrated the manufacture of the catalogue item of FIG. 7 , and it can be seen that when a clothing item as illustrated in FIG. 7 is worn on a person the “perfect” representation as shown in FIG. 9 actually results in a representation which is manipulated to take into account the wearing of the garment by the person as shown in FIG. 8 .
- the pixel points associated with the catalogue item applied to the digital image template as shown in FIG. 7 are mapped to the corresponding reference points when the item is worn as shown in FIG. 4 , to allow a realistic representation of the digital image to be displayed as shown in FIG. 8 .
- a scaling may also be required to scale the manipulated image to the captured image.
- the scaling may be inherent to the mapping process.
- additional information may also be captured.
- Information about the clothing item may be captured, such as the texture of the item of clothing.
- Other information may be captured, such as lighting information.
- the lighting information may be general information about the lighting of the scenario under which the image is captured, as well as lighting information relating to shadowing associated with the wearing of the clothing item by the person.
- this additional information may be utilised.
- the image may be manipulated not only based on the mapping of the reference pixels, but also based on the captured information about the texture of the clothing garment and/or the captured information about the lighting of the scenario.
- the captured image as illustrated in FIG. 6 may capture an image of the reference clothing item fitted to a model, and when a user accesses the system later in order to view an illustration of a particular pattern the pattern may be shown as displayed in combination with the image of the model.
- a user may also possible for a user to provide their own image, being an image taken with them fitted with the reference item, such that catalogue items may be displayed in combination with their own image.
- Such an image may be provided by a user directly, or may be taken by a user on visiting a shop for example.
- the manipulated image is applied to a captured image of an object, such that the manipulated image is applied to a representation of the object which is not the actual object.
- the manipulated image can be considered as merged with the captured image of the object.
- the display of the image as shown in FIG. 8 may be a display of the manipulated image applied to a representation of the object (a captured image) on a computer screen.
- the manipulated image of the catalogue item may be projected onto a person wearing the reference item, and then the displayed catalogue item viewed in a mirror.
- the representation of the object is the actual object (i.e. the person).
- Such an example still needs a reference image to be captured in order to correctly manipulate the catalogue image but the manipulated image is then projected onto the object rather than merged with an image of the object.
- the image may be captured using a camera, and be a static image
- the image may also be captured using a video camera and be a moving image.
- an input being a moving image may be dynamically utilised in order to generate a moving image of the pattern applied to the reference object, so that the image as shown in FIG. 8 is actually a moving image.
- the moving image may be projected onto a person as they move, the capturing of the image comprising capturing video movement, and then manipulating the image in real-time for display onto the person.
- Displaying a manipulated digital image on the reference item fitted to the object may comprise combining the manipulated digital image with the captured image of the reference item fitted to the object.
- the manipulated digital image and the captured image may then be combined by digitally merging them.
- a combined image may be displayed in a virtual reality mirror.
- a person may be positioned in front of a virtual reality mirror, and then a static or moving image of that person fitted with the reference item may be captured.
- a selected catalogue item is then chosen, and mapped accordingly using the above techniques, and displayed on the virtual reality mirror, so the person may view the mirror to see an image as if they were wearing the catalogue item.
- the combined image is displayed on a computer display.
- a person may select multiple catalogue images, or more generally non-reference images, for viewing on the captured image (moving or still).
- Displaying a manipulated digital image on the reference item fitted to the object may comprise projecting the manipulated digital image onto the reference item fitted to the object.
- the manipulated image may be projected onto it directly, and the person can view the projected image in a mirror. The person can thus view the item as if they were wearing an item made based on the image template, and viewing it in the mirror.
- a captured image may be stored.
- Displaying the manipulated digital image on the reference item fitted to the object may comprise retrieving a stored captured image and displaying the manipulated captured image on the retrieved stored captured image.
- the reference item may be an item of clothing.
- the digital image may be a digital image of an item of clothing.
- the reference item may be a clothing item of a type corresponding to the type of clothing item of the digital image.
- a three-part clothing item image may be mapped to a reference item which is made for three parts.
- the number of parts refers to the number of different materials which may make up the item, for example for a knitted item of clothing, three knitted threads.
- the reference item may be made from two distinct materials such that the two different materials are visibly distinct on a captured image of the reference item.
- the reference item may be made from materials of distinct colours.
- the reference item may be made from materials having distinct ultra-violet properties, not being distinct to the naked eye.
- the captured image may or may not be re-touched before coining with the displayed image.
- a step 100 there is created and stored digital image templates for catalogue items.
- the image templates are for non-reference items.
- a step 102 there is created and stored digital image templates for reference items.
- a reference item is manufactured according to the digital template for the reference item.
- a step 106 the manufactured reference item is fitted to an object.
- a digital image of the reference item fitted to the object is captured and stored.
- a digital image template of a catalogue item is retrieved.
- a step 114 reference points on the catalogue item are mapped to locations corresponding to locations of reference points on the fitted reference item (from the captured image).
- a step 116 the digital template of the catalogue image is used to manipulate the catalogue image and apply the captured image of the reference item fitted to the object.
- step 118 the mapping is scaled, if necessary. Scaling may be accounted for in the mapping. Any scaling may be applied before step 114 .
- step 120 the manipulated catalogue image, if necessary appropriately scaled, is displayed on a representation of the object.
- the manipulated images merge with the captured image of the object. If the representation of the object is the object itself, then in step 124 the manipulated image is projected onto the object.
- the system 160 includes a processor 130 and associated memory 132 , an image capture block 142 , a mapping block 144 , a storage block 148 for storing digital image templates of catalogue items, a storage block 150 for storing digital image templates of reference items, and an image generation block 146 . All of the elements of the system are interconnected by a communication bus 154 .
- system 160 is connected to a camera 134 and a video camera 136 , the camera 134 and the video camera 136 being connected to the image capture block 142 .
- system 160 is connected to a display 138 , for example a display associated with a computer, and/or a projector 140 .
- the display 138 and the projector 140 are connected to the image generation block 146 .
- system 160 is connected to a manufacturing block 152 .
- the manufacturing block 152 may be connected to the storage digital image template reference items of the storage block 150 .
- the reference image and its associated digital image are the diagrammatic colour plans of the garment, where the image is clothing item.
- the ultra-violet sensitive material does not influence luminance in a way that it cannot be filtered out.
- Computer vision may be used to automate the process.
- Pattern recognition may be utilised to automatically identify point correspondences.
- pattern recognition and knowledge of the hue or material properties of the reference garment may be used to automatically mask out the non-garment areas.
- the reference pattern that is used to obtain the point correspondences may be embedded into the garment itself.
- the reference garment may be made from the pattern including the embedded reference pattern.
- mappings may be utilised to map different sizes of garments to the same garment.
- the mapping may be crated for one item, and then further mappings crated for different garment sizes. This may be a derivative of scaling.
- the invention is described with an example implementation where the object is a person.
- the object may be any entity.
- the invention is also described in the context of an item of clothing being applied to the object, but in general any item may be applied to an object.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Mathematical Optimization (AREA)
- Mathematical Analysis (AREA)
- Pure & Applied Mathematics (AREA)
- Algebra (AREA)
- Computer Graphics (AREA)
- Computer Hardware Design (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Processing Or Creating Images (AREA)
- Image Analysis (AREA)
Abstract
Description
- The invention relates to the display of a digital image on a representation of an object, and particularly but not exclusively to the display of a digital image representing a design of an item of clothing onto a person.
- It is known to provide for the display of digital images onto a representation of an object. For example a digital image of an item of clothing may be displayed onto an image of a person.
- A problem associated with displaying images onto representations of objects, such as displaying a digital image of a clothing item onto an image of a person, is to facilitate the display in such a way that the resulting displayed image looks realistic. For example a perfect representation of an image, when displayed on a representation of an object, may need not to be perfect, to reflect the fact that the object “carries” the image in a particular way. For example, for a clothing item, the displayed image needs to reflect the fact that the person is wearing an item and the item is adjusted according to how they are wearing it.
- Therefore it is an aim of the invention to provide an improved technique for displaying a digital image on an object.
- A method of displaying a digital image on a representation of an object, comprising: defining reference point positions on a digital image template of a reference item; fitting the reference item to an object; capturing an image of the reference item fitted to the object; mapping positions on a digital image template of a non-reference item to the location positions of the reference points of the captured image to generate a manipulated non-reference digital image; and displaying the manipulated non-reference digital image on the reference item fitted to the object.
- The representation of the object may be the captured image of the object, and the manipulated non-reference digital image is displayed on the captured image.
- The representation of the object may be the object, and the manipulated non-reference digital image is projected on to the object.
- The mapping step may comprise detecting the reference point positions on the captured image and mapping the positions to location points, wherein corresponding points on the digital image template of a non-reference item are mapped to the associated location points to manipulate the non-reference image for display on the representation of the object.
- The non-reference digital image may be a catalogue digital image.
- Displaying the manipulated digital image on the reference item fitted to the representation of the object may further comprise combining the manipulated digital image with the captured image of the reference item fitted to the object. The manipulated digital image and the captured image may be combined by digitally merging them. When the manipulated digital image and the captured image are combined by digitally merging them, the combined image may be displayed in a virtual reality mirror. When the manipulated digital image and the captured image are combined by digitally merging them, the combined image may be displayed on a computer display.
- Displaying the manipulated digital image on the reference item fitted to the object may further comprise projecting the manipulated digital image onto the reference item fitted to the object.
- Capturing the image may include capturing information about the reference item. Manipulating the digital image then includes manipulating the digital image in dependence on the captured information about the reference item. Where the reference item is an item of clothing, the captured information may comprise the texture of the item of clothing.
- Capturing the image may include capturing information about the fitting of the reference item to the object. Manipulating the digital image may then include manipulating the digital image in dependence on the captured information about the fitting of the reference item to the object. The information may include lighting information. The information may include information of the lighting of the reference item fitted to the object. Where the reference item is an item of clothing the information may comprise lighting information and/or illumination information of the clothing on the object.
- Capturing the image may include capturing information about the context of the reference item. Manipulating the digital image then includes manipulating the digital image in dependence on the captured information about the context of the reference item to the object. The information includes luminance information. When the representation of the digital image is displayed the manipulated image is lit appropriately.
- The method may comprise capturing a static image of the reference item fitted to the object. The object may be static. Displaying the manipulated digital image on the reference item fitted to the object may comprise displaying the manipulated data image on a static captured image.
- The method may comprise capturing a moving image of the reference item fitted to the object. The object may be moving. The image may be adapted in real-time in accordance with the movement of the object. Displaying the manipulated digital image on the reference item fitted to the object may comprise dynamically displaying a dynamically manipulated data image on a moving captured image.
- The image may be captured and adapted in real-time, or it may be pre-rendered.
- The method may comprise storing a captured image. Displaying the manipulated digital image on the reference item fitted to the object may comprise retrieving a stored captured image and displaying the manipulated captured image on the retrieved stored captured image.
- The object may be a person.
- The reference item may be an item of clothing. The digital image may be a digital image of an item of clothing. The reference item may be a clothing item of a type corresponding to the type of clothing item of the digital image. For example, a clothing item image comprising three materials may be mapped to a reference item which is made for three materials.
- The reference item may be made from two distinct materials such that the two different materials are distinct on a captured image of the reference item. The reference item may be made from materials of distinct colours with the same luminance. The captured image may be re-touched before combining with the displayed image. The reference item may be made from materials having distinct ultra-violet properties. The captured image may not need to be re-touched.
- A system is configured to provide the method features. The system may be a computer system.
- There may be provided computer program code for performing any described method. There may be provided a computer program product for storing computer program code for performing any described method. The computer product may be a disk or memory device, or a hard drive of a computer system. The method may be implemented by computer program code of more than one computing device operating together.
- The method may be a computer-implemented method.
-
FIG. 1 illustrates a reference matrix for a digital image; -
FIG. 2 illustrates a reference item overlaid on the digital image template ofFIG. 1 ; -
FIG. 3 illustrates a manufactured reference item based on the reference item digital image template ofFIG. 2 ; -
FIG. 4 illustrates the reference item as fitted to a person; -
FIG. 5 illustrates a location mapping of a digital photographic image; -
FIG. 6 illustrates an image of the fitted reference item ofFIG. 4 applied to a digital image template such as that ofFIG. 5 ; -
FIG. 7 illustrates a catalogue item presented on a digital image template; -
FIG. 8 illustrates the display of the catalogue item on the reference image ofFIG. 6 ; -
FIG. 9 illustrates the manufactured catalogue item ofFIG. 7 ; -
FIG. 10 illustrates an exemplary process flow; and -
FIG. 11 illustrates an exemplary computer system architecture. - The invention is now described by way of example with reference in particular to the application of a technique for displaying a digital image on a clothing item fitted to a person. The invention is more broadly applicable, as will be apparent to one skilled in the art.
-
FIG. 1 illustrates adigital image template 10. Thedigital image template 10 is shown as having an array of pixels, p1,1 to pn,m. The size of the array of pixels is implementation dependent. -
FIG. 2 illustrates a reference item applied to the digital image template ofFIG. 2 . In the example described herein, the reference item is an item of clothing. A pattern for the item of clothing is applied to the digital image template.Reference numeral 12 denotes the pattern for a front portion of the item of clothing, andreference numerals - As will be understood with reference to
FIG. 2 , when the pattern is applied to the digital image template pixels of the array of the template coincide with the patterns. Thus each portion of the pattern is associated with a plurality of pixels. The number of pixels which coincide with the portions of the pattern will depend upon the density of the pixel array. - For the purposes of example, it can be seen that four pixels are coincident with the front of the pattern of the
clothing item 12, being denoted as pixels R1, R2, R3 and R4. Two pixels are coincident with thesleeve pattern 14, being denoted as pixels R5 and R6. Two pixels are coincident with asleeve pattern 16 being denoted R7 and R8. - In the pattern for the reference item, the reference points R1 to R8 may be specifically identified on the pattern.
FIG. 3 illustrates an item of clothing made according to the pattern ofFIG. 2 , and it can be seen that the reference points R1 to R8 are located on particular points of the item of clothing. Where the pattern is formed of two colours, for example, then the reference points may be illustrated on the item of clothing as shown inFIG. 3 by the use of different coloured materials to produce the item, so that the reference marks physically stand out. - With reference to
FIG. 4 there is shown the manufactured reference item ofFIG. 3 fitted to a person, the person being an example of an object. As can be seen inFIG. 4 , the reference points R1 to R8 are visible on the reference item of clothing. - With reference to
FIG. 5 there is shown a photographic image template denoted by reference numeral 20. An array of pixels is associated with the photographic image template, denoted by l1,1 to ln,m. The calibration reference for the photographic image template may simply be one corner of the image, such as the top left corner being the pixel l1,1 and the bottom right corner being the pixel ln,m. The pixel array as shown inFIG. 5 may therefore simply be obtained by providing a set of coordinates to the space of the photographic image. - A photographic image is taken of the manufactured reference item fitted to the object (e.g. person) as shown in
FIG. 4 , and the resulting photographic image is shown inFIG. 6 .FIG. 6 thus shows the fitted reference item digitally captured to a photographic template. - As can be understood with reference to
FIG. 5 , the resulting image shown inFIG. 6 can be mapped to coordinates of a physical location system defined by pixel elements l1,1 to ln,m. Thus each of the reference points R1 to R8 may be identified as being located at a physical location of the image being one of the pixels l1,1 to ln,m. - Thus the reference points may be mapped to locations in space of the photographic digital image, for example as follows:
- Point R1 maps to position la,b
- Point R2 maps to position lc,d
- Point R3 maps to position le,f
- Point R4 maps to position lg,h
- Point R5 maps to position li,j
- Point R6 maps to position lk,l
- Point R7 maps to position lo,p
- Point R8 maps to position lq,r
- A user may access a catalogue showing various clothing items, which may have different patterns applied to them. A user may then select a pattern from the catalogue.
- For example
FIG. 7 illustrates a digital image template for a particular catalogue item, comprising afront portion 22, andarm portions FIG. 7 , a pattern comprising acircle 26 is applied to thefront portion 22. - In practice, as a user accesses a catalogue item, they may simply be shown an image representing a pattern, such as an image representing the
front portion 22 of the catalogue item with thepattern 26 applied. - The user then may select the catalogue item such as shown in
FIG. 7 , and the catalogue item is then displayed to the user by applying the digital image template of the catalogue item as shown inFIG. 7 to the digitally captured image of the fitted reference item as shown inFIG. 4 . - Thus as shown in
FIG. 8 , the catalogue item is shown applied to the captured image of a person, but with thepattern 26 suitably adjusted to take into account the wearing of the item by the person, taking into account the location of the reference points of the reference object relevant to the location points of the digital image. - As shown in
FIG. 7 , the catalogue item is shown on a digital image template and therefore has all the reference points which are associated with the reference item. As shown inFIG. 7 , the digital image template of the catalogue item has pixel positions C1 to C8 corresponding to the pixel locations in the reference digital template of points R1 to R8. - Based on the correlation between the reference points R1 to R7 in the location system comprising the location array l1,1 to ln,m, then the points C1 to C8 on the catalogue item associated with the reference points R1 to R8 can be displayed at the appropriate location points, to take into account the adjustment of the item when it is actually applied to a clothing item worn on a person. The digital image of the catalogue item is thus manipulated.
- The points of the catalogue item may be mapped to the captured reference item as follows:
- Point C1 maps to point R1 which maps to position la,b
- Point C2 maps to point R2 which maps to position lc,d
- Point C3 maps to point R3 which maps to position le,f
- Point C4 maps to point R4 which maps to position lg,h
- Point C5 maps to point R5 which maps to position li,j
- Point C6 maps to point R6 which maps to position lk,l
- Point C7 maps to point R7 which maps to position lo,p
- Point C8 maps to point R8 which maps to position lq,r
- With reference to
FIG. 9 there is illustrated the manufacture of the catalogue item ofFIG. 7 , and it can be seen that when a clothing item as illustrated inFIG. 7 is worn on a person the “perfect” representation as shown inFIG. 9 actually results in a representation which is manipulated to take into account the wearing of the garment by the person as shown inFIG. 8 . - Thus the pixel points associated with the catalogue item applied to the digital image template as shown in
FIG. 7 are mapped to the corresponding reference points when the item is worn as shown inFIG. 4 , to allow a realistic representation of the digital image to be displayed as shown inFIG. 8 . - A scaling may also be required to scale the manipulated image to the captured image. Alternatively the scaling may be inherent to the mapping process.
- When the digital image of the person wearing the clothing item is captured, as shown in
FIG. 4 , additional information may also be captured. Information about the clothing item may be captured, such as the texture of the item of clothing. Other information may be captured, such as lighting information. The lighting information may be general information about the lighting of the scenario under which the image is captured, as well as lighting information relating to shadowing associated with the wearing of the clothing item by the person. - When the image as illustrated in
FIG. 8 is generated, this additional information may be utilised. Thus the image may be manipulated not only based on the mapping of the reference pixels, but also based on the captured information about the texture of the clothing garment and/or the captured information about the lighting of the scenario. - The captured image as illustrated in
FIG. 6 may capture an image of the reference clothing item fitted to a model, and when a user accesses the system later in order to view an illustration of a particular pattern the pattern may be shown as displayed in combination with the image of the model. However it may be also possible for a user to provide their own image, being an image taken with them fitted with the reference item, such that catalogue items may be displayed in combination with their own image. Such an image may be provided by a user directly, or may be taken by a user on visiting a shop for example. - In the example described, the manipulated image is applied to a captured image of an object, such that the manipulated image is applied to a representation of the object which is not the actual object. In the described example the manipulated image can be considered as merged with the captured image of the object.
- The display of the image as shown in
FIG. 8 may be a display of the manipulated image applied to a representation of the object (a captured image) on a computer screen. - However the display may take place in other ways.
- For example the manipulated image of the catalogue item may be projected onto a person wearing the reference item, and then the displayed catalogue item viewed in a mirror. In this example the representation of the object is the actual object (i.e. the person). Such an example still needs a reference image to be captured in order to correctly manipulate the catalogue image but the manipulated image is then projected onto the object rather than merged with an image of the object.
- Whilst the image may be captured using a camera, and be a static image, the image may also be captured using a video camera and be a moving image. In this way an input being a moving image may be dynamically utilised in order to generate a moving image of the pattern applied to the reference object, so that the image as shown in
FIG. 8 is actually a moving image. - Where the displayed image is projected onto a person, then the moving image may be projected onto a person as they move, the capturing of the image comprising capturing video movement, and then manipulating the image in real-time for display onto the person.
- Displaying a manipulated digital image on the reference item fitted to the object may comprise combining the manipulated digital image with the captured image of the reference item fitted to the object. The manipulated digital image and the captured image may then be combined by digitally merging them.
- When the manipulated digital image and the captured image are combined by digitally merging them, a combined image may be displayed in a virtual reality mirror. For example, a person may be positioned in front of a virtual reality mirror, and then a static or moving image of that person fitted with the reference item may be captured. A selected catalogue item is then chosen, and mapped accordingly using the above techniques, and displayed on the virtual reality mirror, so the person may view the mirror to see an image as if they were wearing the catalogue item.
- When the manipulated digital image and the captured image are combined by digitally merging them, the combined image is displayed on a computer display. Thus a person may select multiple catalogue images, or more generally non-reference images, for viewing on the captured image (moving or still).
- Displaying a manipulated digital image on the reference item fitted to the object may comprise projecting the manipulated digital image onto the reference item fitted to the object. Thus where a person is wearing the reference item, the manipulated image may be projected onto it directly, and the person can view the projected image in a mirror. The person can thus view the item as if they were wearing an item made based on the image template, and viewing it in the mirror.
- A captured image may be stored. Displaying the manipulated digital image on the reference item fitted to the object may comprise retrieving a stored captured image and displaying the manipulated captured image on the retrieved stored captured image.
- The reference item may be an item of clothing. The digital image may be a digital image of an item of clothing. The reference item may be a clothing item of a type corresponding to the type of clothing item of the digital image. For example, a three-part clothing item image may be mapped to a reference item which is made for three parts. In this respect, the number of parts refers to the number of different materials which may make up the item, for example for a knitted item of clothing, three knitted threads.
- The reference item may be made from two distinct materials such that the two different materials are visibly distinct on a captured image of the reference item. The reference item may be made from materials of distinct colours. The reference item may be made from materials having distinct ultra-violet properties, not being distinct to the naked eye.
- The captured image may or may not be re-touched before coining with the displayed image.
- With reference to
FIG. 10 , there is now further described example implementations in accordance with the invention. In astep 100 there is created and stored digital image templates for catalogue items. In general, the image templates are for non-reference items. - In a
step 102 there is created and stored digital image templates for reference items. - In a step 104 a reference item is manufactured according to the digital template for the reference item.
- In a
step 106 the manufactured reference item is fitted to an object. - In a step 108 a digital image of the reference item fitted to the object is captured and stored.
- In a
step 110 positions of the captured reference marks on the reference item—fitted to the object—are mapped to locations. - In a step 112 a digital image template of a catalogue item is retrieved.
- In a
step 114 reference points on the catalogue item are mapped to locations corresponding to locations of reference points on the fitted reference item (from the captured image). - In a
step 116 the digital template of the catalogue image is used to manipulate the catalogue image and apply the captured image of the reference item fitted to the object. - In
step 118 the mapping is scaled, if necessary. Scaling may be accounted for in the mapping. Any scaling may be applied beforestep 114. - In
step 120, the manipulated catalogue image, if necessary appropriately scaled, is displayed on a representation of the object. - If the representation of the object is not the object itself, for example being an image of the object, then in
step 122 the manipulated images merge with the captured image of the object. If the representation of the object is the object itself, then instep 124 the manipulated image is projected onto the object. - With reference to
FIG. 11 , there is illustrated an example architecture of a system for implementing the described techniques. Thesystem 160 includes aprocessor 130 and associatedmemory 132, animage capture block 142, amapping block 144, astorage block 148 for storing digital image templates of catalogue items, astorage block 150 for storing digital image templates of reference items, and animage generation block 146. All of the elements of the system are interconnected by a communication bus 154. - In addition the
system 160 is connected to acamera 134 and avideo camera 136, thecamera 134 and thevideo camera 136 being connected to theimage capture block 142. - In addition the
system 160 is connected to adisplay 138, for example a display associated with a computer, and/or aprojector 140. Thedisplay 138 and theprojector 140 are connected to theimage generation block 146. - In addition the
system 160 is connected to a manufacturing block 152. The manufacturing block 152 may be connected to the storage digital image template reference items of thestorage block 150. - The reference image and its associated digital image are the diagrammatic colour plans of the garment, where the image is clothing item.
- To the extent that the reference item is made from ultra-violet sensitive material, the ultra-violet sensitive material does not influence luminance in a way that it cannot be filtered out.
- Computer vision may be used to automate the process.
- Pattern recognition may be utilised to automatically identify point correspondences.
- Where the item is a clothing item, pattern recognition and knowledge of the hue or material properties of the reference garment may be used to automatically mask out the non-garment areas.
- Processed forms of the correspondences may be utilised so that remapping, where done, may be done efficiently on external devices
- Preferably the reference pattern that is used to obtain the point correspondences may be embedded into the garment itself. The reference garment may be made from the pattern including the embedded reference pattern.
- An additional mapping may be utilised to map different sizes of garments to the same garment. Thus, where the item is a clothing item, the mapping may be crated for one item, and then further mappings crated for different garment sizes. This may be a derivative of scaling.
- The invention has been described by way of reference to particular example scenarios, and the invention is not limited to these examples.
- For example the invention is described with an example implementation where the object is a person. However the object may be any entity. The invention is also described in the context of an item of clothing being applied to the object, but in general any item may be applied to an object.
- The invention is not limited to the details of any example or embodiment set out hereinabove, nor is the invention limited to the combination of any features of any example or embodiment as set out hereinabove.
Claims (37)
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB1420090 | 2014-11-12 | ||
GB1420090.1 | 2014-11-12 | ||
GBGB1420090.1A GB201420090D0 (en) | 2014-11-12 | 2014-11-12 | Image to item mapping |
PCT/EP2015/076485 WO2016075263A1 (en) | 2014-11-12 | 2015-11-12 | Image to item mapping |
Publications (2)
Publication Number | Publication Date |
---|---|
US20170309075A1 true US20170309075A1 (en) | 2017-10-26 |
US11030807B2 US11030807B2 (en) | 2021-06-08 |
Family
ID=52118354
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/526,629 Active US11030807B2 (en) | 2014-11-12 | 2015-11-12 | Image to item mapping |
Country Status (5)
Country | Link |
---|---|
US (1) | US11030807B2 (en) |
EP (1) | EP3218878B1 (en) |
JP (1) | JP2018500647A (en) |
GB (1) | GB201420090D0 (en) |
WO (1) | WO2016075263A1 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180096506A1 (en) * | 2016-10-04 | 2018-04-05 | Facebook, Inc. | Controls and Interfaces for User Interactions in Virtual Spaces |
US11080912B2 (en) * | 2015-08-10 | 2021-08-03 | Zazzle Inc. | System and method for digital markups of custom products |
US11478033B2 (en) | 2016-11-06 | 2022-10-25 | Global Apparel Partners Inc. | Knitted textile methods |
US11530503B2 (en) * | 2019-07-23 | 2022-12-20 | Levi Strauss & Co. | Three-dimensional rendering preview in web-based tool for design of laser-finished garments |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2559567B (en) | 2017-02-08 | 2022-06-22 | Unmade Ltd | A method of knitting a fabric using a knitting machine and a knitting machine |
US11250572B2 (en) * | 2019-10-21 | 2022-02-15 | Salesforce.Com, Inc. | Systems and methods of generating photorealistic garment transference in images |
TWI770874B (en) * | 2021-03-15 | 2022-07-11 | 楷思諾科技服務有限公司 | Method for displaying simulation images through clicking and rolling operations |
US12100156B2 (en) | 2021-04-12 | 2024-09-24 | Snap Inc. | Garment segmentation |
US11670059B2 (en) | 2021-09-01 | 2023-06-06 | Snap Inc. | Controlling interactive fashion based on body gestures |
US11673054B2 (en) | 2021-09-07 | 2023-06-13 | Snap Inc. | Controlling AR games on fashion items |
US11900506B2 (en) * | 2021-09-09 | 2024-02-13 | Snap Inc. | Controlling interactive fashion based on facial expressions |
US11734866B2 (en) | 2021-09-13 | 2023-08-22 | Snap Inc. | Controlling interactive fashion based on voice |
US11983826B2 (en) | 2021-09-30 | 2024-05-14 | Snap Inc. | 3D upper garment tracking |
US11636662B2 (en) | 2021-09-30 | 2023-04-25 | Snap Inc. | Body normal network light and rendering control |
US11651572B2 (en) | 2021-10-11 | 2023-05-16 | Snap Inc. | Light and rendering of garments |
Citations (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5619799A (en) * | 1992-09-23 | 1997-04-15 | Tpc International | Three-dimensional pattern design method for garments fitted with sleeves |
US5680528A (en) * | 1994-05-24 | 1997-10-21 | Korszun; Henry A. | Digital dressing room |
US6310627B1 (en) * | 1998-01-20 | 2001-10-30 | Toyo Boseki Kabushiki Kaisha | Method and system for generating a stereoscopic image of a garment |
US20040153195A1 (en) * | 2003-01-14 | 2004-08-05 | Watanabe John S. | System and method for inspecting custom-made clothing |
US6907310B2 (en) * | 2001-01-19 | 2005-06-14 | Virtual Mirrors Limited | Production and visualization of garments |
US6968297B1 (en) * | 1999-10-08 | 2005-11-22 | Lectra Sa | Method and device for simulating and representing the dressing of a mannequin |
US7039486B2 (en) * | 2002-03-22 | 2006-05-02 | Kenneth Kuk-Kei Wang | Method and device for viewing, archiving and transmitting a garment model over a computer network |
US7657340B2 (en) * | 2006-01-31 | 2010-02-02 | Dragon & Phoenix Software, Inc. | System, apparatus and method for facilitating pattern-based clothing design activities |
US7663648B1 (en) * | 1999-11-12 | 2010-02-16 | My Virtual Model Inc. | System and method for displaying selected garments on a computer-simulated mannequin |
US7752078B2 (en) * | 2000-03-28 | 2010-07-06 | Sony Corporation | Communication service method and communication apparatus thereof |
US7945343B2 (en) * | 2006-12-18 | 2011-05-17 | Nike, Inc. | Method of making an article of footwear |
US7953648B2 (en) * | 2001-11-26 | 2011-05-31 | Vock Curtis A | System and methods for generating virtual clothing experiences |
US20110298897A1 (en) * | 2010-06-08 | 2011-12-08 | Iva Sareen | System and method for 3d virtual try-on of apparel on an avatar |
US8165711B2 (en) * | 2010-01-05 | 2012-04-24 | Microsoft Corporation | Automated generation of garment construction specification |
US8174521B2 (en) * | 2007-10-26 | 2012-05-08 | Zazzle.Com | Product modeling system and method |
US8275590B2 (en) * | 2009-08-12 | 2012-09-25 | Zugara, Inc. | Providing a simulation of wearing items such as garments and/or accessories |
US20130124156A1 (en) * | 2009-05-26 | 2013-05-16 | Embodee Corp | Footwear digitization system and method |
US20150055085A1 (en) * | 2013-08-22 | 2015-02-26 | Bespoke, Inc. | Method and system to create products |
US20150339853A1 (en) * | 2013-01-02 | 2015-11-26 | Embodee Corp. | Footwear digitization system and method |
US20160180449A1 (en) * | 2014-12-23 | 2016-06-23 | Mihir Naware | Systems and methods for generating virtual contexts from three dimensional models |
US9401023B2 (en) * | 2012-01-30 | 2016-07-26 | Rakuten, Inc. | Clothing image processing system, control method for clothing image processing system, clothing image processing device, control method for clothing image processing device, program, and information recording medium |
US9702071B2 (en) * | 2008-10-23 | 2017-07-11 | Zazzle Inc. | Embroidery system and method |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6307568B1 (en) * | 1998-10-28 | 2001-10-23 | Imaginarix Ltd. | Virtual dressing over the internet |
KR20070111417A (en) | 2007-10-04 | 2007-11-21 | 심형용 | Simulation system for clothes ,using projector |
CA2737067C (en) | 2008-07-29 | 2015-05-19 | Zazzle.Com, Inc. | Product customization system and method |
US10176636B1 (en) * | 2015-12-11 | 2019-01-08 | A9.Com, Inc. | Augmented reality fashion |
-
2014
- 2014-11-12 GB GBGB1420090.1A patent/GB201420090D0/en not_active Ceased
-
2015
- 2015-11-12 WO PCT/EP2015/076485 patent/WO2016075263A1/en active Application Filing
- 2015-11-12 JP JP2017525856A patent/JP2018500647A/en active Pending
- 2015-11-12 EP EP15797285.2A patent/EP3218878B1/en active Active
- 2015-11-12 US US15/526,629 patent/US11030807B2/en active Active
Patent Citations (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5619799A (en) * | 1992-09-23 | 1997-04-15 | Tpc International | Three-dimensional pattern design method for garments fitted with sleeves |
US5680528A (en) * | 1994-05-24 | 1997-10-21 | Korszun; Henry A. | Digital dressing room |
US6310627B1 (en) * | 1998-01-20 | 2001-10-30 | Toyo Boseki Kabushiki Kaisha | Method and system for generating a stereoscopic image of a garment |
US6968297B1 (en) * | 1999-10-08 | 2005-11-22 | Lectra Sa | Method and device for simulating and representing the dressing of a mannequin |
US7663648B1 (en) * | 1999-11-12 | 2010-02-16 | My Virtual Model Inc. | System and method for displaying selected garments on a computer-simulated mannequin |
US7752078B2 (en) * | 2000-03-28 | 2010-07-06 | Sony Corporation | Communication service method and communication apparatus thereof |
US6907310B2 (en) * | 2001-01-19 | 2005-06-14 | Virtual Mirrors Limited | Production and visualization of garments |
US7953648B2 (en) * | 2001-11-26 | 2011-05-31 | Vock Curtis A | System and methods for generating virtual clothing experiences |
US7039486B2 (en) * | 2002-03-22 | 2006-05-02 | Kenneth Kuk-Kei Wang | Method and device for viewing, archiving and transmitting a garment model over a computer network |
US20040153195A1 (en) * | 2003-01-14 | 2004-08-05 | Watanabe John S. | System and method for inspecting custom-made clothing |
US7657340B2 (en) * | 2006-01-31 | 2010-02-02 | Dragon & Phoenix Software, Inc. | System, apparatus and method for facilitating pattern-based clothing design activities |
US7945343B2 (en) * | 2006-12-18 | 2011-05-17 | Nike, Inc. | Method of making an article of footwear |
US8174521B2 (en) * | 2007-10-26 | 2012-05-08 | Zazzle.Com | Product modeling system and method |
US9702071B2 (en) * | 2008-10-23 | 2017-07-11 | Zazzle Inc. | Embroidery system and method |
US20130124156A1 (en) * | 2009-05-26 | 2013-05-16 | Embodee Corp | Footwear digitization system and method |
US8275590B2 (en) * | 2009-08-12 | 2012-09-25 | Zugara, Inc. | Providing a simulation of wearing items such as garments and/or accessories |
US8165711B2 (en) * | 2010-01-05 | 2012-04-24 | Microsoft Corporation | Automated generation of garment construction specification |
US20110298897A1 (en) * | 2010-06-08 | 2011-12-08 | Iva Sareen | System and method for 3d virtual try-on of apparel on an avatar |
US9401023B2 (en) * | 2012-01-30 | 2016-07-26 | Rakuten, Inc. | Clothing image processing system, control method for clothing image processing system, clothing image processing device, control method for clothing image processing device, program, and information recording medium |
US20150339853A1 (en) * | 2013-01-02 | 2015-11-26 | Embodee Corp. | Footwear digitization system and method |
US9639635B2 (en) * | 2013-01-02 | 2017-05-02 | Embodee Corp | Footwear digitization system and method |
US20150055085A1 (en) * | 2013-08-22 | 2015-02-26 | Bespoke, Inc. | Method and system to create products |
US20160180449A1 (en) * | 2014-12-23 | 2016-06-23 | Mihir Naware | Systems and methods for generating virtual contexts from three dimensional models |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11080912B2 (en) * | 2015-08-10 | 2021-08-03 | Zazzle Inc. | System and method for digital markups of custom products |
US11717042B2 (en) | 2015-08-10 | 2023-08-08 | Zazzle, Inc. | System and method for digital markups of custom products |
USRE50039E1 (en) | 2015-08-10 | 2024-07-16 | Zazzle, Inc. | System and method for digital markups of custom products |
US20180096506A1 (en) * | 2016-10-04 | 2018-04-05 | Facebook, Inc. | Controls and Interfaces for User Interactions in Virtual Spaces |
US11478033B2 (en) | 2016-11-06 | 2022-10-25 | Global Apparel Partners Inc. | Knitted textile methods |
US11678706B2 (en) | 2016-11-06 | 2023-06-20 | Global Apparel Partners Inc. | Knitted textile methods |
US11530503B2 (en) * | 2019-07-23 | 2022-12-20 | Levi Strauss & Co. | Three-dimensional rendering preview in web-based tool for design of laser-finished garments |
Also Published As
Publication number | Publication date |
---|---|
WO2016075263A1 (en) | 2016-05-19 |
JP2018500647A (en) | 2018-01-11 |
EP3218878A1 (en) | 2017-09-20 |
US11030807B2 (en) | 2021-06-08 |
EP3218878B1 (en) | 2019-08-21 |
GB201420090D0 (en) | 2014-12-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11030807B2 (en) | Image to item mapping | |
US11961200B2 (en) | Method and computer program product for producing 3 dimensional model data of a garment | |
US11819080B2 (en) | System and method for digital markups of custom products | |
US9552655B2 (en) | Image processing via color replacement | |
CN106548455B (en) | Apparatus and method for adjusting brightness of image | |
US10659750B2 (en) | Method and system for presenting at least part of an image of a real object in a view of a real environment, and method and system for selecting a subset of a plurality of images | |
US11188739B2 (en) | Processing uncertain content in a computer graphics system | |
US10229483B2 (en) | Image processing apparatus and image processing method for setting an illumination environment | |
JP7031697B2 (en) | Information processing device and recognition support method | |
JP2019510297A (en) | Virtual try-on to the user's true human body model | |
ES2743491T3 (en) | Digital overlay of an image with another image | |
JP2016532197A5 (en) | ||
KR102209745B1 (en) | An information display device of a mirror display for advertisement and shopping by recognizing the reflected images on the mirror and method thereof | |
US11900552B2 (en) | System and method for generating virtual pseudo 3D outputs from images | |
WO2018213702A1 (en) | Augmented reality system | |
CN108346178A (en) | Mixed reality object is presented | |
KR20160068186A (en) | Apparatus and method for providingaugmented reality contentents | |
ES2827177T3 (en) | Image processing device, image processing method and program | |
RU2735066C1 (en) | Method for displaying augmented reality wide-format object | |
JP7125847B2 (en) | 3D model display device, 3D model display method and 3D model display program | |
CN107883930B (en) | Pose calculation method and system of display screen | |
JP4696669B2 (en) | Image adjustment method and image adjustment apparatus | |
JP2005107644A (en) | Robot |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: UNMADE LIMITED, UNITED KINGDOM Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WATTS, HAL;ALUN-JONES, BENJAMIN;EMERY, KIRSTY;AND OTHERS;SIGNING DATES FROM 20180207 TO 20180208;REEL/FRAME:045382/0849 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |