US20220130091A1 - Systems and methods for simulating animation of an object - Google Patents
Systems and methods for simulating animation of an object Download PDFInfo
- Publication number
- US20220130091A1 US20220130091A1 US17/391,553 US202117391553A US2022130091A1 US 20220130091 A1 US20220130091 A1 US 20220130091A1 US 202117391553 A US202117391553 A US 202117391553A US 2022130091 A1 US2022130091 A1 US 2022130091A1
- Authority
- US
- United States
- Prior art keywords
- leds
- fully
- lighting apparatus
- microcontroller
- controlled
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/02—Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
- G11B27/031—Electronic editing of digitised analogue information signals, e.g. audio or video signals
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T13/00—Animation
- G06T13/20—3D [Three Dimensional] animation
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B15/00—Special procedures for taking photographs; Apparatus therefor
- G03B15/02—Illuminating scene
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B15/00—Special procedures for taking photographs; Apparatus therefor
- G03B15/02—Illuminating scene
- G03B15/03—Combinations of cameras with lighting apparatus; Flash units
- G03B15/04—Combinations of cameras with non-electronic flash apparatus; Non-electronic flash units
- G03B15/0442—Constructional details of the flash apparatus; Arrangement of lamps, reflectors, or the like
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B15/00—Special procedures for taking photographs; Apparatus therefor
- G03B15/02—Illuminating scene
- G03B15/06—Special arrangements of screening, diffusing, or reflecting devices, e.g. in studio
- G03B15/07—Arrangements of lamps in studios
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/10—Office automation; Time management
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0241—Advertisements
- G06Q30/0276—Advertisement creation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/06—Buying, selling or leasing transactions
- G06Q30/0601—Electronic shopping [e-shopping]
- G06Q30/0641—Shopping interfaces
- G06Q30/0643—Graphical representation of items or shoppers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/06—Buying, selling or leasing transactions
- G06Q30/08—Auctions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/04—Texture mapping
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/50—Lighting effects
- G06T15/506—Illumination models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/50—Lighting effects
- G06T15/60—Shadow generation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/97—Determining parameters from multiple pictures
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/50—Constructional details
- H04N23/55—Optical parts specially adapted for electronic image sensors; Mounting thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/56—Cameras or camera modules comprising electronic image sensors; Control thereof provided with illuminating means
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/74—Circuitry for compensating brightness variation in the scene by influencing the scene brightness using illuminating means
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/743—Bracketing, i.e. taking a series of images with varying exposure conditions
-
- H04N5/2256—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/08—Indexing scheme for image data processing or generation, in general involving all processing steps from image acquisition to 3D model generation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2215/00—Indexing scheme for image rendering
- G06T2215/16—Using real world measurements to influence rendering
-
- H—ELECTRICITY
- H05—ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
- H05B—ELECTRIC HEATING; ELECTRIC LIGHT SOURCES NOT OTHERWISE PROVIDED FOR; CIRCUIT ARRANGEMENTS FOR ELECTRIC LIGHT SOURCES, IN GENERAL
- H05B47/00—Circuit arrangements for operating light sources in general, i.e. where the type of light source is not relevant
- H05B47/10—Controlling the light source
- H05B47/165—Controlling the light source following a pre-assigned programmed sequence; Logic control [LC]
Definitions
- the present invention generally relates to systems and methods for creating an animated visual approximation of a surface of an object. More particularly, such systems and methods relate to creating an animated visual approximation of a surface of a generally flat surface including a disk such as a coin or a coin-like object and a generally planar object such as a collectible card.
- photographs provide static views, photos do not accurately show how light interacts with the surfaces of a collectible object.
- photographs have difficulty rendering complex surface details such as foiling, holograms, and other effects that change as an observer moves or as lighting changes on a collectible.
- photographs can often hide scratches and marks collectors need to see to accurately judge the quality and condition of a collectible.
- photographs do not accurately capture the luster of a coin (i.e., how changes in light reflect or play across the surfaces of a coin) and are prone to concealing details that might otherwise be visible if viewed in hand. Such details might include scratches or other signs of broken luster that could indicate cleaning or alterations.
- manufacturers of newer collectable objects frequently develop new and innovative products having novel features and effects to entice collecting consumers, but as physical card shops and shows become less common, consumers have limited opportunities to physically interact with the products in hand.
- FIG. 1 is a perspective of a light ring assembly according to disclosed embodiments
- FIG. 2 is a perspective of a light ring assembly according to disclosed embodiments
- FIG. 3 is a perspective of a light ring assembly in combination with a camera according to disclosed embodiments
- FIG. 4 is a perspective of a light ring assembly illustrating a direction of lighting according to disclosed embodiments
- FIG. 5 is a perspective of a lens attachment according to disclosed embodiments.
- FIG. 6 is a perspective of a lens attachment according to disclosed embodiments.
- FIG. 7 is a perspective of a light rig assembly in combination with a camera according to disclosed embodiments.
- FIG. 8 is a perspective of a light rig assembly in combination with a camera according to disclosed embodiments.
- FIG. 9 is a cross section of a frame of the light rig assembly of FIG. 8 taken in the plane of line 3 - 3 of FIG. 8 ;
- FIG. 10 is graph of brightness over time for various LEDs in a lighting assembly according to disclosed embodiments.
- FIG. 11 is graph of brightness over time for various LEDs in a lighting assembly according to disclosed embodiments.
- FIGS. 12 and 13 are a graphical user interface within video editing software used during a “Stage 1” video creation process according to disclosed embodiments;
- FIGS. 14 and 15 are a graphical user interface within video editing software showing collectible objects masked off with a black background according to disclosed embodiments;
- FIGS. 16 and 17 are a graphical user interface within video editing software showing collectible objects during animation according to disclosed embodiments
- FIGS. 18 and 19 are a graphical user interface within video editing software showing a final rendering of an animation according to disclosed embodiments.
- FIGS. 20 and 21 are flowcharts illustrating methods of preparing an animated approximation according to disclosed embodiments.
- Embodiments described herein relate to systems and methods for virtually animating an object such as a collectible item.
- the present disclosure provides solutions to the industry-wide problems of remote viewing of such object by rendering an accurate animation of how such objects look in hand via an interactive online experience of such animations where only offline experiences were once available.
- the systems and methods described herein can produce ultra-high-resolution animations depicting how luster moves over an object's surfaces while rotating, allowing possible imperfections to be revealed or that show off the best aspects of the surfaces (e.g., foil, refractor, hologram, rainbow, etc.).
- the systems and methods described herein can provide the rendered animations in industry standard, web-friendly, video file formats customers can use without additional cost or inconvenience.
- the video animations can be displayed using streaming platforms to cater to online video streamers.
- the systems and methods described herein provide animated images that can increase online buyer confidence and increase bids for objects in the marketplace.
- Such systems and methods can also provide object manufacturers high quality online presentations enabling them to better market their products and enables collectors to demonstrate the in-hand appearance of a collectible online.
- the lighting apparatus can include mounting connectors configured to couple a camera to the lighting apparatus at a plurality of positions, a ring or frame positioned above and defining an illumination region and comprising a plurality of light emitting diodes (LEDs), and a microcontroller configured to control the plurality of LEDs to rotate light around a first surface of the object placed within the illumination region.
- the plurality of LEDs can include pure white LEDs which produce more accurate and appealing final animation than color RGB LEDs that simulated white by combining multiple color outputs together.
- the microcontroller can control the plurality of LEDs by dimming and brightening each of the plurality of LEDs in a sequenced order around the ring or frame such that a first of the plurality of LEDs is being dimmed from fully on to fully off while a second of the plurality of LEDs is being brightened from fully off to fully on.
- the microcontroller can control the first of the plurality of LEDs and the second of the plurality of LEDs such that the first of the plurality LEDs reaches fully off simultaneous with the second of the plurality LEDs reaching fully on, and once the first of the plurality LEDs reaches fully off and the second of the plurality LEDs reaches fully on, the microcontroller can begin to dim the second of the plurality LEDs to fully off and begins to control a next one of the plurality of LEDs in the sequenced order to brighten the next one of the plurality of LEDs from fully off to fully on.
- FIGS. 1-4 show an example embodiment of the lighting apparatus as a lighting ring assembly 20 A.
- the lighting ring assembly 20 A can include a RGBW LED ring 22 A that is pivotably connected to a leg 24 by a hinge 26 .
- the RGBW LED ring 22 A can have a 6 inch diameter.
- the RGBW LED ring 22 A can include Adafruit® NeoPixel® 60 ring-5050 RGBW LED having integrated drivers and natural white, 4500K LEDs.
- Adafruit and NeoPixel are federally registered trademarks of Adafruit Industries of New York, N.Y. It is envisioned that other LED rings could be used that include similar or better LED density.
- the lighting ring assembly 20 A can include an enclosure 28 provided at an end of the leg 24 opposite the ring 22 A that houses a microcontroller 30 .
- the microcontroller 30 can include an electrician® 5V microcontroller.
- Arduino is a federally registered trademark of Engineering AG of Switzerland. It is envisioned that other microcontrollers such as USB-based Atmel® 32u4 5 volt or similar microcontrollers can also be used. Atmel is a federally registered trademark of Atmel Corporation of San Jose, Calif.
- the microcontroller 30 can be programmed to smoothly energize or brighten and then fade each of the LEDs around RGBW LED ring 22 A in a sequence.
- the leg 24 and the enclosure 28 can be made from Polylactic acid (PLA) filament using 3D printing.
- FIG. 3 shows the lighting ring assembly 20 A in conjunction with a camera C, and a lens L in a configuration for capturing a plurality of images used in animating an approximation of an object O (e.g., a coin).
- the leg 24 can include multiple connectors 32 for connecting the light ring assembly 20 A to the camera C, such that a position of the light ring assembly 20 A relative to the camera C may be selectively adjusted to provide a preferred lighting setup for the object O.
- the object O can be positioned within an illumination region R of the light ring assembly 20 A such that a first surface of the object O is approximately parallel to a focus plane of the camera C and the lens L.
- light ring assembly 20 A can define the illumination region R from a position above the object O.
- the illumination region R can include an area where light illuminated from the LEDs of the RGBW LED ring 22 A will be directed onto the surface of the object O.
- the RGBW LED ring 22 A can include a plurality of LEDs 40 arranged in a circle and a transparent diffuser 42 (shown partially fragmented to illustrate the LEDs).
- the diffuser 42 can be a 3D printed transparent PLA filament.
- the RGBW LED ring 22 A can have sixty evenly spaced LEDs mounted thereon. However other numbers of LEDs are also contemplated.
- FIGS. 5 and 6 illustrate a clamp 50 sized and shaped for selectively attaching lighting ring assembly 20 A to the lens L of the camera C.
- the clamp 50 can be 3D printed from PLA filament.
- the clamp 50 can include a flexibly resilient clip 52 comprising opposing semi-circular arms 54 extending from a stem 56 and having a base 58 opposite the clip 52 .
- the configuration and materials can provide the clip 52 sufficient flexibility so the clamp 52 can be flexed open for mounting around the lens L and sufficiently resilient to close around the lens L so as to provide a friction fit around the lens L once in a desired position.
- the lighting ring assembly 20 A and the clamp 50 can be used in combination with the camera C to provide a desired lighting setup for the object O.
- the lighting and camera setup described above can eliminate deficiencies inherent in traditional photographic methods used for objects such as coins that hide imperfections and do not capture luster. Further, the setup can enable capture of a plurality of images (e.g. a video) of uniform quality. The setup can also enable a user to perform a procedure for animating coins using a combination of ultra-high resolution videography, custom built and programmed light rigs, and a as discussed more below 3D rendering template providing an improved approximation of how the object O looks when in a viewer's hand.
- FIGS. 7-9 show an example embodiment of the lighting apparatus as a light rig assembly 20 B.
- the light rig assembly 20 B can include an elevated rectangular frame 22 B mounted on selectively removeable legs 60 and the microcontroller 30 .
- the light rig assembly 20 B can be 3D printed using a 3D printer having polylactic acid (PLA) filament.
- PLA polylactic acid
- each side of the frame 22 B can include an open rectangular cross section formed by an opaque top 62 and opposite inner and outer side walls 64 and 66 , which can define an elongated opening 68 .
- the inner side wall 64 can be translucent and the outer side wall 66 can be opaque.
- RGBW LED strips 70 can be mounted on the top 62 inside the opening 68 , and a translucent rectangular diffuser 74 can span the lower edges of the inner and outer side walls 64 and 66 to soften the light emitted from the light rig assembly 20 B.
- translucent rectangular diffuser 74 can be 3D printed.
- the RGBW LED strips 70 can comprise a plurality of light emitting diodes 72 positioned at regular intervals.
- the plurality of LEDs 72 can comprise approximately 144 LEDs per meter and include over 100-120 RGBW LEDs.
- the RGBW LED strips 70 can include Adafruit® RGBW LED light strips. Adafruit is federally registered trademarks of Adafruit Industries of New York, N.Y. It is envisioned that other RGBW LED arrangements that include a similar or higher LED density can also be used.
- each of the plurality of LEDs 72 can include a pure white LED.
- the camera C can be centered above the frame 22 B in a position for capturing the sequence of images used in animating an approximate rendering of a generally planar product P (e.g., a card or collectible) centered between lower ends of the legs 60 .
- a generally planar product P e.g., a card or collectible
- the selectively removeable legs 60 can be removed and replaced with alternative legs having a different length to change a height at which the frame 22 B and LEDs 72 are elevated above the product P.
- the height at which the camera C is positioned above the product P may also be altered by adjusting or changing a support thereof (not shown). Accordingly, the position of the light rig assembly 20 B relative to the camera C may be selectively adjusted to provide a preferred lighting setup for the product P.
- the frame 22 B can define the illumination region R from a position above the object P.
- the illumination region R includes the area where light illuminated from the plurality of LEDs 72 will be directed onto the surface of the object P.
- the lighting and camera setup described above can eliminate deficiencies inherent in traditional photographic methods used for generally planar products such as cards and collectibles that hide imperfections. Further, the setup can enable capturing a uniform quality video for use in a procedure for animating the cards or collectibles to provide an improved approximation of how a card or collectible looks when in a viewer's hand.
- the microcontroller 30 can be configured or programmed to activate the light ring assembly 20 A and/or the light rig assembly 20 B to sequentially fade the LEDs 40 and/or the LEDs 72 on and off to simulate light rotating around the object O and/or the object P.
- the microcontroller 30 gradually dims a first of the LEDs 40 or 72
- the microcontroller gradually brightens a next LED in sequence until the dimming LED is fully off, and the brightening LED is fully lit.
- the microcontroller 30 repeats the control for a next pair of LEDs, gradually dimming the fully lit LED and gradually brightening the next LED in the sequence.
- the sequence of fully energized LEDs progresses around the RGBW LED ring 22 A (see e.g. direction D in FIG. 4 ) and the frame 22 B.
- the microcontroller 30 causes the direction from which the objects O and P are lit to orbit in a corresponding circular or rectangular path respectively.
- the translucent diffusers 42 and 74 can smooth any light transitions from LED to LED to assist in providing a realistic approximation of light moving around the products O and P.
- the microcontroller 30 can be configured to control a different number of active ones of the plurality of LEDs 40 and 72 at once in the sequential operation to produce the rotational effect.
- the microcontroller can be configured to control 2 LEDs at once and in other embodiments can be configured to control 4 LEDs at once.
- FIG. 10 shows a graph of brightness over time for three sequentially ordered and adjacent LEDS 1, 2, and 3 of the LEDs 40 and 72 in a lighting assembly according to disclosed embodiments where the microcontroller 30 controls 2 LEDs at once. As seen in FIG. 10 , at a time T0, LED 1 is at a maximum brightness and LEDs 2 and 3 are off.
- the microcontroller 30 can be configured to dim the LED 1 from a fully on maximum brightness at T0 to fully off at T1 while brightening or energizing the LED 2 from fully off at T0 to the fully on maximum brightness at T1.
- the microcontroller 30 can be configured to switch from controlling the LED 1 to controlling the LED 3 such that from the time T1 to T2 the microcontroller 30 can be configured to dim the LED 2 from a fully on maximum brightness at T1 to fully off at T2 while brightening or energizing the LED 3 from fully off at T1 to the fully on maximum brightness at T2.
- the microcontroller 30 can be configured to repeat this sequential process on all of the LEDs 40 and 72 for a preconfigured amount of time periods sufficient to simulate the rotation of light around the objects O and P at least one full rotation.
- the preconfigured amount of time periods can be sufficient to simulate the rotation of light around the objects O and P at least two full rotations.
- the microcontroller 30 can be configured to modify the length of each time period to speed up or slow down a speed of the simulated light rotation around the objects O and P.
- FIG. 11 shows a graph of brightness over time for five sequentially ordered and adjacent LEDS 1, 2, 3, 4, and 5, of the LEDs 40 and 72 in a lighting assembly according to disclosed embodiments where the microcontroller 30 controls 4 LEDs at once.
- LED 1 is at a maximum brightness
- LED 2 is at an intermediate brightness such as halfway brightened or energized
- LEDs 3, 4, and 5 are off.
- the microcontroller 30 can be configured to dim the LED 1 from a fully on maximum brightness at T0 to an intermediate brightness at T1 while brightening or energizing the LED 2 from the intermediate brightness at T0 to the fully on maximum brightness at T1 and brightening or energizing the LED 3 from fully off at T0 to the intermediate brightness at T1.
- the microcontroller 30 can be configured to begin controlling the LED 4 along with LEDs 1, 2, and 3 such that from the time T1 to T2 the microcontroller 30 can be configured to dim the LED 1 from the intermediate brightness at T1 to fully off at T2, dim the LED 2 from the fully on maximum brightness at T1 to the intermediate brightness at T2 while brightening or energizing the LED 3 from the intermediate brightness at T1 to the fully on maximum brightness at T2 and brightening or energizing the LED 4 from fully off at T1 to the intermediate brightness at T2.
- the microcontroller 30 can be configured to stop controlling the LED 1 and begin controlling the LED 5 along with LEDs 2, 3, and 4 such that from the time T2 to T3 the microcontroller 30 can be configured to dim the LED 2 from the intermediate brightness at T2 to fully off at T3, dim the LED 3 from the fully on maximum brightness at T2 to the intermediate brightness at T3 while brightening or energizing the LED 4 from the intermediate brightness at T2 to the fully on maximum brightness at T3 and brightening or energizing the LED 5 from fully off at T2 to the intermediate brightness at T3.
- the microcontroller 30 can be configured to repeat this sequential process on all of the LEDs 40 and 72 for a preconfigured amount of time periods, including the time periods between T3 and T4 and between T4 and T5, sufficient to simulate the rotation of light around the objects O and P at least one full rotation.
- the preconfigured amount of time periods can be sufficient to simulate the rotation of light around the objects O and P at least two full rotations.
- FIGS. 10 and 11 show a linear transition in brightness of the LEDs 40 and 72 over time
- other embodiments employing non-linear brightness transitions are also contemplated.
- various other embodiments where the microcontroller 30 simultaneously controls an amount of LEDS different from 2 and 4 are also contemplated.
- the LEDs 40 and 72 of the light ring assembly 20 A and/or the light rig assembly 20 B can be substituted for one or more LEDs that the microcontroller 30 can direct to physically move around the light ring assembly 20 A and/or the light rig assembly 20 B to produce the light rotation effect on the surfaces of the objects O and P.
- the microcontroller 30 can be configured to modify the length of each time period to speed up or slow down a speed of the simulated light rotation around the objects O and P.
- the light ring assembly 20 A and/or the light rig assembly 20 B can be used in methods that facilitate creation of files depicting an animated approximation of the objects O and P.
- such methods can include positioning the objects O and P within illumination regions R of the light ring assembly 20 A and/or the light rig assembly 20 B such that a first surface of the objects O and P is approximately parallel to the focus plane of the camera C.
- the methods can include activating the light ring assembly 20 A and/or the light rig assembly 20 B with the microcontroller 30 to rotate light around the first surfaces as described herein.
- the methods can also include capturing first sequences of images of the first surfaces of the objects O and P with the camera C while the light ring assembly 20 A and/or the light rig assembly 20 B are activated.
- the methods can include repositioning the objects O and P within the illumination regions R such that second surfaces of the object are approximately parallel to the focus plane of the camera C, again activating the light ring assembly 20 A and/or the light rig assembly 20 B with the microcontroller 30 , and capturing second sequences of images of the second surfaces.
- the second surfaces can be opposite the first surfaces on the objects O and P.
- the first and/or second sequences of images can include ultra-high-resolution (e.g., 4K resolution taken at 30 frames per second (FPS) or better) videos. In some embodiments, higher resolution videos such as 8K can also be used.
- ultra-high-resolution e.g., 4K resolution taken at 30 frames per second (FPS) or better
- FPS frames per second
- 8K higher resolution videos
- the methods can include importing those sequences of images into “stage I” or first stage video files having a preconfigured size and masked off such that the first surface appears over a first background.
- FIG. 12 shows a graphical user interface display where a sequences of images 76 A of the object O (e.g. a coin) captured by the camera C in the light ring assembly 20 A are processed to mask off the background and generate a first stage video file 78 A shown also in FIG. 14 .
- FIG. 13 shows a graphical user interface display where a sequences of images 76 B of the object P (e.g.
- a card) captured by the camera C in the light rig assembly 20 B are processed to mask off the background and generate a first stage video file 78 B shown also in FIG. 15 .
- additional video post processing (white balance, color and brightness, etc.) can be performed when the first stage video files are created. Processing the sequence of images into the first stage video files can ensure that every animation output by these methods has a consistent size and enables this step of the animation process to be automatically performed by a programable processor.
- the first stage video files can be exported in a resolution of at least 3840 ⁇ 2160 pixels.
- the first stage video files can be formed using off the shelf video editing software such as Adobe Premiere or its equivalents. Adobe Premiere® is a federally registered trademark of Adobe Inc. of San Jose, Calif.
- the second sequence of images of second surfaces of the objects O and P can be imported into the first stage video files 78 A and 78 B after the first sequence of images 76 A and 76 B.
- the second surfaces can appear over the first background in the location of the first surfaces within the first stage video files 78 A and 78 B.
- the transitions in the first stage video files 78 A and 78 B between the first surface and the second surface can be smoothed out with a small time overlap in the images and include an approximate alignment of any light reflections on the first and second surfaces.
- each of the first and second sequences of images can be approximately 10 seconds and any overlap can be approximately one second. However, different lengths for each of these portions of the first stage video files are also contemplated.
- the methods can include incorporating the first stage video files into templates of a three-dimensional (3D) model by aligning the first surface with a video mask of the template such that the video mask obscures the first background of the first stage video file and superimposes a location of the first surface in the first stage video file onto a virtual surface of the 3D model.
- the templates can be selected such that the 3D models therein approximately match the shapes of the objects O and P.
- the 3D model can be configured in the template to cast shadows onto a second background of the template as the 3D model is rotated relative to the second background.
- FIGS. 16 and 17 show the first stage video files 78 A and 78 B of the objects O and P mapped on to corresponding 3D models in a 3D scene having a 3D environment background color B (generally white), a 3D light T casting shadows S from the 3D model to the background but not affecting the background so the background can remain the desired background color, and a 3D camera viewpoint V to adjust the view in which to render the files output by the method.
- a 3D environment background color B generally white
- a 3D light T casting shadows S from the 3D model to the background but not affecting the background so the background can remain the desired background color
- a 3D camera viewpoint V to adjust the view in which to render the files output by the method.
- the 3D models can be vertically positioned on their edges with the solid background colors B.
- the 3D model can then be animated to rotate a specified number of degrees as the 3D light source T casts the realistic shadows S from the 3D models onto the backgrounds.
- the shadows S can naturally change as the 3D model is rotated under the 3D light T.
- the rotation, speed, and direction can all be changed as needed.
- a video mask can be created to mask out the black background from the first stage video files and then applied to virtual surfaces of the 3D models as a video texture.
- the video mask can be various shapes and sizes that match the shapes and sizes of the objects O and P.
- the video mask for the object O can include a circle
- the video mask for the object P can be a rectangle.
- the video mask can be custom made to fit any shape including non-round coins or medals.
- the video masks cast the perfectly rendered 3D shadow S onto the background as the 3D model rotates.
- the rotation of the 3D model combined with the video texture created from rotating light in the first stage video files from the light ring assembly 20 A and/or the light rig assembly 20 B can result in a realistic effect that provides a good approximation of in hand appearance.
- the background colors B can be set to accept the shadows S from the 3D model, but not to accept light from the 3D light source T causing the shadows S of the coin to be shown on the background as it rotates while the rest of the background in the animation remains the selected color.
- rotation of the light as seen on the first surfaces in the first stage video files can be aligned with rotation of the shadows when incorporating the first stage video files into the templates.
- rotation of the 3D models within the templates can be limited to ensure the best possible match and alignment between the shadows S and the light redefected on the first and/or second surfaces.
- rotation of the 3D model can be limited to along the Y-axis in 3D space and further limited to a range of approximately less than 50 degrees. In some embodiments, the range of rotation can be limited from approximately 20 degrees to ⁇ 20 degrees over a 10 second time frame along the Y axis.
- the methods can include output the combined first stage video file and template as the files depicting the animated approximations of the object O and P.
- the outputted file can include a digital 3D object file.
- the outputted file can include a video file.
- the outputted file can include a compressed H.264 video mp4 format file that can be used on websites, television broadcasts, or mobile devices.
- the resulting animations can also be integrated into an official product itself where scanning or looking up the product can display the animation of the product.
- One example use could include a product unboxing video live stream.
- the after the first stage video files are combined with the templates, the combined first stage video files and templates can be passed to a 3 D video editing software for final rendering of the final file output by the methods.
- the 3D video editing software can be used to add logos, modify the background color B, and/or change the video masks in the combined first stage video files and templates.
- the 3D video editing software can include After Effects® or the like. After Effects is a federally registered trademark of Adobe Inc.
- FIG. 20 shows a flow diagram of a method 100 for animating a collectable coin according to disclosed embodiments.
- the method 100 can include starting a submission and activating the ring light assembly 22 A as in 101 and 102 . Then, the method 100 can include placing the coin on a stand, adjusting a focus of the camera C, recording 20 seconds of video of each side of the coin twice and replacing the coin in it submission box, as in steps 104 - 112 . Then, the method 100 can include determining whether the current coin is a last submission, as in 114 . When the coin is not the last submission, the method 100 can include placing a next coin on the stand and repeating steps 104 - 114 .
- the method 100 can include processing the video files from the camera C into a workstation, importing a single obverse and reverse side of a coin into a video editing platform, selecting 10 seconds from each side of the coin, sizing the video file, adjusting the color and/or lighting, exporting a stage 1 video file of the coin in at least a 4K resolution, and naming the exported file according to a customer naming convention and description as in steps 116 - 126 . Then, the method 100 can include determining whether the current coin video being processed is the last submission, as in 128 . When, the current coin video is not the last submission to be processed, the method 100 can include placing repeating steps 118 - 128 on a next coin video.
- the method 100 can include importing the stage 1 video files into a 3D animation template and adjusting the 3D lighting effects as needed, as in 130 and 132 . Then, the method 100 can include determining whether the coin is round, as in 134 . When the coin is not round, the method 100 can include adjusting a video mask, as in 106 . In any event, when the coin is round or the video mask has been adjusted, the method 100 can include sending a render composition to a render queue and setting the render quality, an output folder, a file name, and starting the render, as in 138 and 140 . Then, the method 100 can include determining whether the current 3D coin render is a last to be processed, as in 142 .
- the method 100 can include repeating steps 132 - 142 . However, when the current 3D coin render is the last to be processed, the method 100 can include zipping together all of the 3D rendered file, updating a site database, emailing the customer, and ending the submission process, as in steps 144 - 148 .
- FIG. 21 shows a flow diagram of a method 200 for animating a collectable card according to disclosed embodiments.
- the method 200 is similar to the method 100 and can include starting a submission and activating the light rig assembly 22 B as in 201 and 202 . Then, the method 200 can include placing the card on a stand, adjusting a focus of the camera C, recording 50 seconds of video of each side of the card twice and replacing the card in it submission box, as in steps 204 - 112 . Then, the method 200 can include determining whether the current card is a last submission, as in 214 . When the card is not the last submission, the method 200 can include placing a next card on the stand and repeating steps 204 - 114 .
- the method 200 can include processing the video files from the camera C into a workstation, importing a single obverse and reverse side of a card into a video editing platform, selecting 10 seconds from each side of the card, sizing the video file, adjusting the color and/or lighting, exporting a stage 1 video file of the card in at least a 4K resolution, and naming the exported file according to a customer naming convention and description as in steps 216 - 226 . Then, the method 200 can include determining whether the current card video being processed is the last submission, as in 228 . When, the current card video is not the last submission to be processed, the method 200 can include placing repeating steps 218 - 228 on a next card video.
- the method 200 can include importing the stage 1 video files into a 3D animation template and adjusting the 3D lighting effects as needed, as in 230 and 232 . Then, the method 200 can include determining whether the card is round, as in 234 . When the card is not round, the method 200 can include adjusting a video mask, as in 206 . In any event, when the card is round or the video mask has been adjusted, the method 200 can include sending a render composition to a render queue and setting the render quality, an output folder, a file name, and starting the render, as in 238 and 240 .
- the method 200 can include determining whether the current 3D card render is a last to be processed, as in 242 . When the current 3D card render is not the last to be processed, the method 200 can include repeating steps 232 - 242 . However, when the current 3D card render is the last to be processed, the method 200 can include zipping together all of the 3D rendered file, updating a site database, emailing the customer, and ending the submission process, as in steps 244 - 248 .
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Business, Economics & Management (AREA)
- Computer Graphics (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Accounting & Taxation (AREA)
- Strategic Management (AREA)
- Finance (AREA)
- Development Economics (AREA)
- General Business, Economics & Management (AREA)
- Entrepreneurship & Innovation (AREA)
- Economics (AREA)
- Marketing (AREA)
- Software Systems (AREA)
- Geometry (AREA)
- Human Resources & Organizations (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Quality & Reliability (AREA)
- Tourism & Hospitality (AREA)
- Operations Research (AREA)
- Data Mining & Analysis (AREA)
- Computer Hardware Design (AREA)
- General Engineering & Computer Science (AREA)
- Game Theory and Decision Science (AREA)
- Processing Or Creating Images (AREA)
- Image Processing (AREA)
Abstract
Description
- This application claims priority to U.S. Application No. 63/105,074 filed Oct. 23, 2020 and titled “METHOD OF ANIMATING COINS AND COLLECTIBLES,” and to U.S. Application No. 63/152,134 filed Feb. 22, 2021 and titled “METHOD OF ANIMATING CARDS AND COLLECTIBLES.” U.S. Application Nos. 63/105,074 and 63/152,134 are hereby fully incorporated by reference as if set forth fully herein.
- The present invention generally relates to systems and methods for creating an animated visual approximation of a surface of an object. More particularly, such systems and methods relate to creating an animated visual approximation of a surface of a generally flat surface including a disk such as a coin or a coin-like object and a generally planar object such as a collectible card.
- In recent years, technology and world events have drastically altered the collectible and auction industries such that very few traditional collectible shows or in-person auctions happen and most sales now occur through online marketplaces. With such a major shift in buying practices, collectibles sellers have a need for an improved method of representing products online and visually approximating the appearance of these products, particularly how the products would look in a potential buyer's hand.
- Although photographs provide static views, photos do not accurately show how light interacts with the surfaces of a collectible object. In particular, photographs have difficulty rendering complex surface details such as foiling, holograms, and other effects that change as an observer moves or as lighting changes on a collectible. Further, photographs can often hide scratches and marks collectors need to see to accurately judge the quality and condition of a collectible. For example, photographs do not accurately capture the luster of a coin (i.e., how changes in light reflect or play across the surfaces of a coin) and are prone to concealing details that might otherwise be visible if viewed in hand. Such details might include scratches or other signs of broken luster that could indicate cleaning or alterations. In addition, manufacturers of newer collectable objects frequently develop new and innovative products having novel features and effects to entice collecting consumers, but as physical card shops and shows become less common, consumers have limited opportunities to physically interact with the products in hand.
- In the current collectable market high bid increments are sometimes used when auctioning collectibles. This fact means that a single additional bid can result in a sales price that is thousands or tens of thousands of dollars higher. With many bidders participating online and few being able to view the underlying object in hand, auction houses rely on online presentations to provide buyers the confidence in an object's quality needed to bid. Astute buyers are aware of the limitations of photographs and generally are unimpressed by quick cellphone videography or other such methods commonly deployed by the auction houses.
- In light of the foregoing issues, there is a continuing need for improved systems and methods for fully capturing and representing collectible objects for viewing remotely via an online or similar platform.
- This application will become more fully understood from the following detailed description, taken in conjunction with the accompanying figures, wherein like reference numerals refer to like elements in which:
-
FIG. 1 is a perspective of a light ring assembly according to disclosed embodiments; -
FIG. 2 is a perspective of a light ring assembly according to disclosed embodiments; -
FIG. 3 is a perspective of a light ring assembly in combination with a camera according to disclosed embodiments; -
FIG. 4 is a perspective of a light ring assembly illustrating a direction of lighting according to disclosed embodiments; -
FIG. 5 is a perspective of a lens attachment according to disclosed embodiments; -
FIG. 6 is a perspective of a lens attachment according to disclosed embodiments; -
FIG. 7 is a perspective of a light rig assembly in combination with a camera according to disclosed embodiments; -
FIG. 8 is a perspective of a light rig assembly in combination with a camera according to disclosed embodiments; -
FIG. 9 is a cross section of a frame of the light rig assembly ofFIG. 8 taken in the plane of line 3-3 ofFIG. 8 ; -
FIG. 10 is graph of brightness over time for various LEDs in a lighting assembly according to disclosed embodiments; -
FIG. 11 is graph of brightness over time for various LEDs in a lighting assembly according to disclosed embodiments; -
FIGS. 12 and 13 are a graphical user interface within video editing software used during a “Stage 1” video creation process according to disclosed embodiments; -
FIGS. 14 and 15 are a graphical user interface within video editing software showing collectible objects masked off with a black background according to disclosed embodiments; -
FIGS. 16 and 17 are a graphical user interface within video editing software showing collectible objects during animation according to disclosed embodiments; -
FIGS. 18 and 19 are a graphical user interface within video editing software showing a final rendering of an animation according to disclosed embodiments; and -
FIGS. 20 and 21 are flowcharts illustrating methods of preparing an animated approximation according to disclosed embodiments. - Embodiments described herein relate to systems and methods for virtually animating an object such as a collectible item. In particular, the present disclosure provides solutions to the industry-wide problems of remote viewing of such object by rendering an accurate animation of how such objects look in hand via an interactive online experience of such animations where only offline experiences were once available. In some embodiments, the systems and methods described herein can produce ultra-high-resolution animations depicting how luster moves over an object's surfaces while rotating, allowing possible imperfections to be revealed or that show off the best aspects of the surfaces (e.g., foil, refractor, hologram, rainbow, etc.). Additionally, the systems and methods described herein can provide the rendered animations in industry standard, web-friendly, video file formats customers can use without additional cost or inconvenience. For example, in some embodiments, the video animations can be displayed using streaming platforms to cater to online video streamers. In general, the systems and methods described herein provide animated images that can increase online buyer confidence and increase bids for objects in the marketplace. Such systems and methods can also provide object manufacturers high quality online presentations enabling them to better market their products and enables collectors to demonstrate the in-hand appearance of a collectible online.
- Some embodiments disclosed herein can be directed to a lighting apparatus for use in simulating animation of an object. In particular, the lighting apparatus can include mounting connectors configured to couple a camera to the lighting apparatus at a plurality of positions, a ring or frame positioned above and defining an illumination region and comprising a plurality of light emitting diodes (LEDs), and a microcontroller configured to control the plurality of LEDs to rotate light around a first surface of the object placed within the illumination region. In some embodiments, the plurality of LEDs can include pure white LEDs which produce more accurate and appealing final animation than color RGB LEDs that simulated white by combining multiple color outputs together.
- In some embodiments, the microcontroller can control the plurality of LEDs by dimming and brightening each of the plurality of LEDs in a sequenced order around the ring or frame such that a first of the plurality of LEDs is being dimmed from fully on to fully off while a second of the plurality of LEDs is being brightened from fully off to fully on. Furthermore, the microcontroller can control the first of the plurality of LEDs and the second of the plurality of LEDs such that the first of the plurality LEDs reaches fully off simultaneous with the second of the plurality LEDs reaching fully on, and once the first of the plurality LEDs reaches fully off and the second of the plurality LEDs reaches fully on, the microcontroller can begin to dim the second of the plurality LEDs to fully off and begins to control a next one of the plurality of LEDs in the sequenced order to brighten the next one of the plurality of LEDs from fully off to fully on.
-
FIGS. 1-4 show an example embodiment of the lighting apparatus as alighting ring assembly 20A. Thelighting ring assembly 20A can include aRGBW LED ring 22A that is pivotably connected to aleg 24 by ahinge 26. In some embodiments theRGBW LED ring 22A can have a 6 inch diameter. In some embodiments, theRGBW LED ring 22A can include Adafruit® NeoPixel® 60 ring-5050 RGBW LED having integrated drivers and natural white, 4500K LEDs. Adafruit and NeoPixel are federally registered trademarks of Adafruit Industries of New York, N.Y. It is envisioned that other LED rings could be used that include similar or better LED density. In some embodiments, thelighting ring assembly 20A can include anenclosure 28 provided at an end of theleg 24 opposite thering 22A that houses amicrocontroller 30. In some embodiments, themicrocontroller 30 can include an Arduino® 5V microcontroller. Arduino is a federally registered trademark of Arduino AG of Switzerland. It is envisioned that other microcontrollers such as USB-based Atmel® 32u4 5 volt or similar microcontrollers can also be used. Atmel is a federally registered trademark of Atmel Corporation of San Jose, Calif. In some embodiments, themicrocontroller 30 can be programmed to smoothly energize or brighten and then fade each of the LEDs aroundRGBW LED ring 22A in a sequence. In some embodiments, theleg 24 and theenclosure 28 can be made from Polylactic acid (PLA) filament using 3D printing. -
FIG. 3 shows thelighting ring assembly 20A in conjunction with a camera C, and a lens L in a configuration for capturing a plurality of images used in animating an approximation of an object O (e.g., a coin). As shown, theleg 24 can includemultiple connectors 32 for connecting thelight ring assembly 20A to the camera C, such that a position of thelight ring assembly 20A relative to the camera C may be selectively adjusted to provide a preferred lighting setup for the object O. As seen inFIG. 3 , in some embodiments, the object O can be positioned within an illumination region R of thelight ring assembly 20A such that a first surface of the object O is approximately parallel to a focus plane of the camera C and the lens L. In some embodiments,light ring assembly 20A can define the illumination region R from a position above the object O. In particular, the illumination region R can include an area where light illuminated from the LEDs of theRGBW LED ring 22A will be directed onto the surface of the object O. - In some embodiments, as seen in
FIG. 4 , theRGBW LED ring 22A can include a plurality of LEDs 40 arranged in a circle and a transparent diffuser 42 (shown partially fragmented to illustrate the LEDs). Thediffuser 42 can be a 3D printed transparent PLA filament. In some embodiments, theRGBW LED ring 22A can have sixty evenly spaced LEDs mounted thereon. However other numbers of LEDs are also contemplated. -
FIGS. 5 and 6 illustrate aclamp 50 sized and shaped for selectively attachinglighting ring assembly 20A to the lens L of the camera C. In some embodiments, theclamp 50 can be 3D printed from PLA filament. As seen inFIGS. 5 and 6 , theclamp 50 can include a flexiblyresilient clip 52 comprising opposingsemi-circular arms 54 extending from astem 56 and having a base 58 opposite theclip 52. The configuration and materials can provide theclip 52 sufficient flexibility so theclamp 52 can be flexed open for mounting around the lens L and sufficiently resilient to close around the lens L so as to provide a friction fit around the lens L once in a desired position. As will be appreciated by those skilled in the art, thelighting ring assembly 20A and theclamp 50 can be used in combination with the camera C to provide a desired lighting setup for the object O. - The lighting and camera setup described above can eliminate deficiencies inherent in traditional photographic methods used for objects such as coins that hide imperfections and do not capture luster. Further, the setup can enable capture of a plurality of images (e.g. a video) of uniform quality. The setup can also enable a user to perform a procedure for animating coins using a combination of ultra-high resolution videography, custom built and programmed light rigs, and a as discussed more below 3D rendering template providing an improved approximation of how the object O looks when in a viewer's hand.
-
FIGS. 7-9 show an example embodiment of the lighting apparatus as alight rig assembly 20B. As seen inFIGS. 7 and 8 , in some embodiments, thelight rig assembly 20B can include an elevatedrectangular frame 22B mounted on selectivelyremoveable legs 60 and themicrocontroller 30. In some embodiments, thelight rig assembly 20B can be 3D printed using a 3D printer having polylactic acid (PLA) filament. Furthermore, as shown inFIG. 9 , each side of theframe 22B can include an open rectangular cross section formed by an opaque top 62 and opposite inner andouter side walls elongated opening 68. In some embodiments, theinner side wall 64 can be translucent and theouter side wall 66 can be opaque. RGBW LED strips 70 can be mounted on the top 62 inside theopening 68, and a translucentrectangular diffuser 74 can span the lower edges of the inner andouter side walls light rig assembly 20B. In some embodiments, translucentrectangular diffuser 74 can be 3D printed. - In some embodiments, the RGBW LED strips 70 can comprise a plurality of
light emitting diodes 72 positioned at regular intervals. In some embodiments, the plurality ofLEDs 72 can comprise approximately 144 LEDs per meter and include over 100-120 RGBW LEDs. In some embodiments, the RGBW LED strips 70 can include Adafruit® RGBW LED light strips. Adafruit is federally registered trademarks of Adafruit Industries of New York, N.Y. It is envisioned that other RGBW LED arrangements that include a similar or higher LED density can also be used. In some embodiments, each of the plurality ofLEDs 72 can include a pure white LED. - Furthermore, as seen in
FIGS. 7 and 8 the camera C can be centered above theframe 22B in a position for capturing the sequence of images used in animating an approximate rendering of a generally planar product P (e.g., a card or collectible) centered between lower ends of thelegs 60. As may be appreciated, the selectivelyremoveable legs 60 can be removed and replaced with alternative legs having a different length to change a height at which theframe 22B andLEDs 72 are elevated above the product P. Likewise, the height at which the camera C is positioned above the product P may also be altered by adjusting or changing a support thereof (not shown). Accordingly, the position of thelight rig assembly 20B relative to the camera C may be selectively adjusted to provide a preferred lighting setup for the product P. Similar to thelight ring assembly 20A ofFIGS. 1-4 , in some embodiments, theframe 22B can define the illumination region R from a position above the object P. In particular, the illumination region R includes the area where light illuminated from the plurality ofLEDs 72 will be directed onto the surface of the object P. - The lighting and camera setup described above can eliminate deficiencies inherent in traditional photographic methods used for generally planar products such as cards and collectibles that hide imperfections. Further, the setup can enable capturing a uniform quality video for use in a procedure for animating the cards or collectibles to provide an improved approximation of how a card or collectible looks when in a viewer's hand.
- In operation, the
microcontroller 30 can be configured or programmed to activate thelight ring assembly 20A and/or thelight rig assembly 20B to sequentially fade the LEDs 40 and/or theLEDs 72 on and off to simulate light rotating around the object O and/or the object P. In particular, as themicrocontroller 30 gradually dims a first of theLEDs 40 or 72, the microcontroller gradually brightens a next LED in sequence until the dimming LED is fully off, and the brightening LED is fully lit. Then, themicrocontroller 30 repeats the control for a next pair of LEDs, gradually dimming the fully lit LED and gradually brightening the next LED in the sequence. The sequence of fully energized LEDs progresses around theRGBW LED ring 22A (see e.g. direction D inFIG. 4 ) and theframe 22B. Thus, themicrocontroller 30 causes the direction from which the objects O and P are lit to orbit in a corresponding circular or rectangular path respectively. In some embodiments, thetranslucent diffusers - In some embodiments, the
microcontroller 30 can be configured to control a different number of active ones of the plurality ofLEDs 40 and 72 at once in the sequential operation to produce the rotational effect. For example, in some embodiments the microcontroller can be configured to control 2 LEDs at once and in other embodiments can be configured to control 4 LEDs at once. In particular,FIG. 10 shows a graph of brightness over time for three sequentially ordered andadjacent LEDS LEDs 40 and 72 in a lighting assembly according to disclosed embodiments where themicrocontroller 30controls 2 LEDs at once. As seen inFIG. 10 , at a time T0,LED 1 is at a maximum brightness andLEDs microcontroller 30 can be configured to dim theLED 1 from a fully on maximum brightness at T0 to fully off at T1 while brightening or energizing theLED 2 from fully off at T0 to the fully on maximum brightness at T1. Next, once theLED 1 reaches fully off at T1, themicrocontroller 30 can be configured to switch from controlling theLED 1 to controlling theLED 3 such that from the time T1 to T2 themicrocontroller 30 can be configured to dim theLED 2 from a fully on maximum brightness at T1 to fully off at T2 while brightening or energizing theLED 3 from fully off at T1 to the fully on maximum brightness at T2. Finally, themicrocontroller 30 can be configured to repeat this sequential process on all of theLEDs 40 and 72 for a preconfigured amount of time periods sufficient to simulate the rotation of light around the objects O and P at least one full rotation. In some embodiments, the preconfigured amount of time periods can be sufficient to simulate the rotation of light around the objects O and P at least two full rotations. Furthermore, in some embodiments, themicrocontroller 30 can be configured to modify the length of each time period to speed up or slow down a speed of the simulated light rotation around the objects O and P. - Similar to
FIG. 10 ,FIG. 11 shows a graph of brightness over time for five sequentially ordered andadjacent LEDS LEDs 40 and 72 in a lighting assembly according to disclosed embodiments where themicrocontroller 30controls 4 LEDs at once. As seen inFIG. 11 , at a time T0,LED 1 is at a maximum brightness,LED 2 is at an intermediate brightness such as halfway brightened or energized, andLEDs microcontroller 30 can be configured to dim theLED 1 from a fully on maximum brightness at T0 to an intermediate brightness at T1 while brightening or energizing theLED 2 from the intermediate brightness at T0 to the fully on maximum brightness at T1 and brightening or energizing theLED 3 from fully off at T0 to the intermediate brightness at T1. Next, from time T1 to T2, themicrocontroller 30 can be configured to begin controlling theLED 4 along withLEDs microcontroller 30 can be configured to dim theLED 1 from the intermediate brightness at T1 to fully off at T2, dim theLED 2 from the fully on maximum brightness at T1 to the intermediate brightness at T2 while brightening or energizing theLED 3 from the intermediate brightness at T1 to the fully on maximum brightness at T2 and brightening or energizing theLED 4 from fully off at T1 to the intermediate brightness at T2. Then, from time T2 to T3, themicrocontroller 30 can be configured to stop controlling theLED 1 and begin controlling theLED 5 along withLEDs microcontroller 30 can be configured to dim theLED 2 from the intermediate brightness at T2 to fully off at T3, dim theLED 3 from the fully on maximum brightness at T2 to the intermediate brightness at T3 while brightening or energizing theLED 4 from the intermediate brightness at T2 to the fully on maximum brightness at T3 and brightening or energizing theLED 5 from fully off at T2 to the intermediate brightness at T3. Finally, as seen inFIG. 11 , themicrocontroller 30 can be configured to repeat this sequential process on all of theLEDs 40 and 72 for a preconfigured amount of time periods, including the time periods between T3 and T4 and between T4 and T5, sufficient to simulate the rotation of light around the objects O and P at least one full rotation. In some embodiments, the preconfigured amount of time periods can be sufficient to simulate the rotation of light around the objects O and P at least two full rotations. - It should be noted that while
FIGS. 10 and 11 show a linear transition in brightness of theLEDs 40 and 72 over time, other embodiments employing non-linear brightness transitions are also contemplated. Furthermore, various other embodiments where themicrocontroller 30 simultaneously controls an amount of LEDS different from 2 and 4 are also contemplated. Finally, in some embodiments, theLEDs 40 and 72 of thelight ring assembly 20A and/or thelight rig assembly 20B can be substituted for one or more LEDs that themicrocontroller 30 can direct to physically move around thelight ring assembly 20A and/or thelight rig assembly 20B to produce the light rotation effect on the surfaces of the objects O and P. Furthermore, in some embodiments, themicrocontroller 30 can be configured to modify the length of each time period to speed up or slow down a speed of the simulated light rotation around the objects O and P. - As described herein, the
light ring assembly 20A and/or thelight rig assembly 20B can be used in methods that facilitate creation of files depicting an animated approximation of the objects O and P. For example, such methods can include positioning the objects O and P within illumination regions R of thelight ring assembly 20A and/or thelight rig assembly 20B such that a first surface of the objects O and P is approximately parallel to the focus plane of the camera C. Then, the methods can include activating thelight ring assembly 20A and/or thelight rig assembly 20B with themicrocontroller 30 to rotate light around the first surfaces as described herein. The methods can also include capturing first sequences of images of the first surfaces of the objects O and P with the camera C while thelight ring assembly 20A and/or thelight rig assembly 20B are activated. - In some embodiments, after capturing the first sequences of images of the first surfaces, the methods can include repositioning the objects O and P within the illumination regions R such that second surfaces of the object are approximately parallel to the focus plane of the camera C, again activating the
light ring assembly 20A and/or thelight rig assembly 20B with themicrocontroller 30, and capturing second sequences of images of the second surfaces. In some embodiments, the second surfaces can be opposite the first surfaces on the objects O and P. - In some embodiments, the first and/or second sequences of images can include ultra-high-resolution (e.g., 4K resolution taken at 30 frames per second (FPS) or better) videos. In some embodiments, higher resolution videos such as 8K can also be used.
- After capturing the first sequences and/or second sequences of images, the methods can include importing those sequences of images into “stage I” or first stage video files having a preconfigured size and masked off such that the first surface appears over a first background. For example,
FIG. 12 shows a graphical user interface display where a sequences ofimages 76A of the object O (e.g. a coin) captured by the camera C in thelight ring assembly 20A are processed to mask off the background and generate a firststage video file 78A shown also inFIG. 14 . Similarly,FIG. 13 shows a graphical user interface display where a sequences ofimages 76B of the object P (e.g. a card) captured by the camera C in thelight rig assembly 20B are processed to mask off the background and generate a firststage video file 78B shown also inFIG. 15 . In some embodiments, additional video post processing (white balance, color and brightness, etc.) can be performed when the first stage video files are created. Processing the sequence of images into the first stage video files can ensure that every animation output by these methods has a consistent size and enables this step of the animation process to be automatically performed by a programable processor. In some embodiments, the first stage video files can be exported in a resolution of at least 3840×2160 pixels. In some embodiments, the first stage video files can be formed using off the shelf video editing software such as Adobe Premiere or its equivalents. Adobe Premiere® is a federally registered trademark of Adobe Inc. of San Jose, Calif. - In some embodiments, the second sequence of images of second surfaces of the objects O and P can be imported into the first
stage video files images stage video files stage video files - After generating the first stage video files such as the first
stage video files - In some embodiments, the 3D model can be configured in the template to cast shadows onto a second background of the template as the 3D model is rotated relative to the second background. For example,
FIGS. 16 and 17 show the firststage video files - As seen in
FIGS. 16 and 17 , when the first stage video files are then taken into the 3D editing environment where they are incorporated with the templates, the 3D models can be vertically positioned on their edges with the solid background colors B. The 3D model can then be animated to rotate a specified number of degrees as the 3D light source T casts the realistic shadows S from the 3D models onto the backgrounds. The shadows S can naturally change as the 3D model is rotated under the 3D light T. The rotation, speed, and direction can all be changed as needed. Furthermore, in some embodiments, a video mask can be created to mask out the black background from the first stage video files and then applied to virtual surfaces of the 3D models as a video texture. The video mask can be various shapes and sizes that match the shapes and sizes of the objects O and P. For example, the video mask for the object O can include a circle, the video mask for the object P can be a rectangle. However, the video mask can be custom made to fit any shape including non-round coins or medals. The video masks cast the perfectly rendered 3D shadow S onto the background as the 3D model rotates. The rotation of the 3D model combined with the video texture created from rotating light in the first stage video files from thelight ring assembly 20A and/or thelight rig assembly 20B can result in a realistic effect that provides a good approximation of in hand appearance. Furthermore, in some embodiments, the background colors B can be set to accept the shadows S from the 3D model, but not to accept light from the 3D light source T causing the shadows S of the coin to be shown on the background as it rotates while the rest of the background in the animation remains the selected color. - In some embodiments, rotation of the light as seen on the first surfaces in the first stage video files can be aligned with rotation of the shadows when incorporating the first stage video files into the templates. Similarly, in some embodiments, rotation of the 3D models within the templates can be limited to ensure the best possible match and alignment between the shadows S and the light redefected on the first and/or second surfaces. For example, in some embodiments, rotation of the 3D model can be limited to along the Y-axis in 3D space and further limited to a range of approximately less than 50 degrees. In some embodiments, the range of rotation can be limited from approximately 20 degrees to −20 degrees over a 10 second time frame along the Y axis.
- After the first stage video files are combined with the templates, the methods can include output the combined first stage video file and template as the files depicting the animated approximations of the object O and P. In some embodiments, the outputted file can include a digital 3D object file. Additionally or alternatively, in some embodiments, the outputted file can include a video file. For example, in some embodiments, the outputted file can include a compressed H.264 video mp4 format file that can be used on websites, television broadcasts, or mobile devices. In some embodiments, the resulting animations can also be integrated into an official product itself where scanning or looking up the product can display the animation of the product. One example use could include a product unboxing video live stream.
- In some embodiments, the after the first stage video files are combined with the templates, the combined first stage video files and templates can be passed to a 3D video editing software for final rendering of the final file output by the methods. For example, as seen in
FIGS. 18 and 19 , the 3D video editing software can be used to add logos, modify the background color B, and/or change the video masks in the combined first stage video files and templates. In some embodiments, the 3D video editing software can include After Effects® or the like. After Effects is a federally registered trademark of Adobe Inc. -
FIG. 20 shows a flow diagram of amethod 100 for animating a collectable coin according to disclosed embodiments. As seen inFIG. 20 , themethod 100 can include starting a submission and activating the ringlight assembly 22A as in 101 and 102. Then, themethod 100 can include placing the coin on a stand, adjusting a focus of the camera C, recording 20 seconds of video of each side of the coin twice and replacing the coin in it submission box, as in steps 104-112. Then, themethod 100 can include determining whether the current coin is a last submission, as in 114. When the coin is not the last submission, themethod 100 can include placing a next coin on the stand and repeating steps 104-114. - However, when the current coin is the last submission, the
method 100 can include processing the video files from the camera C into a workstation, importing a single obverse and reverse side of a coin into a video editing platform, selecting 10 seconds from each side of the coin, sizing the video file, adjusting the color and/or lighting, exporting astage 1 video file of the coin in at least a 4K resolution, and naming the exported file according to a customer naming convention and description as in steps 116-126. Then, themethod 100 can include determining whether the current coin video being processed is the last submission, as in 128. When, the current coin video is not the last submission to be processed, themethod 100 can include placing repeating steps 118-128 on a next coin video. - However, when the current coin video is the last video to be processed, the
method 100 can include importing thestage 1 video files into a 3D animation template and adjusting the 3D lighting effects as needed, as in 130 and 132. Then, themethod 100 can include determining whether the coin is round, as in 134. When the coin is not round, themethod 100 can include adjusting a video mask, as in 106. In any event, when the coin is round or the video mask has been adjusted, themethod 100 can include sending a render composition to a render queue and setting the render quality, an output folder, a file name, and starting the render, as in 138 and 140. Then, themethod 100 can include determining whether the current 3D coin render is a last to be processed, as in 142. When the current 3D coin render is not the last to be processed, themethod 100 can include repeating steps 132-142. However, when the current 3D coin render is the last to be processed, themethod 100 can include zipping together all of the 3D rendered file, updating a site database, emailing the customer, and ending the submission process, as in steps 144-148. -
FIG. 21 shows a flow diagram of amethod 200 for animating a collectable card according to disclosed embodiments. As seen inFIG. 21 , themethod 200 is similar to themethod 100 and can include starting a submission and activating thelight rig assembly 22B as in 201 and 202. Then, themethod 200 can include placing the card on a stand, adjusting a focus of the camera C, recording 50 seconds of video of each side of the card twice and replacing the card in it submission box, as in steps 204-112. Then, themethod 200 can include determining whether the current card is a last submission, as in 214. When the card is not the last submission, themethod 200 can include placing a next card on the stand and repeating steps 204-114. - However, when the current card is the last submission, the
method 200 can include processing the video files from the camera C into a workstation, importing a single obverse and reverse side of a card into a video editing platform, selecting 10 seconds from each side of the card, sizing the video file, adjusting the color and/or lighting, exporting astage 1 video file of the card in at least a 4K resolution, and naming the exported file according to a customer naming convention and description as in steps 216-226. Then, themethod 200 can include determining whether the current card video being processed is the last submission, as in 228. When, the current card video is not the last submission to be processed, themethod 200 can include placing repeating steps 218-228 on a next card video. - However, when the current card video is the last video to be processed, the
method 200 can include importing thestage 1 video files into a 3D animation template and adjusting the 3D lighting effects as needed, as in 230 and 232. Then, themethod 200 can include determining whether the card is round, as in 234. When the card is not round, themethod 200 can include adjusting a video mask, as in 206. In any event, when the card is round or the video mask has been adjusted, themethod 200 can include sending a render composition to a render queue and setting the render quality, an output folder, a file name, and starting the render, as in 238 and 240. Then, themethod 200 can include determining whether the current 3D card render is a last to be processed, as in 242. When the current 3D card render is not the last to be processed, themethod 200 can include repeating steps 232-242. However, when the current 3D card render is the last to be processed, themethod 200 can include zipping together all of the 3D rendered file, updating a site database, emailing the customer, and ending the submission process, as in steps 244-248. - From the foregoing, it will be seen that the various embodiments of the present invention are well adapted to attain all the objectives and advantages hereinabove set forth together with still other advantages which are obvious and which are inherent to the present structures. It will be understood that certain features and sub-combinations of the present embodiments are of utility and may be employed without reference to other features and sub-combinations. Since many possible embodiments of the present invention may be made without departing from the spirit and scope of the present invention, it is also to be understood that all disclosures herein set forth or illustrated in the accompanying drawings are to be interpreted as illustrative only and not limiting. The various constructions described above and illustrated in the drawings are presented by way of example only and are not intended to limit the concepts, principles and scope of the present invention.
- Many changes, modifications, variations and other uses and applications of the present invention will, however, become apparent to those skilled in the art after considering the specification and the accompanying drawings. All such changes, modifications, variations and other uses and applications which do not depart from the spirit and scope of the invention are deemed to be covered by the invention which is limited only by the claims which follow.
Claims (20)
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/391,553 US20220130091A1 (en) | 2020-10-23 | 2021-08-02 | Systems and methods for simulating animation of an object |
US17/393,951 US11443470B2 (en) | 2020-10-23 | 2021-08-04 | Systems and methods for simulating animation of an object |
CA3135472A CA3135472A1 (en) | 2020-10-23 | 2021-10-22 | Systems and methods for simulating animation of an object |
GB2115339.0A GB2602548A (en) | 2020-10-23 | 2021-10-25 | Systems and methods for simulating animation of an object |
EP21204559.5A EP3998581A3 (en) | 2020-10-23 | 2021-10-25 | Systems and methods for simulating animation of an object |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US202063105074P | 2020-10-23 | 2020-10-23 | |
US202163152134P | 2021-02-22 | 2021-02-22 | |
US17/391,553 US20220130091A1 (en) | 2020-10-23 | 2021-08-02 | Systems and methods for simulating animation of an object |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/393,951 Continuation US11443470B2 (en) | 2020-10-23 | 2021-08-04 | Systems and methods for simulating animation of an object |
Publications (1)
Publication Number | Publication Date |
---|---|
US20220130091A1 true US20220130091A1 (en) | 2022-04-28 |
Family
ID=78598770
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/391,553 Abandoned US20220130091A1 (en) | 2020-10-23 | 2021-08-02 | Systems and methods for simulating animation of an object |
US17/393,951 Active US11443470B2 (en) | 2020-10-23 | 2021-08-04 | Systems and methods for simulating animation of an object |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/393,951 Active US11443470B2 (en) | 2020-10-23 | 2021-08-04 | Systems and methods for simulating animation of an object |
Country Status (4)
Country | Link |
---|---|
US (2) | US20220130091A1 (en) |
EP (1) | EP3998581A3 (en) |
CA (1) | CA3135472A1 (en) |
GB (1) | GB2602548A (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220343483A1 (en) * | 2021-04-21 | 2022-10-27 | Veery, LLC | System and method for probabilistic determination of likely grade of collectible cards |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5144495A (en) * | 1987-12-03 | 1992-09-01 | Compugrade, Inc | Systems for illuminating and evaluating surfaces |
US5646388A (en) * | 1994-09-30 | 1997-07-08 | Lau Technologies | Systems and methods for recording data |
US20070212038A1 (en) * | 2004-04-15 | 2007-09-13 | Moritex Corporation | Face Imaging Device |
US20110031889A1 (en) * | 2008-05-09 | 2011-02-10 | Shim Hyun-Seop | Ac-powered led lighting system |
US20120062719A1 (en) * | 2010-09-09 | 2012-03-15 | University Of Southern California | Head-Mounted Photometric Facial Performance Capture |
US20140109162A1 (en) * | 2011-05-06 | 2014-04-17 | Benjamin Paul Licht | System and method of providing and distributing three dimensional video productions from digitally recorded personal event files |
US20210042950A1 (en) * | 2019-08-09 | 2021-02-11 | Google Llc | Depth-Aware Photo Editing |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1998044718A2 (en) * | 1997-04-01 | 1998-10-08 | Agris-Schoen Vision Systems, Inc. | High-precision-resolution image acquisision apparatus and method |
US6749310B2 (en) * | 2001-09-07 | 2004-06-15 | Contrast Lighting Services, Inc. | Wide area lighting effects system |
JP4110514B2 (en) * | 2002-05-15 | 2008-07-02 | ソニー株式会社 | Data encoding apparatus and method, data decoding apparatus and method, data transmission apparatus, data reception apparatus, storage medium, and computer program |
US6831641B2 (en) * | 2002-06-17 | 2004-12-14 | Mitsubishi Electric Research Labs, Inc. | Modeling and rendering of surface reflectance fields of 3D objects |
US6791542B2 (en) * | 2002-06-17 | 2004-09-14 | Mitsubishi Electric Research Laboratories, Inc. | Modeling 3D objects with opacity hulls |
JP4335589B2 (en) * | 2002-06-27 | 2009-09-30 | ミツビシ・エレクトリック・リサーチ・ラボラトリーズ・インコーポレイテッド | How to model a 3D object |
US20090296365A1 (en) * | 2008-04-18 | 2009-12-03 | Coinsecure, Inc. | Calibrated and color-controlled multi-source lighting system for specimen illumination |
HU0900478D0 (en) * | 2009-07-31 | 2009-09-28 | Holografika Hologrameloeallito | Method and apparatus for displaying 3d images |
US10445330B2 (en) * | 2015-01-16 | 2019-10-15 | Tag P, Llc. | Computerized technical authentication and grading system for collectible objects |
US11721061B2 (en) | 2016-03-02 | 2023-08-08 | Nutilt Imaging, Llc | Method and apparatus for displaying a collectible |
ES2869878T3 (en) * | 2017-01-17 | 2021-10-26 | Signify Holding Bv | Adjustable spot light position generation |
-
2021
- 2021-08-02 US US17/391,553 patent/US20220130091A1/en not_active Abandoned
- 2021-08-04 US US17/393,951 patent/US11443470B2/en active Active
- 2021-10-22 CA CA3135472A patent/CA3135472A1/en active Pending
- 2021-10-25 EP EP21204559.5A patent/EP3998581A3/en active Pending
- 2021-10-25 GB GB2115339.0A patent/GB2602548A/en active Pending
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5144495A (en) * | 1987-12-03 | 1992-09-01 | Compugrade, Inc | Systems for illuminating and evaluating surfaces |
US5646388A (en) * | 1994-09-30 | 1997-07-08 | Lau Technologies | Systems and methods for recording data |
US20070212038A1 (en) * | 2004-04-15 | 2007-09-13 | Moritex Corporation | Face Imaging Device |
US20110031889A1 (en) * | 2008-05-09 | 2011-02-10 | Shim Hyun-Seop | Ac-powered led lighting system |
US20120062719A1 (en) * | 2010-09-09 | 2012-03-15 | University Of Southern California | Head-Mounted Photometric Facial Performance Capture |
US20140109162A1 (en) * | 2011-05-06 | 2014-04-17 | Benjamin Paul Licht | System and method of providing and distributing three dimensional video productions from digitally recorded personal event files |
US20210042950A1 (en) * | 2019-08-09 | 2021-02-11 | Google Llc | Depth-Aware Photo Editing |
Also Published As
Publication number | Publication date |
---|---|
US11443470B2 (en) | 2022-09-13 |
EP3998581A2 (en) | 2022-05-18 |
US20220130092A1 (en) | 2022-04-28 |
GB202115339D0 (en) | 2021-12-08 |
CA3135472A1 (en) | 2022-04-23 |
GB2602548A (en) | 2022-07-06 |
EP3998581A3 (en) | 2022-10-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11671717B2 (en) | Camera systems for motion capture | |
US7523411B2 (en) | Network-linked interactive three-dimensional composition and display of saleable objects in situ in viewer-selected scenes for purposes of object promotion and procurement, and generation of object advertisements | |
Lanier | Advanced Maya texturing and lighting | |
US11348325B2 (en) | Generating photorealistic viewable images using augmented reality techniques | |
US11443470B2 (en) | Systems and methods for simulating animation of an object | |
JP7405095B2 (en) | image display device | |
BE1019941A3 (en) | DEVICE FOR VIEWING THREE-DIMENSIONAL IMAGES, CREATION OF THREE-DIMENSIONAL IMAGES, AND METHOD FOR CREATING THREE-DIMENSIONAL IMAGES. | |
WO2023116396A1 (en) | Rendering display method and apparatus, computer device, and storage medium | |
Sheng et al. | A spatially augmented reality sketching interface for architectural daylighting design | |
US10079966B2 (en) | Systems and techniques for capturing images for use in determining reflectance properties of physical objects | |
Ponech | The substance of cinema | |
TWM630897U (en) | Article presentation device with label | |
Joon | Principles of Photorealism to develop photorealistic visualisation for Interface Design: A review | |
Pai | An imitation of 3D projection mapping using augmented reality and shader effects | |
Faria et al. | Architecture for video mapping development: The method and the application in NMSPCAM | |
KR20180039508A (en) | Hologram contents production system using chroma key | |
US20230336679A1 (en) | Motion capture using synchronized and aligned devices | |
RU2298838C2 (en) | Advertising device | |
Lanier | Maya Studio Projects Texturing and Lighting | |
Cory et al. | 3D Computer Animated Walkthroughs for Architecture, Engineering, and Construction Applications | |
Goss | Creative Content | |
Arbáizar Rovirosa | Video Mapping | |
Stork et al. | Reflections on Parmigianino's self portrait in a convex mirror: a computer graphics reconstruction of the artist's studio | |
Duarte Ferreira | Beyond Perception: an interaction between graphic design and anamorphosis | |
TW202312081A (en) | Object presentation device with label mainly including an object database, a plan view database, an interactive module, and an operation module |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: STACK'S-BOWERS NUMISMATICS, LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:STENDEBACH, LARRY J.;REEL/FRAME:057080/0217 Effective date: 20210802 |
|
AS | Assignment |
Owner name: CIBC BANK USA, COLORADO Free format text: SECURITY INTEREST;ASSIGNOR:STACKS-BOWERS NUMISMATICS, LLC;REEL/FRAME:058012/0411 Effective date: 20211103 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |