US20150224716A1 - Method and apparatus for embedding a 2-dimensional image in a 3-dimensional model - Google Patents
Method and apparatus for embedding a 2-dimensional image in a 3-dimensional model Download PDFInfo
- Publication number
- US20150224716A1 US20150224716A1 US14/177,933 US201414177933A US2015224716A1 US 20150224716 A1 US20150224716 A1 US 20150224716A1 US 201414177933 A US201414177933 A US 201414177933A US 2015224716 A1 US2015224716 A1 US 2015224716A1
- Authority
- US
- United States
- Prior art keywords
- image
- print matrix
- model
- sub
- embedding
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B29—WORKING OF PLASTICS; WORKING OF SUBSTANCES IN A PLASTIC STATE IN GENERAL
- B29C—SHAPING OR JOINING OF PLASTICS; SHAPING OF MATERIAL IN A PLASTIC STATE, NOT OTHERWISE PROVIDED FOR; AFTER-TREATMENT OF THE SHAPED PRODUCTS, e.g. REPAIRING
- B29C64/00—Additive manufacturing, i.e. manufacturing of three-dimensional [3D] objects by additive deposition, additive agglomeration or additive layering, e.g. by 3D printing, stereolithography or selective laser sintering
- B29C64/30—Auxiliary operations or equipment
- B29C64/386—Data acquisition or data processing for additive manufacturing
-
- B29C67/0088—
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B15/00—Systems controlled by a computer
- G05B15/02—Systems controlled by a computer electric
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/20—Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B33—ADDITIVE MANUFACTURING TECHNOLOGY
- B33Y—ADDITIVE MANUFACTURING, i.e. MANUFACTURING OF THREE-DIMENSIONAL [3-D] OBJECTS BY ADDITIVE DEPOSITION, ADDITIVE AGGLOMERATION OR ADDITIVE LAYERING, e.g. BY 3-D PRINTING, STEREOLITHOGRAPHY OR SELECTIVE LASER SINTERING
- B33Y10/00—Processes of additive manufacturing
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B33—ADDITIVE MANUFACTURING TECHNOLOGY
- B33Y—ADDITIVE MANUFACTURING, i.e. MANUFACTURING OF THREE-DIMENSIONAL [3-D] OBJECTS BY ADDITIVE DEPOSITION, ADDITIVE AGGLOMERATION OR ADDITIVE LAYERING, e.g. BY 3-D PRINTING, STEREOLITHOGRAPHY OR SELECTIVE LASER SINTERING
- B33Y30/00—Apparatus for additive manufacturing; Details thereof or accessories therefor
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B33—ADDITIVE MANUFACTURING TECHNOLOGY
- B33Y—ADDITIVE MANUFACTURING, i.e. MANUFACTURING OF THREE-DIMENSIONAL [3-D] OBJECTS BY ADDITIVE DEPOSITION, ADDITIVE AGGLOMERATION OR ADDITIVE LAYERING, e.g. BY 3-D PRINTING, STEREOLITHOGRAPHY OR SELECTIVE LASER SINTERING
- B33Y50/00—Data acquisition or data processing for additive manufacturing
- B33Y50/02—Data acquisition or data processing for additive manufacturing for controlling or regulating additive manufacturing processes
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/49—Nc machine tool, till multiple
- G05B2219/49023—3-D printing, layer of powder, add drops of binder in layer, new powder
Definitions
- Embodiments of the present invention generally relate to 3-dimensional (3D) printing and, more particularly, to a method and apparatus for embedding a 2-dimensional (2D) image in a 3D model using a single monochrome material.
- a 2D image such as a photo
- a need for printing the 2D image on a 3D image has risen.
- the 2D image is printed on a 3D printer with a material that may be of a different color than the color of the 3D printing material.
- the 2D image is sculpted, not allowing for smooth shading of the 2D image.
- the 2D image may be engraved onto the 3D object after the 3D object is printed.
- both approaches provide an undesirable result to the user who wants to combine the 2D image with the 3D model.
- a method for embedding a 2D image in a 3D model is described.
- the method generates a 3-dimensional (3D) print matrix representing a 2-dimensional (2D) image, wherein the print matrix comprises a plurality of sub-regions, the base plane of each sub-region angled so as to produce a plurality of shades, where each shade represents a shade of the 2D image.
- the method then embeds the print matrix in a (3D) model.
- an apparatus for embedding a 2D image in a 3D model includes a print matrix generator for generating a 3-dimensional (3D) print matrix representing a 2-dimensional (2D) image, wherein the print matrix comprises a plurality of sub-regions, the base plane of each sub-region angled so as to produce a plurality of shades, each shade representing a shade of the 2D image.
- the apparatus also includes an embedding module for embedding the print matrix in a (3D) model.
- a computer readable medium for embedding a 2D image in a 3D model includes instructions to perform the method for embedding a 3D image in a 3D model.
- FIG. 1 is a block diagram of an apparatus for embedding a 2D image in a 3D model, according to one or more embodiments
- FIG. 2 depicts a flow diagram of a method for embedding a 2D image in a 3D model as performed by the image processor, print matrix generator, and embedding module of FIG. 1 , according to one or more embodiments;
- FIG. 3 depicts a flow diagram of a method for generating a print matrix as performed by the print matrix generator of FIG. 1 , according to one or more embodiments.
- FIG. 4 illustrates the processing of the 2D image into the print matrix, according to one or more embodiments.
- the word “may” is used in a permissive sense (i.e., meaning having the potential to), rather than the mandatory sense (i.e., meaning must).
- the words “include”, “including”, and “includes” mean including, but not limited to.
- a 2D image may be any digital image from any source, for example, a scanned photo, or a downloaded image from a user's personal images.
- the 2D image is made up of pixels of many different colors. Due to the fact that 3D printing material is monochromatic, the 2D image is processed in order to convert it to a monochromatic 2D image in a way that preserves significant details of the 2D image while removing extreme details. Significant details are details that are meaningful details to the appearance of the image. Extreme details are details that are not essential to represent the overall appearance of the image.
- the 2D image is made monochromatic by applying a grayscale process to the 2D image.
- Applying the grayscale process converts the image to a black-and-white image that is composed exclusively of a plethora of shades of gray, varying in intensity from black to white.
- 3D printing material due to its reflective properties may only be able to make visible, for example, four different shades of the gray.
- further processing of the grayscale 2D image is required to convert it to a 2-bit image, i.e., an image made up of only four shades of gray.
- the embodiments process the grayscale 2D image to reduce noise in the 2D image.
- Noise in the grayscale image is the plethora of shades of gray in the image.
- a technique for example, Gaussian blurring, may be applied to the grayscale 2D image.
- the grayscale 2D image is reduced to four shades of gray, for example using a technique such as dithering. Dithering reproduces the 2D image using four shades of gray that are required to reproduce the 2D image using the 3D material.
- the dithering process produces a 2D image with a distribution of black pixels in varying density to make the image appear as though there are intermediate colors.
- the result is a 2-bit image (i.e., an image made up of four colors).
- the 2-bit image is associated with four shades of gray.
- the embodiments use the reflective properties of the 3D printing material at different angles to simulate the four shades of gray.
- a print matrix is created that represents the 2D image.
- the print matrix is a 2D rectangle divided into sub-regions. Each sub-region represents one or more pixels of the 2-bit image.
- the base plane of each sub-region is printed at one of four different angles. Each angle in the base plane of sub-regions of a print matrix represents one of the four shades of gray. For example, if each sub-region represents one pixel of the 2-bit image, the shade of gray of the pixel in the 2-bit image is translated into one of the four angles in the print matrix.
- the embodiments create a print matrix that represents the 2D image.
- the reflective properties of the 3D printing material at the four different angles produces the details of the 2D image.
- the print matrix containing a representation of the 2D image is embedded on a surface of the 3D model, the 3D model is ready for printing.
- the embodiments described herein can be employed to allow users to print 2D images on 3D models using monochrome material.
- the reflective properties of the material produce shades of color that are used to reproduce the 2D image.
- such quantities may take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared or otherwise manipulated. It has proven convenient at times, principally for reasons of common usage, to refer to such signals as bits, data, values, elements, symbols, characters, terms, numbers, numerals or the like. It should be understood, however, that all of these or similar terms are to be associated with appropriate physical quantities and are merely convenient labels. Unless specifically stated otherwise, as apparent from the following discussion, it is appreciated that throughout this specification discussions utilizing terms such as “processing,” “computing,” “calculating,” “determining” or the like refer to actions or processes of a specific apparatus, such as a special purpose computer or a similar special purpose electronic computing device.
- a special purpose computer or a similar special purpose electronic computing device is capable of manipulating or transforming signals, typically represented as physical electronic or magnetic quantities within memories, registers, or other information storage devices, transmission devices, or display devices of the special purpose computer or similar special purpose electronic computing device.
- FIG. 1 is a block diagram of an apparatus 100 for embedding a 2D image in a 3D model, according to one or more embodiments.
- the apparatus 100 includes a computer 102 .
- the computer 102 is a computing device, for example a desktop computer, laptop, tablet computer, and the like.
- the computer 102 includes a Central Processing Unit (CPU) 104 , support circuits 106 , and a memory 108 .
- the computer 102 may be connected to a 3D printer 126 .
- the CPU 104 may include one or more commercially available microprocessors or microcontrollers that facilitate data processing and storage.
- the various support circuits 106 facilitate the operation of the CPU 104 and include one or more clock circuits, power supplies, cache, input/output circuits, and the like.
- the memory 108 includes at least one of Read Only Memory (ROM), Random Access Memory (RAM), disk drive storage, optical storage, removable storage and/or the like.
- ROM Read Only Memory
- RAM Random Access Memory
- the memory 108 includes an operating system 110 , an image processor 112 , a print matrix generator 114 , an embedding module 116 , a 2D image 118 , a 3D model 120 , a processed 2D image 122 , and a print matrix 124 .
- the operating system 110 may include various commercially known operating systems.
- the 2D image 118 may be any digital image on the computer 102 .
- the 3D model 120 may be any file representing a 3D model 120 .
- the image processor 112 may be a software plug-in or extension to existing printer software or an Application Programming Interface (API) for a 3D printer 126 .
- the image processor 112 may be a plug-in for 3D model creation software tools.
- the image processor 112 accesses the 2D image 118 .
- the 2D image 118 may be made up of pixels of a plurality of colors.
- the 2D image 118 is processed in order to convert it to a monochromatic 2D image in a way that preserves significant details of the 2D image 118 while removing extreme details.
- the 2D image 118 is made monochromatic by applying a grayscale process to the 2D image 118 . Applying the grayscale process converts the image to a black-and-white image composed exclusively of shades of gray, varying in intensity from black to white.
- 3D printing material due to its reflective properties may only be able to make visible, for example, four different shades of the gray. As such, further processing of the grayscale 2D image is required to convert 2D image 118 to a 2-bit image, in other words, an image made up of only four shades of gray.
- the image processor 112 reduces noise in grayscale 2D image.
- Noise in the grayscale image is the varying shades of gray in the grayscale 2D image.
- the image processor 112 applies a technique, for example, Gaussian blurring, to the grayscale 2D image.
- the image processor 112 reduces the number of shades of gray in the grayscale 2D image using, for example a technique such as dithering. Dithering reproduces the 2D image using four shades of gray that are required to reproduce the 2D image 118 using the 3D material.
- Many 3D printers 126 use a 3D printing material that has reflexive qualities that are capable of showing only four different shades of gray.
- some 3D printers 126 use a 3D printing material that has reflexive qualities that are capable of showing eight different shades of gray.
- the type of 3D printer 126 is known at the time of processing the 2D image 118 .
- the number of different shades of gray produced by the dithering process is predefined.
- the present description describes a dithering process that reduces the grayscale 2D image to four shades of gray, it is appreciated by those of ordinary skill in the art that the dithering process may reduce the number of shades of gray based on the reflexive properties of the 3D printing material.
- the dithering process produces a 2D image using a distribution of black pixels in varying density to make the 2D image appear as though there are intermediate shades of gray.
- the result is a processed 2D image 122 .
- the processed 2D image 122 is a 2-bit image (i.e., an image made up of four colors).
- the processed 2D image 122 is associated with four shades of gray.
- the reflective properties of the 3D printing material at different angles can be used to simulate the four shades of gray in the processed 2D image 122 .
- the print matrix generator 114 creates the print matrix 124 that represents the processed 2D image 122 .
- the print matrix 124 is a 2D rectangle divided into sub-regions. Each sub-region represents one or more pixels of the processed 2D image 122 .
- the base plane of each sub-region is printed at one of four different angles. The four angles are predefined based on the reflexive properties of the 3D printing material, for example, 20, 27, 36, and 45 degrees. Each angle in the base plane of sub-regions of the print matrix 124 represents one of the four shades of gray in the processed 2D image 122 .
- each sub-region represents one pixel of the processed 2D image 122
- the shade of gray of the pixel in the processed 2D image 122 is translated into one of the four angles in the print matrix 124 .
- the print matrix generator 114 creates the print matrix 124 that represents the processed 2D image 122 .
- the reflective properties of the 3D printing material at the four different angles produces the details of the processed 2D image 122 .
- the embedding module 116 determines an area on the surface of the 3D model 120 .
- the embedding module 116 identifies a surface on the 3D model 120 that has at least the volume of the print matrix.
- the surface on the 3D model 120 must be at least of the size of the print matrix and the depth of the print matrix.
- the print matrix 124 for the 2D image 118 may be 5 cm by 5 cm by IA cm deep.
- the surface on the 3D model 120 must have at least the same dimension or larger.
- the area on the surface of the 3D model 120 is selected by a user via a user interface (not shown).
- the user rotates the view of the 3D model 120 to an orientation that shows the area of the 3D model 120 where the user would like the 2D image 118 embedded.
- the user draws, for example, a rectangle on the surface of the 3D model 120 to select the area where the user would like to have the 2D model 118 embedded.
- the volume of the 3D print matrix then replaces a volume at the area of the 3D model 120 with the volume of the 3D print matrix.
- the 3D model 120 is then ready to be printed using any method for printing on the 3D printer 126 .
- the 3D model 120 is printed with the processed 2D image 122 embedded in the surface of the 3D model 120 .
- FIG. 2 depicts a flow diagram of a method 200 for embedding a 2D image in a 3D model as performed by the image processor 112 , print matrix generator 114 , and embedding module 116 of FIG. 1 , according to one or more embodiments.
- the method 200 generates a print matrix that represents the 2D image and embeds the print matrix into a 3D model.
- the method 200 starts at step 202 and proceeds to step 204 .
- the method 200 generates a 3D print matrix representing the 2D image as described in further detail with respect to FIG. 3 below.
- the method 200 then proceeds to step 206 , where the method 200 embeds the 3D print matrix into the 3D model.
- the method 200 finds a surface on the surface of the 3D model that is large enough to hold the print matrix.
- the print matrix may be 5 cm ⁇ 5 cm and 1 ⁇ 2 cm thick. If a large enough surface does not exist on the surface of the 3D model, the method 200 downsizes the print matrix to fit an available area on the surface of the 3D model.
- the area on the surface of the 3D model is selected by a user via a user interface. In such embodiments, the method 200 displays the 3D model.
- the user rotates the view of the 3D model to an orientation that shows the area of the 3D model where the user would like the 2D image embedded.
- the user then draws, for example, a rectangle on the surface of the 3D model to select the area where the user would like to have the 2D model embedded.
- the method 200 uses the selected area to hold the print matrix.
- the method 200 then subtracts a cuboid of the dimensions of the print matrix from the 3D model.
- the method 200 replaces the subtracted cuboid by performing a union of the print matrix with the 3D model.
- the result is a 3D print matrix representative of a 2D image embedded in the surface of a 3D model.
- the method 200 proceeds to step 208 where the method 200 ends.
- FIG. 3 depicts a flow diagram of a method 300 for generating a print matrix as performed by the print matrix generator of FIG. 1 , according to one or more embodiments.
- the method 300 processes a 2D image and generates a print matrix that represents the 2D image.
- the method 300 starts at step 302 and proceeds to step 304 .
- the method 300 accesses the 2D image.
- the 2D image may be any digital image that includes color information for each pixel.
- the method 300 proceeds to step 306 , where the method 300 generates a grayscale image of the 2D image.
- the grayscale digital image identifies an intensity value for each pixel in the 2D image.
- the grayscale image is composed exclusively of shades of gray, varying from black at the weakest intensity to white at the strongest intensity.
- the method 300 proceeds to step 308 , where the method 300 removes noise from the grayscale image.
- the method 300 removes extreme details of the grayscale image through, for example, Gaussian blurring.
- the method 300 may perform selective blurring of low-gradient regions of the grayscale image.
- the method 300 proceeds to step 310 , where the method 300 converts the grayscale image to an indexed color image.
- the method 300 uses an error diffusion technique, such as dithering, to generate an indexed color image.
- the number of colors that are created by the error diffusion process is pre-determined based on the reflective properties of the printing material.
- the method 300 may generate a 2-bit image for a printing material that, due to its reflective qualities, is able to show four distinct shades of gray.
- the method 300 may generate a 3-bit image for a printing material that, due to its reflective properties, is able to show eight distinct shades of gray.
- the method 300 proceeds to step 312 , where the method 300 creates a 3D print matrix from the indexed color image.
- the method 300 creates a 3D print matrix that has a surface area the size of the 2D image.
- the method 300 then creates sub-regions in the print matrix. Each sub-region has its base plane raised at an angle. Due to the reflective properties of the printing material, by angling the base plane of a sub-region, the sub-region produces a shade.
- the angles are pre-defined based on the printing material. For example, for a 2-bit color image, four distinct angles are used to produce four distinct shades.
- the base planes of the sub-regions may be printed at angles of 20, 27, 36, and 45 degrees with respect to a top surface of the print matrix to produce four distinct shades.
- Each sub-region is mapped to a pixel in the indexed color image and each shade is mapped to an angle.
- the method 200 creates the print matrix by defining an angle for each sub-region of the print matrix relative to the top surface of the print matrix. The result is a print matrix that reproduces the 2D image.
- the method 300 proceeds to step 314 and ends.
- FIG. 4 illustrates the stages 400 of the processing of the 2D image into the print matrix, according to one or more embodiments.
- the input image 402 is a grayscale image. However, if the input image 402 was in color, the input image 402 is converted to grayscale. Noise removal is performed on the grayscale image 402 to remove extreme color values. The noise removal process produces a simplified image 404 . Error diffusion is performed on the simplified image 404 . The error diffusion process produces an indexed color image 406 . In this illustrated example, the indexed color image 406 is a 2-bit color image. The four shades of gray in the 2-bit color image are associated with the angles of 20, 27, 36, and 45 degrees.
- a print matrix 408 is created by mapping each shade of gray to an angle.
- a sub-region is created for each pixel of the 2-bit color image 406 .
- the base plane of each sub-region is angled with respect to the top surface of the print matrix 408 at an angle that corresponds to the shade of the corresponding pixel in the indexed color image 406 .
- the result is the print matrix 408 that represents the 2D image.
- a sample area 410 of the print matrix 408 illustrates the effects of angling the base planes of the print matrix 408 .
- Each sub-region 412 , 414 , 416 , 418 has a base plane angled to produce a shade of an indexed color image.
- the sub-region 412 has a base plane angled at 20 degrees.
- the sub-region 414 has a base plane angled at 27 degrees.
- the sub-region 416 has a base plane angled at 36 degrees.
- the sub-region 418 has a base plane angled at 45 degrees. The varying degrees result in different shades.
- a single sub-region 420 is representative of a plurality of sub-regions that make up the print matrix 408 .
- the base plane 422 of the sub-region 420 is angled at an angle 424 relative to the top surface of the print matrix 408 so as to produce a respective one of a plurality of shades of the indexed color image 406 .
- the 3D model 426 is merely a box that may be printed on a 3D printer.
- the print matrix 408 is embedded in the surface of the 3D model 426 producing a 3D model with an embedded 2D image 428 .
- the embodiments of the present invention may be embodied as methods, apparatus, electronic devices, and/or computer program products. Accordingly, the embodiments of the present invention may be embodied in hardware and/or in software (including firmware, resident software, micro-code, etc.), which may be generally referred to herein as a “circuit” or “module”. Furthermore, the present invention may take the form of a computer program product on a computer-usable or computer-readable storage medium having computer-usable or computer-readable program code embodied in the medium for use by or in connection with an instruction execution system.
- a computer-usable or computer-readable medium may be any medium that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
- These computer program instructions may also be stored in a computer-usable or computer-readable memory that may direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer usable or computer-readable memory produce an article of manufacture including instructions that implement the function specified in the flowchart and/or block diagram block or blocks.
- the computer-usable or computer-readable medium may be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium. More specific examples (a non-exhaustive list) of the computer-readable medium include the following: hard disks, optical storage devices, a transmission media such as those supporting the Internet or an intranet, magnetic storage devices, an electrical connection having one or more wires, a portable computer diskette, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, and a compact disc read-only memory (CD-ROM).
- RAM random access memory
- ROM read-only memory
- EPROM or Flash memory erasable programmable read-only memory
- CD-ROM compact disc read-only memory
- Computer program code for carrying out operations of the present invention may be written in an object oriented programming language, such as Java®, Smalltalk or C++, and the like. However, the computer program code for carrying out operations of the present invention may also be written in conventional procedural programming languages, such as the “C” programming language and/or any other lower level assembler languages. It will be further appreciated that the functionality of any or all of the program modules may also be implemented using discrete hardware components, one or more Application Specific Integrated Circuits (ASICs), or programmed Digital Signal Processors or microcontrollers.
- ASICs Application Specific Integrated Circuits
- microcontrollers programmed Digital Signal Processors or microcontrollers.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Chemical & Material Sciences (AREA)
- Materials Engineering (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Manufacturing & Machinery (AREA)
- Mechanical Engineering (AREA)
- Optics & Photonics (AREA)
- General Engineering & Computer Science (AREA)
- Computer Hardware Design (AREA)
- Software Systems (AREA)
- Computer Graphics (AREA)
- Architecture (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Processing (AREA)
- Automation & Control Theory (AREA)
Abstract
Description
- 1. Field of the Invention
- Embodiments of the present invention generally relate to 3-dimensional (3D) printing and, more particularly, to a method and apparatus for embedding a 2-dimensional (2D) image in a 3D model using a single monochrome material.
- 2. Description of the Related Art
- Popularity of 3D printers, either for home printing use or via a remote service, has increased in recent times. Traditionally, a 2D image, such as a photo, is printed using a 2D printer. With the increase in popularity of 3D printers and 3D objects, a need for printing the 2D image on a 3D image has risen. In one approach, the 2D image is printed on a 3D printer with a material that may be of a different color than the color of the 3D printing material. In such approach, the 2D image is sculpted, not allowing for smooth shading of the 2D image. In another approach, the 2D image may be engraved onto the 3D object after the 3D object is printed. However, both approaches provide an undesirable result to the user who wants to combine the 2D image with the 3D model.
- Therefore, there is a need for a method and apparatus for embedding a 2D image in a 3D model using a single monochrome material.
- A method for embedding a 2D image in a 3D model is described. The method generates a 3-dimensional (3D) print matrix representing a 2-dimensional (2D) image, wherein the print matrix comprises a plurality of sub-regions, the base plane of each sub-region angled so as to produce a plurality of shades, where each shade represents a shade of the 2D image. The method then embeds the print matrix in a (3D) model.
- In another embodiment, an apparatus for embedding a 2D image in a 3D model is described. The apparatus includes a print matrix generator for generating a 3-dimensional (3D) print matrix representing a 2-dimensional (2D) image, wherein the print matrix comprises a plurality of sub-regions, the base plane of each sub-region angled so as to produce a plurality of shades, each shade representing a shade of the 2D image. The apparatus also includes an embedding module for embedding the print matrix in a (3D) model.
- In yet another embodiment, a computer readable medium for embedding a 2D image in a 3D model is described. The computer readable medium includes instructions to perform the method for embedding a 3D image in a 3D model.
- The Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
-
FIG. 1 is a block diagram of an apparatus for embedding a 2D image in a 3D model, according to one or more embodiments; -
FIG. 2 depicts a flow diagram of a method for embedding a 2D image in a 3D model as performed by the image processor, print matrix generator, and embedding module ofFIG. 1 , according to one or more embodiments; -
FIG. 3 depicts a flow diagram of a method for generating a print matrix as performed by the print matrix generator ofFIG. 1 , according to one or more embodiments; and -
FIG. 4 illustrates the processing of the 2D image into the print matrix, according to one or more embodiments. - While the method and apparatus is described herein by way of example for several embodiments and illustrative drawings, those skilled in the art will recognize that the method and apparatus for embedding a 2D image in a 3D model is not limited to the embodiments or drawings described. It should be understood, that the drawings and detailed description thereto are not intended to limit embodiments to the particular form disclosed. Rather, the intention is to cover all modifications, equivalents and alternatives falling within the spirit and scope of the method and apparatus for embedding a 2D image in a 3D model defined by the appended claims. Any headings used herein are for organizational purposes only and are not meant to limit the scope of the description or the claims. As used herein, the word “may” is used in a permissive sense (i.e., meaning having the potential to), rather than the mandatory sense (i.e., meaning must). Similarly, the words “include”, “including”, and “includes” mean including, but not limited to.
- As previously explained existing solutions provide undesirable results, such as a lack of smooth shading of a 2D image on a 3D model or create the 2D image by processing the image after the 3D printing is complete.
- Thus, in accordance with an embodiment of the present invention, techniques are disclosed for embedding a 2D image in a 3D model. A 2D image may be any digital image from any source, for example, a scanned photo, or a downloaded image from a user's personal images. The 2D image is made up of pixels of many different colors. Due to the fact that 3D printing material is monochromatic, the 2D image is processed in order to convert it to a monochromatic 2D image in a way that preserves significant details of the 2D image while removing extreme details. Significant details are details that are meaningful details to the appearance of the image. Extreme details are details that are not essential to represent the overall appearance of the image. More specifically, the 2D image is made monochromatic by applying a grayscale process to the 2D image. Applying the grayscale process converts the image to a black-and-white image that is composed exclusively of a plethora of shades of gray, varying in intensity from black to white. However, 3D printing material, due to its reflective properties may only be able to make visible, for example, four different shades of the gray. As such, further processing of the
grayscale 2D image is required to convert it to a 2-bit image, i.e., an image made up of only four shades of gray. - The embodiments process the
grayscale 2D image to reduce noise in the 2D image. Noise in the grayscale image is the plethora of shades of gray in the image. In order to reduce the number of shades of gray (i.e., remove the noise) and also remove extreme details, a technique, for example, Gaussian blurring, may be applied to thegrayscale 2D image. After the noise removal, thegrayscale 2D image is reduced to four shades of gray, for example using a technique such as dithering. Dithering reproduces the 2D image using four shades of gray that are required to reproduce the 2D image using the 3D material. The dithering process produces a 2D image with a distribution of black pixels in varying density to make the image appear as though there are intermediate colors. The result is a 2-bit image (i.e., an image made up of four colors). The 2-bit image is associated with four shades of gray. The embodiments use the reflective properties of the 3D printing material at different angles to simulate the four shades of gray. A print matrix is created that represents the 2D image. The print matrix is a 2D rectangle divided into sub-regions. Each sub-region represents one or more pixels of the 2-bit image. The base plane of each sub-region is printed at one of four different angles. Each angle in the base plane of sub-regions of a print matrix represents one of the four shades of gray. For example, if each sub-region represents one pixel of the 2-bit image, the shade of gray of the pixel in the 2-bit image is translated into one of the four angles in the print matrix. The embodiments create a print matrix that represents the 2D image. In other words, when the 2D image is created using 3D printing material, the reflective properties of the 3D printing material at the four different angles produces the details of the 2D image. After the print matrix containing a representation of the 2D image is embedded on a surface of the 3D model, the 3D model is ready for printing. - Advantageously, the embodiments described herein can be employed to allow users to print 2D images on 3D models using monochrome material. The reflective properties of the material produce shades of color that are used to reproduce the 2D image.
- Various embodiments of a method and apparatus for embedding a 2D image in a 3D model are described. In the following detailed description, numerous specific details are set forth to provide a thorough understanding of claimed subject matter. However, it will be understood by those skilled in the art that claimed subject matter may be practiced without these specific details. In other instances, methods, apparatuses or systems that would be known by one of ordinary skill have not been described in detail so as not to obscure claimed subject matter.
- Some portions of the detailed description that follow are presented in terms of algorithms or symbolic representations of operations on binary digital signals stored within a memory of a specific apparatus or special purpose computing device or platform. In the context of this particular specification, the term specific apparatus or the like includes a general-purpose computer once it is programmed to perform particular functions pursuant to instructions from program software. Algorithmic descriptions or symbolic representations are examples of techniques used by those of ordinary skill in the signal processing or related arts to convey the substance of their work to others skilled in the art. An algorithm is here, and is generally, considered to be a self-consistent sequence of operations or similar signal processing leading to a desired result. In this context, operations or processing involve physical manipulation of physical quantities. Typically, although not necessarily, such quantities may take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared or otherwise manipulated. It has proven convenient at times, principally for reasons of common usage, to refer to such signals as bits, data, values, elements, symbols, characters, terms, numbers, numerals or the like. It should be understood, however, that all of these or similar terms are to be associated with appropriate physical quantities and are merely convenient labels. Unless specifically stated otherwise, as apparent from the following discussion, it is appreciated that throughout this specification discussions utilizing terms such as “processing,” “computing,” “calculating,” “determining” or the like refer to actions or processes of a specific apparatus, such as a special purpose computer or a similar special purpose electronic computing device. In the context of this specification, therefore, a special purpose computer or a similar special purpose electronic computing device is capable of manipulating or transforming signals, typically represented as physical electronic or magnetic quantities within memories, registers, or other information storage devices, transmission devices, or display devices of the special purpose computer or similar special purpose electronic computing device.
-
FIG. 1 is a block diagram of anapparatus 100 for embedding a 2D image in a 3D model, according to one or more embodiments. Theapparatus 100 includes acomputer 102. Thecomputer 102 is a computing device, for example a desktop computer, laptop, tablet computer, and the like. Thecomputer 102 includes a Central Processing Unit (CPU) 104,support circuits 106, and amemory 108. Thecomputer 102 may be connected to a3D printer 126. TheCPU 104 may include one or more commercially available microprocessors or microcontrollers that facilitate data processing and storage. Thevarious support circuits 106 facilitate the operation of theCPU 104 and include one or more clock circuits, power supplies, cache, input/output circuits, and the like. Thememory 108 includes at least one of Read Only Memory (ROM), Random Access Memory (RAM), disk drive storage, optical storage, removable storage and/or the like. - The
memory 108 includes anoperating system 110, animage processor 112, aprint matrix generator 114, an embeddingmodule 116, a2D image 118, a3D model 120, a processed2D image 122, and aprint matrix 124. Theoperating system 110 may include various commercially known operating systems. - When a user wishes to print a
3D model 120 to include a2D image 118, the user submits the3D model 120 and the2D image 118 to theimage processor 112. The2D image 118 may be any digital image on thecomputer 102. The3D model 120 may be any file representing a3D model 120. Theimage processor 112 may be a software plug-in or extension to existing printer software or an Application Programming Interface (API) for a3D printer 126. Alternatively, theimage processor 112 may be a plug-in for 3D model creation software tools. Theimage processor 112 accesses the2D image 118. The2D image 118 may be made up of pixels of a plurality of colors. Due to the fact that 3D printing material is monochromatic, the2D image 118 is processed in order to convert it to a monochromatic 2D image in a way that preserves significant details of the2D image 118 while removing extreme details. The2D image 118 is made monochromatic by applying a grayscale process to the2D image 118. Applying the grayscale process converts the image to a black-and-white image composed exclusively of shades of gray, varying in intensity from black to white. However, 3D printing material, due to its reflective properties may only be able to make visible, for example, four different shades of the gray. As such, further processing of the grayscale 2D image is required to convert2D image 118 to a 2-bit image, in other words, an image made up of only four shades of gray. - The
image processor 112 reduces noise ingrayscale 2D image. Noise in the grayscale image is the varying shades of gray in thegrayscale 2D image. In order to reduce the number of shades of gray (i.e., remove the noise) and also remove details, theimage processor 112 applies a technique, for example, Gaussian blurring, to thegrayscale 2D image. After the noise removal, theimage processor 112 reduces the number of shades of gray in thegrayscale 2D image using, for example a technique such as dithering. Dithering reproduces the 2D image using four shades of gray that are required to reproduce the2D image 118 using the 3D material.Many 3D printers 126 use a 3D printing material that has reflexive qualities that are capable of showing only four different shades of gray. However, some3D printers 126 use a 3D printing material that has reflexive qualities that are capable of showing eight different shades of gray. The type of3D printer 126 is known at the time of processing the2D image 118. As such, the number of different shades of gray produced by the dithering process is predefined. Although the present description describes a dithering process that reduces the grayscale 2D image to four shades of gray, it is appreciated by those of ordinary skill in the art that the dithering process may reduce the number of shades of gray based on the reflexive properties of the 3D printing material. The dithering process produces a 2D image using a distribution of black pixels in varying density to make the 2D image appear as though there are intermediate shades of gray. The result is a processed2D image 122. The processed2D image 122 is a 2-bit image (i.e., an image made up of four colors). The processed2D image 122 is associated with four shades of gray. The reflective properties of the 3D printing material at different angles can be used to simulate the four shades of gray in the processed2D image 122. - The
print matrix generator 114 creates theprint matrix 124 that represents the processed2D image 122. Theprint matrix 124 is a 2D rectangle divided into sub-regions. Each sub-region represents one or more pixels of the processed2D image 122. The base plane of each sub-region is printed at one of four different angles. The four angles are predefined based on the reflexive properties of the 3D printing material, for example, 20, 27, 36, and 45 degrees. Each angle in the base plane of sub-regions of theprint matrix 124 represents one of the four shades of gray in the processed2D image 122. For example, if each sub-region represents one pixel of the processed2D image 122, the shade of gray of the pixel in the processed2D image 122 is translated into one of the four angles in theprint matrix 124. Theprint matrix generator 114 creates theprint matrix 124 that represents the processed2D image 122. In other words, when the processed2D image 122 is created using 3D printing material, the reflective properties of the 3D printing material at the four different angles produces the details of the processed2D image 122. - When the
print matrix 124 is complete, the embeddingmodule 116 determines an area on the surface of the3D model 120. The embeddingmodule 116 identifies a surface on the3D model 120 that has at least the volume of the print matrix. Specifically, the surface on the3D model 120 must be at least of the size of the print matrix and the depth of the print matrix. For example, theprint matrix 124 for the2D image 118 may be 5 cm by 5 cm by IA cm deep. The surface on the3D model 120 must have at least the same dimension or larger. In some embodiments, the area on the surface of the3D model 120 is selected by a user via a user interface (not shown). In such embodiments, the user rotates the view of the3D model 120 to an orientation that shows the area of the3D model 120 where the user would like the2D image 118 embedded. The user then draws, for example, a rectangle on the surface of the3D model 120 to select the area where the user would like to have the2D model 118 embedded. The volume of the 3D print matrix then replaces a volume at the area of the3D model 120 with the volume of the 3D print matrix. The3D model 120 is then ready to be printed using any method for printing on the3D printer 126. The3D model 120 is printed with the processed2D image 122 embedded in the surface of the3D model 120. -
FIG. 2 depicts a flow diagram of amethod 200 for embedding a 2D image in a 3D model as performed by theimage processor 112,print matrix generator 114, and embeddingmodule 116 ofFIG. 1 , according to one or more embodiments. Themethod 200 generates a print matrix that represents the 2D image and embeds the print matrix into a 3D model. Themethod 200 starts atstep 202 and proceeds to step 204. - At
step 204, themethod 200 generates a 3D print matrix representing the 2D image as described in further detail with respect toFIG. 3 below. Themethod 200 then proceeds to step 206, where themethod 200 embeds the 3D print matrix into the 3D model. Themethod 200 finds a surface on the surface of the 3D model that is large enough to hold the print matrix. For example, the print matrix may be 5 cm×5 cm and ½ cm thick. If a large enough surface does not exist on the surface of the 3D model, themethod 200 downsizes the print matrix to fit an available area on the surface of the 3D model. In some embodiments, the area on the surface of the 3D model is selected by a user via a user interface. In such embodiments, themethod 200 displays the 3D model. The user rotates the view of the 3D model to an orientation that shows the area of the 3D model where the user would like the 2D image embedded. The user then draws, for example, a rectangle on the surface of the 3D model to select the area where the user would like to have the 2D model embedded. Themethod 200 uses the selected area to hold the print matrix. Themethod 200 then subtracts a cuboid of the dimensions of the print matrix from the 3D model. Themethod 200 replaces the subtracted cuboid by performing a union of the print matrix with the 3D model. The result is a 3D print matrix representative of a 2D image embedded in the surface of a 3D model. - The
method 200 proceeds to step 208 where themethod 200 ends. -
FIG. 3 depicts a flow diagram of amethod 300 for generating a print matrix as performed by the print matrix generator ofFIG. 1 , according to one or more embodiments. Themethod 300 processes a 2D image and generates a print matrix that represents the 2D image. Themethod 300 starts atstep 302 and proceeds to step 304. - At
step 304, themethod 300 accesses the 2D image. The 2D image may be any digital image that includes color information for each pixel. Themethod 300 proceeds to step 306, where themethod 300 generates a grayscale image of the 2D image. The grayscale digital image identifies an intensity value for each pixel in the 2D image. The grayscale image is composed exclusively of shades of gray, varying from black at the weakest intensity to white at the strongest intensity. - The
method 300 proceeds to step 308, where themethod 300 removes noise from the grayscale image. Themethod 300 removes extreme details of the grayscale image through, for example, Gaussian blurring. In order to ensure that primary images of the grayscale image are retained, themethod 300 may perform selective blurring of low-gradient regions of the grayscale image. - The
method 300 proceeds to step 310, where themethod 300 converts the grayscale image to an indexed color image. Themethod 300 uses an error diffusion technique, such as dithering, to generate an indexed color image. The number of colors that are created by the error diffusion process is pre-determined based on the reflective properties of the printing material. Themethod 300 may generate a 2-bit image for a printing material that, due to its reflective qualities, is able to show four distinct shades of gray. Themethod 300 may generate a 3-bit image for a printing material that, due to its reflective properties, is able to show eight distinct shades of gray. - The
method 300 proceeds to step 312, where themethod 300 creates a 3D print matrix from the indexed color image. Themethod 300 creates a 3D print matrix that has a surface area the size of the 2D image. Themethod 300 then creates sub-regions in the print matrix. Each sub-region has its base plane raised at an angle. Due to the reflective properties of the printing material, by angling the base plane of a sub-region, the sub-region produces a shade. The angles are pre-defined based on the printing material. For example, for a 2-bit color image, four distinct angles are used to produce four distinct shades. In some embodiments, the base planes of the sub-regions may be printed at angles of 20, 27, 36, and 45 degrees with respect to a top surface of the print matrix to produce four distinct shades. Each sub-region is mapped to a pixel in the indexed color image and each shade is mapped to an angle. Themethod 200 creates the print matrix by defining an angle for each sub-region of the print matrix relative to the top surface of the print matrix. The result is a print matrix that reproduces the 2D image. Themethod 300 proceeds to step 314 and ends. -
FIG. 4 illustrates thestages 400 of the processing of the 2D image into the print matrix, according to one or more embodiments. Theinput image 402 is a grayscale image. However, if theinput image 402 was in color, theinput image 402 is converted to grayscale. Noise removal is performed on thegrayscale image 402 to remove extreme color values. The noise removal process produces asimplified image 404. Error diffusion is performed on thesimplified image 404. The error diffusion process produces an indexedcolor image 406. In this illustrated example, the indexedcolor image 406 is a 2-bit color image. The four shades of gray in the 2-bit color image are associated with the angles of 20, 27, 36, and 45 degrees. Aprint matrix 408 is created by mapping each shade of gray to an angle. A sub-region is created for each pixel of the 2-bit color image 406. The base plane of each sub-region is angled with respect to the top surface of theprint matrix 408 at an angle that corresponds to the shade of the corresponding pixel in the indexedcolor image 406. The result is theprint matrix 408 that represents the 2D image. Asample area 410 of theprint matrix 408 illustrates the effects of angling the base planes of theprint matrix 408. Eachsub-region sub-region 412 has a base plane angled at 20 degrees. Thesub-region 414 has a base plane angled at 27 degrees. Thesub-region 416 has a base plane angled at 36 degrees. Thesub-region 418 has a base plane angled at 45 degrees. The varying degrees result in different shades. Asingle sub-region 420 is representative of a plurality of sub-regions that make up theprint matrix 408. Thebase plane 422 of thesub-region 420 is angled at anangle 424 relative to the top surface of theprint matrix 408 so as to produce a respective one of a plurality of shades of the indexedcolor image 406. The3D model 426 is merely a box that may be printed on a 3D printer. Theprint matrix 408 is embedded in the surface of the3D model 426 producing a 3D model with an embedded2D image 428. - The embodiments of the present invention may be embodied as methods, apparatus, electronic devices, and/or computer program products. Accordingly, the embodiments of the present invention may be embodied in hardware and/or in software (including firmware, resident software, micro-code, etc.), which may be generally referred to herein as a “circuit” or “module”. Furthermore, the present invention may take the form of a computer program product on a computer-usable or computer-readable storage medium having computer-usable or computer-readable program code embodied in the medium for use by or in connection with an instruction execution system. In the context of this document, a computer-usable or computer-readable medium may be any medium that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device. These computer program instructions may also be stored in a computer-usable or computer-readable memory that may direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer usable or computer-readable memory produce an article of manufacture including instructions that implement the function specified in the flowchart and/or block diagram block or blocks.
- The computer-usable or computer-readable medium may be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium. More specific examples (a non-exhaustive list) of the computer-readable medium include the following: hard disks, optical storage devices, a transmission media such as those supporting the Internet or an intranet, magnetic storage devices, an electrical connection having one or more wires, a portable computer diskette, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, and a compact disc read-only memory (CD-ROM).
- Computer program code for carrying out operations of the present invention may be written in an object oriented programming language, such as Java®, Smalltalk or C++, and the like. However, the computer program code for carrying out operations of the present invention may also be written in conventional procedural programming languages, such as the “C” programming language and/or any other lower level assembler languages. It will be further appreciated that the functionality of any or all of the program modules may also be implemented using discrete hardware components, one or more Application Specific Integrated Circuits (ASICs), or programmed Digital Signal Processors or microcontrollers.
- The foregoing description, for purpose of explanation, has been described with reference to specific embodiments. However, the illustrative discussions above are not intended to be exhaustive or to limit the invention to the precise forms disclosed. Many modifications and variations are possible in view of the above teachings. The embodiments were chosen and described in order to best explain the principles of the present disclosure and its practical applications, to thereby enable others skilled in the art to best utilize the invention and various embodiments with various modifications as may be suited to the particular use contemplated.
- The methods described herein may be implemented in software, hardware, or a combination thereof, in different embodiments. In addition, the order of methods may be changed, and various elements may be added, reordered, combined, omitted, modified, etc. All examples described herein are presented in a non-limiting manner. Various modifications and changes may be made as would be obvious to a person skilled in the art having benefit of this disclosure. Realizations in accordance with embodiments have been described in the context of particular embodiments. These embodiments are meant to be illustrative and not limiting. Many variations, modifications, additions, and improvements are possible. Accordingly, plural instances may be provided for components described herein as a single instance. Boundaries between various components, operations and data stores are somewhat arbitrary, and particular operations are illustrated in the context of specific illustrative configurations. Other allocations of functionality are envisioned and may fall within the scope of claims that follow. Finally, structures and functionality presented as discrete components in the example configurations may be implemented as a combined structure or component. These and other variations, modifications, additions, and improvements may fall within the scope of embodiments as defined in the claims that follow.
- While the foregoing is directed to embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow.
Claims (20)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/177,933 US9669585B2 (en) | 2014-02-11 | 2014-02-11 | Method and apparatus for embedding a 2-dimensional image in a 3-dimensional model |
US15/498,324 US10373394B2 (en) | 2014-02-11 | 2017-04-26 | Method and apparatus for embedding a 2-dimensional image in a 3-dimensional model |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/177,933 US9669585B2 (en) | 2014-02-11 | 2014-02-11 | Method and apparatus for embedding a 2-dimensional image in a 3-dimensional model |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/498,324 Continuation US10373394B2 (en) | 2014-02-11 | 2017-04-26 | Method and apparatus for embedding a 2-dimensional image in a 3-dimensional model |
Publications (2)
Publication Number | Publication Date |
---|---|
US20150224716A1 true US20150224716A1 (en) | 2015-08-13 |
US9669585B2 US9669585B2 (en) | 2017-06-06 |
Family
ID=53774162
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/177,933 Active 2035-08-11 US9669585B2 (en) | 2014-02-11 | 2014-02-11 | Method and apparatus for embedding a 2-dimensional image in a 3-dimensional model |
US15/498,324 Active 2034-06-11 US10373394B2 (en) | 2014-02-11 | 2017-04-26 | Method and apparatus for embedding a 2-dimensional image in a 3-dimensional model |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/498,324 Active 2034-06-11 US10373394B2 (en) | 2014-02-11 | 2017-04-26 | Method and apparatus for embedding a 2-dimensional image in a 3-dimensional model |
Country Status (1)
Country | Link |
---|---|
US (2) | US9669585B2 (en) |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150317412A1 (en) * | 2014-05-05 | 2015-11-05 | Microsoft Corporation | Fabricating three-dimensional objects with embossing |
CN105599296A (en) * | 2015-12-23 | 2016-05-25 | 上海理工大学 | 2.5D image printing method |
US20180144219A1 (en) * | 2016-11-23 | 2018-05-24 | Simbionix Ltd. | Method and system for three-dimensional print oriented image segmentation |
US10338568B2 (en) * | 2015-04-24 | 2019-07-02 | Hewlett-Packard Development Company, L.P. | Determining halftone schemes |
US10346100B2 (en) * | 2016-01-29 | 2019-07-09 | Hewlett-Packard Development Company, L.P. | Error diffusion |
US10373394B2 (en) | 2014-02-11 | 2019-08-06 | Adobe Inc. | Method and apparatus for embedding a 2-dimensional image in a 3-dimensional model |
US10409264B2 (en) * | 2016-10-27 | 2019-09-10 | Voodoo Manufacturing, Inc. | Fabrication of three-dimensional part with two-dimensional image applied thereon |
CN112848281A (en) * | 2020-12-31 | 2021-05-28 | 浙江闪铸三维科技有限公司 | Light compensation method for photocuring 3D printer |
CN115723333A (en) * | 2021-08-30 | 2023-03-03 | 广州黑格智造信息科技有限公司 | 3D printing method, system, device and storage medium |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170113412A1 (en) * | 2015-10-22 | 2017-04-27 | Adamation, LLC | System and method for printing a three-dimensional object incorporating an image |
CN108460730B (en) * | 2017-02-17 | 2020-06-26 | 京东方科技集团股份有限公司 | Image processing method and device |
CN107999750B (en) * | 2017-11-22 | 2018-09-04 | 江苏久祥汽车电器集团有限公司 | Automobile current generator precision die manufacturing method |
CN109774152B (en) * | 2018-12-18 | 2020-03-17 | 西安交通大学 | Incremental compression storage method for continuous slice images based on three-dimensional model |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5774577A (en) * | 1994-07-15 | 1998-06-30 | Nec Corporation | Compact image conversion system operable at a high speed |
US20050156825A1 (en) * | 2003-12-31 | 2005-07-21 | Lg Electronics Inc. | Method of displaying gray scale in plasma display panel |
US8289318B1 (en) * | 2008-08-29 | 2012-10-16 | Adobe Systems Incorporated | Determining three-dimensional shape characteristics in a two-dimensional image |
US20130328228A1 (en) * | 2012-06-08 | 2013-12-12 | Makerbot Industries, Llc | Color three dimensional printing |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7589868B2 (en) | 2002-12-11 | 2009-09-15 | Agfa Graphics Nv | Method and apparatus for creating 3D-prints and a 3-D printing system |
US20100023155A1 (en) | 2004-10-26 | 2010-01-28 | 2089275 Ontario Ltd. | Method for the automated production of three-dimensional objects and textured substrates from two-dimensional or three-dimensional objects |
KR101123648B1 (en) | 2011-02-21 | 2012-03-20 | 유백현 | Printing method for expressing solid texture of oil painting |
US8579620B2 (en) | 2011-03-02 | 2013-11-12 | Andy Wu | Single-action three-dimensional model printing methods |
US9248623B2 (en) | 2011-10-14 | 2016-02-02 | Makerbot Industries, Llc | Grayscale rendering in 3D printing |
US9669585B2 (en) | 2014-02-11 | 2017-06-06 | Adobe Systems Incorporated | Method and apparatus for embedding a 2-dimensional image in a 3-dimensional model |
-
2014
- 2014-02-11 US US14/177,933 patent/US9669585B2/en active Active
-
2017
- 2017-04-26 US US15/498,324 patent/US10373394B2/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5774577A (en) * | 1994-07-15 | 1998-06-30 | Nec Corporation | Compact image conversion system operable at a high speed |
US20050156825A1 (en) * | 2003-12-31 | 2005-07-21 | Lg Electronics Inc. | Method of displaying gray scale in plasma display panel |
US8289318B1 (en) * | 2008-08-29 | 2012-10-16 | Adobe Systems Incorporated | Determining three-dimensional shape characteristics in a two-dimensional image |
US20130328228A1 (en) * | 2012-06-08 | 2013-12-12 | Makerbot Industries, Llc | Color three dimensional printing |
Non-Patent Citations (1)
Title |
---|
Lee et al. (âShape from Shading with a Linear Triangular Element Surface Modelâ, IEEE Xplore, 1993). * |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10373394B2 (en) | 2014-02-11 | 2019-08-06 | Adobe Inc. | Method and apparatus for embedding a 2-dimensional image in a 3-dimensional model |
US20150317412A1 (en) * | 2014-05-05 | 2015-11-05 | Microsoft Corporation | Fabricating three-dimensional objects with embossing |
US9734264B2 (en) * | 2014-05-05 | 2017-08-15 | Microsoft Technology Licensing, Llc | Fabricating three-dimensional objects with embossing |
US10338568B2 (en) * | 2015-04-24 | 2019-07-02 | Hewlett-Packard Development Company, L.P. | Determining halftone schemes |
CN105599296A (en) * | 2015-12-23 | 2016-05-25 | 上海理工大学 | 2.5D image printing method |
US10346100B2 (en) * | 2016-01-29 | 2019-07-09 | Hewlett-Packard Development Company, L.P. | Error diffusion |
US10409264B2 (en) * | 2016-10-27 | 2019-09-10 | Voodoo Manufacturing, Inc. | Fabrication of three-dimensional part with two-dimensional image applied thereon |
US20180144219A1 (en) * | 2016-11-23 | 2018-05-24 | Simbionix Ltd. | Method and system for three-dimensional print oriented image segmentation |
US10885407B2 (en) * | 2016-11-23 | 2021-01-05 | Simbionix Ltd. | Method and system for three-dimensional print oriented image segmentation |
US11334777B2 (en) * | 2016-11-23 | 2022-05-17 | 3D Systems Inc. | Method and system for three-dimensional print oriented image segmentation |
CN112848281A (en) * | 2020-12-31 | 2021-05-28 | 浙江闪铸三维科技有限公司 | Light compensation method for photocuring 3D printer |
CN115723333A (en) * | 2021-08-30 | 2023-03-03 | 广州黑格智造信息科技有限公司 | 3D printing method, system, device and storage medium |
Also Published As
Publication number | Publication date |
---|---|
US9669585B2 (en) | 2017-06-06 |
US20170225399A1 (en) | 2017-08-10 |
US10373394B2 (en) | 2019-08-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10373394B2 (en) | Method and apparatus for embedding a 2-dimensional image in a 3-dimensional model | |
KR102107709B1 (en) | Spatial transformer modules | |
Cheng et al. | Robust algorithm for exemplar-based image inpainting | |
CN102693527B (en) | Method and apparatus for performing a blur rendering process on an image | |
US8687908B2 (en) | Sample based tone mapping method for high dynamic range images | |
CN111192190B (en) | Method and device for eliminating image watermark and electronic equipment | |
JP2011166217A (en) | Image processing device and image processing program | |
ATE508435T1 (en) | METHOD FOR INTERPOLATION AND SHARPENING OF IMAGES | |
US11641438B2 (en) | Systems and methods for rendering a secure document with a text selection capability | |
Martín et al. | Scale-dependent and example-based grayscale stippling | |
US9741142B2 (en) | Method and apparatus for enabling text editing in a scanned document while maintaining fidelity of the appearance of the text | |
US20130182943A1 (en) | Systems and methods for depth map generation | |
US20170337657A1 (en) | Merging filters for a graphic processing unit | |
JP2017062553A (en) | Three-dimensional model forming device and three-dimensional model forming method | |
TW200943222A (en) | Multi-format support for surface creation in a graphics processing system | |
JP4510743B2 (en) | Tally image generating method and apparatus, tally image generating program, and secret image decoding method | |
Akyüz et al. | Style-based tone mapping for hdr images | |
Pasewaldt et al. | BeCasso: artistic image processing and editing on mobile devices | |
US10423868B2 (en) | Embedding a removable barcode into an image | |
US20170249775A1 (en) | Adaptive depth-guided non-photorealistic rendering method, corresponding computer program product, computer-readable carrier medium and device | |
CN114387315A (en) | Image processing model training method, image processing device, image processing equipment and image processing medium | |
JP6168872B2 (en) | Image processing apparatus, image processing method, and program | |
Siogkas | Visual media processing using matlab beginner's guide | |
US20140354627A1 (en) | Rendering a 3d shape | |
Sadun | IOS Drawing |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: ADOBE SYSTEMS INCORPORATED, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HEMANI, MAYUR;KUMAR, ABHISHEK;GOEL, NAVEEN PRAKASH;REEL/FRAME:032271/0456 Effective date: 20131119 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: ADOBE INC., CALIFORNIA Free format text: CHANGE OF NAME;ASSIGNOR:ADOBE SYSTEMS INCORPORATED;REEL/FRAME:048867/0882 Effective date: 20181008 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |